From: owner-stds-1450-4@majordomo.ieee.org on behalf of Gordon Robinson [Gordon_Robinson@3mts.com]
Sent: Thursday, July 25, 2002 7:10 AM
To: 'don.organ@inovys.com'; Gordon Robinson; 'Micek Tom-ra1370'; 'STIL. 4'; David Dowding (E-mail)
Subject: RE: stds-1450.4: RE: 1450.4 multisite status
The Credence Kalos flash memory tester is an example of
the "complete tester per site" approach. It's also the only tester
good looking enough to get a spread in "Wired" magazine! I've
heard that some Agilent flash memory testers also have that architecture.
I've heard from flash memory gurus that flash testing is not sensible
without a completely separate digital sequencer etc per site because
flash test patterns have an extremely high density of conditional, response-dependent
branching (which STIL is not exactly strong at handling).
 
There is also some trend I've observed to architectures with a sequencer and
pins per board that can be lockstepped together, to make a system have
a serious number of separate digital instruments if needed.
 
I'm fine with adding "optional pieces to guide multisite handling". I've
tried to express "idealism" principles. Multisite gets into trouble when the
idealism breaks down.
 
Should these be UserKeywords? I can easily imagine that each tester will need
its own workarounds.
 
I believe strongly that disabling/enabling is not a test step and should be "invisible"
to the device except for time passing.
 
I'd also like to mention something I don't want us to copy, from Credence's
Agile system. Agile has "conditional" stages, where the flow branches. These
are not available in multisite programs. Instead, multisite flow is handled by
using the "execution flags" that are kept for each site. These flags were originally
intended to allow the operator to set one of a few options, and particular tests would
be obeyed or skippied based on those flag settings. The flags then started to get set/unset
by the test stages themselves, and used as a flow mechanism. That was the only way allowed
for multisite conditional flows. This does have the advantage of placing all of the "fail-priority"
or "pass-priority" in the user's hands because the sequence of stages is used, and if one branch is
placed first in sequence, it will happen first.
 
Gordon
 
-----Original Message-----
From: Don Organ [mailto:don.organ@inovys.com]
Sent: Wednesday, July 24, 2002 10:26 PM
To: Gordon Robinson; 'Micek Tom-ra1370'; 'STIL. 4'; don.organ@inovys.com; David Dowding (E-mail)
Subject: RE: stds-1450.4: RE: 1450.4 multisite status

Wow! Great!
Comments are inline below.
 
-DVO-
-----Original Message-----
From: Gordon Robinson [mailto:Gordon_Robinson@3mts.com]
Sent: Wednesday, July 24, 2002 3:32 PM
To: 'Micek Tom-ra1370'; 'STIL. 4'; 'don.organ@inovys.com'; David Dowding (E-mail)
Subject: RE: stds-1450.4: RE: 1450.4 multisite status

Let me comment on some of the multisite issues, with the
danger that this may add to people's confusion with a difficult topic
before it resolves things.
 
I try to take a couple of extremely strong stances about multisite:
I regard it as OK for a device to experience different time delays "between steps" when being tested multisite.
Different stimulus of any form gets me concerned.
I think we're in trouble when a device has to go through new power-up sequences or be reinitialized whenever it gets enabled after being disabled.
[DVO] I'm not a test-engineer, but different devices have different requirements. I'm told that some must keep a clock going (or else they burn up!) and that some of these "keep-alive" patterns require support from the pattern-sequencer, and many testers can't maintain this "keep-alive" pattern on some sites, while executing the AC tests on other sites. So, the alternative chosen is to power-down these devices for the periods in which they are disabled. Perhaps such devices aren't great candidates for multisite testing.
 
 
There are quite a few general approaches to how multisite testing is performed.
[DVO] We probably want a programming model that makes few assumptions and could be implemented with any of the above approaches. The 2nd bullet is the one I'm most familiar with - and it is the one the raised in issue #2 (disabling/re-enabling) of the working list.
 
The hard issues of multisite occur when the test program has to deal with issues that challenge its model.
[DVO] Bullets #1, #3, and maybe #4 are issues that I think are owned by the ATE vendor and don't require much consideration from us. It is the ATE vendor's responsibility to execute STIL.4 on their tester and if their tester has limitiations such as those listed, then they should be responsible for resolve them. I don't think STIL.4 should need any syntax to address those issues.
 I don't view datalogs as a a simple stream, but as a separate stream for each site. How thats treated visually is a UI design issue.
 
I hope I've stirred up a few things!
[DVO] From reading your email, I think the following are issues we should consider adding to the working list (remembering that the list is not of things we must address, but is of things we should consider addressing):
  1. Syntax for the loadboard connections. 1450 doesn't have a construct for defining the mapping from Signals to tester resources - but I think this is in the 1450.3 charter. Should we consider adding a multisite capable mapping? or at least participate in 1450.3 enough to see that they do it in a way we like?
  2. prematch and postmatch for match mode.
  3. Isolation assumptions/requirements. This is touched on in working list issue #2 (disabling/re-enabling), but maybe this warrants further consideration.
Gordon