| 1. Operator and Work Instructions |
|
| 1.1 |
Is there a revision controlled Test Instruction which contains unique details for the specific product being tested? (Score 0 if any unsigned/undated handwritten instructions or any handwritten instructions more than 48 hrs old) |
| 1.2 |
Are Work Instructions readily available to the test operator and are they followed? |
| 1.3 |
Is the Fixture ID specified on Work Instructions? |
| 1.4 |
Is the Fixture ID traceable to a specific PCBA part number and revision level and to the Unit Under Test? |
| 1.5 |
Is the Batch File specified on Work Instructions? |
| 1.6 |
Is the Batch File traceable to a specific PCBA part number and revision level and to the Unit Under Test? |
| 1.7 |
Is the Vacuum setting and range specified on Work Instructions? |
| 1.8 |
Is the PCB orientation to the Fixture identified in the Work Instructions or on the Fixture? |
| 1.9 |
Does the test Operator have the Standard Operating Procedure (SOP) for the tester available to them at all times? |
| 1.10 |
Is there evidence that the Operator has been trained and certified against the Standard Operating Procedure for the Tester? |
| 1.11 |
Does the Operator know the content of the Standard Operating Procedure for the Tester and do they and follow it? |
| 1.12 |
Are Operators required to log in at the Test station and does this provide an automatic verification of training status? |
| 2. ICT Fixture |
|
| 2.1 |
Is the ICT fixture identified with a name or number? |
| 2.2 |
Is the Preventative Maintenance / Calibration sticker on the ICT fixture current and up to date? |
| 2.3 |
Is there a Preventative Maintenance procedure and schedule for ICT fixtures? |
| 2.4 |
Is there evidence to demonstrate that Preventative Maintenance records are up-to-date? |
| 2.5 |
Are spare test fixture parts (excluding probes) stocked? |
| 2.6 |
Are spare test fixture probes stocked for each design required to support Dell fixtures? |
| 2.7 |
Is the inventory of spare test fixture parts adequately controlled? |
| 2.8 |
Are ICT fixtures adequately stored in such a way that the tester interface is protected? |
| 3. ICT System Hardware |
|
| 3.1 |
Is there a vacuum gauge on the line connected to the test system? |
| 3.2 |
Is a calibrated vacuum gauge visible to the test operator and is it at the correct setting within an acceptable range? |
| 3.3 |
Is the Preventative Maintenance / Calibration sticker on the Test System current and up to date? |
| 3.4 |
Is there a Preventative Maintenance procedure and schedule for ICT Test Systems? |
| 3.5 |
Are the test systems pin electronics verified daily using the internal test? |
| 3.6 |
Is there evidence to demonstrate that Preventative Maintenance records are up-to-date? |
| 3.7 |
Are spare test system parts stocked? |
| 3.8 |
Are spare test system pin electronics boards stocked? |
| 3.9 |
Are the inventory of spare test system parts adequately controlled? |
| 4. ICT Software |
|
| 4.1 |
Are all ICT test systems networked to a server? |
| 4.2 |
Is the ICT software downloaded from a central server when the program is called up or at least compiled once per day? |
| 4.3 |
Is the ICT software revision controlled for program changes? |
| 4.4 |
Are all changes to an ICT program, no matter how insignificant the change is considered, logged in the program or otherwise? |
| 4.5 |
Is it impossible for unapproved ICT s/w changes to remain on the test system longer than 24 hours before recompiling the program? |
| 4.6 |
When changes are made, is there evidence that change details are sent to Dell for approval? |
| 5. Test Operation |
|
| 5.1 |
Is there an automated method of loading test programs (i.e. the batch file). |
| 5.2 |
Is the Fixture ID used to select and automatically load the correct ICT program for the unit under test? |
| 5.3 |
If the same ICT fixture is used for different PCBA part numbers, does the Batch file automatically differentiate? NA may be used. |
| 5.4 |
Is there a foolproof method to ensure that product A will not pass ICT if Batch File B is used? NA may be used. |
| 5.5 |
Is there a foolproof method to ensure that product B will not pass ICT if Batch File A is used? NA may be used. |
| 5.6 |
Is the PCBA orientation to the fixture identified or is required to be checked before board loading? |
| 5.7 |
Is the test program uniquely identified on the test system display after the test program has been loaded? |
| 5.8 |
Is there a documented and agreed convention outlining the storage of untested, failed, and pass boards? |
| 5.9 |
Are boards marked in some way to facilitate the implementation of this convention? (Forced Routing is acceptable.) |
| 5.10 |
Are boards awaiting test identified and stored separately according to the convention OR routed via a dedicated conveyor? |
| 5.11 |
Are passing boards identified and stored separately according to the convention OR routed via a dedicated conveyor? |
| 5.12 |
Are failing boards identified and stored separately according to the convention OR routed via a dedicated conveyor? |
| 5.13 |
Are pass and failed boards stored on different storage carts OR routed via a unique conveyor? |
| 5.14 |
Is SPC data collected and effectively used at this process point? |
| 5.15 |
Is the content of the SPC data chart up-to-date? |
| 5.16 |
Are out of control SPC data points effectively actioned? |
| 5.17 |
Are ICT buffer trigger points established to ensure process shut down should the limits be exceeded? |
| 5.18 |
Has a Gauge R&R study been completed in accordance with l GR&R Procedures? |
| 5.19 |
Has test coverage been calculated using the l Metrics for ICT coverage, and is this readily available and known? |
| 6. ICT WIP Tracking |
|
| 6.1 |
Is a Forced Board Routing system and WIP Tacking system fully deployed throughout the test process? |
| 6.2 |
Does the system verify the ‘last’ step processed and compare it to the expected ‘last’ step? |
| 6.3 |
Do boards that PASS have an identifying mark to indicate their pass status? (Forced Routing is acceptable.) |
| 6.4 |
Do boards that FAIL have an identifying mark to indicate their fail status? (Forced Routing is acceptable.) |
| 6.5 |
Do failed boards have ICT fail listings attached for debug purposes? (Paperless repair is acceptable) |
| 6.6 |
Do all debugged boards have an identifying mark to indicate a debug status? (Forced Routing is acceptable.) |
| 6.7 |
Is there a software link between ICT Test results & the Forced Routing/Quality Data System? |
| 6.8 |
Does this software link eliminate manual intervention to indicate the test result status? |
| 6.9 |
Is this link fully automated and used to log Test Yield/First Pass Yield & Board Yield data? |
| 6.10 |
Is FPY and BY readily known and has it been calculated in accordance with Dell definitions and specified retry conditions? |
| 6.11 |
For boards tested in panel format, is the failed board identified before the panel is removed from the fixture? |
| 7. Debug and FA Capability |
|
| 7.1 |
Are boards awaiting debug identified and stored to one side of the operator or routed via dedicated conveyor? |
| 7.2 |
Are debug buffer trigger points established to ensure process shut down should the limits be exceeded? |
| 7.3 |
Can it be demonstrated that a technician qualification or formal training is required for debug and failure analysis activity? |
| 7.4 |
Can it be demonstrated that all debug personnel meet the above requirements ? |
| 7.5 |
Is there adequate debug equipment available at each debug station ? |
| 7.6 |
Is ICT debug and repair conducted real-time (on-line) whenever possible? |
| 7.7 |
Can it be demonstrated that a tool is used to capture fail codes, their fixes, and then recommend the most common fix? |
| 7.8 |
Does the ICT debug technician access the ICT system via a display unit to better understand to root cause of failure? |
| 7.9 |
Is a test point location map available to the repair/failure analysis operator? |
| 7.10 |
Are there detailed instructions on determining false failures? |
| 7.11 |
Are false failure rates tracked, monitored, and documented? |
| 7.12 |
Are there goals for false failure reduction? (I.e. goals to increase FPY) |