Southern Machinery provides professional SMT peripheral equipment.

Аудит процесса качества электронного производства печатных плат — испытание в цепи

2023-02-09 17:51:23
1. Оператор и рабочие инструкции
1.1 Существует ли инструкция по тестированию с контролируемой редакцией, которая содержит уникальные детали для конкретного тестируемого продукта? (Оценка 0, если есть какие-либо неподписанные/недатированные рукописные инструкции или любые рукописные инструкции старше 48 часов)
1,2 Доступны ли рабочие инструкции оператору-испытателю и соблюдаются ли они?
1,3 Указан ли идентификатор прибора в рабочих инструкциях?
1,4 Is the Fixture ID traceable to a specific PCBA part number and revision level and to the Unit Under Test?
1.5 Is the Batch File specified on Work Instructions?
1.6 Is the Batch File traceable to a specific PCBA part number and revision level and to the Unit Under Test?
1.7 Is the Vacuum setting and range specified on Work Instructions?
1.8 Is the PCB orientation to the Fixture identified in the Work Instructions or on the Fixture?
1.9 Does the test Operator have the Standard Operating Procedure (SOP) for the tester available to them at all times?
1.10 Is there evidence that the Operator has been trained and certified against the Standard Operating Procedure for the Tester?
1.11 Does the Operator know the content of the Standard Operating Procedure for the Tester and do they and follow it?
1.12 Are Operators required to log in at the Test station and does this provide an automatic verification of training status?
2. ICT Fixture
2.1 Is the ICT fixture identified with a name or number?
2.2 Is the Preventative Maintenance / Calibration sticker on the ICT fixture current and up to date?
2.3 Is there a Preventative Maintenance procedure and schedule for ICT fixtures?
2.4 Is there evidence to demonstrate that Preventative Maintenance records are up-to-date?
2.5 Are spare test fixture parts (excluding probes) stocked?
2.6 Are spare test fixture probes stocked for each design required to support Dell fixtures?
2.7 Is the inventory of spare test fixture parts adequately controlled?
2.8 Are ICT fixtures adequately stored in such a way that the tester interface is protected?
3. ICT System Hardware
3.1 Is there a vacuum gauge on the line connected to the test system?
3.2 Is a calibrated vacuum gauge visible to the test operator and is it at the correct setting within an acceptable range?
3.3 Is the Preventative Maintenance / Calibration sticker on the Test System current and up to date?
3.4 Is there a Preventative Maintenance procedure and schedule for ICT Test Systems?
3.5 Are the test systems pin electronics verified daily using the internal test?
3.6 Is there evidence to demonstrate that Preventative Maintenance records are up-to-date?
3.7 Are spare test system parts stocked?
3.8 Are spare test system pin electronics boards stocked?
3.9 Are the inventory of spare test system parts adequately controlled?
4. ICT Software
4.1 Are all ICT test systems networked to a server?
4.2 Is the ICT software downloaded from a central server when the program is called up or at least compiled once per day?
4.3 Is the ICT software revision controlled for program changes?
4.4 Are all changes to an ICT program, no matter how insignificant the change is considered, logged in the program or otherwise?
4.5 Is it impossible for unapproved ICT s/w changes to remain on the test system longer than 24 hours before recompiling the program?
4.6 When changes are made, is there evidence that change details are sent to Dell for approval?
5. Test Operation
5.1 Is there an automated method of loading test programs (i.e. the batch file).
5.2 Is the Fixture ID used to select and automatically load the correct ICT program for the unit under test?
5.3 If the same ICT fixture is used for different PCBA part numbers, does the Batch file automatically differentiate? NA may be used.
5.4 Is there a foolproof method to ensure that product A will not pass ICT if Batch File B is used? NA may be used.
5.5 Is there a foolproof method to ensure that product B will not pass ICT if Batch File A is used? NA may be used.
5.6 Is the PCBA orientation to the fixture identified or is required to be checked before board loading?
5.7 Is the test program uniquely identified on the test system display after the test program has been loaded?
5.8 Is there a documented and agreed convention outlining the storage of untested, failed, and pass boards?
5.9 Are boards marked in some way to facilitate the implementation of this convention? (Forced Routing is acceptable.)
5.10 Are boards awaiting test identified and stored separately according to the convention OR routed via a dedicated conveyor?
5.11 Are passing boards identified and stored separately according to the convention OR routed via a dedicated conveyor?
5.12 Are failing boards identified and stored separately according to the convention OR routed via a dedicated conveyor?
5.13 Are pass and failed boards stored on different storage carts OR routed via a unique conveyor?
5.14 Is SPC data collected and effectively used at this process point?
5.15 Is the content of the SPC data chart up-to-date?
5.16 Are out of control SPC data points effectively actioned?
5.17 Are ICT buffer trigger points established to ensure process shut down should the limits be exceeded?
5.18 Has a Gauge R&R study been completed in accordance with l GR&R Procedures?
5.19 Has test coverage been calculated using the l Metrics for ICT coverage, and is this readily available and known?
6. ICT WIP Tracking
6.1 Is a Forced Board Routing system and WIP Tacking system fully deployed throughout the test process?
6.2 Does the system verify the ‘last’ step processed and compare it to the expected ‘last’ step?
6.3 Do boards that PASS have an identifying mark to indicate their pass status? (Forced Routing is acceptable.)
6.4 Do boards that FAIL have an identifying mark to indicate their fail status? (Forced Routing is acceptable.)
6.5 Do failed boards have ICT fail listings attached for debug purposes? (Paperless repair is acceptable)
6.6 Do all debugged boards have an identifying mark to indicate a debug status? (Forced Routing is acceptable.)
6.7 Is there a software link between ICT Test results & the Forced Routing/Quality Data System?
6.8 Does this software link eliminate manual intervention to indicate the test result status?
6.9 Is this link fully automated and used to log Test Yield/First Pass Yield & Board Yield data?
6.10 Is FPY and BY readily known and has it been calculated in accordance with Dell definitions and specified retry conditions?
6.11 For boards tested in panel format, is the failed board identified before the panel is removed from the fixture?
7. Debug and FA Capability
7.1 Are boards awaiting debug identified and stored to one side of the operator or routed via dedicated conveyor?
7.2 Are debug buffer trigger points established to ensure process shut down should the limits be exceeded?
7.3 Can it be demonstrated that a technician qualification or formal training is required for debug and failure analysis activity?
7.4 Can it be demonstrated that all debug personnel meet the above requirements ?
7.5 Is there adequate debug equipment available at each debug station ?
7.6 Осуществляется ли отладка и ремонт ИКТ в режиме реального времени (онлайн), когда это возможно?
7.7 Можно ли продемонстрировать, что инструмент используется для сбора кодов ошибок, их исправлений, а затем рекомендовать наиболее распространенное исправление?
7,8 Получает ли специалист по отладке ИКТ доступ к системе ИКТ через дисплей, чтобы лучше понять основную причину отказа?
7,9 Доступна ли карта расположения контрольных точек оператору по анализу ремонта/отказов?
7.10 Есть ли подробная инструкция по определению ложных отказов?
7.11 Отслеживаются ли, контролируются и документируются ли показатели ложных отказов?
7.12 Существуют ли цели по уменьшению числа ложных отказов? (т.е. цели по увеличению FPY)

Chat with us