127
Nevada Environmental Restoration Project Industrial Sites Quality Assurance Project Plan Controlled Copy No.: Revision No.: 3 February 2002 Approved for public release; further dissemination is unlimited. U.S. Department of Energy National Nuclear Security Administration Nevada Operations Office Environmental Restoration Division DOE/NV--372--REV. 3

Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

NevadaEnvironmentalRestorationProject

Industrial Sites Quality Assurance Project Plan

Controlled Copy No.: Revision No.: 3

February 2002

Approved for public release; further dissemination is unlimited.

U.S. Department of EnergyNational Nuclear Security Administration

Nevada Operations Office

Environmental Restoration Division

DOE/NV--372--REV. 3

Page 2: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

DOE/NV--372--REV. 3

INDUSTRIAL SITESQUALITY ASSURANCE PROJECT PLAN

U.S. Department of EnergyNational Nuclear Security Administration

Nevada Operations OfficeLas Vegas, Nevada

Controlled Copy No.:

Revision No.: 3

February 2002

Approved for public release; further dissemination is unlimited.

Page 3: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Print ed on r ecycled p aper

Available for public sale, in paper, from:

U.S. Department of CommerceNational Technical Information Service5285 Port Royal RoadSpringfield, VA 22161Phone: 800.553.6847Fax: 703.605.6900Email: [email protected] ordering: http://www.ntis.gov/ordering.htm

Available electronically at http://www.doe.gov/bridge

Available for a processing fee to U.S. Department of Energy and its contractors, in paper,from:

U.S. Department of EnergyOffice of Scientific and Technical InformationP.O. Box 62Oak Ridge, TN 37831-0062Phone: 865.576.8401Fax: 865.576.5728Email: [email protected]

Reference herein to any specific commercial product, process, or service by trade name,trademark, manufacturer, or otherwise, does not necessarily constitute or imply itsendorsement, recommendation, or favoring by the United States Government or anyagency thereof or its contractors or subcontractors.

Page 4: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

INDUSTRIAL SITESQUALITY ASSURANCE PROJECT PLAN

Approved by: Date: Runore C. Wycoff, Division DirectorEnvironmental Restoration Division

Approved by: Date: Janet Appenzeller-Wing, Project Manager Industrial Sites Project

Page 5: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page i of xxiii

Table of Contents

List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi

List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi

List of Acronyms and Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii

Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix

Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix

Project Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx

QAPP Organization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxiii

1.0 Criteria 1 - Program . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.1 Quality Management Policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 Project Organization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.3 NNSA/NV ERD Director . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.3.1 NV ERP Project Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.3.1.1 NV ERP Task Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

1.3.1.2 NV ERP Quality Assurance Coordinator . . . . . . . . . . . . . . . . . . . . 4

1.3.2 Industrial Sites Project Participants . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4

1.3.3 Analytical Laboratories . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.4 Planning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.4.1 Graded Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.4.2 Task Initiation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

1.4.3 Data Quality Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

1.4.3.1 Data Quality Objectives Process . . . . . . . . . . . . . . . . . . . . . . . . . . . 8

1.4.3.2 Data Quality Objectives Reconciliation . . . . . . . . . . . . . . . . . . . . . 9

1.5 Quality Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

1.5.1 Principal Data Quality Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

1.5.1.1 Precision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

1.5.1.2 Accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

1.5.1.3 Representativeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

Page 6: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Table of Contents (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page ii of xxiii

1.5.1.4 Completeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

1.5.1.5 Comparability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

1.5.2 Secondary Data Quality Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

1.5.2.1 Sensitivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

1.5.2.2 Recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.5.2.3 Memory Effects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.5.2.4 Limit of Quantitation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.5.2.5 Repeatability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.5.2.6 Reproducibility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15

1.6 Measurement Quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

1.6.1 Quantitative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

1.6.2 Semiquantitative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

1.6.3 Qualitative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

1.7 Reports to Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16

1.8 Readiness Reviews . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17

1.9 Stop Work Order . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18

2.0 Criteria 2 - Personnel Training and Qualifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.1 Project Personnel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.2 Subcontractor Personnel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

2.3 Analytical Laboratory Personnel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19

3.0 Criteria 3 - Quality Improvement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

3.1 Internal Quality Control Checks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

3.1.1 Field Quality Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

3.1.1.1 Equipment Rinsate Blank Samples . . . . . . . . . . . . . . . . . . . . . . . . 22

3.1.1.2 Field Blank Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22

3.1.1.3 Trip Blank Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

3.1.1.4 Duplicate Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

3.1.2 Laboratory Quality Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

3.1.2.1 Laboratory Control Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

3.1.2.2 Method Blank Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

3.1.2.3 Surrogate-Spike Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

3.1.2.4 Matrix Spike/Matrix Spike Duplicate Samples . . . . . . . . . . . . . . . 25

3.1.2.5 Laboratory Duplicate Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

Page 7: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Table of Contents (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page iii of xxiii

3.2 Data Precision, Accuracy, and Completeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

3.3 Corrective Action . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

3.3.1 Nonconformance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

3.3.2 Root Cause . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26

3.3.3 Trend Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

3.3.4 Lessons Learned . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

4.0 Criteria 4 - Documents and Records . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

4.1 Documents and Records . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

4.1.1 Document Review and Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

4.1.2 Change Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

4.1.3 Records Maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29

5.0 Criteria 5 - Work Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

5.1 Evaluation and Use of Existing and New Data . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

5.2 Computer Hardware and Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

5.2.1 Computer Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

5.2.2 Software Design/Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

5.2.2.1 Code Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

5.2.2.2 Code Verification/Validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

5.2.2.3 Software Documentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32

5.2.3 Peer Review of Software and Code Applications . . . . . . . . . . . . . . . . . . . . 33

5.3 Field Investigation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

5.3.1 Procedures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

5.3.1.1 Sampling Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33

5.3.1.2 Sample Custody . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

5.3.1.3 Chain of Custody Form . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34

5.3.1.4 Custody Seals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

5.3.1.5 Sample Labels and Identification . . . . . . . . . . . . . . . . . . . . . . . . . 35

5.3.1.6 Sample Handling, Preservation, Packaging, and Shipping . . . . . . . 36

5.3.1.7 Decontamination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36

5.3.2 Field Documentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

5.3.3 Investigation-Derived Waste . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37

5.3.4 Photographic Documentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

5.3.5 Identification and Control of Items . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

Page 8: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Table of Contents (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page iv of xxiii

5.3.6 Calibration and Preventive Maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . 38

5.3.6.1 Calibration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38

5.3.6.2 Preventive Maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

5.3.7 Laboratory Operation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

5.3.7.1 Preanalysis Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39

5.3.7.2 Post-Analysis Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

5.4 Analytical Data Usability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

5.4.1 Data Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

5.4.2 Evaluation and Use of Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

5.4.3 Data Reduction, Verification, and Validation . . . . . . . . . . . . . . . . . . . . . . . 41

5.4.3.1 Data Completeness Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

5.4.3.2 Data Review and Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

5.4.3.3 Data Validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

5.4.4 Laboratory Data Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44

5.4.4.1 Data Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45

5.5 Data Quality Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45

6.0 Criteria 6 - Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46

7.0 Criteria 7 - Procurement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

7.1 Procurement Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

7.1.1 Procurement Documents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47

7.1.2 Measurement and Testing Equipment . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

7.1.3 Verification of Quality Conformance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

7.2 Procurement of Laboratory Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

8.0 Criteria 8 - Inspection and Acceptance Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51

9.0 Criteria 9 - Management Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52

10.0 Criteria 10 - Independent Assessments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53

11.0 References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54

Page 9: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Table of Contents (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page v of xxiii

Appendix A - Federal Facility Agreement and Consent Order Outlines

Appendix B - Analytical Table

Page 10: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page vi of xxiii

List of Figures

Number Title Page

I-1 Hierarchy of Documents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxi

I-2 Corrective Action Process Flowpath . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxii

1-1 Organizational Chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2

List of Tables

Number Title Page

1-1 Assessment of Data Quality Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11

B.1-1 General Chemical Analytical Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B-1

B.1-2 General Radiological Analytical Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . B-5

B.1-3 Waste Characterization Chemical Requirements . . . . . . . . . . . . . . . . . . . . . . . . . . B-6

B.1-4 Waste Characterization Radiological Requirements . . . . . . . . . . . . . . . . . . . . . . . B-9

Page 11: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page vii of xxiii

List of Acronyms and Abbreviations

ANSI/ASQC American National Standards Institute/American Society for Quality Control

CADD Corrective Action Decision Document

CAIP Corrective Action Investigation Plan

CAP Corrective Action Plan

CAU Corrective Action Unit

CFR Code of Federal Regulations

CLP Contract Laboratory Program

COC Chain-of-Custody

CR Closure Report

DOE U.S. Department of Energy

DQA Data quality analysis

DQO Data Quality Objective(s)

DQI Data Quality Indicators

EPA U.S. Environmental Protection Agency

ERD Environmental Restoration Division

FFACO Federal Facility Agreement and Consent Order

GC Gas chromatography

GC/MS Gas chromatography/mass spectrometry

H&S Health and Safety

HPLC High-performance liquid chromatography

IDW Investigation-derived waste

LCS Laboratory control sample

LQC Laboratory quality control

M&TE Measurement and test equipment

NAC Nevada Administrative Code

NCR Nonconformance report(s)

NDEP Nevada Division of Environmental Protection

NNSA/NV National Nuclear Security Administration Nevada Operations Office

NTSWAC Nevada Test Site Waste Acceptance Criteria

NV ERP Nevada Environmental Restoration Project

PARCC Precision, accuracy, representativeness, comparability, completeness

QA Quality assurance

QAC Quality Assurance Coordinator

QAPP Quality Assurance Project Plan

Page 12: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

List of Acronyms and Abbreviations (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page viii of xxiii

QC Quality control

RPD Relative percent difference

SAFER Streamlined Approach for Environmental Restoration

SOP Standard Operating Procedure

SOW Scope of Work

SWO Stop Work Order

VOA Volatile organic analysis

% R Percent recovery

Page 13: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page ix of xxiii

Definitions

Acceptance Criteria

Specific characteristics of an item, process, or service defined in codes, standards, or other

requirement documents. (DOE/NV Family Quality Glossary, 1993)

Accuracy

A measure of the closeness of an individual measurement or the average of a number of

measurements to the true value. Accuracy includes a combination of random error (precision)

and systematic error (bias) components that are due to sampling and analytical operations; the

U.S. Environmental Protection Agency (EPA) recommends using the terms “precision” and

“bias,” rather then “accuracy,” to convey the information usually associated with accuracy.

(EPA QA/G-5, 1998)

Activity

An all-inclusive term describing a specific set of operations or related tasks to be performed,

either serially or in parallel (e.g., research and development, field sampling, analytical operations,

equipment fabrication), that in total result in a product or service. (ANSI/ASQC E4-1994)

Assessment

The evaluation process used to measure the performance or effectiveness of a system and its

elements. In this Standard, assessment is an all-inclusive term used to denote any of the

following: audit, performance evaluation, management systems review, peer review, inspection,

or surveillance. (ANSI/ASQC E4-1994)

Audit (Quality)

A systematic and independent examination to determine whether quality activities and related

results comply with planned arrangements and whether these arrangements are implemented

effectively and are suitable to achieve objectives. (ANSI/ASQC E4-1994)

Bias

The systematic or persistent distortion of a measurement process which causes errors in one

direction (i.e., the expected sample measurement is different from the sample’s true value).

(ANSI/ASQC E4-1994)

Page 14: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page x of xxiii

Calibration

Comparison of a measurement standard, instrument, or item with a standard or instrument of

higher accuracy to detect and quantify inaccuracies and to report or eliminate those inaccuracies

by adjustments. (ANSI/ASQC E4-1994)

Certification

The act of determining, verifying, and attesting in writing to the qualifications of personnel,

processes, procedures, or items in accordance with acceptance criteria. (DOE/NV Family Quality

Glossary, 1993)

Characteristic

Any property or attribute of a datum, item, process, or service that is distinct, describable, and/or

measurable. (ANSI/ASQC E4-1994)

Comparability

A measure of the confidence with which one data set can be compared to another.

(ANSI/ASQC E4-1994)

Completeness

A measure of the amount of valid data obtained from a measurement system compared to the

amount that was expected to be obtained under correct, normal conditions.

(ANSI/ASQC E4-1994)

Condition Adverse to Quality

An all-inclusive term used in reference to any of the following: failures, malfunctions,

deficiencies, defective items or nonconformance. (DOE/NV Family Quality Glossary, 1993)

Corrective Action

An action taken to eliminate the causes of an existing nonconformance, deficiency, or other

undesirable situation in order to prevent recurrence. (ANSI/ASQC E4-1994)

Criteria

Rules or tests against which the quality of performance can be measured. They are most effective

when expressed quantitatively. Fundamental criteria are contained in policies and

Page 15: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xi of xxiii

objectives, as well as codes, standards, regulations, and recognized professional practices that

DOE and DOE contractors are required to observe. (DOE/NV Family Quality Glossary, 1993)

Data Quality Objectives (DQOs)

Qualitative and quantitative statements derived from the DQO process that clarify study technical

and quality objectives, define the appropriate types of data, and specify tolerable levels of

potential decision errors that will be used as the basis for establishing the quality and quantity of

data needed to support decisions. (ANSI/ASQC E4-1994)

Data Quality Objectives Process

A systematic strategic planning tool based on the scientific method that identifies and defines the

type, quality, and quantity of data needed to satisfy a specific use. The key elements of the

process include:

� Concisely defining the problem� Identifying the decision to be made� Identifying the key inputs to the decision� Defining the boundaries of the study� Developing the decision rule� Specifying tolerable limits on potential decision errors � Selecting the most resource efficient data collection design

Data quality objectives are the qualitative and quantitative outputs from the DQO process. The

DQO process was developed originally by the EPA, but has been adapted for use by other

organizations to meet their specific planning requirements. (ANSI/ASQC E4-1994)

Data Usability

The process of ensuring or determining whether the quality of the data produced meets the

intended use of the data. (ANSI/ASQC E4-1994)

Deficiency

An unauthorized deviation from acceptable procedures or practices, or a defect in an item.

(ANSI/ASQC E4-1994)

Page 16: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xii of xxiii

Design

Specifications, drawings, design criteria, and performance requirements. Also the result of

deliberate planning, analysis, mathematical manipulations, and design processes.

(ANSI/ASQC E4-1994)

Document

Any written or pictorial information describing, defining, specifying, reporting, or certifying

activities, requirements, procedures, or results. (ANSI/ASQC E4-1994)

Environmental Data

Any measurements or information that describe environmental processes or conditions, or the

performance of environmental technology. (ANSI/ASQC E4-1994)

Environmental Data Operations

Work performed to obtain, use, or report information pertaining to environmental processes and

conditions. (ANSI/ASQC E4-1994)

Graded Approach

The process of basing the level of application of managerial controls applied to an item or work

according to the intended use of the results and the degree of confidence needed in the quality of

the results. (See data quality objectives process.) (ANSI/ASQC E4-1994)

Independent Assessment

An assessment performed by a qualified individual, group, or organization that is not a part of the

organization directly performing and accountable for the work being assessed.

(ANSI/ASQC E4-1994)

Inspection

An activity such as measuring, examining, testing, or gauging one or more characteristics of an

entity and comparing the results with specified requirements in order to establish whether

conformance is achieved for each characteristic. (ANSI/ASQC E4-1994)

Page 17: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xiii of xxiii

Item

An all-inclusive term used in place of any of the following: appurtenance, facility, sample,

assembly, component, equipment, material, module, part, product, structure, subassembly,

subsystem, system, unit, documented concepts, or data. (ANSI/ASQC E4-1994)

Management Assessment

The determination of the appropriateness, thoroughness, and effectiveness of management

processes. (DOE/NV Family Quality Glossary, 1993)

Measurement and Testing Equipment (M&TE)

Tools, gauges, instruments, sampling devices or systems used to calibrate, measure, test, or

inspect in order to control or acquire data to verify conformance to specified requirements.

(ANSI/ASQC E4-1994)

Method

A body of procedures and techniques for performing an activity (e.g., sampling, chemical analysis,

quantification) systematically presented in the order in which they are to be executed.

(ANSI/ASQC E4-1994)

Nonconformance

A deficiency in characteristic, documentation, or procedure that renders the quality of an item or

activity unacceptable or indeterminate; nonfulfillment of a specified requirement.

(ANSI/ASQC E4-1994)

Precision

A measure of mutual agreement among individual measurements of the same property, usually

under prescribed similar conditions, expressed generally in terms of the standard deviations.

(ANSI/ASQC E4-1994)

Procedure

A specified way to perform an activity. (ANSI/ASQC E4-1994)

Page 18: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xiv of xxiii

Process

Any activity or group of activities that takes an input, adds value to it, and provides an output to a

customer. The logical organization or people, materials, energy, equipment, and procedures into

work activities designed to produce a specified end result (work product). (DOE/NV Family

Quality Glossary, 1993)

Quality

The totality of features and characteristics of a product or service that bear on its ability to meet

the stated or implied needs and expectations of the user. (ANSI/ASQC E4-1994)

Quality Assurance (QA)

An integrated system of management activities involving planning, implementation assessment,

reporting, and quality improvement to ensure that a process, item, or service is of the type and

quality needed and expected by the customer. (ANSI/ASQC E4-1994)

Quality Assurance Program

The overall program (management system) established to assign responsibilities and

authorities, define policies and requirements for the performance and assessment of work. (DOE

Order 414.1A, 2001a)

Quality Control

The overall system of technical activities that measures the attributes and performance of a

process, item, or service against defined standards to verify that they meet the stated requirements

established by the customer; operational techniques and activities that are used to fulfill

requirements for quality. (ANSI/ASQC E4-1994)

Quality Improvement

A management program for improving the quality of operations. Such management programs

generally entail a formal mechanism for encouraging work recommendations with timely

management evaluation and feedback or implementation. (ANSI/ASQC E4-1994)

Page 19: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xv of xxiii

Quality Indicators

Measurable attributes of the attainment of the necessary quality for a particular environmental

decision. Indicators of quality include precision, bias, completeness, representativeness,

reproducibility, comparability, and statistical confidence. (ANSI/ASQC E4-1994)

Quality Management Plan (QMP)

A formal document or manual, usually prepared once for an organization, that describes the

quality system in terms of the organizational structure, functional responsibilities of management

and staff, lines of authority, and required interfaces for those planning, implementing, and

assessing all activities conducted. (ANSI/ASQC E4-1994)

Quality System

A structured and documented management system describing the policies, objectives, principles,

organizational authority, responsibilities, accountability, and implementation plan of an

organization for ensuring quality in its work processes, products (items), and services. The

quality system provides the framework for planning, implementing, and assessing work performed

by the organization and for carrying out required QA and QC. (ANSI/ASQC E4-1994)

Readiness Review

A systematic, documented review of the readiness for startup or continued use of a facility,

process, or activity. Readiness reviews are typically conducted before proceeding beyond project

milestones and prior to institution of a major phase of work. (ANSI/ASQC E4-1994)

Record

A completed document that furnishes evidence relating to items or activities. (DOE/NV Family

Quality Glossary, 1993)

Remediation

The process of reducing the concentration of a contaminant (or contaminants) in air, water, or soil

media to a level that poses an acceptable risk to human health. (ANSI/ASQC E4-1994)

Page 20: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xvi of xxiii

Representativeness

A measure of the degree to which data accurately and precisely represent a characteristic of a

population, parameter variations at a sampling point, a process condition, or an environmental

condition. (ANSI/ASQC E4-1994)

Risk

A quantitative or qualitative expression of possible loss which considers both the probability that

an event occurrence will cause harm or loss and the consequences of that event. (DOE/NV

Family Quality Glossary, 1993)

Root Cause

The most basic reason for conditions adverse to quality that, if corrected, will prevent occurrence

or recurrence. (DOE/NV Family Quality Glossary, 1993)

Self Assessment

Assessments of work conducted by individuals, groups, or organizations directly responsible for

overseeing and/or performing the work. (ANSI/ASQC E4-1994)

Service

The result generated by activities at the interface between the supplier and the customer, and by

supplier internal activities to meet customer needs. Such activities in environmental programs

include design, inspection, laboratory and/or field analysis, repair, and installation. (ANSI/ASQC

E4-1994)

Specification

A document stating requirements and which refers to or includes drawings or other relevant

documents. Specifications should indicate the means and the criteria for determining

conformance. (ANSI/ASQC E4-1994)

Standard Operating Procedure

A written document that details the method for an operation, analysis, or action with thoroughly

prescribed techniques and steps, and that is officially approved as the method for performing

certain routine or repetitive tasks. (ANSI/ASQC E4-1994)

Page 21: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Definitions (Continued)

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xvii of xxiii

Surveillance (Quality)

Continual or frequent monitoring and verification of the status of an entity and the analysis of

records to ensure that specified requirements are being fulfilled. (ANSI/ASQC E4-1994)

Technical Review

A documented critical review of work that has been performed within the state of the art. The

review is accomplished by one or more qualified reviewers who are independent of those who

performed the work, but are collectively equivalent in technical expertise to those who performed

the original work. The review is an in-depth analysis and evaluation of documents, activities,

material, data, or items that require technical verification or validation for applicability,

correctness, adequacy, completeness, and assurance that established requirements are satisfied.

(ANSI/ASQC E4-1994)

Traceability

The ability to trace the history, application, or location of an entity by means of recorded

identifications. In a calibration sense, traceability relates measuring equipment to national or

international standard, primary standards, basic physical constants or properties, or reference

materials. In a data collection sense, it relates calculations and data generated throughout the

project back to the requirements for quality for the project. (ANSI/ASQC E4-1994)

Training

The process of providing for and making available to an employee(s) and placing or enrolling an

employee(s) in a planned, prepared, and coordinated program, course, curriculum, subject,

system, or routine of instruction or education, in fiscal, administrative, management, individual

development, or other fields which improve individual and organizational performance and assist

in achieving the agency’s mission and performance goals. (DOE/NV Family Quality Glossary,

1993).

Validation

Confirmation by examination and provision of objective evidence that the particular requirements

for a specific intended use are fulfilled. In design and development, validation concerns the

process of examining a product or result to determine conformance to user needs. (ANSI/ASQC

E4-1994)

Page 22: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xviii of xxiii

Verification

Confirmation by examination and provision of objective evidence that specified requirements have

been fulfilled. In design and development, validation concerns the process of examining a result

of a given activity to determine conformance to the stated requirements for that activity.

(ANSI/ASQC E4-1994)

Page 23: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xix of xxiii

Introduction

This Quality Assurance Project Plan (QAPP) is one of the planning documents used for the

Industrial Sites Project which falls under the oversight of the U.S. Department of Energy (DOE),

National Nuclear Security Administration Nevada Operations Office (NNSA/NV) Environmental

Restoration Project (NV ERP). The NV ERP conducts environmental investigation and

remediation activities at sites under the control of the NNSA/NV. It is the policy of the NV ERP

to conduct all environmental restoration activities in a manner that produces data of a known

quality. Safety shall be integrated into management and work practices at all levels so that

missions are accomplished while protecting the public, the worker, and the environment.

This QAPP describes policies, organization, responsibilities, and objectives of the Industrial Sites

Project and is intended to provide a consistent framework for the collection, evaluation, analysis,

and use of data. The general, common activities of the project are addressed herein. The

information provided is not site- or time-specific but applies throughout the project. This QAPP

must be supplemented with documents that provide site-specific information. In addition, this

QAPP provides for the evaluation of risks associated with the activities to be performed and uses

the graded approach to determine the required level of quality assurance. This document

supplements, and is to be used in conjunction with, the site-specific plans. Site-specific planning

documents must contain quality assurance (QA)and quality control (QC) requirements

appropriate for the site and activities being performed. In the event that project objectives or

regulatory jurisdiction change, this document must be reevaluated for adequacy.

This QAPP meets the requirements provided in DOE Order 414.1A, Quality Assurance

(DOE, 2001a). In addition, the American National Standards, Specifications and Guidelines for

Quality Systems for Environmental Data Collection and Environmental Technology Programs,

ANSI/ASQC E4-1994 (ASQC, 1994) and the U.S. Environmental Protection Agency (EPA)

Requirements for Quality Assurance Plans, EPA/R-5 (EPA, 1999) were used in the development

of this document to ensure consistency with EPA requirements. Environmental Restoration

Project activities shall also be in compliance with DOE Order 440.1A, Worker Protection

Management for DOE Federal and Contractor Employees (DOE, 1998) and DOE Order 450.4,

Safety Management System Policy (DOE, 1996b). Work at hazardous waste sites shall be

conducted in accordance with the applicable sections of Title 29 Code of Federal Regulations

(CFR) Part 1910.120, Hazardous Waste Operations and Emergency Response (CFR, 2001d),

Page 24: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xx of xxiii

and the Nevada Administrative Code (NAC) 444.850-8746 (NAC, 2000) for the disposal of

hazardous waste. Radioactive waste shall be handled and disposed of in accordance with

10 CFR Part 71, Subpart H, Packaging and Transportation of Radioactive Materials - Quality

Assurance (CFR, 2001a); The Nevada Test Site Waste Acceptance Criteria (NTSWAC)

(DOE/NV, 2000b); and DOE Order 435.1, Radioactive Waste Management (DOE, 2001b).

Nuclear activities, as defined by 10 CFR 820.2, Procedural Rules for DOE Nuclear Activities

(CFR, 2001a), shall be subject to the requirements of those sections of 10 CFR 830.120,

Quality Assurance (CFR, 2001c) that apply to the activity being performed.

Corrective action sites, grouped in corrective action units (CAUs), located in the State of Nevada

will be characterized and/or closed under the Federal Facility Agreement and Consent Order

(FFACO, 1996). Characterization of sites outside the state of Nevada will be conducted in

accordance with the applicable state regulations and requirements. This QAPP provides the

structure of the quality program for the project and is used in the development of lower-tier

documents. These documents provide the necessary quality requirements on a site-specific basis.

Figure I-1 delineates the hierarchy of documents for NV ERP activities.

Project Description

Nuclear tests and their associated support activities were conducted at the Nevada Test Site

(NTS), Tonopah Test Range (TTR), and other locations under the control of the NNSA/NV.

The Industrial Sites Project identifies, evaluates, investigates, and remediates corrective action

sites that have been potentially impacted by these activities. Currently, CAUs may be

characterized and closed using one of three corrective action processes: Housekeeping,

Streamlined Approach for Environmental Restoration (SAFER), and the Complex. Figure I-2

depicts these three processes.

The CAUs which may be closed through the housekeeping process are distinguished from other

Industrial Site CAUs because they do not require further investigation prior to final disposition.

Data gathered during record searches and field verification activities sanction the removal of

source materials and directly impacted soil, and subsequent confirmatory sampling without

additional investigation.

The SAFER concept recognizes that technical decisions can be made by experienced project

personnel faced with some procedural uncertainty. This process will be employed where the

parties agree that enough information exists about the nature and extent of contamination to

Page 25: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xxi of xxiii

FFACO

DOE

Orders

NNSA/NV

Orders

ERP Project Plans

ERD

Policy/Plans

Contractor Documents, SOPs,

Instructions, or Work Packages, HASP, SSHASP

Federal

Regulations

State

Regulations

Key

CADD = Corrective Action Decision DocumentCAIP = Corrective Action Investigation PlanCAP = Corrective Action PlanCR = Closure ReportDOE = U.S. Department of EnergyERD = Environmental Restoration DivisionFFACO = Federal Facility Agreement and Consent OrderHASP = Health and Safety PlanNNSA = National Nuclear Security AdministrationNTS = Nevada Test SiteNV = NevadaQAPP = Quality Assurance Project PlanSAFER = Streamlined Approach for Environmental RestorationSOPs = Standard Operating ProceduresSSHASP = Site-Specific Health and Safety Plan

Key

CADD = Corrective Action Decision DocumentCAIP = Corrective Action Investigation PlanCAP = Corrective Action PlanCR = Closure ReportDOE = U.S. Department of EnergyERD = Environmental Restoration DivisionFFACO = Federal Facility Agreement and Consent OrderHASP = Health and Safety PlanNNSA = National Nuclear Security AdministrationNTS = Nevada Test SiteNV = NevadaQAPP = Quality Assurance Project PlanSAFER = Streamlined Approach for Environmental RestorationSOPs = Standard Operating ProceduresSSHASP = Site-Specific Health and Safety Plan

CAIPCADDSAFER

CADD/CRCR

CAPCADD/CAP

IS QAPP

Figure I-1Hierarchy of Documents

Page 26: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xxii of xxiii

Identify CASsIdentify CASs

Group CASs into CAUs

Group CASs into CAUs

WhichCorrective

Action ProcessApplies?

Prioritize CAUsPrioritize CAUs

ImplementCorrective Action

ImplementCorrective Action

Develop CAIPDevelop CAIP

Implement CAIImplement CAI

Develop CADDDevelop CADD

Develop CAPDevelop CAP

Develop ClosureReport

Develop ClosureReport

NDEP IssuedNotice of Completion

NDEP IssuedNotice of Completion

DevelopSAFER

Plan

DevelopSAFER

Plan

ImplementSAFERProcessActivities

ImplementSAFERProcessActivities

UnexpectedComplications?

ImplementCorrective

Action

ImplementCorrective

Action

UnexpectedComplications?

DevelopOptional Work

Plan

DevelopOptional Work

Plan

Housekeeping

Yes

SAFER

Complex

No

Requires NDEP approval

No

Yes

Figure I-2Corrective Action Process Flowpath

Page 27: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: IntroductionRevision: 3Date: 02/07/2002Page xxiii of xxiii

propose an appropriate corrective action prior to the completion of a Corrective Action

Investigation. The SAFER process combines elements of the Data Quality Objectives (DQO)

process and the observational approach to help plan and conduct corrective actions. The DQO

process will be used to define the type and quantity of data needed to complete the SAFER

process. The observational approach will provide a framework for managing uncertainty and

planning decision making.

The Complex process will be used for those CAUs where additional information is needed to

evaluate possible corrective action alternatives. Corrective action alternatives include removing

the existing wastes until a predetermined standard is achieved immobilizing the wastes without

removal or taking no action. A Corrective Action Alternative or a combination of Corrective

Action Alternatives will be selected based on an evaluation of the risk scenarios.

QAPP Organization

The organization of this plan reflects the criteria of DOE Order 414.1A, Quality Assurance

(DOE, 2001a). The ten criteria of this plan cover three major areas: management, performance,

and assessments. Management entails the planning and preparation required for the successful

completion of the project mission. Additionally, this section incorporates quality improvement

processes to enable personnel to detect and prevent quality problems. The performance section

establishes the requirements and procedures to be implemented to ensure that newly collected

environmental data are valid, that uses of existing data are appropriate, and that methods of

environmental modeling are reliable. Assessments provide a feedback loop to project

management whereby the feedback information can be used to evaluate and, if necessary, modify

a system or process to ensure the quality of the product.

Page 28: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 1 of 55

1.0 Criteria 1 - Program

Industrial Sites Project management systems encompass the planning, preparation, and feedback,

necessary to ensure the successful completion of identified objectives. This QAPP has been

prepared to provide the planning and control necessary for effective and efficient work processes.

This document provides the overall QA Program requirements and the general quality practices to

be applied to Industrial Sites activities. Policy is established, roles and responsibilities are defined,

lines of communication are identified, the needs and objectives of the project are confirmed, and

reviews are conducted to ensure (to the extent possible) that all necessary planning and

preparation activities have taken place. Low-level radioactive and mixed waste managed under

the NV ERP must also meet the requirements of the applicable waste acceptance criteria and the

associated waste certification program plan. The following sections describe the quality

management systems to be employed for the effective management of the Industrial Sites Project.

1.1 Quality Management Policy

It is the policy of the NV ERP to provide environmental management that incorporates applicable

regulatory requirements. The Quality Management Program described in this document should be

implemented for all Industrial Sites environmental activities to ensure that work is performed in an

efficient, controlled manner and appropriately documented. Project requirements should be

applied on a graded approach, commensurate with the risk of failure of the items or processes and

the potential harm those risks pose for human health and the environment. Activities shall

conform with applicable federal, state, and local regulations, and contract requirements. Quality

will be part of the normal course of work and incorporated from the earliest planning stages to

completion of the work.

1.2 Project Organization

The NNSA/NV Environmental Restoration Division (ERD) is responsible for the administration

of the NV ERP. The NV ERP is a major project under the DOE Office of Environmental

Management, Southwestern Area Programs. Activities in support of the NV ERP are grouped

into projects based on the media affected or type of contamination. Personnel from the ERD are

assigned project management and technical support responsibilities. All NV ERP Project

Managers are responsible for achieving quality within the specific projects they manage. The

NNSA/NV ERD organization chart is provided in Figure 1-1. Roles and responsibilities for

Page 29: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 2 of 55

Project Integration TeamQuality Assurance Coordinator Assistant Manager for Environmental Management

Environmental Restoration Division Director

UGTA ProjectProject ManagerTask Manager(s)

Industrial Sites ProjectProject ManagerTask Manager(s)

Soils ProjectProject ManagerTask Manager(s)

Offsites ProjectProject ManagerTask Manager(s)

Project Integration TeamQuality Assurance Coordinator Assistant Manager for Environmental Management

Environmental Restoration Division Director

UGTA ProjectProject ManagerTask Manager(s)

Industrial Sites ProjectProject ManagerTask Manager(s)

Soils ProjectProject ManagerTask Manager(s)

Offsites ProjectProject ManagerTask Manager(s)

Project Integration TeamQuality Assurance Coordinator Assistant Manager for Environmental Management

Environmental Restoration Division Director

UGTA ProjectProject ManagerTask Manager(s)

Industrial Sites ProjectProject ManagerTask Manager(s)

Soils ProjectProject ManagerTask Manager(s)

Offsites ProjectProject ManagerTask Manager(s)

UGTA ProjectProject ManagerTask Manager(s)

Industrial Sites ProjectProject ManagerTask Manager(s)

Soils ProjectProject ManagerTask Manager(s)

Offsites ProjectProject ManagerTask Manager(s)

Figure 1-1Organizational Chart

Page 30: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 3 of 55

NV ERP personnel and supporting contractors and organizations (referred to as project

participants) are described in the following sections.

1.3 NNSA/NV ERD Director

The NNSA/NV ERD Director has oversight and management responsibilities for all projects

under the NV ERP and is responsible for the scope and implementation of the QA Program

defined in this document. The Director is the senior management official responsible for ensuring

that this QAPP is established, quality requirements are implemented, and opportunities for

improvement are identified and incorporated.

1.3.1 NV ERP Project Manager

The NV ERP Project Managers report directly to and are the prime point of contact with the

NNSA/NV ERD Director. The NV ERP Project Manager has day-to-day management

responsibilities for technical, financial, and scheduling aspects of his/her assigned project and shall

monitor contractor performance of project activities. At a minimum, the NV Project Manager is

responsible for the following duties:

� Review, approve, and direct the implementation of NV ERP project-specific plans.

� Disseminate pertinent information from NNSA/NV to NV ERP participants.

� Review and approve changes to NV ERP project-specific documents.

� Monitor the activities of participating organizations and provide direction and guidancefor improvement.

� Verify project participants are adequately executing the responsibilities as delineated inSection 1.3.2.

� Notify and apprise the NNSA/NV ERD Director of significant conditions adverse toquality.

1.3.1.1 NV ERP Task Manager

The NV ERP Task Managers report directly to their respective NV ERP Project Managers. The

Task Managers have day-to-day management responsibilities for technical and scheduling aspects

of the assigned project task and shall monitor contractor performance of task activities. At a

minimum, the Task Managers are responsible for the following duties:

� Ensure effective communication among contractors performing work for their assignedtasks.

Page 31: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 4 of 55

� Participate in the organization and planning of activities.

� Perform periodic assessments (such as surveillances) of activities under their purview.

� Monitor the activities of participating organizations and provide direction and guidancefor improvement.

� Notify the NNSA/NV ERD Director, responsible NV ERP Project Manager, NV ERPQuality Assurance Coordinator (QAC), and other involved personnel, of significantconditions adverse to quality.

1.3.1.2 NV ERP Quality Assurance Coordinator

The NV ERP QAC has a direct line of communication with the NNSA/NV ERD Director and the

NV ERP Project Managers. The NV ERP QAC will provide the overall direction of the QA

function. At a minimum, the NV ERP QAC shall have the following duties:

� Identify and respond to QA/QC needs of the NV ERP and provide QA/QC guidance orassistance to individual Project and Task Managers.

� Verify that systems are in place to evaluate data against analytical quality criteria.

� Verify that appropriate corrective actions are taken for nonconforming conditions.

� Notify the NNSA/NV ERD Director, individual NV ERP Project Managers, and otherinvolved personnel, of significant conditions adverse to quality or any adverse trends.

1.3.2 Industrial Sites Project Participants

Project participants, such as supporting contractors and organizations, are responsible for

developing the necessary procedures for their assigned scope of work and ensuring that work is

performed in accordance with applicable federal, state, and local regulations; and approved

NV ERP project plans and procedures consistent with individual contracts and agency

agreements. To fulfill responsibilities specific to QA, participants shall, at a minimum, be

responsible for the following:

� Report information on scope, schedule, cost, technical execution, and quality achievementof task order activities to the NV ERP Project Managers or NV ERP Task Managers.

� Ensure the proper resources are provided for QA activities and that QA activities areintegrated into project activities.

� Evaluate activities to ensure that planning document requirements are implemented.

Page 32: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 5 of 55

� Implement applicable procedures and instructions that govern NV ERP activities.

� Verify that work is technically sound, of acceptable quality, and consistent with projectobjectives.

� Ensure personnel are trained and qualified to achieve initial proficiency, maintainproficiency, and adapt to changes in technology, methods, or job responsibilities.

� Perform assessments to verify compliance with applicable requirements.

� Identify deficient areas and implement effective corrective action for quality problems.

� Notify the NV ERP Project Managers, NV ERP Task Managers, and other involvedpersonnel, of significant conditions adverse to quality or any adverse trends.

� Verify that appropriate corrective actions are taken for nonconformances.

1.3.3 Analytical Laboratories

Analytical laboratories used to support the NV ERP are responsible for ensuring that samples are

received, handled, stored, and analyzed according to the analytical laboratory’s QA program and

contract requirements. Analytical laboratories performing data analysis shall participate in a

Performance Evaluation Sample Programs appropriate for analyses performed and be subject to

periodic audits. Subcontracted analytical services are subject to the same requirements.

Verification of subcontractor conformance is the responsibility of the contracting organization.

1.4 Planning

The NV ERP and participant personnel responsible for oversight of data collection operations

should verify that the data-collection system design is defined, controlled, verified, and

documented. All planning shall incorporate the principle of Integrated Safety Management to

mitigate hazards to workers. A graded approach to data quality requirements shall be used to

meet the sampling objectives and data needs of a given site and the dynamic nature of the

program. Work assignments should be clearly communicated with lines of communication

established among all participants. Organizations assigned lead responsibilities shall coordinate

project planning with decision makers and participating organizations.

1.4.1 Graded Approach

The graded approach is defined as, “The process of basing the level of application of managerial

controls applied to an item or work according to the intended use of the results and the degree of

confidence needed in the quality of the results.” (ASQC, 1994). To achieve maximum

Page 33: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 6 of 55

effectiveness and satisfy customer expectations, it is essential that the degree of quality applied be

appropriate to the type of activity and intended use of the results being provided.

This QAPP provides for the evaluation of risks associated with the activities to be performed and

uses the graded approach to determine the required level of quality assurance. Environmental

program activities can range from the simple to diverse or complex. As a result, the level of

QA/QC applied to each project must be flexible enough to accommodate each activity. For

example, the level of QA/QC applied to samples collected to verify clean closure is much more

rigorous than data collected for a preliminary assessment that is meant to provide information for

planning purposes. When determining the level of rigor required, the following issues should be

taken into consideration:

• The environmental decision to be made

• The impact on human health and the environment

• The regulatory requirements for the site-specific environmental problem

The DQO process for each CAU must distinguish between measurements required to achieve

project objectives (or determine limits on decision errors) and data collected for informational or

background purposes (such as information used to guide field investigations). All measurements

should be classified as to whether they are required to achieve project objectives or for

informational purposes. Critical measurements will undergo closer scrutiny during the data

gathering and review processes. Participants in the DQO process shall determine the required

data quality for each type of measurement to ensure the integrity of information used to make the

required decisions. This data quality defined for each measurement shall be stated in the DQO

documents. If data quality is not assigned during the DQO process, then the rational for not

making this determination shall be documented.

1.4.2 Task Initiation

A project kickoff meeting should be conducted at the beginning of each task. This meeting

should brief key personnel assigned to the task on the purpose, expected outcome, schedule, and

personnel responsibilities for completion of the effort. The planning process should be monitored

by responsible managers to ensure communication of status, assess progress, and implement any

corrective action needed to achieve timely completion.

Page 34: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 7 of 55

1.4.3 Data Quality Objectives

When appropriate, planning and scoping for environmental data/information needs will include the

use of the DQO process to develop a scientific and resource-effective data collection design. This

is a systematic planning process for defining the criteria that a data collection design should satisfy

in order to provide the information necessary to make environmental decisions. The DQO

process entails detailed up-front planning that:

� Identifies and clarifies the critical decisions necessary to achieve project objectives (thosedecisions necessary to reach an end point).

� Identifies all data needs (e.g., type, quantity, and quality).

� Ensures the efficient and timely completion of requirements at minimal cost.

The most current version of U.S. Environmental Protection Agency (EPA) QA/G-4, Guidance

for the Data Quality Objectives Process (EPA, 2000b), will be used to develop DQOs and the

results of the process documented. Participants in the DQO process for each site-specific project

should include representatives of all data users and decision makers involved with that project.

The appropriate NNSA/NV ERD personnel, NV ERP participants, and state regulators jointly

establish DQOs for each site, or group of similar sites, to allow the work to be planned in a

manner that will ensure data will meet the needs of the end users. During the DQO process,

representatives from the involved organizations will:

� Clarify the study objective

� Develop a conceptual site model

� Identify critical decisions

� Identify the data need

� Determine how the data will be used

� Determine what the data represent

� Define the most appropriate type of data to collect

� Determine the most appropriate conditions from which to collect the data

� Specify tolerable limits on decision errors which will be used as the basis for establishingthe quantity and quality of data needed to support the decision

Page 35: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 8 of 55

Tables B.1-1, B.1-2, B.1-3, and B.1-4 (see Appendix B of this document) represent the default

analytical requirements. The DQO process may determine alternative requirements.

1.4.3.1 Data Quality Objectives Process

The DQO Process is a seven-step iterative planning approach that defines the purpose of the data

collection effort, clarifies what the data should represent to satisfy this purpose, and specifies the

performance requirements for the quality of information to be obtained from the data. The

following sections briefly describe each of the seven steps:

Step 1: State the Problem

This step defines the environmental problem(s). Process knowledge, combined with

professional judgment, is used to develop a conceptual site model that describes the contaminant

source, method of release, potential contaminant migration pathways, and transport mechanisms

that would drive migration.

Step 2: Identify the Decision

This step identifies what questions the study will attempt to resolve and what actions may result.

This usually involves a number of decisions and pathways to end points. Often, these pathways

are organized into a decision logic flow diagram.

Step 3: Identify the Inputs to the Decision

This step identifies the information needed to resolve each decision statement. This part of the

process identifies the type, quantity, and quality of each information source. The basis for

determining the potential contaminants, action levels for each contaminant, and identification of

sampling and analysis methods that can meet the data requirements are delineated. These data

requirements define the metrics that will need to be met in order to justify that decisions are

valid.

Step 4: Define the Study Boundaries

In this step we define the target population to be sampled and the population (or spatial extent)

to which the decision will apply. Any temporal and spatial constraints that are pertinent for

decision making are identified, along with conditions that would warrant suspension of work and

a reassessment of the conceptual site model if constraints are exceeded.

Page 36: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 9 of 55

Step 5: Develop a Decision Rule

Under Step 5 the logical basis for choosing among alternative actions is described. This is

generally an "If..., then..." statement that defines the choice of actions for the decision maker and

describes the conditions under which each possible alternative action would be chosen.

Step 6: Specify Tolerable Limits on Decision Errors

The decision makers’ tolerable decision error rates are identified, based on a consideration of the

consequences of making an incorrect decision. A decision error rate is the probability of making

an incorrect decision based on data that inaccurately estimate the true state of nature. Decision

errors are always possible unless 100 percent of the population is measured. A conservative

baseline condition (i.e., null hypothesis) is assumed that must be proved to be wrong. The two

possible decision errors are false rejection and false acceptance of the baseline condition.

For example, the false rejection of this baseline condition would mean accepting that

contaminants of potential concern are not present above PALs when they really are, increasing

risk to human health and environment. The false acceptance error would mean accepting that

contaminants of concern are above PALs when they are not, resulting in increased costs for

unneeded characterization.

Step 7: Optimize the Design

The purpose of Step 7 is to develop a resource-effective sampling and analysis design for

generating data that are expected to satisfy the DQOs developed in Steps 1 through 6 of the

DQO process. The optimum number of samples to be collected, sampling locations, and

sampling methods are determined along with the rationale behind the sampling and analysis

design. This step provides the data needs metrics that will be used to evaluate the validity of the

decisions.

1.4.3.2 Data Quality Objectives Reconciliation

An evaluation, subsequent to data collection and validation, to assess whether the collected data

met the performance criteria specified in the DQO process must be performed as the final phase of

the project. This evaluation should determine whether the results obtained from the project or

task can be reconciled with the requirements defined by the data user or decision maker.

Section 5.5, “Data Quality Assessment,” addresses reconciliation with DQOs.

1.5 Quality Indicators

Data Quality Indicators (DQIs) are qualitative and quantitative descriptors used in interpreting the

degree of acceptability or utility of data. The principal DQIs are precision, accuracy,

Page 37: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 10 of 55

representativeness, comparability, and completeness. Secondary DQIs include sensitivity,

recovery, memory effects, limit of quantitation, repeatability, and reproducibility.

Sampling and analytical data goals are determined during the DQO process and are based on the

intended use of the data, current field procedures, instrumentation, and available resources.

Quality indicator goals should be established during the site-specific DQO process to properly

support the overall project or sampling task objectives (see Table 1-1). If DQIs fail to meet

project objectives, then the impact to the decision shall be assessed and documented.

Appendix B of this document specifies default values for precision and accuracy. Those values

that either are not identified, or different from the table, must be defined during the DQO process.

An evaluation of the DQIs shall be performed during the assessment of data to determine if the

goals set during the DQO process have been accomplished. An EPA document, currently being

drafted on DQIs (EPA QA/G-5i), was consulted to ensure consistency with anticipated EPA

guidance.

1.5.1 Principal Data Quality Indicators

The principal DQIs are precision, accuracy, representativeness, comparability, and completeness.

The five DQIs are also referred as the PARCC parameters. Principal DQIs are discussed in detail

in the following sections and are identified on Tables B.1-1 and B.1-2 (see Appendix B of this

document).

1.5.1.1 Precision

Precision measures the reproducibility of data under a given set of conditions. Specifically,

precision is a quantitative measurement of the variability of a population of measurements

compared to their average value. Precision shall be assessed by collecting, preparing, and

analyzing duplicate field samples and by creating, preparing, and analyzing laboratory duplicates

from one or more field samples. Precision will be reported as relative percent difference (RPD).

The RPD is calculated as the difference between the measured concentrations of Sample 1 and

Sample 2, divided by the average of the two concentrations, and multiplied by 100.

Precision goals are method, matrix, analyte, and laboratory-specific. The tables in Appendix B

present the most common parameters that are analyzed during Industrial Sites projects, and the

associated precision goals. In general, the desired precision goals should be within the range of

RPD �30 percent; however, some precision goals are outside this range as the tables in

Appendix B demonstrate. Specific precision goals listed in the tables represent the values for the

Page 38: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 11 of 55

Table 1-1Assessment of Data Quality Indicators

(Page 1 of 2)

Parameter Minimum Requirementsa Potential Impacts of Failure to

Meet Objectives Corrective Action(s)

PrecisionLaboratory

Field

Laboratory duplicates and/or matrixspike/matrix spike duplicate peranalytical batch on required samples; variation between duplicates shallnot exceed analytical method-specific criteria agreed to during theDQO process; if precision is notmethod-specific, 20 to 25 RPD fromthe average value

Variations between fieldduplicates should not exceed20 percent relative deviationfrom the average value

b

Increased variability in data cancause loss of completenesswhich diminishesrepresentativeness so thatlaboratory QA objectives cannotbe satisfied

Unacceptable levels ofuncertainty

Erroneous final decision(s)

Identify source of problem, thenreanalyze, resample, reviewlaboratory protocols, or use adifferent analytical technique, asnecessary; assess and documentimpacts to decisions

Identify source of problem, thenresample or add more samples, asnecessary; reevaluate performanceobjectives; assess and documentimpacts to decisions

AccuracyLaboratory

Laboratory control sample resultsand matrix spikes within qualitycontrol criteria as specified in theanalytical method

Laboratory method blanks belowrequired detection limit

Decrease in reliability can causeloss of completeness whichdiminishes representativenessso that laboratory QA objectivescannot be satisfied

Identify source of problem, thenreanalyze or use differentanalytical approach, as necessary; assess and document impacts todecisions

RepresentativenessLaboratory

Field

Sample preparation requirementsthat will not introduce a bias

Samples must be representative ofthe potentially contaminated medium

Inaccurate identification orestimate of concentration ofcontaminant; insufficient data tomake decision(s)

Generation of false positive ornegative data; data biased eitherhigh or low for contaminantconcentration

Identify source of problem, thenreanalyze, resample, or documentwhich site areas are poorlycharacterized, as necessary;assess and document impacts todecisions

Redesign sample collectionmethods, as necessary. Additionalsampling may be required. Assess and document impacts todecisions

CompletenessLaboratory

Field

80 percent of measurements

Sufficient data to satisfy data needsfor critical decisions

Incomplete content and site characterization

Lack of complete content andsite characterization

Determine whether the missingdata are needed; can remainingsample volumes be used or dothey need to be resampled? Assess and document impacts todecisions

Identify the source of problem;determine how the missing dataimpact the investigation andwhether they are needed; resampleif necessary or provide additionalanalysis of samples at laboratory tocomplete data set for lost samples;assess and document impacts todecisions

Page 39: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Table 1-1Assessment of Data Quality Indicators

(Page 2 of 2)

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 12 of 55

Parameter Minimum Requirementsa Potential Impacts of Failure to

Meet Objectives Corrective Action(s)

ComparabilityLaboratory

Field

Equivalent samples analyzed as perthe Sampling design; same analyticalmethods, the same units ofmeasurement and detection limitsmust be used for like analyses

All samples are collected as per thesampling design developed in theDQO process; approved proceduresthat are standard industry protocolshall be used to ensure comparabilityof sample collection

Increase in overall error

Reduction in ability to measureagainst required standards

Implement comparable analyticaltechniques or change laboratories; assess and document impacts todecisions

Identify source of problem, thenredesign sample collectionmethods and/or add more samplesto increase confidence asnecessary; assess and documentimpacts to decisions

aAdditional Information is presented in the text.

bMay not be attainable for soil samples and analyte concentrations that are near the detection limit.

methods given in such nationally recognized compendia as SW-846 and the Contract Laboratory

Program (CLP) Scopes of Work (SOW). Project-specific precision goals should be established

during the DQO process.

The RPD shall be within the limits defined in site-specific plans. Values exceeding the

acceptance criteria, established during the site-specific DQO process, must be evaluated for

usability.

1.5.1.2 Accuracy

Analytical accuracy is defined as the nearness of a measurement to the true or accepted reference

value. It is the composite of the random and systematic components of the measurement system

and measures bias in a measurement system. Accuracy measurements for spike samples and

laboratory control samples shall be calculated as percent recovery (%R), which is calculated by

dividing the measured sample concentration by the true concentration and multiplying the quotient

by 100.

Analytical accuracy goals are method, matrix, analyte, and laboratory-specific. The tables in

Appendix B present the most common parameters that are analyzed during the Industrial Sites

projects, and the accuracy goals associated with those parameters. In general, desired analytical

accuracy goals should be within the range of %R = 70-130 percent. However, some accuracy

Page 40: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 13 of 55

goals are outside this range as the tables in Appendix B demonstrate. Specific accuracy goals

listed in the tables represent the values for the methods given in such nationally recognized

compendia as SW-846 and the CLP SOW. Project-specific analytical accuracy goals should be

established during the DQO process.

The percent recovery shall be within the limits defined in site-specific plans. Values exceeding the

acceptance criteria, established during the site-specific DQO process, must be evaluated for

usability.

1.5.1.3 Representativeness

Representativeness expresses the degree to which sample data accurately and precisely represent

a characteristic of a sample population, parameter variation at a sampling point, process

condition, or an environmental condition (EPA, 1998). Representativeness depends on the proper

design and execution of a sampling program and it is achieved through careful selection of

sampling intervals and locations as well as analytical parameters and the correct collection

methods.

Representativeness is a qualitative term that should be evaluated to determine whether in situ and

other measurements are made and physical samples collected in such a manner that the resulting

data appropriately reflect the media and phenomenon measured or studied. The number of

samples collected must be sufficient to demonstrate that the data represent the population of

interest to the statistical certainty required by the DQOs. Collection, storage, handling, and

transport of samples should be performed in a manner that preserves the in situ characteristics of

the samples and maintains the representativeness of the sample to the site.

1.5.1.4 Completeness

Completeness is the amount of valid data obtained that satisfies the data requirements. The DQO

process shall identify the critical decisions to be made and the data needed to support those

decisions.

Completeness is not intended to be a measure of representativeness; that is, it does not describe

how closely the measured results reflect the actual concentration or distribution of the pollutant in

the media sampled. A project could produce 100 percent data completeness (i.e., all samples

planned were actually collected and found to be valid), but the results may not be representative

of the pollutant concentration actually present. Alternatively, there could be only 70 percent data

Page 41: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 14 of 55

completeness (30 percent lost or found invalid), but due to the nature of the sample design, the

results could still be representative of the target population and yield valid estimates.

Completeness is affected by unexpected conditions that may occur during the data collection

process. The number of samples prescribed for an activity must be sufficient to meet data

requirements identified in the DQO process. Lack of completeness may require reconsideration

of the limits for the false negative and positive error rates.

1.5.1.5 Comparability

Comparability is the qualitative term that expresses the confidence that two data sets can

contribute to a common analysis and interpolation. Comparability must be carefully evaluated to

establish whether two data sets can be considered equivalent in regard to the measurement of a

specific variable or group of variables. In a laboratory analysis, the term comparability focuses on

method type comparison, holding times, stability issues, and aspects of overall analytical

quantitation.

There are a number of characteristics that can make two data sets comparable, these

characteristics vary in importance depending on the final use of the data. The closer these data

sets are with regard to these characteristics, the more appropriate it will be to compare them. For

example, large differences between characteristics may be of only minor importance, depending

on the decision that is to be made from the data.

Comparability is maximized by using standard techniques and procedures (e.g., standard operating

procedures) to collect and analyze representative samples and by reporting analytical results in

appropriate units. Comparability is limited by the other quality indicators because only when

precision and accuracy are known can datasets be compared with confidence.

1.5.2 Secondary Data Quality Indicators

Secondary DQIs are sensitivity, recovery, memory effects, limit of quantitation, repeatability, and

reproducibility. Generally, these DQIs are applied only to laboratory measurement processes.

The site-specific DQO process will determine which, if any, secondary DQIs apply to the project.

1.5.2.1 Sensitivity

Sensitivity is the capability of a method or instrument to discriminate between measurement

responses representing different levels of a variable of interest. Sensitivity is determined from the

value of the standard deviation at the concentration level of interest. It represents the

Page 42: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 15 of 55

minimum difference in concentration that can be distinguished between two samples with a high

degree of confidence. Sensitivity must be sufficient to detect contaminates at or below decision

levels.

1.5.2.2 Recovery

Recovery is an indicator of bias in a measurement. This is best evaluated by the measurement of

reference materials or other samples of known composition. In the absence of reference

materials, spikes or surrogates may be added to the sample matrix. The recovery is often stated

as the percentage measured with respect to what was added. This means that control charts or

some other means should be used for verification. The Industrial Sites uses matrix spike and

matrix spike duplicates of selected environmental samples as an indicator of bias.

1.5.2.3 Memory Effects

A memory effect occurs when a relatively high-concentration sample influences the measurement

of a lower concentration sample of the same analyte. This can occur when the higher

concentration sample precedes the lower concentration sample in the same analytical instrument.

This represents a fault in an analytical measurement system that reduces accuracy.

1.5.2.4 Limit of Quantitation

The limit of quantitation is the minimum concentration of an analyte or category of analytes in a

specific matrix that can be identified and quantified above the method detection limit and within

specified limits of precision and bias during routine analytical operating conditions.

1.5.2.5 Repeatability

Repeatability is the degree of agreement between independent test results produced by the same

analyst using the same test method and equipment on random aliquots of the same sample within a

short time period.

1.5.2.6 Reproducibility

Reproducibility is the precision that measures the variability among the results of measurements of

the same sample at different laboratories. It is usually expressed as a variance and low values of

variance indicate a high degree of reproducibility.

Page 43: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 16 of 55

1.6 Measurement Quality

The DQO process defines the quality of data needed to support decisions for the CAU under

investigation. The quality required of a data set is determined by the intended use in the decision

making. All data to be collected are classified into one of three measurement quality categories:

quantitative, semiquantitative, and qualitative. The categories for measurement quality are

defined as follows:

1.6.1 Quantitative

Quantitative data results from direct measurement of a characteristic or component within the

population of interest. These data require the highest level of QA/QC in collection and

measurement systems because the intended use of the data is to resolve primary decisions

(i.e., rejecting or accepting the null hypothesis) and/or verifying closure standards have been met.

Laboratory analytical data are usually assigned as quantitative data.

1.6.2 Semiquantitative

Semiquantitative data is generated from a measurement system that indirectly measures the

quantity or amount of a characteristic or component of interest. Inferences are drawn about the

quantity or amount of a characteristic or component because a correlation has been shown to exist

between results from the indirect measurement and the quantitative measurement. The QA/QC

requirements on semiquantitative collection and measurement systems are high but may not be as

rigorous as a quantitative measurement system. Semiquantitative data contribute to decision

making, but are not generally used alone to resolve primary decisions. The data are often used to

guide investigations toward quantitative data collection.

1.6.3 Qualitative

Qualitative data identifies or describes the characteristics or components of the population of

interest. The QA/QC requirements for qualitative data are the least rigorous on data collection

methods and measurement systems. Professional judgement is often used to generate qualitative

data. The intended use of the data is for information purposes, to refine conceptual models, and

guide investigations rather than resolve primary decisions. This measurement of quality is

typically associated with historical information and data where QA/QC may be highly variable or

not known.

1.7 Reports to Management

Participant management and NV ERP Project Managers shall be made aware of project activities

and shall participate in the development, review, and operation of these activities. The

Page 44: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 17 of 55

management of all participants shall be informed of quality-related activities through the receipt,

review, and/or approval of any of the following:

• Project-specific plans and procedures• Assessment reports• Corrective action requests, corrective actions, and schedules• Nonconformance reports (NCR)

Individuals identifying nonconforming conditions or deficiencies are responsible for documenting

and reporting said conditions. All nonconformances and findings related to quality shall be

corrected as required, documented, and properly reported. In addition, periodic assessment of

QA/QC activities and data quality parameters shall be evaluated and reported to the participating

project field and laboratory management.

1.8 Readiness Reviews

Readiness reviews verify that all planning documents and systems are in place for the

successful and efficient accomplishment of the mission. The verification that personnel are

qualified and knowledgeable in the activities they are assigned to perform is included in

readiness reviews.

Readiness reviews shall be performed by participating organizations prior to the start of any

major scheduled activity and prior to restarting work (following stop work orders) to verify and

document that project planning and prerequisites have been satisfactorily completed. At a

minimum, readiness reviews shall verify that the following issues have been addressed:

• The scope of work is compatible with project objectives.

• The planned work is appropriate to meet objectives.

• Work instructions have been reviewed for adequacy and appropriateness, formallyapproved, and issued to personnel who will be performing the work.

� Hazards have been identified, analyzed, categorized, and controls implemented.

• Proper resources (e.g., personnel, equipment, and materials) have been identified and areavailable.

• Assigned personnel have read the applicable work instructions and have been trained andqualified.

Page 45: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 1.0Revision: 3Date: 02/07/2002Page 18 of 55

• Internal and external interfaces have been defined.

• Proper work authorizations and permits have been obtained.

• The calibration of all material and test equipment is current.

� A feedback mechanism has been established to facilitate process improvement.

1.9 Stop Work Order

All NV ERP personnel and project participants are authorized and have the responsibility to stop

work when a condition adverse to safety, quality, or the environment is identified that, if allowed

to continue, would result in personal injury, damage to NV ERP equipment or property, or have

an adverse impact on mission accomplishment, budget, schedule, or cause damage to the public

and/or environment. If imminent danger exists, a Stop Work Order (SWO) may be verbally

imposed. An SWO may be limited to a specific activity, item, or design, or it may be broad in

scope and encompass all activities relating to the deficiency or violation.

Resumption of work shall begin only upon completion of the necessary actions to eliminate the

adverse condition specified in the SWO and with approval of the NNSA/NV ERD Director or

designee. Health and Safety (H&S)-related SWOs shall require the additional signature of the

appropriate H&S Manager. If the SWO involves a quality issue, resumption of work shall also

require the approval signature of the NV ERP QAC.

The issuance of an SWO shall be in accordance with the identifying organization’s protocol.

Page 46: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 2.0Revision: 3Date: 02/07/2002Page 19 of 55

2.0 Criteria 2 - Personnel Training and Qualifications

The NV ERP and project participant management shall ensure that personnel are qualified and

knowledgeable in the activities they perform. Training should emphasize correct performance of

assigned work and provide an understanding of why quality requirements exist. Personnel

qualification and training records shall be maintained as quality documents in accordance with

DOE Order 414.1A, Quality Assurance (DOE, 2001a).

2.1 Project Personnel

Personnel shall be trained and qualified to perform the tasks to which they are assigned.

Objective evidence of qualifications may include academic credentials, personal resumes,

registrations and/or certifications, licenses, and training records. The qualifications of personnel

shall be evaluated against assigned responsibilities and any identified training needs must be

addressed.

Training should be provided to achieve and maintain proficiency; adapt to changes in technology,

methods, or job description; and allow for feedback and effectiveness of job performance.

Training may take the form of orientation and/or indoctrination, formal classroom training, or

on-the-job training. This training should include regulatory requirements, scopes of work,

QA/QC requirements, and applicable work instructions.

On-the-job training should be conducted and documented by personnel experienced in the task

being performed in accordance with each organization’s requirements. Any work performed by a

trainee should be under the supervision of an experienced individual. Trainees should

demonstrate capability prior to performing work independently.

2.2 Subcontractor Personnel

Subcontractor personnel shall be qualified and trained to perform the duties for which they were

contracted. The contracting organization shall be responsible for verifying the qualifications of

their subcontractors.

2.3 Analytical Laboratory Personnel

Laboratories shall be prequalified prior to performing work for the Industrial Sites Project.

During the prequalifing assessment, contractors shall verify that analytical laboratories have

Page 47: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 2.0Revision: 3Date: 02/07/2002Page 20 of 55

established job descriptions for positions affecting data quality. These descriptions shall provide

the minimum qualifications in terms of education, experience, and skills necessary for an analyst

to carry out duties in the laboratory. Laboratories shall provide appropriate orientation and/or

training of the laboratory QA program and project requirements and must implement a

performance-based qualification program which includes periodic requantification for analysts

performing work for the Industrial Sites Project.

Page 48: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 21 of 55

3.0 Criteria 3 - Quality Improvement

The objective of the Industrial Sites Project is to produce quality products and to continuously

seek methods to improve both processes and products. Processes shall be established with the

objective of preventing problems and improving quality. Peer reviews of various work products

should be built into the work processes to ensure the quality of the products prior to release. All

personnel are encouraged to identify and suggest improvements in all areas of Industrial Sites

activities.

Management shall seek to cultivate an atmosphere which fosters the belief that improvement is

always possible, and accountability and excellence must be established at all levels. It is equally

important to identify and implement process improvements and efficiencies. Successful

techniques should be evaluated to determine the potential for performance improvements in other

areas or projects. The following sections identify processes that, at a minimum, shall be

implemented.

3.1 Internal Quality Control Checks

Quality control checks shall be performed for data collected in the field and data obtained

through on-site and/or off-site analysis. Existing and historical data used for input into a model

shall have quality control checks to the extent that it is possible. Information shall be reviewed

by someone other than the initiator to ensure correct collection, transcription, and manipulation.

Transcribed data shall be verified to ensure the correctness of the transcription. Data that has

been manipulated shall be checked to ensure the manipulation process was performed as the

originator intended.

Proprietary computer applications used for the evaluation of historical data maintained or

transferred via electronic media shall have QC checks performed that are appropriate to the

application being used. These checks must be documented and maintained in accessible files.

Field sampling and laboratory analytical activities shall incorporate QC procedures. All field and

laboratory operations and systems shall be evaluated for their potential to impact the quality of

generated data. System quality controls that meet the requirements of this QAPP shall be

established and documented through the use of approved procedures, plans, or instructions.

Page 49: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 22 of 55

Quality control samples shall be incorporated into the analytical stream to assess the overall data

quality produced by the program. The QC samples consist of field- and laboratory-generated

samples which are used to evaluate sampling and analytical precision and accuracy as well as the

levels of potential contamination introduced by the sampling and analytical effort. The following

paragraphs describe the QC samples that will be generated, as applicable, for the samples to be

collected.

3.1.1 Field Quality Control

The field data collection QC program is designed to provide confidence that data collected

during field activities adequately represent the area of interest. For sampling activities, field QC

samples provide a mechanism for assessing and documenting that the collection process meets

the QA objectives of the project. The number and type of field QC samples required shall be

determined during the DQO process for each site. Field QC samples may include trip blanks,

equipment rinsate blanks, source blanks, field blanks, and field duplicates. Field QC samples

shall be submitted to the laboratory in such a manner that the laboratory is not aware that the

sample is for QC purposes. Collection and documentation of field QC samples shall be in

accordance with approved procedures and site-specific plans. Other types of data collected, such

as observational data and measurements, shall have the appropriate quality control checks

applied to ensure the information collected is of a quality that meets the objectives of the activity.

3.1.1.1 Equipment Rinsate Blank Samples

An equipment rinsate blank is collected from the final rinse solution from the equipment

decontamination process to determine the effectiveness of the decontamination process. The

blanks shall be prepared by pouring deionized water through or over a sampling device after it

has been decontaminated and prior to using the device for environmental sample collection. If

equipment rinsate blank analytical results indicate possible contamination of samples, potentially

affected environmental sample results shall be reviewed to determine whether qualifiers should

be assigned to the data or whether the source should be resampled. Results of rinsate blank

analyses shall be maintained with the corresponding sample analytical data in the laboratory

records file and reported in the laboratory data package.

3.1.1.2 Field Blank Samples

Field blanks are collected and analyzed by the laboratory to determine if contamination in the air

during sample collection and packaging may have contaminated the samples. The field blanks are

prepared by pouring deionized water into clean sample containers in the field near the sampling

locations, or by exposing a clean swipe to the same ambient as those present during

Page 50: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 23 of 55

sampling. Field blanks should be collected as closely in time and space to the environmental

sample as possible. If field blank analytical results indicate possible contamination of associated

samples, environmental sample results shall be reviewed to determine whether qualifiers should be

assigned to the data or whether the source should be resampled.

3.1.1.3 Trip Blank Samples

A trip blank is a 40-milliliter volatile organic analysis (VOA) container of organic-free water that

is shipped to the field along with the other VOA sample containers. The blank is not opened, but

is otherwise maintained, handled, stored, packaged, and shipped as if it were collected in the

field. The purpose of the trip blank is to determine if contaminants have entered the sample

through diffusion across the Teflon™-faced, silicone rubber septum of the sample vial during the

performance of laboratory, field, or shipping procedures. The trip blank is only analyzed for

volatile organic constituents. Trip blanks shall be submitted for analysis at a frequency of one

sample per shipping container that contains field VOA samples.

Following the analyses, if the trip blanks indicate possible contamination of the samples,

the appropriate project personnel shall be notified. Results of trip blank analyses shall be

maintained with the corresponding sample analytical data in the laboratory records file and

reported in the laboratory data package.

3.1.1.4 Duplicate Samples

Field duplicates are QC samples that are collected as closely in time and space to the

environmental sample as possible to assess sample variability and to measure sampling and

analytical variability. The field duplicates shall mirror the sampling and analytical profile of the

original sample and be assigned a unique sample number. The duplicate sample number shall

not indicate that it is a QC sample to minimize handling, analysis, and data-evaluation bias.

Parameters to be analyzed shall be the same as those analyzed for the corresponding

environmental sample. Sample management and documentation procedures for duplicates shall

be the same as those used for environmental samples.

3.1.2 Laboratory Quality Control

All on-site and off-site laboratories performing analyses for the Industrial Sites Project shall

conduct their activities in accordance with a written and approved QA plan. Laboratory quality

control (LQC) samples shall be analyzed using the same analytical procedures used to analyze

environmental samples. Each analytical laboratory shall generate QC samples during each

analytical run to assess and document accuracy and precision associated with each analytical

Page 51: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 24 of 55

measurement in accordance with the laboratory QA plan. All data from concurrently analyzed

LQC samples and other quality controls which are used to demonstrate analytical control shall be

included in the laboratory’s analytical report. The requirements for the types and number of LQC

samples will depend on the analytical procedure or method and the laboratory’s QA objective for

each test. Laboratory quality control samples include method blanks, surrogate-spike, and matrix

spike/matrix spike duplicate samples.

3.1.2.1 Laboratory Control Samples

One laboratory control sample (LCS) shall be prepared and analyzed with each batch of samples

per matrix. The LCS shall be carried throughout the sample preparation and analysis procedures

to assess laboratory accuracy and precision. The LCS shall be analyzed concurrently with each

analytical batch for each analyte of interest and shall be prepared from standards independent of

the calibration standard. Control limits for recovery shall be established, and recovery data shall

be plotted on internal control charts. The LCS data outside these recovery limits shall be

considered "out of control," and the laboratory shall initiate corrective action(s) that shall be

performed in accordance with the laboratory's QA plan. Results of duplicate LCS analyses shall

be reported as RPD and %R and included with the associated analytical report.

3.1.2.2 Method Blank Samples

Method blanks shall be analyzed by the laboratory to check for instrument contamination and

contamination and interference from reagents used in the analytical method. A method blank

shall be concurrently prepared and analyzed for each analyte of interest for each analytical batch.

Method blank data outside statistical control limits shall be considered "out of control," and

corrective action(s) shall be performed in accordance with the laboratory's QA plan. Method

blank data shall be reported in the same units as the corresponding environmental samples, and

the results shall be included with each analytical report.

3.1.2.3 Surrogate-Spike Samples

Surrogate-spike sample analysis shall be performed for all samples analyzed by gas

chromatography (GC), gas chromatography/mass spectrometry (GC/MS), and high-performance

liquid chromatography (HPLC) to monitor the percent recovery of the sample preparation and

analytical procedure on a sample-by-sample basis. Surrogate standards are nontarget compounds

added to GC, GC/MS, HPLC standards, blanks, and samples prior to extraction or purging.

Surrogate compounds and concentrations added shall be those specified in the applicable

analytical method. Recovery values for surrogate compounds shall be within the control limits

specified by the laboratory and in accordance with assessment procedures in the laboratory's QA

Page 52: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 25 of 55

plan, or the analysis shall be repeated. Results of surrogate spike sample analyses shall be

reported as % R. Surrogate spike samples are not used on radiological samples.

3.1.2.4 Matrix Spike/Matrix Spike Duplicate Samples

Project site-specific matrix spike/matrix spike duplicates shall be analyzed by the laboratory to

determine interferences of the sample matrix on the analytical methods and subsample variance of

the laboratory data. A separate sample aliquot shall be spiked with the analytes of interest and

analyzed with every 20 samples per matrix or, if fewer than 20 samples were collected, at least

one of the samples shall be spiked. Results of the matrix spike/matrix spike duplicate analyses

shall be reported as percent recovery and RPD and included with the analytical report. Results

that are outside the established recovery or reproducibility limits for the analytical method shall be

considered “out of control,” and the laboratory shall initiate corrective action(s) that shall be

performed in accordance with the laboratory’s QA plan.

3.1.2.5 Laboratory Duplicate Samples

Two aliquots of the same sample per matrix shall be prepared and analyzed for inorganic analysis

and the duplicate results will be used to calculate the precision as defined by the RPD. If the

precision value exceeds the control limit, the appropriate laboratory personnel will identify the

root cause of the nonconformance and implement corrective actions. A laboratory duplicate

analysis shall be performed for every 20 samples.

3.2 Data Precision, Accuracy, and Completeness

Quality control sample results are used to evaluate laboratory and field precision and accuracy.

Precision shall be determined by comparing the concentrations of the various constituents

between duplicate analyses. Accuracy shall be determined by comparing analytical results with

the known (true) value of a reference standard (e.g., a laboratory control sample). The accuracy

of the spiked samples must be within the accepted accuracy of the method of analysis for the

analyte of interest. Sample results falling outside of acceptable ranges for precision and accuracy

shall be brought to the attention of laboratory management for evaluation and corrective action(s)

as needed. Data precision, accuracy, and completeness requirements shall be dependent on the

end use of the data and determined during the DQO process for each site.

Laboratory results shall be checked upon receipt. If there appears to be an error in the analysis,

the laboratory shall be contacted immediately, and corrective action(s) must be taken. If

investigation reveals that processes were not in control, corrective action(s) shall be taken and the

resulting data evaluated to determine any impacts.

Page 53: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 26 of 55

3.3 Corrective Action

This section establishes the methods and responsibilities for identifying, reporting, controlling, and

resolving conditions of nonconformance and conditions adverse to quality for activities performed

in support of the Industrial Sites Project.

3.3.1 Nonconformance

A nonconformance is a deficiency in characteristic, documentation, or procedure that renders the

quality of an item or activity unacceptable or indeterminate. The NV ERP policy encourages all

personnel to identify and document nonconforming items and processes. It is also NV ERP policy

to identify nonconformances in a manner that focuses on solutions and discourages fault-finding in

order to encourage the open identification and resolution of problems. Individuals identifying

nonconforming conditions or items are responsible for documenting and reporting the

nonconformance. Responsible personnel should be notified at the time the nonconformance is

identified so that, when possible, corrective measures may be taken immediately.

All NCRs shall be handled in accordance with each organization’s internal processes. An NCR

shall specify:

• Originator• Date of the nonconformance• NCR number (unique)• Responsible organization• Requirement(s)• Nature of the nonconformance• Disposition• Technical justification for disposition

When an NCR affects cost, schedule, scope, or is a H&S issue, the applicable NV ERP Project

Manager and the NV ERP QAC must be notified.

3.3.2 Root Cause

A root cause is the most basic element that, if corrected, will prevent recurrence of the same

(or similar) problem. Root-cause analysis should be used where the understanding of the basic

underlying cause is important to the prevention of similar or related problems. The root-cause

analysis should be used to gain an understanding of the deficiency, its causes, and the necessary

corrective actions to prevent recurrence. The level of effort expended should be based on the

possible negative consequences of a repeat occurrence of a problem.

Page 54: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 3.0Revision: 3Date: 02/07/2002Page 27 of 55

3.3.3 Trend Analysis

Trend analyses should be performed on nonconforming conditions, deficiencies, root causes, and

the results of improvement initiatives to identify any possible trends. Adverse trends shall be

brought to the attention of the appropriate management. Positive trends, such as improved

performance or cost savings resulting from enhancements or the application of new technology,

should be shared to facilitate improvement in other areas or projects. As appropriate, information

obtained from trend analyses should be included in a lessons learned system.

3.3.4 Lessons Learned

A lessons learned system has been established at NNSA/NV as a focal point for reporting and

retrieving important information concerning experiences gained through previous activities.

Continuous improvement can be fostered through incorporation of applicable lessons learned into

work processes and project planning activities, including work plan development, budget

development, and strategic planning. The lessons learned program should be used interactively

with other management tools such as critiques, assessments, readiness reviews, and evaluations of

field activities.

Page 55: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 4.0Revision: 3Date: 02/07/2002Page 28 of 55

4.0 Criteria 4 - Documents and Records

This QAPP supplements requirements for site-specific activities of the NV ERP. Each project

also has planning documents and work plans, as deemed necessary, for the work to be performed.

Contractors may determine that additional procedures are necessary to further define the

responsibilities and activities of specific scopes of work. Figure I-1 is a flowchart of the guidance

documents.

4.1 Documents and Records

Systems and controls shall be implemented by project participants for identifying, preparing,

reviewing, approving, revising, collecting, indexing, filing, storing, maintaining, retrieving,

distributing, and disposing of pertinent quality documentation and records. For characterization

and remediation at sites located in the state of Nevada, the following documents shall be

developed in accordance with the FFACO (1996) outlines contained in Appendix A of this plan:

� Corrective Action Investigation Plan (CAIP)� SAFER� Corrective Action Decision Document (CADD)� CADD/Closure Report� Corrective Action Plan (CAP)� CADD/CAP� Closure Report (CR)

The format for documents pertaining to sites outside of the state of Nevada where NV ERP is

conducting project work shall be established in cooperation with the appropriate state agencies.

4.1.1 Document Review and Control

Plans and reports shall be reviewed for quality requirements, technical adequacy, completeness,

and accuracy prior to their approval and issuance. The NV ERP documents shall be reviewed in

accordance with the procedure DOE/NV EM-02-002, Document Review and Coordination

(DOE/NV, 2000a).

A system or process for identifying documents that require control, and controlling those

documents, shall be implemented to ensure that the latest revision of a document is used. The

Industrial Sites participant management is responsible for ensuring that personnel who perform

work are in possession of the most current version of the documents applicable to the activities

being conducted.

Page 56: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 4.0Revision: 3Date: 02/07/2002Page 29 of 55

Revisions to controlled documents shall be approved by the same level of authority or

organization as the original. Documents no longer in use should have their status clearly

indicated, and record copies should be maintained in accordance with DOE Order 200.1,

Information Management Program (DOE, 1996a), and the applicable records inventory and

disposition schedule.

4.1.2 Change Control

Changes or modifications to approved procedures or plans may be necessary in order to adjust an

activity to actual field conditions or to revise programmatic methods of implementing project

requirements. Industrial Sites participants shall ensure that changes are properly identified,

documented, approved, and controlled in accordance with the individual procedures of each

participant organization. Verbal authorization of changes must be documented and followed up

with a written change notice in a timely manner. Changes shall be approved commensurate with

the original document prior to implementation of the change. Changes to the site-specific health

and safety plan shall be in accordance with the governing organization’s requirements. The

NV ERP shall be notified of changes that impact the technical scope, cost, or schedule of the

project.

4.1.3 Records Maintenance

Sufficient records of Industrial Sites activities shall be prepared, reviewed, and maintained.

Project records shall be maintained in accordance with DOE Order 200.1, Information

Management Program. Contractors and other agency participants shall have a system in place for

the storage and retrieval of quality records that is consistent with environmental regulations and

DOE Order 200.1. (DOE, 1996a)

Page 57: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 30 of 55

5.0 Criteria 5 - Work Processes

The performance of activities shall be based upon the objectives of the project. Details of specific

environmental data-collection activities will be discussed in the applicable site-specific planning

documents. Appropriate technical methods or a scientific rationale shall be employed. Activities

shall be performed in accordance with approved procedures and site-specific plans that comply

with the applicable requirements of DOE Orders, procedures, and project planning documents.

Upon request, contractors and participating organizations shall supply the NV ERP with copies of

applicable procedures. Deviations from the applicable approved project plans and procedures

shall be approved and documented.

5.1 Evaluation and Use of Existing and New Data

Existing and new data shall be evaluated against current requirements for their intended use. This

analysis consists of editing, screening, checking, auditing, verification, and review. Methods shall

be in place for the control and transfer of data, control of interpretive work products, and the

control of data within the central database. The process should provide guidance for gathering,

manipulating, and distributing data. The quality of existing data shall be determined, based on the

traceability of data and the level of QA/QC applied to the data during initial collection, prior to

inclusion into the central database. Reports, models, or interpretative works shall indicate the

quality of the data being used. Prior to use, newly acquired analytical data will be evaluated

against predetermined objectives and criteria.

5.2 Computer Hardware and Software

Computer hardware/software configurations are defined as the combination of computer program

software version, operating software version, and model of computer hardware. Computer

software and hardware/software configurations used in the acquisition, modeling, or storage of

environmental data shall be installed, tested, used, maintained, controlled, and documented to

meet the requirements of the user and/or data management criteria. Compatibility between

software and hardware systems must be achieved for long-term retrievability. To the extent

possible, contractor’s and project participants’ hardware and software should be compatible with

the NV ERP.

Page 58: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 31 of 55

5.2.1 Computer Systems

Computer hardware/software configurations for the storage and manipulation of environmental

data should be tested by knowledgeable individuals prior to actual use and the results documented

and maintained. Changes to hardware/software configurations should be assessed to determine

the impact of the change on the technical and quality objectives of the environmental program. If

any of the components are changed or modified and a new configuration results, or if program

requirements change so that the capability of the hardware/software configurations to meet the

new requirements is uncertain, then the configuration should be retested and redocumented.

Computer hardware/software configurations integral to measurement and test equipment

(M&TE) that are calibrated for specific uses do not require further testing unless the software

uses change or the configuration is modified.

The physical media on which software is stored shall be controlled and protected so that software

and data are physically retrievable and protected from loss or compromise by catastrophic events.

Back-up copies shall be maintained so that a single event will not cause a significant loss of

software or data.

5.2.2 Software Design/Development

Project participants involved in the development or use of major-use software for modeling or

technical computations will develop and implement processes for the development, modification,

verification/validation, and control of computer software codes. Code criteria should be clearly

defined prior to development or purchase and should be consistent with applicable national

standards. Software will be qualified for use, based on its ability to provide acceptable results for

its intended application. The configuration of software should be controlled and documented so

traceability is maintained through the developmental history. Documentation of the development

or modification of software codes must include the appropriate peer reviews and verification/

validation.

Page 59: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 32 of 55

5.2.2.1 Code Evaluation

Newly developed computer codes or modification to existing software shall be reviewed and the

reviews documented by individuals who are knowledgeable in the area of code development.

Reviewers should consider the following aspects:

� Assumptions are reasonable and valid

� Correctness of the mathematical model

• Conformance of methods to accepted and published concepts (recognizing thatalternative methods and interpretations other than those of the evaluators may beacceptable)

• Consistency of results with known data

• Reasonable and prudent use of data and analysis tools

• Appropriateness for intended purpose

5.2.2.2 Code Verification/Validation

Software should be qualified for use based on its ability to provide acceptable results for the

intended application. Software verification and validation activities will include provisions for

providing confidence that the software adequately and correctly performs all intended functions.

The extent of verification/validation required shall depend on the complexity, risk, and uniqueness

of the code. Computer software code modifications shall be verified and validated according to

the same requirements as the original code. Verification of changes may be limited to the scope

of the modification if the rest of the code is not affected. Acquired technical software used

without modification must have operational checks performed through test cases to verify that the

software is functioning as intended.

Computer application used for the evaluation of historical data maintained or transferred via

electronic media shall have QC checks performed as appropriate to the application being used.

These checks must be documented and maintained in project files.

5.2.2.3 Software Documentation

All developed or procured computer codes shall be uniquely identified. Software code

documentation shall be maintained with associated calculations and reference material.

Documentation will consist of software design and reference material, verification/validation

records, operational test records, and user-oriented information.

Page 60: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 33 of 55

5.2.3 Peer Review of Software and Code Applications

The peer review is an assessment of the assumptions, calculations, extrapolations, alternate

interpretations, methodology, acceptance criteria, and conclusions pertaining to interpretive work

products generated through use of computer software. Peer reviews shall be performed and

documented to ensure that interpretive work products are technically adequate, properly

documented, and satisfy established technical and quality requirements. Peer reviewers shall

possess the appropriate subject matter/technical expertise and not have participated in preparing

the original work. All review comments and the attendant comment responses shall be recorded

on review sheets and maintained in the project files. The acceptable level of accuracy of each

interpretive work product should be established by project management.

5.3 Field Investigation

Field activities generally involve the collection of data for the purpose of decision making. Field

data acquisition shall be accomplished through the use of approved plans and contractor-specific

procedures or instructions, by qualified personnel, using appropriate tools and calibrated

equipment. Additionally, all work shall be performed safely within the controls established to

prevent/mitigate hazards. Details of specific environmental data collection activities shall be

delineated in the associated project plans and instructions. Contractors and participating

organizations shall supply NNSA/NV and Nevada Division of Environmental Protection (NDEP)

with copies of the procedures and plans applicable to the activities being performed in the state of

Nevada. Additionally, project participants should provide NNSA/NV and the NDEP with site-

specific information that specifies what procedures will be used for a project or activity. Data

acquisition methods for which a procedure does not exist (those that are unique, experimental, or

under development) shall be detailed in the project-specific plans or instructions. Deviations from

the applicable project plans and procedures shall be approved and documented.

5.3.1 Procedures

Specifically, EPA QA/R-5, EPA Requirements for Quality Assurance Project Plans, and

EPA QA/G-5, EPA Guidance for Quality Assurance Project Plans, recommend procedures be

developed for the following described in Sections 5.3.1.1 to 5.3.1.7.

5.3.1.1 Sampling Methods

The planning documents shall specify the type of samples to be obtained (e.g., grab, composite,

subsurface drilling) together with the method of sample collection. Although the list below is not

Page 61: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 34 of 55

all inclusive, the Industrial Sites projects will typically require procedures of the following data

collection methods:

• Ground-penetrating radar • Global positioning system• Surface magnetic survey• Surface electromagnetic resistivity• Radiological surveys• Surface soil sampling• Shallow subsurface sampling• Subsurface sampling• Sampling during drilling• Sediment sampling• Soil sample compositing• Surface water sampling• Swipe sampling• Field-screening analysis

5.3.1.2 Sample Custody

Sample custody procedures are necessary to prove that the sample data correspond to the sample

collected, if data are intended to be legally defensible. Chain-of-custody (COC) for each field

sample collected must be documented to provide the traceability of possession from the time the

samples are collected until disposal. A sample is considered to be in custody if it meets any of the

criteria listed below. For custody to exist, the sample must be in:

• Person's actual possession• Person's unobstructed view, after being in the person's physical possession• Secured area to prevent tampering, after having been in physical possession• Designated secured area, restricted to authorized personnel only

5.3.1.3 Chain of Custody Form

Project participants conducting sampling activities should have procedures or instructions in place

to ensure that COC is properly executed. Each individual who possesses a sample is responsible

for sample custody until the sample is relinquished to another individual or an authorized secure

storage area via the COC form. Field teams shall initiate COC forms for samples collected during

field activities. The basic COC requirements are as follows:

� Whenever samples are transferred to a new sample custodian, the new custodian shall signhis or her name, company name, and note the time and date that the transfer occurs.

Page 62: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 35 of 55

� Each sample shall be assigned a unique identification number that is recorded on both theindividual sample label and the COC form.

� There shall be no gaps on the record of custody.

� The COC form shall accompany the samples during handling and shipment, and it shallchronicle the history of custody.

� If field samples are split for shipment to separate laboratories or samples are composited,after a COC has been initiated, a new COC form must be generated. The original and thenew COC forms must cross-reference each other using the unique COC form number. Copies of the original field COC form shall be attached to the new form.

� In the event interim storage is necessary, a method of securing and maintaining theintegrity of the sample shall be employed.

� Sample information on the original COC form shall not be deleted, marked out, orobscured in any way.

� One copy of the COC form shall be retained by the field sampling personnel for trackingpurposes.

� While the transfer of samples to FedEx or a common carrier does not require the carrier’ssignature on the COC form, the carrier’s name and way bill should be noted on the form.

Custody is initiated by field personnel at the sampling location, and completed by laboratory

personnel upon final disposition of the sample. Without exception, sample custody shall be

maintained for all samples.

5.3.1.4 Custody Seals

To ensure that tampering is easily detectable, each sample container shall be individually sealed

with a custody seal. The seal shall be placed over or around the lid of the sample container so

that the container cannot be opened without breaking the seal. Each custody seal shall be initialed

and dated by a member of the sampling team.

5.3.1.5 Sample Labels and Identification

Identification and traceability of samples collected as part of a data-collection task are necessary

to ensure the success of the Industrial Sites Project. Sample labels shall be affixed to containers in

a manner that does not obscure any data preprinted on the containers. Sampling information

Page 63: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 36 of 55

shall be printed on the labels using indelible ink. The sample label shall contain the following

required information:

� Project name• Unique sample number• Sampling date and time (military)• Sample location and depth interval (if applicable)• Sample medium• Requested analyses• Name of the individual collecting the sample• Preservation or conditioning of the sample

Each sample number shall be indicated on both the container and field data/sample collection

forms. For samples requiring multiple containers, the same sample identification numbers shall be

required on each container. Labels that are not plastic coated and have the potential to smear or

deteriorate shall be covered with clear tape.

5.3.1.6 Sample Handling, Preservation, Packaging, and Shipping

The site-specific planning documents shall identify the appropriate sample containers,

preservation procedures, and holding times for specific analyses. Where applicable, sample

containers shall be certified clean per EPA protocol and shall remain sealed until ready for use.

Contractor procedures for sample handling, preservation, packaging, and shipping should include

instructions for maintaining the integrity of the sample, such as wrapping glass containers in

bubble wrap and maintaining the samples by the applicable preservation methods described in

the contractor site-specific plans. In addition, the procedure should specify that samples

transferred to an appropriate shipping container be cooled, if required; and protected from

breakage by using shock-absorbent packing material. Procedures should also specify all steps

necessary for packing the shipping containers. Procedures for the shipment of samples must

comply with Title 49 CFR, Parts 170 to 177 (CFR, 2000), for the packaging, labeling/placarding.

If a temporary storage is utilized, the procedure should state under what conditions the samples

may be stored.

5.3.1.7 Decontamination

Decontamination is primarily applicable in situations of sample acquisition from solid, semi-

solid, or liquid media. To prevent cross-contamination of samples, equipment coming in contact

with samples shall be decontaminated prior to use, between sampling locations, and before

leaving the site. The project must consider the appropriateness of the decontamination

Page 64: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 37 of 55

procedures for the project at hand. Contractor procedures should address decontamination

methods for the sample matrixes involved and the potential contaminants of concern. Heavy

equipment, such as a drill rig or backhoe, shall be decontaminated prior to arriving and departing

the site.

Equipment rinsate blanks shall be submitted to the analytical laboratory to assess the

effectiveness of the decontamination process. If the rinsate blank results indicate possible

contamination, corrective actions shall be implemented to preclude recurrence. Sample results

obtained using the suspect sampling equipment shall be reviewed to determine whether analytical

qualifiers should be assigned to the data.

5.3.2 Field Documentation

Field documentation should be of sufficient detail to facilitate the reconstruction of field

activities. Field personnel shall document activities on a daily activity report, a logbook, or on

the appropriate form as required by each contractor doing work for the NV ERD. Documentation

shall be made in indelible ink and include all information applicable to the activity being

performed. Types of field information include:

• The project name• The date and start time of each field activity• Names and affiliations of field personnel• The equipment used, including identification number• A general description of the day's field activities, showing the sequence of events• Problems encountered• Changes or modifications to the approved sample collection plan• Nonconformances and any corrective actions taken• Weather conditions• Field equipment calibration data• Field measurements or tests performed• References to associated forms for details of each activity conducted• The signature of the individual completing the report

Records shall be preserved and maintained in accordance with Section 4.1.3 of this document.

5.3.3 Investigation-Derived Waste

All investigation-derived waste (IDW) generated in support of NNSA/NV ERP activities shall be

managed in accordance with DOE Orders, U.S. Department of Transportation regulations,

Resource Conservation and Recovery Act (RCRA) regulations, Nevada laws and regulations, the

FFACO (1996), State/DOE agreements, relevant permits, and site-specific requirements.

Page 65: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 38 of 55

5.3.4 Photographic Documentation

With the approval of the NNSA/NV and in accordance with Nevada Test Site and Tonopah Test

Range requirements, photographs may be taken of the corrective action investigation and

corrective action. Photographs shall be documented on a photographic log in accordance with

contractor procedures. The photographs and negatives shall be processed and stored in

accordance with NNSA/NV security procedures.

5.3.5 Identification and Control of Items

The NV ERP participants shall establish and document sufficient controls to ensure that quality-

affecting items (such as equipment, components, and material) can be readily identified. These

controls shall be established to prevent incorrect use, to retain integrity of materials, and to

preserve the desired operating characteristics of equipment. Controls shall be applied that are

based on the risk to the project if control of the item is lost. Appropriate controls shall be applied

prior and subsequent to use. Specific requirements for preservation and packaging shall be

identified in project documents.

Hazardous materials shall be properly controlled and transported in accordance with

Title 49 CFR Part 173, Shippers—General Requirements for Shipments and Packaging

(CFR, 2000).

5.3.6 Calibration and Preventive Maintenance

The M&TE used in NV ERP projects shall be uniquely identified and controlled. A system of

calibration and preventive maintenance shall be employed by project participants to ensure the

proper operation of M&TE. Reference standards of the correct type, range, and acceptable

uncertainty shall be used for collecting data consistent with the project objectives.

5.3.6.1 Calibration

Approved procedures or the manufacturer’s recommendations shall be used to calibrate M&TE

prior to use and at prescribed intervals thereafter. The frequency of periodic or factory

calibrations shall be based on the manufacturer’s recommendations, national standards of

practice, equipment type and characteristics, and past experience. Operational or in-house

calibrations and/or source-response checks shall be performed on the appropriate M&TE prior to

the start of work and at prescribed intervals to verify the equipment’s continued accuracy and

operational function.

Page 66: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 39 of 55

Equipment for which the periodic calibration period has expired, equipment that fails calibration,

or equipment that becomes inoperable shall be tagged "out-of-service" and, when possible,

segregated to prevent inadvertent use. Results of activities performed using equipment that is out

of calibration shall be evaluated for adverse affects and the appropriate personnel notified.

Physical and chemical standards shall have certifications traceable to National Institute of

Standards and Technology, EPA, or other nationally recognized agencies. Supporting

documentation on all reference standards and equipment shall be maintained.

5.3.6.2 Preventive Maintenance

To avoid preventable breakdowns and work delays, project participants shall perform periodic

preventive maintenance on field and laboratory equipment in accordance with manufacturers’

specifications and warranties. The frequency of preventive maintenance shall be based on

manufacturers' recommendations and the users' professional knowledge and experience.

5.3.7 Laboratory Operation

Laboratories performing analytical work for the Industrial Sites Project must operate in

accordance with an acceptable written QA program. Plans and procedures relevant to the NV

ERP must be made available upon request. Deviations from approved procedures shall be

documented. See Section 7.0, “Procurement,” for additional requirements contractors should

verify during the procurement process.

All Industrial Sites participants who subcontract analytical services must ensure quality of

services through established procurement practices and oversight activities. Laboratories must

participate in an Interlaboratory Performance Evaluation program appropriate to sample types

and analyses. The laboratory must maintain participation in the DOE interlaboratory quality

assurance programs appropriate for the samples analyzed and in the EPA programs which are

most appropriate based on sample types and analyses. The laboratory must provide the results of

these performance evaluation studies along with the laboratory’s response to any deficiencies

which were identified.

5.3.7.1 Preanalysis Storage

Samples received at the analytical laboratory that have been entered into the sample tracking

system shall be placed into a storage refrigerator or secure storage area until analyzed. The

methods of storage are generally intended to:

Page 67: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 40 of 55

• Retard biological action• Retard hydrolysis of chemical compounds and complexes• Reduce volatility of constituents• Reduce adsorption effects• Reduce light exposure

Preservation methods are generally limited to pH control, preservative addition, and refrigeration.

Radiological samples do not require preservation. Preanalysis sample storage procedures shall be

documented and described in laboratory-specific procedures.

5.3.7.2 Post-Analysis Storage

The possibility of reanalysis requires that proper environmental control for post-analysis samples

be provided. These controls shall be described in laboratory-specific procedures. In general,

samples shall not be kept longer than one year. The samples shall be properly disposed of by the

laboratory unless other arrangements have been made to return them to the site.

5.4 Analytical Data Usability

Analytical data received for input into a project should be assessed for acceptability against the

requirements stipulated in the applicable project document. Personnel should verify that

analytical data reports have been reviewed by appropriate individuals other than those generating

the analytical data or the report and that all forms of the report (printed or electronic) carry a

notice of any limitations on the use of the data.

5.4.1 Data Management

Analytical data shall be controlled and managed to guarantee data integrity throughout

acquisition and development. Systems must be established for directing analytical data results

into a controlled data management system. Requirements shall be established for identification,

collection, selection, control, and transfer of analytical data. Analytical data that are submitted

shall be qualified and traceable to original data records and procedures established for

processing, storage, and control of data. Analytical data users are responsible for determining if

the data are sufficient for their intended use.

The management of data shall include procedures for the verification and validation of data to

ensure that all data used to support decisions made under the NV ERP are of known and

Page 68: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 41 of 55

documented quality. Procedures shall be used to optimize the detection and correction of errors

and prevent data loss during data reduction, reporting, and data entry into databases.

5.4.2 Evaluation and Use of Data

Participating organizations shall have a system in place for the control and transfer of data and

interpretive work products, which will provide guidance for gathering, manipulating, and

distributing data. The quality of existing data shall be determined, based on the traceability of

data and the level of QA/QC applied to the data during initial collection and current requirements

for their intended use. This analysis consists of editing, screening, checking, auditing,

verification, and review. Reports, models, or interpretative works shall indicate the quality of the

data being used. Prior to use, newly acquired analytical data will be evaluated against

predetermined objectives and criteria.

Computer applications used for the evaluation of data maintained or transferred via electronic

media shall have quality control checks performed as appropriate to the application being used.

5.4.3 Data Reduction, Verification, and Validation

Computations performed on raw data are considered data reductions. Numerical reduction of

field and analytical data shall be formally checked in accordance with approved procedures, and

this checking must be performed prior to the presentation of results. If unchecked results are to

be presented, transmittals or subsequent calculations based on these results must be marked

"preliminary" until the results are checked and determined to be correct.

Verification is the process of checking and reviewing the data reduction process. Data

verification is a systematic review of data by qualified individuals to check data reduction and

ensure that data meet specified guidelines.

Validation of analytical data is a comprehensive verification which includes complete review of

raw data. The site-specific DQO process shall establish what percentage of analytical data

packages shall be validated. Qualifiers may be attached to the data to indicate the results of the

verification process. These qualifiers may restrict or limit certain uses of the data.

Page 69: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 42 of 55

5.4.3.1 Data Completeness Review

A completeness review should be conducted to ensure that field and laboratory data and

documentation are present and complete. This review is designed to be expeditiously conducted

upon receipt of the data from the laboratory. During this review, problems should be identified

and documented. Information from this review should accompany the data. The review should

include the verification that:

• Chain of custody documents are complete and legible• The cover letter (case narrative) indicates significant problems• All requested analyses were performed on all samples• Holding times were not exceeded• The cooler temperature at time of sample receipt• The proper preservation and pH for each matrix and parameter was used• The laboratory log-in report• All field forms are present and complete• Report forms inventory includes all CLP or CLP-like forms• The report format reflects appropriate reporting level

5.4.3.2 Data Review and Summary

Selected QC checks and procedures shall be evaluated for compliance or noncompliance with

DQO standards. Deficiencies in the data package shall be communicated to the laboratory, and

additions or corrections to the data package shall be controlled by a document control form. This

review is designed to be conducted by personnel with training in, and a technical understanding

of, laboratory methods and data quality, but without the extensive experience required of

professional, trained data validators. Calculations of results from raw data will not be verified,

and data validation qualifiers will not be assigned at this review level. The results of this review

and a summary of parameter detections shall be forwarded to the appropriate subproject manager

for evaluation. The following items shall be checked:

• The laboratory sample identification corresponds to the client sample identification.

• There is a one-to-one correlation of laboratory sample numbers with client samplenumbers.

• Each QC sample was assigned a unique laboratory sample number.

• The correct sample matrix was identified for all the sample results.

• All discrepancies identified in the Data Completeness review have been accounted for.

• Critical items meet the Industrial Sites Project requirements.

Page 70: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 43 of 55

• Preparation and analysis dates and analytical batch numbers were assigned to each batch.

• Preparation dates and QC batch numbers were assigned to each QC batch.

• It is clear which QC batches correspond to which analytical batches.

• The appropriate units were reported for the environmental and QC sample results.

• Sample dilutions are properly noted.

• Sample detection limits were properly adjusted for dilutions.

• Detection limits meet Industrial Sites Project requirements.

• Laboratory data qualifiers are correct and explained or a key is included.

• Blanks are clean or flagged by the laboratory.

• QC samples have been analyzed for appropriate analytes, and results are within the limits.

• Laboratory control and matrix spike samples have been analyzed for the appropriateanalytes.

• The correct analytical method was used for the analysis.

5.4.3.3 Data Validation

Data validation encompasses a complete validation of the analytical results according to EPA

functional guidelines or an equivalent industry-standard protocol. Data validation and review of

CLP and CLP-like data packages shall be performed in accordance with the Laboratory Data

Validation Functional Guidelines for Evaluating Inorganic Analyses (EPA, 1994a) and

Laboratory Data Validation Functional Guidelines for Evaluation Organic Analyses

(EPA, 1994b). For non-CLP parameters, the data validation shall be performed following the

requirements outlined in DOE Standard Operating Procedures (SOPs), which are based on the

Hazardous Waste Remedial Action Program, Requirements for Quality Control of Analytical

Data (DOE, 1988). Data validation shall include a check of the calculation of all QC sample

results and the third party confirmation of a minimum of 5 percent (based on direction from the

Radioactive Waste Acceptance Program) of the sample result calculations from characterization

samples or samples intended to demonstrate that the contaminant(s) of concern have been

isolated, stabilized, and/or removed. Data validation shall also include a check of all the

functional guideline parameters not included in lower level reviews. Summary results from

previous reviews shall be provided as part of the data package transferred from designated site

Page 71: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 44 of 55

personnel to the third-party validation organization. The results of lower-level reviews shall be

used to minimize redundancy and duplication of effort.

The percentage of data packages to be validated for the Industrial Sites Project shall be dependent

on the end use of the data and established during the site-specific DQO process. Sample results

selected for validation shall be determined by use of a random number generator or may be

selected by project management in cases where special criteria exist. The Industrial Sites Project

Manager shall maintain the option of having additional data packages reviewed.

5.4.4 Laboratory Data Reporting

Analytical data reports must contain, at a minimum, the following information:

• Cover page with the reviewer's signature, data qualifiers, and a description of anytechnical difficulties encountered during the analyses

• Date the sample was received

• Date the sample was prepared

• Date the sample was analyzed

• Sample identification number

• Laboratory sample identification number

• Analytical method reference number

• Analytical results

• Tabulated QC sample results

• Instrument tuning and calibration results

• Final copy of the COC form, with appropriate signatures

Data packages shall be required for all analytical results unless sample results are excluded from

data validation by NV ERP project management. Validated data shall be reviewed to determine

whether they meet the DQOs of the investigation. The data shall be reviewed to ensure that the

required number of samples were collected, critical samples were collected and analyzed, and the

results passed data validation criteria. The data shall also be reviewed to determine whether

detection limits were met. Data reporting techniques shall be in accordance with the project data

Page 72: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 5.0Revision: 3Date: 02/07/2002Page 45 of 55

reporting requirements; data reporting procedures shall be consistent with those found in the

User’s Guide to the Contract Laboratory Program (EPA, 1991).

5.4.4.1 Data Reporting

Data shall be reported in accordance with standardized formats. Electronic data transfers shall be

delivered, along with the hard copy, on 3.5-inch diskettes or other methods agreed upon with the

NV ERP Common Data Repository custodial organization.

5.5 Data Quality Assessment

A Data Quality Assessment (DQA) shall be performed to determine if the DQOs have been

satisfied. The DQA is built on the fundamental premise that data quality is meaningful only when

it relates to the intended use of the data. The context in which a data set is to be used must be

known in order to establish a relevant yard stick for judging whether or not the data set is

adequate. The DQA process should answer the fundamental question of whether the quality of

the data set enables a decision (or estimate) to be made with the desired confidence.

For example, if the data provide evidence strongly in favor of one course of action over another,

then the decision maker can proceed knowing that the decision will be supported by unambiguous

data. However, if the data do not show sufficiently strong evidence to favor one alternative, then

the data analysis alerts the decision maker to this uncertainty. The decision maker now is in a

position to make an informed choice about how to proceed (such as collect more or different data

before making the decision, or proceed with the decision despite the relatively high, but

acceptable, probability of drawing an erroneous conclusion).

The DQA shall be documented and included in the appropriate decision document. Information

on DQA can be found in EPA QA/G-9, Guidance for Data Quality Assessments: Practical

Methods for Data Analysis (EPA, 2000a).

Page 73: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 6.0Revision: 3Date: 02/07/2002Page 46 of 55

6.0 Criteria 6 - Design

Engineering designs in support of the Industrial Sites Project shall be in accordance with a

documented design control process and based on sound engineering and scientific principles using

the appropriate standards. The acceptability and adequacy of the design product shall be verified

or validated by qualified individual(s) other than those who performed the original design.

Verification and validation shall be completed prior to approval and implementation of the design.

Design records shall include the design steps and sources of input that support the final output.

The final design output shall be approved in accordance with the participants’ internal procedures.

Changes or modifications to the final design shall be subject to the same control measures and

approvals as applied to the original design.

Page 74: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 7.0Revision: 3Date: 02/07/2002Page 47 of 55

7.0 Criteria 7 - Procurement

Procurement of items and services for the Industrial Sites Project shall be consistent with standard

commercial purchase order terms and conditions, and performed in cooperation with the

NNSA/NV Contracts Management Division. Project participants must have processes in place

that meet the requirements of their contracts or agreements and applicable federal requirements.

7.1 Procurement Control

Items and services of a technical nature procured in support of the Industrial Sites Project shall be

of a quality that meets the requirements of the project. Project participants shall establish controls

to ensure that, as a minimum, procured items and services meet specifications delineated in the

procurement documents. Each participating organization shall have systems in place to track

items and confirm the delivery of procured items and services as specified. Project participants

shall have a program in place, invoking the appropriate quality requirements of the contractor’s

QA program and specifying any project requirements for the procurement of items and services.

Subcontractors procured for Industrial Sites activities must be evaluated for prior experience,

ability to perform specific tasks, and cost. The capabilities of subcontractor personnel shall be

assessed by the procuring contractor to verify qualifications and determine the type and amount of

training and supervision needed for environmental restoration activities.

7.1.1 Procurement Documents

Procurement documents for the Industrial Sites Project shall define the scope of work for the item

or service being procured and provide specifications, acceptance criteria, shipping and handling

requirements, health and safety requirements, and any documentation required. Technical

specifications shall either be directly included in the procurement documents or included by

reference to specific drawings, specifications, procedures, regulations, or codes that describe the

items or services to be furnished. Procurement documents shall be reviewed for accuracy and

completeness by qualified personnel prior to initial issue. Changes to a procurement document

require the same level of review and approval as the original document.

Page 75: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 7.0Revision: 3Date: 02/07/2002Page 48 of 55

7.1.2 Measurement and Testing Equipment

Procurement documents shall also require that all purchased and rented M&TE be calibrated to

existing national standards prior to acceptance and that calibration documentation is provided.

Calibration certification and instrument manufacturer’s manuals should be available in project files

for M&TE. Schedules for recalibration shall be established and implemented for M&TE requiring

periodic calibration.

7.1.3 Verification of Quality Conformance

If applicable, procurement documents for Industrial Sites Project-related items or services shall

require access to the subcontractor’s or vendor’s facilities, including their subtier facilities, work

areas, and records for assessments to verify acceptability. Upon delivery, procured items or

services shall be inspected for conformance to procurement specifications and requirements prior

to using items or placing them in service. Project personnel have the authority to stop work if

significant quality problems are identified. Procured items should be evaluated for suspect/

counterfeit parts. If there are indications that suppliers knowingly supplied substandard items or

services, the DOE Office of Inspector General shall be notified.

7.2 Procurement of Laboratory Services

During the procurement process, the organization contracting the laboratory must verify that the

laboratory has procedures in place that are appropriate for the analysis to be performed.

Procedures shall be required for the following laboratory activities:

� Sample Receipt and Handling - Describes the precautions to be used when receivingsample shipment containers, instruct how to verify that chain-of-custody has beenmaintained, examine samples for damage, check for proper preservatives and temperature,and log samples into the laboratory sample tracking system.

� Sample Scheduling - Describe the sample scheduling in the laboratory and include theprocess used to ensure that holding time requirements are met.

� Sample Storage - Prescribe storage conditions for all samples, verify and document dailystorage temperature, and describe how to ensure that custody of the samples is maintainedwhile in the laboratory.

� Reagent/Standard Preparation - Describe how to prepare, handle, and use standards andreagents. Information should include specific grades of materials used in reagent andstandard preparation, appropriate glassware and containers for preparation and storage,and labeling and recordkeeping for standards and dilutions.

Page 76: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 7.0Revision: 3Date: 02/07/2002Page 49 of 55

� Test Methods - Test methods describing how the analyses are actually performed in thelaboratory.

� Sample Preparation and Analysis Procedures - These procedures include applicableholding times, extraction, digestion, or preparation steps as appropriate to the method;procedures for determining the appropriate dilution to analyze; and any other informationrequired to perform the analysis accurately and consistently.

� Instrument Standardization - This includes concentration(s) and frequency of analysis ofcalibration standards, linear range of the method, and calibration acceptance criteria.

� Sample Data - Directions on recording requirements and documentation including sampleidentification number, analyst, data verification, date of analysis and verification, andcomputational method(s).

� Precision and Bias - This procedure includes all analytes for which the method isapplicable and the conditions for use of this information.

� Test-Specific QC - The QC activities applicable to the specific test and references to otherapplicable QC procedures.

� Equipment Calibration and Maintenance - These procedures describe how to ensure thatlaboratory equipment and instrumentation are in working order. Information to beprovided includes: calibration methods and schedules, maintenance and schedules,maintenance logs, service arrangements for all equipment, and spare parts available in-house. Calibration and maintenance of laboratory equipment and instrumentation shouldbe in accordance with manufacturers’ specifications or applicable test specifications.

� QC - The type, purpose, and frequency of QC samples to be analyzed in the laboratoryand the acceptance criteria should be stated in procedures. Information should include theapplicability of the QC sample to the analytical process, the statistical treatment of thedata, and the responsibility of laboratory staff and management in generating and using thedata.

� Corrective Action - Systems shall be in place to identify and correct deficiencies in thelaboratory system. The procedures should specify how corrective action will be handledto include a description of the deficiency, the corrective action taken, verificationrequirements, and should include the person(s) responsible for implementing the correctiveaction.

� Data Reduction and Validation - Procedures should describe how to review and verify thedata, and provide instruction on how to compute and interpret the results from QCsamples to verify that the analytical results are reported correctly.

� Method Proficiency - Procedures should be in place for demonstrating proficiency witheach analytical method routinely used in the laboratory. All terminology, procedures, and

Page 77: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 7.0Revision: 3Date: 02/07/2002Page 50 of 55

frequency of determinations associated with the laboratory’s establishment of the methoddetection limit and the reporting limit should be well-defined and well-documented.

� Control Limits - Procedures should describe the establishment and updating of controllimits for analysis. This should include control charts and/or tabular presentations of data.

� Data Handling - Data resulting from the analyses should be reduced according toprotocols described in the laboratory procedures. Computer programs used for datareduction should be validated before use and verified on a regular basis.

Page 78: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 8.0Revision: 3Date: 02/07/2002Page 51 of 55

8.0 Criteria 8 - Inspection and Acceptance Testing

Inspections and acceptance testing shall be accomplished in accordance with approved inspection

documents and test procedures that reflect acceptance and performance criteria. Individuals

performing inspections and acceptance testing shall be independent of those who performed the

work. Quality-affecting materials used during characterization, corrective action, or sampling

activities shall be inspected upon receipt for adequacy. The M&TE used in the performance of

inspections or acceptance tests shall be calibrated and properly maintained. Any item or work

determined to be defective shall be controlled to avoid inadvertent use.

Page 79: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 9.0Revision: 3Date: 02/07/2002Page 52 of 55

9.0 Criteria 9 - Management Assessment

Planned and periodic assessments shall be conducted and shall involve the participation of

management at all levels. The primary emphasis of management assessments is to evaluate the

implementation of the integrated QA program and identify problems that hinder the achievement

of objectives. Contractor management should conduct periodic assessments that focus on such

issues as the:

• Adequacy of implementation of the integrated QA program, with particular emphasis onquality improvement

• Existence of any management biases or organizational barriers that impede theimprovement process

• Adequacy of the appraised organization's structure, staffing, and physical facilities

• Existence of effective training programs

The results of the assessment shall be documented in an assessment report and issued to the

appropriate managers. Senior management has the primary responsibility to ensure the timely

follow-up of corrective actions, including an evaluation of the effectiveness of management's

actions. Results of the management assessment should be entered into a tracking system to

ensure satisfactory closure.

Page 80: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 10.0Revision: 3Date: 02/07/2002Page 53 of 55

10.0 Criteria 10 - Independent Assessments

Independent management and technical assessments shall be performed to verify compliance with

applicable quality requirements, DOE policies, and procedures. Assessments shall be conducted

to measure item and service quality, the adequacy of work performance, and to promote

improvement. The scheduling of the assessments and resource allocation for independent

assessments should be based on the status, risk, and complexity of work being assessed.

The group performing the independent assessment shall be composed of individuals that are not

directly involved in the work being assessed. Each group performing independent assessments

shall have sufficient authority and freedom to carry out the activities necessary to effectively

conduct the assessment. Assessments should focus on improving the quality of the processes that

lead to the end product.

Results of each assessment should be tracked and resolved by responsible management with

follow-up of deficient areas. Assessment responses should include: corrective action,

identification of the root cause, actions to prevent recurrence, lessons learned, and actions for

improvement.

Page 81: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 11.0Revision: 3Date: 02/07/2002Page 54 of 55

11.0 References

American Society of Quality Control. 1994. American National Standard, Specifications andGuidelines for Quality Systems for Environmental Data Collection and EnvironmentalTechnology Programs, ANSI/ASQC E-4 1994. Milwaukee, WI.

ASQC , see American Society of Quality Control.

CFR, see Code of Federal Regulations.

Code of Federal Regulations. 2000. Title 49, Parts 171 through 177, "Hazardous MaterialsRegulations." Washington, DC.

Code of Federal Regulations. 2001a. Title 10, Part 71, “Packaging and Transportation ofRadioactive Material: Subpart H–Quality Assurance.” Washington, DC.

Code of Federal Regulations. 2001b. Title 10, Part 820, “Procedural Rules for DOE NuclearActivities.” Washington, DC.

Code of Federal Regulations. 2001c. Title 10, Section 830.120-122.. “Nuclear SafetyManagement: Subpart A–Quality Assurance Requirements.” Washington, DC.

Code of Federal Regulations. 2001d. Title 29, Part 1910.120, “Hazardous Waste Operationsand Emergency Response.” Washington, DC.

DOE, see U.S. Department of Energy.

EPA, see U.S. Environmental Protection Agency.

FFACO, see Federal Facility Agreement and Consent Order.

Federal Facility Agreement and Consent Order. 1996 (as amended). Agreed to by the State of Nevada, the U.S. Department of Energy, and the U.S. Department of Defense.

Nevada Administrative Code. 2000. NAC 444, 850-8746, “Disposal of Hazardous Waste.” Carson City, NV.

U.S. Department of Energy. 1988. Requirements for Quality Control of Analytical Data. Washington, DC.

U.S. Department of Energy. 1994. Analytical Laboratory Quality Assurance Guidance inSupport of EH Environmental Sampling and Analysis Activities, DOE/EM-159P. Washington, DC.

Page 82: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPSection: 11.0Revision: 3Date: 02/07/2002Page 55 of 55

U.S. Department of Energy. 1996a. Information Management Program, DOE O 200.1. Washington, DC.

U.S. Department of Energy. 1996b. Safety Management System Policy, DOE P 450.4. Washington, DC.

U.S. Department of Energy. 1998. Worker Protection Management for DOE Federal andContractor Employees, DOE O 440.1A. Washington, DC.

U.S. Department of Energy. 2001a. DOE O 414.1A, Quality Assurance. Washington, DC.

U.S. Department of Energy. 2001b. Waste Management, DOE O 435.1, CHG 1. Washington, DC.

U.S. Department of Energy, Nevada Operations Office. 1993. DOE/NV Family QualityGlossary. Las Vegas, NV.

U.S. Department of Energy, Nevada Operations Office. 2000a. Document Review and

Coordination, Rev. 1, DOE/NV EM-02-002. Las Vegas, NV.

U.S. Department of Energy, Nevada Operations Office. 2000b. Nevada Test Site WasteAcceptance Criteria, NTSWAC, DOE/NV--325–Rev. 3. Las Vegas, NV.

U.S. Environmental Protection Agency. 1991. User’s Guide to the Contract LaboratoryProgram, EPA/540/P-91/002. Washington, DC.

U.S. Environmental Protection Agency. 1994a. Laboratory Data Validation FunctionalGuidelines for Evaluating Inorganics Analyses, EPA/540/R-94/083. Washington, DC.

U.S. Environmental Protection Agency. 1994b. Laboratory Data Validation FunctionalGuidelines for Evaluating Organic Analyses, EPA/540/R-94/082. Washington, DC.

U.S. Environmental Protection Agency. 1998. EPA Guidance for Quality Assurance ProjectPlans, EPA QA/G-5, EPA/600/R-98/018. Washington, DC.

U.S. Environmental Protection Agency. 1999. EPA Requirements for Quality Assurance ProjectPlans, EPA QA/R-5. Interim Final. Washington, DC.

U.S. Environmental Protection Agency. 2000a. Guidance for Data Quality Assessments:Practical Methods for Data Analysis, EPA/QA/G-9. Washington, DC.

U.S. Environmental Protection Agency. 2000b. Guidance for the Data Quality ObjectivesProcess, EPA QA/G-4, EPA/600/R-96/055. Washington, DC.

Page 83: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Appendix A

Federal Facility Agreement andConsent Order Outlines

(This information is included as provided in the

Federal Agreement and Consent Order [1996]; therefore, refer to the

FFACO for any references cited in this appendix.)

Page 84: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 85: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 86: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 87: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 88: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 89: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 90: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 91: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 92: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 93: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 94: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 95: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 96: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 97: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 98: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 99: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 100: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 101: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 102: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 103: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 104: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 105: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 106: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 107: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 108: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 109: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 110: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 111: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 112: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 113: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC
Page 114: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-1 of B-10

Table B.1-1General Chemical Analytical Requirements

(Page 1 of 4)

Parameter orAnalyte

Medium orMatrix

AnalyticalMethod

Minimum ReportingLimit Regulatory Limit

PrecisionRelative Percent Difference

(RPD)a

AccuracyPercent Recovery (%R)b

ORGANICS

Total Volatile Organic Compounds (VOCs)

Water8260Bc Analyte-specific estimated

quantitation limitsd NALab-specifice Lab-specifice

Soil Lab-specifice Lab-specifice

Toxicity Characteristic Leaching Procedure (TCLP) VOCs

Benzene

Aqueous 1311/8260Bc

0.050 mg/Ld 0.5 mg/Ld

Lab-specifice Lab-specifice

Carbon Tetrachloride

0.050 mg/Ld 0.5 mg/Ld

Chlorobenzene 0.050 mg/Ld 100 mg/Ld

Chloroform 0.050 mg/Ld 6 mg/Ld

1,2-Dichloroethane 0.050 mg/Ld 0.5 mg/Ld

1,1-Dichloroethene 0.050 mg/Ld 0.7 mg/Ld

Methyl Ethyl Ketone 0.050 mg/Ld 200 mg/Ld

Tetrachloroethene 0.050 mg/Ld 0.7 mg/Ld

Trichloroethene 0.050 mg/Ld 0.5 mg/Ld

Vinyl Chloride 0.050 mg/Ld 0.2 mg/Ld

Total Semivolatile Organic Compounds (SVOCs)

Water8270Cc Analyte-specific estimated

quantitation limitsd NALab-specifice Lab-specifice

Soil Lab-specifice Lab-specifice

TCLP SVOCs

o-Cresol

Aqueous 1311/8270Cc

0.10 mg/Ld 200 mg/Ld

Lab-specifice Lab-specifice

m-Cresol 0.10 mg/Ld 200 mg/Ld

p-Cresol 0.10 mg/Ld 200 mg/Ld

Cresol (total) 0.30 mg/Ld 200 mg/Ld

1,4-Dichloro-benzene

0.10 mg/Ld 7.5 mg/Ld

2,4-Dinitrotoluene 0.10 mg/Ld 0.13 mg/Ld

Page 115: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-2 of B-10

Hexachloro-benzene

Aqueous 1311/8270Cc

0.10 mg/Ld 0.13 mg/Ld

Lab-specifice Lab-specifice

Hexachloro-butadiene

0.10 mg/Ld 0.5 mg/Ld

Hexachloro-ethane

0.10 mg/Ld 3 mg/Ld

Nitrobenzene 0.10 mg/Ld 2 mg/Ld

Pentachloro-phenol 0.50 mg/Ld 100 mg/Ld

Pyridine 0.10 mg/Ld 5 mg/Ld

2,4,5-Trichloro-phenol

0.10 mg/Ld 400 mg/Ld

2,4,6-Trichloro-phenol

0.10 mg/Ld 2 mg/Ld

TotalPesticides

Water8081Ac Analyte-specific

(CRQL)f NALab-specifice Lab-specifice

Soil Lab-specifice Lab-specifice

TCLP Pesticides

Chlordane Aqueous 1311/8081Ac 0.0005 mg/Lf 0.03 mg/Ld Lab-specifice Lab-specifice

Endrin 0.001 mg/Lf 0.02 mg/Ld

Heptachlor 0.0005 mg/Lf 0.008 mg/Ld

HeptachlorEpoxide

0.0005 mg/Lf 0.008 mg/Ld

gamma-BHC (Lindane)

0.0005 mg/Lf 0.4 mg/Ld

Methoxychlor 0.005 mg/Lf 10 mg/Ld

TCLP Pesticides

Toxaphene 0.05 mg/Lf 0.5 mg/Ld

Polychlorinated Biphenyls (PCBs)Water

8082c Analyte-specific (CRQL)f NA Lab-specifice Lab-specifice

Soil

TotalHerbicides

Water8151Ac

1.3 µg/Lc

NA Lab-specifice Lab-specifice

Soil 66 µg/kgc

TCLP Herbicides

2,4-DAqueous 1311/8151Ac 0.002 mg/Ld 10 mg/Ld

Lab-specifice Lab-specifice

2,4,5-TP 0.00075 mg/Ld 1 mg/Ld

Table B.1-1General Chemical Analytical Requirements

(Page 2 of 4)

Parameter orAnalyte

Medium orMatrix

AnalyticalMethod

Minimum ReportingLimit Regulatory Limit

PrecisionRelative Percent Difference

(RPD)a

AccuracyPercent Recovery (%R)b

Page 116: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-3 of B-10

Total Petroleum Hydrocarbons (TPH)

WaterGasoline

8015B modifiedc

0.1 mg/L

NA Lab-specifice Lab-specificeSoil Gasoline 0.5 mg/kg

Water Diesel 0.5 mg/L

Soil Diesel 25 mg/kg

ExplosivesWater

8330c14 µg/Lc

NA Lab-specifice Lab-specifice

Soil 2.2 mg/kgc

Polychlorinated Dioxins and FuransWater

8280A/8290c 0.05 µg/Lc

NA Lab-specifice Lab-specifice

Soil 5 µg/kgc

INORGANICSTotal Resource Conservation and

Recovery Act (RCRA) Metals

ArsenicWater 6010Bc 10 µg/Lh

NAWaters - 20g

Soils - 35h

Matrix Spike75-125g

Laboratory Control Sample

80 - 120g

Soil 6010Bc 1 mg/kgh

BariumWater 6010Bc 200 µg/Lh

Soil 6010Bc 20 mg/kgh

CadmiumWater 6010Bc 5 µg/Lh

Soil 6010Bc 0.5 mg/kgh

ChromiumWater 6010Bc 10 µg/Lh

Soil 6010Bc 1 mg/kgh

LeadWater 6010Bc 3 µg/Lh

Soil 6010Bc 0.3 mg/kgh

MercuryWater 7470Ac 0.2 µg/Lh

Soil 7471Ac 0.1 mg/kgh

SeleniumWater 6010Bc 5 µg/Lh

Soil 6010Bc 0.5 mg/kgh

SilverWater 6010Bc 10 µg/Lh

Soil 6010Bc 1 mg/kgh

TCLP RCRAMetalsArsenic

Aqueous 1311/6010Bc 1311/7470Ac

0.10 mg/Lh 5 mg/Ld

20g

Matrix Spike75-125g

Laboratory Control Sample

80 - 120g

Barium 2 mg/Lh 100 mg/Ld

Cadmium 0.05 mg/Lh 1 mg/Ld

Chromium 0.10 mg/Lh 5 mg/Ld

Lead 0.03 mg/Lh 5 mg/Ld

Mercury 0.002 mg/Lh 0.2 mg/Ld

Selenium 0.05 mg/Lh 1 mg/Ld

Silver 0.10 mg/Lh 5 mg/Ld

Table B.1-1General Chemical Analytical Requirements

(Page 3 of 4)

Parameter orAnalyte

Medium orMatrix

AnalyticalMethod

Minimum ReportingLimit Regulatory Limit

PrecisionRelative Percent Difference

(RPD)a

AccuracyPercent Recovery (%R)b

Page 117: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-4 of B-10

Cyanide

Water

9010Bc

0.01 mg/Lg

NAWater - 20g

Soil - 35h

Matrix Spike75-125g

Laboratory Control Sample

80 - 120g

Soil 1.0 mg/kgh

SulfideWater

9030B/9034c

0.4 mg/Lc

NA Lab-specifice Lab-specificeSoil orSediment 10 mg/kgg

pH/CorrosivityWater 9040Bc

NApH >2 i

Lab-specifice Lab-specifice

Soil 9045Cc pH<12.5 i

Ignitability

Water 1010c

NA

Flash Point <140o Fd, i

NA NASoil 1030c

Burn Ratec, i

>2.2 mm/sec nonmetals;>0.17 mm/sec metals

a RPD is used to Calculate PrecisionPrecision is estimated from the relative percent difference of the concentrations measured for the matrix spike and matrix spike duplicate analyses of unspiked field samples, or field duplicates of unspiked samples. It is calculated by: RPD = 100 x {(|C1-C2|)/[(C1+C2)/2]}, where C1 = Concentration of the analyte in the first sample aliquot, C2 = Concentration of the analyte in the second sample aliquot.

b %R is used to Calculate AccuracyAccuracy is assessed from the recovery of analytes spiked into a blank or sample matrix of interest, or from the recovery of surrogate compounds spiked into each sample. The recovery of each spiked analyte is calculated by: %R = 100 x (Cs-Cu/Cn), where Cs = Concentration of the analyte in the spiked sample, Cu = Concentration of the analyte in the unspiked sample, Cn = Concentration increase that should result from spiking the sample

c EPA Test Methods for Evaluating Solid Waste, 3rd Edition, Parts 1-4, SW-846, CD ROM, Washington, DCd Estimated Quantitation Limit as given in SW-846 (EPA, 1996)e In-House Generated RPD and %R Performance Criteria

It is necessary for laboratories to develop in-house performance criteria and compare them to those in the methods. The laboratory begins by analyzing 15-20 samples of each matrix and calculating the mean %R for each analyte. The standard deviation (SD) of each %R is then calculated, and the warning and control limits for each analyte are established at ± 2 SD and ± 3 SD from the mean, respectively. If the warning limit is exceeded during the analysis of any sample delivery group (SDG), the laboratory institutes corrective action to bring the analytical system back into control. If the control limit is exceeded, the sample results for that SDG are considered unacceptable. These limits are reviewed after every 20-30 field samples of the same matrix and are updated at least semiannually. The laboratory tracks trends in both performance and control limits by the use of control charts. The laboratory’s compliance with these requirements is confirmed as part of an annual laboratory audit. Similar procedures are followed in order to generate acceptance criteria for precision measurements

f EPA Contract Laboratory Program Statement of Work for Organic Analysis (EPA, 1988b; 1991; and 1994b)g EPA Contract Laboratory Program Statement of Work for Inorganic Analysis (EPA, 1988a and 1994a)h Sampling and Analysis Plan (Field Sampling and Quality Assurance Project Plan) with guidance EPA Region IX (EPA, 2000).i RCRA Regulations and Keyword Index, 2000

Definitions:µg/kg = Microgram(s) per kilogrammg/kg = Milligram(s) per kilogrampCi/L = Picocurie(s) per litermg/L = Milligram(s) per literµg/L = Microgram(s) per liter

CRQL = Contract-required quantitation limitNA = Not applicable°F = Degrees Fahrenheitmm/sec = Millimeters per secon

Table B.1-1General Chemical Analytical Requirements

(Page 4 of 4)

Parameter orAnalyte

Medium orMatrix

AnalyticalMethod

Minimum ReportingLimit Regulatory Limit

PrecisionRelative Percent Difference

(RPD)a

AccuracyPercent Recovery (%R)b

Page 118: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-5 of B-10

Table B.1-2General Radiological Analytical Requirements

Parameter orAnalyte

Medium orMatrix

AnalyticalMethod

Minimum ReportingLimit Regulatory Limit

PrecisionRelative Percent Difference (RPD)a

AccuracyPercent

Recovery (%R)b

RADIOCHEMISTRY

Gamma-emitting Radionuclidesc

Water EPA 901.1d 10 pCi/L (Cs-137)f

NA20 Laboratory Control

Sample Yield 80-120iSoil HASL-300e 2.14 pCi/g (Cs-137) 35

Isotopic Plutoniumc

Water ASTM D3865-97g 0.1 pCi/LNA

20

Chemical Yield30-105i

Laboratory Control Sample 80 - 120i

Soil HASL-300e 0.05 pCi/g 35

Isotopic Uraniumc

WaterHASL-300e

ASTM D3972-97g 0.1 pCi/L

NA

20

SoilHASL-300e

ASTM E1000-90g 0.05 pCi/g 35

Strontium - 90cWater ASTM D5811-95g 1.0 pCi/L

NA20

Soil HASL-300e 0.5 pCi/g 35

Americium - 241cWater ASTM-1205-97g 0.1 pCi/L NA 20

Soil HASL-300e 0.05 pCi/ge NA 35

TritiumcWater EPA 906.0d 400

NA20 Laboratory Control

Sample Yield 80-120qSoil (Sludge) PAI 754h 5 pCi/g 20

a RPD is used to calculate precision.Precision is estimated from the relative percent difference of the concentrations measured for the matrix spike and matrix spike duplicate analyses of unspiked field samples, or field duplicates of unspiked samples. It is calculated by: RPD = 100 x {(|C1-C2|)/[(C1+C2)/2]}, where C1 = Concentration of the analyte in the first sample aliquot, C2 = Concentration of the analyte in the second sample aliquot.

b %R is used to calculate accuracy.Accuracy is assessed from the recovery of analytes spiked into a blank or sample matrix of interest, or from the recovery of surrogate compounds spiked into each sample. The recovery of each spiked analyte is calculated by: %R = 100 x (Cs-Cu/Cn), where Cs = Concentration of the analyte in the spiked sample,Cu = Concentration of the analyte in the unspiked sample, Cn = Concentration increase that should result from spiking the sample

c Isotopic minimum detectable concentrations are defined during the DQO process and specified in the CAIP as applicabled Prescribed Procedures for Measurements of Radioactivity in Drinking Water, EPA 901, 901.1, 906, and 908.0 (EPA, 1980)e Manual of Environmental Measurements Laboratory Procedures, HASL-300 (DOE, 1997)f Detection limit is based upon method and isotope. Isotope-specific minimum reporting limit to be specified in CAIP.g American Society for Testing and Materialsh PAI - Paragon Analytics, Inc.i U.S. EPA Laboratory Program, National Functional Guidelines, EPA 540/R-94/013 (EPA, 1994a)

Page 119: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-6 of B-10

Table B.1-3Waste Characterization Chemical Requirementsa

(Page 1 of 3)

Parameter Method Medium Preservative Holding Time TCLP Analyte

Minimum Detectable

Concentration (mg/L)

Regulatory Limit(mg/L)

TCLPb Volatile

Organics1311/

8240Bc

WaterpH < 2 with HCl,

Cool to 4�C

14 days to TCLP extraction; 14 days from extraction to

analysis

BenzeneCarbon TetrachlorideChlorobenzeneChloroform1,2-Dichloroethane1,1-DichloroethyleneMethyl Ethyl KetoneTetrachloroethyleneTrichloroethyleneVinyl Chloride

0.0500.0500.0500.0500.0500.0500.200.0500.0500.100

0.50.5

100.06.00.50.7

200.00.70.50.2

Soil Cool to 4�C

TCLP Semivolatile Organics

1311/8270B

c

Water

Cool to 4�C

14 days to TCLP extraction; 7 days from TCLP extraction to

preparative extraction; 40 days from preparative extraction to

analysis

o-Cresolm-Cresolp-CresolCresol (total)1,4-Dichlorobenzene2,4-DinitrotolueneHexachlorobenzeneHexachlorobutadieneHexachloroethaneNitrobenzenePentachlorophenolPyridine2,4,5-Trichlorophenol2,4,6-Trichlorophenol

0.0500.0500.0500.1500.0500.0500.0500.0500.0500.0500.2500.1000.0500.050

200.0200.0200.0200.07.50.130.130.53.02.0100.05.0400.02.0

Soil

TCLP Metals1311/6010, 1311/7000

seriesc

WaterpH < 2, HNO3Cool to 4�C

180 daysd

or28 daysd for Mercury

AsBaCdCrPbHgSeAg

0.3000.2000.0050.0100.1000.00020.2500.010

5.0100.01.05.05.00.201.05.0

Soil Cool to 4�C

Page 120: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-7 of B-10

TCLP Pesticides1311/

8080Ac

Water

Cool to 4�C

14 days to TCLP extraction; 7 days from TCLP extraction to

preparative extraction; 40 days from preparative extraction to

analysis

ChlordaneEndrinHeptachlorHeptachlor epoxideLindaneMethoxychlorToxaphene

0.0050.00050.00050.00050.00050.0010.020

0.030.020.0080.0080.410.00.5

Soil

TCLP Herbicides1311/

8150Bc

Water

Cool to 4�C

14 days to TCLP extraction; 7 days from TCLP extraction to

preparative extraction; 40 days from preparative extraction to

analysis

2,4-D2,4,5-TP

0.120.017

10.01.0Soil

Ignitability 1010c Water

Cool to 4�C 180 days NA NAFlash Point< 140 �FSoil

Corrosivity9040

cWater

Cool to 4�C Immediate NA NApH > 2pH < 12.59045

cSoil

Cyanide 9010c Water

pH > 12 with NaOH. Cool to

4�C 14 days NA0.005

NA

Soil Cool to 4�C 0.5 mg/kg

Table B.1-3Waste Characterization Chemical Requirementsa

(Page 2 of 3)

Parameter Method Medium Preservative Holding Time TCLP Analyte

Minimum Detectable

Concentration (mg/L)

Regulatory Limit(mg/L)

Page 121: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-8 of B-10

Sulfide 9030c

Water

pH > 9 with NaOH. 4 drops

2N zinc acetate/100 mL.

Cool to 4�C. 7 days NA

1

NA

SoilWet surface with 2N zinc acetate.

Cool to 4�C10 mg/kg

Free Liquids Paint Filter Test

9095c

Soil Cool to 4�C NA NA NANo free liquid

present

aOther parameters may be required based on knowledge of the Corrective Action Site.

bToxicity Characteristic Leaching Procedure

cEPA Test Methods for Evaluating Solid Waste, 3rd Edition, Part 1-4, SW-846 (EPA, 1996)

dHolding time from collection to TCLP extraction is 180 days and from extraction to analysis is also 180 days (360 days total elapsed time). For mercury, corresponding holding

times are 28 and 28 days, respectively.

G - GlassNA - Not applicableHCI - Hydrochloric acidNaOH - Sodium hydroxideHNO3 - Nitric acidmg/kg - Milligrams per kilogramPE - Polyethylenemg/L - Milligrams per litermL - Milliliter�C - Degrees Celsius

Table B.1-3Waste Characterization Chemical Requirementsa

(Page 3 of 3)

Parameter Method Medium Preservative Holding Time TCLP Analyte

Minimum Detectable

Concentration (mg/L)

Regulatory Limit(mg/L)

Page 122: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-9 of B-10

Table B.1-4Waste Characterization Radiological Requirements

Parameter Method MediumHolding

TimePreservative

Minimum DetectableConcentration

Gross AlphaEPA 900.0a Water 6 months HN03 to pH<2 3 pCi/L

SM 7110b

Soil 6 months None 3 pCi/g

Gross BetaEPA 900.0 Water 6 months HNO3 to pH<2 4 pCi/L

SM 7110 Soil 6 months None 4 pCi/g

Gamma Spectroscopy

(based on 137Cs

EPA 901.1a

Water 6 months HNO3 to pH<2 10 pCi/L

HASL 300, 4.5.2.3

c Soil 6 months None 0.50 pCi/g

Tritium

EPA 906.0a

Water

6 months None

400 pCi/L

PAI 704/754d Soil/

Sediment1 pCi/g

Sr-90EPA 905.0 Water 6 months HNO3 to pH<2 1.0 pCi/L

HASL Sr-02 Soil 6 months None 0.50 pCi/g

Isotopic Pu

HASL 300Pu-10

c Water 6 months HNO3 to pH<2 0.10 pCi/L

HASL 300Pu-02

Soil 6 months None 0.05 pCi/g

aPrescribed Procedures for Measurement of Radioactivity in Drinking Water, U.S. Environmental Protection Agency (EPA, 1980)

bStandard Methods for the Examination of Water and Wastewater, American Public Health Association, 1992cEnvironmental Measurements Laboratory Procedure Manual, HASL-300 (DOE, 1997)dParagon Analytical Laboratories, Inc.

G - GlassHNO3 - Nitric acidoz - OuncepCi/g - Picocuries per grampCi/L - Picocuries per literPE - Polyethylene137Cs - Cesium

Page 123: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPAppendix BRevision: 3Date: 02/07/2002Page B-10 of B-10

References

American Public Health Association. 1992. Standard Methods for the Examination of Water and Wastewater, 18th Edition. Washington, DC.

RCRA Regulations and Keyword Index. 2000. NAC444.850-8746, “Disposal of Hazardous Waste.” Carson City, NV.

U.S. Department of Energy. 1997. Environmental Measurements Laboratory Procedures Manual, HASL-300, 28th Ed., Vol. I. New York, NY.

U.S. Environmental Protection Agency. 1980. Prescribed Procedures for Measurement of Radioactivity in Drinking Water, EPA 600/4-80-032 (NTIS/PB80-224744; CD ROM). Washington, DC.

U.S. Environmental Protection Agency. 1988a. Contract Laboratory Program Statement of Work for Inorganic Analysis, SOW No. 788, EPA/540/R-94/093. Washington, DC.

U.S. Environmental Protection Agency. 1988b. Contract Laboratory Program Statement of Work for Organic Analysis, SOW No. 2/88, EPA/540/R-94/096. Washington, DC.

U.S. Environmental Protection Agency. 1991. Contract Laboratory Program Statement of Work for Organic Analysis, OLMO 1.8, EPA/540/R-94/078. Washington, DC.

U.S. Environmental Protection Agency. 1994a. Contract Laboratory Program National Functional Guidelines for Inorganic Data Review, EPA/540/R-94/013. Washington, DC.

U.S. Environmental Protection Agency. 1994b. Contract Laboratory Program Statement of Work for Inorganic Analysis, ILMO 3.0, EPA/540/R-94/076. Washington, DC.

U.S. Environmental Protection Agency. 1996. Test Method for Evacuating Solid Waste Physical/Chemical Methods, SW-846, 3rd Edition, CD-ROM, PB97-501928GEI. Washington, DC.

U.S. Environmental Protection Agency. 2000. Memorandum from S.J. Smucker to PRG table mailing list regarding Region IX Preliminary Remediation Goals (PRGs), 1 August. San Francisco, CA.

Page 124: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPDistributionRevision: 3Date: 02/07/2002Page 1 of 4

Distribution

Copies

Janet Appenzeller-Wing 1 (Controlled)Environmental Restoration DivisionU.S. Department of Energy National Nuclear Security Administration Nevada Operations OfficeP.O. Box 98518, M/S 505Las Vegas, NV 89193-8518

Sabine Curtis 1 (Controlled)Environmental Restoration DivisionU.S. Department of Energy National Nuclear Security Administration Nevada Operations OfficeP.O. Box 98518, M/S 505Las Vegas, NV 89193-8518

Sean Kosinski 1 (Controlled)Environmental Restoration DivisionU.S. Department of Energy National Nuclear Security Administration Nevada Operations OfficeP.O. Box 98518, M/S 505Las Vegas, Nevada 89193-8518

Kevin Cabble 1 (Controlled)U.S. Department of Energy National Nuclear Security Administration Nevada Operations OfficeP.O. Box 98518, M/S 505Las Vegas, Nevada 89193-8518

Diana Hovey-Spencer 1 (Controlled)Desert Research InstituteP.O. Box 19040, M/S 433Las Vegas, Nevada 89132

June Simms 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Page 125: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPDistributionRevision: 3Date: 02/07/2002Page 2 of 4

Copies

Dustin Wilson 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Rick Deshler 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Jeanne Wightman 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Candice Fillmore 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Linda Linden 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Sue Zvoda 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Jill Dale 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Bob Sobocinski 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

John Davis 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Page 126: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPDistributionRevision: 3Date: 02/07/2002Page 3 of 4

Copies

Jeff Johnson 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Lynn Kidman 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Robert McCall 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

John Fowler 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Steve Adams 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Grant Evenson 1 (Controlled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Central Files 1 (Uncontrolled)IT Corporation, M/S 439P.O. Box 93838Las Vegas, Nevada 89193-3838

Manager, Southern Nevada FFACO 1 (Controlled)*Public Reading Room 1 (Uncontrolled)*P.O. Box 98521, M/S NLV040Las Vegas, NV 89193-8521

Page 127: Industrial Sites Quality Assurance Project Plan Library... · following: audit, performance evaluation, management systems review, peer review, inspection, or surveillance. (ANSI/ASQC

Industrial Sites QAPPDistributionRevision: 3Date: 02/07/2002Page 4 of 4

Copies

Manager, Northern Nevada FFACO 1 (Uncontrolled)*Public Reading RoomNevada State Library and Archives Federal Publications100 North Stewart StreetCarson City, NV 89701-4285

Technical Information Resource Center 1 (Uncontrolled)U.S. Department of Energy National Nuclear Security Administration Nevada Operations OfficeP.O. Box 98518, M/S 505Las Vegas, NV 89193-8518

U.S. Department of Energy 1 (Uncontrolled, electronic copy)Office of Scientific and Technical InformationP.O. Box 62Oak Ridge, TN 37831