8.1 Metrology and Measurement
8.1 Metrology and Measurement
1. Measurement Principles
Definition of Measurement:
Quantitative comparison of unknown quantity with predefined standard.
Process of obtaining magnitude of physical quantity.
Fundamental Elements:
Measurand: Quantity to be measured.
Standard: Reference for comparison.
Instrument: Device for comparison.
Observer: Person/system performing measurement.
Basic Requirements:
Standardization: Using accepted references.
Traceability: Chain of comparisons to national/international standards.
Reproducibility: Same results under same conditions.
Measurement Methods:
Direct: Value obtained directly from instrument.
Indirect: Value calculated from other measurements.
Comparative: Comparison with known standard.
Null Method: Balancing to achieve zero difference.
2. Errors and Calibration
Error Types:
Systematic Errors:
Consistent, predictable deviations.
Due to instrument defects, environmental factors, observer bias.
Correctable through calibration.
Random Errors:
Unpredictable, statistical variations.
Due to unknown causes, environmental fluctuations.
Reduced by averaging multiple readings.
Gross Errors:
Large, obvious mistakes.
Due to misreading, incorrect recording.
Eliminated by careful technique.
Calibration Process:
Comparing instrument with higher accuracy standard.
Establishing correction factors.
Documenting instrument performance.
Calibration Terms:
Standards Hierarchy: Primary → Secondary → Working standards.
Calibration Interval: Time between calibrations.
Calibration Certificate: Documented results.
Error Reduction Techniques:
Regular calibration.
Environmental control.
Proper training.
Statistical analysis.
3. Accuracy and Precision
Accuracy:
Closeness to true value.
Measures correctness.
Affected by systematic errors.
Precision:
Repeatability of measurements.
Measures consistency.
Affected by random errors.
Key Differences:
High Accuracy, Low Precision: Average correct but scattered.
Low Accuracy, High Precision: Consistent but offset.
High Accuracy, High Precision: Ideal measurement.
Low Accuracy, Low Precision: Poor measurement.
Related Terms:
Resolution: Smallest detectable change.
Sensitivity: Output change per input change.
Linearity: Constant sensitivity over range.
Hysteresis: Different readings for increasing/decreasing inputs.
Quantification:
Accuracy: Often expressed as ± percentage of full scale.
Precision: Measured by standard deviation.
4. Inspection and Acceptance Testing
Inspection Purpose:
Verify conformance to specifications.
Ensure quality standards.
Prevent defective products.
Inspection Types:
100% Inspection: Every item checked.
Sampling Inspection: Statistical sample checked.
First-article Inspection: First production item.
In-process Inspection: During manufacturing.
Final Inspection: Before shipment.
Acceptance Testing:
Formal verification against requirements.
Often involves destructive/non-destructive testing.
Decision to accept/reject batch.
Statistical Quality Control:
Control charts for process monitoring.
Acceptance sampling plans.
AQL (Acceptable Quality Level) determination.
Measurement Uncertainty:
All measurements have associated uncertainty.
Must be reported with measurement results.
Affects acceptance decisions.
Documentation:
Inspection reports.
Test certificates.
Non-conformance reports.
Corrective action records.
Last updated