Dust / Particulate Monitor (Optical Scattering AFE & Node)
← Back to: Industrial Sensing & Process Control
This page explains how to build a predictable optical-scattering dust monitor: a stable light source, low-noise AFE, temperature-aware calibration, and evidence-driven edge processing so PM readings stay trustworthy in real deployments.
It also provides a practical validation chain (factory → production test → field diagnostics) so drift, saturation, and interference can be traced to measurable signals and fixed fast.
Center idea: make an optical-scattering dust monitor predictable
An optical-scattering dust monitor becomes reliable only when the system is designed as an evidence chain: stable light source → low-noise AFE → temperature-aware calibration → edge filtering that preserves truth → wireless reporting that is traceable.
Definition of “done” (mechanically verifiable evidence fields)
- Baseline is stable and explainable: both dark baseline (light off/blocked) and ambient baseline (light on with scatter removed or equivalent) remain bounded, and their drift over temperature is modeled instead of hand-waved.
- Response is repeatable: repeated exposures to a reference aerosol (or a controlled surrogate standard) produce consistent outputs across days and across temperature—without hidden saturation.
- Telemetry is sufficient to reproduce any reported number: raw photodetector current (or proxy), TIA output, light-source current (set & measured), temperature, optional fan/flow status, plus a fault bitmask and the current gain/baseline state.
Why these fields matter: most “mysterious drift” is not dust—it is baseline drift, gain state changes, or light-source instability. If the log can show what the AFE saw and what state the system was in, troubleshooting becomes deterministic.
Use this diagram as a reference architecture for optical-scattering dust monitors (laser/LED drive → PD/TIA/ADC → edge MCU → wireless), including the minimum telemetry fields.
Measurement targets and constraints (PM1/PM2.5/PM10, dynamics, accuracy)
In optical-scattering designs, specification defines architecture. If targets are vague (“measure PM2.5”), the design will drift into untestable assumptions. This chapter converts PM targets into hard constraints on dynamic range, time response, and ambient rejection—each tied to measurable evidence.
2.1 Decide the outward-facing output (what the product promises)
- Mass bins: PM1 / PM2.5 / PM10 as reported values (what users will trend and compare).
- Event vs baseline behavior: whether short “dust bursts” must be preserved as events or smoothed into a stable trend.
- Health / self-diagnosis: whether the device must expose sensor health (contamination suspicion, saturation frequency, light-source stability).
A practical rule: use two time constants—a slow window for stable reporting and a fast channel for diagnostics/events. This prevents “fixing noise” by hiding real dust dynamics.
2.2 Dynamic range: treat saturation as a state, not a surprise
- Low end (clean air): noise floor must be below the smallest meaningful change in background; otherwise PM1/PM2.5 becomes dominated by baseline jitter.
- High end (heavy dust): define what happens when the signal grows beyond headroom: auto-ranging (gain steps), integration control (sampling strategy), or flagged clipping (explicit saturation reporting).
- Non-negotiable evidence: saturation must set a loggable flag and preserve the gain state so later analysis can separate “true dust” from “measurement ceiling.”
2.3 Time response: reporting latency is an engineering choice
- 1–3 s class: reacts quickly to bursts but is sensitive to EMI, wireless TX artifacts, and ADC/reference perturbations.
- 5–10 s class: produces a calm trend line, but can hide short dust events unless a parallel fast channel exists.
- Design implication (evidence-based): the chosen window must be visible in telemetry (window length, filter mode), so “firmware changes” cannot silently change product behavior.
2.4 Environmental constraints: map each stressor to a mandatory mitigation
- Temperature range: requires temperature-aware calibration (coefficients or table) and explicit baseline rules (freeze vs adapt).
- Ambient light: requires either modulation/synchronous sampling or a two-baseline strategy (dark + ambient) with documented recovery behavior.
- Humidity / contamination exposure: requires drift detection and maintenance triggers (baseline slope, saturation frequency, light-source efficiency proxy).
- Vibration / electrical noise: requires a coupling test plan (wireless TX on/off, supply ripple injection) with recorded deltas.
2.5 Hard mapping: targets → architecture → validation items
- Dynamic range → TIA gain strategy + ADC resolution + gain/saturation states logged.
- Time response → sampling rate + filtering window(s) + window mode visible in telemetry.
- Ambient light robustness → modulation/synchronous sampling OR baseline subtraction + sunlight/flicker stress test results recorded.
Acceptance mindset: if this mapping cannot be written clearly, later chapters (TIA design, temperature compensation, filtering) will turn into “tuning,” which is not production-stable engineering.
Use this map to translate PM targets into hard constraints (dynamic range, response time, ambient robustness) and to define the minimum evidence fields for validation.
Optical scattering signal chain overview (from photons to counts)
Optical-scattering monitors do not “measure particles directly.” They measure time-varying scattered light, translate it into a detector current, and then convert that electrical signal into features and counts that can be mapped to PM bins under a specific calibration version. This chapter defines where the signal is born, where it is corrupted, and which evidence fields keep the design testable.
3.1 From photons to electrical current: the three-component model
- Baseline component: detector dark current + light leakage + slow optical/aging drift. This component must be bounded and modeled, not “filtered away.”
- Event component: particle-driven fluctuations (pulses/variance changes). This component must survive filtering windows without being mistaken for drift.
- Corruption component: EMI pickup, rail ripple, wireless TX coupling, and sampling/alias artifacts. This component must be detected and labeled in logs.
A stable system always separates “baseline vs events vs corruption” using states and telemetry. If a reading cannot be reproduced from logged raw channels and states, it is not a measurement—it is a guess.
3.2 Noise injection points mapped to evidence fields
- Ambient light injection: track ambient baseline and recovery time after step changes.
- Laser intensity drift: correlate output drift with light current (set/measured) and temperature.
- Detector dark current drift: maintain a dark-baseline model vs temperature.
- TIA noise / headroom limits: estimate noise floor and expose saturation + gain range states.
- EMI pickup: run TX on/off A/B tests and record deltas, not anecdotes.
Practical mapping: “counts to PM” is an algorithmic mapping step; “photons to current” is physics and electronics. When these are mixed without states, field drift is often misdiagnosed as dust.
Use this block diagram to explain the optical-scattering signal path and the five major noise injection points (ambient light, laser drift, dark current, TIA noise, EMI pickup).
Light-source driver design (laser/LED drive, safety, modulation)
In scattering systems, the light source is part of the measurement instrument. Stability is achieved by choosing an appropriate drive mode (CW or modulated), enforcing constant-current behavior across temperature and aging, and exposing telemetry and fault states that make field behavior traceable.
4.1 Drive modes: CW vs modulated (selection rule, not preference)
- CW drive: simplest implementation, but more susceptible to ambient-light drift and slow baseline shifts.
- Modulated/pulsed drive: enables synchronous sampling and improves ambient rejection; requires timing alignment between modulation and ADC windows.
- Engineering rule: if operation includes strong ambient changes (sunlight, flickering indoor lighting) or low-concentration stability is required, prefer modulated drive with explicit sampling windows and logged mode states.
4.2 Constant-current driver requirements (must be defined and measured)
- Compliance headroom: insufficient headroom forces non-linear current behavior and amplifies temperature/aging effects. Evidence comes from I_set vs I_meas tracking at temperature extremes.
- Edge behavior (rise/fall, overshoot): edges that are too sharp inject EMI into the AFE; edges that are too slow reduce usable synchronous windows. Evidence comes from “light toggle” tests and TIA-out impulse signatures.
- Current monitor path: a measurable I_meas is mandatory. Without it, drift cannot be separated into optical vs electrical causes.
4.3 Safety and shutdown (system-level, traceable)
- Interlocks: lid/open-path, over-temperature, timeout. These must drive a hardware-visible shutdown.
- Fault shutdown: open/short detection and thermal faults must be latched or timestamped for field diagnosis.
- Current limit: a hard upper bound prevents software misconfiguration from damaging the light source.
Field reliability depends on fault transparency: if a light source turns off, the reason must be identifiable from telemetry (fault bit + timestamp + temperature + I_meas history).
4.4 Evidence captures that prevent “mystery drift”
- Optical power proxy vs temperature: use monitor photodiode or closed-loop proxy to show stability across temperature, not just current stability.
- Modulation vs sampling plan: define modulation frequency, ADC sampling rate, and window length as a coherent plan to avoid aliasing artifacts. Keep the plan visible in telemetry (Drive_mode, Mod_freq, Window_len).
This chapter intentionally focuses on signals and states (I_set/I_meas, mode, faults, interlocks). Power-stage topology details belong to dedicated driver-topology pages.
Use this diagram to specify a light-source driver for optical-scattering dust monitors: drive modes, current monitoring, safety interlocks, and the telemetry required for traceable field behavior.
Detector and analog front-end (PD/SiPM + TIA + filtering)
The analog front-end (AFE) turns detector current into digital features while preserving evidence: low noise at clean-air baseline, headroom during dust bursts, and traceable gain/offset calibration. AFE design is successful only when noise floor, saturation behavior, and state changes are observable and repeatable.
5.1 Detector choice tradeoffs: photodiode vs SiPM
- Photodiode (PD): predictable linear behavior and simpler biasing; performance depends heavily on TIA noise and leakage control.
- SiPM: effective gain improves weak-signal visibility; requires bias management and tighter temperature control due to stronger temperature-dependent behavior.
- Selection rule: PD favors stable baselines and straightforward calibration; SiPM favors very low-light detection but demands more bias/temperature evidence fields.
5.2 TIA architecture: gain strategy, bandwidth, and stability margin (conceptual)
- Gain strategy: single high gain improves clean-air resolution but risks clipping; dual-gain or auto-ranging preserves headroom while keeping low-end sensitivity.
- Bandwidth strategy: bandwidth must align with the modulation/sampling plan (see H2-6). Excess bandwidth amplifies EMI and switching edges without adding measurement value.
- Stability margin concept: insufficient margin appears as ringing or “false pulses,” especially when temperature shifts component values. Verification uses light-toggle step tests and impulse signatures at TIA output.
5.3 Input protection and leakage control
- Protection must be invisible in-band: input clamps and ESD paths must not introduce measurable leakage or temperature-dependent offsets that resemble dust.
- Placement matters: protection and guard paths should prevent transient damage while avoiding extra parasitics at the summing node.
5.4 Saturation management: make clipping a state, not a surprise
- Dual-gain path: low-gain channel preserves bursts; high-gain channel resolves clean-air baseline.
- Auto-ranging: change gain with hysteresis; log the gain range state and the reason for switching.
- Clamp strategy: if clipping occurs, it must raise a saturation flag and preserve context (gain state + headroom margin).
5.5 Analog filtering targets (scoped)
- Reject supply ripple and switching edges: prevent rail-coupled artifacts from entering the ADC.
- Suppress EMI pickup: reduce high-frequency bursts that correlate with wireless TX and digital transitions.
- Do not “erase” baseline reality: slow drift is handled by baseline strategy (H2-6). Analog filtering should prevent corruption, not hide evidence.
Acceptance mindset: a “good AFE” is not a schematic—it is a measurable profile: noise floor at clean air, headroom during bursts, and explicit gain/saturation states.
Use this AFE diagram to describe how PD/SiPM current becomes robust digital data using leakage-aware input protection, dual-gain or auto-ranging TIA, analog filtering, and loggable evidence outputs.
Ambient-light rejection and synchronous detection strategy
Most field drift is caused by ambient/light-leak combined with baseline tracking failure. Robust designs treat ambient light as a measurable input and use synchronous sampling plus a baseline estimator to separate net scattering from ambient conditions.
6.1 Product-class methods (choose and document the assumption)
- Optical/mechanical shielding: treated as a constraint and leakage limit, not a guarantee. Designs still require electrical observability of ambient conditions.
- Electrical rejection: modulation (pulsed light) paired with correlated/synchronous sampling to remove ambient components.
- Baseline estimator: maintain dark and ambient samples and apply subtraction rules with explicit freeze/recovery states.
6.2 Baseline estimator rules (prevent event corruption)
- Dark sample: capture when light is off (or path blocked) to estimate detector/AFE baseline.
- Ambient sample: capture ambient-only component (light off) and track its changes over time.
- Subtraction path: compute net scattering as (ON-window − OFF-window) under a known timing plan. Baseline updates must freeze during particle events to avoid “learning the dust.”
- Recovery behavior: after a light-step (sunlight / switching on room lights), recovery time must be bounded and logged.
6.3 Validation checklist (stress tests that reveal real drift sources)
- Sunlight stress: different angles/intensity; verify ambient baseline tracking and recovery time bounds.
- Flicker stress: indoor flicker and dimmers; verify no alias patterns appear in net scattering output.
- Light on/off step response: measure step response and recovery curve; confirm baseline-state transitions are logged.
Acceptance mindset: ambient rejection is proven by a timing diagram, window alignment, and logged recovery behavior—not by “it seems stable indoors.”
Use this timeline to document correlated sampling for ambient rejection: modulation, ON/OFF ADC windows, baseline samples, and the net subtraction path with validation items.
Temperature compensation and drift management (the practical playbook)
Drift becomes manageable only when it is classified, measured, and constrained by guardrails. Practical designs combine temperature-aware calibration with slow baseline adaptation and strict rules that prevent real dust events from being absorbed as “drift.”
7.1 Drift sources (symptoms mapped to evidence fields)
- Light-source efficiency drift: long-term baseline slope. Compare output trends to Light_I_set/Light_I_meas and temperature.
- Detector dark current vs temperature: baseline changes persist even with light off. Track Dark_baseline vs Temp.
- TIA offset / leakage drift: low-end bias shifts and temperature sensitivity. Track baseline offset and stability over a temperature sweep.
- ADC reference / gain drift: apparent scaling across the range. Track ADC_ref_proxy and calibration gain/offset versions.
- Contamination (optical path): reduced responsiveness and longer recovery after light steps. Track Recovery_time and contamination heuristics.
7.2 Compensation layers (stacked, versioned)
- Temperature sensing placement & sampling: measure temperatures that correlate with drift sources (light source, detector/AFE region). Log both raw and filtered temperature to avoid time-constant confusion.
- Per-temperature calibration: use coefficients or LUT tables to correct gain/offset over temperature. Store as versioned assets (Cal_version, Temp_coeff/LUT_id).
- Slow baseline adaptation: allow bounded long-term adjustments with a maximum drift rate and a rollback checkpoint.
7.3 Guardrails: “Do not chase dust as drift”
- Freeze during events: when event detection triggers, baseline updates must stop (Baseline_state = FREEZE).
- Recovery mode for light steps: allow controlled recovery after ambient steps (Baseline_state = RECOVERY) while preventing dust misclassification.
- Rate limits: baseline slope per hour must be bounded to avoid runaway adaptation.
- Traceable reasons: baseline state transitions must carry a reason code (event, light step, stable, fault).
7.4 Chamber-test evidence package (minimum acceptance)
- Baseline vs temperature plots: up/down sweep with hysteresis observation; compare “before vs after compensation.”
- Temp coefficient logs: coefficients or LUT version attached to each firmware build and telemetry packet.
- Residual drift metric: bounded drift after compensation under stable dust conditions.
A drift strategy is complete only when baseline-state transitions, coefficients/LUT versioning, and chamber plots can explain the output across temperature without manual retuning.
Use this diagram to explain drift management for optical-scattering dust monitors: drift sources, temperature-aware calibration layers, and guardrails that prevent real dust events from being absorbed as drift.
Digitization and edge MCU processing (sampling, filtering, binning)
Edge processing succeeds when the digital pipeline is reproducible: a telemetry packet can replay the same derived metrics. This chapter defines ADC requirements, a scoped DSP pipeline, mapping dependencies, and fault detection that turns “odd behavior” into traceable flags.
8.1 ADC requirements (beyond resolution)
- Resolution (effective bits): sets the clean-air quantization floor and low-end sensitivity.
- Sampling rate: must align with modulation and windowing (ON/OFF) to avoid alias patterns.
- Reference stability: drifting Vref creates “scale drift” that looks like dust. Track a reference proxy where possible.
8.2 Scoped DSP pipeline (replayable steps)
- Pre-clean: remove clipped samples and impossible codes; tag them for diagnostics.
- Smoothing: short-window smoothing for stability without large latency.
- Outlier rejection: suppress spikes correlated with EMI or wireless TX.
- Event detection: detect bursts using variance/peaks and emit Event_flag and event counters.
- Rolling windows: compute 1–10 s windows and longer trends for reporting.
8.3 Mapping: counts → concentration (explicit model dependence)
- Concentration mapping depends on the calibration aerosol/model and optical geometry. Therefore mapping must be versioned and attached to telemetry.
- Store and report Mapping_model_id and Cal_version so metrics remain comparable across builds and batches.
8.4 Fault detection (bitmask, not anecdotes)
- Saturation: ADC/TIA clipping detected and flagged with context (gain range and headroom).
- Stuck-at baseline: baseline does not move and noise becomes abnormal—indicates disconnection or algorithmic lockup.
- Light-source failure: mismatch between I_set and I_meas, or monitor proxy not responding to control changes.
- Contamination heuristic: responsiveness decreases and recovery time increases over time.
A minimal telemetry packet should enable replay: raw/net channels, filtering outputs, baseline states, temperature, driver current, and fault bitmask with versions.
Use this diagram to document the edge DSP pipeline (pre-clean, smoothing, outlier rejection, event detection, rolling windows), versioned mapping, and the minimum telemetry schema required to replay derived dust metrics.
Wireless + edge node architecture (MCU/wireless, OTA, interface-level integrity)
A dust monitor becomes deployable when sampling and diagnostics remain deterministic under wireless activity, and when every field unit reports enough context to explain its readings. This chapter focuses on node partitioning, power states, OTA hooks, and interface-level integrity—without RF/antenna design details.
9.1 Node partitioning templates (choose by measurement sensitivity)
- Split architecture (sensing MCU + wireless SoC): sensing stays deterministic; wireless stack bursts are isolated from the sampling timeline.
- Combo architecture (single SoC): requires strict scheduling: sampling windows must preempt wireless tasks, and TX bursts must be rate-limited around measurement windows.
9.2 Scheduling and data path (deterministic sampling, buffered reporting)
- Sampling as a real-time task: ON/OFF windows and baseline state transitions must stay phase-stable.
- Wireless as a background task: batch reporting from a ring buffer reduces frequent TX bursts that correlate with measurement noise.
- TX-aware guard time: optional policy: avoid TX during critical sampling windows and log TX activity when unavoidable.
9.3 OTA hooks (update strategy with rollback behavior)
- Versioned assets: firmware version and calibration/mapping versions must be reported together.
- Rollback point: update failures should revert to a known-good image and preserve calibration constants.
- Update window policy: schedule updates outside critical sampling periods to avoid baseline disturbance.
9.4 Interface-level integrity (basic, practical)
- Monotonic counter: detects replay and duplicates.
- Packet integrity check: detect corruption and drop invalid payloads.
- Version attachment: each packet carries FW/Cal/Model versions for traceability.
9.5 Field diagnostics (minimum remote triage capability)
- Last fault: fault code and timestamp (or sample index).
- Sensor health score: derived from noise floor, saturation rate, recovery time, and contamination heuristics.
- Calibration version: ensures cross-node comparability and correct model interpretation.
A deployable node must explain itself remotely: deterministic sampling, buffered reporting, versioned calibration, and a compact diagnostic footprint.
Use this node diagram to describe deployable dust monitors: sensing-domain determinism, interface integrity (counter + checksum), OTA hooks with rollback behavior, and field diagnostics (last fault + health score + versions).
Power integrity, EMC/ESD, and robustness (what breaks the measurement)
Dust monitors often fail silently: the device stays “alive,” but power noise and EMI corrupt micro-amp measurement paths. Robust designs treat power/EMI as measurable inputs: define domains, identify coupling paths, and prove immunity using A/B tests such as wireless TX on/off impact on baseline and noise floor.
10.1 Rail planning (domains and reference cleanliness)
- Separate domains: keep analog sensing rails isolated from digital/wireless rails where possible.
- Reference stability: ADC reference cleanliness is often more important than bulk supply regulation.
- Ground boundaries: define where noisy return currents flow so TIA and Vref nodes are not modulated by system activity.
10.2 Coupling paths (what injects false dust)
- Laser switching edges → rail/ground bounce → TIA corruption: residual artifacts can survive ON/OFF subtraction if windows are misaligned.
- Wireless TX bursts → Vref/ADC disturbance: raises the noise floor and can create periodic patterns in net_scatter.
- Digital clocks → electric/magnetic coupling: introduces narrowband artifacts that appear as false events.
10.3 Robustness strategy (separation + filtering + time-slotting)
- Separation: keep the measurement chain’s reference and summing nodes protected from noisy domains.
- Filtering: limit high-frequency junk entering ADC inputs and references.
- Time-slotting: avoid measuring during high-interference activities (TX bursts, high di/dt edges) and log activity when overlap occurs.
10.4 ESD/surge philosophy for exposed ports (conceptual)
- Energy termination at the boundary: clamp and route energy at the port entrance, not through the sensing area.
- Leakage discipline: protection must not create temperature-dependent offsets at the TIA input.
- Post-event detectability: after ESD, self-tests and health scores must reveal silent degradation.
10.5 Evidence fields and field tests (minimum acceptance)
- Wireless TX on/off A/B test: compare noise floor, baseline slope, and false event rate under fixed conditions.
- Baseline disturbance observations: tag measurement artifacts when rails or system states change.
- Recovery-time checks: verify bounded recovery after interference steps.
Robustness is proven by evidence: TX on/off deltas, baseline stability under system activity, and detectability after ESD events—not by “it seems stable on the bench.”
Use this coupling-path map to explain silent measurement corruption in dust monitors: noise sources, rail/ground/field/leakage paths, victim nodes (TIA/ADC/Vref/baseline), and evidence fields from TX on/off A/B testing.
Calibration, production test, and field validation (the evidence chain)
This chapter turns “it seems to work” into a manufacturing and field method. The objective is a closed evidence chain: calibration artifacts are versioned, production tests are fixture-reproducible, and field validation can trigger maintenance or recalibration.
11.1 What calibration covers (scope lock)
- Dark baseline: light off + optical blocked (or shuttered) to measure detector/TIA offsets and noise.
- Ambient baseline: light off + optical exposed to quantify light leak and baseline tracking burden.
- Electrical gain/offset: AFE+ADC chain check using a known injection/loopback path.
- Reference aerosol point(s): at least one optical anchor point to bind mapping_model_id to real behavior.
Out of scope here: RF/antenna tuning and deep scattering theory derivations.
11.2 Factory calibration flow (step outputs are evidence fields)
- Step A — Dark: record dark_mean, dark_noise, dark_temp, and ensure no clipping.
- Step B — Ambient: record ambient_mean and optional ambient_step_recovery to quantify light-leak sensitivity.
- Step C — Electrical gain/offset: inject a known equivalent input; write cal_gain, cal_offset, and adc_ref_proxy.
- Step D — Optical reference point(s): capture ref_point_id, ref_counts, label_concentration, and bind to mapping_model_id.
- Step E — Seal & readback: store a versioned payload in NVM and verify with an integrity check; publish cal_version.
11.3 Production test fixtures (optical + electrical + fault injection)
- Optical sanity: light proxy responds to a controlled stimulus; detect weak/failed sources and gross contamination.
- Electrical loopback: isolate AFE+ADC; verify gain/offset/noise without relying on the optical chamber.
- Fault injection: open/short and saturation paths must raise a deterministic fault_bitmask and reason code.
Fixture outputs should be per-unit logs: PASS / REWORK / SCRAP + reason_code.
11.4 Acceptance limits and rework criteria (make decisions reproducible)
- Acceptance: bounded dark_noise, bounded ambient_leak, valid cal_gain/offset range, no saturation under fixture profile.
- Rework triggers: optical sanity fails, ambient leak too high, abnormal recovery time, repeated saturation at reference point.
- Rework actions: clean chamber/optics, reseat source/detector, re-run calibration, then re-test with the same profile.
11.5 Field validation (co-location, drift tracking, maintenance prompts)
- Co-location comparison: compare to a trusted node/instrument for a defined duration; log colocation_delta_mean and tail stats.
- Drift tracking: publish baseline_drift_slope, noise_floor_est, and recovery_time trends.
- Maintenance prompts: trigger clean/filter/service when contamination heuristics or drift patterns exceed guardrails.
11.6 Deliverables (what must exist in software + logs)
- Calibration payload format: versioned structure, integrity check, readback verification.
- Per-unit production test log: fixture_id, test_profile_id, PASS/REWORK/SCRAP, reason_code, key stats.
- Field policy table: triggers → actions (maintenance_flag, recal_trigger, baseline freeze rules).
A unit is “manufacturable” only when calibration parameters are versioned, test fixtures produce deterministic PASS/FAIL decisions, and field telemetry can explain drift and maintenance needs.
Example material numbers (MPNs) for calibration/test building blocks
These are commonly used parts that fit the roles below. Select equivalents based on availability, optical geometry, and target noise floor.
- Photodiode (detector examples): Vishay BPW34; OSRAM SFH 2704.
- Low-noise/TIA-capable op-amp examples: TI OPA380 (TIA-friendly); Analog Devices ADA4530-1 (ultra-low bias for high impedance); Analog Devices LTC6268 (high-speed, low bias).
- Precision ADC examples: TI ADS1220 (24-bit delta-sigma); Analog Devices AD7685 (16-bit SAR).
- Temperature sensor examples: TI TMP117 (high-accuracy digital); Microchip MCP9808 (digital temp).
- Calibration / log NVM examples: Microchip 24LC256 (I²C EEPROM); Infineon/Cypress S25FL064L (SPI NOR flash).
- RTC for timestamped evidence (optional): Micro Crystal RV-3028-C7.
- Secure identity / integrity helper (optional): Microchip ATECC608B.
- ESD protection for exposed ports: TI TPD4E1U06; Nexperia PESD5V0S1UL.
- Load switch / rail gating: TI TPS22910A.
- System buck regulator (node power example): TI TPS62162 (low-Iq buck).
- Wireless SoC (node examples, RF details excluded): Nordic nRF52840; TI CC2642R; Espressif ESP32-C3.
- Low-power MCU (if split architecture): ST STM32L432 (L4 family example).
Use this flow to document a complete evidence chain for dust monitors: factory calibration artifacts, fixture-based production test decisions, QA audits, and field validation that triggers maintenance or recalibration with version updates.
FAQs (Troubleshooting entry points, evidence-driven)
Each FAQ maps back to the evidence chain: light-source stability, AFE integrity, ambient rejection, temperature drift control, edge processing, wireless/node behavior, EMC robustness, and calibration/validation artifacts.
Readings drift with temperature — laser intensity drift or PD/TIA offset?
Open
Short answer: Separate source drift from analog offset drift by correlating temperature with a light-power proxy and with dark/baseline metrics.
- What to measure: light_i_meas (or light_proxy) vs temperature during stable air; check if net_scatter follows it.
- What to measure: dark_mean/baseline_state vs temperature with light off (or blocked) to expose PD/TIA/ADC offset drift.
- First fix: Freeze baseline updates during temperature ramps, then apply a temperature compensation table (Example parts: TMP117, MCP9808).
Works in lab, fails in sunlight — ambient leak or sampling/aliasing?
Open
Short answer: Sunlight failures usually come from ambient light leakage plus a baseline estimator that cannot recover, or from misaligned sampling windows that leave residual artifacts.
- What to measure: ambient_mean with source off and optics exposed; compare indoor vs sunlight.
- What to measure: Window alignment markers (mod_phase or sample-index tags) and alias_suspect_flag around modulation edges.
- First fix: Add a “light-off ambient capture” step and adjust sample windows/guard time to avoid edge residue (Example parts: OPA380 for stable TIA, ADS1220 for ref stability).
Sudden spikes when wireless transmits — coupling or rail dip?
Open
Short answer: Treat TX as a controlled disturbance and prove causality with TX on/off A/B while watching reference-sensitive telemetry.
- What to measure: tx_state aligned with net_scatter spikes; compute spike rate when TX is active vs inactive.
- What to measure: adc_ref_proxy (or internal reference sample) and noise_floor_est with TX on/off.
- First fix: Enforce TX-aware guard time around sampling windows and separate analog/digital rails where possible (Example parts: TPS22910A for rail gating, TPD4E1U06 for port ESD).
Saturates only during dust bursts — TIA headroom or gain switching logic?
Open
Short answer: Burst-only saturation is usually headroom at the TIA/ADC or a gain switch that changes too late (or oscillates) during fast events.
- What to measure: saturation_flag and clipped-sample count; record gain_range_id at the moment of saturation.
- What to measure: Peak net_scatter and recovery time after burst; compare with fixed low-gain mode.
- First fix: Lock gain to the lowest range for a controlled burst test, then tune switch thresholds/hysteresis (Example parts: OPA380, AD7685).
Baseline slowly climbs over days — contamination or baseline update bug?
Open
Short answer: A slow baseline rise is contamination when recovery-time and noise trends worsen, but it is a baseline algorithm bug when the estimator “chases dust” during events.
- What to measure: baseline_drift_slope, noise_floor_est, and recovery_time trends over days.
- What to measure: Baseline update state (baseline_state) and event gating (whether baseline updates continue during dust events).
- First fix: Add guardrails: freeze baseline updates during events and trigger maintenance when contamination heuristics persist (Example parts: 24LC256 for versioned cal storage).
Two units disagree by 30% — calibration constants or optics variance?
Open
Short answer: First isolate electrical-chain gain/offset differences via loopback, then blame optics only if electrical alignment matches.
- What to measure: cal_version and mapping_model_id on both units; check mismatched calibration artifacts.
- What to measure: Fixture loopback metrics (loopback_gain_meas, loopback_noise) to confirm AFE/ADC equivalence.
- First fix: Re-run the same factory calibration profile and verify readback integrity before changing optics (Example parts: ATECC608B for integrity, S25FL064L for logs).
Nighttime readings are higher — humidity/condensation or ambient IR?
Open
Short answer: Night shifts often come from condensation changing optical paths or from ambient IR leakage; discriminate using source-off ambient captures and recovery patterns.
- What to measure: ambient_mean (source off) at night vs day; look for elevated background.
- What to measure: Temperature trend and recovery time trend after step changes; condensation often increases hysteresis and recovery time.
- First fix: Add a scheduled source-off ambient sample and tighten baseline update rules during humidity-prone periods (Example parts: RV-3028-C7 for RTC scheduling).
“Zero air” still shows PM — dark subtraction wrong or ADC reference drift?
Open
Short answer: Zero-air offsets are usually dark/ambient subtraction errors or reference drift that moves the ADC scale without obvious saturation.
- What to measure: dark_mean/dark_noise and the subtraction result at a stable temperature.
- What to measure: adc_ref_proxy drift across time and across power states (sleep/active/TX).
- First fix: Re-run the dark and ambient capture sequence and lock reference sampling during calibration (Example parts: TMP117, ADS1220).
After a firmware update, the trend changed — filter window or mapping coefficients?
Open
Short answer: Trend changes after updates are almost always pipeline changes (windowing/outlier rules) or a swapped mapping_model_id rather than “new physics.”
- What to measure: FW_version and filter_window_id (or equivalent) plus mapping_model_id before/after.
- What to measure: Replay the same stored raw segment through old vs new pipeline and compare output deltas.
- First fix: Pin mapping_model_id to the calibration artifact and treat filter changes as a versioned parameter (Example parts: S25FL064L for raw/log capture).
Laser occasionally shuts down — interlock false trip or driver thermal fault?
Open
Short answer: Distinguish false interlock trips from true thermal faults by capturing fault bitmasks and temperature context right before shutdown.
- What to measure: fault_bitmask (interlock_flag vs otp_flag) and light_i_meas in the 5 seconds before/after the event.
- What to measure: Temperature at driver and near optics; check repeatability under identical duty cycles.
- First fix: Add latching fault logs with timestamps and ensure derating/OTP is reported distinctly from interlock logic (Example parts: TMP117, TPS62162 for stable rail).
Fast response looks noisy — need bandwidth or smarter filtering?
Open
Short answer: A faster response always trades noise unless the pipeline uses event-aware windows; the correct choice is set by response-time targets and acceptable false-event rates.
- What to measure: Response-time target vs noise_floor_est and event_false_rate under a fixed stimulus profile.
- What to measure: Window length and outlier rules; check whether variance spikes are treated as events or noise.
- First fix: Keep analog bandwidth stable and tune digital windows (rolling + outlier rejection) to meet the response/noise budget (Example parts: ADS1220 or AD7685 depending on architecture).
Field units degrade faster — ESD events or EMI stress on the AFE?
Open
Short answer: Field degradation is ESD-related when step changes or permanent leakage shifts appear after events, and EMI-related when noise rises with system activity and correlates with TX/state changes.
- What to measure: Event logs (ESD/surge counters if available) and post-event shifts in dark_mean/offset or persistent baseline jumps.
- What to measure: TX on/off deltas in noise_floor_est and adc_ref_proxy plus alias flags under identical conditions.
- First fix: Improve boundary protection and ensure post-event self-test/loopback marks health degradation (Example parts: TPD4E1U06, PESD5V0S1UL).