Edge Backhaul Node (Microwave/mmWave/Fiber)
← Back to: 5G Edge Telecom Infrastructure
An Edge Backhaul Node bridges microwave/mmWave or fiber physical links to Ethernet/OTN transport while distributing low-noise frequency/time across internal modules. In practice, success is proven by controlling LO/PLL jitter and interface stability—and by exporting counters/logs that make EVM, BER, and switchover behavior measurable in the field.
H2-1 · What it is: system role and hard boundaries
An edge backhaul node bridges physical transport (microwave/mmWave RF/IF or fiber optics) to Ethernet/OTN bearer interfaces, while distributing frequency/time inside the node to keep modulation quality, link stability, and transport counters within spec.
What is visible in the field (ports → KPIs)
Start from what technicians and buyers can verify on day one: ports, counters, and pass/fail signals.
- RF / IF ports → link quality signals (EVM, spurs), stability under temperature and power cycling.
- Ethernet ports → throughput, link flap rate, PRBS/BER measurements, retimer/PHY margins.
- OTN ports → pre/post-FEC BER, FEC correction counters, framing alarms, long-haul stability evidence.
- Timing I/O (SyncE/PTP/1PPS/ToD) → in/out availability, holdover behavior, time/frequency alarms.
- Management / OAM → logs, counters, alarm history (lock/unlock, switchover events, temperature excursions).
Boundary checks (anti-overlap, fixed wording)
- Bearer + physical mapping only: focuses on PHY/RF/OTN/Ethernet integration and validation; does not cover MEC/UPF compute and service pipelines.
- Time/frequency I/O and distribution: supports SyncE/PTP/1PPS/ToD interfaces and node-level holdover; does not cover GNSS disciplining and time-source internals.
- OTN/fiber interfaces (not fiber-panel switching): covers transport framing/FEC/counters; does not cover optical power/loss AFE or relay-based panel switching.
What “done” looks like (evidence-first)
- Every claimed capability maps to at least one counter/log/alarm that can be exported and trended.
- Transient events (link retrain, reference switch, temperature step) are captured with pre/post windows to convert “rare” issues into repeatable evidence.
H2-2 · Deployment patterns and link budgets (engineering scenarios)
This section maps real deployment shapes to measurable KPIs, then connects each KPI to the internal block that usually dominates it.
Three common patterns (kept separate to avoid mixing)
Microwave / mmWave backhaul (point-to-point, high capacity, often E-band)
- Dominant KPIs: EVM margin, spurs, link stability under temperature and reference changes.
- Most fragile block: LO/PLL phase-noise and conversion chain isolation.
- What proves it: EVM trend vs. temperature; spur scan results; lock/unlock and switchover event logs.
Fiber / OTN backhaul (long reach bearer, strong emphasis on statistics)
- Dominant KPIs: pre/post-FEC BER, FEC correction margin, framing alarms, long-run stability.
- Most fragile block: PHY/retimer margin and OTN framer/FEC behavior under stress.
- What proves it: PRBS/BER reports, FEC counter baselines, link flap frequency, alarm timelines.
Hybrid backhaul (wireless at the edge + fiber aggregation)
- Dominant KPIs: availability, switchover hit, multi-interface recovery time.
- Most fragile block: reference/time distribution and interface recovery behavior during transitions.
- What proves it: before/after windows around interface retrain and reference switch; alarm + counter correlation.
KPI → dominant block → verification (field-friendly)
| KPI (what users feel) | Dominant block (where it usually enters) | How to verify (evidence to capture) |
|---|---|---|
| Throughput (Gbps) | Ethernet PHY / retimer margin, FEC overhead behavior, interface retrain frequency | Traffic test + link flap count; PRBS; FEC counters baseline vs. stress |
| BER (pre/post-FEC) | SerDes/retimer equalization, optical/RF interface margin, framer/FEC robustness | PRBS BER report; post-FEC error trend; alarm timeline (frame loss, LOS/LOF) |
| EVM margin | LO/PLL phase noise, spurs and leakage paths, conversion chain isolation | EVM vs. temperature; spur scan; lock/unlock logs; switchover disturbance windows |
| PDV / jitter (packet/time) | Timing distribution (SyncE/PTP I/O), holdover behavior, reference transitions | Offset/jitter statistics (node I/O); reference change logs; alarm correlation |
| Availability (hitless expectation) | Interface recovery and switchover handling, alarm thresholds, event logging completeness | Drop/restore timestamps; counters around events (pre/post); reproducible switchover script |
Budget thinking (three parallel chains)
- RF chain budget: phase noise / spurs → EVM margin → achievable modulation/capacity under temperature and aging.
- Bearer chain budget: SerDes margin → BER (pre/post-FEC) → long-run stability with counters as proof.
- Sync chain budget: time/frequency I/O quality → holdover + transitions → alarms and evidence windows for field issues.
H2-3 · Reference architecture: a port-to-module profile (field-ready)
This architecture is organized as a repeatable profile from ports to internal blocks, so every KPI can be tied to a measurable counter, log, or alarm. The same block names are reused in later troubleshooting sections to keep the page consistent and searchable.
Fixed layers (do not reorder)
- Port / Front-end → connectors, protection boundaries, and what enters the node.
- Conversion / PHY → RF/IF conversion or optical-electrical PHY/retimers.
- Framing / Mapping → OTN framing/FEC and mapping into Ethernet services.
- Timing → SyncE/PTP/1PPS/ToD I/O, distribution, node-level holdover alarms.
- Mgmt / OAM → counters/logs/alarms that prove the above layers.
Module checklist (Inputs / Outputs / KPIs / Evidence)
KPIs: EVM sensitivity, spur levels
Evidence: EVM trend, spur scan list, temperature-linked alarms
KPIs: image rejection, LO leakage, group delay ripple
Evidence: spur signatures vs LO power, IQ calibration residuals
KPIs: phase noise, lock stability, switchover disturbance
Evidence: lock/unlock logs, switchover events, jitter/offset windows
KPIs: BER margin, equalization stability, retrain frequency
Evidence: PRBS BER report, link flap counters, LOS/LOF alarms
KPIs: corrected/uncorrected errors, latency overhead budget
Evidence: pre/post-FEC BER, corrected counters, framing alarms timeline
KPIs: throughput stability, loss events around transitions
Evidence: traffic tests, drops around retrain/switchover, service counters
KPIs: offset/jitter stats, holdover drift profile, alarms
Evidence: offset logs, holdover alarms, reference change records
KPIs: stability under heat, repeatability of counters
Evidence: alarm history + counters correlated to temperature steps
KPIs: completeness and time-correlation
Evidence: event timelines, “before/after” windows, trendable baselines
H2-4 · Up/down conversion chain: where EVM and spurs are decided
In microwave/mmWave backhaul, modulation quality is most often limited by the conversion chain and its LO behavior. The goal here is not broad RF theory, but a repeatable mapping from visible symptoms to the first measurement that reduces uncertainty.
Typical paths (minimum necessary view)
- Tx: Baseband/IF → Mixer → RF → PA → RF out
- Rx: RF in → LNA → Mixer → IF → baseband
Architecture choice (only what changes EVM/spurs/filtering)
| Option | What it improves | Where it becomes fragile |
|---|---|---|
| Super-heterodyne | Image filtering opportunities, spur separation, more controllable IF shaping | More blocks and conversions; more places for LO feedthrough and group delay ripple to enter |
| Direct conversion | Simpler frequency plan and fewer conversions; potentially lower latency | Highly sensitive to LO leakage, I/Q imbalance, and DC/near-DC impairments that show up as EVM loss |
Failure modes (symptom → likely cause → first measurement)
First measurement: EVM vs temperature step; PLL lock/unlock log; spur scan before/after
First measurement: spectrum spur list; spur amplitude vs LO power; confirm image frequency location
First measurement: EVM distribution histogram; IQ calibration residual; power sweep vs EVM
First measurement: noise floor vs LO on/off; spur scan; compare with shield/grounding states
First measurement: event window capture (pre/post); lock events; EVM/BER correlation in time
H2-5 · Low phase-noise PLL/LO: from L(f) to jitter to EVM/BER
Phase noise L(f) becomes integrated jitter after the loop and distribution chain, and that jitter reduces constellation separation (EVM margin) and effective demodulation thresholds—especially at high-order modulation and tight link budgets. This section focuses on node-local reference choices, loop-bandwidth trade-offs, and where jitter cleaning is most effective.
The practical chain: what to measure and what it impacts
Why it matters: different offset regions dominate different impairments (close-in vs far-out)
Evidence: spur/phase-noise report + lock stability logs
Why it matters: reduces sampling/LO stability and increases demodulation uncertainty
Evidence: jitter metrics + event-window correlation (reference switch → jitter/EVM spike)
Why it matters: sets achievable modulation order and capacity at a given link budget
Evidence: EVM histogram/trend, BER trend, capacity step-down events
Node reference choice (TCXO vs OCXO): boundary and decision cues
This page stays within node-local behavior: short disruptions, brief reference instability, and holdover alarms—without expanding into GNSS disciplining or atomic sources.
| Reference | Best for (node-local) | Typical risks / what to verify |
|---|---|---|
| TCXO | Cost/power-sensitive nodes, moderate holdover needs, fast warm-up | Temperature sensitivity shows as drift/offset during transitions; verify reference switch windows + alarms |
| OCXO | Stricter stability during short reference disruptions and tighter EVM margins | Power/thermal constraints; verify warm-up behavior, stability under enclosure temperature ramps |
Verification anchor: capture “before/after” windows around reference changes (or induced disturbances) and correlate with EVM/BER and PLL lock status.
Loop bandwidth: the three dominance regions (how to tune and how to accept)
Engineering move: improve reference quality and spur control; validate lock stability
Acceptance: lock/unlock frequency, spur list stability, “no surprise” events
Engineering move: set BW/ damping to balance tracking vs noise injection
Acceptance: reference switch window shows limited EVM/BER spike and fast settle
Engineering move: choose lower-noise VCO/distribution and improve isolation
Acceptance: noise floor and spur baseline remain stable across load/temperature
Where to place jitter cleaning inside the node (most effective first)
Prove it: EVM improves without changing channel conditions; spur list stabilizes
Prove it: PRBS BER improves; link retrain counters drop; FEC pressure trends lower
Prove it: timing alarms reduce; offset/jitter logs tighten during transitions
H2-6 · Ethernet / OTN interfaces: bearer mapping and clock recovery boundaries
This section covers interface integrity and bearer mapping evidence (PHY/retimer, framer/FEC, counters and alarms). It does not cover switching, queue behavior, or TSN scheduling, which belong to edge switching and boundary-clock switch topics.
Ethernet side (PHY / Retimer / SerDes): what breaks stability first
What to capture: retrain counters, link-up/down timeline, temperature correlation
What to capture: PRBS BER reports, eye margin summaries (if available), error counters
What to capture: corrected/uncorrected trend and whether it climbs with temperature/load
OTN side (Framer / FEC / mapping): why it exists in this node
OTN is treated here as an evidence-rich bearer layer: it provides framing alarms and FEC statistics that prove long-run stability under real stress. The focus stays on what it adds inside the node (verification items), not on network-level service engineering.
Use: tie alarm bursts to retrain events and environmental steps
Use: distinguish “margin shrinking” vs “rare transient” behavior
Use: event-window capture (retrain/switchover) to prove no hidden instability
Interface checklist: run full, run stable, prove stable
| Checklist item | Failure symptom | Primary evidence |
|---|---|---|
| Run full rate (sustained) | Rate fallback, unstable lane bring-up | Throughput test + link training status + lane error counters |
| Run stable BER (stress/temperature) | BER spikes, intermittent drops | PRBS BER report; time-correlated error counter trend |
| FEC pressure stays bounded | Corrected errors climb continuously | Corrected/uncorrected counters; pre/post-FEC BER trend |
| No hidden instability around events | Drops during retrain/switchover | Event windows (pre/post) + drop counters + alarm timeline |
| Alarms are actionable | “No alarm” yet link is unstable | Alarm taxonomy + timestamped logs; cross-check with counters |
H2-7 · Precision timing in-node: SyncE + PTP landing points (node scope only)
Precision timing in this node is implemented as a closed loop: Inputs → Selection → Distribution → Monitoring → Alarms. SyncE provides frequency, PTP provides time, and 1PPS/ToD provides auxiliary alignment/verification. The goal is stable delivery to RF/clock consumers with evidence-based alarms and event-window proofs.
Inputs: three timing signal classes and what each means
Interface focus: SyncE-capable port/clock recovery domain
Common misuse: treating “frequency OK” as “time OK” without time validation
Interface focus: PTP packet path and timestamp domain (TSU boundary)
Common misuse: relying on a single snapshot instead of trends + event windows
Interface focus: 1PPS pin + ToD interface (format dependent)
Common misuse: ignoring distribution path noise and treating the pin as “ideal”
Selection and switchover: reference priority, hitless intent, and real disturbance paths
What to prove: selector decisions match alarms and measurable degradations
Realistic case: brief disturbance may show as PLL transient, PTP offset spike, or RF EVM glitch
Evidence: event-window capture around switch action
Prove it: holdover-enter/exit logs + drift/offset trend stays bounded
Distribution: who needs what quality (RF LO vs SerDes vs timestamp domain)
| Consumer | Primary sensitivity (what degrades first) | Fastest proof (node evidence) |
|---|---|---|
| RF LO / sampling clock | EVM margin and demodulation threshold; micro-glitches show during disturbances | EVM trend + spur stability + switch-window correlation |
| SerDes / retimer domain | BER margin, retrain probability, and FEC pressure trends | PRBS BER + link flap/retrain counters + corrected/uncorrected trend |
| Timestamp unit (TSU) domain | PTP offset/jitter stability; spikes reveal reference and distribution events | Offset/jitter time series + holdover/selector logs on the same timeline |
Monitoring and alarms: the minimum evidence set (actionable, not decorative)
H2-8 · Redundancy & reliability: proving switchovers and rollbacks with evidence
“Switched over” is not the same as “service is unaffected”. Every redundancy action must be validated with an event window and a minimal KPI set: drops, EVM micro-glitch, PLL unlock, and PTP offset spike. This section provides acceptance scripts that bind each action to counters and logs.
Redundancy dimensions (kept separate to avoid mixed conclusions)
Risk: retrain/link flap, short packet drops, FEC pressure climb
Evidence: link timeline + PRBS/BER + FEC counters + drop window
Risk: PLL transient, offset spike, holdover enter/exit oscillation
Evidence: selector logs + lock/holdover + offset/jitter windows + (optional) EVM window
Risk: “recovered” but margins shift (BER/EVM/offset baseline changes)
Evidence: version log + baseline comparison before/after
Acceptance method: one event window, four KPIs, one decision
| KPI | What “bad” looks like | Fast evidence (node scope) |
|---|---|---|
| Drops | burst drops during switch and settle phases | drop counters + throughput trace aligned to switch timestamp |
| EVM micro-glitch | brief constellation expansion or spur pop during switchover | EVM trend/histogram with event markers (if RF chain is present) |
| PLL unlock | unlock/relock events or long settle time | lock log + settle duration + holdover status |
| PTP offset spike | offset exceeds thresholds or slow return to baseline | offset/jitter time series with pre/post window summary |
Acceptance scripts (each action bound to counters and logs)
H2-9 · Observability & operations: using OAM/telemetry to debug jitter, errors, and link drops
The most expensive field faults are intermittent: jitter bursts, transient BER/FEC events, and short link drops. The node becomes diagnosable only when it exposes a minimal evidence set and records pre/post event windows so “cannot reproduce” turns into a replayable timeline.
Minimal observability set (grouped by what it can prove)
Proves: whether timing disturbances align with service symptoms (EVM spikes, offset spikes, BER stress).
Proves: whether errors originate in the physical link and how quickly recovery stabilizes.
Proves: whether the link is “alive but stressed” and how close it is to uncorrectables.
Proves: whether environment or node-local power events coincide with drops or retrains.
Proves: whether “same symptom” is actually a different software/config regime.
The most valuable log: pre/post event windows (make intermittents replayable)
Approach: always capture a pre-window and post-window around a trigger and align all evidence onto one timeline.
Operator: manual “mark event” during observed degradation, then export the buffered window.
Result: the dominant cause becomes visible through consistent co-occurrence patterns.
Symptom → evidence mapping (what to pull first, and what often misleads)
| Symptom | First evidence to pull | Second evidence (confirm/triage) | Common misread |
|---|---|---|---|
| EVM gets worse | PLL lock/unlock + jitter proxy + selector switches (event window) | Temperature trend + any RF chain status snapshots (if exported) | Blaming RF hardware without checking timing disturbances and settle events |
| Throughput drops | Drop counters + link retrain/link flap timeline | FEC pressure trend (corrected growth) + BER window stats | Assuming “traffic” is the cause while ignoring a recovering link or rising FEC stress |
| Intermittent disconnect | Link up/down + flap counters + uncorrected events + reset reasons | Node-local power alarms (PG/UV) + thermal alarm markers | Chasing a remote network issue when the node is cycling a local protection path |
| Delay jitter spikes | PTP offset/jitter series + selector/holdover events | Interface error windows (BER/FEC) that correlate with packet timing variance | Over-attributing to switching/queues without first proving a timebase disturbance |
H2-10 · Validation & production test: what “done” means and how stability is proven in the field
Validation is complete only when key KPIs remain stable across disturbance scenarios and the node can produce field evidence. Testing is structured in three layers: engineering validation, production screening, and field self-check (node scope only).
Three-layer structure (engineering → production → field self-check)
Output: repeatable procedures + windowed KPI summaries tied to firmware/config identity.
Output: pass/fail records + minimal counters captured per unit.
Output: self-check results + aligned evidence export to operations.
Test template (test item → fixture/input → pass criteria → record fields)
| Test item | Fixture / input | Pass criteria (windowed) | Record fields (evidence) |
|---|---|---|---|
| EVM stability | Modulated RF link / controlled channel; optional thermal steps | No abnormal expansion/spikes during steady state and during disturbance windows | EVM trend + event markers + temperature snapshot + active reference |
| PRBS / BER | PRBS generator/checker or link test mode; cable/optical path variations | BER stays within margin; no burst errors after settle | BER windows + retrain/link flap counters + eye/eq summary (if available) |
| FEC margin / pressure | Stress link (attenuation/noise); framing active | Corrected counter growth acceptable; uncorrected events absent or bounded | Corrected/uncorrected counters + margin proxy + timeline correlation |
| PLL lock time | Power cycle / ref re-apply; repeated runs | Lock time consistent; no unlock storms | Lock/unlock log + settle duration + holdover transitions |
| Reference switchover disturbance | Force Ref1 loss/degrade; trigger selector decision | Bounded offset spike; fast return to baseline; no flapping | Selector reason code + offset window stats + PLL events |
| Holdover vs temperature | Thermal sweep (engineering) / limited hot/cold points (production) | Drift trend remains within operational tolerance for expected durations | Holdover duration + drift/offset summary + temperature trace |
| Hot-plug / retrain recovery | Repeated plug/unplug or link drop injection | Recovery time consistent; service stabilizes quickly | Link timeline + retrain counts + drops window + FEC/BER after settle |
| Evidence export completeness | Trigger windows; export to remote collector | All required fields present and aligned to one timeline | Event window bundle: timing + link + FEC + thermal + audit |
H2-11 · BOM / IC selection checklist (criteria + example part numbers)
This section is a selection-by-evidence checklist for an edge backhaul node: each device class is judged by criteria → why it matters → how to verify, then a short list of representative IC part numbers is provided for fast sourcing and design reviews.
1) PLL / network synchronizer / jitter cleaner (node-internal clock quality)
- Criterion: phase noise profile + integrated jitter for the target consumer (LO vs SerDes vs TSU).
Why: jitter directly degrades EVM/BER or creates offset spikes under switchover. Verify: jitter/phase-noise report + BER/EVM correlation; log “jitter cleaner state”. - Criterion: loop bandwidth controls (reference-noise vs VCO-noise dominant regions).
Why: wrong bandwidth “looks locked” but fails under real traffic / temperature. Verify: reference step response + spur scan + stability over temperature. - Criterion: hitless reference switching + switch reason codes + anti-flap mechanisms.
Why: “switched” is not equal to “service unaffected.” Verify: packet loss / EVM glitch window + lock/unlock counters + time-offset spike. - Criterion: holdover drift vs temperature & aging (node-level, not GNSS/atomic).
Why: backhaul must survive short ref loss without large time/frequency error. Verify: controlled ref removal + temperature sweep; record drift trend fields. - Criterion: output format flexibility (LVDS/LVPECL/CML/CMOS) + fanout + skew controls.
Why: LO/SerDes/TSU often need different signaling and deterministic skew. Verify: clock integrity at consumers + skew/phase alignment logs. - Criterion: observability: lock state, DPLL state, reference quality alarms exportable to telemetry.
Why: field issues are solved by evidence, not lab assumptions. Verify: event logs include pre/post window + timestamps aligned to node time.
- TI LMK05318 / LMK05318B — network synchronizer clock with jitter cleaning + hitless switching.
- Skyworks/Silicon Labs Si5345 — jitter attenuator / clock multiplier family.
- Microchip ZL30772 — SyncE / IEEE 1588 network synchronizer (DPLL-based).
- Renesas 8V97003 — wideband RF synthesizer with low phase jitter (LO-grade reference option).
- TI LMX2594 — 15GHz wideband PLL synthesizer (LO generation, low integrated jitter).
- Analog Devices ADF4371 — wideband microwave synthesizer (up to 32GHz with dividers/multipliers).
- Analog Devices HMC7044 — dual-loop jitter attenuator / reference selection & distribution.
- Analog Devices LTC6952 — ultralow jitter PLL + multi-output distribution (clock tree building block).
2) Mixer / upconverter / LO distribution (EVM + spur control zone)
- Criterion: LO leakage + isolation (LO→RF/IF feedthrough).
Why: leakage can dominate spurs and degrade EVM; it may appear only at specific temperatures. Verify: spur scan vs temperature + correlate to LO power. - Criterion: image / sideband suppression and calibration hooks (I/Q balance controls if applicable).
Why: poor suppression forces heavier filtering and reduces margin. Verify: image rejection test, store calibration coefficients in logs. - Criterion: linearity (IIP3 / P1dB) under expected channel loading.
Why: intermod rises at high load/strong signal and looks like “random EVM collapse.” Verify: two-tone IMD + EVM under traffic-like waveforms. - Criterion: group delay / amplitude flatness across band (especially wideband microwave).
Why: group delay ripple increases demod penalty and tightening equalization. Verify: swept response + EVM across channels. - Criterion: LO distribution sensitivity to supply noise and layout coupling.
Why: supply ripple can upconvert to spurs. Verify: injected ripple test + spur response + supply monitor time alignment.
- Analog Devices ADMV1013 — wideband microwave upconverter (24–44GHz class), suited for point-to-point microwave radios.
- Analog Devices HMC7044 — can serve as low-noise reference selection + LO/clock distribution hub for RF chain blocks.
- TI LMX2594 / ADI ADF4371 — common LO synthesizer building blocks for microwave/mmWave backhaul nodes.
3) Ethernet PHY / retimer / SerDes (throughput + BER stability)
- Criterion: supported line rates & host interfaces (KR/KR4, CAUI, 25GAUI, etc.).
Why: mismatched host-side modes create training failures and intermittent flaps. Verify: link bring-up matrix + stable autoneg/training. - Criterion: equalization + CTLE/DFE training robustness across worst-case channel loss/crosstalk.
Why: “works in lab” fails with field cabling/backplane variance. Verify: stress channel sweep; record training status and margins. - Criterion: jitter tolerance (input) and jitter transfer (output) behavior.
Why: jitter accumulation can push BER over threshold during ref switch or thermal drift. Verify: BER vs injected jitter; correlate with clock-tree state. - Criterion: low latency retiming and deterministic behavior under retrain/relock.
Why: recovery time is a field KPI. Verify: scripted unplug/replug + retrain duration histogram in logs. - Criterion: FEC visibility: counters and error distribution accessible to telemetry.
Why: corrected errors rising is an early warning. Verify: counters streamed with timestamps + event window snapshots.
- TI DS280DF810 — 28Gbps multi-rate 8-channel retimer (ultra-low latency class) for long/lossey links.
- Marvell Alaska C 88X5113 — 25G×4 / 100G Ethernet PHY family with RS-FEC support and training.
4) OTN framer / FEC / mapping (transport interface boundary)
- Criterion: supported client mapping & frame modes required by the target backhaul network.
Why: mapping gaps turn into “works on one network only.” Verify: interoperability checklist + bring-up evidence. - Criterion: FEC overhead vs latency budget + margin visibility.
Why: stronger FEC can save BER but costs latency; stability is proven by margin/counter trends. Verify: corrected/uncorrected counters + margin proxy under stress. - Criterion: counter quality and alarms (thresholds, latching, time-stamped export).
Why: counters are the primary field evidence for “why throughput fell” or “why link degrades.” Verify: exported counters with timestamps and pre/post window snapshots. - Criterion: recovery behavior (reconfig/retrain time) and deterministic state transitions.
Why: field KPI is downtime and recovery time. Verify: scripted failure injection + recovery-time distribution.
- Microchip PM5990 / PM5991 — DIGI-G4 OTN processor family (OTN processing / mapping class).
- Microchip PM6010 — DIGI-G4 ecosystem device (high-capacity transport interface building block).
5) Sensors / power monitors / supervisors (field evidence enablers)
- Criterion: sensor placement coverage: RF chain hot spots + SerDes/optics zone + clock section.
Why: many faults are thermal; without coverage, root cause is missed. Verify: thermal step test + event window snapshots. - Criterion: rail monitoring accuracy and alert latency (for brownout/overcurrent correlation).
Why: supply events can masquerade as jitter or BER issues. Verify: alert timestamp aligned with BER/FEC/PLL state. - Criterion: multi-rail supervisor with programmable delays + watchdog hooks (node-local stability).
Why: clean resets prevent silent corruption and repeated flaps. Verify: reset cause, watchdog events, and restart timeline logged. - Criterion: event logging format: pre/post trigger window, minimum fields, deterministic timestamps.
Why: field issues become reproducible only with windowed evidence. Verify: forced trigger produces a complete evidence bundle.
- TI INA238 — 85V digital current/voltage/power monitor with alert (rail evidence & correlation).
- TI TMP117 — high-accuracy digital temperature sensor (thermal evidence anchor).
- TI TPS386000 — multi-rail voltage supervisor with programmable delay + watchdog (reset integrity).
How to use this checklist in a design review
- Pick the dominant KPI (EVM / BER / FEC margin / ref switch disturbance) and map it to the consumer block (LO / SerDes / framer).
- Select candidates only if evidence is exportable (counters + lock states + windowed logs) and can be time-aligned.
- Write the verification script before final BOM: ref switch + temperature sweep + PRBS/BER + FEC trend + recovery time histogram.
H2-12 · FAQs (Edge Backhaul Node)
Each answer follows a field-ready pattern: likely causes → fastest evidence checks → a short verification path (counters, logs, and targeted tests).