Small Cell / DAS RU: RF Tx/Rx, Timing/Sync, Ethernet & PoE
← Back to: Telecom & Networking Equipment
Small Cell/DAS RUs succeed when RF performance, timing/sync, Ethernet robustness, and PoE power/thermal limits are engineered as one measurable system. This page shows where each function lands inside the RU and how to validate stability with evidence (counters, waveforms, and temperature correlation) for production and field troubleshooting.
H2-1 · What this page covers (Small Cell / DAS RU as a deployable box)
This page focuses on how a Small Cell / DAS Remote Unit turns three constraints—RF chain, timing/sync, and PoE-powered Ethernet—into a manufacturable, maintainable field box. The goal is not a textbook RF overview, but an engineering map of what must be inside, where failures show up, and how to validate the design before deployment.
- In-scope: RU box partitioning (RF/IF/timing/Ethernet/PoE), power & thermal limits, sync integration as an endpoint, and field validation.
- Out-of-scope: DU/CU baseband compute, O-RU system stack (eCPRI/JESD/DPD deep dive), and AAS/massive-MIMO multi-channel beamforming.
Small Cell RU vs DAS RU (differences that change the hardware)
| Design dimension | Small Cell RU (typical impact) | DAS RU (typical impact) |
|---|---|---|
| Output power & thermal density | Higher risk of PA derating in outdoor enclosures; thermal path dominates EVM/ACLR stability. | Power may be lower, but distributed nodes amplify maintenance/ESD exposure and long-term drift. |
| Channel count & RF partitioning | Often 1–2 chains; RF performance is limited by PA/LNA bias, filtering, and port protection. | Multiple remote endpoints make uniformity and field swapability key: calibration & monitoring must be simple. |
| Interface & power delivery | Ethernet backhaul/fronthaul with strict uptime; PoE limits can cap Tx power or duty cycle. | Longer cabling and more touch points increase surge/ESD and mis-wiring risk; protection must be tolerant. |
Three bottlenecks to design around (and how they show up in the field)
- Why it bites: cable loss + temperature derating + PD current limits reduce usable power under peak load.
- Typical symptoms: unexpected resets, Tx power caps, link flaps during bursts.
- What to log/measure: PD state, input droop, DC/DC current limit events, reboot counters.
- Why it bites: PA and power stages drift with temperature; small heatsinks turn minor inefficiency into major EVM/ACLR loss.
- Typical symptoms: EVM rises after warm-up, ACLR degrades, PA derates at repeatable case temperatures.
- What to log/measure: hotspot sensors, derating events, Tx power vs temperature curves.
- Why it bites: jitter/wander and sync loss turn into RF impairment or service drops; recovery behavior matters as much as lock accuracy.
- Typical symptoms: PTP lock loss, holdover entry/exit “glitches”, brief outages that look like RF faults.
- What to log/measure: lock state, holdover active time, offset threshold crossings, link reconnect count.
- A one-glance block diagram (RF / IF AFE / timing & sync / Ethernet & PoE power).
- Practical budgets: PoE usable power vs peak load, thermal derating triggers, sync holdover targets.
- Where to place timing/sync blocks in an RU, and what to monitor for stable field behavior.
- A validation and troubleshooting path from symptoms → evidence → root cause.
H2-2 · Deployment-driven requirements (turn installation reality into measurable targets)
Small cell and DAS remote units fail in the field for reasons that rarely appear in lab-only block diagrams: cabling losses, temperature swings, ESD/surge events, and sync recovery behavior. This section converts deployment conditions into six measurable requirements that drive every later design choice.
Deployment scenarios (what they break first)
- Risk pattern: frequent touch points and long cable runs amplify ESD/surge exposure and mis-wiring.
- Common “silent failure”: protection/grounding changes increase Ethernet errors or raise RF noise floor over time.
- Early validation focus: post-ESD functional drift checks (EVM/BER vs baseline) and remote alarms/logging completeness.
- Risk pattern: solar load and enclosure thermal resistance push PA and power stages into derating.
- Common misdiagnosis: thermal drift looks like an RF design problem (EVM/ACLR worsens after warm-up).
- Early validation focus: temperature sweeps with repeatable derating thresholds and stable sync recovery behavior.
- Risk pattern: “fix one thing, break another” (e.g., added EMI parts increase PHY loss or PoE droop sensitivity).
- Common failure: borderline designs pass the lab but fail under combined stress (heat + load + sync disturbances).
- Early validation focus: A/B testing of protection/filtering changes against BER, EVM, and reboot counters.
Six measurable requirements (each must have a pass/fail check)
- P_budget (usable PoE power): measured under peak traffic, worst-case cable, and temperature—validated by droop + PD/DC/DC event counters.
- T_case (enclosure thermal limit): pass/fail defined by derating onset and EVM/ACLR stability after warm-up.
- Surge/ESD robustness: not only “no damage”, but “no performance drift” (BER/EVM baseline compare after strikes).
- Holdover target: how long the RU maintains service when sync disappears, and how cleanly it recovers (no large step/glitch).
- Ethernet link quality: beyond link-up—BER/packet loss/reconnect rates under EMI and cable stress.
- Remote management coverage: alarms + logs must explain field symptoms (sync state, thermal derate, PD state, error counters).
With these requirements defined, the rest of the page maps them to concrete blocks: RF chain (H2-3/4), IF AFE (H2-5), clocking and sync (H2-6/7), Ethernet robustness (H2-8), and the PoE power tree (H2-9).
H2-3 · RF signal chain partition (Tx/Rx blocks, responsibility, and test points)
A remote unit RF front-end should be partitioned as measurable blocks with clear spec ownership and test points. This prevents “RF mystery failures” by tying field symptoms (EVM drift, power drop, receive sensitivity loss) to where to measure and which block is responsible.
- Spec focus: output power, linearity, ACLR/EVM stability across temperature.
- Failure signature: warm-up drift (EVM/ACLR worsens after minutes) often correlates with PA thermal state and supply headroom.
- Evidence to collect: PA output coupler power vs temperature, and coupler power vs supply droop under peak load.
- Spec focus: noise figure (NF), blocker tolerance, and intermodulation behavior under strong adjacent signals.
- Failure signature: “good in lab, bad in site” often indicates blocker-driven compression or front-end leakage rather than weak-signal issues.
- Evidence to collect: post-LNA noise floor and gain under controlled blocker injection.
- TDD switching: switch isolation and leakage determine how much Tx energy re-enters the Rx chain.
- Common symptom: Rx degradation that appears “random” but is repeatable at specific duty cycles or Tx power levels.
- Evidence to collect: leakage proxy measurements (before LNA) correlated with Rx sensitivity/EVM changes.
- Coupler placement: PA-output vs antenna-side coupling measures different “power truths” (including filter/duplex effects).
- Detector role: provides a stable observable for closed-loop power control and protection triggers.
- Validation action: compare detector reading vs external power meter over temperature and band to confirm calibration granularity.
- TPs exist and are accessible: PA output coupler, post-LNA, and IF out.
- Each block has a primary owner: PA linearity, LNA noise, filter out-of-band rejection, switch isolation.
- Each symptom has evidence: EVM drift ↔ thermal/power; Rx loss ↔ compression/leakage; power drop ↔ derating/protection.
H2-4 · PA/LNA biasing & protection (from “can transmit” to “stays stable in the field”)
In small cell and DAS remote units, the highest field failure rate often comes from biasing and protection, not from the RF block diagram itself. A robust design treats PA/LNA bias as a controlled system with measured sensors, debounced thresholds, and explicit recovery behavior.
- Ramp behavior matters: soft enable prevents supply droop and avoids false protection triggers.
- Temperature compensation goal: stabilize linearity and reduce warm-up drift, not maximize bias current.
- Validation action: capture input droop, PA current peak, and time-to-stable EVM for multiple start/stop profiles.
- Noise path: supply ripple and ground movement can raise the post-LNA noise floor or inject modulation artifacts.
- Blocker realism: ripple effects become visible under strong adjacent signals where headroom is reduced.
- Validation action: measure post-LNA noise floor while sweeping supply ripple and verifying sensitivity degradation thresholds.
- Trip: immediate shutoff for safety-critical events; highest service impact.
- Derate: controlled reduction (power, duty, bias) to keep service stable; requires smooth curves to avoid oscillation.
- Latch-off: prevents repeated damage, but demands clear remote diagnostics and safe recovery procedures.
- Power drop: check thermal derate events → current limit → VSWR triggers → bias stability.
- EVM degradation: check PA bias/thermal stability → rail noise → leakage/compression evidence.
- Intermittent resets: check input droop and protection trip loops; confirm the event log explains every restart.
- No false trips: thresholds include debounce/blanking, and normal transients do not trigger protection.
- Predictable behavior: trip/derate/latch-off actions are deterministic and recover cleanly.
- Actionable logs: every protection event writes a reason code, sensor snapshot, and a timestamp.
H2-5 · IF AFE (up/down conversion + VGA/AGC) — dynamic range and interference recovery
The IF analog front end determines whether a remote unit keeps working under real interference: it sets the usable dynamic range, controls how quickly saturation recovers, and shapes which noise/spur paths become visible as EVM/ACLR degradation. This section stays on analog chain + clocks + dynamic range + lightweight calibration, and does not cover JESD or baseband protocol details.
- Image vs filtering: IF placement shifts how hard image suppression and channel filtering must work.
- LO leakage visibility: poor isolation can create fixed “signature tones” that look like unexplained spurs.
- Spur classification: distinguish fixed-location spurs (reference/divider leakage) from configuration-linked spurs (synth/mixing products).
- Loop bandwidth: too fast causes gain “hunting”; too slow causes long outage after bursts or blockers.
- Saturation recovery: the most practical metric is time-to-stable after a strong interferer disappears.
- Detection point choice: gain control is only as good as the measurement point used for AGC decisions.
- Noise floor path: IF noise + VGA gain distribution sets the baseline EVM floor under weak signal.
- Compression path: blocker → IF stage compression → in-band distortion products → EVM rises under load.
- Phase noise path (at IF): LO phase noise can translate into close-in noise skirts that look like “mysterious” EVM loss.
- DC offset: measure a “quiet” baseline condition and apply a small correction to avoid false clipping and biased AGC decisions.
- IQ imbalance: detect image residue and apply a minimal coefficient update (factory + periodic service window).
- When to rerun: temperature transitions, reference changes, or repeated saturation events that shift baseline behavior.
- No “hidden outage”: saturation recovery time is bounded and repeatable under burst/blocker tests.
- AGC is stable: no sustained oscillation or gain hunting across expected blocker levels.
- Spurs are classifiable: fixed vs configuration-linked spurs can be separated and tied to a path.
- Calibration is serviceable: DC/IQ calibration can run without deep baseband involvement and leaves logs/flags.
H2-6 · Frequency synthesis & clock tree — phase noise, spurs, and domain-aware distribution
In a compact RU, clocking problems often appear as “RF issues” (EVM drift, ACLR shoulders, intermittent lock loss). A workable design treats frequency synthesis as a system: reference source choice, synthesizer spur management, and a clock tree that respects domain sensitivity. This section builds a minimum jitter-cleaning chain and prepares the ground for timing/sync sections without diving into PTP/SyncE protocol mechanics.
- Power/thermal limits: PoE and sealed enclosures constrain warm-up power and steady dissipation.
- Stability goal: pick the source based on required drift/hold behavior of the RU, not as a standalone “best clock.”
- Verification: track drift and lock behavior across temperature ramps and power cycles.
- Phase-noise path: LO phase noise can translate into modulation error and spectral regrowth.
- Spur management: classify spurs by behavior (fixed vs configuration-linked) and tie them to reference/divider/mixing origins.
- Verification: record spur presence and EVM/ACLR correlation across channel plans and temperature corners.
- RF LO domain: sensitive to phase noise and spur injection.
- IF sampling domain: sensitive to jitter that degrades sampling accuracy and error vector stability.
- Ethernet PHY domain: sensitive to wander and lock/relock behavior over long time scales.
- Ref in: stable reference delivered to a single “cleaning boundary.”
- Jitter cleaner: provides a controlled output and (optionally) a hold behavior for short disturbances.
- Fanout: isolates domains so noisy loads do not contaminate RF/IF sensitive paths.
- Spurs are explainable: each major spur class can be tied to a known origin and configuration dependency.
- Domain isolation works: PHY activity or management clocks do not inject spurs/jitter into RF/IF domains.
- Lock behavior is deterministic: relock time and temperature drift behavior are measurable and repeatable.
H2-7 · Timing & sync integration — PTP/SyncE at the RU endpoint
A Small Cell / DAS RU must treat timing as an endpoint engineering problem: where sync lands in hardware, how lock health is monitored, and how the RU behaves under link disturbances. The focus here is placement + observability + holdover + controlled degradation inside the RU, not network-wide timing switch design.
- Hardware timestamps: timestamps must be taken at a deterministic boundary (near MAC/PHY) to avoid variable latency.
- 1PPS / ToD hooks: a simple pulse/time interface makes step events and drift measurable during commissioning.
- Clock health exposure: jitter-cleaner lock state and reference selection state must be readable by management firmware.
- Input relationship: SyncE arrives with Ethernet PHY frequency and becomes one candidate for the RU reference selector.
- Output relationship: once selected and cleaned, the RU clock tree distributes frequency to RF/IF and transport blocks.
- Health signals: “present/absent,” “quality change,” and “switch events” are more actionable than protocol terminology.
- Triggering: holdover starts when SyncE/PTP inputs are lost or declared unhealthy (link-down is not the only trigger).
- Business constraints: define acceptance using RU-visible KPIs (frequency error budget, EVM stability, relock behavior).
- Policy: short holdover may freeze settings; prolonged holdover should enter a controlled degradation mode.
- State: locked / holdover / free-run, plus selected reference source and cleaner lock state.
- Events: ToD step detected, lock-loss event, reference switch event, relock attempt event.
- Counters: PTP offset over-limit count, holdover enter count, relock attempts, ToD step count.
- Deterministic lock behavior: relock time and reference switching are repeatable across power cycles and temperature ramps.
- ToD step visibility: time-step detection produces a log event and a counter increment with a timestamp.
- Holdover policy is enforced: after a defined holdover duration or KPI drift, the RU enters a controlled degrade mode.
H2-8 · Ethernet fronthaul/backhaul interface — PHY, isolation, and surge-resilient stability
RU Ethernet design is not just “link up”: it must remain stable under outdoor ESD/surge events, unpredictable cabling, and tight EMI constraints in a sealed enclosure. This section focuses on RU-side port shape, PHY selection, isolation placement, and protection parasitics that can quietly degrade signal integrity.
- Dual ports: require two complete protection + magnetics chains to avoid cross-coupled failures.
- Service reality: field plug/unplug events and cable unknowns demand robust ESD handling and link retraining.
- Ring mention only: treat ring as “fast recovery needed,” without diving into ring protocol details.
- Temperature range: stable behavior through outdoor thermal swings and hot enclosure conditions.
- EMI behavior: common-mode noise tolerance and predictable emissions under real cable conditions.
- Power budget: PHY power and heat must fit a sealed RU thermal budget.
- Cable margin: tolerance to return loss and cable quality variance (field reality, not lab cables).
- Why isolate: ground potential differences and surge common-mode currents must not enter sensitive RU grounds.
- Placement: port → protection zone → magnetics → PHY defines a clear boundary for energy diversion vs signal integrity.
- Shield bonding: the shield-to-chassis point must be deliberate to prevent common-mode current roaming.
- TVS capacitance: protection capacitance can reduce eye margin and increase intermittent link drops.
- CMC behavior: poor placement or selection can convert common-mode energy into differential distortion.
- Verification: correlate link flap/CRC counters with surge tests and check for retrain storms after events.
- After ESD/surge: link retrains and returns to steady state without repeated flapping.
- SI margin: protection parasitics do not cause chronic CRC/PCS error growth under worst-case cabling.
- Logs: link flap count, retrain count, CRC error count, and event timestamps are collected for field diagnosis.
H2-9 · PoE PD power tree (802.3af/at/bt) — budget, sequencing, isolation, and rail ownership
In Small Cell / DAS RUs, PoE is not just “power delivery.” It sets the real ceiling for RF output, thermal headroom, and field stability. A usable design starts with worst-case available power, then builds a PD + isolated conversion chain with inrush control, multi-rail sequencing, and fault visibility.
- Classified vs usable: PoE class sets the upper bound, but cable loss and hot conditions reduce usable power.
- Worst-case planning: budget with long/poor cabling and elevated ambient, then bind RF output modes to that budget.
- Margin discipline: reserve power for inrush, transient bursts, and recovery states to avoid repeated brownouts.
- Classification/handshake: negotiation outcomes must be readable (class, power granted, retry reason).
- Inrush limiting: uncontrolled inrush often looks like “random reboot” during cold starts or plug-in events.
- Thermal protection: PD/bridge temperature limiting can silently cap power and trigger cascading rail drops.
- Power allocation: enforce rail priorities so PA bursts do not starve low-noise analog or management rails.
- Isolation boundary: keep the boundary explicit and treat it as the anchor for EMI filtering and return control.
- EMI filtering: place filtering to control where common-mode and differential energy flows, not just to “add parts.”
- Ground partition: high-power PA returns must not share sensitive analog return paths by accident.
- PA rail: owns peak power, transient load steps, and derating behavior under thermal constraints.
- Low-noise analog rails: own noise floor and spur cleanliness for LNA/AFE/clock-sensitive blocks.
- Digital & management rails: own determinism: watchdog, reset, PG gating, fault logging, and safe retry policies.
- Startup determinism: cold/hot starts succeed without repeated brownouts or negotiation loops.
- Rail priority holds: PA bursts do not collapse low-noise or management rails.
- Fault visibility: PD class, inrush events, PG faults, and latch reasons are logged with timestamps.
H2-10 · Thermal, mechanics & environmental hardening — derating as a closed loop
RU reliability is decided by heat and environment. The goal is not a generic “thermal discussion,” but a closed-loop derating system: identify dominant heat sources, define thermal paths to the enclosure, place sensors that reflect both case and hotspots, and execute deterministic actions that protect RF performance without oscillating between states.
- PA: highest power density, directly affects output capability and linearity under heat stress.
- DC/DC: efficiency and switching losses create localized hotspots near magnetics and power stages.
- PHY/SoC: management heat matters in sealed boxes, especially when cable conditions force higher transmit effort.
- Path chain: silicon → PCB copper → interface material → enclosure → ambient.
- Bottlenecks: interface pressure, pad aging, and enclosure coupling often dominate long-term drift.
- Design intent: thermal paths must be explicit so sensor readings can explain behavior under load.
- Tiered actions: reduce PA output power first, then apply broader service limitations if temperature continues rising.
- Stability: use hysteresis and minimum dwell times to avoid rapid oscillation between states.
- Recoverability: define clear exit conditions for each derate level and record the transition reason.
- Case sensor: tracks enclosure and environment, useful for long-term trend and site conditions.
- Hotspot sensors: near PA and power stages to protect silicon and prevent runaway heating.
- False readings: wrong placement causes false derating or missed overheating—both create field failures.
- Worst-case stability: RU reaches a stable operating point or a stable derated state without oscillation.
- Explainable transitions: every derate entry/exit is explainable via sensor readings and logged reasons.
- Field-aligned logging: log temperature peaks, dwell times, and derate levels for site diagnosis.
H2-11 · Validation & troubleshooting checklist — turning field failures into repeatable lab workflows
This section defines “done” using evidence: deterministic bring-up, RF stability under temperature and VSWR stress, time/sync resilience (lock → holdover → relock), and EMC survival without hidden degradation. Each checklist block follows the same format: Setup → Procedure → Pass/Fail → Evidence.
- Test at cold and hot conditions (ambient extremes + sealed enclosure steady state).
- Use at least two cable cases: short/good cable and long/worst cable.
- Enable RU logging for PoE states, PG/reset causes, and Ethernet link events.
- Power-cycle (≥10 times) and record success/failure rate and time-to-ready.
- Perform fast unplug/plug events and observe negotiation retries and brownout behavior.
- Force peak load during bring-up (management + PHY active + RF idle then RF enable).
- Verify sequencing: management rails stable first, then analog/PLL rails, then PA high-power rail.
- Pass: deterministic bring-up without repeated resets or negotiation loops.
- Pass: PG chain is monotonic (no oscillation); no “phantom” PG drop under normal transients.
- Fail: periodic resets, repeated PoE renegotiation, or link flapping during rail enable.
- PoE state timeline: class/allocated power, inrush event markers, retry reason codes.
- Key waveforms: inrush current, isolated bus droop, PG edges, reset line, rail ramp timing.
- Ethernet: link-up time, link flap count, CRC/PCS error counters during and after bring-up.
- Use a calibrated VSA/power meter and a controlled thermal condition (steady-state at multiple points).
- Prepare a controlled mismatch / VSWR injection method (step-by-step, reversible).
- Ensure PA bias/protection telemetry is readable (overtemp, overcurrent, VSWR trip/derate).
- Measure output power and modulation quality at nominal temperature.
- Repeat after thermal soak (hot enclosure) and after cold start (if applicable).
- Apply VSWR steps: normal → moderate mismatch → fault injection; record RU response each step.
- Confirm protection policy: derate vs trip vs latch-off, and verify recovery conditions.
- Pass: performance drifts are explainable (temperature/rail/bias) and do not cause uncontrolled oscillations.
- Pass: VSWR events trigger the intended action (derate/trip) and are logged with clear reason codes.
- Fail: EVM/linearity collapses with no corresponding thermal/rail evidence or protection state change.
- RF results: power, spectrum, and modulation quality snapshots per temperature and VSWR step.
- Bias/rails: PA rail droop, bias DAC/driver state, protection trip counters, hotspot temperatures.
- Event correlation: timestamped link between RF degradation and thermal/PoE/sync state transitions.
- Enable RU sync status outputs: lock state, holdoverActive, reference selection state.
- Prepare a controlled link disturbance (PTP interruption and/or link flap) and a ToD step scenario.
- Keep RF and Ethernet traffic at representative load (not idle-only validation).
- Record steady lock health under normal link conditions (baseline counters and offset behavior).
- Interrupt PTP and observe: lock loss detection → holdover entry → alarm behavior → RF impact.
- Restore PTP and measure: relock time and stability (no repeated lock oscillation).
- Inject a time-step event and verify that step is detected, logged, and handled as defined.
- Pass: holdover entry/exit is deterministic and visible; relock completes within the defined window.
- Pass: ToD step events generate counters/logs and do not create silent time jumps.
- Fail: repeated lock oscillation, unlogged step events, or sync failures that masquerade as “RF issues.”
- Lock state timeline: lostPTP, lostSyncE (if used), holdoverActive, reference switch events.
- Counters: offset over-limit count, step count, relock attempts, holdover enter count.
- Correlation: RF/Ethernet KPI changes aligned to sync state transitions.
- Run “before” baselines: RF quality, Ethernet error counters, sync stability, and thermal behavior.
- Enable persistent logs with timestamps and retain them across resets.
- Apply the defined ESD/surge events at the port and chassis points per the RU test plan.
- Re-run the same baseline measurements and compare deltas (not just “still alive”).
- Check for latent issues: link flap storms, CRC growth, sync offsets becoming noisy, RF quality drift.
- Pass: no persistent KPI regression and no growth in error counters beyond defined tolerance.
- Fail: “works” but with higher EVM, higher packet errors, or unstable sync/holdover behavior.
- Delta report: pre/post event RF snapshots + Ethernet/sync counter deltas.
- Event markers: ESD/surge time stamps aligned with resets, link flaps, and protection trips.
- PoE / power: class/granted power, inrush events, bus undervoltage events, PG fault source, retry reason.
- Thermal: case temp, hotspot temp peaks, derate level, dwell time per derate level, thermal limit events.
- Sync: lock state, holdoverActive, reference selection, offset over-limit count, step count, relock attempts.
- Ethernet: link flap count, CRC/PCS errors, retrain count, timestamped link transitions.
- RF protection: VSWR trip/derate count, PA overtemp count, PA overcurrent count, power-backoff state.
These are examples to anchor BOM searches and validation mapping. Always verify current datasheets and availability.
- TI: TPS2372 / TPS2373 families
- Analog Devices: LT4294 / LT4293 families
- Analog Devices: LT4295 (PD + isolated conversion controller class)
- Skyworks: Si5345 family
- Analog Devices: AD9546 family
- Marvell Alaska: 88E15xx families
- Microchip: VSC85xx families
- Analog Devices: ADL5902 (RMS detector class)
- Analog Devices: AD8318 (log detector class)
H2-12 · FAQs (Small Cell / DAS RU)
These FAQs target common field symptoms (EVM drift, link errors, sync loss, intermittent resets) and map each answer to the relevant section for deeper troubleshooting steps.