Gateway / Bridge Controller for Industrial Ethernet & TSN
← Back to: Industrial Ethernet & TSN
A Gateway / Bridge Controller preserves VLAN, QoS, and PTP time semantics while aggregating multiple field protocols into Ethernet uplinks. This page turns forwarding, buffering, observability, and recovery into measurable budgets and checklists so performance and diagnosability remain deterministic in real deployments.
H2-1. Definition & Scope Guard (Gateway / Bridge Controller)
Lock the page boundary upfront: define gateway/bridge behavior, establish measurable success criteria, and prevent “encyclopedic” overlap with sibling topics.
Working definition (behavior-level)
A Gateway/Bridge Controller aggregates multiple ingress domains into Ethernet uplink by enforcing segmentation (VLAN), priority behavior (QoS/queues), and time semantics preservation (PTP-aware forwarding), with built-in observability for field diagnosis.
Scope guard (hard boundary)
- Bridge pipeline: classify → queue → schedule/shaping → forward.
- VLAN segmentation and field-to-uplink mapping strategy.
- QoS policy mapping (PCP/DSCP → queues) and congestion behavior.
- PTP-aware forwarding behavior: timestamp tap points and time semantics preservation.
- Latency/jitter budgeting within the gateway/bridge box (p99/p50 and worst-case bounds).
- Observability: counters, mirroring hooks, black-box fields for forensics.
- TSN scheduling tables and mechanisms (Qbv/Qci/GCL) → TSN Switch / Bridge
- PTP/SyncE/WR clocking theory and templates → Timing & Sync
- Industrial protocol stacks internals (PROFINET/EtherCAT/CIP) → Industrial Ethernet Stacks
- PHY/magnetics/ESD/surge layout details → PHY Co-Design & Protection
- PoE/PoDL power classification and thermals → PoE / PoDL
Gateway vs Bridge vs Router (behavior differences)
| Dimension | Bridge | Gateway / Bridge Controller | Router |
|---|---|---|---|
| Primary function | L2 forwarding between ports/domains. | Multi-domain aggregation to Ethernet with policy + time semantics + observability. | L3 forwarding across IP subnets (routing control plane not expanded here). |
| State kept | MAC learning table (optional). | MAC/VLAN mapping + per-class queues + policy counters + black-box logs. | IP routes, ARP/ND, ACL/NAT (not expanded on this page). |
| Segmentation | Basic VLAN tagging/untagging. | Field-to-uplink VLAN strategy, trunking, and isolation guarantees. | Subnet boundaries and routing policies. |
| QoS behavior | Limited queues, basic priority. | Deterministic queue mapping + congestion behavior (no starvation) with measurable bounds. | Policy-based forwarding and shaping at L3/L4. |
| Time semantics | Usually unaware of PTP residency effects. | PTP-aware forwarding: defined timestamp taps, bounded queue-induced variation, and auditability. | Time handling depends on platform; not expanded here. |
| Diagnostics | Basic port counters. | Per-class counters + mirroring triggers + black-box event fields (field forensics). | Depends on routing stack; not expanded here. |
This page stays at the behavior level. Protocol-stack internals, TSN scheduling, and clocking theory are intentionally linked out.
Typical placement and I/O boundary
- Placement: Aggregation point between cell/line networks and plant/edge uplink (the “narrow waist” where policy and observability matter).
- Inputs: Multiple ingress domains (field networks, serial/IO aggregation, or segmented device clusters).
- Output: Ethernet uplink (often VLAN trunk) with explicit QoS behavior and preserved PTP time semantics.
Success criteria (measurable, with placeholders)
Owner map (prevent topic overlap)
| Topic / Question | Owned here? | Go to | Reason (boundary rule) |
|---|---|---|---|
| Field-to-uplink VLAN mapping and trunk rules | YES | This page | Bridge-level segmentation strategy; avoids switch-implementation details. |
| Queue mapping and QoS behavior under congestion | YES | This page | Policy intent and measurable bounds at the gateway boundary. |
| TSN schedule tables (Qbv/Qci/GCL) and admission control | NO | TSN Switch / Bridge | Scheduling mechanics belong to TSN switching, not gateway boundary definition. |
| PTP/SyncE algorithm and clock templates | NO | Timing & Sync | Clocking theory is separate; this page focuses on PTP-aware forwarding behavior only. |
| PHY/magnetics/ESD/surge layout and component selection | NO | PHY Co-Design & Protection | Signal integrity and protection are owned by the PHY co-design page. |
If only three things matter
- Policy boundary: VLAN/QoS mapping must be explicit and testable under congestion.
- Time boundary: PTP-aware forwarding needs defined timestamp taps and bounded queue-induced variation.
- Forensics boundary: Counters + black-box fields must make field failures reproducible.
The diagram highlights a single boundary: multi-domain ingress → bridge pipeline → Ethernet uplink, with explicit VLAN/QoS behavior, PTP-aware handling, and mandatory observability hooks.
H2-2. Where It Sits: Topologies & Traffic Profiles
Map the gateway/bridge to real plant layouts and define traffic buckets early, so later VLAN/QoS and latency criteria have a clear measurement denominator.
Placement by topology (what typically breaks first)
- Bridge sits at the uplink exit of chained device clusters.
- First failure mode: burst accumulation (queue fill) at uplink chokepoint.
- Primary control: queue headroom + strict separation of control traffic.
- Bridge is the aggregation center with many ingress ports.
- First failure mode: uplink congestion causing priority inversion/starvation.
- Primary control: explicit QoS mapping and measurable fairness bounds.
- Bridge is either a ring node or a boundary out to uplink.
- First failure mode: loops and broadcast amplification (“storm”).
- Primary control: storm guards + fast isolation + forensic visibility.
Traffic buckets (four-bucket model)
The four buckets define a stable denominator for policy: VLAN/QoS mapping must protect cyclic control and time-sensitive traffic from bursty diagnostics or bulk flows.
Conflict matrix (symptom → first bridge-layer check)
| Conflict | Observable symptom | First bridge-layer check | Pass criteria (X) |
|---|---|---|---|
| Cyclic control vs bulk burst | Control jitter spikes while average utilization looks low. | Peak-window utilization + per-queue depth + priority mapping. | p99 jitter ≤ X µs at Y% peak. |
| Diagnostics storm vs steady traffic | Queue drops rise; recovery oscillates (“flap”). | Rate-limit diagnostics class + drop reason counters + event triggers. | Drops bounded ≤ X/hour. |
| Broadcast/loop amplification | Network appears “stuck” despite low average load. | Peak frame rate + storm counters + fast isolation behavior. | Isolation ≤ X ms; service recovers. |
| Monitoring overhead vs determinism | Enabling mirroring increases jitter. | Mirror sampling/trigger policy + reserved queue headroom. | Jitter delta ≤ X µs when enabled. |
Measurement denominators (define before tuning)
- Utilization: track both average and peak-window (e.g., X ms window) to reveal burst-driven queue fill.
- Queue depth: per-class depth and drop reason (overflow vs policing) as the primary congestion truth.
- Latency/jitter: report p50/p95/p99 under the defined traffic mix and topology.
- Event traces: config changes, link flaps, and guard actions must be timestamped for correlation.
Deliverable · Traffic profile table (template)
| Traffic class | Rate (avg) | Burst (peak window) | Priority / Queue | Max jitter | Pass criteria (X) |
|---|---|---|---|---|---|
| Cyclic control | X Mbps | X Mbps @ X ms | Queue-0 (strict) | ≤ X µs | p99 jitter ≤ X |
| Event | X Mbps | X Mbps @ X ms | Queue-1 | ≤ X ms | drops ≤ X/hour |
| Diagnostics | X Mbps | X Mbps @ X ms | Queue-2 (policed) | ≤ X ms | util peak ≤ X% |
| Bulk data | X Mbps | X Mbps @ X ms | Queue-3 (best-effort) | ≤ X ms | no starvation |
The overlay establishes topology-specific denominators: burst-driven queue fill (line), uplink congestion and fairness (star), and storm amplification + fast isolation (ring).
H2-3. Functional Architecture: Data Plane vs Control/Management Plane
Separate deterministic forwarding hardware from configuration and operations paths. Later chapters bind performance topics to the data plane and maintainability topics to the control/management plane.
Two-plane model (engineering definition)
- Per-packet path: ingress → classify → queue → schedule/shaping → egress.
- Primary metrics: throughput, p99 latency, p99 jitter, drops, timestamp variation.
- Primary requirement: bounded behavior under peak-window bursts.
- Configuration and policy lifecycle: apply, audit, rollback.
- Operations: counters, telemetry, alerts, secure firmware update.
- Primary requirement: reproducible forensics without disturbing determinism.
Management interfaces (interface layer only)
SPI / UART / PCIe / SGMII are entry paths for configuration and observability. The interface itself does not guarantee determinism; the bottlenecks are typically sampling granularity, DMA/CPU contention, and logging policy.
Plane coupling risks (what changes determinism)
Check: jitter delta when enabled ≤ X µs.
Check: track peak-window utilization + max queue depth.
Check: no CRC/drop correlation with CPU load peaks.
Deliverable · Module responsibility matrix (HW vs SW)
| Module | Must be HW | Can be SW | Hybrid | Boundary rule |
|---|---|---|---|---|
| Classifier + queue mapping | YES | — | Config in SW | Affects p99 jitter directly; path must be bounded. |
| Queue/buffer enforcement | YES | — | Policy in SW | Drops/headroom must be reproducible under bursts. |
| Timestamp tap points | YES | — | Export in SW | Time semantics must be fixed and auditable. |
| Counters + drop reasons | Fast path | Aggregation | YES | HW counts; SW exports with bounded overhead. |
| FW update + rollback | — | YES | Health gate | Operability and recovery belong to management plane. |
The diagram highlights a controlled coupling: configuration flows down to the data plane, while counters/telemetry/events flow up with bounded overhead to avoid disturbing tail latency.
H2-4. Bridging Pipeline: Classification, Forwarding, and Buffering
Build a deterministic mental model of how a packet moves through the bridge. Later QoS/VLAN and observability decisions attach to explicit checkpoints (drop, queue, tx).
Classification inputs (field-level, not ACL internals)
- L2 identity: MAC destination/source and ingress port.
- Segmentation: VLAN ID (tag/untag decisions at the boundary).
- Type hint: EtherType (service separation without stack coupling).
- Priority hint: DSCP/PCP mapping to traffic class/queue.
Forwarding database behavior (resources and failure modes)
- FDB capacity: X MAC entries per VLAN domain.
- Aging time: X s (audit impact on mobility vs churn).
- Overflow behavior: flood / drop (must be measurable via counters).
Buffering model (where jitter and drops are born)
- Higher utilization under mixed traffic.
- Needs explicit headroom rules to prevent tail-latency blow-ups.
- Requires drop-reason counters per class.
- More predictable bounds for cyclic control traffic.
- Resource fragmentation must be budgeted per class.
- Backpressure thresholds can be class-specific.
Backpressure and drop policy (capability checklist)
Deliverable · Buffer budget template (queue depth + burst + headroom)
| Traffic class | Peak burst (X ms) | Target queue depth | Reserved headroom | Drop threshold | Pass criteria (X) |
|---|---|---|---|---|---|
| Cyclic control | X Mbps @ X ms | X pkts / X KB | Min reserved | X% | p99 jitter ≤ X |
| Event | X Mbps @ X ms | X pkts / X KB | Shared + cap | X% | drops ≤ X/hour |
| Diagnostics | X Mbps @ X ms | X pkts / X KB | Policed | X% | peak ≤ X% |
| Bulk data | X Mbps @ X ms | X pkts / X KB | Best-effort cap | X% | no starvation |
Budget inputs must use a peak-window definition. Average utilization is insufficient to prove bounded queueing variation.
Three checkpoints converge most field failures: Drop Counter (why frames are rejected), Queue Depth (burst-driven contention), and TX Counter (actual egress behavior).
H2-5. VLAN Strategy: Segmentation, Trunking, and Field-to-Uplink Mapping
Define gateway-side VLAN roles and mapping rules to isolate field domains while converging traffic into a controlled uplink trunk. Focus is on mapping strategy and verifiable boundary behavior.
Scope guard (strategy, not switch internals)
- In scope: access/trunk roles, tag/untag boundaries, field-to-uplink VLAN mapping, native VLAN policy, and audit-friendly templates.
- Out of scope: switch silicon implementation details (TCAM/ACL/snooping), TSN time-window mechanisms, and ring protocol algorithms.
Port roles and boundary semantics
- Field devices often send untagged frames.
- Gateway applies VLAN tagging at ingress or strips tags at egress.
- Boundary must be explicit: which VLAN is assigned to untagged ingress.
- Uplink carries multiple VLANs with an allow-list.
- Native VLAN (if used) must be declared and audited.
- Mapping rules define how many field VLANs converge into how many uplink VLANs.
Field-to-uplink VLAN mapping strategies (behavior-level)
Common pitfalls and first checks (gateway-side)
First check: uplink native VLAN explicitly set and documented.
First check: trunk allow-list contains the mapped VLANs; port role is correct.
First check: isolate suspected VLAN domain; compare broadcast counters per port group.
Deliverable · VLAN mapping table (template)
| Field port group | Ingress role | Ingress VLANs | Tag action | Uplink trunk VLANs | Allow-list | Native VLAN | Observability hook | Pass criteria |
|---|---|---|---|---|---|---|---|---|
| Cell-A devices | Access | Field VLAN A | Tag at ingress | Uplink VLAN 9xx | Explicit list | None / X | Ingress tag + uplink tag | No leakage; untagged=0 |
| Maintenance ports | Trunk | Field VLAN B/C | Translate | Uplink VLAN 9xx | Explicit list | X | Per-VLAN counters | No tag gaps |
Use explicit port roles and trunk allow-lists. Treat native VLAN as a controlled exception with audits and counters to prevent silent cross-domain leakage.
H2-6. QoS & Queueing: Priorities, Shaping, and Congestion Behavior
Convert QoS into measurable engineering behavior: how labels map into queues and how congestion changes jitter and drops. Validation focuses on worst-case uplink load, not nominal averages.
Priority inputs (mapping, not full policy stacks)
- 802.1p PCP: L2 priority hint used for deterministic queue selection.
- DSCP: L3 priority hint that can be mapped into L2 traffic classes at the gateway boundary.
- Rule: one service class must map to one queue consistently across field and uplink; mixed mappings break tail latency predictability.
Minimal QoS class set (4 classes)
Queueing behavior (what happens under load)
- Protects Control/Sync tail latency.
- Risk: Diagnostics/Data starvation during sustained congestion.
- Requires explicit minimum service or shaping for non-critical queues.
- Prevents starvation by assigning service shares.
- Risk: Control jitter increases if weights do not reflect peak-window bursts.
- Requires validation at worst-case uplink occupancy.
Congestion behavior (symptom-level, protocol-agnostic)
Deliverable · Minimal QoS policy set + validation checklist
- Control and Sync require dedicated queues with reserved headroom.
- Diagnostics must be rate-bounded to prevent control timing collapse.
- Data may degrade but must not starve indefinitely.
| Class | Ingress label | Queue ID | Scheduling | Shaping | Drop preference | Pass criteria (placeholders) |
|---|---|---|---|---|---|---|
| Control | PCP/DSCP → X | Q0 | SP / WRR | Reserved | Last | p99 jitter ≤ Y @ X% uplink |
| Sync | PCP/DSCP → X | Q1 | SP / WRR | Protected | Late | residence var ≤ Y @ X% |
| Diagnostics | PCP/DSCP → X | Q2 | WRR | Rate cap | Earlier | drops ≤ X/hour @ X% |
| Data | PCP/DSCP → X | Q3 | WRR | Best-effort cap | First | no indefinite starvation |
- Uplink occupied at X%: Control p99 jitter ≤ Y.
- Queue max depth remains below headroom threshold for Control/Sync.
- Drop reasons are separated by class; no ambiguous “unknown drops”.
- Diagnostics is rate-bounded; Data does not starve indefinitely.
The ladder clarifies a minimal deterministic policy: map labels into four queues, enforce headroom for critical classes, and validate behavior at worst-case uplink occupancy.
H2-7. PTP-Aware Forwarding: Timestamp Points and Time Semantics
Preserve time semantics across the bridge pipeline by defining where timestamps are taken, how variable delay is observed, and which forwarding behaviors keep timing interpretable under queueing and shaping.
Scope guard (bridge behavior, not PTP algorithms)
- In scope: one-step/two-step semantic requirements at the bridge, timestamp tap points, and observability of queue-induced variability.
- Out of scope: servo/BMCA/filtering, SyncE/WR derivations, and TSN gate control list mechanics.
Time semantics (engineering definitions)
One-step vs two-step (semantic requirements at the bridge)
- Requires a precise egress tap close to transmission.
- Any last-moment queue variability directly impacts semantics.
- Needs hardware timestamp + fast field update or equivalent mechanism.
- Records the precise egress time and reports it via a follow-up mechanism.
- Allows more flexibility in datapath but requires stable association between frames.
- Still needs consistent tap definition and queue visibility for validation.
Bridge-side capabilities for time-aware forwarding
- MAC ingress / MAC egress tap definitions.
- Optional PHY-side tap exposure (listed as a capability).
- Consistent timestamp domain and rollover behavior.
- Per-class queue depth and residence distribution hooks.
- Drop reasons separated from timing misses.
- PTP frames can be prioritized to avoid deep queues.
- Support for delay-compensation or correction updates.
- Export of residence time metrics for validation.
- Stable association between timestamps and forwarded frames.
Deliverable · Timestamp tap points table (template)
| Tap point | Represents | Dominant error sources | Observability hook | Validation method | Pass criteria (X/Y) |
|---|---|---|---|---|---|
| MAC ingress TS | Rx boundary | CDC/FIFO alignment | Rx TS counter | Dual-ended capture | TS error ≤ X |
| Queue entry TS | Queue boundary | Depth variation | Depth + residence | Load sweep test | p99 ≤ Y |
| MAC egress TS | Tx boundary | Scheduler/shaper | Tx TS counter | Worst-case occupancy | TS error ≤ X |
A stable tap definition plus queue-variability observability keeps timing semantics interpretable even when congestion and shaping introduce variable residence time.
H2-8. Latency & Determinism Budget: What You Can Control
Turn latency and determinism into a budget table by separating fixed processing delay, queueing delay, and shaping delay, then assigning margins and validation conditions for worst-case uplink occupancy.
Delay components (budgetable parts)
What the bridge can control vs external conditions
- Queue separation by service class and reserved headroom.
- Scheduler choice and shaping limits aligned with worst-case bursts.
- PTP/control priority handling and measurement hooks.
- Telemetry rate limits to avoid data-plane interference.
- Uplink bandwidth, upstream congestion, and topology changes.
- Traffic burstiness and broadcast amplification outside the bridge.
- Link-level errors and retransmission patterns in connected equipment.
- Clock stability and environmental variation impacting timing margins.
How to write an end-to-end budget (repeatable format)
- Budget per service class (Control/Sync/Diagnostics/Data) instead of a single blended number.
- For each segment: record baseline, p99, upper bound, and a margin placeholder.
- Attach conditions: uplink occupancy X%, shaping on/off, telemetry rate caps, and worst-case burst window.
- Require a validation method per segment: dual-point capture, counters, or controlled load sweep.
Deliverable · Latency waterfall table (template)
| Segment | Latency baseline | Latency p99 | Jitter p99 | Upper bound | Control lever | Validation method | Pass criteria (X) |
|---|---|---|---|---|---|---|---|
| Field port → ingress | Base | p99 | p99 | Max | Ingress hooks | Dual-point capture | ≤ X |
| Ingress → queue | Base | p99 | p99 | Max | Queue split | Load sweep | ≤ X |
| Queue → egress | Base | p99 | p99 | Max | Scheduler | Worst-case occupancy | ≤ X |
| Egress → uplink | Base | p99 | p99 | Max | Shaping cap | Traffic replay | ≤ X |
Budget segments isolate what the bridge controls (queue/scheduling/shaping) from external conditions. Record baseline, p99 add-on, and margin under explicit uplink occupancy assumptions.
H2-9. Reliability & Recovery: Redundancy Interaction and Fail-Safe Behavior
Define fail-safe behaviors and recovery hooks so link loss, loops, and storms are contained locally instead of cascading into system-wide instability.
Scope guard (behavior hooks, not protocol deep-dives)
- In scope: symptoms, containment points, default fail-safe actions, and recovery/rollback + evidence logging.
- Out of scope: MRP/HSR/PRP packet/state details, TSN scheduling mechanics, and PHY waveform root-cause analysis.
Common failure modes and bridge-side containment points
- Symptom: frequent up/down events, throughput collapse.
- Containment: isolate the port, freeze learning, protect control class.
- Evidence: link-event timeline + per-port error summary.
- Symptom: broadcast/unknown unicast surge, queues saturate.
- Containment: storm control capability, unknown-unicast limit, isolation trigger.
- Evidence: storm meter + queue watermark snapshots.
- Symptom: p99 latency explodes, drops spike under uplink pressure.
- Containment: class protection + rate caps + circuit-breaker rules.
- Evidence: per-queue depth distribution + drop reasons.
Fail-safe defaults (verify as rules, not as slogans)
- Rate limit broadcast/unknown unicast first.
- Isolate a port/VLAN/class if pressure persists.
- Circuit-break (fuse) when thresholds are sustained.
- Degrade to keep Control/Sync alive under crisis.
- Trigger on queue watermark > X or storm meter > X.
- Release only after stable window > Y with counters returning to baseline.
- Record state transitions into black-box with a single event ID.
Restart recovery (persist → rollback → re-enable)
Deliverable · Recovery checklist (template)
| Fault | First observation | Isolation action | Containment check | Recovery steps | Pass criteria (X/Y) |
|---|---|---|---|---|---|
| Link flap | Link events + error spike | Isolate port, freeze learning | Queue depth returns | Re-enable gradually | Stable > Y |
| Loop / storm | Storm meter + drops | Rate cap, isolate offender | Broadcast falloff | Restore in steps | p99 ≤ X |
| Misconfig | Config change event | Rollback snapshot | Counters normalize | Validate baseline | Stable > Y |
Fail-safe rules are validated by explicit triggers (X) and release conditions (Y), with a single event ID capturing state transitions for post-mortem analysis.
H2-10. Monitoring & Diagnostics: Counters, Mirroring, and Black-Box Forensics
Build an observability loop: counters that support triage, mirroring/sampling that can be triggered by anomalies, and a minimal black-box record that enables root-cause reconstruction.
Triage flow (counters organized by decisions)
Counter taxonomy (layers with consistent accounting)
- Link up/down timeline
- Drop totals + error summary
- Utilization in fixed windows
- Depth + watermark
- Drop reason (tail/police)
- Service rate / scheduler stats
- Control / Sync / Diag / Data
- Latency markers (p99)
- Priority mapping hits
Every counter must declare its time window and denominator (per-packet / per-byte / per-second) to avoid “good-looking” metrics that cannot be compared across ports or time.
Mirroring & capture (capability points with triggers)
- Port mirroring by port/VLAN/class
- Sampling mode for long-term monitoring
- Event correlation via a shared event ID
- Drop spike > X per window
- Queue watermark > X
- Link flap within Y
- Storm detected by meter
Deliverable · Minimal black-box fields (MVP template)
| Field name | Purpose | Retention | Trigger | Correlation key |
|---|---|---|---|---|
| event_id | Join counters/log/mirror | N events | Any anomaly | event_id |
| ts_utc | Timeline reconstruction | N days | Any anomaly | event_id |
| port_id / vlan / class | Scope & blast radius | N events | On trigger | event_id |
| temp / power | Environment correlation | N events | On anomaly | event_id |
| queue_watermark / drop_reason | Congestion fingerprint | N events | Depth>X | event_id |
Prefer event-driven snapshots with rate limits to prevent “logging storms”. Store compact fingerprints plus a correlation key that links to mirrored samples.
Taps feed counters/logs and trigger-driven captures. A minimal black-box record stores compact fingerprints plus a correlation key so remote operations can reconstruct the incident and refine thresholds.
H2-11. Engineering Checklist: Design → Bring-up → Production
Convert gateway/bridge requirements into an executable closure loop: define accounting rules in Design, validate minimal risk set in Bring-up, and freeze consistency + thresholds for Production acceptance.
Scope guard (execution-first)
- In scope: VLAN/QoS/PTP-aware behaviors, queue/buffer/latency budgeting, observability hooks, fault injection, and acceptance thresholds.
- Out of scope: magnetics/layout and PHY waveform debugging, TSN Qbv/Qci parameter details, and protocol stack deep dives.
Deliverable · Design Gate checklist (template)
| Item | How to verify | Pass criteria (X) |
|---|---|---|
| Traffic accounting rules defined | Declare time windows + denominators for utilization/drops/latency markers. | All counters have window + denom |
| VLAN mapping table frozen | Field VLANs → uplink trunk mapping reviewed; native VLAN rules explicit. | No ambiguous “untag” paths |
| QoS minimal set defined | Control/Sync/Diag/Data mapping to queues + scheduler behavior documented. | 4 classes mapped end-to-end |
| Queue & buffer budget drafted | Compute burst headroom + queue depth plan; define drop behavior policy. | Budget template filled |
| PTP semantics plan | Define timestamp tap points + visibility; define queue-induced variable delay observability. | All tap points enumerated |
| Observability MVP locked | Per-port/per-queue/per-class counters + event IDs + black-box fields defined. | Schema documented + versioned |
Deliverable · Bring-up Gate checklist (template)
| Item | How to verify | Pass criteria (X) |
|---|---|---|
| Minimum connectivity | Validate L2 forwarding + VLAN tag behavior across all port groups. | No unexpected leakage |
| Throughput + burst absorption | Drive bursts while monitoring queue depth/watermark and drop reasons. | Drops ≤ X, watermark ≤ X |
| Congestion behavior under uplink pressure | Set uplink utilization to X% and verify control class remains serviced. | Control p99 jitter ≤ X |
| PTP-aware semantics visibility | Confirm timestamp taps are observable at defined points; verify variable delay markers exist. | Taps present + stable |
| Fault injection closure | Inject link flap/storm/reboot; verify fail-safe triggers + black-box event ID capture. | Containment ≤ X, recovery ≤ Y |
Deliverable · Production Gate checklist (template)
| Item | How to verify | Pass criteria (X) |
|---|---|---|
| Version + config consistency | Freeze firmware version + config hash; verify export/import round-trip. | Hash match 100% |
| Manufacturing test suite | Run minimal forwarding + congestion + fault injection + counters sanity. | All tests pass |
| Log schema frozen | Verify black-box fields and event IDs remain stable across builds. | Schema version locked |
| Acceptance thresholds applied | Verify throughput/drops/latency markers/recovery time against limits. | p99 ≤ X, recovery ≤ Y |
Engineering inventory (example material part numbers)
The list below provides reference parts commonly used in gateway/bridge controller builds. Final selection must match required ports, time semantics, queue/buffer needs, and compliance targets.
- Texas Instruments AM6442 (Sitara AM64x)
- Texas Instruments AM2434 (Sitara AM243x)
- NXP LS1028A (Layerscape)
- Renesas R9A07G084 (RZ/N2L)
- Microchip LAN9662 (TSN switch family)
- Microchip LAN9373 (TSN switch family)
- Microchip KSZ9477 (managed switch)
- NXP SJA1105T (TSN switch)
- Texas Instruments DP83867 (Gigabit PHY)
- Texas Instruments DP83869 (Gigabit PHY)
- Microchip KSZ9031RNX (Gigabit PHY)
- Microchip ATECC608B (secure element)
- NXP SE050 (secure element)
- Infineon SLB9670 (TPM 2.0 family)
- Winbond W25Q128JV (QSPI NOR)
- Macronix MX25L12835F (QSPI NOR)
- Texas Instruments TPS62130 (buck regulator)
- Texas Instruments TPS62840 (buck regulator)
- Texas Instruments TPS7A20 (LDO)
- Texas Instruments TPD4E05U06 (4-channel ESD)
- Littelfuse SP3012-04UTG (ESD array)
H2-12. Applications & IC Selection Logic
Connect use-case → constraints → required capabilities → verification method → threshold placeholders. Part numbers are included as reference anchors (not as a shopping list).
Application buckets (bridge-controller-centric)
The 12 must-ask questions (to land on the right specs)
- How many field ports and what uplink speed?
- What is the worst-case burst size and cadence?
- At uplink utilization X%, what control jitter limit is required?
- Is storm control / circuit-break behavior mandatory?
- Is one-step required or is two-step acceptable?
- Where must timestamp taps be visible (ingress/egress)?
- Must queue-induced variable delay be observable?
- What is the allowable timestamp error budget (X)?
- Is config hash + rollback required?
- Which black-box fields are mandatory for forensics?
- Is secure boot + key storage required?
- Which remote management channel is planned (local/remote)?
Deliverable · Selection scorecard (template)
| Spec item | Why it matters | How to verify | Threshold (X) |
|---|---|---|---|
| Ports & uplink speed | Defines bottleneck and worst-case contention. | Sustained load + utilization window checks. | Uplink ≥ X |
| Queue model & buffer headroom | Determines burst survival and drop behavior. | Watermark + drop reason under bursts. | Drops ≤ X |
| QoS / class protection | Prevents control/sync starvation. | Uplink pressure test + class latency markers. | p99 jitter ≤ X |
| PTP-aware forwarding hooks | Preserves time semantics through the bridge. | Timestamp tap visibility + delay markers. | Error ≤ X |
| Observability & black-box | Makes failures diagnosable in the field. | Counter schema + event ID + trigger capture validation. | Retention ≥ X |
Reference IC shortlist (by design style)
Request a Quote
H2-13. FAQs (Gateway / Bridge Controller)
These FAQs are designed to close long-tail field issues without expanding scope. Each answer is always four lines: Likely cause → Quick check → Fix → Pass criteria (with X/Y placeholders and explicit measurement windows).