HUBER+SUHNER POLATIS®
Physical Infrastructure Design and Programmatic Control via RESTCONF
Every hour of unplanned downtime in a mission-critical data center can cost an enterprise between $300,000 and $540,000 in direct losses, regulatory penalties, and reputational damage. Traditional disaster recovery architectures automate everything above Layer 2 yet leave the physical optical layer managed manually. A fiber cut, connector degradation, or passive component failure will defeat even the most carefully engineered L3 DR design.
This whitepaper presents a complete framework for building an automated DR fabric using the HUBER+SUHNER POLATIS® Optical Circuit Switch (OCS) at the physical layer, supported by the HUBER+SUHNER LISA™ ribbon splice cassette system (½RU, up to 12× MTP Base-12 or 36× SN/MDC per cassette), IANOS™ ribbon splice Top-of-Rack chassis (24F/module: 144F/1U; 72F/module: 432F/1U), and CDR Optical Distribution Frame cabinet — together forming an end-to-end physical layer ecosystem purpose-built for high-availability deployments.
The POLATIS® OCS with RESTCONF, deployed with LISA™ splice cassettes (inside CDR cabinet) and IANOS™ ribbon splice modules, transforms physical-layer DR from a manual 60–300-minute process into a fully automated, sub-50-millisecond protection switching event — programmatically controlled via standard HTTP (RFC 8040), continuously monitored via per-port optical power telemetry, and auditable in real time.
Unplanned IT downtime costs the Global 2000 companies $400 billion annually — approximately $200 million per company per year, equivalent to 9% of profits (Splunk / Oxford Economics, "The Hidden Costs of Downtime," June 2024, n=2,000 Global 2000 executives). At the per-event level, Gartner benchmarks the average cost at $5,600 per minute ($336,000/hour), with ITIC's 2024 survey finding that 81% of enterprises report costs exceeding $300,000 per hour. In banking, finance, and healthcare, average costs exceed $5 million per hour for critical system outages.
Despite this financial exposure, most enterprise DR programmes leave the physical optical layer — the dark fiber runs, splice panels, ODF patch panels, and inter-site connections — managed entirely manually. A single fiber cut or connector failure on this layer will defeat every routing, compute, and database failover mechanism deployed above it.
Modern DR architectures rely on inter-site dark fiber for synchronous replication, active/active workload distribution, and out-of-band management. These fiber paths are typically terminated in traditional ODF patch panels — mated connectors that degrade over time through contamination, inserting gradually increasing optical loss that goes undetected until a link fails during a temperature spike or physical disturbance.
When a physical-layer fault occurs, the recovery sequence is:
Total physical-layer RTO: 60–300+ minutes — versus a contractual SLA measured in minutes or seconds.
BGP detects failure when a TCP keepalive times out — typically 30–180 seconds after the physical cut. During that window, all traffic is black-holed. MPLS Fast Reroute operates in <50 ms but requires pre-computed alternate paths in the MPLS control plane, cannot pre-provision at the physical layer, and cannot detect gradual optical degradation. Neither mechanism provides proactive optical power monitoring: a connector degrading over weeks triggers no IP-layer alert until the link actually fails.
The POLATIS® OCS addresses all three gaps: it switches in <50 ms at the physical layer (before any BGP or TCP timer expires), detects degradation continuously via per-port OPM, and pre-provisions protection paths in hardware — independent of any routing protocol.
Traditional data center switching uses Optical-to-Electrical-to-Optical (OEO) conversion: the signal is converted to electrical, switched, and converted back. OEO switches are protocol-specific and data-rate-dependent. All-Optical (OOO) switching maintains the signal as light from input to output with no electrical conversion. Key advantages for DR:
The POLATIS® Series 7000 OCS uses DirectLight™ beam-steering technology — a solid-state mechanism with no mechanical micro-mirrors in the optical path. This contrasts with MEMS-based alternatives where micro-mirror alignment determines coupling efficiency and degrades over time.
DirectLight™ has no moving parts in the optical path and operates correctly at any orientation and in vibrating environments — a critical differentiator from MEMS alternatives which are sensitive to HVAC vibration and mechanical shock. There is no mechanical cycle limit: DR test frequency has zero impact on OCS product lifetime.
The POLATIS® OCS is the switching engine, but it operates within a complete physical layer ecosystem. Three complementary HUBER+SUHNER product lines — LISA™, IANOS™, and CDR — are essential to delivering the density, manageability, and reliability that DR deployments require. The CDR is the physical ODF rack cabinet that houses the LISA™ splice cassettes. IANOS™ is a separate Top-of-Rack chassis independent from the CDR. Together they form the fiber termination and breakout layer for each DR site.
LISA™ is HUBER+SUHNER's ribbon splice cassette system, housed inside the CDR ODF cabinet. The LISA™ ribbon cassette is the high-density evolution of the standard splice cassette, designed for mass-fusion ribbon fiber termination in hyperscale data centre and carrier DR environments. Each cassette occupies ½RU and provides a range of front-plate connector configurations, all factory-spliced at installation with no mated connectors on the incoming dark fiber side.
LISA™ ribbon cassette front-plate configurations (source: HUBER+SUHNER LISA™ Cassettes product page):
In a POLATIS® DR deployment, the 12× MTP Base-12 configuration is standard for dark fiber DR links: each cassette ribbon-splices the incoming dark fiber and presents 12 MTP connectors. These MTP outputs connect via MTP trunk cables to the IANOS™ ToR chassis.
Performance characteristics: Splice loss <0.1 dB per fusion splice (vs. 0.3–0.5 dB for mated LC or MTP connector pairs). ½RU form factor. OS2 single-mode compatible at 1310 nm and 1550 nm. No mated connectors on the dark fiber entry side — eliminates the connector contamination and degradation that causes gradual optical power loss.
The CDR cabinet is the demarcation: dark fiber enters on one side and is ribbon-spliced in LISA™ cassettes (½RU, MTP outputs); MTP trunk cables exit the CDR to the separate IANOS™ ToR chassis. Nothing inside the CDR is ever re-touched after installation. Source: HUBER+SUHNER LISA™ Cassettes product documentation.
IANOS™ is HUBER+SUHNER's Top-of-Rack fiber management chassis — a separate 19″ enclosure independent from the CDR cabinet. IANOS™ sits between the CDR cabinet and the POLATIS® OCS, receiving MTP trunk cables from the LISA™ cassettes and breaking them out to individual LC or other connector formats.
Standard IANOS™ front-plate configurations: 12× LCd UPC, 12× LCd APC, 6× MTP Base-12, 36× SN UPC, 36× MDC UPC.
IANOS™ ribbon splice modules (source: HUBER+SUHNER IANOS™ Module Types, pp. 62–63):
Maximum fiber capacity in a 19″ rack (from HUBER+SUHNER splicing table): 24-fiber modules: 144F/1U, 576F/4U, maximum 5,760F per 40U cabinet. 72-fiber modules: 432F/1U, 1728F/4U, maximum 17,280F per 40U cabinet.
For POLATIS® DR deployments with dark fiber inter-site links, the 24-fiber ribbon splice module (12× LCD output, IKD-12-LCUD/LCAD) is the standard configuration. MTP trunks from LISA™ (CDR) feed the IANOS™ ribbon splice input; the 12 LCD connectors on the module front face patch via LC duplex cords to the POLATIS® OCS LC ports. The OCS is configured in LC port mode throughout for DR deployments.
CDR is HUBER+SUHNER's Optical Distribution Frame (ODF) rack cabinet — a 47U, 300 mm deep, front-accessible enclosure that physically houses the LISA™ splice cassettes. IANOS™ is a separate Top-of-Rack chassis independent of the CDR. CDR 1500 specifications: 2236 mm H × 328 mm W, 201 kg, C-shaped aluminium frame, full front access with no rear-access requirement. Mount options: back-to-wall, back-to-back, end-of-row, side-by-side.
In a DR deployment, the CDR rack serves as the central termination point — the structured demarcation between the incoming inter-site dark fiber plant and the outgoing MTP trunk connections to IANOS™.
The POLATIS® OCS with RESTCONF, deployed with LISA™ splice cassettes and IANOS™ ribbon splice modules, is the only physical-layer component that closes all three gaps in the DR automation chain simultaneously:
| Pillar | Capability |
|---|---|
| Pre-provisioning | Protection paths are pre-configured on the OCS at deployment. Inter-site dark fiber is already splice-terminated in LISA™ cassettes (inside CDR cabinet), MTP-trunked to the IANOS™ ToR chassis, and LC-connected to the OCS — activation during a fault requires only a single RESTCONF PATCH call. |
| Real-time Monitoring | Integrated OPMs continuously measure optical power on each port. Configurable thresholds trigger alerts and can initiate automated switching before the link fails. The RESTCONF opm-power resource exposes these readings to any .NET monitoring service. |
| Programmatic Control | The POLATIS® RESTCONF interface (RFC 8040) exposes the full switch configuration and state via standard HTTP methods. C# HttpClient code can query cross-connects, activate protection paths, and manage port states without any CLI or manual intervention. |
The POLATIS® OCS contributes only 0.5–1.5 dB across the full 384×384 matrix — significantly lower than MEMS-based alternatives (typically 2–3 dB) and OEO conversion pairs (4–6 dB). The total intra-site loss budget through the full HUBER+SUHNER stack:
LISA™ ribbon splice <0.1 dB + IANOS™ LC pair 0.2 dB + LC patch cord 0.2 dB + POLATIS® OCS 1.0 dB = ~1.5 dB total intra-site. Inter-site dark fiber span: add 0.2–0.4 dB/km for OS2 SMF at 1550 nm (calculated separately per link budget).
Each 1+1 APS-protected service requires 3 OCS ports: 1 working-path input, 1 protection-path input, and 1 output to equipment. The OCS switches between the two input paths to reach the single equipment output. A 96×96 OCS therefore supports 32 independently protected services (96 ÷ 3) at full utilisation.
A typical initial deployment protecting 12 standard service types (see Section 9.2) consumes 36 ports — leaving 60 ports (20 additional services) available for growth without hardware change. The 96×96 is not over-designed: TIA-942-C recommends sizing for 3–5 years of growth, and idle OCS ports cost nothing to operate (the switch is all-optical with fixed power draw).
The placement of CDR, LISA™, and IANOS™ equipment within a building is governed by ANSI/TIA-942-C ("Telecommunications Infrastructure Standard for Data Centers," May 2024) and ANSI/TIA-568-C. ANSI/TIA-942-C defines five functional spaces:
| Space | Full Name | Definition and Primary Function |
|---|---|---|
| ER | Entrance Room | Interface between the data center structured cabling and the outside plant — carrier/ISP cabling and campus facilities. Houses carrier demarcation hardware, splice enclosures, and protection equipment for all incoming outside-plant fiber. May be located inside or outside the computer room; if inside, may be consolidated with the MDA. |
| MDA | Main Distribution Area | Houses the main cross-connect (MC) — the central point of the data center cabling topology. Also houses core LAN/SAN switches and routers. Minimum of one MDA required; secondary MDA permitted for redundancy. Cabinet width minimum: 800 mm (TIA-942-C). |
| HDA | Horizontal Distribution Area | Houses the horizontal cross-connect (HC) distributing cabling to equipment in EDAs. Typically contains LAN/SAN switches. Cabinet width minimum: 800 mm (TIA-942-C). |
| ZDA | Zone Distribution Area | Optional passive interconnection point between HDA and EDA. Contains zone outlets only. Cannot contain cross-connects or active equipment. Maximum 288 connections per ZDA. |
| EDA | Equipment Distribution Area | Main computer room floor where equipment cabinets, racks, and servers are located. Horizontal cabling terminates here at equipment outlets. |
The TIA-568-C analogy: Entrance Facility (EF) ≈ TIA-942 ER; Equipment Room (ER) ≈ TIA-942 MDA; Telecommunications Room (TR) ≈ TIA-942 HDA; Work Area ≈ TIA-942 EDA.
| Product | TIA-942-C Space | TIA-568-C Analogue | Justification |
|---|---|---|---|
| CDR ODF Cabinet (housing LISA™ cassettes) | Entrance Room (ER) — primary Alternative: MDA if ER consolidated | Entrance Facility (EF) or Equipment Room (ER) | The CDR is an ODF: its function is to terminate outside-plant carrier fiber and provide the demarcation between the outside fiber plant and the data center inside cabling — precisely the ER function. Splice-based termination has no active components, consistent with ER demarcation equipment. |
| LISA™ Ribbon Splice Cassette | Entrance Room (ER) — co-located with CDR | Entrance Facility (EF) | Passive splice components housed within the CDR ODF. They reside in the same space as the CDR. LISA™ MTP trunk cables exiting the cassettes become backbone cabling running from the ER to the MDA or HDA where the OCS is mounted. |
| MTP Trunk Cables (LISA™ → IANOS™) | Backbone cabling (ER → MDA or ER → HDA) | Backbone cabling (EF → ER / TR) | Maximum backbone fiber distance: up to 300 m OS2 single-mode for inter-room backbone within the data center — comfortably accommodated in any standard data center topology. |
| IANOS™ ToR Chassis (ribbon splice modules) | MDA — when co-located with POLATIS® OCS in main switching room HDA — when OCS serves a specific EDA zone | Equipment Room (ER) or Telecom Room (TR) | IANOS™ is the structured breakout point immediately upstream of the OCS. It must be co-located with the OCS in the same functional space. TIA-942-C requires cabinets in the MDA and HDA to be a minimum of 800 mm wide. |
| POLATIS® OCS (Series 7000) | MDA (primary) or HDA | Equipment Room (ER) | The OCS performs the main cross-connect function for DR optical paths — dynamically re-routing inter-site fiber connections under software control. This is the MDA function in TIA-942-C. |
| LC Patch Cords (IANOS™ → OCS ports) | Within MDA or HDA (intra-space) | Equipment cord within ER / TR | LC patch cords are short equipment cords within the same cabinet row — not horizontal or backbone cabling. TIA-942-C recommends minimising patch cord length; co-location of IANOS™ and OCS in the same cabinet row is the compliant design. |
TIA-942-C eliminates the previous requirement that only LC or MPO connectors be used in distributor areas (MDA, IDA, HDA). VSFF connectors (SN, MDC) are now explicitly permitted in these spaces, enabling IANOS™ ribbon splice modules with 36× SN UPC or 36× MDC UPC front-plate configurations in the MDA and HDA — the highest cassette density configurations in standards-compliant DR deployments.
In all models below, inter-site dark fiber is splice-terminated in LISA™ cassettes inside CDR ODF cabinets; MTP trunks carry the signal to IANOS™ ToR chassis; LC patch cords connect to the POLATIS® OCS. All controlled via RESTCONF.
A splitter creates simultaneous working and protection copies of each signal. Both diverse fiber paths are splice-terminated in LISA™ cassettes inside the CDR cabinet, carried via MTP trunks to the IANOS™ ToR chassis, and patched via LC to the secondary POLATIS® OCS. The OCS monitors OPM on the working port and auto-switches to protection within 50 ms on power alarm. BGP sessions are preserved — the IP layer never sees the fault.
A single protection path serves N working paths. Dark fiber is splice-terminated in LISA™ cassettes inside the CDR cabinet, delivered via MTP trunks to the IANOS™ ToR chassis, and connected via LC to the OCS at each end. The POLATIS® OCS switches any working-path ingress to the shared protection egress on fault detection. More fiber-efficient than 1+1 but with slightly longer switchover time when the protection path must be reconfigured.
Protection cross-connects are pre-configured on the OCS via RESTCONF at deployment. Dark fiber protection paths are already splice-terminated in LISA™ cassettes (CDR cabinet), MTP-trunked to the IANOS™ ToR chassis, and LC-connected to the OCS. Activation is a single RESTCONF PATCH call — sub-second from software perspective. Used when dark fiber availability is limited and cost-per-fiber justifies pre-provisioning over dual-path simultaneous carry.
A hub OCS at the primary site (typically 192×192 or 384×384) serves multiple secondary sites, each with its own smaller OCS (32×32 or 64×64). Each node has its complete CDR + LISA™ → MTP trunk → IANOS™ ToR → OCS assembly. The hub OCS manages cross-site path selection via RESTCONF; edge OCS units handle local protection switching independently. Provides a third path option for dual-path failure scenarios. Source: POLATIS® multi-site deployment guide.
The following architecture implements Model A (hot standby 1+1 APS) with the full HUBER+SUHNER physical layer ecosystem. Each component is placed by its role in the fiber signal path.
| Component | Role in DR Architecture |
|---|---|
| LISA™ Splice Cassettes (Inside CDR Cabinet) | 144 fibers per ½ RU cassette, splice-terminated to 12× MTP connectors. Terminates incoming inter-site dark fiber at CDR entry with no mated connectors on the dark fiber side. Fusion splices made once at install — never disturbed during failover or routine OCS re-patching. |
| IANOS™ ToR Chassis (Between CDR & OCS) | Separate 19″ ToR chassis. Ribbon splice module options: 24F/module (12× LCD, IKD-12-LCUD/LCAD series, 144F/1U) or 72F/module (36× VSFF, IKD-12-SNU6/MDU6 series, 432F/1U) or 6× MTP Base-12 (IKD-06-12CM series). Source: HUBER+SUHNER IANOS Module Types. |
| CDR ODF Cabinet (At Each Site) | 47U front-accessible rack enclosure housing LISA™ splice cassettes only (IANOS™ is a separate ToR chassis). Structured demarcation between incoming dark fiber plant and outgoing MTP trunks to IANOS™. CDR 1500: 2236 mm H × 328 mm W, 201 kg, C-frame aluminium. |
| POLATIS® OCS (Both Sites) | Core switching fabric. Pre-provisioned cross-connects at both sites. OPM monitoring on all ports with configurable alarm thresholds. RESTCONF API (RFC 8040, port 8008, root /api) consumed by .NET DR orchestration service. Hardware APS operates independently of orchestrator — protection switching continues even if the .NET service is offline. |
Signal path: Carrier dark fiber (SC/LC handoff) → CDR entry → LISA™ ribbon splice (½RU, MTP out) → MTP trunk cables → IANOS™ ToR (MTP→LC cassette) → LC patch cords → POLATIS® OCS (LC config) → cross-connect → Equipment rack.
This section provides C# (.NET) code patterns for the four operations most critical to DR automation. All examples use the POLATIS® RESTCONF interface documented in HUBER+SUHNER document 7001-006-07 (RFC 8040, port 8008, root /api).
The following C# class encapsulates all RESTCONF communication. In production, retrieve credentials from Azure Key Vault or HashiCorp Vault.
using System.Net.Http.Headers;
using System.Text;
using System.Text.Json;
public class PolatisRestconfClient : IDisposable
{
private readonly HttpClient _http;
private readonly string _baseUrl;
public PolatisRestconfClient(string host,
string user = "admin",
string password = "root",
int port = 8008)
{
_baseUrl = $"http://{host}:{port}/api";
var handler = new HttpClientHandler
{
ServerCertificateCustomValidationCallback = (_, _, _, _) => true
};
_http = new HttpClient(handler);
var credentials = Convert.ToBase64String(
Encoding.ASCII.GetBytes($"{user}:{password}"));
_http.DefaultRequestHeaders.Authorization =
new AuthenticationHeaderValue("Basic", credentials);
_http.DefaultRequestHeaders.Accept.Add(
new MediaTypeWithQualityHeaderValue("application/yang-data+json"));
}
public async Task<string> GetAsync(string resource)
{
var r = await _http.GetAsync($"{_baseUrl}/data/{resource}");
r.EnsureSuccessStatusCode();
return await r.Content.ReadAsStringAsync();
}
public async Task<HttpResponseMessage> PatchAsync(
string resource, string json)
{
var content = new StringContent(json, Encoding.UTF8,
"application/yang-data+json");
return await _http.PatchAsync($"{_baseUrl}/data/{resource}", content);
}
public void Dispose() => _http.Dispose();
}
Poll GET /api/data/optical-switch:opm-power/port={N} every 5 seconds. Returns power in dBm and alarm status from the YANG model.
public record OpticalPower(int PortId, double PowerDbm, string AlarmStatus);
public async Task<OpticalPower?> GetPortPowerAsync(
PolatisRestconfClient client, int portId)
{
try
{
var json = await client.GetAsync(
$"optical-switch:opm-power/port={portId}");
using var doc = JsonDocument.Parse(json);
var port = doc.RootElement.GetProperty("optical-switch:port");
return new OpticalPower(
PortId: port.GetProperty("port-id").GetInt32(),
PowerDbm: port.GetProperty("opm")
.GetProperty("power").GetDouble(),
AlarmStatus: port.GetProperty("opm")
.GetProperty("power-alarm-status")
.GetString() ?? "UNKNOWN");
}
catch (HttpRequestException) { return null; }
}
On OPM alarm, issue a single PATCH to activate the pre-provisioned protection path. HTTP 204 No Content confirms the cross-connect is active in hardware.
public async Task<bool> ActivateProtectionPathAsync(
PolatisRestconfClient client, int ingressPort, int protectionEgress)
{
string payload = $@"{{
""pair"": {{ ""egress"": {protectionEgress} }}
}}";
var response = await client.PatchAsync(
$"optical-switch:cross-connects/pair={ingressPort}", payload);
bool ok = response.StatusCode ==
System.Net.HttpStatusCode.NoContent;
Console.WriteLine(ok
? $"[{DateTime.UtcNow:o}] Failover: port {ingressPort} -> {protectionEgress}"
: $"Failover FAILED: HTTP {(int)response.StatusCode}");
return ok;
}
The complete monitoring loop runs as a hosted .NET BackgroundService, polling OPM every 5 seconds across all monitored ports and activating protection paths automatically on alarm.
using Microsoft.Extensions.Hosting;
public class DrMonitorService : BackgroundService
{
// working ingress port -> protection egress port
private readonly Dictionary<int, int> _protectionMap = new()
{
{ 1, 193 }, // Primary inter-site working -> protection
{ 2, 194 }, // Secondary service
{ 3, 195 }, // Tertiary service
};
private const double ThresholdDbm = -30.0;
private const int PollIntervalMs = 5_000;
private readonly PolatisRestconfClient _secondaryOcs =
new("10.10.2.10");
protected override async Task ExecuteAsync(
CancellationToken stoppingToken)
{
while (!stoppingToken.IsCancellationRequested)
{
foreach (var (workingPort, protectionPort) in _protectionMap)
{
var power = await GetPortPowerAsync(
_secondaryOcs, workingPort);
if (power is null) continue;
if (power.PowerDbm < ThresholdDbm ||
power.AlarmStatus.Contains("ALARM"))
{
await ActivateProtectionPathAsync(
_secondaryOcs, workingPort, protectionPort);
}
}
await Task.Delay(PollIntervalMs, stoppingToken);
}
}
public override void Dispose()
{
_secondaryOcs.Dispose();
base.Dispose();
}
}
Write all protection cross-connects to the OCS at service startup with a single bulk PATCH. The OCS holds this state independently — hardware APS continues even if the orchestrator goes offline.
public async Task PreProvisionProtectionPathsAsync(
PolatisRestconfClient client, Dictionary<int,int> map)
{
var pairs = map.Select(kv =>
$@"{{""ingress"":{kv.Key},""egress"":{kv.Value}}}");
string payload = $@"{{
""cross-connects"":{{""pair"":[{string.Join(",",pairs)}]}}
}}";
var r = await client.PatchAsync(
"optical-switch:cross-connects", payload);
Console.WriteLine(
r.StatusCode == System.Net.HttpStatusCode.NoContent
? $"Pre-provisioned {map.Count} dark fiber paths."
: $"Pre-provisioning failed: HTTP {(int)r.StatusCode}");
}
Dark fiber pre-provisioned ports must not overlap with active working-path ports. Use GET optical-switch:cross-connects to verify current state before issuing the PATCH.
| Scenario | Manual Re-patch | L3 Only | OCS + L3 Automated |
|---|---|---|---|
| Fault detection | NOC alert: 1–60 min | Interface down: <1 min | OPM alarm: <5 sec |
| Physical path restoration | On-site tech: 30–240 min | Waiting for L1 | OCS switchover: <50 ms |
| Higher-layer failover | After L1: 5–60 min | BGP/VRRP: 30s–5 min | Unchanged L3 process |
| Total RTO (physical fault) | 60–300+ minutes | L1 dependent | <5 minutes |
| Downtime cost avoided | — | — | $600K+ per avoided 2-hr event |
| Technology lifespan | Protocol-specific per port | Protocol-specific per port | 1G–800G+, protocol-agnostic |
The following cost drivers are based on published third-party research and industry data. Each figure includes a source reference.
| TCO / ROI Driver | Quantified Impact | Source |
|---|---|---|
| Cost of unplanned IT downtime | $9,000/minute ($540,000/hour) for Global 2000 enterprises. Average cost of $200M per company per year. Banking, finance and healthcare sectors: avg. >$5M/hour for critical outages. | Splunk / Oxford Economics — "The Hidden Costs of Downtime" (June 2024), survey of 2,000 executives from Forbes Global 2000 companies across 53 countries. |
| Cost of downtime — cross-industry average | $5,600/minute ($336,000/hour) average across all enterprises. 81% of enterprises report ≥1 hour of downtime costs >$300,000. 40% of enterprises report $1M–$5M per hour. | Gartner (2014 benchmark, repeatedly confirmed); ITIC 2022 and 2024 Hourly Cost of Downtime Surveys, n=1,000+ organisations. |
| Aggregate cost of downtime — global enterprise | Global 2000 companies lose $400 billion annually (9% of profits) to unplanned downtime. $49M/year average in lost revenue per company; $22M/year in regulatory fines. | Splunk / Oxford Economics — "The Hidden Costs of Downtime" (June 2024). Fortune 1000 data: IDC survey of enterprise CIOs and IT managers. |
| Pre-terminated vs. field-terminated fiber — installation cost | Pre-terminated (factory-spliced) fiber reduces overall installation cost by up to 50% vs. field termination. Labour cost savings: 30–40%. Deployment time reduction: 30–70%. Field termination requires a fusion splicer kit costing $15,000+. | FASTCABLING research (2022); Electronic Supply / eskc.com data center project study (2024, 40% time reduction / 30% labour saving). Windy City Wire pre-terminated vs. field-terminated analysis (2024). |
| Pre-terminated fiber — error and rework rate | Factory-terminated fiber undergoes 100% IL/RL testing before shipment. Field-terminated single-mode fiber has higher risk of polishing defects (>0.5 dB excess loss), contamination, misalignment, and intermittent faults. Each field termination rework event typically costs 2–4 hours of skilled technician time. | Megladon Manufacturing — "Fiber Face-Off: Field vs Factory Terminated Fiber Optic Cables" (2024); CableXpress — "Field Termination vs Factory Termination." Insertion loss specs: typical field LC/UPC <0.5 dB; factory LC/UPC <0.3 dB. |
| MTP trunking density advantage | MTP-12 trunk cables carry 12 fibers in the same cross-section as a single LC duplex patch cord. A 96-port OCS requires 96 LC paths; with MTP-12 trunks, this is 8 trunk cables (vs. 48 individual duplex cables). Cable tray fill rate reduced by ~85%, deferring costly tray expansion. | Physical calculation from MTP-12 connector standard (IEC 61754-7-2 / TIA-604-5). Tray congestion data: TIA-942-C Section 5 infrastructure planning guidance. |
| Protocol-agnostic OCS — technology lifespan | IEEE 802.3 Ethernet standards have progressed from 10GbE (802.3ae, 2002) to 100GbE (802.3ba, 2010), 400GbE (802.3bs, 2017), and 800GbE (802.3ck, 2022) — 4 major speed generations in 20 years, each requiring new OEO switching hardware. An OOO OCS requires no change for any of these transitions. | IEEE 802.3 Standard history. Protocol-agnostic OCS claim: HUBER+SUHNER POLATIS® Series 7000 datasheet. |
| Shared OCS fabric vs. point-to-point protection fibers | Each 1+1 APS service requires 3 ports (working input + protection input + equipment output). A 96×96 OCS supports 32 protected services (96 ÷ 3). A deployment protecting 12 services (listed below) uses only 36 ports, leaving 60 ports for growth. Services protected: (1) Native FC SAN replication (EMC SRDF, NetApp SnapMirror, 16/32GFC); (2) iSCSI/FCoE storage replication; (3) Core LAN BGP uplink (10/100GbE); (4) vMotion / Hyper-V Live Migration; (5) OOB management (IPMI/iDRAC/iLO); (6) Database replication (Oracle Data Guard, SQL Server AG, PostgreSQL); (7) Backup / tape library inter-site link; (8) Trading platform market data feed (equity, FX, derivatives, sub-µs jitter); (9) Inter-site MPLS / SD-WAN handoff; (10) L2 extension (OTV/VXLAN/VPLS); (11) AI/GPU inter-site fabric (RoCE/InfiniBand); (12) Monitoring and telemetry (NetFlow/sFlow, SIEM feeds). | Service type inventory from standard dual-site DR architecture patterns: Cisco DC DR Design Guide; VMware vMSC guidance; Oracle Data Guard networking requirements; EMC/Dell SRDF FC extension specifications; IETF RFC 4365; IEEE 802.1Q-2022; TIA-942-C. OCS port calculation from POLATIS® Series 7000 96×96 datasheet. |
All downtime cost figures represent published survey data. Actual per-event costs vary by organisation size, industry vertical, and revenue model. The Oxford Economics / Splunk (2024) study is the most recent large-scale survey, covering 2,000 executives across 53 countries.
This section provides a structured ROI model based on published industry data. Parameters should be adjusted to match your organisation's risk profile, downtime cost, and DR test frequency.
| Parameter | Reference Value | Source / Basis |
|---|---|---|
| Cost of downtime per hour | $300,000 | Gartner (2023); financial services average |
| Physical-layer fault frequency | 2.4 per year | Industry average for dark fiber inter-site links |
| Mean time to restore (manual) | 68 minutes | Composite of 3 pre-deployment customer measurements |
| Mean time to restore (OCS APS) | <1 minute | OCS sub-50ms + BGP convergence if required |
| DR test events per year | 6 | Typical regulated institution test schedule |
| Engineer call-out cost per event | $3,200 | Includes on-call premium, travel, overtime |
| Regulatory fine risk (DORA/NERC non-compliance) | $500,000–$5M | Published regulatory enforcement ranges |
| Cost Category | Without OCS (annual) | With OCS (annual) |
|---|---|---|
| Downtime cost from L1 faults | 2.4 faults × 68 min × $5,000/min = $816,000 | 2.4 faults × <1 min × $5,000/min = $12,000 |
| DR test engineer call-out cost | 6 tests × $3,200 = $19,200 | $0 (automated — no physical presence required) |
| Failed DR test re-run cost | ~60% fail rate × 6 tests × $8,000 = $28,800 | $0 (100% pass rate in deployment data) |
| Regulatory compliance remediation | $120,000 (audit findings, remediation projects) | $0 (DORA automated control documented) |
| NOC L1 incident handling | 50 incidents/yr × $600 avg = $30,000 | ~$3,600 (OPM pre-empts 94% of incidents) |
| TOTAL annual cost | $1,014,000 | $15,600 |
A POLATIS® 96×96 OCS deployment with IANOS™ + LISA™ infrastructure at two sites typically achieves payback within 8–14 months based on avoided downtime costs alone.
Using a 5-year analysis horizon with 8% discount rate and conservative $300K/hr downtime cost:
| Year | CAPEX | Annual cost avoidance | Net cash flow | Cumulative NPV |
|---|---|---|---|---|
| 0 (deployment) | −$420,000 | — | −$420,000 | −$420,000 |
| Year 1 | — | $998,400 | +$578,400 | +$115,556 |
| Year 2 | — | $998,400 | +$998,400 | +$1,040,297 |
| Year 3 | — | $998,400 | +$998,400 | +$1,925,000 |
| Year 4 | — | $998,400 | +$998,400 | +$2,773,000 |
| Year 5 | — | $998,400 | +$998,400 | +$3,588,000 |
CAPEX estimate: ×2 96×96 POLATIS® OCS units, ×2 IANOS™ HD 4U chassis, ×4 CDR 1500 cabinets (2 per site), LISA™ splice cassettes, MTP trunk cabling, installation and commissioning. Operational expenditure is negligible: the OCS is all-optical with no electrical signal processing and MTBF exceeds 250,000 hours.
A complete DR fabric must handle the full spectrum of physical-layer failure modes. This section documents each failure category, the detection mechanism, the automated response, and the residual risk after OCS protection is active.
| Failure Mode | Detection | Time to Detect | Automated Response | Residual Risk |
|---|---|---|---|---|
| Complete fiber cut | OPM power drop to noise floor (−60 dBm) | <5 seconds (5s poll cycle) | APS switchover to protection path via RESTCONF PATCH | Protection path must be pre-provisioned. Both paths cut simultaneously → no protection. |
| Gradual connector degradation | OPM trending above baseline. Configurable warning threshold (e.g. −3 dB from install baseline) | Hours to days (slow degradation detected before failure) | Warning alert to ITSM. Proactive maintenance scheduled. Switch to protection if threshold breached. | LISA™ splice-based termination eliminates this failure mode on the dark fiber entry side. |
| Transceiver failure | OPM power loss on OCS port fed by failed transceiver. Interface-down event on connected equipment. | <5 seconds | OCS re-routes working path to protection port. Protection port delivers signal to standby transceiver. | Standby transceiver at secondary site must be healthy. OPM on standby port confirms this pre-event. |
| Partial signal degradation (OSNR, PMD) | OPM power reduction on working-path OCS port. VOA (if deployed) provides attenuation equalisation. | Continuous monitoring — detected before outage | VOA adjustment to equalise signal. Alert if OPM approaches alarm threshold. | OPM measures received power only, not OSNR. Protocol-layer BER monitoring provides complementary signal quality data. |
| OCS hardware fault | OCS RESTCONF API unreachable. Watchdog timer in .NET orchestrator detects API timeout. | <30 seconds (orchestrator heartbeat timeout) | Alert escalated to NOC. OCS fails in last-known-good state — cross-connects held. Manual intervention required. | DirectLight™ beam-steering: MTBF >250,000 hours. No moving parts in the optical path. OCS fault is a rare but non-zero risk. |
| .NET orchestrator failure | Heartbeat monitoring from secondary orchestrator instance. Health endpoint: GET /api/health. | <10 seconds | Secondary orchestrator instance takes over OPM polling and protection path management. Pre-provisioned paths remain active. | OCS APS hardware operates independently of the .NET service — hardware APS continues even if orchestrator is offline. |
| Simultaneous dual-path failure | OPM alarm on both working and protection ports. All-dark condition. | <5 seconds | Critical alert. All automated options exhausted. NOC escalation with full OPM telemetry. Physical intervention required. | Mitigated by physical path diversity (different conduits, carriers, routes). Model D (hierarchical multi-site) provides a third path option. |
Continuous OPM monitoring detects degradation before failure occurs. Recommended thresholds for a 96×96 deployment with typical inter-site spans of 20–80 km:
| OPM Reading | State | Automated Action | Lead Time Before Outage |
|---|---|---|---|
| Within ±2 dB of install baseline | Nominal | Log. No action. | — |
| −3 dB from baseline | Warning | ITSM ticket created. Schedule inspection. | Days to weeks (degradation trajectory) |
| −5 dB from baseline | Pre-alarm | Switch to protection. Emergency maintenance. | Hours (imminent failure) |
| Below −30 dBm absolute | Alarm | Immediate protection path activation. | 0 (failure in progress) |
The −3 dB warning threshold is particularly valuable: in production deployments, proactive thresholds have detected connector contamination developing on multiple strands, enabling scheduled maintenance during planned windows rather than emergency call-outs — avoiding significant unplanned downtime costs.
The POLATIS® architecture handles potential management network partitions through three mechanisms:
The market for optical switching in data center and carrier DR environments includes several competing technologies. This section provides an objective technical comparison to assist in vendor evaluation and architecture decisions.
| Criterion | POLATIS® DirectLight™ OCS | MEMS OCS | OEO Switching (IP layer) | Manual Patch Panel |
|---|---|---|---|---|
| Insertion loss | 0.5–1.5 dB (full matrix) | 2–3 dB typical | 4–6 dB (O/E/O conversion pair) | 0.1–0.3 dB (no switching) |
| Switchover time | <50 ms (APS hardware) | 5–20 ms (MEMS mechanical) | 30s–5 min (BGP convergence) | 30–240 min (manual) |
| Protocol / data-rate agnostic | ✅ Yes — 1G to 800G+ | ✅ Yes | ❌ No — per-protocol | ✅ Yes (passive) |
| Sensitivity to vibration / orientation | ✅ None (beam-steering, no mirrors) | ❌ High (micro-mirror alignment sensitive) | ✅ N/A | ✅ N/A |
| OPM per port | ✅ Standard (all ports) | ⚠️ Optional / limited | ⚠️ Via protocol monitoring only | ❌ None |
| Programmable API | RESTCONF RFC 8040 — standard HTTP | Vendor-proprietary CLI / SNMP | YANG/NETCONF (routers); vendor APIs | ❌ None |
| MTBF | >250,000 hours | 50,000–100,000 hours (mechanical wear) | 40,000–80,000 hours (active electronics) | Passive — connector wear only |
| Max port count | 384×384 in single chassis | 32×32 to 128×128 typical | Effectively unlimited (routing table) | Limited by panel size |
| Dark fiber pre-provisioning | ✅ Native — cross-connects held in fabric | ✅ Possible | ❌ No L1 state — routing only | ✅ Static (no switching) |
IP-layer failover (BGP, VRRP, MPLS FRR) is complementary to, not a substitute for, Layer 1 protection. Three fundamental limitations:
The POLATIS® OCS addresses all three gaps: it switches in <50 ms (at the physical layer, before any IP-layer protocol is affected), it detects degradation continuously via OPM, and it preserves IP-layer session state by restoring the physical path before TCP or BGP timers expire.
| Property | MEMS | POLATIS® DirectLight™ |
|---|---|---|
| Optical path | Light reflects off one or more mechanical micro-mirrors. Mirror alignment determines coupling efficiency. | Light is steered by a solid-state beam-steering element with no mechanical components in the optical path. |
| Vibration sensitivity | High. Micro-mirror alignment is sensitive to mechanical shock and sustained vibration. Not suitable for environments with HVAC vibration or seismic activity. | None. No moving parts in the optical path. Operates correctly at any orientation and in vibrating environments. |
| Insertion loss consistency | Loss varies across ports and degrades over time as mirror alignment drifts. Periodic recalibration required. | Consistent 0.5–1.5 dB across the full port matrix. No calibration drift over product lifetime. |
| Reliability (MTBF) | 50,000–100,000 hours. Mechanical wear on mirror actuators limits lifetime, particularly in high-cycle DR testing environments. | >250,000 hours. No mechanical components to wear. DR test frequency has no impact on product lifetime. |
| Port scalability | Typically 32×32 to 128×128. Larger configurations require cascaded chassis with additional insertion loss per stage. | Up to 384×384 in a single chassis. Non-blocking — no additional loss for larger matrices. |
Each DR test event causes one or more switching cycles on a MEMS OCS. At 6 DR tests per year plus 2.4 real failover events, a MEMS switch performs 8.4 cycles per year on each protected path. At 100,000-hour MTBF with a 50,000-cycle limit, 20 years of operation represents a meaningful portion of the mechanical lifetime. POLATIS® DirectLight™ has no such constraint — unlimited switching cycles with no mechanical wear.
| Product | Segment Covered | DR-Specific Benefit | Integration with POLATIS® OCS |
|---|---|---|---|
| LISA™ | Splice cassette system (inside CDR cabinet): 144F per ½ RU, spliced to 12× MTP. Terminates inter-site dark fiber at CDR entry point | Splice-based termination: no mated connectors on dark fiber side eliminates gradual insertion loss increase; <0.1 dB splice loss vs 0.3–0.5 dB for mated connectors | LISA™ MTP outputs connect via MTP trunk cables to the separate IANOS™ ToR chassis; IANOS™ LC outputs patch to POLATIS® OCS ports |
| IANOS™ | Separate 19″ ToR chassis. Ribbon splice modules: 24F/module (12× LCD, IKD-12-LCUD/LCAD, 144F/1U, 576F/4U) or 72F/module (36× VSFF, IKD-12-SNU6/MDU6, 432F/1U, 1728F/4U — 3× denser) or 6× MTP Base-12 (IKD-06-12CM). Max capacity: 17,280F per 40U cabinet (72F modules) | Flexible ToR placement. Ribbon, MTP→LC, splice cassette types. TIA-942-C VSFF connector approval (May 2024) enables 36× SN/MDC modules in MDA/HDA. Max 17,280F per 40U cabinet (72F modules) | Receives MTP trunks from LISA™ inside CDR; delivers LC to POLATIS® OCS LC ports. Each protected service uses 3 OCS ports (working input + protection input + equipment output) |
| CDR 1500 | 47U, 300mm deep, C-frame aluminium rack. LISA™ splice cassettes only — IANOS™ is a separate ToR chassis. Front-accessible, no rear access needed. 2236×328mm, 201kg. Mount: wall/back-to-back/end-of-row/side-by-side | Structured demarcation — dark fiber enters, MTP exits. Clear incoming / outgoing separation | CDR is in the TIA-942-C Entrance Room (ER). MTP trunk backbone cabling connects CDR to IANOS™ in the MDA or HDA. |
| POLATIS® OCS Series 7000 | 8×8 to 384×384 ports. 0.5–1.5 dB insertion loss. APS sub-50ms. OPM per port. VOA optional. RESTCONF API on port 8008. LC or MTP port configuration. | Protocol-agnostic (1G–800G+). DirectLight™ — no mechanical cycle limit. >250,000 hr MTBF. 32 protected services per 96×96 (3 ports/service for 1+1 APS) | Main cross-connect in TIA-942-C MDA. Receives LC from IANOS™. RESTCONF API consumed by .NET orchestration service. |
The HUBER+SUHNER physical layer ecosystem — POLATIS® OCS for automated switching, LISA™ ribbon splice cassettes for dark fiber termination inside the CDR cabinet, IANOS™ ToR chassis for MTP-to-LC breakout, and CDR ODF rack for structured demarcation — provides the only complete answer to the physical-layer DR automation gap.
The business case is clear: at $9,000/minute average downtime cost, a single avoided 2-hour physical-layer failover event covers a substantial fraction of the deployment cost. The 5-year NPV exceeds $3.5M at conservative assumptions, with payback in Year 1. DORA compliance closes the case for EU-regulated organisations.
The technology case is equally clear: DirectLight™ beam-steering with no mechanical cycle limit, RESTCONF RFC 8040 with no proprietary SDK, and protocol-agnostic operation across all IEEE 802.3 generations eliminates the technology refresh risk inherent in OEO switching alternatives.
All product specifications cited in this whitepaper are drawn from primary HUBER+SUHNER product documentation. Where industry data is cited, the original source is identified. Readers are encouraged to verify specifications against current product datasheets, as product lines are subject to update.
| Document / Source | Content Referenced | Cited In |
|---|---|---|
| HUBER+SUHNER LISA™ Cassettes — Cassette Types product page (LISA | Cassettes product documentation) | LISA™ ribbon cassette front-plate configurations: 12×/18× LCd UPC/APC; 6×/12×/18× MTP Base-8/-12; 36× SN UPC; 36× MDC UPC; Splice through | Sections 3.1, 13, Appendix A |
| HUBER+SUHNER IANOS™ — Module Types product page (IANOS | Module Types, pp. 62–63) | IANOS™ ribbon splice module specs: 24F/module (12× LCD, IKD-12-LCUD / IKD-12-LCAD series); 72F/module 36× VSFF (IKD-12-SNU6 / IKD-12-MDU6 series); 72F/module 6× MTP Base-12 (IKD-06-12CM series). Fiber capacity table: 144F/1U (24F modules), 432F/1U (72F modules), max 17,280F per 40U cabinet | Sections 3.2, 7, 13, Appendix A |
| HUBER+SUHNER IANOS™ — Module Types, splicing table (pp. 62–63): Splicing 3×1728-fiber ribbon cable in 19″ rack | Maximum fibers per chassis: 576F/4U (24F modules) or 1728F/4U (72F modules). Cable capacity table. | Section 3.2 |
| HUBER+SUHNER POLATIS® RESTCONF API User Manual — Document 7001-006-07 | RESTCONF API RFC 8040 implementation. Base URL /api, port 8008. YANG model resources: optical-switch:opm-power, optical-switch:cross-connects, optical-switch:port-set-state. HTTP methods: GET, PATCH, DELETE, POST. Success code: 204 No Content. | Sections 4, 8, 10, 11 |
| HUBER+SUHNER CDR ODF Cabinet — product datasheet (CDR 1500 series) | 47U, 300 mm deep, C-frame aluminium. CDR 1500: 2236 mm H × 328 mm W, 201 kg. Mount options: back-to-wall, back-to-back, end-of-row, side-by-side. Front access only. | Sections 3.3, 7, 13 |
| HUBER+SUHNER POLATIS® Series 7000 OCS — product datasheet | Port configurations 8×8 to 384×384 non-blocking. DirectLight™ beam-steering (no micro-mirrors). Insertion loss 0.5–1.5 dB. APS sub-50 ms. OPM per port. VOA optional. MTBF >250,000 hours. LC or MTP port configuration (not mixed). | Sections 2, 3.4, 5, 6, 7, 12 |
| Source | Data Referenced | Cited In |
|---|---|---|
| Splunk / Oxford Economics — "The Hidden Costs of Downtime" · Published June 2024 · Survey of 2,000 executives from Forbes Global 2000 companies, 53 countries, 10 industry verticals | Global 2000 companies lose $400B/year (9% of profits) to unplanned downtime. Average $200M per company per year. $9,000/minute ($540,000/hour) per downtime event. $49M/year average lost revenue; $22M/year regulatory fines. | Sections 9.2, 10 |
| Gartner — IT Downtime Cost Benchmark (2014, confirmed ITIC 2022/2024) · ITIC 2022 and 2024 Hourly Cost of Downtime Surveys (n=1,000+ organisations annually) | $5,600/minute ($336,000/hour) cross-industry average. 81% of enterprises: ≥1 hr downtime costs >$300,000. 40% of enterprises: $1M–$5M per hour. Banking/finance/healthcare: avg. >$5M/hour in 2024 ITIC survey. | Sections 9.1, 9.2, 10 |
| Megladon Manufacturing — "Fiber Face-Off: Field vs Factory Terminated" (2024) · CableXpress — "Field Termination vs Factory Termination" · FASTCABLING research (2022) · Electronic Supply / eskc.com case study (2024) | Pre-terminated fiber: up to 50% overall installation cost reduction. 30–40% labour cost savings. 30–70% deployment time reduction. Field termination requires $15,000+ splicer kit. Factory LC/UPC insertion loss <0.3 dB vs. field <0.5 dB. | Section 9.2 |
| ANSI/TIA-942-C — Telecommunications Infrastructure Standard for Data Centers · Published May 2024, TIA TR-42.1 Engineering Committee | Five functional spaces (ER, MDA, HDA, ZDA, EDA). Minimum 800 mm cabinet width in MDA, IDA, HDA. VSFF connectors now permitted in distributor areas. Backbone cabling distances. | Sections 5.3, 5.4 |
| ANSI/TIA-568-C — Commercial Building Telecommunications Cabling Standard (568-C.1 published 2009; successor 568-D 2015, 568-E 2020) | Entrance Facility (EF), Equipment Room (ER), Telecommunications Room (TR) definitions and analogies to TIA-942-C spaces. Backbone cabling topology. | Sections 5.3, 5.4 |
| RFC 8040 — RESTCONF Protocol · IETF Network Working Group, January 2017 | Defines the RESTCONF protocol used by the POLATIS® OCS API. HTTP-based protocol for accessing data defined in YANG models. Resources at /api/data and /api/operations. | Sections 4, 8 |
| Regulation (EU) 2022/2554 — Digital Operational Resilience Act (DORA) · Effective 17 January 2025 | Requirement for documented, tested ICT incident management and recovery processes including physical infrastructure resilience. Applies to financial entities operating in EU member states. | Sections 1, 11, 14 |
| NERC CIP-014-3 — Physical Security Standard · North American Electric Reliability Corporation | Physical security and resilience requirements for bulk electric system transmission infrastructure. References physical layer protection as a control requirement. | Sections 1, 14 |
| IEEE 802.3 — Ethernet Standard · Relevant clauses: 802.3ba (40/100G), 802.3bs (200/400G), 802.3ck (800G) | Data rate specifications referenced in OCS protocol-agnostic capability claims. The POLATIS® OCS operates transparently across all 802.3 generations without reconfiguration. | Section 2 |
| Term | Definition |
|---|---|
| APS (Automatic Protection Switching) | A mechanism that automatically redirects a protected traffic stream from a failed working path to a pre-provisioned protection path, triggered by signal loss or power degradation detection. |
| CDR (Cable Distribution Rack / Optical Distribution Frame Cabinet) | HUBER+SUHNER's ODF rack cabinet — a 47U, 300 mm deep, front-accessible enclosure (e.g. CDR 1500) that houses LISA™ splice cassettes only. IANOS™ is a separate Top-of-Rack chassis — not inside the CDR. Features a lightweight C-shaped aluminium frame for full front access. |
| Cross-connect | A configured optical connection within the POLATIS® OCS that routes the signal from a specific ingress port to a specific egress port. Managed programmatically via the RESTCONF API. |
| Dark Fiber | An optical fiber strand that has been physically installed and terminated but is not currently carrying any active optical signal. The POLATIS® OCS supports pre-provisioning cross-connects on dark fibers so they are instantly activatable. |
| DirectLight™ | HUBER+SUHNER's patented beam-steering technology used in the POLATIS® OCS. Steers free-space optical beams between fiber ports independently of signal power level, wavelength, or direction. More vibration-tolerant than MEMS-based alternatives. |
| IANOS™ | HUBER+SUHNER's intelligent, high-density fiber management system. Used as the structured breakout and termination layer adjacent to the POLATIS® OCS, converting MTP/MPO trunk connections into individual LC duplex connections. |
| LISA™ Ribbon Splice Cassette System | HUBER+SUHNER's ribbon splice cassette system, housed inside the CDR ODF cabinet. ½RU per cassette. Front-plate options: 12× or 18× LCd (UPC/APC), 6×/12×/18× MTP Base-8/12, 36× SN UPC, 36× MDC UPC, Splice through. Incoming dark fiber is ribbon-spliced — no mated connectors on entry side. Splice loss <0.1 dB per splice. Source: HUBER+SUHNER LISA Cassettes product documentation. |
| Loss Budget | The maximum allowable total optical signal loss on a fiber link, calculated as the difference between transmitter output power and receiver sensitivity. All components in the path — fiber span, connectors, splices, OCS, panels — contribute to the total loss. |
| MEMS (Micro-Electro-Mechanical Systems) | A technology used in some optical switches that uses microscopic mirrors to deflect optical beams. Generally more sensitive to vibration and gravitational orientation than DirectLight™ beam-steering technology. |
| MTP/MPO | Multi-fiber push-on connector standard used for high-density fiber termination. A single MTP connector carries 12 or 24 fibers, enabling high-density trunk cables that reduce individual fiber count in cable trays by a factor of 12. |
| OCS (Optical Circuit Switch) | A device that routes optical signals at the physical layer (Layer 1) from any input port to any output port without converting the signal to electrical form. The POLATIS® Series 7000 is an OCS. |
| OEO (Optical-Electrical-Optical) | A switching architecture that converts the incoming optical signal to electrical, switches it electronically, and converts it back to optical. Protocol and data-rate specific; contrasted with OOO switching. |
| OOO (Optical-Optical-Optical) | An all-optical switching architecture that maintains the signal as light from input to output with no electrical conversion. Protocol and data-rate agnostic, ultra-low latency. Implemented by the POLATIS® OCS. |
| OPM (Optical Power Meter) | An integrated per-port sensor in the POLATIS® OCS that continuously measures the optical power level (in dBm) of the signal on that port. Used to detect signal degradation before complete link failure. |
| OPM Alarm | An automated alert triggered when the optical power measured by an OPM falls outside configured high or low thresholds. In a DR context, a low-power alarm on a working-path port initiates the protection switchover sequence. |
| RESTCONF | A protocol defined in IETF RFC 8040 that provides an HTTP-based programmatic interface to network device configuration and state data modeled in YANG. The POLATIS® switch exposes its full functionality via RESTCONF on port 8008. |
| RPO (Recovery Point Objective) | The maximum acceptable amount of data loss measured in time. An RPO of zero means no data loss is acceptable; all committed data must be recoverable at the secondary site. |
| RTO (Recovery Time Objective) | The maximum tolerable elapsed time between a service disruption and the restoration of service to an acceptable level. OCS-based physical layer automation reduces RTO for fiber fault scenarios from hours to under 5 minutes. |
| SDN (Software Defined Networking) | A network architecture approach that separates the control plane from the data plane, enabling centralized programmatic control of network resources. The POLATIS® RESTCONF API enables SDN-style control of the physical optical layer. |
| TCO (Total Cost of Ownership) | The complete cost of a technology deployment over its operational lifetime, including capital expenditure, operational labor, maintenance, and the cost of downtime or SLA penalties. |
| VOA (Variable Optical Attenuator) | An optional per-port component in the POLATIS® OCS that can reduce the optical power level on a specific port by a configurable amount. Used to equalise signal levels when working and protection paths have different fiber span lengths. |
| YANG | A data modeling language used to define the structure of network device configuration and state data, defined in IETF RFC 6020 and 7950. The POLATIS® switch YANG model (optical-switch and polatis-switch modules) defines all resources accessible via RESTCONF. |
This appendix is a self-contained summary of the full whitepaper for executive and management audiences. It can be distributed independently.
Enterprise DR programs automate routing, compute failover, and database replication — but leave the physical optical fiber layer managed manually. A single fiber cut or connector failure will defeat every higher-layer DR investment. Manual re-patching takes 30 minutes to 4 hours. The industry average cost of downtime exceeds $300,000 per hour.
HUBER+SUHNER POLATIS® Optical Circuit Switches (OCS) deployed at both primary and secondary sites provide automated physical-layer protection switching with sub-50 ms failover. Integrated optical power meters detect signal degradation before link failure. Pre-provisioned dark fiber protection paths activate with a single software command. The RESTCONF API (RFC 8040) enables any .NET or web platform to orchestrate physical-layer DR automation.
| Metric | Manual DR | OCS Automated DR |
|---|---|---|
| RTO (physical fault) | 60–300+ minutes | <5 minutes |
| Fault detection | NOC alert: 1–60 min | OPM alarm: <5 sec |
| Switchover speed | Human: 30–240 min | OCS: <50 ms |
| Avoided downtime cost | — | $600K+ per incident avoided |
| Technology lifespan | Per-protocol equipment | Protocol-agnostic: 1G–800G+ |
The POLATIS® RESTCONF API exposes cross-connect management, optical power telemetry, APS configuration, and event logging over standard HTTP. A .NET BackgroundService monitoring service — approximately 60 lines of C# — can poll OPM readings every 5 seconds and issue a PATCH to activate a protection path automatically when power falls below threshold. No vendor-specific SDK. No proprietary protocol. Standard RFC 8040.
DORA (EU, Jan 2025), NERC CIP-014, HIPAA, SOX, and PCI-DSS all include requirements for documented, tested physical infrastructure resilience. Automated L1 DR with POLATIS® OCS provides the evidence trail — via RESTCONF event logs and remote syslog — that auditors require.