With the rapid expansion of the Johor-Singapore Special Economic Zone (JS-SEZ) and the growing demand for AI, financial trading, and real-time cloud applications, extending infrastructure from Singapore into Johor is no longer optional—it is a strategic requirement.
However, for latency-sensitive workloads, distance directly impacts performance.
This leads to a critical question for infrastructure teams:
Can sub-1ms latency between Singapore and Johor actually be achieved—and how do you deploy it in real-world conditions?
In 2026, the answer is yes. But achieving it requires more than just the right architecture—it requires hardware that can actually be delivered on time.
This guide breaks down the technical feasibility, architecture design, hardware selection, and deployment strategy required to deliver sub-1ms cross-border connectivity.
Table of Contents
- Part 1: The Physics of Sub-1ms Latency
- Part 2: Architecture Design for Deterministic Low Latency
- Part 3: Hardware Selection — Where Most Designs Fail
- Part 4: From Design to Deployment — The Real Bottleneck
- Part 5: Deployment Best Practices for Sub-1ms Performance

Part 1: The Physics of Sub-1ms Latency (For CTOs & Architects)
Sub-1ms latency is not theoretical—it is physically achievable.
- Distance (SG ↔ Johor): ~15–30 km
- Fiber propagation: ~5 microseconds per km
This results in:
- ~0.1–0.15 ms one-way latency
- ~0.2–0.3 ms round-trip baseline
Where Latency Actually Comes From
In real deployments, latency is dominated by:
- Switch forwarding delay
- Number of hops
- Buffering and congestion
- Protocol overhead
The fiber is not the bottleneck—the network design is.
Part 2: Architecture Design for Deterministic Low Latency
To consistently achieve sub-1ms latency, both the physical layer and logical architecture must be optimized.
Physical Layer: Dark Fiber + DWDM
For mission-critical workloads, relying on shared carrier networks introduces unpredictable latency and jitter.
Instead, enterprises deploy:
- Dark fiber across SG–Johor links
- DWDM (Dense Wavelength Division Multiplexing)
- Coherent optics (400ZR / 800ZR)
This allows:
- Direct 400G/800G transport
- No intermediate carrier switching
- Deterministic, ultra-low latency
Logical Layer: Spine-Leaf + EVPN-VXLAN
On top of the physical layer:
- Extend Spine-Leaf architecture across sites
- Use EVPN-VXLAN for scalable Layer 2 extension
- Maintain consistent east-west traffic performance
Architecture Selection
Comparison of common cross-border architectures:
| Architecture | Latency | Use Case |
| L2 Extension | Lowest | Simple deployments |
| EVPN-VXLAN | Low & scalable | Enterprise / cloud |
| DCI (DWDM) | Ultra-stable | AI / financial workloads |
Part 3: Hardware Selection — Where Most Designs Fail
At this point, most architectures look perfect on paper.
But in real deployments, this is where projects break.
The 2026 Reality
- High-performance switches (400G / low-latency ASICs) are in short supply
- Models like MSN3700 and high-end IB switches are frequently backordered
- Lead times can delay projects by months
The biggest risk is not latency—it’s hardware availability.
Practical Hardware Strategy
- Select ultra-low-latency switches (<1µs forwarding)
- Use coherent optics (400ZR) to eliminate extra devices
- Prioritize deployable hardware over theoretical best-fit models
Reference hardware for SG–Johor interconnect:
| Role | Model | Key Benefit | Availability |
| Leaf | MSN2410 / MSN2700 | Low-latency edge | 7–10 days |
| Spine | MSN4600-CS2FC | High-bandwidth core | 7–10 days |
| Advanced Spine | MSN3700 | Ultra-low latency | Limited |
Part 4: From Design to Deployment — The Real Bottleneck
In theory, sub-1ms latency is straightforward.
In practice, projects are delayed because:
- Hardware is unavailable
- Procurement cycles are too slow
- Vendor lead times are unpredictable
This has led to a shift in how infrastructure teams approach deployment:
Design is no longer based on ideal hardware—it is based on what can actually be deployed now.
To support this shift, teams increasingly rely on platforms like Router-switch to identify available switching hardware that meets latency and bandwidth requirements.
At the same time, tools like IT-Price allow procurement teams to check inventory and generate real-time quotes.
Part 5: Deployment Best Practices for Sub-1ms Performance
To ensure consistent low latency:
- Minimize hop count across the network
- Use direct fiber routes where possible
- Avoid oversubscription in critical paths
- Enable ECN and PFC for AI workloads
- Design active-active redundancy without adding latency
Example CLI command to verify software version:
switch# show version
Final Thoughts
From a technical perspective, achieving sub-1ms latency between Singapore and Johor is entirely feasible.
The real challenge in 2026 is execution.
- Hardware shortages delay projects
- Procurement inefficiencies increase risk
- Ideal designs often fail in real-world conditions
The most successful deployments follow a simple principle:
Build your network around what you can deploy—not what you wish you had.
To move forward without delays:
- Use IT-Price to verify availability
- Source hardware via Router-switch
By combining low-latency architecture with real-world procurement strategy, enterprises can deliver sub-1ms cross-border connectivity—reliably and on schedule.

Expertise Builds Trust
20+ Years • 200+ Countries • 21500+ Customers/Projects
CCIE · JNCIE · NSE7 · ACDX · HPE Master ASE · Dell Server/AI Expert





















































































































