Networking & CommunicationsProtocols & Standards

Network Latency

Overview

Direct Answer

Network latency is the measurable delay, typically expressed in milliseconds, between the initiation of a data transmission and the receipt of that data at its destination. It encompasses the cumulative time required for data to traverse physical network infrastructure, routers, switches, and intermediate nodes.

How It Works

Data packets travel across physical and logical network paths, experiencing delays at each hop due to transmission time, propagation delay through cables or wireless channels, processing time at network devices, and queuing when congestion occurs. The total delay is the sum of these components: transmission delay (packet size divided by link bandwidth), propagation delay (distance divided by signal speed), processing delay at routers, and queuing delay in buffers.

Why It Matters

Latency directly impacts user experience in real-time applications such as video conferencing, online gaming, and financial trading, where sub-100 millisecond response times are critical. High latency reduces responsiveness, degrades application performance, increases transaction settlement times, and can create synchronisation problems in distributed systems.

Common Applications

Latency is a primary concern in cloud computing, video streaming services, VoIP telephony, multiplayer gaming, autonomous vehicle control systems, and high-frequency trading platforms where millisecond-level performance directly influences operational effectiveness and user satisfaction.

Key Considerations

Latency and bandwidth are distinct but interdependent; high bandwidth does not guarantee low latency. Geographic distance, network topology, congestion patterns, and hardware capabilities all influence latency, and optimising it often requires addressing multiple layers simultaneously rather than focusing on a single network segment.

More in Networking & Communications