Multithreading weaves its way into nets

Network systems increasingly need to be application-aware to control access, allocate resources and prioritize traffic.

Maintaining stateful packet flow information at gigabit-per-second line speeds requires a rate of random memory access that is beyond the capability of today’s traditional processors.

And application-specific integrated circuits (ASICs), while fast, can’t keep pace with constant changes in network protocols and applications.

A new architectural approach for application-aware networks has demonstrated tangible benefits: massive multithreading (MMT).

Understanding this technology is key to evaluating the next wave of network infrastructure.

MMT maximizes memory throughput by letting a greater number of memory requests to be active simultaneously.

Because of this, MMT is able to perform sophisticated protocol processing in software at throughput levels that previously required one or more dedicated ASICs.

This optimization of RAM access also enables MMT to overcome the stateful packet throughput limitations of traditional multiprocessors.

Each memory operation introduces processing latency. To maintain low latency and high throughput in the face of demanding memory access requirements, network-oriented multiprocessing architectures need to support a very high number of simultaneous threads and execution pipelines, each with its own dedicated processing resources.

By pushing packets in parallel through 100 or more threads, deep packet inspection can be sustained at 10Gbps data rates with a latency of less than one millisecond — an impossible task for two (or even two dozen) threads operating at today’s high-end clock rate of 4GHz.

This allows MMT to accommodate VOIP and other delay-sensitive applications on high-speed backbones.

As bandwidth continues to increase, an even greater number of simultaneous threads may become necessary.

Initially, the demand will be met with higher thread counts implemented in tribes of multiple streams served by separate cores.

Such advances are occurring already for network access control and identity-based network applications in LANs, where 128 threads is state of the art. Over time, advances in technology will permit higher levels of protocol processing parallelism with a greater number of streams and execution pipelines.

QuickLink: 066118

–Nemirovsky is chief scientist for ConSentry Networks.

Related Download
Virtualization: For Victory Over IT Complexity Sponsor: HPE
Virtualization: For Victory Over IT Complexity
Download this white paper to learn how to effectively deploy virtualization and create your own high-performance infrastructures
Register Now