Arista jammed a lot of power into an 11RU chassis. 384 10Gb ports and 10Tb of switching capacity with 3.4 microseconds port-to-port with 64 byte packets. The switch can move 5.76 million packet per second. Each line card as 2.3 GB of RAM for packet buffering- -enough for 50 ms of packets per/10Gb switch port. The 7500 has front to back cooling, an important feature for managing cooling in equipment racks and dual supervisor modules for redundant management. Pricing starts at $140,000.
Who needs that kind of power? The obvious answer is financial institutions in engaged in trading where microseconds matter and HPC applications like research and modeling. Phil Papadapoulos, director of UC Systems, San Diego Computing Center, has been using an Arista 7500 for several months. The researchers he supports often need to analyze terabytes of data that are striped across many disks on a SAN. "The 7500 makes a good interconnect between server clusters and storage. What's important to us is that we can bi-sect the 7500 and burst 192 ports to 192 ports."
Your data center may start to look more like an HPC cluster in the future. Virtualization lets you consolidate several servers onto a single platform. To maintain adequate performance, the hypervisor server needs high-capacity networking and storage to keep up with the virtual machines. One of the reasons why companies want to stick with fiber channel is that it is high-capacity network separate from the data network that has a predictable low latency required for good storage performance. "Approximately 4.5 microsecond port-to-port latency," Papadapoulos said, "is close to the 1.7 to 2.2 microseconds latency with Infiniband. In many cases, the approximately 10 percent difference is not important." With a low latency and the large buffers per port, the 7500 should be able to keep up with any data you throw at it.