Testing Details and Highlights
Performance testing compared both highend and midrange categories of products from the top three ADC vendors, based on market share. In the highend category, the comparative test included F5's VIPRION® chassis with one PB200 and another with two PB200 blades, the BIG-IP 8900, Citrix's NetScaler MPX-17000 Classic and nCore products, and Cisco's ACE20. For the midrange category F5 compared the BIG-IP 3900 and Cisco's ACE 4710. Citrix's midrange product was not tested because it is the same architecture and technology as the MPX-17000, which is included in the report. Customers can extrapolate their midrange product performance based on the results in the report.
-Testing measured a range of realworld use cases including requests per second for various Layer 7 response sizes, compression, caching, SSL, and Layer 4 performance.
-At a similar or lower price point compared to others, BIG-IP products demonstrated the highest processing capabilities for Layer 7 requests, making F5 solutions the best value for enterprises and service providers.
For midrange ADCs, F5 solutions achieved up to 15x the Layer 7 performance of the nearest competitor.
For highend ADCs, F5 solutions achieved up to 9x the Layer 7 performance of the other tested devices.
-With F5's unique Clustered Multiprocessing (CMP) capability and chassis design architecture, BIG-IP solutions provide truly linear scaling capabilities where other solutions do not. This enables F5 customers to add ondemand capacity and processing as required by their users, applications, and infrastructures.
-F5 devices have 63%-3700% greater energy efficiency compared to the other vendors' products.
"The guiding principles of a comparative report are that it must be accurate, transparent, and reproducible," said Karl Triebes, SVP of Product Development and CTO at F5. "We strive to provide as much factual information as possible so the customer can make informed purchasing decisions, and that others in good faith can reproduce these tests and see the same results. F5 conducted these tests with our own performance experts, as opposed to contracting with an 'independent' third party hired to produce favorable results. In addition, we invite constructive feedback on all materials or product configurations to ensure testing is accurate. We're committed to publishing test results following this methodology, and it is our hope that the industry will follow our lead in adopting an open approach to technology evaluations."
Because F5's comparative testing methodology is designed to be reproducible, customers can make more informed assessments of vendors' offerings and trust the results to be reliable, since they can be independently verified and reflect realworld scenarios. In contrast to this comparative approach, thirdparty competitive testing organizations often make unqualified assertions or do not disclose their testing methodologies, leading to inaccurate and sometimes misleading conclusions. And because competitive test methodologies and configuration information are rarely disclosed, performance claims can be difficult to reproduce and validate, making a fair and balanced analysis unachievable.
"Historically, ADC vendors have often measured and categorized performance capabilities individually, resulting in a level of ambiguity for customers attempting to interpret or compare one vendor's claims to another's," said Zeus Kerravala, SVP of Enterprise Research at Yankee Group. "By releasing both their performance results and testing methodology, F5 is going a long way toward encouraging prospective customers, business partners, and even competitors to evaluate solutions on a level playing field. This approachalong with a wide range of testing parametershelps minimize the effect of variable performance measuring techniques and definitions, selective disclosure, and marketing hype so that the facts can be borne out."