About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

STAC Benchmarks IBM’s Hadoop

Subscribe to our newsletter

STAC – aka the Securities Technology Analysis Center – has benchmarked IBM’s proprietary Platform Symphony implementation of Hadoop MapReduce, versus the standard open source offering, to compare their respective performance. On average, IBM’s implementation performed jobs 7.3 times faster than the standard, reducing total processing time by a factor of six.

Better known for its benchmarking of low-latency trading platforms, STAC leveraged the Statistical Workload Injector for MapReduce (SWIM), developed by the University of California at Berkeley. SWIM provides a large set of diverse MapReduce jobs based on production Hadoop traces obtained from Facebook, along with information to enable characterisation of each job. STAC says it undertook the benchmarking because many financial markets firms are deploying Hadoop.

The hardware environment for the testbed consisted of 17 IBM compute servers and one master server communicating over gigabit Ethernet. STAC compared Hadoop version 1.0.1 to Symphony version 5.2. Both systems ran Red Hat Linux and used largely default configurations.

IBM attributes the superior performance of its offering in part to its scheduling speed. IBM’s Hadoop is API-compatible with the open source offering but runs on the Symphony grid middleware that became IBM’s with its aquisition of Platform Computing, which closed in January of this year.

For more information on STAC’s IBM Hadoop benchmark, see here.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: The challenges and potential of data marketplaces

Data is the lifeblood of capital markets. It is also a valuable commodity providing financial institutions with additional insight when gathered in an internal data marketplace, or packaged and sold externally to other institutions. While the theory is sound, the practice of setting up a data marketplace can be challenging. Internally, vast amounts of data...

BLOG

Vendor Strategy: How SpaceTime is Distilling Complex Markets into Actionable Insights

With regulators placing an ever-greater focus on best execution, firms across the buy and sell side are making increasing use of transaction cost analysis (TCA) to demonstrate to customers how their execution performance measures up against various benchmarks. But TCA is not just a regulatory tool. Real-time TCA can also be an incredibly useful resource...

EVENT

Data Management Summit New York City

Now in its 12th year, the Data Management Summit (DMS) in New York brings together the North American, capital markets enterprise data management community, to explore the evolution of data strategy and how to leverage data to drive compliance and business insight.

GUIDE

The Data Management Challenges of Client Onboarding and KYC

This special report accompanies a webinar we held on the popular topic of The Data Management Challenges of Client Onboarding and KYC, discussing the data management challenges of client onboarding and KYC, and detailing new technology solutions that have the potential to automate and streamline onboarding and KYC processes. You can register here to get immediate...