About a-team Marketing Services

A-Team Insight Blogs

Researchers Harness Supercomputers for Game Changing Financial Markets Analysis

Subscribe to our newsletter

Researchers from a number of U.S. universities, tapping into supercomputing power and optimising code, believe they have made a “game changing” step in the speedy analysis of financial markets, reducing operations from hours or even days, down to minutes.

The researchers from the University of Illinois, the University of Pittsburgh and the University of San Diego are making use of parallel processing supercomputer capacity provided via XSEDE – the Extreme Science and Engineering Discovery Environment – in order to run analytics, and have optimised their code to provide as much as a 126x speedup compared to previous analysis undertaken in 2010.

That performance boost, which the researchers reported at last month’s XSEDE conference in San Diego, will allow them to analyse market phenomenon with nanosecond time granularity across the entire Nasdaq equities market in just a couple of hours. This includes the impact of high frequency and other low-latency trading strategies, and could be used to detect whether frowned upon strategies, such as quote stuffing, have been deployed.

XSEDE provides the research community with access to a variety of IT resources, including 16 supercomputers, visualisation and data analysis systems. The supercomputers used by this particular research comprise:

* Blacklight at the Pittsburgh Supercomputing Centre. Blacklight is an Intel-based SGI shared memory system intended for applications that require a large memory space for computational tasks.
* Gordon at the San Diego Supercomputer Centre. Gordon is a Flash-based supercomputer, also incorporating Intel chips, designed in partnership with Appro (now Cray) for data intensive workloads.
* Stampede at the Texas Advanced Computing Centre, University of Texas at Austin. Stampede – currently cited as the sixth most powerful supercomputer in the world – was designed in collaboration with Dell, and includes Intel multi-core and Xeon Phi many-core processors for highly parallel computational processing.

Code optimisation has thus far focused on several areas required to build a limit order book for each security, including the input and output of data and its pre-processing to convert it into suitable formats for computation.

That optimised code is then executed in parallel for each security in market, with only a small subset requiring a lengthy order book construction phase. Once an order book has been constructed for each security, then analytics can be run. 

Previous research by members of the team and others – using just Blacklight – into the impact of the non-reporting of ‘odd lot’ trades that are commonly executed by HFT strategies, on overall volume, suggested that nearly 5% of volume was omitted. Partly as a result of that research, regulators plan to introduce reporting of such trades this October.

The deployment of supercomputing technology – especially parallel processing and data intensive/in-memory computing – is likely to become more common for such functions as back-testing and strategy construction, as trading firms adopt intelligent approaches that are not simply dependent of low-latency execution.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Trade the Middle East & North Africa: Connectivity, Data Systems & Processes

In Partnership With As key states across the region seek alternatives to the fossil fuel industries that have driven their economies for decades, pioneering financial centres are emerging in Egypt, United Arab Emirates (UAE), Saudi Arabia and beyond. Exchanges and market intermediaries trading in these centres are adopting cutting-edge technologies to cater to the growing...

BLOG

Survey Highlights Challenges in Investment Research Data Amid Rising Demand for Systematic Strategies

The growing adoption of quantitative and AI/Machine Learning (ML) techniques, alongside the rise of systematic investment strategies, has elevated the importance of investment research data, according to a recent Bloomberg survey of over 150 quants, research analysts, and data scientists. The survey, conducted during a global series of client workshops, identified data coverage, timeliness, and...

EVENT

Data Management Summit New York City

Now in its 15th year the Data Management Summit NYC brings together the North American data management community to explore how data strategy is evolving to drive business outcomes and speed to market in changing times.

GUIDE

AI in Capital Markets: Practical Insight for a Transforming Industry – Free Handbook

AI is no longer on the horizon – it’s embedded in the infrastructure of modern capital markets. But separating real impact from inflated promises requires a grounded, practical understanding. The AI in Capital Markets Handbook 2025 provides exactly that. Designed for data-driven professionals across the trade life-cycle, compliance, infrastructure, and strategy, this handbook goes beyond...