About a-team Marketing Services
The leading knowledge platform for the financial technology industry
The leading knowledge platform for the financial technology industry

A-Team Insight Blogs

Researchers Harness Supercomputers for Game Changing Financial Markets Analysis

Subscribe to our newsletter

Researchers from a number of U.S. universities, tapping into supercomputing power and optimising code, believe they have made a “game changing” step in the speedy analysis of financial markets, reducing operations from hours or even days, down to minutes.

The researchers from the University of Illinois, the University of Pittsburgh and the University of San Diego are making use of parallel processing supercomputer capacity provided via XSEDE – the Extreme Science and Engineering Discovery Environment – in order to run analytics, and have optimised their code to provide as much as a 126x speedup compared to previous analysis undertaken in 2010.

That performance boost, which the researchers reported at last month’s XSEDE conference in San Diego, will allow them to analyse market phenomenon with nanosecond time granularity across the entire Nasdaq equities market in just a couple of hours. This includes the impact of high frequency and other low-latency trading strategies, and could be used to detect whether frowned upon strategies, such as quote stuffing, have been deployed.

XSEDE provides the research community with access to a variety of IT resources, including 16 supercomputers, visualisation and data analysis systems. The supercomputers used by this particular research comprise:

* Blacklight at the Pittsburgh Supercomputing Centre. Blacklight is an Intel-based SGI shared memory system intended for applications that require a large memory space for computational tasks.
* Gordon at the San Diego Supercomputer Centre. Gordon is a Flash-based supercomputer, also incorporating Intel chips, designed in partnership with Appro (now Cray) for data intensive workloads.
* Stampede at the Texas Advanced Computing Centre, University of Texas at Austin. Stampede – currently cited as the sixth most powerful supercomputer in the world – was designed in collaboration with Dell, and includes Intel multi-core and Xeon Phi many-core processors for highly parallel computational processing.

Code optimisation has thus far focused on several areas required to build a limit order book for each security, including the input and output of data and its pre-processing to convert it into suitable formats for computation.

That optimised code is then executed in parallel for each security in market, with only a small subset requiring a lengthy order book construction phase. Once an order book has been constructed for each security, then analytics can be run. 

Previous research by members of the team and others – using just Blacklight – into the impact of the non-reporting of ‘odd lot’ trades that are commonly executed by HFT strategies, on overall volume, suggested that nearly 5% of volume was omitted. Partly as a result of that research, regulators plan to introduce reporting of such trades this October.

The deployment of supercomputing technology – especially parallel processing and data intensive/in-memory computing – is likely to become more common for such functions as back-testing and strategy construction, as trading firms adopt intelligent approaches that are not simply dependent of low-latency execution.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Integrating Intelligent Machine Readable News

Intelligent machine readable news is a powerful tool in the arsenals of trading and investment firms seeking competitive advantage. It turns unstructured data into actionable insight and can be used, for example, to uncover market trends, identify correlations and evaluate sentiment. In turn, it can inform quant strategies and predictive models. While machine readable news...

BLOG

The Current and Future Landscape of Real-Time Enterprise Market Data Distribution

Distributing real-time market data across the enterprise is a complex proposition, with many moving parts. Banks and other financial institutions spend vast sums to ensure that the correct market data, the lifeblood of their trading activities, is sourced, normalised, controlled and distributed out to end users and applications in the timeliest manner, with minimal downtime....

EVENT

Data Management Summit New York

Now in its 12th year, the Data Management Summit (DMS) in New York brings together the North American, capital markets enterprise data management community, to explore the evolution of data strategy and how to leverage data to drive compliance and business insight.

GUIDE

ESG Handbook 2021

A-Team Group’s ESG Handbook 2021 is a ‘must read’ for all capital markets participants, data vendors and solutions providers involved in Environmental, Social and Governance (ESG) investing and product development. It includes extensive coverage of all elements of ESG, from an initial definition and why ESG is important, to existing and emerging regulations, data challenges...