About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Nanosecond Market Data Feeds – FPGA Centric vs. FPGA Accelerated Designs

Subscribe to our newsletter

The “Microburst” Problem

Trading architects face the constant challenge of lowering latency.  Today, this challenge is focused on achieving nanosecond speeds in a deterministic way, even during periods of high market activity, so called ‘microbursts’.

In fast moving ‘bursty’ markets, commodity CPUs are no longer capable of offering the lowest latency or tolerating microbursts. Memory bottlenecks, OS overheads, poor parallelism, compiler limitations and slow networking stacks are among the many factors constraining performance.

So, having exhausted conventional methods using commodity CPUs, architects of trading systems have been turning to FPGA technologies to improve latencies and address this microburst problem.

Applying FPGAs for Pipeline Processing

Because FPGAs chips are able to compute thousands of operations per clock cycle, they can be programmed to process market data using a technique known as pipelining. Pipelining allows for a message to begin being processed before the previous ones have been fully dealt with.

A properly pipelined FPGA design can guarantee to process each byte of incoming traffic with the same latency, even at 100% network saturation. However, not all FPGA approaches are the same.

Alternative FPGA Architectures

Today, there are two main approaches in use by the trading community:

* A CPU-based system, using FPGA acceleration technologies
* A pure FPGA system, using a matrix of FPGA’s

Architecture 1 – CPU System with FPGA Acceleration:

Collectively known as ‘hardware acceleration’ techniques, the system remains primarily a CPU centric architecture, with the selective use of FPGAs to off-load certain functions and accelerate the data processing.

In this case, FPGA acceleration is predominately achieved by the use of a PCIe card with an embedded FPGA. With this CPU centric architecture, the FPGA is used as an off-load technology. Since some market data processing does occur in the FPGA, improvements in latency are delivered. However this approach is still affected by the inherent bottlenecks of a traditional CPUs system.

For example, the limitation in PCIe slots, space and power makes it difficult to scale the number FPGAs used. In addition, the communication between two PCIe devices generally takes several microseconds, which is a very long time in the world of market data processing.

Therefore, this solution scales only by sharing the CPU further. Thus, performance will be degraded when data rates increase, when more feeds are being processed, or when more distribution interfaces are added.

This architecture is efficient for locally consumed market data, but it is still susceptible to microbursts.

Architecture 2 – A Pure FPGA Matrix Architecture:

When deterministic latency is critical, the goal is to avoid the bottlenecks of the CPU and maximise the use of FPGA’s throughout the whole market data processing cycle. This FPGA centric approach uses an expandable matrix of FPGA nodes that are linked with raw binary interconnections. This matrix architecture offers the flexibility to efficiently add FPGA resources as the need for processing increases.

In addition, to avoid bottlenecks that can occur with other system resources, each FPGA node comes with its own set of I/Os and memory. This modular matrix allows the system to grow proportionally across multiple feeds.

As a result, all functions of the market data parsing, book building, filtering and distribution can all be done in hardware with no bottleneck, irrespective of the number of feeds received, the rate of data  or the number of downstream consumers.

A Platform for the Future

When looking for optimised speed, it is important consider the complete set of functions needed to be performed to receive, manage and distribute market data. A solution making partial use of FPGAs may deliver some acceleration; however bottlenecks are moved to another part of the system, usually the function that is implemented in software.

A pure FPGA centric design that utilises a modular FPGA approach to scale capacity, can maintain nanosecond speeds even during market microbursts. Processing and normalising market data is the first step. But a matrix of FPGAs can then be extend to enrich fields, trigger order executions, conduct risk checks and even host the actual trading algorithm.

Subscribe to our newsletter

Related content

WEBINAR

Upcoming Webinar: Leveraging interoperability: Laying the foundations for unique best-of-breed trading solutions

Date: 23 May 2024 Time: 10:00am ET / 3:00pm London / 4:00pm CET Duration: 50 minutes Interoperability on the trading desk promises more actionable insights, real-time decision making, faster workflows and reduced errors by ensuring data consistency across frequently used applications. But how can these promises be kept in an environment characterised by multiple applications...

BLOG

New DTCC Report Recommends Best Practices to Achieve T+1 Settlement Success

In anticipation of the transition to a T+1 settlement cycle in the US, the Depository Trust & Clearing Corporation (DTCC) has released a new report, “Hitting 90% Affirmation by 9:00 PM ET on Trade Date: The Key to T+1 Success”, which highlights the importance of automating post-trade processes to achieve success in the upcoming T+1...

EVENT

AI in Capital Markets Summit London

The AI in Capital Markets Summit will explore current and emerging trends in AI, the potential of Generative AI and LLMs and how AI can be applied for efficiencies and business value across a number of use cases, in the front and back office of financial institutions. The agenda will explore the risks and challenges of adopting AI and the foundational technologies and data management capabilities that underpin successful deployment.

GUIDE

Entity Data Management Handbook – Fourth Edition

Welcome to the fourth edition of A-Team Group’s Entity Data Management Handbook sponsored by entity data specialist Bureau van Dijk, a Moody’s Analytics company. As entity data takes a central role in business strategies dedicated to making the customer experience markedly better, this handbook delves into the detail of everything you need to do to...