About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

NovaSparks’ STAC-M1 Benchmark Highlights Determinism Under Load

Subscribe to our newsletter

A just released STAC Report covering the performance of NovaSparks’ FPGA market data platform highlights not just its processing latency but also the deterministic nature of that latency under different data loads.

The STAC-M1 benchmark (as defined by financial markets participants and administered by the Securities Technology Analysis Center) measures the performance of direct data feed processing solutions according to a number of different criteria, including end-to-end latency and throughput.

The NovaSparks solution uses only FPGA microprocessors in its architecture, in contrast to offerings that augment mainstream x86 processors with FPGA acceleration of certain functions. As such, the company claims its platform is less prone to latency variance – or jitter – compared to its competitors.

The predictable – or deterministic – nature of the NovaSparks platform was borne out by the benchmark tests conducted by STAC, which simulated a Nasdaq TotalView ITCH feed being received at 2x and 20x a typical data rate at market open and close.

According to STAC: “During replay at 20 times recorded market data volumes, the NovaSparks solution demonstrated mean latency of just 1.4 microseconds, along with 99.9th percentile latency of just 2.8 microseconds. Jitter (standard deviation) was just 0.12 microseconds at 2x market rate and 0.15 microseconds at 20x market rate.” See this chart:

 

While for many the push to reduce latency further is not as big a focus as it once was, maintaining deterministic latency is still important for many trading strategies. Keeping latency constant under extreme market conditions has historically been a challenge, and its one that NovaSparks is looking to solve with its FPGA platform.

“Deterministic processing of market data at ultra-low latency rates is a breakthrough for an industry that is constantly re-assessing their ability to trade across all market conditions,” says Michal Sanak, CIO at proprietary trading firm RSJ.

Subscribe to our newsletter

Related content

WEBINAR

Upcoming Webinar: Agility as Alpha: How Trading Infrastructure Determines Who Wins in Volatile Markets

Date: 21 May 2026 Time: 10:00am ET / 3:00pm London / 4:00pm CET Duration: 50 minutes Tariff shocks, geopolitical realignment and macroeconomic regime shifts are redrawing the investment landscape faster than most firms’ technology stacks can keep up. For hedge funds and asset managers, the ability to move quickly into new asset classes, geographies or...

BLOG

Beyond the Blueprint: Integrating Data Fabric and Data Mesh in Capital Markets

The demands placed upon modern trading infrastructures, driven by increasing data volumes, the mandate for real-time processing, and stringent regulatory requirements, are exposing the limitations of historical data architectures. In response, capital markets firms are accelerating the re-evaluation of their data strategies to secure greater agility, scalability, and enhanced governance. A recent webinar hosted by...

EVENT

AI in Capital Markets Summit London

Now in its 3rd year, the AI in Capital Markets Summit returns with a focus on the practicalities of onboarding AI enterprise wide for business value creation. Whilst AI offers huge potential to revolutionise capital markets operations many are struggling to move beyond pilot phase to generate substantial value from AI.

GUIDE

Evaluated Pricing

Valuations and pricing teams are facing a much higher degree of scrutiny from both the regulatory community and the investor community in the glare of the post-crisis data transparency spotlight. Fair value price transparency requirements and the gradual move towards a more harmonised accounting standards environment is set within the context of the whole debate...