About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

BittWare’s TeraBox Bulks Up FPGA Processing for Trading Scale and Analytics

Subscribe to our newsletter

FPGA specialist BittWare has introduced TeraBox, an appliance that delivers up to 16 FPGAs, and targeted at high scale trading and analytics applications.

TeraBox supports up to eight BittWare S5-PCIe-DS cards, each of which hosts two Altera Stratix V FPGAs, 64 gigabytes of RAM and 16 10gE network ports. Thus, each 5U appliance can scale to 16 FPGAs, 512GB of RAM and 128 10gE ports. The appliance can also optionally host a traditional x86 processor, perhaps for monitoring or co-ordination functionality.

According to BittWare’s vice president of systems and solutions Ron Huizen, TeraBox has two likely applications in the financial markets:

* For trading systems where the entire application logic is hosted on the FPGA card, TeraBox offers high scale in one appliance, thus reducing cost of deployment compared with server-hosted approaches. Algorithmic trading and real-time risk control are applications that can likely be deployed more cost effectively with TeraBox.
* For analytics applications, such as algo back testing, pre-trade analytics and risk managemnt, TeraBox’s multiple FPGAs can work together to provide parallelised performance. Connectivity between the FPGAs can be acheived via the chassis PCIe bus, or via the 10gE network.

The latter analytics example would be to some extent breaking new ground for FPGAs, which typically have been dedicated to specific functions, such as data feed processing. For the most part, parallel analytics applications have been targeted at multi-core x86 processors, or at GPUs.

A key to deploying FPGAs for parallel applications will be the introduction of the next version of the OpenCL programming framework. OpenCL 2.0 – the specification for which was released this July – calls for support for dynamic parallelism and shared virtual memory.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Data platform modernisation: Best practice approaches for unifying data, real time data and automated processing

Financial institutions are evolving their data platform modernisation programmes, moving beyond data-for-cloud capabilities and increasingly towards artificial intelligence-readiness. This has shifted the data management focus in the direction of data unification, real-time delivery and automated governance. The drivers of this transition are improved operational efficiency as manual processes are replaced by faster, more accurate automated...

BLOG

When Margin Moves Upstream: How TT is Reworking Trading Decisions After the OpenGamma Deal

More than a month after completing its acquisition of OpenGamma, Trading Technologies is beginning to articulate how the deal is intended to change the way firms think about margin, capital efficiency, and trading decision-making. Rather than positioning margin as a downstream risk or treasury concern, TT is now framing capital efficiency as a front-office variable...

EVENT

TEST Event page 1

Now in its 15th year the TradingTech Summit London brings together the European trading technology capital markets industry and examines the latest changes and innovations in trading technology and explores how technology is being deployed to create an edge in sell side and buy side capital markets financial institutions.

GUIDE

The Reference Data Utility Handbook

The potential of a reference data utility model has been discussed for many years, and while early implementations failed to gain traction, the model has now come of age as financial institutions look for new data management models that can solve the challenges of operational cost reduction, improved data quality and regulatory compliance. The multi-tenanted...