About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

QuantHouse Offers Historical Data on-Demand to Algo Traders

Subscribe to our newsletter

QuantHouse has released Historical Data on-Demand, a service designed to speed up the research, development and back-testing phase of any trading strategy, and allow clients to implement new trading ideas within days rather than weeks or months.

The company is offering up to 10 years of historical data on-demand for the US, European and Asia-Pacific markets. Access to the data is available via a web portal, so clients can search for the data they need and purchase it online using a web browser of choice. The historical datasets purchased are delivered as flat files and are available for immediate integration into any system, without the need to integrate an API. Historical data can be replayed over prior time periods with the results being refined and adjusted to optimise trading performance.

While the time taken to fulfil the research, development and back-testing cycle of a trade can push execution beyond optimal timings, QuantHouse says giving research and development teams Historical Data on-Demand will enable them to rapidly test new and current trading strategies, and detect potential losses or degradation of the strategies within days, not weeks.

Stephane Leroy, chief revenue officer and co-founder of QuantHouse, explains: “The trading landscape has changed significantly in the past few years, it is no longer about how fast your trades are sent, but how quickly your trading strategy can be ready. To move away from speed trading to smart trading, you need access to trusted, reliable and consistent data on-demand, so that you can spot changes and emerging patterns in the market quickly and evaluate and adjust your trading strategy accordingly. Our Historical Data on-Demand service gives clients an advantage by moving them into a much more real-time environment.”

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: The Role of Data Fabric and Data Mesh in Modern Trading Infrastructures

The demands on trading infrastructure are intensifying. Increasing data volumes, the necessity for real-time processing, and stringent regulatory requirements are exposing the limitations of legacy data architectures. In response, firms are re-evaluating their data strategies to improve agility, scalability, and governance. Two architectural models central to this conversation are Data Fabric and Data Mesh. This...

BLOG

TXSE Selects Exegy FPGA Technology for Market Data Infrastructure

The Texas Stock Exchange (TXSE) has selected Exegy to provide FPGA-based market data feed handlers as part of its launch infrastructure. TXSE is positioning itself as the first fully integrated U.S. equities exchange built from scratch in more than 25 years. As part of that ground-up approach, the venue has opted to deploy FPGA technology...

EVENT

RegTech Summit New York

Now in its 9th year, the RegTech Summit in New York will bring together the RegTech ecosystem to explore how the North American capital markets financial industry can leverage technology to drive innovation, cut costs and support regulatory change.

GUIDE

Connecting to Today’s Fast Markets

At the same time, the growth of high frequency and event-driven trading techniques is spurring demand for direct feed services sourced from exchanges and other trading venues, including alternative trading systems and multilateral trading facilities. Handling these high-speed data feeds its presenting market data managers and their infrastructure teams with a challenge: how to manage...