About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Datawatch Adds Panopticon Streams Real-Time Stream Processing Engine

Subscribe to our newsletter

Datawatch has increased the speed of real-time streaming and time series data analytics with stream processing engine Panopticon Streams. The engine can be used as a stand-alone solution or in conjunction with Panopticon’s Visual Analytics platform.

Peter Simpson, vice president of visualisation strategy at Datawatch Panopticon, says: “Capital markets customers will benefit from Panopticon Streams’ support of several key use cases, including best execution, real-time P&L, transaction cost analysis and trader and trading surveillance.

“The addition of the engine’s capabilities means we now offer a general purpose streaming analytics platform. It has applications anywhere organisations need to identify anomalies and outliers, investigate their causes, back test potential solutions, and then alter their business processes to address the issue. Given the software’s ability to handle real-time and time series data, we believe it will be most useful in electronic trading, telecommunications, energy, and IoT applications.”

The combination of stream processing, rapid data comprehension through visual analysis, faster investigation through time series analysis and playback down to the individual tick, is designed to help organisations make timely, more informed decisions that have immediate financial impact.

Built on the Apache Kafka platform, Panopticon’s solutions enable business users to build sophisticated Kafka data flows with no coding. Users who understand the business problems can create their own data flows, which can use information from a number of sources and incorporate joins, aggregations, conflations, calculations, unions and mergers, and alerts. Analysts can visualise processed data using Panopticon Visual Analytics and deliver it to Kafka, Kx kdb+, InfluxDb, or any SqL database.

Subscribe to our newsletter

Related content

WEBINAR

Upcoming Webinar: From Data to Alpha: AI Strategies for Taming Unstructured Data

Date: 16 April 2026 Time: 9:00am ET / 2:00pm London / 3:00pm CET Duration: 50 minutes Unstructured data and text now accounts for the majority of information flowing through financial markets organisations, spanning research content, corporate disclosures, communications, alternative data, and internal documents. While AI has created new opportunities to extract signals, many firms are...

BLOG

From Silos to Sequencers: Why Core Trading Architectures Are Being Rewritten for 24/7 Markets

The most consequential changes facing financial markets technology in 2026 will not be driven by new asset classes or incremental latency gains, but by a fundamental rethinking of how trading systems are architected at their core. For decades, market participants have organised technology around functional silos: execution, risk, middle office, post-trade. These boundaries were reinforced...

EVENT

AI in Capital Markets Summit London

Now in its 3rd year, the AI in Capital Markets Summit returns with a focus on the practicalities of onboarding AI enterprise wide for business value creation. Whilst AI offers huge potential to revolutionise capital markets operations many are struggling to move beyond pilot phase to generate substantial value from AI.

GUIDE

Data Lineage Handbook

Data lineage has become a critical concern for data managers in capital markets as it is key to both regulatory compliance and business opportunity. The regulatory requirement for data lineage kicked in with BCBS 239 in 2016 and has since been extended to many other regulations that oblige firms to provide transparency and a data...