About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Datawatch Adds Panopticon Streams Real-Time Stream Processing Engine

Subscribe to our newsletter

Datawatch has increased the speed of real-time streaming and time series data analytics with stream processing engine Panopticon Streams. The engine can be used as a stand-alone solution or in conjunction with Panopticon’s Visual Analytics platform.

Peter Simpson, vice president of visualisation strategy at Datawatch Panopticon, says: “Capital markets customers will benefit from Panopticon Streams’ support of several key use cases, including best execution, real-time P&L, transaction cost analysis and trader and trading surveillance.

“The addition of the engine’s capabilities means we now offer a general purpose streaming analytics platform. It has applications anywhere organisations need to identify anomalies and outliers, investigate their causes, back test potential solutions, and then alter their business processes to address the issue. Given the software’s ability to handle real-time and time series data, we believe it will be most useful in electronic trading, telecommunications, energy, and IoT applications.”

The combination of stream processing, rapid data comprehension through visual analysis, faster investigation through time series analysis and playback down to the individual tick, is designed to help organisations make timely, more informed decisions that have immediate financial impact.

Built on the Apache Kafka platform, Panopticon’s solutions enable business users to build sophisticated Kafka data flows with no coding. Users who understand the business problems can create their own data flows, which can use information from a number of sources and incorporate joins, aggregations, conflations, calculations, unions and mergers, and alerts. Analysts can visualise processed data using Panopticon Visual Analytics and deliver it to Kafka, Kx kdb+, InfluxDb, or any SqL database.

Subscribe to our newsletter

Related content

WEBINAR

Upcoming Webinar: How to move to a modern, component based trading architecture using a Buy AND Build approach

Date: 7 May 2026 Time: 10:00am ET / 3:00pm London / 4:00pm CET Duration: 50 minutes To remain competitive in today’s electronic markets, firms need trading architectures that support rapid innovation, effortless integration of new capabilities, and the agility to respond to shifting market demands. This is prompting technology leaders to move beyond the traditional...

BLOG

LSEG and Bank of America Target AI-ready, Governed Data Integration in Multi-Year Partnership

London Stock Exchange Group (LSEG) and Bank of America have agreed a multi-year strategic partnership centred on embedding governed, AI-ready data and analytics directly into the bank’s core workflows. Rather than a distribution agreement focused on access, the collaboration reflects a broader architectural shift: integrating unified, rights-cleared content, analytics and risk intelligence across advisory, trading,...

EVENT

Data Management Summit New York City

Now in its 15th year the Data Management Summit NYC brings together the North American data management community to explore how data strategy is evolving to drive business outcomes and speed to market in changing times.

GUIDE

The Data Management Implications of Solvency II

This special report accompanies a webinar we held on the popular topic of The Data Management Implications of Solvency II, discussing the data implications for asset managers and their custodians and asset servicers. You can register here to get immediate access to the Special Report.