About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Talking Reference Data with Andrew Delaney: Feeling the Collars of the Libor Traders

Subscribe to our newsletter

Financial institutions are bracing themselves for further regulatory clampdown on market pricing on the wake of the LIBOR and FX market manipulation scandals. As regulators draft new guidelines for benchmark data contributions, and UK Chancellor George Osborne signals increased scrutiny of the London FX market’s pricing procedures, data managers and compliance officers are assessing their processes to institute best practice – and guard against the risk of further financial penalties.

So far the price has been high, with recent fines relating to LIBOR alone exceeding $7 billion and the Financial Times predicting an eventual total of as much as $22 billion!

In response to the ensuing public outrage, regulators are starting to act. Several have prescribed new measures to safeguard a repetition, and many institutions are assessing what they should be doing to ensure their contributions are accurate, timely and honest.

But what is best practice? And what can firms do to secure control over the prices and rates their traders – and increasingly, pricing engines and chat lines – distribute to the marketplace?

As you know these questions have been vexing me for quite some time, not least because we’ve been noticing a number of prospective industry solutions emerging that make use of monitoring technologies to keep tabs on the proprietary pricing data that makes its way to the markets in various ways.

We’ve investigated possibilities in a number of ways, the most recent of which has been our preparation for next week’s webinar on the topic, ‘Bracing for the Wave—or Sailing Ahead of It?’; Reducing Risk Through Benchmark Data Controls. Featuring speakers from Verint, BT and Z/Yen, this webinar looks at how firms can do more to monitor their traders’ voice, data and chat activities and institute controls to ensure compliance with internal policy and new regulatory guidelines. Our panel of experts will discuss approaches to instituting robust and consistent monitoring across all pricing channels, and offer practical advice on technology solutions and organisational methodologies to ensure best practice.

It’s a topic I plan to explore more at our Data Management and Intelligent Trading Summits in London and New York later in the year. In the meantime, however, be sure to join us at 3pm London time (10am New York), next Wednesday July 2. You can register free of charge here.

Relatedly, we just completed a white paper with infrastructure monitoring specialist ITRS that looks more broadly at firms’ data contributions to industry benchmarks, including indices, and fragmented price-driven markets, like Swap Execution Facilities (SEFs).

Changing market dynamics are placing new emphasis on market-makers and other generators of proprietary contributed market pricing to ensure the integrity of their data. Structural shifts like the introduction of exchange-like (SEFs) are exacerbating the pressures on financial institutions to get their pricing systems in good shape.

Concerns about data integrity are impacting a wide range of asset classes, including over-the-counter markets – like fixed income, foreign exchange and derivatives – and exchange-listed so-called structured instruments – like exchange-traded funds (ETFs), warrants, turbo-warrants, certificates and indices.

Meanwhile, as firms seek to monetise the data they generate from their business activities, new demands for quality control are driving a renewed look at data integrity. With data sales emerging as a significant element of the overall revenue picture, financial institutions are striving to adopt best practices in response to customer demand for higher- quality data services.

Our new paper – ‘Ensuring Integrity of Proprietary & Derived Market Information’ – discusses the obstacles to establishing robust pricing processes, and examines the market dynamics driving efforts towards improved data integrity and where best practices are evolving. You can download it free of charge here.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: The Role of Data Fabric and Data Mesh in Modern Trading Infrastructures

The demands on trading infrastructure are intensifying. Increasing data volumes, the necessity for real-time processing, and stringent regulatory requirements are exposing the limitations of legacy data architectures. In response, firms are re-evaluating their data strategies to improve agility, scalability, and governance. Two architectural models central to this conversation are Data Fabric and Data Mesh. This...

BLOG

A-Team Group Announces Winners of the 2025 TradingTech Insight Awards USA

A-Team Group is delighted to announce the winners of the 2025 TradingTech Insight Awards USA, recognising the leading providers of trading technology, infrastructure, and consultancy services for capital markets across North America. This year’s awards highlight the technology providers driving innovation and performance across the trading lifecycle, from market connectivity and execution management to analytics,...

EVENT

Eagle Alpha Alternative Data Conference, Spring, New York, hosted by A-Team Group

Now in its 8th year, the Eagle Alpha Alternative Data Conference managed by A-Team Group, is the premier content forum and networking event for investment firms and hedge funds.

GUIDE

Regulatory Data Handbook 2025 – Thirteenth Edition

Welcome to the thirteenth edition of A-Team Group’s Regulatory Data Handbook, a unique and practical guide to capital markets regulation, regulatory change, and the data and data management requirements of compliance across Europe, the UK, US and Asia-Pacific. This year’s edition lands at a moment of accelerating regulatory divergence and intensifying data focused supervision. Inside,...