The leading knowledge platform for the financial technology industry
The leading knowledge platform for the financial technology industry

A-Team Insight Blogs

Talking Reference Data with Andrew Delaney: Feeling the Collars of the Libor Traders

Financial institutions are bracing themselves for further regulatory clampdown on market pricing on the wake of the LIBOR and FX market manipulation scandals. As regulators draft new guidelines for benchmark data contributions, and UK Chancellor George Osborne signals increased scrutiny of the London FX market’s pricing procedures, data managers and compliance officers are assessing their processes to institute best practice – and guard against the risk of further financial penalties.

So far the price has been high, with recent fines relating to LIBOR alone exceeding $7 billion and the Financial Times predicting an eventual total of as much as $22 billion!

In response to the ensuing public outrage, regulators are starting to act. Several have prescribed new measures to safeguard a repetition, and many institutions are assessing what they should be doing to ensure their contributions are accurate, timely and honest.

But what is best practice? And what can firms do to secure control over the prices and rates their traders – and increasingly, pricing engines and chat lines – distribute to the marketplace?

As you know these questions have been vexing me for quite some time, not least because we’ve been noticing a number of prospective industry solutions emerging that make use of monitoring technologies to keep tabs on the proprietary pricing data that makes its way to the markets in various ways.

We’ve investigated possibilities in a number of ways, the most recent of which has been our preparation for next week’s webinar on the topic, ‘Bracing for the Wave—or Sailing Ahead of It?’; Reducing Risk Through Benchmark Data Controls. Featuring speakers from Verint, BT and Z/Yen, this webinar looks at how firms can do more to monitor their traders’ voice, data and chat activities and institute controls to ensure compliance with internal policy and new regulatory guidelines. Our panel of experts will discuss approaches to instituting robust and consistent monitoring across all pricing channels, and offer practical advice on technology solutions and organisational methodologies to ensure best practice.

It’s a topic I plan to explore more at our Data Management and Intelligent Trading Summits in London and New York later in the year. In the meantime, however, be sure to join us at 3pm London time (10am New York), next Wednesday July 2. You can register free of charge here.

Relatedly, we just completed a white paper with infrastructure monitoring specialist ITRS that looks more broadly at firms’ data contributions to industry benchmarks, including indices, and fragmented price-driven markets, like Swap Execution Facilities (SEFs).

Changing market dynamics are placing new emphasis on market-makers and other generators of proprietary contributed market pricing to ensure the integrity of their data. Structural shifts like the introduction of exchange-like (SEFs) are exacerbating the pressures on financial institutions to get their pricing systems in good shape.

Concerns about data integrity are impacting a wide range of asset classes, including over-the-counter markets – like fixed income, foreign exchange and derivatives – and exchange-listed so-called structured instruments – like exchange-traded funds (ETFs), warrants, turbo-warrants, certificates and indices.

Meanwhile, as firms seek to monetise the data they generate from their business activities, new demands for quality control are driving a renewed look at data integrity. With data sales emerging as a significant element of the overall revenue picture, financial institutions are striving to adopt best practices in response to customer demand for higher- quality data services.

Our new paper – ‘Ensuring Integrity of Proprietary & Derived Market Information’ – discusses the obstacles to establishing robust pricing processes, and examines the market dynamics driving efforts towards improved data integrity and where best practices are evolving. You can download it free of charge here.

Related content

WEBINAR

Recorded Webinar: Embedding AI & Machine Learning into your trading operations

It’s no secret that AI and Machine Learning are changing the landscape of the trading industry. As these tools and techniques become more mainstream, their availability to even the smallest of financial firms is becoming a more viable proposition.  From speech recognition and natural language processing (NLP) to deep learning algorithms and predictive analytics, these...

BLOG

DSB Calls for New Technology Advisors as it Pivots to Address New Data Challenges

The Derivatives Service Bureau (DSB) has had an interesting few years. Founded by the Association of National Numbering Agencies (ANNA) to facilitate the allocation and maintenance of International Securities Identification Numbers (ISINs), Classification of Financial Instrument codes (CFIs) and Financial Instrument Short Names (FISNs), in 2019 the group was also designated by the Financial Stability...

EVENT

Data Management Summit New York City

Now in its 10th year, the Data Management Summit (DMS) in NYC explores the shift to the new world where data is redefining the operating model and firms are seeking to unlock value via data transformation projects for enterprise gain and competitive edge.

GUIDE

Regulatory Data Handbook 2020/2021 – Eighth Edition

This eighth edition of A-Team Group’s Regulatory Data Handbook is a ‘must-have’ for capital markets participants during this period of unprecedented change. Available free of charge, it profiles every regulation that impacts capital markets data management practices giving you: A detailed overview of each regulation with key dates, data and data management implications, links to...