The leading knowledge platform for the financial technology industry
The leading knowledge platform for the financial technology industry

A-Team Insight Blogs

LIBOR: Could We Have Seen it Coming?

The repercussions of the LIBOR price fixing scandal are likely to run and run.

We now know that from around 2005 onwards a number of banks were attempting to manipulate rates. Over $6 billion has already been levied in fines, and with $300 trillion-worth of contracts pegged to this all-important benchmark, the figure could well run to $20-30 billion over coming years as those affected recalculate their damages on the basis of what LIBOR should have been at the time.

Both the regulators and banks have been slow to react, but it is now clear that the wave of LIBOR-inspired compliance requirements that have resulted from the scandal have particular implications for capital markets.

Participants are concerned that they will be ‘unfairly’ penalised by regulation, which will cause them to lose competitive traction – and they are struggling to rationalise compliance processes and platforms over many disparate geographies and wildly different jurisdictions.

What is also clear is that at the time of the price fixing scandal, off the shelf automated surveillance systems already existed that could have picked up anomalous trading patterns and alerted compliance officers at the time to what was happening.

“Somebody in compliance should have been running some of these basic statistical predictive analytics tests on what is clearly an intrinsically important benchmark in a major market area,” says Professor Michael Mainelli of Z/Yen. “I would contend that the industry has a lot to learn, and to start learning you have to admit what you don’t know. We clearly don’t seem to know what’s going on out on the trading floors in even a basic sense and certainly we’re not measuring.”

CFOs and corporate treasurers are now waking up to this fact and are anxious to stop history repeating itself. Proactive management of a firm’s risk and compliance environments is desperately needed to stay ahead of the compliance wave.

But what action can and should they take? The new regulations call for ongoing proactive monitoring, but voice remains central to much trading floor activity, and because these data records are unstructured they need to be analysed once captured.

Existing voice surveillance systems used by banks tend to be manual and therefore extremely resource intensive, expensive and time-consuming. It can take four hours to listen to one hour of voice content, says Verint’s Robert Simpson; fewer than 15 per cent of people involved in the process can be monitored at any one time – so this is clearly not a scalable process.

Speech analytics technology is capable of delivering up to 75 per cent accuracy. But banks have been hesitant to adopt it, claiming that the technology is bleeding edge, hence unproven, and unsuitable for wholesale banking environments where there is so much background noise.

Tim Furmidge, Head of Product Management, Financial Technology Services, BT, illustrates how such systems can be introduced in a staged way in this thought provoking video webinar that demonstrates how proactive management of firms’ risk and compliance environments can not only help them stay ahead of the regulatory wave, but also deliver significant productivity gains.


This video webinar was moderated by Andrew Delaney, Editor-in-Chief of the A-Team Group and featured Tim Furmidge, Head of Product Management, Financial Technology Services, BT; Robert Simpson, Vice President, Global Financial Compliance Practice, Verint and Professor Michael Mainelli FCCA FCSI FBCS, Executive Chairman, Z/Yen Group.

Related content

WEBINAR

Upcoming Webinar: Evolution of data management for the buy-side 2021

Date: 27 May 2021 Time: 10:00am ET / 3:00pm London / 4:00pm CET Duration: 50 minutes The buy-side faced a barrage of regulation in 2020 and is now under pressure to make post-Brexit adjustments and complete LIBOR transition by the end of 2021. To ensure compliance and ease the burden of in-house data management, many...

BLOG

n-Tier Blockchain-Based Reference Data Consensus Solution Aims to Drive Down Errors and Costs

As reference data volumes continue to soar, bringing with them huge data cleansing, validation and management costs, financial institutions are beginning to consider collaborative solutions that can improve data accuracy while reducing cost. n-Tier, a New York headquartered company that helps firms ensure accuracy and completeness of reference data, has joined the party with a...

EVENT

RegTech Summit Virtual

The RegTech Summit Virtual is a global online event that will be held in June 2021 with an exceptional guest speaker line up of RegTech practitioners, regulators, start-ups and solution providers to collaborate and discuss innovative and effective approaches for building a better regulatory environment.

GUIDE

Entity Data Management Handbook – Seventh Edition

Sourcing entity data and ensuring efficient and effective entity data management is a challenge for many financial institutions as volumes of data rise, more regulations require entity data in reporting, and the fight again financial crime is escalated by bad actors using increasingly sophisticated techniques to attack processes and systems. That said, based on best...