About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Taylor Discusses Why LCH.Clearnet Has Invested in its Data Architecture

Subscribe to our newsletter

UK-based clearer LCH.Clearnet is mid-way through a significant project to revamp its internal data architecture, according to Martin Taylor, group chief information officer at the clearer. Taylor told attendees to last week’s Thomson Reuters organised data management roundtable that the clearer is six months into a project that will deliver a more standardised approach to data management via the move to a master data management (MDM) structure and the implementation of a new data warehouse.

Taylor discussed the impact that the Lehman default had on the clearing house in September 2008, at a time when it was in the throes of a significant platform upgrade within its energy derivatives business. The events around the largest default it had ever handled forced the clearing house to re-evaluate its approach to data management, he explained. “It provided us with a deeper understanding of the role of clearing in the market and highlighted that we need earlier access to data,” he continued.

One of the biggest problems Lehman’s collapse, caused immediately after the event, was the lack of available staff that understood where the data was kept in order to be able to unwind the investment bank in an orderly manner. Ordinarily there are people around to assist in the unwinding process by extracting vital counterparty data from the relevant systems, but in Lehman’s case, everyone had literally left the building, he explained. Taylor stressed that the upcoming regulation around living wills would further highlight these issues to the benefit of data management overall.

For its part, LCH.Clearnet is improving the security, integrity and auditability of its internal data management systems. The data management function actually sits within the clearer’s risk management department and thus also puts emphasis on the speed and accuracy of the data for risk management purposes.

Taylor emphasised that the people involved in the data management process are as important as the systems. He highlighted the public and government concerns about data security as an indicator of how important data security has become as an issue for the industry as a whole. “We need to make sure that we understand who is handling the data and that there is an audit trail and checks and balances in place to regulate this,” he said.

To this end, he feels a single group should be put in charge of managing data across an enterprise, but noted: “the bigger the enterprise, the harder this is to achieve”.

He concluded by suggesting, like a number of the other panellists, that the financial services industry could learn data standardisation tips from other industries such as pharmaceuticals. “They are much more ruthless in their application of data management and we don’t need to reinvent the wheel. After all, they need to be able to track down components very quickly just in case one of them kills somebody,” he said.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Unpacking Stablecoin Challenges for Financial Institutions

The stablecoin market is experiencing unprecedented growth, driven by emerging regulatory clarity, technological maturity, and rising global demand for a faster, more secure financial infrastructure. But with opportunity comes complexity, and a host of challenges that financial institutions need to address before they can unlock the promise of a more streamlined financial transaction ecosystem. These...

BLOG

S&P Global Data via Cloud: Unlocking Real-Time, Scalable Insights with Snowflake and Databricks Delta Sharing

As organisations accelerate their cloud migration strategies to manage growing volumes of structured and unstructured data, demand is rising for secure, real-time, cloud-native access to trusted datasets. Leveraging Snowflake and Databricks Delta Sharing, S&P Global provides a scalable, agile foundation that allows organizations to directly access and query S&P Global and curated third-party datasets without...

EVENT

TradingTech Summit New York

Our TradingTech Summit in New York is aimed at senior-level decision makers in trading technology, electronic execution, trading architecture and offers a day packed with insight from practitioners and from innovative suppliers happy to share their experiences in dealing with the enterprise challenges facing our marketplace.

GUIDE

Regulatory Data Handbook 2020/2021 – Eighth Edition

This eighth edition of A-Team Group’s Regulatory Data Handbook is a ‘must-have’ for capital markets participants during this period of unprecedented change. Available free of charge, it profiles every regulation that impacts capital markets data management practices giving you: A detailed overview of each regulation with key dates, data and data management implications, links to...