About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

ICFR Meeting Calls for Increased Consistency of Supervisory Data Collection, Global Coordination

Subscribe to our newsletter

The International Centre for Financial Regulation’s (ICFR) recent meeting on macroprudential data scoping highlighted a number of concerns related to the collection of data for the monitoring of systemic risk, not least of which was the need for global coordination among regulators. Attendees to the meeting, which took place earlier this month in London under Chatham House rules, noted there is a strong argument for globally consistent data reporting formats, standard data definitions and a more established frequency of reporting in order to ease the process for financial institutions in terms of “data programming and compliance”.

The ICFR meeting attendees also noted that choices on the data required for systemic risk monitoring should be made using a risk-based methodology. This is in order to ensure that the most important data series are captured first, and the cost of collection can be justified. Some participants said it might be useful to implement a common accounting or reporting language across the regulators, although this would take an unprecedented degree of coordination. Whereas others felt that a ‘one in, one out’ rule should be used when it comes to data collection – that is, once a new method of data collection is adopted, the old one should be abandoned. The process of data collection should be more harmonised, argued these attendees, who stressed the need for comparability among different data items and across regulatory borders. This is essential in the ability to track risks related to systemically important financial institutions that operate across the globe.

To this end, regulators should use existing and new data collection bodies in order to gather these data items, such as newly established central counterparties (CCPs) within the OTC markets. The focus of the meeting was also on determining the correct level of data disclosure to adopt, taking into account privacy concerns and data protection considerations. It was agreed, for instance, that as increased data granularity is becoming more important to the regulatory community in the face of the increased complexity of the financial markets. Although some regulators advocate maximum granularity, some attendees indicated that this might create distortions in the market, since close monitoring of one financial instrument incentivises the banking industry to design a ‘new instrument’ which will not be captured by the data. On the subject of disclosure, attendees notes that a project on the optimal disclosure of information on the efficiency of price setting would be useful, as there is little academic work on this topic. For example to determine how much information can be put in the market before the price mechanism no longer functions.

The issue of the frequency of data collection was also addressed, with the general stance being that the time span of each data collection should be reduced to days or weeks, rather than quarters or years. In addition, it was proposed that clearing and registration of derivatives would naturally increase standardisation and, hence, ease supervision. All of this feedback reflects some of the discussions going on within the regulatory community with regards to the introduction of more standard data formats and identifiers in the market, for example, the US Office of Financial Research’s (OFR) proposals on legal entity identification. Hopefully, participants at the meeting will be incentivised to join the wider debate on the subject of data standards.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Best practices for creating an effective data quality control framework

Data quality is critical to capital markets processes from identifying counterparties to building customer relationships, regulatory reporting, and ultimately improving the bottom line. It can also be extremely difficult to achieve. One solution is a data quality control framework that includes an automated and systematic process that monitors the state of data quality and ensures...

BLOG

ESMA Recommends Digital Token Identifier for Pilot of Distributed Ledger Technology Regulation

ESMA has recommended use of the ISO standard Digital Token Identifier (DTI) for the pilot of its Distributed Ledger Technology (DLT) regulatory regimen. DTIs are issued by the DTI Foundation, a non-profit division of Etrading Software, with a view to bringing standardisation to digital asset and crypto markets. The DLT pilot will apply from 23...

EVENT

TradingTech Briefing New York

TradingTech Insight Briefing New York will explore how trading firms are innovating and leveraging technology as a differentiator in today’s cloud and digital based environment.

GUIDE

Regulatory Data Handbook 2022/2023 – Tenth Edition

Welcome to the tenth edition of A-Team Group’s Regulatory Data Handbook, a publication that has tracked new regulations, amendments, implementation and data management requirements as regulatory change has impacted global capital markets participants over the past 10 years. This edition of the handbook includes new regulations and highlights some of the major regulatory interventions challenging...