About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Understanding Risk is Key to Effective Data Management, Says BNY M’s Cox

Subscribe to our newsletter

Institutions need to invest in automated processes and good quality staff in order to adequately assess their risk exposure within data management process, said BNY Mellon’s Matthew Cox, head of securities data management, EMEA. They also need to understand the internal risks that are posed by bad quality data and examine what vendors can and can’t provide in terms of data accuracy, he explained to the FIMA 2008 delegation.

“Identifying the source of inaccurate data is key to mitigating risk in the current environment,” he explained. “BNY Mellon does this by dual sourcing data and producing golden copy.”

All models have their issues, however, but “common sense” and stepping back from the problem rather than merely reacting to issues that arise is the best approach, he said. “We need to understand what the end client wants to use the data for and whether they pass that data onto another party,” Cox explained. “This is in order to more accurately measure client and financial risk, which must both be taken into account at the end of the day.”

When servicing multiple clients, this will have a significant impact on reputational risk, he continued. In the custody business, for example, client service is of critical importance and controls must be in-built to account for data discrepancies and increased risk exposure.

There is obviously more risk involved in manual processes and therefore automation and STP are one method of dealing with these risks, said Cox. “There is also a need for good quality people to be involved in the data management process but these people should not be put in a position where they often have to make judgement calls about data,” he explained.

He warned delegates that there is “no hiding place” from problems caused by data, as institutions cannot simply blame vendors: the institution must take the responsibility for its services. To make sure the data is correct, tolerance levels must be set and data must be regularly checked, he explained. “Checks and controls must be built around the areas that the most problems occur and where the risks are greatest. Finger on the pulse rather than single snapshots allow institutions to react in a more timely manner.”

It is also unrealistic to expect 100% integrity of data from vendors, he contended, as inaccuracies can be down to issues with the underlying source data.

BNY Mellon uses a data vendor scorecard with red, amber and green scores to measure the metrics being met (or not) by its vendors. “The facts speak for themselves in this way and we have control over our vendor relationships – we can prove that they need to improve in certain areas with hard evidence,” Cox explained.

Reprising Largier’s earlier point, Cox also discussed the benefits of working in partnership with the vendor community and producing detailed service level agreements to adequately measure performance at a basic level.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Unlocking Transparency in Private Markets: Data-Driven Strategies in Asset Management

As asset managers continue to increase their allocations in private assets, the demand for greater transparency, risk oversight, and operational efficiency is growing rapidly. Managing private markets data presents its own set of unique challenges due to a lack of transparency, disparate sources and lack of standardization. Without reliable access, your firm may face inefficiencies,...

BLOG

As Finance Sector Workers Embrace AI, Study Warns ‘Be Careful What You Wish For’

The potential real-world impacts of hastily deployed artificial intelligence rollouts have been highlighted in new reports that underscore the need for better-quality data and greater literacy in the technology. Financial firms that don’t invest in creating greater workforce awareness of how AI tools can be used are at risk not only of failing to optimise...

EVENT

RegTech Summit London

Now in its 9th year, the RegTech Summit in London will bring together the RegTech ecosystem to explore how the European capital markets financial industry can leverage technology to drive innovation, cut costs and support regulatory change.

GUIDE

BCBS 239 Data Management Handbook

Our 2015/2016 edition of the BCBS 239 Data Management Handbook has arrived! Printed copies went like hotcakes at our Data Management Summit in New York but you can download your own copy here and get access to detailed information on the  principles and implications of BCBS 239 on Data Management. This Handbook provides an at-a-glance...