About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Improving Operational Efficiency in the Data Management Process

Subscribe to our newsletter

By: Hugo Boer, Senior Product Manager, Asset Control

As the financial services industry starts to harness a raft of new data sources for fast, effective and usable insights, the bottleneck for financial institutions becomes how well they really understand their data management processes.

How many firms, for example, can answer the following questions: Do we understand the impact of bad data quality? Can we measure this quality, and do we have full oversight over steps in the data management process? Can we pre-empt data issues? When data issues arise, can we take restorative action quickly and track adjustments along the way without losing oversight of those changes?

If your firm can’t answer all these questions, the onus is on developing transparent, end-to-end financial data management that enables real-time insight into daily data sourcing, mastering and distribution processes, improves workflows and increases operational efficiency. This will allow firms to unlock the value from new data sources.

Data scrutiny

New regulatory drivers and business pressures have led to increased scrutiny on the data management process. For example, the European Central Bank’s Targeted Review of Internal Models (TRIM) was introduced with the aim of assessing whether internal model results to calculate risk-weighted assets were reliable and comparable. The TRIM guide contained a specific Data Quality Framework focusing on data accuracy, consistency, completeness, validity, availability and traceability as a precondition for these models.

This regulatory focus is, however, just one aspect of the growing recognition among financial institutions of the need to improve insight into data management processes. There is huge business pressure on data management teams not only to manage increasing numbers of data sources, but also to deliver accurate and consistent datasets in ever decreasing time windows.

Despite overlap between the data used by different departments, many teams are still operating in functional silos, from finance to risk. In an increasingly joined up and overlapping corporate data environment, these dispersed data management activities are inherently inefficient, from parallel data sourcing teams buying the same data multiple times to expending the same effort on data preparation. The result is not only high data sourcing and preparation costs, but also unnecessary data storage and, critically, unacceptable operational risk.

Transparent process

What is required is a single overview of the data management process; the ability to track data collection and verification progress, and gain rapid insight into any problems that could affect delivery of Service Level Agreements (SLAs). While companies have attempted to deliver point oversight via existing management information tools, they have failed to provide an intuitive single view over the entire data management process across the business. What data management teams require is transparency across the diverse data silos and deliveries to data consumers and insight into the status of every process of data sourcing, cleansing and verification through to delivery to downstream systems. Essentially, data management teams need a single view into the health of corporate data.

The implications of enhanced data transparency are significant. In addition to meeting the regulatory requirements associated with increased data scrutiny, including data quality, visibility and completeness, with a single view of the entire data management process, organisations can begin to drive significant operational change and create a culture of continuous data improvement.

For example, a complete perspective of any overlap in data resources will enable streamlining of data acquisition, reducing both purchase costs as well as data cleansing and delivery costs. It will also overcome risks associated with a lack of data understanding between different areas, which can create significant federation issues that can affect both operational performance and regulatory compliance. Simple steps such as calibrating consistently applied rules for datasets or asset classes, and ensuring changes to data cleansing rules are documented, will further reinforce the value of acquired data to the business.

Extended data understanding

Transparency into the status of data sourcing, processing and delivery should not be limited to data management experts; transparency of the data supply chain should be shared with everyone in the company, providing end users with insight into the quality of the data used for risk, finance, post-trade-reporting and so on. Data confidence is a fundamental requirement in post financial crisis trading and providing end users with access to a simplified view of the data acquisition, cleansing and provisioning process for each data source will play a key role in fostering a common, companywide understanding of the data and how it is used.

For example, showing users that Bloomberg data is used as the primary source for US corporate bonds, Thomson Reuters data for foreign exchange and Six Financial data for corporate actions; capturing comments from data analysts when this hierarchy is changed; noting what data cleansing rules have been used; and when manual intervention took place can all be valuable information. This transparency will support better data knowledge and confidence and can also overcome some of the data misalignment that has evolved over the past couple of decades.

With better understanding of the end-to-end process for each data source, firms can begin to spot trends in the relative quality of different sources per market and asset class. Are there repeat errors in a data source? Is there an alternative data source already being used somewhere else in the business? Or is it time to onboard a new provider? End-to-end data management visibility will enable firms to drive a culture of continual improvement, addressing data quality issues and seeking out the most effective data sources for the business.

Conclusion

The total cost associated with end-to-end data management is becoming far more apparent, especially given the growing overlap in data usage across the business and the rise in new data sources available. Add in the escalating regulatory expectations for robust processes and the operational risk associated with siloed data management teams and the implications of a lack of transparency become very apparent.

To maximise the value of new data sources, financial institutions need to evolve from departmental data silos and achieve end-to-end to transparency of the data management process. Further, while this will significantly improve the data management operation, it is also essential to push data responsibility and knowledge to the end users. Data quality is a business issue and providing data transparency to business teams will be key in creating a strong culture of continuous improvement and leveraging feedback to drive up data quality and confidence across the organisation.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Trade South Africa: Considerations for Connecting to and Trading the Johannesburg Markets

Interest among the international institutional community in trading South African markets is on the rise. With connectivity, data and analytics options for trading on the Johannesburg Stock Exchange growing more sophisticated, and the emergence of A2X as a credible alternative equity market, South Africa is shaping up as a financial centre that can offer a...

BLOG

FCA Criticism of Funds’ SDR Approach Stirs Controversy

The UK regulator has criticised fund managers for failing to abide by interim guidance on ESG disclosures just days before it’s expected to announce details of a formal regulation. But the comments have been met with frustration and caution within the ESG data and technology community, with one practitioner arguing that the Financial Conduct Authority...

EVENT

Buy AND Build: The Future of Capital Markets Technology, London

Buy AND Build: The Future of Capital Markets Technology London on September 19th at Marriott Hotel Canary Wharf London examines the latest changes and innovations in trading technology and explores how technology is being deployed to create an edge in sell side and buy side capital markets financial institutions.

GUIDE

Regulatory Data Handbook 2023 – Eleventh Edition

Welcome to the eleventh edition of A-Team Group’s Regulatory Data Handbook, a popular publication that covers new regulations in capital markets, tracks regulatory change, and provides advice on the data, data management and implementation requirements of more than 30 regulations across UK, European, US and Asia-Pacific capital markets. This edition of the handbook includes new...