About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Greater Importance Being Placed on Assumptive Data for Valuations, Agree DMRAV Panellists

Subscribe to our newsletter

Increased importance is being placed on providing the assumptive data inputs that are going into pricing and valuations calculations, according to panellists at last month’s A-Team Group Data Management for Risk, Analytics and Valuations (DMRAV) conference in NYC. This data is required as a result of changes to accounting rules related to fair value by both the Financial Accounting Standards Board (FASB) and the International Accounting Standards Board (IASB) in the post-crisis landscape.

Last month, the two bodies finally agreed on a global definition for “fair value” and firms will now be compelled to disclose more about their Level 3 assets, which comprise hard to value, illiquid or riskier assets that require evaluated pricing methodologies rather than market prices. Firms will therefore have to disclose more about the models, processes and assumptions that have gone into the fair valuation methodology used to price these assets.

Firms will also have to disclose whether any of the securities they value have moved from the Level 1 bracket to Level 2, or vice versa. This is to ensure the regulatory community is apprised of these market dynamics and any potential systemic risks in the making are flagged.

These developments were noted by DMRAV panellists and John Lynch, global head of pricing at Alliance Bernstein (who previously worked for Bloomberg), said: “The more assumptive data you have to hand, the better armed you are to be able to meet client and auditor questions about pricing and valuations. We are involved in increasingly academic conversations about this data.”

Firms such as Alliance Bernstein are therefore putting pressure on their vendor partners to provide a higher level of granular data about the inputs into and the methodologies used to determine an evaluated price. Lynch highlighted the importance of this data, given that pricing is critical for trading opportunities and feeds mission critical risk systems. Transparency into these prices is therefore paramount, Lynch explained: “It all comes back to the vendors being able to provide information about the colour that has been added to the pricing data and how the price relates to those for other securities.”

SIX Telekurs’ head of evaluated pricing research and development Perry Beaumont indicated that his firm has seen a lot more of a focus on the Level 1, 2 and 3 data inputs from firms in order to enable clear differentiation between these categories. Brian Buzzelli, head of pricing and reference data for the Americas region at rival vendor Thomson Reuters, noted a particularly strong push to this end in the fixed income and derivatives space, given the level of regulatory scrutiny being directed at these markets, which has resulted in the need for more “credible ingredients” to go into the pricing mix via “higher dimensions of transparency”.

Mark Abramowitz, director of US taxables at S&P Securities Evaluations, added that the desire to “kick the tyres” of the pricing models has significantly increased over recent years.

The increased pressure for timeliness with regards to data was also noted by vendor panellists, all of who indicated that they have received more requests for intraday pricing recently than ever before. Buzzelli added, however, that a balance between timeliness and accuracy of data needs to be reached in order to ensure the reliability of pricing data.

This focus on data quality and reliability was, in fact, a recurring theme throughout the conference and many other speakers highlighted the new industry imperative to prioritise data management concerns.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Best practice approaches to integrating legacy data with the cloud

Acceleration of cloud adoption, increasing demand for digital transformation and real-time data management have led financial institutions to rethink their data infrastructure to enable more agile operating models that can respond faster to change and make data a competitive advantage. For many, integrating data from legacy systems and data across the business landscape with a...

BLOG

FinScan Combines Data Quality Experience with Technological Expertise to Deliver Agile AML Solution

FinScan has combined experience in data quality with technological expertise in screening to provide an Anti-Money Laundering (AML) solution designed to help financial institutions develop more efficient AML programmes that generate fewer false positives and support better detection of true alerts. The company is part of Pittsburgh-based Innovative Systems, which was founded in 1968 and...

EVENT

Data Management Summit London

Now in its 14th year, the Data Management Summit (DMS) in London brings together the European capital markets enterprise data management community, to explore how data strategy is evolving to drive business outcomes and speed to market in changing times.

GUIDE

Regulatory Data Handbook 2023 – Eleventh Edition

Welcome to the eleventh edition of A-Team Group’s Regulatory Data Handbook, a popular publication that covers new regulations in capital markets, tracks regulatory change, and provides advice on the data, data management and implementation requirements of more than 30 regulations across UK, European, US and Asia-Pacific capital markets. This edition of the handbook includes new...