About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Wachovia Expands Use of AC Plus To Boost Risk Data Quality

Subscribe to our newsletter

Wachovia Corp. has once again expanded its use of Asset Control’s AC Plus data management platform to boost the capabilities of its internal risk management function. Wachovia originally implemented AC Plus in its risk management operation in 2004 (Reference Data Review, January 2005) and subsequently extended its use to other areas of the enterprise as part of a three-year project (Reference Data Review, September 2005).

The latest expansion involves the addition of sources of market data used to support Wachovia’s risk management systems. The bank has added several undisclosed “complex data feeds” that it says will help “improve the quality of market data utilized within risk management” as well as offering that data throughout the bank.

Wachovia’s risk solution makes use of snapshot, end-of-day and time-series pricing information for interest rates, credit spreads, equities, FX and commodities, gathered and managed by the AC Plus platform. Additionally, AC Plus consolidates and validates data from external sources, including Reuters, Bloomberg, FT Interactive Data and Markit Group, to provide consistency and reliability.

Wachovia is making use of Asset Control’s range of four-dimensional graphing capabilities. This will allow the bank to analyze and survey data anomalies and trends over time.

Speaking at the ISIPS conference in London this month, Martijn Groot, head of product management at Asset Control, outlined how Asset Control’s audit and backtracking functions allow clients to standardize and consolidate disparate and often-conflicting price data from multiple sources into a single consolidated price that can be published to internal application.
The process involves applying client-defined business rules to incoming and internal data sources. These rules reflect the client’s approach to data management, and may range in complexity from a sophisticated algorithm to a simple average in order to arrive at a figure that the institution is comfortable with.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Unpacking Stablecoin Challenges for Financial Institutions

The stablecoin market is experiencing unprecedented growth, driven by emerging regulatory clarity, technological maturity, and rising global demand for a faster, more secure financial infrastructure. But with opportunity comes complexity, and a host of challenges that financial institutions need to address before they can unlock the promise of a more streamlined financial transaction ecosystem. These...

BLOG

Governance to be Scrutinised at Inaugural AI in Data Management Summit NYC

Ensuring artificial intelligence deployments are securely governed without stymieing their potential is a delicate balancing act. It requires carefully drawn policies, frameworks and processes. As deployment of the technology expands and its capabilities and complexity multiply, the governance structure must adapt and evolve. How to get this right is among the most important topics swirling...

EVENT

RegTech Summit London

Now in its 9th year, the RegTech Summit in London will bring together the RegTech ecosystem to explore how the European capital markets financial industry can leverage technology to drive innovation, cut costs and support regulatory change.

GUIDE

Entity Data Management Handbook – Sixth Edition

High-profile and punitive penalties handed out to large financial institutions for non-compliance with Anti-Money Laundering (AML) and Know Your Customer (KYC) regulations have catapulted entity data management up the business agenda. So, too, have industry and government reports on the staggering sums of money laundered on a global basis. Less apparent, but equally important, are...