About a-team Marketing Services
The knowledge platform for the financial technology industry
The knowledge platform for the financial technology industry

A-Team Insight Blogs

Wachovia Expands Use of AC Plus To Boost Risk Data Quality

Subscribe to our newsletter

Wachovia Corp. has once again expanded its use of Asset Control’s AC Plus data management platform to boost the capabilities of its internal risk management function. Wachovia originally implemented AC Plus in its risk management operation in 2004 (Reference Data Review, January 2005) and subsequently extended its use to other areas of the enterprise as part of a three-year project (Reference Data Review, September 2005).

The latest expansion involves the addition of sources of market data used to support Wachovia’s risk management systems. The bank has added several undisclosed “complex data feeds” that it says will help “improve the quality of market data utilized within risk management” as well as offering that data throughout the bank.

Wachovia’s risk solution makes use of snapshot, end-of-day and time-series pricing information for interest rates, credit spreads, equities, FX and commodities, gathered and managed by the AC Plus platform. Additionally, AC Plus consolidates and validates data from external sources, including Reuters, Bloomberg, FT Interactive Data and Markit Group, to provide consistency and reliability.

Wachovia is making use of Asset Control’s range of four-dimensional graphing capabilities. This will allow the bank to analyze and survey data anomalies and trends over time.

Speaking at the ISIPS conference in London this month, Martijn Groot, head of product management at Asset Control, outlined how Asset Control’s audit and backtracking functions allow clients to standardize and consolidate disparate and often-conflicting price data from multiple sources into a single consolidated price that can be published to internal application.
The process involves applying client-defined business rules to incoming and internal data sources. These rules reflect the client’s approach to data management, and may range in complexity from a sophisticated algorithm to a simple average in order to arrive at a figure that the institution is comfortable with.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: Unpacking Stablecoin Challenges for Financial Institutions

The stablecoin market is experiencing unprecedented growth, driven by emerging regulatory clarity, technological maturity, and rising global demand for a faster, more secure financial infrastructure. But with opportunity comes complexity, and a host of challenges that financial institutions need to address before they can unlock the promise of a more streamlined financial transaction ecosystem. These...

BLOG

Reconciliation and the Silent Revolution Reshaping Financial Operations

By Sarva Srinivasan, head of global strategy and managing director at NeoXam, Americas. In most financial institutions, reconciliation has traditionally lived quietly in the background. It is often viewed as a necessary control process that ensures transactions, positions and balances match across systems and counterparties. Important, yes, but rarely considered fundamental to the business. But...

EVENT

AI in Data Management Summit New York City

Following the success of the 15th Data Management Summit NYC, A-Team Group are excited to announce our new event: AI in Data Management Summit NYC!

GUIDE

Solvency II Data Management Handbook

Want to get a handle on Solvency II and what it means for data management? Need to make sure you have all the bases covered for the looming January 2016 deadline? Our Solvency II Data Management Handbook is now available for free download to help you. This Handbook is the ultimate guide to all things...