The leading knowledge platform for the financial technology industry
The leading knowledge platform for the financial technology industry

A-Team Insight Blogs

Know Your Partner – Cicada on Counterparty Data

Share article

Following on from the risk data work it had conducted with clients, including a global insurance company, Cicada’s strategic alliance with risk management vendor Algorithmics – signed last spring (Reference Data Review, May 2003) – marked a milestone in the company’s foray into the world of risk. While Algo brought to the table its suite of risk management systems, Cicada was to focus on data integrity, cleansing data where necessary and facilitating the integration of internal and external data sources.

As it began work with clients, Cicada found that it was increasingly focusing on the quality – or lack of it – of partner data: the descriptive information relating to a firm’s counterparties and potential counterparties. Firms, it seems, have a poor grasp of precisely whom they are trading with, and this can be a major obstacle to the secure settlement of transactions, either electronically or otherwise.

According to Richard Bennett, vice president, Europe, for Cicada – and former Algorithmics executive – partner data involves establishing an understanding of the legal entity structure of the firms an institution does business with.

That, he says, involves looking at where processes reside within an organization, where perhaps certain trades are made in London but settled in Hong Kong, for example. This also includes linkages between counterparties and counterparty structures. “The idea is to get a clear view, from a risk perspective, of the exposure you have,” he says. The results have repercussions for many areas of the business, including client account management, risk management, trading and back office, and client relationship management.

This focus on knowing your counterparty has been driven by a number of factors. A major one has been the collapse of Enron. A large number of financial institutions had direct exposure to Enron. But when, wanting to get a picture of their overall exposure to the bankrupt energy conglomerate, they dug a little deeper, their financial controllers found much broader exposure, resulting from Enron’s use of SPVs (Special Purpose Vehicles) and the issue of certain securities by different subsidiaries.

Another driver of the need for a process to manage partner data is the growing complexity of the business. The corporate landscape is constantly changing, making it difficult – and even more essential – to keep on top of whom one is dealing with.

Finally, a raft of new regulatory measures has raised the bar in terms of requirements for institutions, among them Sarbanes-Oxley, the Patriot Act and, perhaps most importantly in the risk area, Basel II.

As with other areas of reference data, Cicada prefers to think of partner data management as a process, rather than embracing the concept of the Golden Copy as an end-all. “It’s difficult to get one copy of the data to fit all downstream applications,” says Bennett. Moreover, there is no single source and, once collected, the data doesn’t sit still. So, while some of the partner data requirement may be bought in from outside – from business information suppliers such as Standard & Poor’s or Dun & Bradstreet – much of it will be generated internally.

The difficulty of putting the two together and managing the result on an ongoing basis is good news for Cicada. “Each firm, in practice, is completely different,” says Bennett, “on an overall data quality management basis. We have been doing this ourselves for some time and thought we could get more leverage through the relationship with Algo. Clients want a combination of process, cleansed data and analytics.”

This generates two indicators – loss-given defaults and probability of default – that need to factored into a firm’s overall credit view, a crucial requirement of Basel II. This view is essentially a picture of a firm’s risk of operations, the risk it faces each time it opens its doors in the morning.

Cicada and Algo are now offering both risk data and managed risk solutions, or a combination of the two, initially to Algorithmics’ existing client base. “Implementing any risk solution is long, laborious and expensive, with a low level of certainty of outcome,” says Bennett. “One of the objectives is to give a certainty of outcome, and to do that in a way that is fully integrated into the Algo suite means a reduced set-up time.” Bennett expects overall return for clients to improve as the two partners gain more expertise through market experiences.

In partnering with Algo, Cicada was looking for a risk vendor with the “right footprint” – both geographically and with respect to user types. Algo has clients both sides of the Atlantic and beyond, primarily in the banking and brokerage businesses. This presence has helped Cicada differentiate itself from others offering similar services in other segments, such as Barra within the buy-side community and RiskMetrics in the asset management, hedge funds and treasury areas.

So far, he says, the business is getting a warm response. Cicada and Algo pitch their services to the chief risk officers within financial institutions and their teams. Bennett says a number of bids have been made, adding that the Basel II timeframe is a substantial factor in determining the timing of such projects.

Currently, the Basel II requirements are slated for introduction by the end of 2006. This implies that firms will need to have systems in place by the end of 2005, in order to collect the required year’s worth of data in order to comply. This, in turn, suggests a decision-making period over the next 18 months. That schedule could be sent off course, however, if Basel II’s timeframe is delayed, which is to be determined next month.

Either way, sooner or later, firms will be required to apply counterparty descriptive data to their overall credit risk positions. Cicada hopes to be waiting in the wings when that happens.

Related content

WEBINAR

Upcoming Webinar: Entity identification and client lifecycle management – How financial institutions can drive $4 billion in cost savings

Date: 21 January 2021 Time: 10:00am ET / 3:00pm London / 4:00pm CET Duration: 50 minutes A new model in Legal Entity Identifier (LEI) issuance has created significant opportunities for financial institutions to capitalise on their KYC and AML due diligence. By becoming Validation Agents and obtaining LEIs on behalf of their clients, financial institutions...

BLOG

J.P. Morgan Steps Forward as First Validation Agent in Global LEI System

J.P. Morgan has become the first validation agent in the Global LEI System, signalling early interest in the scheme as a means of improving client onboarding and Know Your Customer (KYC) . The validation agent framework was introduced by GLEIF in September 2020 and is designed to help financial institutions improve customer experience, accelerate client...

EVENT

RegTech Summit Virtual

The RegTech Summit Virtual is a global online event that will be held in May 2021 with an exceptional guest speaker line up of RegTech practitioners, regulators, start-ups and solution providers to collaborate and discuss innovative and effective approaches for building a better regulatory environment.

GUIDE

The Global LEI System – A Solution for Entity Data?

The Global LEI System – or GLEIS – has been in development since the middle of last year. Development has been patchy at times, but much has been done, leaving fewer outstanding issues, but also raising new questions. What’s emerging is a structure for the GLEIS going forward, complete with a mechanism for registering and...