I’ve been struck in recent months by the changing attitudes within our marketplace to the concept of a golden copy, or single, accepted source of the truth as far as reference data is concerned. The industry seems to be moving away from established principles of on-site enterprise data management, most visibly – noisily, perhaps? – toward the utility model of outsourced, mutualised reference data management.
But there seems to be another, less tangible theme emerging as firms seek that Holy Grail. This occurred to me as I chatted this week with friends at 1View, whose Paul Kennedy recently appeared on our Hot Topic webinar on utility approaches to data management. 1View is one of several companies I’ve come across in my travels that seem to be taking a different approach to golden copy, one that doesn’t require major infrastructure investment but instead focuses on control of existing data sources and a fair degree of cross-referencing.
1View’s approach is simple – which can be dangerous. In short, it allows clients to upload data extracts from their key systems to 1View, which then ranks them into an aggregated superset. All exceptions are flagged, and mappings made to deal with data formats and other discrepancies. The result, according to 1View chief Nigel Pickering, is that all common data across source systems is synchronised over time, without the need for major infrastructure or process change.
1View’s approach seems to resonate with those of other suppliers who appear to be taking a bottom-up rather than top-down approach to data standardisation or synchronisation. Thomson Reuters’ response to the EU ruling on its RIC code, the efforts of companies like Simplified Financial Information, and Skyler with its LiveCache platform all spring to mind.
No, they’re not addressing the exact same problem, but they are taking a similar approach by attacking data control from the ground up. Putting aside Thomson Reuters for now, SFI has teamed with performance measurement specialist TS Associates to monitor where data is being sourced and used across the financial enterprise, and Skyler’s LiveCache can be used to create a kind of Universal Bank ID that clients can use to access data pertaining to a given instrument or entity from any internal or external source used by the bank.
You may think the link here is tenuous, and I’ll concede that my thinking on this far fully baked. I do think, however, that the topic warrants further exploration, and may make for a good panel discussion at one of our future Data Management Summits or even a webinar. What do you reckon?