
Bloomberg and Kaiko have announced a collaboration to provide access to Bloomberg’s reference and pricing data within blockchain-based tokenised market environments.
The initiative will initially support tokenised U.S. Treasuries and repo use cases on the Canton Network. Bloomberg Data License content will be delivered on-chain through Kaiko’s infrastructure, with access governed by existing licensing and entitlement controls.While framed as a tokenisation initiative, the announcement centres on a more specific challenge: how to integrate licensed institutional data into blockchain workflows without fragmenting valuation, risk and reconciliation processes.
Aligning Data Across On-Chain and Off-Chain Systems
For trading technology teams, one of the structural issues in tokenised market pilots has been the separation between execution and settlement logic on-chain and pricing and reference data sourced off-chain.
In fixed income and repo markets, valuation inputs underpin margin calculations, collateral eligibility and settlement finality. Where counterparties rely on different data feeds or timing conventions, discrepancies can emerge that require manual reconciliation or additional controls.
By enabling authorised participants to access Bloomberg data directly within a permissioned blockchain environment, the collaboration aims to reduce divergence between on-chain transaction records and the data sets used in traditional trading, risk and operations systems.
The emphasis on entitlement management indicates that blockchain distribution is being adapted to existing enterprise data governance models rather than replacing them.
Infrastructure Rather Than Marketplace
The choice of Canton highlights the institutional positioning of the initiative. Designed to support privacy-enabled, permissioned financial workflows, Canton differs from public blockchain networks in its participant controls and interoperability features.
The collaboration between Bloomberg and Canton aims to address the data layer underpinning tokenised instruments, an area that has become increasingly relevant as institutions move from proof-of-concept projects toward production-grade infrastructure.
From a trading architecture perspective, embedding licensed reference data on-chain may simplify integration between distributed ledger systems and core platforms such as order management, collateral management and risk engines. It may also support more automated smart contract logic where pricing or eligibility rules depend on authoritative data sources.
Operational and Architectural Considerations
A number of factors are likely to be central to broader adoption, including refresh frequency, latency characteristics, and synchronisation mechanisms between blockchain and traditional data distribution channels, as well as commercial models for on-chain data consumption. In markets such as Treasuries and repo, timing precision and data consistency affect margining, liquidity management and regulatory reporting. Ensuring that on-chain data feeds operate within acceptable performance and control thresholds will therefore be a key implementation consideration.
The collaboration reflects a wider industry effort to integrate tokenised market structures with established trading and data standards. The key is whether blockchain-based workflows can be incorporated into existing enterprise environments without creating parallel data silos.
As institutions continue to test tokenised fixed income and collateral models, the availability of licensed reference data within on-chain environments may become an increasingly relevant component of production readiness.
Subscribe to our newsletter



