A couple of weeks ago, together with Reuters, we conducted a survey of the marketplace to find out how well latency measurement is entrenched, and what users thought of datafeed providers and measurement tool vendors. And we asked them to comment on their current market data handling infrastructures too.
One figure that came back – which I had an instant gut reaction against – is that 39 per cent of respondents reckon their systems are adequate to cope with increasing market data volumes. To me, this number seemed just too high, given what we are being told to expect about data rates from OPRA, etc. A further 31 per cent reckoned their systems would be able to cope following planned upgrades.
Contrast those figures with just 24 per cent suggesting that measurement/monitoring tools are currently adequate (the rest say that tools are not good enough, or are still investigating tools, including finding out what’s out there because they don’t know of any). There’s an old adage in the IT industry and I’ve heard it a lot of late. It’s that unless one can monitor/measure it, one can’t manage it.
Personally, I reckon many of the 39 per cent are large investors in silicon – not the silicon of microprocessors, but the silicon found in large buckets of sand.
Am I being harsh? Of course, I’d be delighted to hear more from you on this subject.
Until next time … here’s some good music.
[tags]OPRA,benchmarks,latency measurement,latency monitoring[/tags]