About a-team Marketing Services

A-Team Insight Blogs

Talking Reference Data with Andrew Delaney: Feeling the Collars of the Libor Traders

Subscribe to our newsletter

Financial institutions are bracing themselves for further regulatory clampdown on market pricing on the wake of the LIBOR and FX market manipulation scandals. As regulators draft new guidelines for benchmark data contributions, and UK Chancellor George Osborne signals increased scrutiny of the London FX market’s pricing procedures, data managers and compliance officers are assessing their processes to institute best practice – and guard against the risk of further financial penalties.

So far the price has been high, with recent fines relating to LIBOR alone exceeding $7 billion and the Financial Times predicting an eventual total of as much as $22 billion!

In response to the ensuing public outrage, regulators are starting to act. Several have prescribed new measures to safeguard a repetition, and many institutions are assessing what they should be doing to ensure their contributions are accurate, timely and honest.

But what is best practice? And what can firms do to secure control over the prices and rates their traders – and increasingly, pricing engines and chat lines – distribute to the marketplace?

As you know these questions have been vexing me for quite some time, not least because we’ve been noticing a number of prospective industry solutions emerging that make use of monitoring technologies to keep tabs on the proprietary pricing data that makes its way to the markets in various ways.

We’ve investigated possibilities in a number of ways, the most recent of which has been our preparation for next week’s webinar on the topic, ‘Bracing for the Wave—or Sailing Ahead of It?’; Reducing Risk Through Benchmark Data Controls. Featuring speakers from Verint, BT and Z/Yen, this webinar looks at how firms can do more to monitor their traders’ voice, data and chat activities and institute controls to ensure compliance with internal policy and new regulatory guidelines. Our panel of experts will discuss approaches to instituting robust and consistent monitoring across all pricing channels, and offer practical advice on technology solutions and organisational methodologies to ensure best practice.

It’s a topic I plan to explore more at our Data Management and Intelligent Trading Summits in London and New York later in the year. In the meantime, however, be sure to join us at 3pm London time (10am New York), next Wednesday July 2. You can register free of charge here.

Relatedly, we just completed a white paper with infrastructure monitoring specialist ITRS that looks more broadly at firms’ data contributions to industry benchmarks, including indices, and fragmented price-driven markets, like Swap Execution Facilities (SEFs).

Changing market dynamics are placing new emphasis on market-makers and other generators of proprietary contributed market pricing to ensure the integrity of their data. Structural shifts like the introduction of exchange-like (SEFs) are exacerbating the pressures on financial institutions to get their pricing systems in good shape.

Concerns about data integrity are impacting a wide range of asset classes, including over-the-counter markets – like fixed income, foreign exchange and derivatives – and exchange-listed so-called structured instruments – like exchange-traded funds (ETFs), warrants, turbo-warrants, certificates and indices.

Meanwhile, as firms seek to monetise the data they generate from their business activities, new demands for quality control are driving a renewed look at data integrity. With data sales emerging as a significant element of the overall revenue picture, financial institutions are striving to adopt best practices in response to customer demand for higher- quality data services.

Our new paper – ‘Ensuring Integrity of Proprietary & Derived Market Information’ – discusses the obstacles to establishing robust pricing processes, and examines the market dynamics driving efforts towards improved data integrity and where best practices are evolving. You can download it free of charge here.

Subscribe to our newsletter

Related content

WEBINAR

Recorded Webinar: How to maximise the use of data standards and identifiers beyond compliance and in the interests of the business

Data standards and identifiers have become common currency in regulatory compliance, bringing with them improved transparency, efficiency and data quality in reporting. They also contribute to automation. But their value does not end here, with data standards and identifiers being used increasingly for the benefit of the business. This webinar will survey the landscape of...

BLOG

Nasdaq CSD Renews LEI Service Platform

Nasdaq CSD, a central securities depository providing access to the Estonian, Icelandic, Latvian and Lithuanian markets, and an accredited Legal Entity Identifier (LEI) issuer in the Baltic states and Nordics, has released an enhanced LEI service platform, Nasdaq LEI, that provides more straightforward and effective LEI code issuance and management to help firms meet regulatory...

EVENT

TradingTech Briefing New York

Our TradingTech Briefing in New York is aimed at senior-level decision makers in trading technology, electronic execution, trading architecture and offers a day packed with insight from practitioners and from innovative suppliers happy to share their experiences in dealing with the enterprise challenges facing our marketplace.

GUIDE

Data Lineage Handbook 2019

Welcome to our latest handbook on data lineage, a critical concern for data managers working to achieve regulatory compliance, deliver operational gains, and provide meaningful value to the business. The handbook covers the complete scope of data lineage, with a view to helping you win management buy-in and budget, decide whether to build or buy...