Beginning in the early 2000s, when the algorithms and software capable of performing transaction cost analysis (TCA) on a semi-automated basis first became prevalent, the definition of the function was always: a method of determining the effectiveness of a set of transactions performed by a counterparty – the key word within that definition being ‘effectiveness.’
In 2019, the global financial services industry is set to spend an estimated USD 50bn on the raw, historical markets and transactions data inputs required to fuel a broad spectrum of daily trading activities across all major asset classes.
In 2019, the wholesale FX market is not the market of yore. Formerly, the defining characteristic of the currencies trading landscape was liquidity fragmentation and siloed pools of cash held within geography-specific currency pairings.
Banks are one of the greatest engines for generating data: daily, they collectively produce petabytes of transactions, prices, risk metrics, customer information….
A decade after the financial crisis, the buyside (asset managers, hedge funds, institutional investors and large corporates) have changed at least as much as the investments banks that serve them.
Gold-i CEO Tom Higgins recently claimed in Finance Magnates Magazine (“Cryptocurrency liquidity, past, present, future” article) that “the most challenging factor continues to remain the access to and quality of liquidity.”