ChatGPT is the fastest growing app of all time, gaining more than 100 million users just two months after its launch in November last year. It enables users to have human-like conversations that include reasonable-sounding and often correct answers to all sorts of questions. Like humans, it can ask for more information and explain reasoning.
The first academic research about the use of ChatGPT in finance is now being released. Two recent studies make GPT seem like a promising technology, to improve investment decisionmaking and to explain its decisions. Perhaps the long-held dream of replacing humans in finance is coming true.
In December last year, I wrote that “a tireless machine able to digest all information and immune to biases should be clearly superior to humans when it comes to investing. Except it’s not.”
Financial management was one of the earliest goals of artificial intelligence (AI) research because it seemed like an easy and highly rewarding task.
Yet so far, AI has succeeded only in niche applications in finance.
GPT stands for Generative Pre-trained Transformer, a five-year-old idea that might be a game-changer in AI applications.
Very broadly, there are three approaches to extracting useful information from data. With structured data, such as accounting numbers or price histories, statistics and formal models can be applied. With completely unstructured data — series of bits that could be photographs, physical measurements, text or anything else — there are algorithms that can extract patterns and predict inputs.
Language is somewhere in between. There is structure, meaning that only certain letter combinations are intelligible words and there are grammar rules for stringing words together.
Yet there are exceptions to rules and nuances beyond the literal text. A person needs a lot of domain knowledge and context to understand text.
There is an old story — it has been tracked back to 1956 at which time it was already old — about an AI worker who built a program to translate between English and Russian. She gave it the phrase “out of sight, out of mind” to translate to Russian, and then translated the Russian back to English and got “invisible idiot.”
There are no rules of language that tell us the phrase is an aphorism about forgetfulness rather than a description of an individual, but no native speakers would make the mistake.
GPT models are the hottest approach to working with language data, but quantitative trading and investment have used cruder language models for many years. A human researcher reads relevant information such as company statements, news stories, surveys and research reports carefully and slowly.
Computers can read vast quantities of information in many languages and come up with instant conclusions. This is essential for high-frequency trading when being a millisecond sooner to determine whether a news headline is good or bad news for a stock price is the name of the game.
Most of the language models used in quantitative finance today treat it as structured data. Algorithms look for certain words, or just measure the number of words in a headline or news release. Some algorithms look for certain patterns or structures, but none of the major ones try to understand the meaning of the text, and none of them can explain why they reach their conclusions or hold further conversation on the subject.
Now come two papers titled “Can ChatGPT Decipher Fedspeak?” and “Can ChatGPT Forecast Stock Price Movements?”
This is not a discussion about SkyNet taking over Wall Street, but whether ChatGPT beats older models — many of which treat language as structured — in making fast decisions about short texts.
The first paper asked ChatGPT to determine if an individual sentence from a US Federal Reserve statement was “dovish,” suggesting the central bank was more likely to cut than raise interest rates, or “hawkish,” suggesting the opposite.
A high-frequency trading algorithm might rate each sentence in the Fed release and use the output along with other data to trade federal funds futures or other instruments before the human analysts had finished reading the first word in the release.
In this study, ChatGPT clearly did better at matching conclusions of human analysts than dictionary-based models that looked only for certain words. When the researchers fine-tuned ChatGPT by giving it extra training on Fed statements with feedback on how humans rated the statements, it agreed with human researchers about as often as two human researchers agreed with each other, and its explanations for its decisions were plausible.
This is not immediately useful for trading. The paper did not disclose how fast the model ran, nor whether overall interpretations of entire Fed releases agreed well with human conclusions.
Whether they agreed with reality is not the point, as high-frequency traders are trying to beat the market to the new consensus, not to the theoretically correct place.
The paper said that GPT models might have turned a corner with regards to actually understanding language. If that is true — and one study does not prove anything — they could be unleashed on a much wider range of text to generate theses, such as whether inflation is likely to continue to be a problem over the next 12 months, rather than flash signals for high-frequency trading.
Instead of binary buy/sell signals, ChatGPT could also hold a conversation with a human analyst to improve investment decisions.
If this seems to be working, a future generation of GPT models could be trained on the entire history of texts and financial price movements.
The second paper is more directly relevant for trading. It used ChatGPT to rate news headlines as good or bad for stock prices. It tested the strategy of buying a stock with good news at the open after the headline was released and selling at the close; or selling at the open and buying back at the close if the headline was bad.
The results were inconclusive. The ChatGPT signal had a 0.01 correlation with the next day’s raw stock return.
However, to evaluate a signal one needs to compare to the residual return after adjusting for the market return, and perhaps for known factors. A 0.01 correlation could be valuable in combination with other signals, or it might not.
The tested strategy did have positive returns from October 2021 to December last year without transaction costs, but the authors do not provide data on whether it beat a market strategy, nor whether the positive return was significant statistically.
A reported 0.13 percent gross profit per trade suggests that it might not overcome transaction costs.
The authors also report a regression that includes future information, so it cannot be used to evaluate effectiveness for making decisions based on information known at the time.
The ChatGPT signal supplies no additional information to the three decimal places, although it does seem to have some small positive value, the authors said.
Yet inconclusive does not mean failure.
The study did say that ChatGPT was better than popular alternative models, and research on GPT and other large language models is continuing.
GPT is an AI tool that can work with humans, learn from them and teach them, rather than some incomprehensible black box. At the very least, it seems poised to replace older algorithms and to increase the use of AI in quantitative and qualitative investing.
It is a long way from taking over Wall Street, but there is no reason to think it cannot.
Aaron Brown is a former managing director and head of financial market research at AQR Capital Management. He might have a stake in the areas he writes about.
This column does not necessarily reflect the opinion of the editorial board or Bloomberg LP and its owners.
In their New York Times bestseller How Democracies Die, Harvard political scientists Steven Levitsky and Daniel Ziblatt said that democracies today “may die at the hands not of generals but of elected leaders. Many government efforts to subvert democracy are ‘legal,’ in the sense that they are approved by the legislature or accepted by the courts. They may even be portrayed as efforts to improve democracy — making the judiciary more efficient, combating corruption, or cleaning up the electoral process.” Moreover, the two authors observe that those who denounce such legal threats to democracy are often “dismissed as exaggerating or
The Chinese Nationalist Party (KMT) caucus in the Legislative Yuan has made an internal decision to freeze NT$1.8 billion (US$54.7 million) of the indigenous submarine project’s NT$2 billion budget. This means that up to 90 percent of the budget cannot be utilized. It would only be accessible if the legislature agrees to lift the freeze sometime in the future. However, for Taiwan to construct its own submarines, it must rely on foreign support for several key pieces of equipment and technology. These foreign supporters would also be forced to endure significant pressure, infiltration and influence from Beijing. In other words,
“I compare the Communist Party to my mother,” sings a student at a boarding school in a Tibetan region of China’s Qinghai province. “If faith has a color,” others at a different school sing, “it would surely be Chinese red.” In a major story for the New York Times this month, Chris Buckley wrote about the forced placement of hundreds of thousands of Tibetan children in boarding schools, where many suffer physical and psychological abuse. Separating these children from their families, the Chinese Communist Party (CCP) aims to substitute itself for their parents and for their religion. Buckley’s reporting is
As Taiwan’s domestic political crisis deepens, the opposition Chinese Nationalist Party (KMT) and Taiwan People’s Party (TPP) have proposed gutting the country’s national spending, with steep cuts to the critical foreign and defense ministries. While the blue-white coalition alleges that it is merely responding to voters’ concerns about corruption and mismanagement, of which there certainly has been plenty under Democratic Progressive Party (DPP) and KMT-led governments, the rationales for their proposed spending cuts lay bare the incoherent foreign policy of the KMT-led coalition. Introduced on the eve of US President Donald Trump’s inauguration, the KMT’s proposed budget is a terrible opening