Prof Blake Steenhoven studies the ever-changing corporate disclosure space and reveals what concerned IROs can do in response
Prof Blake Steenhoven
In most cases, effective communication is intuitive – you know it when you hear it. As an IR professional, you’re likely a natural communicator in your day-to-day life and investors are, after all, just people. But what does an effective message sound like when your audience is a machine?
Investors are increasingly adopting advanced tools that use AI and machine learning to analyze disclosures, trying to find the hidden meaning in the numbers, words and even non-verbal behavior in corporate disclosures. These tools are being used during your earnings calls, roadshow presentations and even private phone calls with investors. So how do you communicate effectively in this new disclosure landscape?
In this article, I’ll provide insights from academic literature to explain what some of these tools are, why investors are using them and what IR professionals can do about it.
Much of the content of corporate disclosure is also available in the financial statements, which suggests something most IR professionals know: how something is said can be as important as what is said. And investors looking for an edge over their competitors are using a variety of methods to analyze the words and vocal cues in disclosure.
Textual analysis Methods of analyzing the language in a disclosure range from simple word counts to complex probabilistic modeling. Word counts can be used to measure the frequency of specific words, like ‘revenue’ or ‘risk’, as well as categories of words, like personal pronouns. Specific dictionaries can also be used for characteristics like tone, which compare the number of positive and negative words in a disclosure.
While these approaches are simple to use, a limitation is that they can fail to capture the context of the disclosure. For example, consider the simple sentence: ‘Bad debt expense decreased.’ Although this clearly conveys positive information, a simple word count approach would likely classify each of the words as negative, resulting in a tone measure that doesn’t capture the meaning of the sentence.
More sophisticated approaches try to overcome these issues using tools like machine learning. These methods typically begin by collecting a large sample of historical data, such as the transcripts from a company’s previous earnings calls.
This data is used to ‘train’ a model to look for relationships between groups of words and stock movements. The model can then be used in real time to predict stock prices using the words in new disclosures.
Although machine-learning approaches have some advantages over simpler methods, they also have limitations. Machine learning involves complex computations and algorithmic analysis, which can be costly to implement. The complexity of these approaches can also increase the risk of user error. When combined with high-frequency trading strategies, the speed and efficiency of the analysis means these failures can be catastrophic.
And because they’re based purely on historical data, algorithms can fail when novel circumstances arise. Even in the best-case scenario, the complexity of statistical models can make it difficult or impossible for investors to know what the model is reacting to, unlike simpler metrics or qualitative analysis.
Acoustic analysis In addition to the words you use, investors are also listening to vocal characteristics like pitch, volume and how quickly you talk. These cues can be informative about a speaker’s personality, emotional state and intentions, and research has shown that they can predict financial performance, stock returns and even CEO career outcomes.
Non-verbal communication has always mattered, but the availability of free open-source applications like Praat and Python have allowed these cues to be quantified with greater precision and speed than ever before. While it’s clear that there’s useful information in vocal cues, it’s important to note that this area is still new and investors are still learning how to use this data.
Although there’s some evidence that individual cues like vocal pitch are associated with stock price movements, the general consensus is that these cues need to be looked at in combination with each other.
With little guidance on how to combine these measures, investors may rely on a purely statistical machine-learning approach or off-the-shelf software. These approaches are costlier but the larger issue, particularly for proprietary software, is that they’re opaque and difficult for users to understand. Without being able to see what’s happening under the hood, investors are left to rely on the developers’ claims.
For example, many voice analysis tools are marketed as effective lie detectors, and claim they can determine whether a chief executive is being deceptive. While this idea might sound appealing to investors, these claims should be treated with skepticism; the proprietary nature of these tools prevents independent verification, and the evidence overwhelmingly shows that vocal cues cannot reliably detect deception.
Even so, research consistently shows that stock prices and KPIs are associated with managers’ vocal characteristics, and investors are likely to continue analyzing and trading on these cues.
What we can do about it The disclosure landscape is changing and a common concern I hear from IR professionals is not knowing how to respond. How do you communicate effectively when even subtle changes in vocal pitch can move stock prices? Before you cancel your earnings calls, here are a few recommendations.
Blake Steenhoven is an assistant professor of accounting at the Smith School of Business at Queen’s University, Canada