Fundamental Analysis Via Machine Learning
Fundamental Analysis Via Machine Learning
To cite this article: Kai Cao & Haifeng You (2024) Fundamental Analysis via Machine Learning,
Financial Analysts Journal, 80:2, 74-98, DOI: 10.1080/0015198X.2024.2313692
To link to this article: https://doi.org/10.1080/0015198X.2024.2313692
of a comprehensive set of financial statement items,2 fail to consider even though our ML forecasts are
let the algorithms “learn” from the historical data to based on financial statement data only.
determine the underlying relationships, and generate
Our paper contributes to the burgeoning literature
out-of-sample forecasts for 134,154 firm-year observa-
on the application of ML in finance and accounting
tions from 1975 to 2019. We compare the accuracy,
research (e.g., Chen et al. 2022; Gu, Kelly, and Xiu
information content, and investment value of the out-
2020; Bao et al. 2020; Bertomeu et al. 2021; Ding
of-sample ML forecasts to the forecasts obtained from
et al. 2020) by demonstrating the decisional useful-
the RW model and five other models developed in the
ness of ML technology in one of the most important
extant literature: the (first-order) autoregressive model
tasks in fundamental analysis: corporate earnings
(AR), two models (HVZ and SO) developed by Hou, van
forecasting. We show that by identifying economi-
Dijk, and Zhang (2012) and So (2013), respectively; and
cally important predictors and capturing subtle non-
the earnings persistence (EP) and residual income (RI)
linear relationships, ML can better utilize financial
models proposed by Li and Mohanram (2014).
statement information and produce significantly more
We find that the ML forecasts are significantly more accurate and informative earnings forecasts.
accurate than not only the RW model but also all
Our paper is also of interest to investment professio-
extant models. The mean absolute forecast errors of
nals. First, our model can be used to generate earn-
the extant models are approximately 7.79% to
ings forecasts at a low cost even for firms with a
26.53% greater than that of the ML model. Cross-
short history and those without analyst coverage. It
sectional analyses indicate that the ML model leads
not only offers a less biased alternative to analyst
to even greater accuracy improvements among firms
forecasts as a valuation input by stock pickers (e.g.,
with more difficult-to-forecast earnings. The ML fore-
Ohlson 1995; Ohlson and Juettner-Nauroth 2005)
cast also has greater information content as measured
but also can be directly used to value many firms
by its predictive power with respect to future actual
without analyst coverage. Second, investors may also
earnings changes (ECH). Forecasted earnings changes
easily modify our models to forecast other funda-
(FECH) based on the ML forecast explain 18.61% of
mental variables (e.g., sales, gross profit) that are of
the variation in ECH, whereas FECH based on the
considerable importance for investors. Third, our
extant models only explain about 8.07% to 12.22%.
paper also offers a potential systematic trading strat-
We then test whether the new information uncov- egy to quantitative investors, who can build on our
ered by the ML model can lead to significant study and refine the model by, for example, forecast-
improvements in investment decision-making. To this ing future quarterly earnings and other fundamentals
end, we orthogonalize the ML forecast against the or by incorporating non–financial statement informa-
contemporaneous forecasts from the extant models tion to further improve returns to the strategy.
and use the residuals to measure the new informa-
tion uncovered by the ML forecasts (beyond the
extant models). The results show that the new infor- Related Literature and Extant
mation component has significant predictive power Earnings Forecasting Models
with respect to future stock returns. The top quintile As Monahan (2018) summarizes, early research
of stocks with the most favorable new information mostly adopted the time-series approach to forecast
significantly outperforms the least favorable new future earnings (e.g., Ball and Watts 1972; Watts and
information quintile by approximately 34 to 77 bps Leftwich 1977). Overall, their results suggest that the
per month on a risk-adjusted basis. simple RW model, which predicts expected future
earnings to equal current earnings, generates more
We also find that the ML forecasts perform well
accurate out-of-sample forecasts than the more
against analyst consensus forecasts, even though ana-
sophisticated autoregressive integrated moving aver-
lysts have access to much more information than the
age (ARIMA) models (e.g., Brown 1993; Kothari
financial statements. First, the ML forecast has signifi-
2001).3 The superiority and simplicity of the RW
cantly lower mean absolute forecast errors than analyst
model make it a natural benchmark in evaluating
consensus earnings forecasts for all three-year forecast
other earnings forecasting models.
horizons. Second, the ML forecast has greater relative
information content than analyst forecasts in predicting There are several potential reasons for the poor per-
future earnings changes. Third, the ML forecast con- formance associated with the ARIMA models. First,
tains a significant amount of information that analysts these models require a long time series to yield
reliable parameter estimates, but the earnings pro- assets; NDDi, t indicates a zero dividend; DIVi, t is
cess may not be stationary over a long period. dividends per share; BTMi, t is the book-to-market
Second, these firm-specific time-series models ignore ratio; and Pricei, t is the stock price at the end of
the rich information in other financial statement line the third month after the end of fiscal year t. In
items. To overcome these limitations, subsequent addition to financial statement items, the SO model
studies turn to cross-sectional or panel-data uses the stock price and book-to-market ratio,
approaches, which use a pooled cross-section of allowing it to utilize more forward-looking
firms to estimate forecasting models. These models information.
are considered as state-of-the-art earnings prediction
models in the literature (e.g., Gerakos and Gramacy The final two cross-sectional models are proposed by
2013; Call et al. 2016).4 We therefore employ them Li and Mohanram (2014):
as alternative benchmarks. We discuss these models Ei, tþ1 ¼ a0 þ a1 NegEi, t þ a2 Ei, t þ a3 NegEi, t Ei, t þ ei, tþ1
below and summarize them with detailed variable (4)
definitions in Appendix 1 Panel A.
Ei, tþ1 ¼ a0 þ a1 NegEi, t þ a2 Ei, t þ a3 NegEi, t Ei, t
The first cross-sectional model that we examine is þ a4 BVEi, t þ a5 TACCi, t þ ei, tþ1 (5)
the first-order AR model:
Ei, tþ1 ¼ a0 þ a1 Ei, t þ ei, tþ1 (1) Equation (4) allows loss firms to have different
earnings persistence (the EP model) from profitable
where Ei, t is firm i’s earnings in year t. Gerakos and firms. Equation (5) is based on the residual income
Gramacy (2013) show that the AR model performs model (the RI model) proposed by Feltham and
well relative to the RW model and is more accurate Ohlson (1996), which further augments the EP
than many sophisticated models. model with the book value of equity (BVEi, t ) and
total accruals (TACCi, t ) from Richardson et al.
The second model, the HVZ model, proposed by (2005).
Hou, van Dijk, and Zhang (2012), extends the Fama
and French (2000, 2006) model and forecasts future Although the above models are the state-of-the-art
earnings as follows: models in the literature, they fail to consistently out-
perform the RW model (e.g., Monahan 2018; Easton,
Ei, tþ1 ¼ a0 þ a1 Ai, t þ a2 Di, t þ a3 DDi, t þ a4 Ei, t
Kelly, and Neuhierl 2018). Taken at face value, these
þ a5 NegEi, t þ a6 ACi, t þ ei, tþ1 (2) results seem to suggest that there is not much incre-
mental information in financial statements beyond
This model forecasts future earnings with total assets the bottom-line earnings, which contradicts both the
(Ai, t ), the dividend payment (Di, t ), the dividend-paying conventional wisdom and the fundamental tenet of
dummy (DDi, t ), historical earnings (Ei, t ), and an indica- financial statement analysis. Given that “the question
tor variable for negative earnings (NegEi, t ) and of whether historical accounting numbers are useful for
accruals (ACi, t ). forecasting earnings is central to accounting research”
(p. 183, Monahan 2018), both Monahan (2018) and
The third model, the SO model, is developed by So Easton, Kelly, and Neuhierl (2018) call for further
(2013). So modifies the Fama and French (2006) research in this area. We posit that, rather than indi-
model and forecasts future earnings per share (EPS) cating the lack of the decisional usefulness of finan-
using the following model: cial statements or fundamental analysis, the above
results might be driven by the methodological limita-
EPSi, tþ1 ¼ a0 þ a1 EPSþ − þ
i, t þ a2 NegEi, t þ a3 ACi, t þ a4 ACi, t tions of the extant models, which can be overcome
þ a5 AGi, t þ a6 NDDi, t þ a7 DIVi, t þ a8 BTMi, t by ML algorithms.
þ a9 Pricei, t þ ei, tþ1
(3) ML-Based Earnings Forecasts
where EPSþ
is the EPS for positive earnings and is
i, t The extant models do not make the best use of
zero otherwise; NegEi, t is an indicator variable for information in financial statements to forecast future
negative earnings; AC−i, t is accruals per share for earnings. First, the extant models focus on a small
negative accruals and is zero otherwise; ACþ i, t is number of aggregate financial statement items, such
accruals per share for positive accruals and is zero as bottom-line earnings and total assets, and fail to
otherwise; AGi, t is the percentage change in total fully consider many other financial statement line
76
Fundamental Analysis via Machine Learning
items that could be highly valuable for earnings pre- their inclusion are provided in Appendix 1 Panel B.
diction (e.g., Fairfield, Sweeney, and Yohn 1996; The input predictors (a total of 60) can be broadly
Chen, Miao, and Shevlin 2015). Second, even categorized as follows:
though economic theories and empirical evidence
suggest the prevalence of nonlinear relationships i. Historical earnings and their major compo-
between historical accounting information and nents (8). We include earnings components,
future earnings (e.g., Lev 1983; Freeman and Tse as prior literature has shown that disaggregat-
1992; Baginski et al. 1999; Chen and Zhang 2007), ing earnings provides additional information
these models mostly adopt linear functional forms and improves earnings forecasts (Fairfield,
(or with some simple interactions) and are therefore Sweeney, and Yohn 1996; Chen, Miao, and
unlikely to capture these subtle yet important rela- Shevlin 2015).
tionships. ML algorithms are designed to handle ii. Other important individual income statement
high-dimensional data and are rather flexible with items (5). Specifically, we include advertising
respect to the functional forms of the underlying and R&D expenses, as they tend to generate
relationships. Thus, they can potentially overcome long-term future benefits (e.g., Lev and
the above limitations and generate better earnings Sougiannis 1996; Dou et al. 2021). We also
forecasts. Below we describe the development of include special and extraordinary items and dis-
our ML models. continued operations, as firms may shift core
expenses to these items to manipulate core
Financial Statement Line Items as earnings (e.g., Barnea, Ronen, and Sadan 1976;
Predictors. Both the financial ratios and raw val- McVay 2006). Finally, we include common divi-
ues of financial statement items can be used in dends, as they may signal firms’ future earning
forecasting future earnings.5 A priori, it is unclear power (e.g., Nissim and Ziv 2001).
which choice is better. On the one hand, financial iii. Summary and individual balance sheet items
ratios are grounded in economic theories, which (16). The balance sheet summarizes resources
may help filter out the noise in the raw values of with potential future economic benefits and
financial items and better capture the key drivers of may contain incremental information. For exam-
future earnings. On the other hand, the transforma- ple, the literature shows that the book value of
tion from raw values to ratios may cause a loss or equity is an important driver of future earnings
distortion of information. In a recent study, Bao (e.g., Feltham and Ohlson 1995; Li and
et al. (2020) compare the predictive power of ML Mohanram 2014), and Ball et al. (2020) argue
models based on raw accounting numbers versus that retained earnings may measure average
those based on financial ratios with respect to earning power better than the book value of
accounting fraud. They find that raw value–based equity.
models perform significantly better. Following their iv. Operating cash flows (1). We include cash flows
study, we rely on raw accounting values as our pri- from operating activities, as prior studies have
mary set of predictors and examine alternative pre- found that the cash flow component of earn-
dictor sets of financial ratios in additional analyses. ings is more persistent than the accrual compo-
Furthermore, we follow prior literature (e.g., Li and nent, and separating them helps predict future
Mohanram 2014) and scale all raw accounting val- earnings (e.g., Sloan 1996; Call et al. 2016).
ues (including both the input features and the tar- v. The first-order differences of the above items
get variable) by the number of common shares (30). We include them, as prior studies show
outstanding. that changes in financial statement items often
contain incremental information beyond the
As Richardson, Tuna, and Wysocki (2010) suggest, a levels (e.g., Kothari 1992; Ohlson and Shroff
concern with the early literature of fundamental anal- 1992; Richardson et al. 2005).
ysis is the in-sample identification of predictive varia-
bles. To mitigate this concern, we select a ML Algorithms. Following prior studies (e.g.,
comprehensive list of key financial statement items Rasekhschaffe and Jones 2019; Cao et al. 2024),
and let ML algorithms “learn” from historical data in we develop a ML earnings prediction model by
terms of how to optimally select and combine these ensembling the out-of-sample earnings forecasts
items. The resulting models are then used to gener- generated from several popular ML algorithms.
ate out-of-sample earnings predictions. The detailed Among them, two algorithms are based on decision
list of financial statement items and the rationale for trees and one is based on ANNs. Our two decision
tree–based algorithms are the standard RF and the Data, Sample Selection, and Model
GBR algorithms. In implementing our ANN-based
algorithm, we adopt the bootstrap aggregating (i.e.,
Estimation Procedure
bagging) technique by constructing 10 bootstrap Our initial sample comprises 267,777 firm-year
samples, with each sample randomly drawing 60% observations obtained from the intersection of the
of the observations from the training set. Compustat fundamentals annual file7 and the Center
Thereafter, we train an ANN model for each boot- for Research in Security Prices (CRSP) data up to fis-
strapped sample and then average the 10 models cal year 2019. We further impose the following data
to generate predictions.6 requirements: (1) the following financial statement
items must be non-missing: total assets, sales reve-
Cross-Validation and Hyperparameter nue, income before extraordinary items, and common
Tuning. In the implementation of ML, it is impera- shares outstanding; (2) the stocks must be ordinary
tive to select a model with an appropriate level of common shares listed on the NYSE, AMEX, or
complexity because overly simple models tend to NASDAQ; (3) the firms cannot be in the financial (SIC
underfit the data, while overly complex models tend 6000–6999) or regulated utilities (SIC 4900–4999)
to have an overfitting problem, and both lead to poor industries; and (4) the stock prices at the end of the
out-of-sample predictability. The level of model com- third month after the end of the fiscal year must be
plexity is largely determined by the value of certain greater than US$1. Among the remaining firm-year
hyperparameters, which must be set before estimat- observations, we replace missing values of the cash
ing other parameters such as regression coefficients flow from operating activities with the corresponding
and neural network weights. For example, values of numbers computed from the balance sheet approach
the hyperparameters such as the maximum depth of (Sloan 1996). We then set the missing values of the
the decision trees in the RF and GBR models deter- remaining line items to zero before computing the
mine the overall model complexity as well as the first-order differences of the 30 items in Appendix 1.
number/fraction of input features effectively used in This leaves us with a final sample of 156,256 obser-
the model. vations from 1965 to 2019. Because we need data
from the past 10 years to estimate the models, our
We search for the “optimal” hyperparameter values testing sample (i.e., prediction set) starts from 1975
through hyperparameter tuning via cross-validation. and consists of 142,592 firm-year observations.
We use fivefold cross-validation to identify the opti- Table 1 presents the number of firms in the final
mal hyperparameter values that generate the most testing sample by year, where the number of annual
accurate forecasts on the validation samples. observations ranges from 2,299 in 2019 to 4,976 in
Specifically, for each year in our test sample, we use 1997.
data from the previous 10 years to train the ML mod-
els. We randomly split the 10 years of data into five At the third month end after fiscal year t, we gener-
groups/folds of validation sets, with each fold includ- ate the out-of-sample forecasts of one-year-forward
ing 20% of the data. For each fold, we use the earnings Etþ1 for the above testing sample using the
remaining 80% of the firm-year observations as the aforementioned machine learning algorithms and the
training sample. For each of the ML algorithms, we 60 predictors. Following prior literature (e.g., Hou,
provide a set of reasonable candidate values for the van Dijk, and Zhang 2012; Li and Mohanram 2014),
key hyperparameters (see details provided in for each year t between 1975 and 2019, we use all
Appendix 2). For each combination of the hyperpara- observations from the previous 10 years (i.e., year
meters, we compute the mean squared forecast t − 10, t − 9, … , t − 1) to estimate the models. As dis-
errors of the five validation sets using the model esti- cussed earlier, we use fivefold cross-validation to
mated from the remaining 80%, which forms the identify the optimal hyperparameters. Using these
training set. The mean squared forecast errors on the optimal values, we retrain the model using the obser-
validation sets are used as the basis to select the vations from the previous 10 years and then apply
optimal hyperparameters, which are then used to the trained models to the predictors of year t to gen-
train a new model on the training set. We then apply erate earnings forecasts for year t þ 1. For consis-
the model to the current year’s financial statement tency, all extant models are also estimated using the
data to generate out-of-sample earnings forecasts for data of the same previous 10 years, and the resulting
the following year, which are then compared to the linear models are applied to their respective predic-
subsequent actual earnings to evaluate the relative tors in year t to generate earnings forecasts for
performance of various models. year t þ 1.8
78
Fundamental Analysis via Machine Learning
highest quintiles of the ROA volatility, magnitude of univariate regression results, confirming the results in
total accruals, R&D expense, and loss makers, respec- Panel A that ML brings greater improvement in accu-
tively. In Panel B of Table 3, we conduct a regression racy among firms with higher earnings volatility, a
analysis of the difference in the forecast accuracy larger magnitude of accruals, more R&D expense,
between RW and ML forecasts on the above deter- and loss makers. The multivariate regression results
minants. Columns (1) through (5) report the in column (6) are largely consistent, except that the
ML 6.87 2.91
RW 7.64 0.77 7.85 11.20% 3.09 0.18 5.73 6.12%
AR 7.55 0.68 8.96 9.93% 3.08 0.17 6.16 5.87%
HVZ 7.43 0.55 8.31 8.07% 3.11 0.20 7.95 6.93%
EP 7.42 0.55 8.06 8.03% 3.13 0.22 7.89 7.63%
RI 7.41 0.54 7.61 7.79% 3.11 0.20 7.82 6.90%
SO 8.70 1.82 12.72 26.53% 3.47 0.57 14.81 19.45%
ML 9.09 4.42
RW 10.28 1.19 8.07 13.06% 4.73 0.31 5.25 6.93%
AR 10.18 1.09 9.50 11.99% 4.70 0.28 9.76 6.41%
HVZ 9.71 0.62 7.71 6.84% 4.62 0.19 6.68 4.41%
EP 9.64 0.55 6.03 6.06% 4.66 0.23 6.56 5.31%
RI 9.56 0.47 5.29 5.22% 4.60 0.18 6.01 4.14%
SO 10.31 1.22 11.87 13.42% 4.91 0.49 12.80 11.01%
ML 10.88 5.58
RW 12.25 1.37 7.28 12.54% 5.92 0.34 3.38 6.16%
AR 12.27 1.38 9.86 12.69% 5.93 0.35 8.02 6.28%
HVZ 11.42 0.53 6.96 4.88% 5.73 0.15 4.07 2.64%
EP 11.38 0.50 5.04 4.56% 5.79 0.21 5.03 3.73%
RI 11.21 0.33 3.73 3.03% 5.70 0.12 3.37 2.20%
SO 12.03 1.14 9.62 10.50% 6.11 0.53 11.41 9.47%
continued
80
Fundamental Analysis via Machine Learning
Table 2. (continued)
D: Alternative measures of forecast accuracy (N = 134,154 firm-years)
MAE of Ranks MSE of De-medianed Forecasts
Comparison with ML Comparison with ML
Average*100 DIFF t stat. %DIFF Average*100 DIFF t stat. %DIFF
ML 10.07 1.76
RW 10.58 0.51 10.92 5.03% 2.21 0.45 3.52 25.67%
AR 10.63 0.56 12.46 5.55% 2.13 0.37 3.52 20.87%
HVZ 10.71 0.63 15.04 6.30% 2.06 0.29 3.32 16.66%
EP 10.58 0.51 10.95 5.04% 2.07 0.31 3.72 17.51%
RI 10.63 0.56 12.26 5.53% 2.05 0.29 3.34 16.54%
SO 11.54 1.47 11.79 14.54% 2.77 1.00 8.77 56.90%
Notes: This table compares the accuracy between the machine learning (ML) forecast and the extant models over the sample
period of 1975–2019. Panels A, B, and C report the time-series average of the mean and median absolute forecast errors of one-,
two-, and three-year-ahead earnings forecasts, respectively. The absolute forecast error is calculated as the absolute value of the
difference between the actual future earnings and the earnings forecasts, scaled by the market equity at the end of three months
after the end of the last fiscal year. Panel D reports alternative measures of forecast accuracy, i.e., the time-series average of the
mean absolute errors of the scaled rank of forecasts difference and the mean squared errors of the de-medianed forecasts differ-
ence. DIFF is the time-series average of the difference, calculated as the mean (median) absolute forecast error of each model
minus that of the ML model. The t statistic of DIFF time-series is reported accordingly. The percentage difference (%DIFF) is DIFF
divided by the time-series average of the annual mean (median) absolute forecast error of the ML model.
Table 3. (continued)
B: Multivariate regression analysis of the improvement in the accuracy of ML forecasts relative to the RW forecasts (N =
129,310)
Y = |RW forecast errors| − |ML forecast errors|
(1) (2) (3) (4) (5) (6)
Notes: This table presents a cross-sectional analysis of the improvement in the forecast accuracy of the machine learning (ML)
model, relative to that of the random walk (RW) model. The percentage improvement in Panel A is defined as the time-series aver-
age of the annual difference in the mean absolute forecast errors (MAFE) between the ML and RW models, divided by the MAFE
of the RW model. A positive number indicates improved accuracy of the ML model. In Panel A, we sort all firms into quintiles for
each year based on the magnitude of the partition variable (ROA volatility, absolute value of total accruals divided by total assets,
and absolute value of working capital accruals divided by total assets, respectively). For R&D expense, we classify all firms with
missing R&D expense into a separate group and sort the remaining firms into quartiles for each year based on their R&D expense
divided by total assets. We also divide all firms into two groups for each year, depending on whether their earnings are negative.
In Panel B, we regress the difference in the forecast accuracy on the determinants each year and report the mean coefficients of
the annual regressions, as well as the corresponding Fama–MacBeth t statistics.
ROA volatility becomes statistically insignificant and ECH on the FECH of different models. To facilitate
the coefficient on R&D expense flips its sign in the the comparison of the coefficients, we follow the lit-
presence of other determinants. erature to standardize FECH so that it has a zero
mean and unit variance each year. The three columns
Information Content Analysis. Forecast accu- in the middle panel of Table 4 show that the coeffi-
racy is not the sole determinant of the decision use- cients on FECH for the extant models range from
fulness of earnings forecasts. For example, although 0.0304 to 0.0480, explaining between 8.07% and
the RW forecast is more accurate than other fore- 12.22% of the cross-sectional variation in realized
casts, it provides no information with respect to earnings changes. In contrast, the FECH based on
future earnings changes. In this section, we evaluate the ML forecast has a regression coefficient of
the information content of various models by investi- 0.0606 and an explanatory power of 18.61%. Next,
gating their (out-of-sample) predictive power with
we estimate a multivariate regression of ECH on
respect to the future earnings change, ECH. ECH is
FECH based on the ML model by controlling for the
computed as the difference between earnings in year
FECH of all extant models. The right panel of Table 4
t þ 1 and those in year t, scaled by market capitaliza-
shows that the coefficient on FECHML is significantly
tion at the end of the third month after the end of
positive, with a t statistic of 17.98. In contrast, most
fiscal year t. We calculate the forecasted earnings
of the FECH coefficients based on the extant models
change, or FECH, as the predicted earnings for year
become statistically insignificant (or have the wrong
t þ 1 minus the actual earnings for year t, scaled by
sign), except for that of the SO model, which also
market capitalization at the third month end after the
uses forward-looking nonfinancial statement predic-
end of fiscal year t.
tors such as the stock price and book-to-market
We first compare the mean correlation coefficients ratio.
between the ECH and FECH calculated from differ-
ent models. The left two panels of Table 4 show that Economic Significance Analysis. The above
the Pearson (Spearman) correlation coefficients results suggest that ML technology helps generate
between FECH and ECH range from 0.199 to 0.321 more accurate earnings forecasts of (statistically) sig-
(0.117 to 0.179) for the extant models, which are all nificant incremental information beyond the extant
lower than the correlations between the ECH and models. However, it is unclear whether such new
FECH of the ML forecast, 0.413 (0.3). We then esti- information is economically significant. To shed light
mate the univariate Fama–MacBeth regression of on the economic significance of the results, we test
82
Fundamental Analysis via Machine Learning
Notes: This table performs the information content analysis of the machine learning (ML) forecast against the extant models. The
left panel reports the average annual cross-sectional Pearson (Spearman) correlation coefficients between the forecasted earnings
changes calculated using various models and the actual earnings changes for a total of 134,154 firm-year observations, with an
average of 2,981 firms per year over the 45-year sample period from 1975 to 2019. The middle panel reports the univariate
Fama–MacBeth regression results. In the regression, all forecasted earnings changes are standardized to have a zero mean and
unit variance each year. The right panel reports the multivariate Fama–MacBeth regression results. Specifically, we regress earn-
ings changes (ECH) on the forecasted earnings changes (FECH) of the ML forecasts and control for all earnings changes predicted
using the extant models. All independent variables are standardized to have a zero mean and unit variance each year. All earnings
changes are scaled by the market equity at the end of three months after the end of the last fiscal year. The table presents the
average coefficients, along with the Fama–MacBeth t statistics and the average adjusted R2. The subscripts are omitted for
brevity.
whether the new information in the ML forecasts has total assets) in light of Fama and French (2015). The
(statistically and economically) significant predictive timeline for the computation of EXRET12Mi, tþ1 and
power with respect to future stock returns.10 To cap- ML RESDi, t is provided in Figure 1, and the detailed
ture the new information uncovered by ML models, variable definitions are provided in Panel C of
we orthogonalize the ML-based forecasts against the Appendix 1. If the new information component has
forecasts generated using the RW and extant models. already been fully priced (when the year t financial
Specifically, we run an annual cross-sectional regres- statements are announced), the coefficient on
sion of the ML forecasts on the RW forecasts and ML RESDi, t , i.e., b1 would be statistically
the forecasts of the five extant models each year and insignificant.
use the residual to measure the new information
Table 5 presents the Fama–MacBeth regression
uncovered by the ML models. Then, we estimate the
results of model (6). Column (1) reports the regres-
following models to test whether the residual fore-
sion results using the same set of control variables as
casts predict future stock returns:
in Bartram and Grinblatt (2018). The results show
XS
EXRET12Mi, tþ1 ¼ b0 þ b1 ML RESDi, t þ cX that the new information component of the ML fore-
s¼1 s i, s, t
cast, ML_RESD, is significantly associated with future
þ IndustryFE þ ei, tþ1 12-month excess returns, even after controlling for
(6) various return-predictive factors. In column (2), we
further control for ROE and INV (Fama and French
where EXRET12Mi, tþ1 is the one-year-ahead cumula-
2015). The results remain robust, showing that the
tive excess return starting from the fourth month of
new information component still exhibits a positive
fiscal year t þ 1 for firm i. ML RESDi, t is the residual
and statistically significant association with future
from the regression that orthogonalizes the ML fore-
stock returns.11
cast against the RW model and the five extant mod-
els. Xi, s, t is the end-of-year t value of firm i’s control We also conduct a portfolio analysis. Specifically, at the
characteristics. We follow Bartram and Grinblatt beginning of each month, we estimate the new infor-
(2018) to select the list of control variables and also mation component as the residual from the regression
control for industry fixed effects. Furthermore, we of the ML forecasts on the contemporaneous forecasts
add ROE (return on equity) and INV (growth rate of generated from the RW model and the five extant
Notes: This figure presents the timeline for the variables used in the future return prediction analysis. Assume that fiscal year t þ 1
(2011) of firm i ends on 12/31/2011. In the future return prediction analysis, we regress EXRET12Mi, tþ1 , which is the excess cumu-
lative return over the period of 04/01/2011 to 03/31/2012 on ML RESDi, t , which is a proxy for the new information component
of a machine learning forecast, EtMODEL ½Xi, tþ1 , estimated on 03/31/2011.
models. We then sort all stocks into quintiles based on hedge portfolio. For example, the hedge portfolio
the residuals for each three-digit SIC industry. We con- yields a Fama–French five-factor alpha of approxi-
struct a hedge portfolio that takes long positions in mately 45 bps per month. In Figure 2, we plot the
quintiles with the most favorable new information and cumulative log returns (value-weighted) for the five
short positions in quintiles with the least favorable new quintiles and the hedge portfolio. The plot shows
information. Table 6 reports the mean monthly return, that the cumulative returns increase monotonically
Sharpe ratio (annualized), CAPM alpha, Fama–French with the quintile rank, and the hedge portfolio
three-factor alpha, Carhart four-factor alpha, Fama– returns are reasonably consistent over time.12
French five-factor alpha (i.e., the three-factor model
plus the Conservative Minus Aggressive (CMA) and
Robust Minus Weak (RMW) factors), and the alpha Additional Analyses
after controlling for all factors in the Fama and French
dataset, for the five new information component quin- The Importance of Nonlinear Effects. We
tiles, as well as the hedge portfolios that take long conduct several additional analyses to better under-
positions in the top quintiles and short positions in the stand the underlying reasons for the superior perfor-
bottom quintiles. mance of the ML forecasts. We first plot the feature
(Gini) importance charts of the RF model to check
Panel A of Table 6 reports the results for the equally whether they use economically sensible features to
weighted portfolios, showing that the mean monthly generate predictions. Figure 3 shows that past earn-
excess returns increase monotonically from 0.50% ings and operating cash flows are important predic-
for the lowest quintile to 1.23% for the highest quin- tors for future earnings, ranked first and third,
tile. The hedge portfolio generates a monthly mean respectively. Interestingly, total income tax and its
return of 0.73%. The Sharpe ratio also increases first-order difference are the second and fourth most
monotonically from 0.25 to 0.69 for the five quin- important features, respectively. These findings are
tiles. The hedge portfolio generates a Sharpe ratio of consistent with recent literature on the important
1.29. Furthermore, we find that the risk-adjusted
role of tax income or expenses in capturing the qual-
returns (or alphas) increase consistently with the
ity of earnings and predicting future fundamentals
quintile rank. Finally, even after controlling for all fac-
and stock returns (e.g., Lev and Nissim 2004; Hanlon
tors in the Fama and French dataset, the hedge port-
2005; Hanlon, Laplante, and Shevlin 2005; Thomas
folio still earns a monthly alpha of 51 bps.
and Zhang 2011, 2014). Panels A through C of
The results for the value-weighted portfolios appear- Figure 4 present the accumulated local effects (ALE)
ing in Panel B of Table 6 are slightly weaker but still plots (Apley and Zhu 2020) of the top five most
significant both statistically and economically. The important features of the RF models13 for 1975,
hedge portfolio generates a monthly mean return of 1995, and 2015, respectively. The figures show obvi-
0.48%, with a Sharpe ratio of 0.6. Furthermore, all ous nonlinear relationships between these input fea-
Fama–French factor alphas are still significant for the tures and future earnings.14
84
Fundamental Analysis via Machine Learning
Notes: This table summarizes the monthly return/alpha to quintiles sorted based on the new information uncovered using the
machine learning (ML) models. At the beginning of each month, we estimate the new information component as the residual from
the cross-sectional regression of ML forecasts on the forecasts generated from the random walk (RW) model and the five extant
models. We sort the stocks into quintiles based on the resulting residuals for each three-digit SIC industry and report the return
performance of the hedge portfolio, which takes long positions in quintiles with the most favorable new information and short
positions in quintiles with the least favorable new information. Panel A reports the results for the equal-weighted portfolios. Panel
B reports the results for the value-weighted hedge portfolios. We report the mean monthly returns and risk-adjusted returns
(alpha) to the portfolios with the corresponding t statistics in the parentheses.
models 1 and 2) but also has significant incremental Transaction Costs and Other
information beyond them (see models 3 and 4 of the Implementation Considerations. In this sec-
panel). These results suggest that (i) analysts do not tion, we discuss the effect of transaction costs and
fully incorporate financial statement information into other implementation issues. Our investment strategy
their forecasts and (ii) investors can benefit from the is based on the ML earnings forecasts derived from
ML forecasts even if they already have access to annual financial statement data and has relatively low
analysts’ forecasts, as the ML forecasts contain a turnover. Untabulated analyses show that for the
considerable amount incremental information beyond value-weighted portfolio, the monthly portfolio turn-
analyst forecasts.15 over is less than 30% most of the time and averages
86
Fundamental Analysis via Machine Learning
Notes: This figure plots the cumulative (log) returns to the value-weighted quintile portfolios sorted on the new information compo-
nent of the ML forecasts (i.e., regression residual of the ML forecast on the extant forecasts) and the hedge portfolio, which is the
difference between the two extreme quintiles. The left vertical axis is for the cumulative log returns to the five quintile portfolios
while the right vertical axis is for the cumulative log return to the hedge portfolio (the solid red line).
Notes: This figure plots the average feature importance extracted from the fitted models of the random forest (RF) that we train
with data from the 1975–2019 period. The higher the importance score, the more important the feature is. To facilitate representa-
tion, we set the maximum of the y axis at 0.05, while the average feature importance values for earnings (“E”) is 0.7932.
Figure 4. Accumulated Local Effects (ALE) of the Top Five Most Influential Features
Notes: (A) Accumulated local effects (ALE) plots of the random forest (RF) model for 1975. (B) Accumulated local effects (ALE) plots
of the random forest (RF) model for 1995. (C) Accumulated local effects (ALE) plots of the random forest (RF) model for 2015.
88
Fundamental Analysis via Machine Learning
Notes: This table examines the importance of nonlinear and interaction effects in explaining the superior information content in
the machine learning (ML) forecasts. Panel A presents the Fama–MacBeth regression of earnings changes (ECH) on the forecasted
earnings changes (FECH) of both the linear and nonlinear composite forecasts, without (Model 1) and with (Model 2), controlling
for the FECHs of the extant models. Panel B presents the Fama–MacBeth regression of ECH on the nonlinear component of the
ML forecasts (i.e., the residual from the annual cross-sectional regression of the ML forecasts on the 60 input features) and the
FECH of the fitted value from the regression, without (Model 1) and with (Model 2), controlling for the FECHs of the extant mod-
els. All independent variables are standardized to have a zero mean and unit variance each year. All earnings changes are scaled
by the market equity at the end of three months after the end of the last fiscal year. The table presents the average coefficients,
along with the corresponding Fama–MacBeth t statistics in brackets and the average adjusted R-squares. The subscripts are omit-
ted for brevity.
approximately 28.5%. Using an aggressive estimate information, and forecasting quarterly earnings, etc.
of the round-trip cost of 50 bps, the annualized We leave these further refinements to interested
return would be reduced by approximately 7 bps per readers.
month (¼ 50 bps / 2 28.5%).16 However, for retail
investors who trade aggressively with market orders
and buy (sell) stocks at the ask (bid) prices, the aver-
age monthly return to the value-weighted strategies
Conclusions
would drop substantially to about 20 bps, with a
Our exploration of ML in the context of fundamental
Sharpe ratio of 0.25.
analysis, particularly in forecasting corporate earn-
However, it is worth pointing out that simple decile-/ ings, yields enlightening results. ML models outstrip
quintile-ranked portfolios are far from an efficient contemporary earnings prediction models from the
way to implement a quantitative strategy. literature in terms of forecast accuracy and the rich-
Sophisticated investors mostly utilize risk and trans- ness of information provided. This enhanced perfor-
action cost models together with portfolio optimizers mance of ML models is primarily due to their ability
to construct more (risk and t cost) efficient portfolios, to unearth new economically significant information
which tend to yield much better performance (e.g., from the publicly available financial statement data.
Sivaramakrishnan, Brown, and Kasturi 2018). They achieve this by identifying the key predictors
Furthermore, investors may also be able to improve and capturing complex nonlinear relationships that
their strategy performance by using more advanced traditional models and methodologies might overlook
ML technology, incorporating non-accounting or be unable to process effectively.
Table 8. Comparison between the Machine Learning Forecasts and Analyst Forecasts
A. Forecast accuracy comparison
Mean absolute forecast errors*100 Median absolute forecast errors*100
t+1 t+2 t+3 t+1 t+2 t+3
Notes: This table compares the composite forecasts based on the machine learning model (ML) and analyst consensus forecasts.
Panel A reports the time-series average of the mean and median absolute forecast errors of the one- to three-year-ahead earnings
forecasts for the ML model and the analyst consensus forecasts issued in the third month after the last fiscal year end. Panel B
reports the Fama–MacBeth regression results. All independent variables are standardized to have a zero mean and unit variance
each year. The panel presents the average coefficients, along with the Fama–MacBeth t statistics in brackets and the average
adjusted R-square.
The new information that ML models bring to light is (ii) ML forecasts contain significant incremental
not merely statistically significant; it bears consider- information beyond analyst consensus forecasts
able economic value for investors. The residuals from even if analysts have access to all the financial
ML forecasts—representing information not captured statements used in ML models (and much more),
by existing models—show a robust predictive rela- suggesting that analysts fail to fully incorporate
tionship with future stock returns. This finding indi- the information in key financial statement items
cates that the market has not fully priced in the into their forecasts.
information revealed by ML models, thus allowing
investors using ML-derived insights to potentially The overarching conclusion is that ML technology
achieve superior returns. holds considerable promise in refining investment
decision-making processes. By more effectively
Additionally, the comparison between ML forecasts extracting and utilizing value-relevant information
and consensus analyst forecasts reveals notewor- from financial statements, ML can play a transfor-
thy findings: (i) ML forecasts are as accurate as mative role in enhancing the accuracy and efficacy
consensus analyst forecasts over a one-year fore- of earnings forecasts and, by extension, in the
cast horizon and are significantly more accurate broader domain of financial analysis and
than them over longer forecast horizons and investment.
90
Fundamental Analysis via Machine Learning
Appendix 1:
Models and Variable Definitions
A: Summary of extant models and variable definitions
Extant Models
Variables Definition
continued
92
Fundamental Analysis via Machine Learning
Dependent variable
EXRET12Mtþ1 One-year-ahead excess return, computed as the 12-month cumulative return less that of
the risk-free rate, starting from the fourth month after the end of the last fiscal year
Main independent variables
ML RESDt The new information uncovered by machine learning models, estimated as the residual by
regressing the one-year-ahead machine learning-based forecasts on the one-year-ahead
earnings forecasts from the RW model and the five extant models in year t
Controls
Betat Annual market beta using the market model calculated in WRDS
SIZEt Logarithm of market capitalization at the end of the third month after the end of the
last fiscal year
BMt Book-to-market ratio, calculated as the book value of equity divided by the market equity
at the end of three months after the end of the last fiscal year
MOMt Momentum calculated as the cumulative return during the 11-month period starting
12 months ago
ACCt Accruals scaled by total assets
ST Reversalt Return in the prior month
LT Reversalt Return in the prior five years, excluding the prior year
SUEt Quarterly unexpected earnings surprise based on a rolling seasonal random walk model
(Livnat and Mendenhall 2006)
Gross Profitabilityt (sale (sale) − cost of goods sold (cogs)) / total assets (at)
Earnings yieldt Earnings to price (Penman et al. 2015)
ROEt Earnings (ib − spi) divided by common equity (ceq)
INVt Growth rate of total assets (att =att−1 − 1)
IndustryFE Three-digit SIC industry fixed effects
Appendix 2:
Tuning of Hyperparameters for the Machine Learning Models
Editor's Note
Submitted 28 September 2022
Accepted 30 January 2024 by William N. Goetzmann
Notes
1. For example, the law of diminishing returns predicts a 3. Subsequent research finds that the RW model performs
nonlinear relationship between capital investment and well even when compared with analyst forecasts. For
future earnings. Prior literature has also shown that the example, Bradshaw et al. (2012) find that analysts’
relationship between current and future earnings is earnings forecasts are not economically more accurate
nonlinear and varies across firms (e.g., Lev 1983; Freeman than the naïve RW forecasts, and those for longer
and Tse 1992; Baginski et al. 1999; Chen and Zhang horizons are even less accurate than the naïve RW
2007). forecasts.
2. The financial statement items include (i) earnings and 4. There was also a flurry of early studies on fundamental
their major components; (ii) other income statement items analysis using financial ratios (see reviews by Kothari [2001]
that prior studies have shown to produce long-term and Richardson et al. [2010]). However, most of these
benefits or be associated with earnings manipulation; (iii) studies examine in-sample associations and are subject to
balance sheet accounts that are important for earnings concerns of the in-sample identification of predictive variables
prediction; and (iv) operating cash flows, which will be (p. 424 of Richardson et al. 2010). They provide little (if any)
discussed in greater detail in Section 3. evidence on the accuracy or informativeness of out-of-
94
Fundamental Analysis via Machine Learning
sample forecasts. Furthermore, these studies focus on the 11. It is possible that the results might not be robust if we control
sample period before the 1990s. It is unclear whether their for a large number of other characteristics that the literature
conclusions still hold in recent years. has discovered. Similarly, although we have controlled for all
factors in the Fama–French database in the subsequent time-
5. We limit the input variables to the financial statement data series test, the risk-adjusted returns may not be significant if
because we are interested in the decision usefulness of we control for other additional factors that are theoretically or
fundamental analysis using financial statement information. empirically related to the new predictor. We thank one of the
Furthermore, we compare the ML models with the extant referees for pointing this out.
models, most of which also only use financial statement
items (with the exception of the SO model). 12. The left vertical axis is for the cumulative log returns to
the five quintile portfolios, while the right vertical axis is
6. These ML algorithms have been widely adopted in the for the cumulative log return to the hedge portfolio. The
literature. For the sake of brevity, we do not discuss the two axes have different scales. It is worth pointing out
technical details about them. Interested readers can refer that the hedge portfolio has a lower volatility than the
to Gu et al. (2020) for these details. We select these highest quintile. The annualized volatility of the two
models rather than other more complex ones, which portfolios is 9.5% and 16.3%, respectively.
incorporate time-series dynamics such as Recurrent
Neural Network (RNN), Long Short-Term Memory (LSTM), 13. The plots for the GBR models are similar but untabulated
or reinforcement learning because all benchmark extant for the sake of brevity.
models are cross-sectional. To the extent that more
14. The nonlinear models also accommodate the interaction
sophisticated models may be able to produce better
effects between predictors. Untabulated analyses show
forecasts and that other inputs such as industry and that the interaction effects between the following
macroeconomic variables might contain additional economically related pairs contribute the most to the
information, our results should be interpreted as a lower explanatory power of the GBR model: the change in sales
bound that ML can attain in earnings forecasting tasks. revenue and change in the cost of goods sold, change in
short-term debts and change in total current liabilities,
7. Ideally, we would like to use the point-in-time data to
sales revenue and accounts payable, depreciation and
avoid any subsequent restatements in accounting
amortization expense and net property, plant and
numbers. However, we do not have access to the data.
equipment, and cost of goods sold and inventories. The
We therefore follow related studies (e.g., Hou et al. 2012;
finding that the ML models pick up the interaction
So 2013; Li and Mohanram 2014) and use the Compustat
between income statement items and the corresponding
fundamental annual file. gross accrual items (e.g., depreciation and amortization
(DA) and net PP&E, COGS and inventories) resonates
8. For each round of training, cross-validation (CV) and
remarkably well with the recent call for research by
prediction, we normalize all independent variables to [0,1] for
Dichev (2020) on the role of gross accruals in determining
the training set. We also transform the independent variables
earnings quality.
in the prediction set using the same scaler obtained from the
training set. Furthermore, to alleviate the influence of outliers, 15. In untabulated analyses, we find that ML forecasts are
we winsorize the input features (i.e., per-share accounting also a useful benchmark to assess the ex-ante bias in
numbers) at the top and bottom 1% for the training sample. analyst forecasts. When analysts’ forecasts are
substantially higher (lower) than the ML forecasts, these
9. We also conduct robustness tests and compare the
forecasts tend to be overly optimistic (pessimistic) than
forecast accuracy for earnings scaled by total assets and
the actual earnings. Furthermore, these stocks tend to
shares outstanding (i.e., EPS), and the inference remains have significantly negative (positive) returns over
the same. Furthermore, we partition the sample into subsequent periods.
subgroups based on the market cap, and the results still
show that the ML forecast is significantly more accurate 16. The literature has yet to arrive at a consensus on the level
than the RW and extant model forecasts in all subgroups. of trading costs. While Novy-Marx and Velikov (2016)
The results are untabulated but available upon request. suggest that “[r]ound-trip transaction costs for typical value-
weighted strategies average in excess of 50 basis points
10. Note that this is a test of two joint hypotheses: (i) the (bps),” Frazzini et al. (2018) argue that this estimate is “an
new information uncovered by the ML model is order of magnitude larger than [what] our model or live
economically significant and (ii) the market does not fully costs suggest.” They suggest that for a large institutional
understand the new information. This test also sheds light trader, “the effective bid-ask spread across all trades
on the potential decision usefulness of fundamental averages less than 0.015% per year,” and the market impact
analysis and financial information. As Fama (1965) is just under nine basis points on average for all trades
suggests, fundamental analysis is only of value if it completed within a day. The impact of the transaction cost
provides new information not yet fully priced. Thus, this would be much lower if we use Frazzini et al.’s estimate.
analysis also sheds light on whether fundamental analysis
is useful for investors.
References
Apley, D. W., and J. Zhu. 2020. “Visualizing the Effects of Journal of the Royal Statistical Society Series B: Statistical
Predictor Variables in Black Box Supervised Learning Models.” Methodology 82 (4): 1059–1086. doi:10.1111/rssb.12377.
Baber, W. R., S. H. Kang, and Y. Li. 2011. “Modeling Chen, P. F., and G. Zhang. 2007. “How Do Accounting
Discretionary Accrual Reversal and the Balance Sheet as an Variables Explain Stock Price Movements? Theory and
Earnings Management Constraint.” The Accounting Review 86 Evidence.” Journal of Accounting and Economics 43 (2-3): 219–
(4): 1189–1212. doi:10.2308/accr-10037. 244. doi:10.1016/j.jacceco.2007.01.001.
Baginski, S. P., K. S. Lorek, G. L. Willinger, and B. C. Branson. Chen, S., B. Miao, and T. Shevlin. 2015. “A New Measure of
1999. “The Relationship between Economic Characteristics and Disclosure Quality: The Level of Disaggregation of Accounting
Alternative Annual Earnings Persistence Measures.” The Data in Annual Reports.” Journal of Accounting Research 53 (5):
Accounting Review 74 (1): 105–120. doi:10.2308/accr.1999.74. 1017–1054. doi:10.1111/1475-679X.12094.
1.105. Chen, X., Y. H. Cho, Y. Dou, and B. Lev. 2022. “Predicting
Ball, R., J. Gerakos, J. T. Linnainmaa, and V. Nikolaev. 2020. Future Earnings Changes Using Machine Learning and Detailed
“Earnings, Retained Earnings, and Book-to-Market in the Cross Financial Data.” Journal of Accounting Research 60 (2): 467–515.
Section of Expected Returns.” Journal of Financial Economics doi:10.1111/1475-679X.12429.
135 (1): 231–254. doi:10.1016/j.jfineco.2019.05.013. Dichev, I. D. 2020. “Fifty Years of Capital Markets Research in
Ball, R., and R. Watts. 1972. “Some Time Series Properties of Accounting: Achievements so Far and Opportunities Ahead.”
Accounting Income.” The Journal of Finance 27 (3): 663–681. China Journal of Accounting Research 13 (3): 237–249. doi:10.
doi:10.1111/j.1540-6261.1972.tb00991.x. 1016/j.cjar.2020.07.005.
Bao, Y., B. Ke, B. Li, Y. J. Yu, and J. Zhang. 2020. “Detecting Ding, K., B. Lev, X. Peng, T. Sun, and M. A. Vasarhelyi. 2020.
Accounting Fraud in Publicly Traded US Firms Using a Machine “Machine Learning Improves Accounting Estimates: Evidence
from Insurance Payments.” Review of Accounting Studies 25 (3):
Learning Approach.” Journal of Accounting Research 58 (1):
1098–1134. doi:10.1007/s11142-020-09546-9.
199–235. doi:10.1111/1475-679X.12292.
Dou, W. W., Y. Ji, D. Reibstein, and W. Wu. 2021. “Inalienable
Barnea, A., J. Ronen, and S. Sadan. 1976. “Classificatory
Customer Capital, Corporate Liquidity, and Stock Returns.” The
Smoothing of Income with Extraordinary Items.” The Accounting
Journal of Finance 76 (1): 211–265. doi:10.1111/jofi.12960.
Review 51 (1): 110–122.
Easton, P. D., P. Kelly, and A. Neuhierl. 2018. “Beating a
Bartram, S. M., and M. Grinblatt. 2018. “Agnostic Fundamental
Random Walk.” Available at SSRN: https://ssrn.com/abstract=
Analysis Works.” Journal of Financial Economics 128 (1): 125–
3040354 or doi:10.2139/ssrn.3040354.
147. doi:10.1016/j.jfineco.2016.11.008.
Fairfield, P. M., R. J. Sweeney, and T. L. Yohn. 1996.
Barua, A., S. Lin, and A. M. Sbaraglia. 2010. “Earnings
“Accounting Classification and the Predictive Content of
Management Using Discontinued Operations.” The Accounting
Earnings.” The Accounting Review 71 (3): 337–355.
Review 85 (5): 1485–1509. doi:10.2308/accr.2010.85.5.1485.
Fama, E. F. 1965. “Random Walks in Stock Market Prices.”
Bertomeu, J., E. Cheynel, E. Floyd, and W. Pan. 2021. “Using Financial Analysts Journal 21 (5): 55–59. doi:10.2469/faj.v21.n5.
Machine Learning to Detect Misstatements.” Review of 55.
Accounting Studies 26 (2): 468–519. doi:10.1007/s11142-020-
09563-8. Fama, E. F., and K. R. French. 2000. “Forecasting Profitability
and Earnings.” The Journal of Business 73 (2): 161–175. doi:10.
Bradshaw, M. T., M. S. Drake, J. N. Myers, and L. A. Myers. 1086/209638.
2012. “A Re-Examination of Analysts’ Superiority over Time-
Series Forecasts of Annual Earnings.” Review of Accounting Fama, E. F., and K. R. French. 2006. “Profitability, Investment
Studies 17 (4): 944–968. doi:10.1007/s11142-012-9185-8. and Average Returns.” Journal of Financial Economics 82 (3):
491–518. doi:10.1016/j.jfineco.2005.09.009.
Brown, L. D. 1993. “Earnings Forecasting Research: Its
Implications for Capital Markets Research.” International Journal Fama, E. F., and K. R. French. 2015. “A Five-Factor Asset
of Forecasting 9 (3): 295–320. doi:10.1016/0169- Pricing Model.” Journal of Financial Economics 116 (1): 1–22.
2070(93)90023-G. doi:10.1016/j.jfineco.2014.10.010.
Bublitz, B., and M. Ettredge. 1989. “The Information in Feltham, G. A., and J. A. Ohlson. 1995. “Valuation and Clean
Discretionary Outlays: Advertising, Research, and Surplus Accounting for Operating and Financial Activities.”
Development.” The Accounting Review 64 (1): 108–124. Contemporary Accounting Research 11 (2): 689–731. doi:10.
1111/j.1911-3846.1995.tb00462.x.
Cao, S., W. Jiang, J. L. Wang, and B. Yang. 2024. “From Man
vs. Machine to Man þ Machine: The Art and AI of Stock Feltham, G. A., and J. A. Ohlson. 1996. “Uncertainty Resolution
and the Theory of Depreciation Measurement.” Journal of
Analyses.” Journal of Financial Economics. Forthcoming.
Accounting Research 34 (2): 209–234. doi:10.2307/2491500.
Call, A. C., M. Hewitt, T. Shevlin, and T. L. Yohn. 2016. “Firm-
Frazzini, A., R. Israel, and T. J. Moskowitz. 2018. “Trading
Specific Estimates of Differential Persistence and Their
Costs.” Available at SSRN 3229719.
Incremental Usefulness for Forecasting and Valuation.” The
Accounting Review 91 (3): 811–833. doi:10.2308/accr-51233. Freeman, R. N., and S. Y. Tse. 1992. “A Nonlinear Model of
Security Price Responses to Unexpected Earnings.” Journal of
Chan, L. K., J. Lakonishok, and T. Sougiannis. 2001. “The Stock
Accounting Research 30 (2): 185–209. doi:10.2307/2491123.
Market Valuation of Research and Development Expenditures.”
The Journal of Finance 56 (6): 2431–2456. doi:10.1111/0022- Gerakos, J. J., and R. Gramacy. 2013. “Regression-Based
1082.00411. Earnings Forecasts.” Chicago Booth Research Paper, 12–26.
96
Fundamental Analysis via Machine Learning
Thomas, J., and F. X. Zhang. 2014. “Valuation of Tax Expense.” Science 60 (1): 227–245.
Review of Accounting Studies 19 (4): 1436–1467. doi:10.1007/ doi:10.1287/mnsc.2013.1748.
s11142-013-9274-3.
Watts, R. L., and R. W. Leftwich. 1977. “The Time Series of
Vitorino, M. A. 2014. “Understanding the Effect of Annual Accounting Earnings.” Journal of Accounting Research 15
Advertising on Stock Returns and Firm Value: Theory (2): 253–271. doi:10.2307/2490352.
and Evidence from a Structural Model.” Management
98