Wednesday, February 12, 2020

A Simple Equity Volatility Estimator

While short-term asset returns are unpredictable, volatility is highly predictable theoretically and practically. The VIX index is a forward-looking estimate of volatility based on index option prices. Though introduced in 1992 it has been calculated back to 1986, because when released they wanted people to understand how it behaved.



Given the conditional volatility varies significantly over time it is very useful to generate a VIX proxy for cases where one does not have VIX prices. This includes pre-1986 US, countries that do not have VIX indices, and when trying to estimate the end-of-day VIX. This latter problem is subtle but important because historical closing VIX prices are taken from the 4:15 ET in the US while the market closes at 4:00, and so using VIX prices for daily strategies can generate a subtle bias when used in daily trading strategies.

First, we must understand the VIX because there's some subtlety here. It is really not a volatility estimate, but a variance estimate presented as volatility. VIX is calculated as the square root of the par SP500 variance swap with a 30-day term, multiplied by 100 and annualized (ie, 19.34 means 19.34% annualized). That is, it would be the strike volatility in a 30-day variance swap at inception:


On September 22, 2003, the CBOE changed the VIX calculation in two ways. First, they began to use SP500 rather than SP100 option prices. This lowered the volatility to about 97% of its old vol level because the SP500 is more diversified and less volatile. Second, instead of just taking the average volatilities of nearby puts and calls, they used explicit call and put prices in a more rigorous way. This is because a variance swap's replicating portfolio consists of the following weights for out-of-the-money puts and calls. 



The VIX futures started trading in 2004, and options on these futures started in 2008. Liquid markets make index prices more efficient because nothing motivates like the profit motive (eg, regardless of your preferences, more money will help you achieve them). The net result is that one should use data since 2004 when analyzing the VIX even though there is data back to 1986 (which, is still useful for some applications).

One can see that the old VIX index was significantly more biased upwards than after these changes. This implies abnormal volatility trading strategies prior to 2004 if you assumed the VIX was a true par variance swap price.  Now, there should be a slight positive bias in the VIX due to the variance premium, where shorting variance generates a positive return over time. Personally, I think this variance premium is really a consequence of the equity premium, in that short variance strategies have very strong correlations with being long the market. That is, the variance premium is not an independent priced risk factor, just a consequence of the equity premium given its high beta. 

VIX
Var(VIX)
Actual Vol
Actual Variance
1986-2003
20.91
4.96
17.67
3.12
2004-2019
18.20
4.05
17.99
3.24
VIX/ActVol
Var(VIX)/ActVar
1986-2003
1.18
1.59
2004-2019
1.01
1.25


As a liquid market price, the VIX is a good benchmark for any equity volatility model. The most common academic way to estimate volatility is some variant of a Garch(1,1) model, which is like an ARMA model of variance:


The problem is that you need to estimate the parameters {w, α, β} using a maximum likelihood function, which is non-trivial in spreadsheets. Further, there is little intuition as to what these parameters should be. We know that α plus β should be less than 1, and that the unconditional variance is w/(1-α-β). That still leaves the model highly sensitive to slight deviations, in that if you misestimate them you often get absurd extrapolations.

For daily data, a simple exponentially weighted moving average (EWMA) version of Garch(1,1) works pretty well, with w=0, α=0.05, and  β=0.95. This generates a decent R2 with the day and month-ahead variance.

EWMA Vol Estimator on Daily Data


Alas, this has two problems. First, there is a predictable bias in the EWMA because it ignores mean reversion in volatility. Garch models address this via the intercept term, but as mentioned it is tricky to estimate and creates non-intuitive and highly sensitive parameters. We can see this bias by sorting the data by VIX into deciles, and take the average EWMA, where the relative difference in the VIX and the EWMA increases the lower the EWMA. As this bias is fairly linear, we can correct for this via the function 



US data sorted into VIX deciles
2004-2019

VIX EWMA EWMA*
Low 11.1 8.1 10.6
2 12.7 10.2 13.2
3 14.0 11.4 14.7
4 15.6 12.4 15.8
5 17.1 13.4 16.9
6 18.7 15.0 18.7
7 20.7 17.3 21.2
8 23.0 19.1 23.1
9 25.9 21.3 25.3
High 40.3 39.5 39.7

Secondly, there's the correlation between returns and VIX movements that are asymmetric: positive index returns decrease implied volatility while negative movements increase implied volatility. Further, the strength of the relationship is asymmetric, in that down moves are twice as strong as up moves.  Here are the contemporaneous changes in the VIX and SPY using daily returns since 2003. I sorted by SPX return into 20 buckets and took the average SPX and VIX percent changes.



An EWMA would generate a symmetric U-pattern between asset returns and volatility as 0.012 = (-0.01)2,  a huge mismatch with real daily VIX changes.

There are a couple of good reasons for this asymmetric volatility response to price changes. As recessions imply systematic economic problems, there's always a chance that negative news is not just a disappointment, but reveals a flaw in your deepest assumptions (e.g., did you know you don't need 20% down to buy a house anymore?).  This does not happen in commodities because for many of these markets higher prices are correlated with bad news, such as oil shocks or inflation increases. Another problem is that many large-cap companies are built primarily of exponential growth assumptions. Companies like Tesla and Amazon need sustained abnormal growth rates to justify their valuations, so any decline could mean an inflection point back to normal growth, lowering their value by 90%. Again, this has no relevance for commodities.

One can capture this by the following function


For example, if the return was +1%, yesterday's vol is multiplied by 0.975, while if it was down 1%, the adjustment factor is 1.05. While the empirical relation of returns on volatility is not just asymmetric but non-linear (the absolute returns have a diminishing marginal impact), putting in a squared term creates problems as they extrapolate poorly, and so this piecewise linear approximation is applied to make the model more robust.

These two adjustments--one for mean reversion, one for the return-implied volatility correlation--generates the following function for adjusting the simple EWMA: 



The first term captures the volatility-return correlation, the second mean reversion. The term 0.2 adjusts the speed to which our volatility estimate moves towards its long-run target given its current level. I'd like to give this a cool name with Latin roots but given two adjustments it would become German-sized, so I'm just going to call this transformed estimate of the EWMA 'EricVol' for simplicity and clarity. After this transformation, the bias to our vol estimate is diminished:


Vol Estimators sorted by VIX

VIX
EricVol
EWMA
Low
11.1
10.8
8.1
2
12.7
13.6
10.2
3
14.0
15.0
11.4
4
15.6
16.1
12.4
5
17.1
17.2
13.4
6
18.7
19.2
15.0
7
20.7
21.9
17.3
8
23.0
24.0
19.1
9
25.9
26.7
21.3
High
40.3
43.2
39.5

Comparing the daily correlations with the VIX changes, we see EricVol is much more correlated than the simple EWMA, especially in the most volatile times


Daily Correlation with VIX Changes
2004-2019

EWMA
EricVol
2008
29%
82%
Oct-08
-19%
84%
total
37%
75%

As most volatility trading strategies are linear functions of variance, and the VIX itself is really the square root of its true essence, we predict returns squared and square our vol estimates in these equations. 

Regression R2 for predicting forward day-ahead and 21-day ahead variance


VIX
EWMA
EricVol
day-ahead
33.0%
26.9%
34.3%
Month-ahead
61.1%
58.4%
61.8%

If we look at regressions that predict future variance given our estimates, we see EricVol is significantly better than a simple EWMA. While it does slightly better than the VIX, I doubt this generates significant profits trading, say, the VXX, though readers are free to try.  

You can download a spreadsheet with all this data and the model here. You need to have two columns in a spreadsheet because you have to keep a time-series of EWMA and EricVol, which is annoying, but it's much simpler than fitting a Garch model. Most importantly, its parameters are intuitive and stable.  

Tuesday, February 04, 2020

Factor Risk and Return

Factor returns should reflect risk, in that they have traditionally been interpreted as proxies for some kind of risk not measured by beta. The idea is that perhaps what people really care about is whether there will be another oil shock, and nothing matters as much. Stocks that have a high dependence on cheap oil would have more risk than other stocks. In the early 1980s, this was a common hypothesis, though later people would add things like consumption growth and inflation.

Remember that our conception of risk comes from the idea that as utility functions are concave (have decreasing marginal utility), the higher the expected variance in our wealth, the lower the expected utility. Thus, $1 for certain tomorrow is worth more than a payoff with equal probabilities for {$0.50, $1.50}. Because of decreasing marginal utility, one enjoys an extra 50 cents less than one suffers from the 50 cent deficit, so you have to bribe people to make them indifferent to such an alternative, what is called the risk premium.

While most presentations of the CAPM use normal distributions, the initial creators did not simply jump on this assumption blindly because it was convenient, as there was a lot of focus on the best distributional assumption for asset returns.  I took the weekly returns and normalized them by a rolling volatility forecast to capture the heteroskedasticity (otherwise the tails are much fatter). While these deviations from normality are statistically significant, they aren't large. You can see the weekly returns are too frequent at the bottom, offset by the extra number of slightly-above-average returns.

Markowitz considered downside volatility and maximum drawdown among other metrics, and Eugene Fama's dissertation and several subsequent papers investigated the degree to which downside risk outside the normal distribution--aka fat tails--was the big driver. This was all motivated by Benoit Mandelbrot's documentation that many commodity markets had large downside tail returns. If you can get through this paper by Fama (1965), you'll understand why he stopped investigating it: it's boring. [In contrast, the degree to which market crashes overstate 'average' returns is much more interesting, see Barro (2005). But note, this is all about whether the average equity return premium should be reduced by 3%, not a criticism of the CAPM or its spawn.]

You can capture the above deviation from normality in many different ways but they don't add much of any practical value. Levy and Markowitz (1979) documented that the gaussian distribution, for all its faults, is pretty innocuous. All you have to do is increase a person's risk aversion so that they 'feel' the downside risk more strongly, and you capture the effects of fat declines just as you would with some mixed distribution with more parameters. The costs are low, but the benefits are large. Normal distributions are additive (one normal plus another is still normally distributed), which makes statistics a lot easier. A mere 2 parameters describe the entire distribution (all the higher moments are linear functions of the mean and variance). Further, when x is normally distributed, E[ex]=eμ+1/2σ2, which is useful in all sorts of models (eg, exponential utility is of the form 1-e-αx).

I pulled a bunch of portfolio factor data from Ken French's website. It's an awesome resource for anyone into equity risk because it's free and easy to access (no sign-ins or passwords).  For the US I pulled seven factors, the high and low quintiles among the largest 2 size quintiles, for 7 factors: book/market, cashflow/price, earnings/price, volatility, investment (asset growth), prior returns, and accruals.

As a practical matter, most metrics of risk are correlated. Beta and portfolio volatility have incredibly high correlations, so much so they are almost redundant. Minimizing beta or vol is basically the same thing.  The correlation between portfolio volatility and the maximum drawdown is not as high, but still highly significant. To the degree risk is the expected worst-case scenario, volatility/beta should capture a good deal of this.


When we look at the relationship between these risk metrics for various factor portfolios, there is nothing close to a strong correlation. Indeed, the correlation is negative to a first approximation.

For international portfolios, I used book/market, cashflow/price, and earnings/price from several regions. For these, I normalized by subtracting the region-specific volatility and average returns, so that factor performance is comparable.  While risk is no longer negatively correlated with returns, the correlation is weak.


Factor risk premiums do not reflect risk, they explain themselves. Perhaps they are capturing something institutional, like taxes. I came across Phil DeMuth's latest book, The Overtaxed Investor, and he emphasizes the tax advantage of capital gains over dividends. Now if in equilibrium investors have to be indifferent, perhaps the extra 2% earned by high dividend stocks is compensated by the higher tax rate. Transaction costs could be another issue, especially for small-cap stocks, in that it may cost one a couple of percent more to get in and out of these positions, which would not be reflected in end-of-day returns but would your for small-cap investor. Such explanations have nothing to do with risk.

In the latest American Finance Association Presidential Adress, David Hirshleifer presented the following theory by way of an analogy. Moths (supposedly) fly into flames kamikaze-style, as they are hard-wired to fly towards the moon to get their bearings. Clearly, this isn't good for the moths, but rather, a rule-of-thumb that does not extrapolate well, especially after man discovered fire. Perhaps our monkey brains have similar lacunae. For example, many remember dating someone crazy because she was super hot, not because men like crazy girls, but rather, crazy girls do hot things that cause men's limbic system to shut down the cerebral cortex (this is now a meme, but young fools will only learn by experience). Similarly, investors could be attracted to risk, not directly, but rather, the fact that risky stocks have attractive attributes like stories about transforming the workplace, or they have pot-smoking CEOs, or that they specifically focus more on marketing to investors as opposed to consumers.

While I agree with most people that many investors are ignorant and susceptible to biases, these suboptimal decision-makers are a poor explanation for equilibrium asset prices; their bad decision-making is mainly evinced by excessive trading. It doesn't take many smart traders to figure out that if they steel themselves against those risky-stock promoters they can make large risk-adjusted returns. It's not complicated, you don't have to tie yourself against a mast like Odysseus, just don't listen to conference calls or watch CNBC, and instead just crunch datasets where all you know about ABC corp are its objective metrics.  Then buy the wallflower stocks and sell the crazy/hot ones. Arbitrageurs should counteract behavior biases in asset markets with free entry.

Biases, departures from normality, and new factors don't explain why anything reasonably correlated with something related to wealth volatility has no correlation with average returns. This leaves the utility function. With relative risk, there should be no risk premium just as there isn't one. This makes factor investing much less attractive because, without the risk adjustment that accentuates the alpha, you just have the potential for making 2% extra but have to take on more risk to do so (by deviating from the market, creating benchmark risk).  It's not like there are $20 bills lying on the floor, but rather, a bunch of change in a fountain, and you can grab a handful if you don't mind getting soaking wet and mean looks. 

Tuesday, January 28, 2020

Is the Fama-French Model Dead?

When I was in graduate school at Northwestern in the early 90s the hot financial topics were all related to finding and estimating risk factors: Arbitrage Pricing Theory via latent factors  (Connor and Koraczyk 1986), Kalman filter state-space models (eg, Stock and Watson 1989), and method of moment estimators (Lars Hansen 1982). These appealed to central limit theorem proofs, which is the academic dream, as math proofs decisively prove objective meta-knowledge, how to know important things. Learning such methods involved intelligence and hard work, promising an instant barrier to entry, and certainly not from those boorish B-school students who had much better parties but didn't know statistics, let alone measure theory.

Around that same time, Eugene Fama and Ken French were looking at the returns of low and high dividend stocks by simply sorting them and looking at the basic statistics of these portfolios (see here). While Fama was well-respected for his theoretical work on defining efficient markets and his seminal 1973 paper confirming the capital asset pricing model (ie, the CAPM), he seemed like one of those antediluvian economists who typified the 1970s. Indeed, if you read those old papers, you will be amazed by how simple they were--you can understand them on one read--making the many inevitable false-positives glaringly obvious (with hindsight).

But then in 1992 Fama and French published their blockbuster 3-factor model documenting that if you sort stocks first by size and then by beta there is no correlation between beta and stocks. The previous finding was the product of an omitted variable, size. That is, one might look at humans and find hair length inversely correlated with height: longer hair correlates with shorter humans. But if you note women have longer hair than men and control for that, the correlation goes away; hair has nothing to do with height.

Size and value were identified in the late 70s, often in accounting journals. Fama and French's analysis was simple but the results were too strong to dismiss, especially as Fama was not some accounting noob, but rather, someone who understood risk and expected returns. Fama and French hypothesized size and value premiums implied they proxied risks that affect investors but academics can't measure, just as we can see the effects of dark matter but not the dark matter itself.

Still, many researchers were burned by false findings, and value and size were discovered in the same naive way of simply finding a strong pattern and presuming it's real as opposed to an order statistic (see here DeBondt and Thaler's list of anomalies circa 1989). While F-F said they were risk proxies, it seemed ad hoc. Economic theory implies risk premiums in that the sine qua non of economics, the utility function, implies a risk premium in an if and only if relation (which implies B ⇒ A and A ⇒ B). [my take: the error here is applying it to an individual's wealth, it works great for ice cream and shoes.] Add to this 20 years where economists learned that the CAPM model had been proven ('a peer-reviewed fact,' they would say today), and you have a strong bias that is not going to be discarded after some simple analysis. You can identify such researchers in the past--many still plying their trade--by how they would introduce their findings via moment conditions (see Harvey and Siddique, 1999), or factor models that identify loadings, factor returns, and the price of risk, λ (see Ang et al 2006).

Yet size and value investing became popular. Dimensional Fund Advisers created a highly successful and influential quant shop, and size and value became ubiquitous as either fund signatures or risk factors, depending on the application. There were dark times, such as when the size factor faltered in the late 80s and mid-90s, or when value stocks got clobbered in the tech bubble, but they always came back: they were risky! The rigorous methodology of latent factors models, Kalman filters, and GMM, meanwhile, produced nothing practical in finance.

Simple models win in the long run, but the short run is always dominated by the obfuscators. Socrates and Jesus criticized the sophists and Pharisees, respectively, because these official intellectuals were sanctimonious hair-splitters, and missed the forest for the trees. Christian apologists were the foremost Medieval scholars, and though now apologist is a derogatory term for a dogmatic thinker, the key is that sophisticated reasoning has always been used to mask dogmatic thinking. Sophisticates use their superior knowledge to out-argue, not find the truth. Complex arguments that are long and hard to follow, and have the support of the most other intellectuals are impossible to rebut. No one is going to spend a lot of time understanding such arguments if they think it's all a waste of time, so everyone who could argue with them was a fellow-believer (eg, I originally wanted to be a macroeconomist, but quickly discerned it was pointless and so today I cannot give a thorough critique of dominant macro models).

Simple models are better at finding the truth because they are easier for others to replicate and test, and this is the best evidence of something that is true and important, because things that work tend to be copied over time. They also expose dopey reasoning and overfitting, because readers can understand them. Complexity does not overcome overfitting, it just hides it better. GMM estimators are rarely criticized for overfitting because it's too hard to tell, and when no one is getting caught for the world's oldest statistical sin you can be sure it is rampant (eg, when the UFC had a steroid problem no one was getting caught).

More importantly, there just are not that many really profound financial ideas one would call complex. Off the top of my head, I think about how volatility is proportional to the square root of time, Black-Scholes, or the convexity adjustment in forwards vs. futures for interest-rates. Yet, these can all be communicated in one sitting to a person who understands basic statistics. That's not true for the fancy methods above.

While the Fama-French approach thrived due to its simplicity, we are now in another dark time. Specifically, the 5-factor model introduced by Fama-French in 2015 has reported the following  monthly return averages for their factors (excluding the uncontroversial market risk premium):


Since publication, however, the results look like this (note: they should be positive):



Unlike prior dark times, we now also have a boatload of factors in addition to Fama and French's 5.  Overfitting is a perennial problem. Unlike 'big data' used by Amazon and Netflix, factor data is orders of magnitude smaller and grows very slowly, all the while many thousands of full-time researchers are looking at the same data for new patterns. They are all highly motivated to find new, better factors, and we all know what happens when you torture the data long enough. The simplicity of the 3-factor Fama-French model was not an equilibrium, as this invited many to ignore some factors and create their own (even Fama and French joined in, adding 2 more). There is now just as much tendentious and disingenuous reasoning among the factor promoters as there was with those old opaque and complex methods, we have just gotten used to it.

This especially goes for equity factors applied at hedge funds, which number at least 20 and include orthogonally estimated style factors but also industry factors that no one thinks have a return premium. Argument by authority is ubiquitous among authorities of all stripes, only among scientists, this tactic is masked by having multiple subtle assumptions justified by references to various well-cited peer-reviewed articles, each of which itself does the same thing, making it impossible to refute because no one has that much time. I have never seen a good argument for how these factors help manage risk, but I understand how they are very useful in explaining why returns were down last month. Pretentious and parochial jargon makes it easier for the common (ie, dumb) risk manager, senior executive, fund advisor, etc., to feel valuable: they know something outsiders do not understand, and those outsiders can't prove it's not blather.

No financial researcher agrees on the canonical set of factors or how best to proxy them.  While no one is suggesting Kalman filters, everyone has some subtle twist for their value factor, how to time factors, or their own creation. Failure now has led many to reexamine their factors, worried that they either overfit their models in the past or would be doing so now by changing them. The risk of jumping off a factor just before it rebounds is symmetric to the risk of holding on to one that will join price, accruals, and inflation betas in the history dustbin.


Monday, January 20, 2020

Diversification

I was interested in calculating what the portfolio volatility would be for a portfolio given various correlation assumptions, and also the number of assets. So I took two portfolio of the S&P500 in two very different years: 2008 and 2017. The VIX had one of its highest average levels in 2008, at 31.5, while its lowest in 2017, at 11.0.  Because I'm interested in low vol portfolios, I took the stocks with below-average volatility from the prior year.  The statistics on these 250 stocks were as follows:

Lowest 250 vol stocks in S&P500

2008
2017
AvgVol
60%
19%
AvgCorr
0.48
0.14

Both total volatility and correlation were about 3 times higher in the crisis year of 2008. This is because the total market volatility was so much higher. One could say correlations rise in bad times, but one could also say that mathematically, given a 'one-factor' model (the market), if the market has higher volatility its assets will have a higher correlation.

The equation I presented last week calculates portfolio volatility as follows:
  • Volatility (n assets, correlation c)=sqrt(avgVol^2/n + (n-1)/n*c*avgVol^2)
Here c is the average correlation, so the average of the correlation matrix off the diagonal. Volatility and correlation are taken from the same period here, that in the years 2008 and 2017.

In the equation, most of this diversification benefit happens rather quickly, with 90% of it coming after n=15. Given many people look at multiple factors, not just the market factor, this could make this formula less relevant. So I took all those 250 stocks and created thousands random of sub-portfolios to get an estimate of the actual, or empirical, volatility. 

Comparing this to the formula generated the data in the above chart: they are basically identical. You only see two lines because there are two sets of lines that overlap. In both scenarios, the 10 asset portfolio had insignificantly higher volatility than the 250 asset portfolio. While the correlations are quite different, generating different levels of diversification as a function of the number of assets, they both almost perfectly matched the simple formula.  

Bottom line: 10 assets really is enough.  


Wednesday, January 15, 2020

Helmut Schoeck's Envy

Helmut Schoeck published Envy in 1966. It remains one of the few systematic treatises on the subject and highlights a profound aspect of human nature. Envy is of keen interest to me because ever since I wrote my dissertation on the low volatility effect back in 1994, I have been interested in why. In standard economic theory, individuals are assumed to maximize their utility, which is independent of other's wealth. With this assumption, you get a risk premium, and it should be ubiquitous. The risk premium is actually quite rare, limited to a couple of asset classes (though not within them). For more on envy in economics, see here.

Envy is eternal and ubiquitous, but also taboo. This is because envy is one of those few vices that is neither an excess or deficit of virtue, in that cowardice and recklessness are a deficit and excess of the virtue courage. The literature defending the rationality and value of emotions usually excludes envy as irredeemable.

Thus, no one candidly admits to being envious. We really do not even have a word for sincerely saying, 'I envy John,' in that one can only say this in a way that implies a benign, emulative, or admiring variety of envy, rather than the invidious form.  We do not even like to analyze it very deeply.  For example, Schoeck notes that Herman Melville's Billy Budd is about a good man who arouses the envy of the ship's master-at-arms, motivating many spiteful actions. The book's theme is clearly about envy, yet of the many written analyses of the novel they rarely mention envy, and never focus upon it, but rather highlight evil or injustice

Ordinary language tends to conflate envy and jealousy, yet they are distinct emotions. A man can be jealous of another man because of their mutual affection for a woman, and here the rival is fungible and the beloved is not fungible. In envy, it is the other way around, resenting a rival, not because of a specific woman, but rather, that he can obtain a woman of that type.

Most observers of envy, from Aristotle on, have noted it is felt toward those with whom the subject perceives himself as in competition. We only feel envious of our peers, or who we consider being our peers. Thus, we do not envy a world-class athlete, in that we know we are simply not in that class of people. In medieval times you had peasants, knights, lords, and a king, while in Ancient Greece you had slaves, freedmen, metics, and the upper class. Envy was restricted to one's class and perhaps those in the class above.

Essence of envy

Schoeck presents some theories for the origin of envy, not all convincing. First, from our competition for parental resources. As the runt of the litter gets fewer parental resources, or the weak bird is left to naked fratricide in the presence of the mother, all children wish to dominate their parent's attention out of self-preservation. Yet this is the definition of jealousy, in that the instinct presupposes a non-fungible desired object: parental attention. Another cause mentioned is in keeping society well-functioning. In his view, without envy, people with power would be haughty and overbearing. Yet, simple reciprocal altruism would seem to be at play, in that a successful alpha baboon realizes that if they are too overbearing, their underlings can and will kill him (see Frans de Waal).

Another interesting cause of envy is the ignorance of chance. The most envy-ridden tribes do not possess a concept of luck. Everything is determined by spirits, fortune, or the 'evil eye.' In primitive societies, the success of others is equated with the betrayal of the tribe, because it could have only have occurred by leveraging evil spirits or stealing from others. A certain degree of rationality, or at least a freedom from a magical view of things, is required before the envious man can fully realize that the man he envies does not possess something which, but for the possessor's existence, he, the envious man, might otherwise have.

I see envy as the result of our base, animalistic, status-seeking human nature. When we focus on life on this Earth, we want status, and what that implies, justification. Only a high status can prove, to ourselves and others, that we are good, important, and admired. Status, alas, is in fixed supply. While this constraint can be ameliorated by the concept of having many status silos--football, math, Fortnight--ultimately there are a limited number of positional goods: lakefront property, the 'best' cars or vacations, and most importantly, attractive mates.

Note that our greatest suffering is not caused by physical pain or privation, but rather when the will of another inflicts it. This is because random accidents do not reflect on us as individuals, as they happen by chance to everyone. But suffering caused by another implies inferiority, in that the tormentor had the ability to force their will against ours. Our status is diminished when a malefactor successfully torments us, as it highlights our impotence, and thus, lack of status.

Utopia and Envy

The completely just society where envy is absent is doomed because it is based on the false premise that once there is justice, there will be nothing left for anyone to envy. This situation can never occur because man inevitably discovers something new to envy. Further, by raising envy to the status of virtue (as righteous resentment over injustice), political entrepreneurs are motivated to reveal new inequalities that are to blame for people's impotence. Thus we now have the virtue of intersectionalism, where a transgendered black man is more righteous than a cis-black straight man.

There are so many ways to align individuals at some point we realize the ultimate minority is the individual. Yet, when envy is elevated it inverts all values that compete, resulting in the de-individualization of individuals in the name of equality. Simple utilitarianism implies that justice for many is more important than for any one person, and so what counts are the superficial categories we consider 'diverse.' The individual is unimportant.

Most utopias presume that reducing inequality will eradicate envy, though in fact, the opposite is more certain. Utopias are places where people are so equal that they will only have less than others by conscious choice, the way I have less strawberry ice cream than my neighbor because I prefer chocolate. A sad fact of life is that equality of opportunity comes to grief because individuals do not all have the same ability to make use of their opportunities with comparable success. This is a more bitter experience than one for which one can blame others rather than one's self.

The most egalitarian societies, hunter-gatherer societies, actually have more envy that less egalitarian ones. This isn't obvious because their lack of wealth seems to imply they are not greedy, and thus, selfish like Westerners, but this is just wishful naïve thinking. There is no tribe where social harmony prevails because each man has as little as the next. For primitive egalitarian societies, envy is both a cause and an effect. Schoeck notes that such communities tend to have no conception of luck, and so see any objective relative prosperity as a sign of theft or alignment with evil forces. Schoeck discusses Native American, African, and Polynesian tribes, and note the strong social pressure for anyone better off to be lavish in hospitality and generous with gifts. He knows that if he fails in this, 'the voice of envy will speak out in the whispers of witchcraft' which would make his life very unpleasant.

Envy prevents people from accumulating the wealth needed to create technology, and thus the free time needed to create art, science, literature, and philosophy. A lack of differentiation makes more apparent one's inadequacies, in that some will be more clever, athletic, or brave. When people are made equal in endowments, the only reason for not attaining a higher status is within us, our character, our essence.

In earlier times, the upper classes were different in many objective ways. An elite's risk of starvation was insignificant compared to the peasant, reflected in the fact that they were 10 centimeters taller than a commoner in 18th century England; they could read, sometimes in foreign languages, which implied a great deal of useful education. Today those barriers are gone. Almost everyone knows how to read and write, the basics of mathematics and history, the poor are fatter than the rich, and height disparities between individuals of the same race are absent.

 Equality Increases Envy

Our unprecedented wealth and comfort should have led to less envy, in that we all have access to the knowledge that brings self-actualization and self-esteem, but instead, it has lead to greater envy. The poorly paid op-ed writer of your local newspaper considers themselves not just equal to the rich, but better informed and more articulate, as proven by their many eloquent and insightful essays. Their lack of status motivates envy because they feel just as worthy, as competent, as those who are doing better. David Brooks highlighted this in his book Bobos in Paradise.

Consider our very best intellectuals. One might think they have acquired wisdom, and so a sense of self-actualization and self-esteem. Yet professors in the social sciences are an unhappy lot, reflected in Sayre's law: 'academic politics is the most vicious and bitter form of politics, because the stakes are so low.' This is because only those at the top of each specialty have the power to confer prestige and honor on favored individuals, subjects, and style of discourse. They influence taste, favor certain methodologies, and define the boundaries of their disciplines. To be chief consecrator is the intellectual’s dream, and it only can be apportioned to a few.

The idea that men are intrinsically equal in every admirable trait, combined with the idea that all we need is nourishment, warmth, and education, is simply wrong. Sure, to be engaged in a desperate struggle for food and shelter is to be wholly free from a sense of futility, but such environments are rare for anyone reading this blog. The desire for praise is more imperative than the desire for food and shelter. Praise from the praiseworthy is beyond all rewards, why we are more hurt by the lukewarm approval of men we respect than the contempt of fools.

Social justice justifies not having many things so that others may not have them either. In the politics of envy, all that is needed is to promise the envious the destruction or the confiscation of assets enjoyed by others; behind that, there is no need to promise anything constructive. Thus it really does not matter how the government spends punitive taxation. As shown in primitive societies, when creative ambition is punished, you get stasis, not just in economic productivity, but everything else: art, literature, science.

Resentment is held to be a potentially morally justifiable emotion, whereas envy former is not.  Resentment may come from real injustice, such as slavery. Yet good reasons for resentment are few, edge cases, while much resentment comes from a desire to blame others for our inadequacies. A passionate obsession with inequality is an attempt to compensate for a lack of meaning in one's own life, and you can never get enough of what you don't need to make you happy.

It is easy to lie to ourselves about our motivations, creating a bad equilibrium, as poor hunter-gatherer societies show. These are not idyllic communes free of self-interest, but rather, ignorant and envious cultures. Social flourishing is most fruitful when envious considerations unencumber man's creative faculties. Envy is something that should be discouraged. Pride is a sense of worth derived from the perceived appreciation of others, while self-esteem derives from the belief that we are profoundly appreciated, as a unique individual, by those we profoundly appreciate. Note that such a feeling comes out of a relationship of individuals, not some aggregate metric of status.

While there is nothing wrong with having wealth or wanting more, this should be of secondary importance, and wisdom is all about priorities. We should be encouraged to appreciate excellence in others to build excellence in ourselves. This takes faith, in that we have to believe our excellence in character, even if not reflected in the current status hierarchy, is appreciated by someone we admire, if not now, then in the future.