April 5, 2011
“Deep sea oil, dirty coal mining, obsolete nuclear plants, arms trade, human rights abuses – your bank could be financing environmentally and socially destructive businesses. It doesn’t have to be that way. In July 2011, the European Commissioner for banking, Michel Barnier, will publish a proposal to implement the new “Basel III” rules for banks into European law. These rules aim to make the banking system more robust and stable. Tell Michael Barnier to include sustainability criteria in the proposal – to encourage banks to reconsider dangerous investments and to invest more into sustainable businesses, such as renewable energy producers and social entrepreneurs. Sign the petition now!” (from makefinancework.org April 2011)
October 11, 2010
Reading through the New York Times and the Wall Street Journal, I was struck by a glaring omission. While the public has been mesmerized by currency wars and mortgage moratoriums, along with the usual sex, drugs and rock-n-roll, financial service reform has been largely forgotten. Although a few observers continue to follow the fate of the toothless financial reforms passed a few months ago, Wall Street has returned to business as usual.
Unfortunately, business as usual remains extremely dangerous. The financial system today relies upon a volatile mixture of leveraged finance and socially-driven irrationalities.
Leveraged finance is scary enough, illustrated by a recent conversation I had with a salesperson working for a major prime broker. For those of you who haven’t heard the term before, a prime broker lends securities and money to hedge funds, allowing some to invest more than $30 for every $1 they actually possess. Where does this money actually come from? Despite working at the epicenter of leverage finance, the salesperson seemed to have little idea.
What institutional investors do with borrowed money is even scarier. Simon, Millo, Kellard and Ofer (2010) find that hedge fund managers experienced groupthink in one spectacular financial episode, VW-Porsche. Being over-embedded with one another, one powerful group of managers talked each other into a “consensus trade”. Later, they collectively refused to heed warnings that the trade was becoming dangerous. When this trade inevitably exploded, the hedge fund managers stampeded out with their billions of dollars, briefly creating a spectacular bubble. This episode is consistent with my own research, which shows a follow-the-leader pattern amongst hedge fund managers. Not only are institutional investors gossipy and panicky, but they also imitate the most prestigious investors (e.g. the Tiger Cubs) using their social ties. Statistical analyses suggest that such imitation may occur despite imitators systematically harming themselves.
These cases illustrate why the social studies of finance (SSF) are so important. When socially-mediated irrationalities affect people who control hundreds of billions of leveraged dollars, they could very easily create financial bubbles and crashes impacting the real economy. Understanding these socialized irrationalities remains our best defense against future bubbles and collapses, an eventuality as long as “business as usual” continues.
The credit crisis has imposed on Americans a crash course on the risks of financial models. If derivatives, as Warren Buffet famously put it, are “financial weapons of mass destruction,” models are now seen as the nuclear physics that gave rise to the bomb — powerful, incomprehensible and potentially lethal. Given their dangers, what should Wall Street do with its models?
At one extreme, skeptics have attacked models for their unrealism, lack of transparency, and limited accountability. Models, they charge, are black boxes that even expert users fail to understand. Models become dangerously inaccurate when the world changes. And whenever a bad model fails, it is all too easy for traders to conjure up the “perfect storm” excuse. Wall Street, the skeptics conclude, needs to curtail its addiction to models.
At the other extreme, academics in finance and Wall Street practitioners dismiss the backlash as barking up the wrong tree. Models certainly produce the wrong results when fed the wrong assumptions. But the real culprit in this case is not the model, but the over optimistic trader in his greedy quest for the bonus. Paraphrasing the National Rifle Association (“guns don’t kill people, people kill people”), defenders of models place the blame with bad incentives: “models don’t kill banks,” we hear them saying; “bankers kill banks.” To the proponents of modeling, then, the crisis underscores the need for yet more calculations. That is, for bigger and better models.
Does Wall Street need more models or less models? We see this as a false choice. The debate, in our view, needs to shift from the models themselves to the organization of modeling. We have identified a set of organizational procedures, which we call “reflexive modeling,” that lead to superior financial models.
Consider, first, what a financial model ultimately is. Whether as an equation, an algorithm or a fancy Excel spreadsheet, a financial model is no more than a perspective, a point of view about the value of a security. Models are powerful: they reveal profit opportunities that are invisible to mom-and-pop investors. But there’s a catch: they do not always work. Because stock prices are the outcome of human decisions, financial models do not actually work like the iron law of Newtonian gravity.
Models, then, pose a paradox. They hold the key to extraordinary profits, but can inflict destructive losses on a bank. Because a model entails a complex perspective on issues that are typically fuzzy and ambiguous, they can lock traders into a mistaken view of the world, leading to billionaire losses. Can banks reap the benefits of models while avoiding their accompanying dangers?
Our research suggests how. We conducted a sociological study of a derivatives trading room at a large bank on Wall Street. The bank, which remained anonymous in our study, reaped extraordinary profits from its models, but emerged unscathed from the credit crisis. For three years, we were the proverbial fly on the wall, observing Wall Street traders with the same ethnographic techniques that anthropologists used to understand tribesmen in the South Pacific (The study can be downloaded at http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1285054).
The key to outstanding trades, we found, lies outside the models. Instead, it is a matter of culture, organizational design and leadership.
The bank that we observed introduced reflexivity in every aspect of its organization. From junior traders to their supervisors, everyone at the bank was ready to question their own assumptions, listen for dissonant cues, and respect diverse opinions.
How? As many have already suggested, individuals certainly matter. The bank hired people with a healthy dose of humility and an appreciation for the limits of their smarts. This often meant older traders rather than younger hotshots.
But the key to the bank’s reflexiveness did not just lie in individuals. By reflexiveness we don’t mean super-intelligent traders engaged in some heroic mental feat – splitting and twisting their minds back on themselves like some intellectual variant of a contortionist. Reflexivity is a property of organizations.
The architecture of the bank, for instance, was crucial. The open-plan trading room grouped different trading strategies in the same shared space. Each desk focused on a single model, developing a specialized expertise in certain aspect of the stocks.
To see why this was useful, think of a stock as a round pie. Investors on Main Street often eat the pie whole, with predictably dire consequences. The professionals that we saw, by contrast, sliced stocks into different properties. Each desk was in charge of a different property, and the different desks then shared their insights with each other. This could happen in a one-minute chat between senior traders across desks, or in an overheard conversation from the desk nearby. This communication allowed traders to understand those aspects of the stock that lay outside their own models — the unexpected “black swans” that can derail a trade.
Sharing, of course, is easier said than done. The bank made it possible with a culture that prized collaboration. For instance, it used objective bonuses rather than subjective ones to ensure that envy did not poison teamwork. It moved teams around the room to build the automatic trust that physical proximity engenders. It promoted from within, avoiding sharp layoffs during downturns.
Most importantly, the leadership of the trading room had the courage to punish uncooperative behavior. Bill, the manger of the room, made it abundantly clear that he would not tolerate the view, prominent among some, that if you’re great at Excel, “it’s OK to be an asshole.” And he conveyed the message with decisive clarity by firing anti-social traders on the spot — including some top producers.
In other words, the culture at the bank was nothing like the consecration of greed that outsiders attribute to Wall Street. We refer to it as “organized dissonance.”
The bank went so far as to use its own models to be reflective upon modeling. The traders translated stock prices into the model estimates developed by their competitors. This information often planted healthy doubts on the traders’ own estimates, sending them back to the drawing board when necessary. Interestingly, this form of “reverse engineering” was accomplished by using the traders’ own models in reverse, much as one can flip a telescope to make something close-up look like it is far away.
Our study suggests that a lack of reflexivity –that is, the lack of doubt on the part of banks– may be behind the current credit crisis. We are reminded of infantry officers who instructed their drummers to disrupt cadence while crossing bridges. The disruption prevents the uniformity of marching feet from producing resonance that might bring down the bridge. As we see it, the troubles of contemporary banks may well be a consequence of resonant structures that banished doubt, thereby engendering disaster.
This blog post was coauthored with David Stark. David Stark is chair of the Department of Sociology at Columbia and is the author of The Sense of Dissonance (Princeton University Press, 2009).
September 17, 2008
Martha’s post (and particularly, the reference to the FT’s decade of moral hazard) made me think about the notions of moral hazard and systemic risk from a sociological perspective. From a financial economic perspective, moral hazard and systemic risk are categorised as different ‘species’ in the markets’ ecosystem: the former is mostly a bilateral risk while the latter is, well, systemic. However, when looked at sociologically, an interesting connection may appear.
Moral hazard may affect mostly bilateral contractual connections, but its source is rooted in a widely accepted, acted-upon belief or what Bourdieu would have called a field. Continuing Bourdieu’s lexicology, in the case of moral hazard, the field in which many financial actors seemed to have operated included the following habitus: counterparty to a financial contract would have a reason to believe that they can default on their contractual obligation and not suffer the consequences of that action, as the governmental regulator would step in. Of course, not just any actor in the financial market could develop and maintain such habitus. The events of the last few days show us that not even big and well-respected players, such as Lehman Brothers, could count on a bailout.
What is it, then, that allowed AIG to be saved and left Lehman Brothers to the market forces, that is, to go bankrupt? This brings us to systemic risks, or as the case may be, the lack thereof. Maybe the demise of Lehman Brothers was a result of a miscalculation on their part. That is, maybe they assumed that they were big enough to pose systemic risk in case of failure, but they were, in fact, ‘less than systemic’.
Which brings us back also to AIG. Is it the case that AIG was really too big and constituted too many inter-connections to be allowed to fail? The answer, as everyone can recite by now, is a resounding ‘yes’. The collapse of AIG would have been the crystallization of a systemic risk scenario and the Federal Reserve would not have allowed it to unfold. There is no denying that AIG plays a major role in the market’s immune system, as it were. Its share in default protection contracts is substantial. However, it is not only the actual market share that turned AIG into a potential systemic risk; it was the fear, fuelled by uncertainty, about who exactly were the ‘infected’ counterparts of AIG and to what extent they were affect, that drove the Fed to give AIG the unprecedented loan. The Fed, like many other market participants, regulatory and otherwise, is playing in the field of financial markets not according to a fully prescribed set of rules, but more through acknowledgments achieved through practice-based trials and errors.
August 31, 2008
As promised, here are some notes following the Market Devices session that Daniel Beunza, Dan Gruber and Klaus Webber arranged (thanks again!). I refer here mostly to the comments our discussant, Bruce Kogut, made. He made some excellent points there. In fact, they made me think critically about the core elements of the performativity approach and, as a result, sharpen the argument. Actually, having read some of the comments to this post in orgtheory, especially Ezra Zuckerman’s, I think that this follow up corresponds with that discussion too.
Bruce refered to the empirical point in my point that says that Black-Scholes-Merton was not accurate and he asked something along the lines of ‘how can one say that model X was not accurate if there was no alternative (or that there were and model X was the least inaccurate one). Here Bruce touched one of the core points of performativity. On one hand, the historical data shows that Black-Scholes-Merton was never very accurate and, as you rightly pointed out, the actors were (or became) fully aware of that fact. So, do we have here a case whereby Black-Scholes-Merton was simply the least inaccurate model in existence?
This question is penetrating because it drills to a core concepts that the performativity approach has been presenting (but, until now, was not explicit enough about it) of ‘scientific’ and ‘organisational’ accuracies. Originating in the sociology of science, performativity ‘plays against’ the scientific concept of accuracy and validity. That is, the concept according to which predictions are valid if, and to the degree to which, they correspond to a universal set of assumptions. Taking this concept to an extreme, predictions can never ‘become accurate’ as a result of interactions between the prediction and the world around it. Hence, theories or models can become better predictors of reality only as a result of improvements in the theories themselves.
The sociology of scientific knowledge claims that ‘scientific’ accuracy is created and maintained by an ‘organisational’ element. Predictive models are typically subjected to a continuous set of public experiments and persuasion trials where their predictive powers are challenged. Hence, to have scientific accuracy, a stable set of ties has to emerge and exist between different actors. Such a network of ties represents an organisational form that, if structurally stable, makes the model ‘organisationally accurate’. That is, enough actors (and ones in influential positions) share opinions regarding the usefulness, efficacy of the practices and/or technologies that use the model.
So, was the Black-Scholes-Merton model simply the best prediction method available, in spite of the fact it that was not accurate scientifically? The interdependency between scientific accuracy and organisational accuracy tells us that we cannot judge the scientific accuracy of a predictive model while separating it from its organisational accuracy (especially in a ‘public experiment’ environment such as financial markets). In fact, the important element here, as you rightly pointed out in your comments, is that market participants decided to develop applications based on the Black-Scholes-Merton model (e.g. margin calculations, calculation of required capital) and, crucially, to develop interdependencies based on the model. This structural dynamic is what made Black-Scholes-Merton ‘organisationally accurate’: an inter-organisational space, composed of options exchanges, the clearinghouse and the SEC, emerged where the model, its results and their validity were accepted (and acted upon). Note that this is not an ‘anything goes’ approach; it does not predict that any model could be accepted and made into an industry standard. It does suggest, however, that inter-subjectivity among organisational actors is crucial for the acceptance of risk management models and that we should examine the dynamics of that process when analysing such inter-organisational environments, such as modern financial markets.
June 3, 2008
I was asked the question above, in different variations, many times during the workshop we did in NY in April. ‘what if, instead of Black-Scholes, the traders would have used some less-useful prediction mechanism (for example, astrology) would you then expect a performative effect to take place?’ or (the flip side of the previous question) ‘if Black-Scholes was not accurate to begin with and only became so as a result of traders using it, then why would anyone use a theoretical pricing model that was not producing accurate results?’
An answer, specific to the Black-Scholes case and how it became popular in spite of its inaccuracy during time of financial stress, can be found in this paper, that’s now making its way through a journal review process. However, the concept of performativity of expert knowledge in organisations alludes implicitly to a more general mechanism through which the process unfolds. As I did very briefly in the workshop and, as I now develop a paper version, I would like to offer here an initial set of theoretical definitions that describe the conditions necessary for performativity to take place.
To generalise performativity of predictive expertise, it is necessary to refer to reflexivity. Naturally, actors’ reflexivity is at the core of performativity: actors’ reactions to the ‘predictive content’ of a theory are the engine of performativity. What then, affects actors’ reactions to the theory?
First, given that actors are aware of the content of a theory, the actors’ ability to intervene in the field for which the prediction is made determines how effective would be the efforts to act in accordance with the theory. For example, we may have accurate theories about the rate at which the universe is expanding, but we are virtually helpless when it comes to intervening in this field. In this case, our reflexivity (after all, we are part of the universe) cannot be translated to intervention. In contrast, the public nature of financial markets makes them a field that is open to theory-driven intervention. In fact, such interventions – ‘taking positions in the market’ – are the lifeblood of the financial system.
Second, to serve as a basis for performativity, the connections between the field and the theory have to form a ‘well-designed’ public experiment. By ‘well-designed public experiment’ I mean that it is necessary for the predictive theory to provide items of information that could be confirmed or refuted unambiguously by items of information coming from the field. Again, the Black-Scholes pricing model and the early options exchange provided a nice public experiment. The model predicted a numerical item referring to a future market price and the market produced the actual price. In comparison, astrology can also be used a market-predictive methodology, but it is not likely to create to a good public experiment, as astrological predictions are very difficult to confirm or refute unambiguously.
Let us go back now to question about the predictive quality of the theory. That is, what would happen if the theory produces inaccurate predictions: is it possible that performativity would take place? Note that the two conditions above do not refer to an a priori validity of theories; they simply refer to the mechanisms through which performativity of such theories may evolve. Thus, ‘false’ theories can be performed; and that is because the second half of the mechanism described above, the public experiment, is exactly the process by which the validity of such theories is tested. Theories that produce predictions about the social, and especially ones that refer to intervention-prone areas in society, are not examined in isolated laboratories, but in public experiments.
However, a simple counter-argument can be presented here: maybe what we witness are the actions of the actors that simply activate a theory that had been correct? Such an argument, of course, does not take into account the full meaning of a predictive theory. Predictions that derive from theories do not provide arbitrary predictions of the future, but refer to a causal mechanism that stands at the basis of the predictions. So, when a theory does not incorporate the effect that its own prediction have on actors that can (and do) intervene in the predicted field, the predictive ability of that theory is reduced significantly. Of course, it is possible that such a ‘deaf’ theory would work, even for a considerable period of time, but when the actors (both human and non-human, as it were) would stop ‘supporting’ it, the theory would no longer produce accurate predictions. A good example, of course is, this is what happened (with some simplifications) to the Black-Scholes pricing model after the 1987 crash.
If we take this argument to its logical extreme then we could hypothesise situations whereby any theory could be performed, just as long as actors find its predictions useful and act in accordance with these predictions. However, the notion of usefulness underlying this argument is, in essence, a pragmatist one. That is, we know that the usefulness of a theory does not equate with its accuracy (the same way a public experiment is different from a classic laboratory experiment) and yes, we can hypothesise such a situation. But, it would be safe to assume that it is unlikely that actors would find theory that consistently produces grossly inaccurate predictions useful and thus, it is unlikely that such theory would be performative.