Opinion

Felix Salmon

What would happen if investments in people succeeded?

Felix Salmon
Oct 22, 2012 20:54 UTC

Daniel Friedman has a question:

There are now a few companies — Upstart and Lumni to name two — trying to create a market where we can invest in people in exchange for a percentage of future income. If they succeed and reach scale, will student debt problems be alleviated? Will our future incomes be reliably predicted by newly developed algorithms? Will it encourage increased risk-taking?

The first thing to note here its that this is a market which does not exist. Lumni tried to gin up some interest in it and failed; at this point they won’t even return my calls or emails. Upstart has a bunch of high-profile venture backing (Google Ventures, Kleiner Perkins), and has even managed to launch a pilot with a few hand-picked students and investors. But it’s still very, very early days. Equity-in-people is an idea which comes around occasionally, but so far it has never really taken off, and there’s not much reason to believe that this time is different.

But still, it’s an interesting question: what would the consequences be if Upstart, and/or companies like it, became big and successful?

Firstly, Friedman asks, would student-debt problems be alleviated? The answer to this one, I think, is quite clearly no, for two reasons. The main reason is that tuition fees are induced, in much the same way that traffic is. If you build a nice big new road, traffic will appear to fill it up. And if you build a nice new source of funding for students, universities will raise their tuition fees so as to capture as much of that funding as possible. Yes, high student debt is a consequence of high tuition fees — but at the same time, high tuition fees are a consequence of the ready availability of student loans.

It’s worth pointing out here that Upstart’s model gives out funds at the end of a student’s four-year education, rather than at the beginning. If the student has lots of debt, and wants less debt, then she can use the funds to pay down some of that debt. But many students will choose to spend the money in other ways, simply layering the new equity funding on top of their old debt funding. Alternatively, if the expectation is that students will use their Upstart funds to pay down debt, then that only means, at the margin, that lenders will feel free to extend even more credit in the run-up to the early cash-out.

The other reason why student-debt problems won’t be alleviated is that Upstart funds are debt, if you look hard enough. The documentation is subject to change, but at heart would-be Upstarts are always going to find something along the lines of this in any Upstart contract:

YOU UNDERSTAND AND ACKNOWLEDGE THAT THE FUNDING AMOUNT YOU RECEIVE FROM US, IS A DEBT TO US.

The debt is not a typical loan, of course, and it can end up being repaid at less than face value, if the Upstart enters a poorly-remunerated career. But it’s still a debt. And if the student ends up defaulting on that debt, the penalties are generally enormous. When people take out loans, they nearly always have the best of intentions, and don’t think there’s any way they will end up defaulting — especially if the repayment amounts are tied to their income. But defaults happen. And when they do, anybody who has defaulted to Upstart is going to find out the hard way that Upstart loans are the most expensive of them all.

So at the margin, these kind of schemes are likely to exacerbate the student-loan problem: they’re hurting, rather than helping. Remember that ultimately the money for these schemes is coming from investors, who are taking a substantial risk and being promised returns significantly greater than anything seen in the debt markets. Remember too that a significant portion of the students will end up repaying less than they’re originally given — because they go into low-paying occupations, perhaps, or maybe because they just decide they’d rather settle down and have a family instead of a career. As a result, those students who do go into decently-paying careers will end up repaying sometimes double or triple the amount of money they were originally given. That doesn’t seem like an alleviation of student debt problems to me — not when we’re talking about sums in the $30,000 range. It sounds like piling an extra $60,000 of liabilities on top of all the existing student loans you might have.

Friedman’s second question is whether companies like Upstart will start being able to predict students’ future incomes. That’s certainly something that Upstart is spending quite a lot of time on, but I think the effort is premature: first you need to find out whether students are happy repaying the funds at all. And I, for one, don’t think that Upstart’s algorithms are going to be particularly good at predicting incomes.

For one thing, the students with predictable incomes — the doctors and lawyers and the like — are never going to accept Upstart’s offer. For another, there will always be a certain number of students who are looking to game the system. All these schemes have an adverse-selection problem: Upstart funding is going to be much more attractive to students who are pretty sure they’re never going to make much money. If you aspire to being a part-time primary-school substitute teacher, for instance, while spending the rest of your time working on the Great American Novel, then Upstart might be great for you — and it’s unlikely that any algorithm would be able to capture that.

What’s more, the biggest risk, from Upstart’s point of view, is that they’ll end up funding a student who marries someone very successful, gets pregnant pretty early on, and then never returns to the workforce. This still happens frequently enough that an honest algorithm would charge higher interest rates to women than to men. As a result, the algorithm can’t actually be honest.

Finally, there’s the group of students who want to follow the Silicon Valley dream and become entrepreneurs straight out of college. That’s generally a great time to found a company, and Upstart (slogan: “The startup is you“) is positively encouraging such people. But equally, the chances that any given person will have success as an entrepreneur are pretty much random: such things come mostly down to luck, and can’t be reduced to some algorithm. Especially when the really successful entrepreneurs are going to be the ones who can afford lawyers to shield their income from Upstart.

There is no algorithm which can tell you who’s going to be a successful entrepreneur and who isn’t. Some college dropouts are college dropouts; other college dropouts are Steve Jobs, or Bill Gates, or Mark Zuckerberg. Being dyslexic is generally not a great thing when it comes to total lifetime income, and yet dyslexics are massively overrepresented among highly-successful CEOs. If you’re looking for outliers — and successful entrepreneurs, pretty much by definition, are all outliers — then you can’t really start generalizing.

Friedman finishes by asking if this model will encourage increased risk-taking. I think the answer there depends on what you mean by risk-taking. It certainly encourages people to accept low-paying jobs with payoffs which get put into corporate shell vehicles or which don’t end up cashing out for a decade. And at the other end of the spectrum, it encourages people to take low-paying jobs which pay non-financial dividends: ski instructor, say. But in general, I don’t think that a new form of private income tax is a great way to encourage innovation. Generally, if you start taxing something, you get less of it, not more of it. And I see no reason why this model should work out any differently.

When Taleb met Davies

Felix Salmon
Oct 19, 2012 21:59 UTC

This morning, Nassim Taleb returned to Twitter, posting one of the technical appendices to his new book. And immediately he got into a wonderfully wonky twitterfight/conversation with Daniel Davies.

I don’t pretend to understand all the subtleties of the conversation between the two, but, for Tom Foster, here’s an attempt. Davies has promised a Crooked Timber post on other parts of the appendix; I’m really looking forward to that.

How to protect New York from disaster

Felix Salmon
Sep 11, 2012 18:51 UTC

Today, September 11, is a day that all New Yorkers become hyper-aware of tail risk — of some monstrous and tragic disaster appearing out of nowhere to devastate our city. And so it’s interesting that the NYT has decided to splash across its front page today Mireya Navarro’s article about the risk of natural disaster — flooding — in New York.

Beyond the article’s publication date, Navarro doesn’t belabor the point. But in terms of the amount of death and destruction caused, a nasty storm hitting New York City could actually be significantly worse than 9/11. Ask anybody in the insurance industry: a hurricane hitting New York straight-on is the kind of thing which reinsurance nightmares are made of. And as sea levels rise in coming decades, the risks will become much worse: remember, it’s flooding from storm surges which causes the real devastation, rather than simply things blowing over in high winds.

So, what can or should be done? One option is to basically attempt to wall New York City off from the Atlantic Ocean:

A 2004 study by Mr. Hill and the Storm Surge Research Group at Stony Brook recommended installing movable barriers at the upper end of the East River, near the Throgs Neck Bridge; under the Verrazano-Narrows Bridge; and at the mouth of the Arthur Kill, between Staten Island and New Jersey. During hurricanes and northeasters, closing the barriers would block a huge tide from flooding Manhattan and parts of the Bronx, Brooklyn, Queens, Staten Island and New Jersey, they said.

Needless to say, this solution is insanely expensive: the stated price tag right now is $10 billion — well over $1,000 per New Yorker — and I’m sure that if such a project ever happened, the final cost would be much higher. And such barriers don’t last particularly long, either. London built the Thames Barrier in 1984, and there’s already talk about when and how it should be replaced. And building a single barrier across the Thames is conceptually and practically a great deal simpler than trying to hold back the many different ways in which the island of Manhattan is exposed to the water.

What’s more, there’s an environmental cost associated with barriers, as well as a financial cost. Which cuts against the kind of things which New York has been doing. They’re smaller, and much less robust. But they improve the environment, rather than making it worse. And they’re relatively cheap. For instance: installing more green roofs to absorb rainwater. Expanding wetlands, which can dampen a surging tide, even in highly-urban places like Brooklyn Bridge Park. Even “sidewalk bioswales”. (I’m a little bit unclear myself on exactly what those are, but they sound very green.)

Adam Freed, the outgoing deputy director of New York’s Office of Long-Term Planning and Sustainability, talks about making “a million small changes”, while always bearing in mind that “you can’t make a climate-proof city”. That’s a timely idea: we can’t make New York risk-free, and it’s not clear that it would make sense to do so even if we could. After all, as we all learned 11 years ago today, it’s impossible to protect against each and every source of possible devastation.

Other cities have similar ideas:

In Chicago, new bike lanes and parking spaces are made of permeable pavement that allows rainwater to filter through it. Charlotte, N.C., and Cedar Falls, Iowa, are restricting development in flood plains. Maryland is pressing shoreline property owners to plant marshland instead of building retaining walls.

Still, all of this green development does feel decidedly insufficient in comparison to the enormous risks that New York is facing. I like the idea of a “resilience strategy”, but there are still a lot of binary outcomes here, especially when it comes to tunnels. Either tunnels flood or they don’t — and if they do, the consequences can be really, really nasty. Imagine a big flood which took out all of the subway and road tunnels into Manhattan, or even just the subway tunnels across New York Bay as well as the Holland Tunnel. As such a flood becomes more likely, the cost of protecting against it with some big engineering work — insofar as such a thing is possible — becomes increasingly justifiable.

And this is just depressing:

Consolidated Edison, the utility that supplies electricity to most of the city, estimates that adaptations like installing submersible switches and moving high-voltage transformers above ground level would cost at least $250 million. Lacking the means, it is making gradual adjustments, with about $24 million spent in flood zones since 2007.

Lacking the means? What is that supposed to mean? New York City has a credit rating of Aa1 from Moody’s; ConEd has a crediting rating of A3. Interest rates are at all-time lows. There has never been a better time to invest a modest $250 million in helping to ensure that New York can continue to have power in the event of a storm. Doing lots of small things is all well and good, and I’m not convinced that the huge things are necessarily worthwhile — or even, in the case of moving people to higher ground, even possible. But the medium-sized things? Those should be a no-brainer right now.

Chart of the day, employment-status edition

Felix Salmon
Sep 7, 2012 14:59 UTC

status.jpg

There are two ways in which the national employment situation influences the election. The first is, simply, the effect of unemployment and underemployment on America’s animal spirits. People who are unemployed, or who are so discouraged that they’re not even looking for work any more, don’t tend to be very happy with their lot, and as a result are more likely to vote against the current president. It may or may not be fair, but the president does get blamed for current economic conditions, and arguments about first derivatives (“it’s bad, but it’s getting better”) or counterfactuals (“it’s bad, but it’s better than it would have been under the other guys”) tend to be pretty unpersuasive to voters.

On this level, today’s employment report is pretty gruesome. According to the establishment survey, employers added just 96,000 jobs this month — less than the amount needed just to keep up with population growth. According to the household survey, the size of the civilian labor force shrank by 368,000 people last month. And the number of people not in the labor force grew by an absolutely massive 581,000.

Right now, the proportion of Americans with a job is lower than it has been in over 30 years. America’s getting older, and you’d expect the number to be falling — but it shouldn’t be falling nearly as fast as this. We’re well below trend, when it comes to the employment-to-population ratio, and that’s really bad for the economy as a whole: it means we have fewer productive workers, and as a result the country is creating much less wealth than it could be creating if more people had jobs. At the margin, of course, anything that depresses the amount of wealth in the country is bad for the incumbent president.

So anybody trying to use the jobs report to handicap the result of the election should probably see a tick down, this morning, in the chances of Obama’s re-election. My feeling is, however, that the size of the tick is likely to be very small. These things depend much more on levels than on deltas, and in any case the current electorate is more polarized than ever, with political convictions which are hard to shake.

Which brings me to the second way that employment affects electoral outcomes. The employment numbers are reported, on the first Friday of every month, and political parties try to use the numbers to their best advantage. On this front, there’s really only one number that matters, and that’s the headline unemployment rate. Financial types care more about the payrolls number, because it’s more accurate and less fuzzy. The unemployment rate, by contrast, is harder to calculate, and is based on the idea that you’re only unemployed if you’re looking for work. But the fact is that from a rhetorical perspective, the unemployment rate is the thing which counts. And so in terms of the optics of today’s report, it’s good for Obama, just because the unemployment rate fell — to 8.1% this month from 8.3% last month and 9.1% a year ago.

That’s still well above the 7% at which the psephologists will tell you that it’s very hard for an incumbent to get reelected. And it still starts with an 8 — although there’s now a small chance that on the day we actually vote, the unemployment rate might start with a 7. But the Republicans can’t say that the unemployment rate is rising, and the Democrats can say that it is falling. Will that change votes? Again, not very many. But insofar as arguments have an effect on elections, this report — bad though it is — has failed to give the Republicans the kind of rhetorical ammunition they might have hoped for.

Underlying both of these dynamics is the way in which the story of discouraged workers — people falling out of the labor force entirely — has become increasingly important, to the point at which it makes the headline unemployment rate much less useful as an economic indicator. Once upon a time, if you didn’t have a job, you fell into one of two categories: either you didn’t want to work, or else you were looking for work. Nowadays, however, there’s a huge third category of discouraged workers who would love a job but don’t even see the point of looking any more.

The Bureau of Labor Statistics has an interesting-if-obscure data series called “labor force status flows”. Most of the people interviewed in the survey measuring the unemployment rate, it turns out, were also interviewed the previous month. So it’s possible to look at the number of people, on a month-to-month basis, who were unemployed last month and who were no longer in the labor force this month. Historically, that number has been somewhere between 1.5 million and 2 million per month, on a seasonally-adjusted basis. But when the recession hit, it spiked to more than 2.5 million, and even more than 3 million at the peak. And it’s still extremely high.

It’s natural for lots of unemployed people to move out of the labor force each month: the Boomers are retiring, after all. But a glance at this chart is all it takes to see that we’re well outside normal territory, and that we’re still seeing millions of people leave the labor force not because they want to but because they feel that there’s simply no point in looking for work any more. I don’t know when or whether this line will come back down to its historical levels. But so long as it’s as elevated as this, the Federal Reserve has its work cut out. Because it means that even if the unemployment rate comes down substantially, we still won’t have really reached full employment — not unless the size of the labor force increases substantially at the same time.

Why you won’t find hyperinflation in democracies

Felix Salmon
Sep 4, 2012 03:29 UTC

There are those who believe that the length of a mathematical paper is inversely proportionate to how interesting it is. Something similar can be said about the new paper — short and absolutely first-rate — from Steve Hanke and Nicholas Krus, entitled “World Hyperinflations“. It’s technically 19 pages long, but the first 12 are basically just throat-clearing, and the last two are references. The meat is the five pages in the middle: three pages of tables, and another two of footnotes, detailing every instance of hyperinflation that the world has ever seen.

Hyperinflation, here, has a clear quantitative definition: prices rising by at least 50% per month. (Remember that, the next time some scaremonger starts talking about how US monetary policy risks causing hyperinflation.) And after some three years’ work, Hanke and Krus have managed to come up with an exhaustive list of every hyperinflationary episode in history — 56 in all, or 57 if you include North Korea in early 2010, where the data aren’t solid enough to merit inclusion in the list.

Every entry gets its own footnote, and while there are a lot of relatively easy-to-obtain IMF publications in there, there’s also no shortage of much more obscure source material: Simeun Vilendecic’s Banking in Republika Srpska in the late XX and early XXI century, for instance, or Abram van Heyningen Hartendorp’s 1958 History of Industry and Trade of the Philippines.

The earliest hyperinflation on the list came in France, at the end of the 18th Century, when inflation hit a monthly rate of 304% in mid-August 1796. The famous Weimar hyperinflation in Germany is pegged as taking place between August 1922 and December 1923; it reached a monthly peak of 29,500% in October 1923, with prices increasing at 20.9% per day, and doubling every 3.7 days. And the longest period of hyperinflation comes in Greece, which saw hyperinflation for a whopping 55 months, from May 1941 to December 1945. There’s no particular reason, looking at this list, why Germany should have been particularly scarred by hyperinflation, to the point at which it fiercely attacks even the possibility of relatively modest inflation, where France and Greece (not to mention Hungary or China or Argentina) have been much less deeply affected.

There is, however, a very strong correlation between the length of time that a period of hyperinflation goes on, and the levels that it can reach at its height. If you look at the top six hyperinflations on the list — which include both Germany and that 55-month period in Greece — all but one lasted for longer than a year. Meanwhile, five of the bottom six hyperinflations took place in just a single month, with the sixth lasting just three months.

At their highest, the numbers start to beggar the imagination: in mid-November 2008, for instance, inflation in Zimbabwe reached a monthly rate of 79,600,000,000%. That’s 79 billion percent per month. At that rate, prices pretty much double every day. And Zimbabwe doesn’t even manage to grab the top spot: in July 1946, Hungary saw hyperinflation of 41,900,000,000,000,000%. That’s 42 quadrillion percent in one month, with prices doubling every 15 hours.

The real value of this paper is its exhaustive nature. By looking down the list you can see what isn’t there — and, strikingly, what you don’t see are any instances of central banks gone mad in otherwise-productive economies. As Cullen Roche says, hyperinflation is caused by many things, such as losing a war, or regime collapse, or a massive drop in domestic production. But one thing is clear: it’s not caused by technocrats going mad or bad.

For that matter, there are no hyperinflations at all in North America: the closest we’ve come, geographically speaking, was in Nicaragua, from 1986-91. In fact, if you put to one side the failed states of Zimbabwe and North Korea, there hasn’t been a hyperinflation anywhere in the world since February 1997, more than 15 years ago, despite the enormous number of heterodox central-bank actions in that time.

All of which is to say that hyperinflation, in and of itself, really isn’t anything to worry about. It’s pretty much impossible to predict — and if your country has hyperinflation, it almost certainly has even bigger other problems. In fact, I’d hesitate to categorize hyperinflation as a narrowly economic phenomenon at all, as opposed to simply being a symptom of much bigger failures at the geopolitical level. Those failures are exacerbated by hyperinflation, of course: there’s very much a vicious cycle in these episodes. But you only ever find hyperinflation under extreme conditions, and, with a single exception (Peru), I’m not even sure I can find any genuine democracies on this list.

Update: As many people have helpfully pointed out, Weimar can definitely be considered a genuine democracy, even if it was suffering extraordinary geopolitical burdens.

Annals of dubious research, 401(k) loan-default edition

Felix Salmon
Aug 13, 2012 05:13 UTC

Bob Litan, formerly of the Kauffman Foundation and the Brookings Institution, has recently taken up a new job as director of research for Bloomberg Government, where he’s going to have to be transparent and impartial. But one of his last gigs before moving to Bloomberg — a paper on the subject of people borrowing money from their 401(k) accounts — was neither of those things.

To understand what’s going on here, first check out Jessica Toonkel’s article from Friday about Tod Ruble and his company, Custodia.

Tod Ruble is trying to sell retirement plan insurance that employers say they do not want and their employees may not need.

But the Dallas-based veteran commercial real estate investor is not letting that stop him. Since late 2010, he has started up a company, Custodia Financial, and spent more than $1 million pushing for legislation that would allow companies to automatically enroll employees who borrow from their 401(k) plans in insurance that could cost hundreds of dollars a year.

Once you’ve read that, go back and check out a spate of stories that hit a series of major news outlets in July. Alan Farnham of ABC News, for instance, ran a story under the headline “401(k) Loan Defaults Skyrocket”:

A new study estimates that such defaults might total $37 billion a year, a sharp increase from 2007, when defaults totaled only $665 million.

Similarly, check out Walter Hamilton, in the Chicago Tribune (and LA Times): the headline there is “Defaults on 401(k) loans reach $37 billion a year”. At Time, Dan Kadlec also ran with the $37 billion number, saying that “the default rate on these loans has skyrocketed since the recession”. Similar stories came from Blake Ellis at CNN Money (“Loan defaults drain $37 billion from 401(k)s each year”), Mitch Tuchman at MarketRiders (“401k Loan Default Time Bomb Is Ticking”), and many others.

The only hint of skepticism came from Barbara Whelehan at BankRate. She noted that the study cited Kevin Smart, CFO of Custodia Financial, as a source — and she also noted that “it would be a boon for the insurance industry to get the rules changed, and it is working behind the scenes to do just that. In April, Custodia Financial submitted a statement to the House & Ways Committee arguing for automatic enrollment into insurance coverage for 401(k) loans.”

Whelehan also smelled something fishy in the way the paper was paid for:

This paper by Navigant Economics, which made a big splash in the press, was financially supported by Americans for Retirement Protection. That organization has a website, ProtectMyRetirementBenefits.com, but no “about us” link. It does give you the opportunity to sign a petition demanding protection of retirement funds through insurance. Take a look at it, and see if you think the website was created by average Americans or by the insurance industry.

Whelehan was actually breaking news here: there’s no public linkage between Americans for Retirement Protection, the organization which paid for the paper, and the astroturf website. In fact, Americans for Retirement Protection seems to have no public existence at all, beyond a footnote in the paper, which was co-authored by Bob Litan and Hal Singer.

Enter Toonkel, writing her story about Custodia. In the course of her reporting, she discovered — and Custodia confirmed — that Americans for Retirement Protection, and ProtectMyRetirementBenefits.com, are basically alter egos of Custodia itself. Custodia would welcome other organizations joining in, but that’s unlikely to happen, because Custodia owns the patents on the big idea that the paper and the website are pushing — the idea that 401(k) loans should come bundled with opt-out insurance policies.

Once you’re armed with this information, it’s impossible not to look at the Litan-Singer paper in a very different way. Its abstract concludes: “We demonstrate that the social benefits of steering (but not compelling) plan participants towards insurance when they borrow are likely positive and economically significant.” And yet nowhere in the paper is there any indication that it was bought and paid for by the very company which has a patent on doing exactly that.

And what about that $37 billion number? Are defaults on 401(k) loans really as big a problem as the paper says that they are? After all, the smaller the problem, the less important it is to introduce an expensive fix for it.

The simple answer is no: 401(k) loan defaults are not $37 billion per year. But the fact is that nobody knows for sure exactly where they are, which makes it much easier to come up with exaggerated estimates. As the paper itself admits, “the sum total of 401(k) defaults ought to be an easily accessible statistic, but it is not”. And the $37 billion, far from being a good-faith estimate, in fact looks very much like an attempt to get the largest and scariest number possible.

So how did Litan and Singer arrive at their $37 billion figure? Let’s start with the only concrete numbers we have — the ones from the Department of Labor, whose most recent Private Pension Plan Bulletin gives a wealth of information about all private pension plans in the country. Every pension plan has to file something called a Form 5500, and the bulletin aggregates all the numbers from all the 5500s which are filed; the most recent bulletin gives data from 2009.

This bulletin has two datapoints which are germane to this discussion. First of all, there’s Table A3, on page 7 of the bulletin (page 11 of the PDF). That shows that loans from defined-contribution pension plans to their own participants totaled $51.7 billion in 2009. Secondly, there’s Table C9, the aggregated income statement for the year. If you look at page 32 of the bulletin (page 35 of the PDF), you’ll see a line item called “deemed distribution of participant loans”, which came to $670 million for the year. If you borrow money from your 401(k) and you don’t pay it back, then that money is deemed to have been distributed to you, and counts as a default. So we know that the official size of 401(k) defaults in 2009 was $670 million — a far cry from Litan and Singer’s $37 billion.

Now the $670 million figure does not account for all 401(k) defaults. Most importantly, in some situations, if you default on a 401(k) loan after having been fired from your job, then the money is counted as an “actual distribution” rather than as a “deemed distribution”.

The Litan-Singer paper goes into some detail about this. “According to a recent study by Smart (2012),” they write, “although Form 5500 reflects actual distributions, there is no way to determine the amount of actual defaults.” They then look in detail at Smart’s figures, footnoting him five consecutive times, and treating him as an undisputed authority on such matters. Their citation is merely “Kevin Smart, The Hidden Problem of Defined Contribution Loan Defaults, May 2012.”

Where might someone find this paper? Here, since you ask: it’s helpfully hosted at CustodiaFinancial.com. And on the front page of the paper, Kevin Smart is identified as the “Chief Financial Officer, Custodia Financial”.

There’s no indication whatsoever in the Litan-Singer paper that the “Smart” they cite so often is the CFO of Custodia Financial, the company which has the most to gain should their recommendation be accepted. And there’s certainly no indication that he’s essentially their employer: that Custodia paid them to write this paper. In fact, the name Custodia appears nowhere in the Litan-Singer paper at all.

It’s instructive to look at the Smart paper’s attempt to estimate the magnitude of the 401(k) default problem. I’ll simplify a little here, but to a first approximation, Smart assumes that 12% of people with 401(k) loans lose their jobs. He also assumes that if you lose your job when you have a 401(k) loan, there’s an 80% chance you’ll default on that loan. As a result, he comes up with a 9.6% default rate on 401(k) loans. He then multiplies that 9.6% default rate by total 401(k) loans of $51.7 billion, adds in some extra defaults due to death and disability, and comes up with a grand total of $6.2 billion in loan defaults per year, excluding the “deemed distributions” of $670 million. Call it $7 billion in total, of which $6 billion could be protected by insuring loans against unemployment, death, and disability.

Now remember that this is a paper written by the CFO of Custodia Financial — someone who clearly has a dog in this race. It’s in Smart’s interest to make the loan-default total look as big as possible, since the bigger the problem, the more likely it is that Congress will agree to implement Custodia’s preferred solution.

But when Litan and Singer looked at Smart’s paper, they weren’t happy going with his $6 billion figure. Indeed, as we’ve seen, their paper comes up with a number six times larger. So if even the CFO of Custodia only managed to estimate defaults at $6 billion, how on earth do Litan and Singer get to $37 billion?

It’s not easy. First, they double the total amount of 401(k) loans outstanding, from $52 billion to $104 billion. Then, they massively hike the default rate on those loans, from 9.6% to 17.9%. And finally, they add in another $12 billion or so to account for the taxes and penalties that borrowers have to pay when they default on their loans.

It’s possible to quibble with each of those changes — and I’ll do just that, in a minute. But it’s impossible to see Litan and Singer compounding all of them, in this manner, without coming to the conclusion that they were systematically trying to come up with the biggest and scariest number they could possibly find. It’s true that whenever they mention their $37 billion figure, it’s generally qualified with a “could be as high as” or similar. But they knew what they were doing, and they did it very well.

When the Chicago Tribune and the LA Times say in their headlines that 401(k) loan defaults have reached $37 billion a year, they’re printing exactly what Custodia and Litan and Singer wanted them to print. The Litan-Singer paper doesn’t exactly say that defaults have reached that figure. But if you put out a press release saying that “the leakage of funds in 401(k) plans due to involuntary loan defaults may be as high as $37 billion per year”, and you don’t explain anywhere in the press release that “leakage of funds” means something significantly different to — and significantly higher than — the total amount defaulted on, then it’s entirely predictable that journalists will misunderstand what you’re saying and apply the $37 billion number to total defaults, even though they shouldn’t.

But let’s backtrack a bit here. First of all, how on earth did Litan and Singer decide that the total amount of 401(k) loans outstanding was $104 billion, when the Department of Labor’s own statistics show the total to be just half that figure? Here’s how.

First, they decide that they need the total number of active participants in defined-contribution pension plans. They could get that number — 72 million — from the Labor Department bulletin: it’s right there in the very first table, A1. But the bulletin isn’t helpful to them, as we’ve seen, so instead they find the same number in a different document from the same source.

That’s as much Labor Department data as Litan and Singer want to use. Next, they go to the Investment Company Institute, which has its own survey, covering some 23 million of those 72 million 401(k) participants. According to that survey, in 2011, 18.5% of active participants had taken out a loan; Litan and Singer extrapolate that figure across the 401(k) universe as a whole.

Finally, Litan and Singer move on to Leakage of Participants’ DC Assets: How Loans, Withdrawals, and Cashouts Are Eroding Retirement Income, a 2011 report from Aon Hewitt which is based on less than 2 million accounts, of the 72 million total. According to the Aon Hewitt report, which doesn’t go into any detail about methodology, when participants took out loans, “the average balance of the outstanding amount was $7,860″. Needless to say, that number was never designed to be multiplied by 72 million, as Litan and Singer do, to generate an estimate for the total number of loans outstanding.

If you want an indication of just how unreliable and unrepresentative the $7,860 number is, you just need to stay on the very same page of the Aon Hewitt report, which says that 27.6% of participants have a loan. If Litan and Singer think that the $7,860 figure is reliable, why not use the 27.6% number as well? If they did that, then the total number of 401(k) loans outstanding would be $7,860 per loan, times 72 million participants, times 27.6% of participants with a loan. Which comes to $156 billion.

But of course we know that there were just $51.7 billion of loans outstanding in 2009; evidently Litan and Singer reckoned that it just wouldn’t pass the smell test if they tried to get away with saying that number might have trebled in a single year. So they confined themselves to merely doubling the number, instead.

Litan and Singer give no reason to mistrust the official $51.7 billion number, except to say that it’s “outdated”. But if it’s outdated, it’s only outdated by one year: it’s based on 2009 data, while the much narrower surveys that Litan and Singer cite are generally based on 2010 data. At one point, they cite the ICI survey to declare that there is “an estimated $4.5 trillion in defined contribution plans”, despite the fact that the much more reliable Labor Department report shows that there was just $3.3 trillion in those plans as of 2009. This, I think, quite neatly puts the lie to the Litan-Singer implication that the problem with the Labor Department numbers is merely that they are out of date, and that when we get numbers for 2010 or 2011, they might well turn out to be in line with the Litan-Singer estimates. There’s simply no way that total DC assets rose from $3.3 trillion to $4.5 trillion in the space of a year or two.

In other words, whatever advantage the ICI and Aon Hewitt surveys have in terms of timeliness, they more than lose in terms of simply being based on a vastly smaller sample base. Litan and Singer adduce no reason whatsoever to believe that the ICI and Aon-Hewitt surveys are in any way representative or particularly accurate, despite the fact that the discrepancies between their figures and the Labor Department figures are prima facie evidence that they’re not representative or particularly accurate. If the ICI and Aon Hewitt surveys were all we had to go on, then I could understand Litan-Singer’s decision to use them. But given that the Labor Department already has the number they’re looking for, it just doesn’t make any sense that they would laboriously try to recreate it using less-reliable figures.

It’s true that the Labor Department’s figures do undercount in one respect: they cover only plans with 100 or more participants — and therefore cover “only” 61 million of the 72 million active participants in DC plans. If Litan and Singer had taken the Labor Department’s numbers and multiplied them by 72/61, or 1.18, that I could understand. But disappearing into a rabbit-warren of private-sector surveys of dubious accuracy, and emerging up with a number which is double the size of the official one? That’s hard to justify. So hard to justify, indeed, that Litan and Singer don’t even attempt to do so.

That, indeed, is the strongest indication that the Litan-Singer paper can’t really be taken seriously. For all their concave borrower utility functions and other such economic legerdemain, they simply assert, rather than argue, that they “believe” it is “more appropriate” to use private-sector surveys rather than hard public-sector data. Such decisions cannot be based on blind faith: there have to be reasons for them. And Litan-Singer never explain what those reasons might be.

Now the move from public-sector to private-sector data merely doubles the total size of the purported problem, while Litan-Singer are much more ambitious than that. So their next move is to bump up the default rate on loans substantially.

There’s no official data on default rates at all, so Litan and Singer, following Smart’s lead, decide to base their sums on a Wharton paper from 2010. Once again, they have to extrapolate from a very small sample: the Wharton researchers had at their disposal a dataset covering 1.5 million plan participants (just 2% of the total). Looking at what happened over a period of three years, from July 2005 to June 2008, the researchers found that the number of terminations, and the number of defaults, remained pretty steady:

defaults.tiff

These are the numbers that Smart used in his paper: roughly 12% of loan holders being terminated each year, and roughly 80% of those defaulting on their loans.

But these are not the numbers that Litan-Singer use. Instead, they notice that the overall default rate, as a percentage of overall loans outstanding, was roughly double the national unemployment rate at the time. And so since the unemployment rate doubled after June 2008, they conclude that the default rate on outstanding 401(k) loans probably doubled as well.

Do they have any evidence that the default rate on 401(k) loans might have doubled after 2008? No. Well, they have a tiny bit of evidence: they look at the small variations in default rates in each of the three years covered in the Wharton study, and see that those variations move roughly in line with the national unemployment rate. Never mind that the default rate fell, from 9.9% to 9.7%, between 2006 and 2008, even as the unemployment rate rose, from 4.8% to 5.0%. They’ve still somehow managed to convince themselves that it’s reasonable to assume that the default rate today is nowhere near the 9.6% seen in the Wharton survey, and in fact is probably closer to — get this — 17.9%.

This doesn’t pass the smell test. The primary determinant of the default rate, in the Wharton study, was the percentage of loan holders who wound up having their employment terminated, for whatever reason. And so what Litan-Singer should be looking at is the increase in the probability that any given employee will end up being terminated in any given year.

Remember that in any given month, or year, the number of people fired is roughly equally to the number of people hired. When the former is a bit larger than the latter for an extended period, then the unemployment rate tends to go up; when it’s smaller, the rate goes down. But the churning in the employment economy is a constant, even when the unemployment rate is very low.

When the unemployment rate rose after 2008, that was a function of the fact that the number of people being fired was a bit higher than normal, while the number of people being hired was a bit lower than normal. But looked at from a distance, neither of them changed that much. In terms of the Wharton study, what we saw happening to the unemployment rate is entirely consistent with the percentage of loan-holders being terminated, per year, staying pretty close to 12%. Of course it’s possible that number rose sharply, but it’s really not possible that number rose as sharply as the unemployment rate did. And so I find it literally incredible that Litan and Singer should decide to use the national unemployment rate as a proxy for the number of people whose employment is terminated each year.

Well, maybe not literally incredible — the fact is there’s one very good reason why they might do that. Which is that they were being paid by Custodia to use any means possible to exaggerate the number of annual 401(k) loan defaults.

Litan and Singer do actually provide a mini smell test of their own: they say that their hypothesized rise in 401(k) loan defaults is more or less in line with the rise in, say, student-loan defaults or in mortgage defaults over the same period. But those statistics aren’t comparable at all, because Litan and Singer are already assuming that the default rate on 401(k) loans, among people who lose their job, was a whopping 80% before the financial crisis. There’s a 100% upper bound here: you can’t have a default rate of more than 100%. Remember that the whole point of this paper is to provide the case that people taking out 401(k) loans should insure themselves against unemployment: any rise in the default rate from people who don’t lose their job (or die, or become disabled) is more or less irrelevant here. And when your starting point is a default rate of 80%, there really is a limit to how much that default rate can rise; it’s certainly going to be difficult to see it rise by more than 85%, even if you allow a simultaneous increase in the number of people being terminated.

All of this massive exaggeration has an impressive effect: if you take $104 billion in loans, and apply a 17.9% default rate, then that comes to a whopping $18.6 billion in 401(k) loan defaults every year. A big number — but still, evidently, not big enough for Litan and Singer. After all, their number is $37 billion: double what we’ve managed to come up with so far. We’ve already doubled the size of the loan base, and almost-doubled the size of the default rate, so how on earth are we going to manage to double the total again?

The answer is that LItan and Singer, at this point, stop measuring defaults altogether, and turn their attention to a much more vaguely-defined term called “leakage”. Once again, they decide to outsource all their methodology to Custodia’s CFO, Kevin Smart. The upshot is that if you borrowed $1,000 from your 401(k) and then defaulted on that loan, the amount of “leakage” from your 401(k) is deemed to be much greater than $1,000. Litan and Singer first add on the 10% early-withdrawal penalty that you get charged for taking money out of your plan before you retire. They also add on the income tax you have to pay on that $1,000, at a total rate of 30%. (They reckon you’ll pay 25% in federal taxes, and another 5% in state taxes.) So now your $1,000 default has become a $1,400 default.

How does that extra $400 count as leakage from the 401(k), rather than just something that gets added to your annual tax bill? Smart explains:

Most participants borrow from their retirement savings because they are illiquid and do not have access to other sources of credit. This clearly demonstrates that participants who default on a participant loan do not have the financial means to pay the taxes and penalty. Unfortunately, their only source of capital is their retirement savings plan so many take the remaining account balance as an additional early distribution to pay the taxes and penalty, further increasing the amount of taxes and penalties due. These taxes and penalties become an additional source of leakage from retirement assets.

Smart’s 16-page paper has no fewer than 24 footnotes, but he fails to provide any source at all for his assertion that “many” people raid their 401(k) plans in order to pay the taxes on the money they’ve already borrowed. In any event, Smart (as well as Litan and Singer, following his lead) makes the utterly unjustifiable assumption that not only many but all 401(k) defaulters end up withdrawing the totality of their penalties and extra taxes from their retirement plan. And then, just for good measure, because that withdrawal also comes with a penalty and taxes, they apply a “gross-up” to that.

By the time all’s said and done, the $1,000 that you lent yourself from your 401(k) plan, and failed to pay back in a timely manner, has become $1,520 in “leakage”. Add in some extra “leakage” for people who default due to death or disability (apparently even dead people raid their 401(k) plans to pay income tax on the money they withdrew), and somehow Litan and Singer contrive to come up with a total of $37 billion.

It’s an unjustifiable piling of the impossible onto the improbable, and the press just lapped it up — not least because it came with the imprimatur of Litan, a genuinely respected economist and researcher. Custodia hired him for precisely that reason: they knew that if his name was on the front page of a report, that would give it automatic credibility. But for exactly the same reason, Litan had a responsibility to be intellectually honest when writing this thing.

Instead, he never even questioned any of the assumptions made by Custodia’s CFO. For instance: if you’re terminated, and you default on your 401(k) loan, what are the chances that the money you received will end up being counted as an “actual distribution” rather than as a “deemed distribution”? Smart and Litan and Singer all implicitly assume that the answer is 100%, but they never spell out their reasoning; my gut feeling is that it’s not nearly as clear-cut as that, and that it all depends on things like when you lost your job, when you defaulted, and who your pension-plan administrator is.

Custodia’s business, and the Litan-Singer paper, are based on the idea that if people who borrowed money from their 401(k) plans had insurance against being terminated from their jobs, then that would have significant societal benefit. In order for the societal benefit to be large, the quantity of annual 401(k) loan defaults due to termination also has to be large. But right now, there’s not a huge amount of evidence that it actually is: in fact, we really have no idea how big it is.

I can say, however, that Custodia has already won this battle where it matters — in the press. “Protecting 401(k) savings from job loss makes a lot of sense,” said Time’s Kadlec in his post — and so long as Custodia can present lawmakers with lots of headlines touting the $37 billion number and supporting their plan, Litan and Singer will have done their job. The truth doesn’t matter: all that matters is the headlines, and the public perception of what the truth is.

Come to think, maybe this makes Litan the absolutely perfect person to run the research department at Bloomberg Government. On the theory that it takes a thief to catch a thief, Bloomberg has hired someone who clearly knows all the tricks when it comes to writing papers which come to a predetermined conclusion. And he also has a deep understanding of the real purpose of most of the white papers floating around DC: it’s not to get closer to the truth, but rather to stamp a superficially plausible institutional imprimatur onto a policy that some lobbyist or pressure group desperately wants enacted. I can only hope that in the wake of using his talents in order to serve Custodia Financial, Litan will now turn around and use them in order to serve rather greater masters. Like, for instance, truth, and transparency, and intellectual honesty.

Why social mobility is important

Felix Salmon
Jul 30, 2012 21:46 UTC

Tim Harford is a fan of the clear way in which Alex Tabarrok has couched the debate — which started with a Tyler Cowen post back in January — about the desirability of intergenerational economic mobility. Or, in English, is it a good thing if quite a lot of poor people become rich?

The Marginal Revolution guys say that looking at economic mobility is overrated; Cowen, also in January, linked to a bunch of critics of that position, including John Quiggin, Brad DeLong, and Paul Krugman. Recently, DeLong resuscitated the discussion, and Krugman came back for a second go-round as well, all of which resulted in Cowen being rude about Krugman, and Tabarrok trying to clear things up.

Tabarrok’s post is indeed clear, but it’s clear in an invidious way. He basically starts with his conclusion, saying that if a high-mobility society has no better outcome, in general, than low-mobility society, then there’s not very much to choose between them. And similarly, he says, if both a high-mobility society and a low-mobility society have the same very good outcome, then again there’s not much to choose between them.

But this obtusely misses the fundamental reason why high mobility is a good thing: that it improves outcomes. A sclerotic society where no rich people become poor and where no poor people become rich is never going to be a hive of creative destruction. Cowen even comes close to admitting this, when he says that “if the general standard of living is rising, mobility takes care of itself over time” — except he has the causality largely backwards. If you have lots of social mobility, then the general standard of living is going to go up: you’ll have lots of poor people becoming richer, and you’ll also have the rich protecting their downside, in the likely event that they become poorer, by doing their best to improve the lot of the poor.

So when Cowen talks about economic mobility not mattering much “for a given level of income”, or when Tabarrok talks about “some simple societies” with fixed levels of income, they’re taking the variable in the equation and they’re turning it into a constant. What they should be doing is looking at two societies, equal in all respects except that one is high-stasis and the other is high-churn, then fast-forwarding to see which one turns out better. The answer, of course, is the high-churn society — which means, working backwards, that if you want growth, you also want social mobility.

As a result, it’s reasonable to conclude that anything which impedes social mobility — like rising inequality, say — also impedes growth. The effect might not be huge, but it’s there. And the only way not to see it is to effectively assume your conclusions.

The sensible hunt for manufacturing jobs

Felix Salmon
Jul 12, 2012 16:29 UTC

Michael Kinsley tackles outsourcing today, complaining that Barack Obama is a protectionist who doesn’t understand its value, and that Mitt Romney is keener to pander to protectionists than he is to defend free-market principles. He writes:

Romney or Obama? “I don’t want the next generation of manufacturing jobs taking root in countries like China or Germany.” Early in the Republican primary campaign, China was the one subject Romney seemed genuinely agitated about. Imposing tariffs on Chinese goods was on the long list of things Romney said he was going to do on Day One of his presidency. Maybe he still is, but he doesn’t play it up the way he used to.

Meanwhile, if Romney is a free trader at heart, faking a bit of protectionism, Obama seems to be a protectionist at heart, faking a belief in free trade. That quote in the previous paragraph is from Obama, and shows a fundamental misunderstanding of how markets work. Trade is not a zero-sum game. There isn’t a certain number of manufacturing jobs that will either go to China or Germany, or come to us. We want China and Germany to have lots of manufacturing jobs. The more they have, the richer they are, the better off we will be as well. Beggar-thy-neighbor policies don’t work.

Kinsley is probably right on the politics, here, but he’s wrong on the economics. Here’s Obama’s quote, in context:

I was able to sign trade agreements with Korea and Colombia and Panama so our businesses can sell more goods to those markets. That’s why I’ve fought for investments in schools and community colleges, so that our workers remain the best you’ll find anywhere, and investments in our transportation and communication networks, so that your businesses have more opportunities to take root and grow.

I don’t want America to be a nation that’s primarily known for financial speculation and racking up debt buying stuff from other nations. I want us to be known for making and selling products all over the world stamped with three proud words: “Made in America.” And we can make that happen. (Applause.)

I don’t want the next generation of manufacturing jobs taking root in countries like China or Germany. I want them taking root in places like Michigan and Ohio and Virginia and North Carolina. And that’s a race that America can win.

This, it seems to me, is an entirely coherent economic policy if what you’re trying to do is maximize the number of good working-class jobs in America. There’s no doubt that US employment, as a whole, is on a long-term secular trend away from goods and towards services. And at the same time, the two countries with the world’s biggest trade surpluses — China and Germany — are precisely the two countries with the healthiest manufacturing industries. What’s more, neither of them is suffering a jobs crisis.

So the question arises: should the US continue to accept the Ricardian bargain whereby it concedes to China and Germany the comparative advantage in manufacturing, while keeping for itself the comparative advantage in borrowing-to-consume and constructing synthetic CDOs? The answer is no, and not only because there’s something hollow and dangerous about an economy which is too reliant on financial whiz-bangery. There’s something more important at stake here, and that’s employment.

US manufacturing in fact is extremely competitive on a global scale; the problem is that output has lagged productivity improvements, with the result that we’re making more stuff with ever fewer people.

There’s no particular reason why that should be the case: when manufacturers in China and Germany become more efficient, that’s their sign to employ more people, rather than fewer. As each employee becomes increasingly profitable, it makes perfect sense to keep on adding more employees. Or at least it does in some countries. In the US, by contrast, capital is cheap and plentiful, and there’s much more incentive here to replace people with capital goods wherever possible.

But at that point, why even invest the money in the manufacturing industry at all? Everything becomes a question of opportunity cost, and if you can get higher returns in say the financial sector, then the rational thing to do is to start an investment bank, make lots of money from your trading desk, and then take the proceeds and spend them on manufactured goods from China and Germany. That’s how markets work: goods and money become interchangeable, and if you have money then you don’t need to be able to actually make things any more. Money gives you all the competitive advantage you need.

Except, that’s a strategy which works until it doesn’t. I’m reminded of this bit from Kurt Eichenwald’s piece on Microsoft’s Steve Ballmer in the latest Vanity Fair:

The Microsoft CEO used to proclaim that it would not be first to be cool, but would be first to profit — in other words, i would be the first to make money by selling its own version of new technologies. But that depended on one fact: Microsoft could buy its way into the lead, because it always had so much more cash on hand than any of its competitors.

No more. The advantage that Ballmer relied on for so long is now nonexistent. Google has almost the same amount of cash on its books as Microsoft — $50 billion to Microsoft’s $58 billion. Apple, on the other hand, started the year with about $100 billion. Using superior financial muscle won’t work for Microsoft or Ballmer anymore.

A technology company’s ability to innovate is not a bad metaphor for an economy’s ability to manufacture things and employ people while doing so. If it’s lacking, then for a certain amount of time that hole can be patched with money. But not forever.

And so I think that Barack Obama’s push to bring manufacturing employment back to Michigan and Ohio and Virginia and North Carolina makes all the sense in the world. Trade is not a zero-sum game — Kinsley is right about that — but at the same time that’s no reason to feel sanguine when you see good working-class jobs get exported to countries where the idea of building a blue-collar career in the manufacturing sector is still a perfectly sensible and reasonable one. America does not have a jobs crisis among college graduates, even if the employment situation for recent graduates right now is grim. It does have a jobs crisis in areas where factories have closed and industrial skills are no longer valued.

If you run a company, one of your jobs is to ensure that your company’s money isn’t wasted. And similarly, if you run an economy, one of your jobs is to ensure that your country’s labor force isn’t wasted. There are far too many Americans, right now, who could be working and aren’t. That’s downright inefficient. Their skills are well suited to the manufacturing industry. And so, if new manufacturing jobs are to be created, somewhere in the world, it makes a huge amount of sense that they be created in places like Michigan and Ohio and Virginia and North Carolina: that’s where the low-hanging fruit lies, in terms of hard-working employees with enormous potential productivity gains.

Kinsley’s right that we want China and Germany to have manufacturing jobs. But here’s the thing: China and Germany have manufacturing jobs. New manufacturing jobs, at least in the short term, should move where there’s both a shortage of such things right now, and a large potential labor force willing to get to get to work tomorrow. Certainly it’s the job of the president to encourage precisely that. And if he succeeds, he will have built a powerful economic engine in a part of America which is not doing well at the moment.

Note that Obama talked about bringing those jobs to America not in any protectionist context, but rather in the context of a series of free trade agreements which, at the margin, make outsourcing easier rather than harder. His goal is not to steal jobs from elsewhere, but rather to make America a place where the infrastructure and workers are so attractive that companies around the world decide to source their manufacturing here. No one cares, any more, about the nationality of the employers in these states: a job is a job, whether your employer is American or Brazilian or Swedish.

The computer I’m writing this on was made by an American company in China, but there’s no particular reason why other items in my household shouldn’t be made by a Chinese company in America. If the US can create the conditions for that to happen — if Chinese companies voluntarily move some of their manufacturing here because that’s how effective the US manufacturing sector has become — then everybody wins. That’s what Obama wants. And wanting that requires no misunderstanding whatsoever of how markets work.

Adventures with marginal pricing, auto edition

Felix Salmon
Jul 2, 2012 23:23 UTC

Brian Chen has the news today that Uber is rolling out a cheaper version of its service:

Uber’s convenience comes with a cost. People are paying not just for the service, but also the gas used by the big sedans. That’s where hybrid vehicles will help bring down the price: drivers will spend less time and money fueling up…

In San Francisco, for example, the hybrid cars will cost $5 for the base fee, and then $3.25 a mile after that. By contrast, the town cars cost $8 for the base fee and then $4.95 a mile.

A quick back-of-the-envelope calculation shows that this has very little to do with the amount of money that drivers spend fueling up. Compare a Prius (51 miles per gallon) to an Escalade (10 miles per gallon): if gas is $3.78 per gallon, that puts the cost of gas per mile at 7.4 cents for the Prius and 37.8 cents for the Escalade — a difference of 30 cents per mile. Whereas Uber’s price for the Escalade is a premium of $1.70 per mile.

What’s more, since the drivers of these cars can’t pick up hails on the street, they have a lot of downtime waiting for the next gig. As a result, it doesn’t really cost the Escalade driver extra money if she ends up having to refuel once a day rather than once a week. Obviously, the fuel costs are higher — but the opportunity cost of her time is negligible.

Here’s Chen:

The company convinced its car-service partners to buy a total of 50 hybrids just for customers coming through Uber — a sign that drivers are making money with the start-up.

But of course it’s more complicated than that. If drivers were happy with the money they were making with Uber, then they’d stick happily with what they’ve got. In order to be persuaded to switch over, they have to believe that they’ll make more money in a hybrid than they would in a sedan. And that’s despite the fact that “in general”, according to Uber’s Scott Munro, “hybrids will cost 30 to 40 percent less than Uber’s black town cars”.

If that’s the case, then if you compare a sedan driver and a hybrid driver, the hybrid driver will need to be making three trips for every two the sedan driver makes, just to end up with the same amount of money. In order for the hybrid to be more attractive than the sedan, and taking into account the fact that at the margin you’d rather make fewer trips than more trips, a typical driver would realistically be hoping to double the number of fares she was getting before she was happy switching to the cheaper car.

But I suspect that the real relationship here is not between Uber and its drivers, so much as it is between Uber and car-service companies. Any given driver might well prefer to continue driving a sedan, rather than being moved over to a hybrid. But the car-service companies make money on every fare, and so their best interest is served just by increasing the total number of fares, rather than the average income received per driver per day.

As a result, I suspect that this move is going to decrease Uber drivers’ take-home income, on average, rather than increase it. As you might expect, when prices drop. But it will increase income for both the car-service companies and for Uber itself — and it will increase the total number of Uber drivers.

It’s easy to sign up with Uber if you’re a company; much harder if you’re a single driver. The Uber model is that Uber contracts with the owners of capital, who then employ the labor needed to provide the service. And once again, the rich will end up making more, the not-rich will end up making less, and the rich will present the whole thing as a victory for all concerned.

But there’s something else going on here, too, which is the way that companies love to push the idea that we’re paying for extra costs, even when we’re not. Uber sedans are more expensive than Uber hybrids because Uber reckons that’s the way it can best maximize its revenues and profits — not because the sedans are significantly more expensive to drive. Another example of this? Gas stations which offer different prices for cash and credit.

I like this idea, in theory, because gas prices are the most salient prices in America: we’re much more conscious of how much gas costs than we are of how much anything else costs. And if the price for gas on credit is significantly more than the price for cash, then that will help drive home just how big those credit interchange fees are.

Except, gas stations have no particular reason to charge just the interchange fee as a premium. Is the difference 10 cents a gallon? That’s about 3%, which is at the high end of credit interchange fees. After that, it’s all just pure profit for the gas station — and sometimes the difference can be as much as 2 dollars a gallon.

That isn’t a condign surcharge; it’s price gouging. And even a relatively common 20-cent surcharge is basically a convenience or ignorance fee, a way of extracting extra money from people who don’t have the cash or who don’t realize how much extra they’re paying. The rate of paying-with-plastic ranges between 60% and 100%, which means that realistically what we’re talking about here is essentially a bait-and-switch: attract customers with a low headline price, and then charge them a higher one.

Part of modern life is the way in which we naturally gravitate towards easy and automatic ways of paying. If you give Uber your card number once, you never really need to pay at all; you just find the charge on your credit-card statement. It’s certainly convenient — but it also allows Uber to charge quite enormous sums for what they provide. And similarly, at the gas pump, we just want to swipe our cards and get out of there, rather than faffing about with cash. And so there’s an incentive for companies like Uber and gas stations to inexorably increase the implied convenience fee we get charged for using easy payments methods — even if those payments work out cheaper for them. (After all, it would cost Uber a fortune if we paid our drivers in cash and then Uber had to try to reclaim its share from those drivers.)

My radical new universal payments system would help a little bit here, since it would make it impossible for vendors to claim that the more convenient payments method was somehow more expensive for them. But it wouldn’t solve the deeper problem, which is that the more painless payments are, the less we feel the pain. And so merchants will always find ways to charge us more now, if we’re not going to really feel how much we paid until much later. And then, when customers start revolting at the high prices they’re paying, the merchants will act like they’re doing us a favor by offering an inferior and cheaper option.

Bishop vs Krugman

Felix Salmon
Jun 18, 2012 23:16 UTC

Paul Krugman was not happy with the choice of Matthew Bishop to review his new book in the NYTBR, and the main locus of the disagreement seems to be, at heart, how much respect Krugman should give to people who disagree with him.

Here’s Bishop:

No opportunity to preach to the choir is missed by the populist Mr. Krugman, nor any chance to mock those he calls the “Very Serious People” who disagree with him. This is often entertaining: during a stern speech in 2010 by Germany’s finance minister, Krugman’s wife dismissed those who regard austerity as a sort of moral purification with the whispered aside, “As we leave the room, we’ll be given whips to scourge ourselves.” But the book’s preachiness gives those politicians and economists who most need to read this book an easy excuse to ignore it.

To this Moderately Serious Reviewer, Krugman’s habit of bashing anyone who does not share his conclusions is not merely stylistically irritating; it is flawed in substance… The austerians may be excessively fearful of so-called “bond vigilantes,” but that does not mean there is no need to worry about what investors think about the health of a government’s finances. Sure, ridicule those fundamentalists who believe it is theoretically impossible for an economy ever to suffer a shortage of demand, but does Krugman really need to take passing shots at Erskine Bowles and Alan Simpson, the chairmen of the widely respected bipartisan Bowles-Simpson Commission on deficit reduction appointed by President Obama? Maybe his case for stimulating the economy in the short run would be taken more seriously by those in power if it were offered along with a Bowles-­Simpson-style plan for improving America’s finances in the medium or long term. Instead, Krugman suggests cavalierly that any extra government borrowing probably “won’t have to be paid off quickly, or indeed at all.”

I can see why Krugman finds this annoying. Krugman’s whole point is that Bowles, Simpson, and the like are wrong and dangerous. And as he reminds us today, he was right and they were wrong, two years ago. He should get credit for that. But Bishop, the kind of person who loves nothing more than schmoozing important people at Davos, thinks that Krugman “would be taken more seriously” if he were more polite to “widely respected” people with the word “chairman” in their names.

This criticism is off-base for three different reasons, I think. Jared Bernstein deals with the substance very well:

Krugman has been consistently empirical on this point. His argument is not that investors’ sentiments don’t matter. It’s that they’re embedded in prices and can be followed on an hourly basis. Those numbers—the bond yields on sovereign debt—show that markets judge US debt to be safe and Spanish and Greek debt to be risky. If you want to criticize Krugman on this count, you need to explain what’s wrong with the markets themselves—why they’re giving the wrong signals. Otherwise, you’re into phantom-menace land, just across the way from where the confidence fairy hangs out.

This is a point I myself tried making to Bishop back in April, with no visible success: Bishop’s convinced that when it comes to gauging future inflation expectations, we should for some reason trust the volatile and largely-insane gold market at least as much as we should trust the most liquid and efficient market in the world, that for US Treasury bonds.

As for the style, there is no shortage of Serious liberals willing to do exactly what Bishop suggests. Indeed, Erskine Bowles probably counts as one himself, even as he sits on the board of Morgan Stanley. Pretty much the entire Obama administration deals constantly with calls for fiscal prudence and austerity, and takes them very seriously. There’s something of a bipartisan consensus on the issue — so if like Krugman you think that the consensus is bonkers, the only real way to get your point across is to be very clear that no matter how grand these people are, they’re simply wrong, and do not deserve to be taken seriously.

And then there’s the whole class-based undertone to the discussion, which I think if anything Krugman doesn’t make forcefully enough. The thing that Serious liberals and Serious conservatives have in common — the thing which in large part makes them “widely respected” in the first place — is that they’re rich. Usually, very rich. And rich people, as I said in my own review of Krugman’s book, don’t actually worry much about unemployment: it doesn’t really hurt them, even if they lose their jobs. What they do worry about is inflation, since that erodes the value of their dollars. And so when Krugman calls for a nice dose of inflation to help cure the economy’s ills, what he’s really calling for is for a significant chunk of the fixed-income portfolios of the rich to be devalued in real terms.

The rich don’t like that, and the austerity consensus is in large part a closing of ranks — one of the few areas where left and right can agree, at least at the upper end of the income spectrum. And that’s why my own review of Krugman’s book was a pessimistic one. When rich liberals and rich conservatives agree on something, that thing is going to happen. Especially when that thing is in their own self-interest.

Lessons in pricing a scarce resource

Felix Salmon
Jun 9, 2012 00:08 UTC

There’s a fine art to pricing any scarce resource. Ex ante, it’s impossible to do a precise calibration of supply and demand, but being able to do so is crucial to getting things right. If you’re in a business where you can make more of whatever you’re selling when demand rises, that’s one thing. But when you’re selling tickets, or Facebook shares, that’s not the case.

In a world where you have to set the price in advance, and then it can’t be changed, the calculus is simple. Set the price too high, and you end up with insufficient demand and a general feeling of failure; you don’t attract the number of people you were hoping for, and even those people are likely to end up feeling ripped off. On the other hand, set the price too low, and you create disappointment among people who wanted to give you their money and can’t, quite aside from the fact that you’re clearly leaving money on the table.

In the past couple of days, we’ve seen good examples of both. At Yankee Stadium, the price of tickets is way too high, as is evidenced by the huge number of empty seats, and by the fact that on the secondary market, two thirds of tickets are sold for less than face value. Lots of seats are being sold by people asking less than $5 a pop, and at the top of the ticket-price range, the average discount to face value is more than $90.

At the same time, sold-out $125 tickets for the Big Apple Barbecue Block Party are being hawked on Craigslist for significant premiums to face value, prompting Ryan Sutton to declare that they should be more expensive next year.

The Yankees are taking a shoot-the-messenger approach to their attendance problems, blaming the secondary market in tickets, rather than the fact that the tickets cost a fortune. That’s just silly, and it’s a no-brainer that the price of Yankees tickets should come down. Just like an IPO, you want to price season tickets so there’s a small implied “pop” in there — people with season tickets should be able to sell them on the secondary market for a little bit more than they paid. That helps keep demand for season tickets healthy, year in and year out.

What’s more, the Yankees have the same stupid pricing as the Metropolitan Opera: every game or opera costs the same amount, no matter how in-demand or run-of-the-mill the matchup. Pretty much every other baseball team has pricing variable enough that at least the big games cost more; the Yankees should take a leaf out of Broadway’s book and do the same. Broadway pretty much always sells out every show, these days, at whatever the clearing price is, and scalping is way down. That would make Yankees games much less desolate.

Pricing the barbecue tickets is trickier, but Sutton is right: when you’re raising money for charity, it’s a little heartbreaking to see tickets being flipped for profit. And the cost of setting the price too high is small: if the tickets look as though they’re not going to sell out, you just run some kind of special offer where people can buy them at a discount for a limited period of time. No harm, no foul.

And what about IPOs? With them, there’s no do-over, and the process tends to be driven very much by big investment banks with more than half an eye on their reputation in the equity capital markets. They care about making money on every deal, but they care much more about getting a healthy stream of fee income from future deals. While the barbecue can overprice its tickets without too much damage, investment banks don’t have the same luxury.

And that’s probably the real reason why there’s an IPO pop. Underpricing the IPO might mean that the issuing company is leaving money on the table — but overpricing the IPO is much worse, as Facebook and its underwriters are still discovering. So banks always err on the side of underpricing. Except, as in this case, when the issuer has too much power, and gets too greedy.

Greg Ip’s risk hairball

Felix Salmon
Jun 4, 2012 17:42 UTC

Greg Ip is getting in on the probabilities game, this time looking at the three big risks facing the global macroeconomy.

Ip puts the chance of a Chinese hard landing at 20%; of the euro falling apart at 40%; and of the US fiscal cliff actually happening at 30%. Individually, each of these risks is bearable. But make the reasonable assumption that they’re independent variables, and it turns out that if you put them all together, the chances of none of them happening are just one in three.

But let’s go a bit further. Let’s say that if none of these things happen, that’s Good. If one of these things happen, that’s Bad. If two of these things happen, that’s Dreadful. And if all three of these things happen, that’s Apocalypse.

Then this is the result that you get. The chances of a good outcome are 33.6%, a bad outcome is 45.2%, a dreadful outcome is 18.8%, and the chances of apocalypse are a small but still scary 2.4%.

pie.png

You can also look at each possible outcome individually, like this:

Europe US China Probability
happy.jpg happy.jpg happy.jpg 33.6%
sad.jpg happy.jpg happy.jpg 22.4%
happy.jpg sad.jpg happy.jpg 14.4%
sad.jpg sad.jpg happy.jpg 9.6%
sad.jpg happy.jpg happy.jpg 8.4%
sad.jpg happy.jpg sad.jpg 5.6%
happy.jpg sad.jpg sad.jpg 3.6%
sad.jpg sad.jpg sad.jpg 2.4%

The most likely single outcome is the Good one, where everything goes well in all three regions. The next most likely outcome is the bad one where Europe falls apart but the US and China keep things together — that’s 22%. Then comes the bad outcome with a US fiscal cliff while Europe and China muddle through: that’s 14%. And in fourth place is the dreadful outcome where you get the US fiscal cliff and the euro falling apart: that has a substantial 10% probability.

This is what Ip calls the “big hairball of risk”, and it basically explains the flight to quality that we’re seeing globally, with long-term yields below 2% in every major currency in the world. How do you invest in such a world? It’s really hard, but most sensible strategies involve a pretty large degree of downside protection in the form of risk-free assets. “And that sort of disengagement,” says Ip, “can make economic pessimism self-fulfilling.”

Unless, of course, governments take advantage of their ultra-cheap funding to step in with massive economic stimulus.

Silicon Valley hubris watch, TJ Rodgers edition

Felix Salmon
Jun 1, 2012 20:33 UTC

Chrystia Freeland has found a classic example of Silicon Valley hubris in TJ Rodgers, the CEO of Cypress Semiconductor. Rodgers reckons his taxes shouldn’t go up; his reasons are pretty simple. If that happened, he explains, he’d have less money to invest in Silicon Valley. And since “taking money from the investments, my investments, out of Silicon Valley, where they have been very, very good for the economy” is self-evidently a really bad idea, then raising his taxes can’t possibly make any sense.

The first problem with this is that Silicon Valley is not a place where TJ Rodgers’s money is magically transmogrified into growth and jobs. Yes, Silicon Valley is a very innovative place. But it would be a very innovative place even if it had significantly less money than it has right now. I actually lived in Palo Alto in the mid-80s, when Silicon Valley had long since matured as a vibrant technology center, and as I recall it cost about $500 a month to rent a nice family home. Since then, the amount of money in Silicon Valley has increased enormously, along with the price of Palo Alto real estate. But it’s far from clear that all that new money has made the Valley any more innovative.

In fact, we know exactly what would happen if a lot of the wealth in Silicon Valley suddenly got taken away: we know because it happened, when the dot-com bubble burst in 2000-01. The researchers kept on researching, the innovators kept on innovating, and the main effect of the dot-com bust was that bits of San Francisco looked, briefly, as though they might actually become affordable. That didn’t last for long.

But Chrystia herself provides the much stronger rebuttal to Rodgers, when she quotes Nandan Nilekani, the co-founder of Infosys, pointing out that huge amounts of Silicon Valley wealth are built on US government spending.

“It’s the role of governments to create public goods which are platforms for innovation. If you look at the U.S., the Internet was a government defense program on which today you have this huge innovation ecosystem. GPS is another example.” That system “was designed for military applications. But today it’s used for maps or car navigating systems or whatever. So the ideal is to create these global public goods or these national public goods that are platforms. And then make them open so that people can innovate.”

The big money in Silicon Valley these days is very much around anything that can credibly call itself a “platform”. Facebook might be worth a lot less than it was worth a couple of weeks ago, but it’s still worth vastly more than it would be if it were merely a media company making $1 billion a year selling ads. The bulk of Facebook’s value is embedded in the fact that it’s a platform: it’s the tool we use, all over the internet, to be ourselves and to have connections to our friends.

Innovation is in large part about building things on other things: build an iPad location tool on the GPS architecture, for instance, or build a social network using existing internet architecture. And underneath it all is a system of base-level infrastructure which needs to be carefully tended lest it all fall apart. One difference between Rodgers and Nilekani is that Rodgers has been living in Silicon Valley long enough that he takes a huge amount for granted: he has everything he needs to run a great business right on his doorstep. Nilekani, by contrast, needs to deal with obstructive Indian bureaucracy all the time: he knows that India, as a platform, is vastly inferior to the Bay Area.

Once you build a platform, you never know how you might be able to extract value from it. I was talking to the SecondMarket guys yesterday, for instance, and they talked a bit about what they’re calling their new marketing platform. They spent a lot of time building up a huge database of accredited investors interested in putting money into private companies — and now they’re looking to monetize that database by essentially renting it out to other shops looking for that kind of investor.

Art funds, diamond funds, wine funds, distressed-mortgage funds, tax-lien funds, you name it — somewhere in that SecondMarket database there’s a group of people who are likely to be very interested. And SecondMarket itself needs to do very little work, since they’re basically just acting as a high-tech dating agency, matching investors with fund managers. The investors came for the pre-IPO equity; they’ll stay for all the other goodies that SecondMarket can introduce them to.

And the USA is, in many ways, the ultimate platform — a massive market and currency zone, which can provide enormous demand for great products while also providing peace, prosperity, strong and stable institutions, and everything else that someone like Rodgers needs to be successful. It makes sense to invest in that platform — and it makes sense, too, that the people who get the most out of it should be asked to reinvest the most back into it.

You can call that reinvestment “stimulus” if you like, although that word seems to have gone out of favor these days. But what’s sure is that unless the USA keeps its infrastructure up to date, it’s going to lose a lot of competitiveness over the long term. We need a smart energy grid, we need a much better transportation network, we need to improve our educational system, and ultimately we need to have a much more constructive legislature than we have right now. If we fail in that, the whole country will decline, and it will take the likes of Rodgers down with it.

Nick Hanauer, another Silicon Valley multi-millionaire, uses the gardening metaphor: we need to maintain our garden if we’re going to reap abundant crops. So long as people like Rodgers think that the government is good for nothing but misguided cash-for-clunkers schemes, and that the best thing it can do is just get out of their way, they’re going to be the worst kind of free-rider: the kind who doesn’t even know they’re free-riding. Maybe we should tell him to try to build a great technology company in, I dunno, Greece, and see how that works. Only then might he appreciate just how much of his net worth he owes to his country.

When monetary policy needs to incorporate fiscal policy

Felix Salmon
Apr 13, 2012 01:38 UTC

I’m in Berlin this week, at the annual INET meetings, where the big theme this year seems to be an attempt to rope in everyone from anthropologists to neuroscientists in an attempt to solve the big economic problems which are proving intractable to economists. But still, it’s the economics and finance folk who are top of the agenda. And since George Soros is footing a large part of the bill for this conference, he and his latest op-ed are getting star billing. Sadly, however, most of the delegates have been at the conference all day and therefore haven’t had the opportunity to read Mohamed El-Erian’s speech in St Louis, which is equally germane.

The two, in fact, complement each other quite nicely. El-Erian’s main point is that central banks can’t solve the crisis on their own; Soros has an intriguing idea which addresses that fact, and attempts to add some fiscal-policy bite to the operation of monetary policy.

The EU’s fiscal charter compels member states annually to reduce their public debt by one-twentieth of the amount by which it exceeds 60% of GDP. I propose that member states jointly reward good behavior by taking over that obligation.

The member states have transferred their seignorage rights to the ECB, and the ECB is currently earning about €25 billion ($32.7 billion) annually. The seignorage rights have been estimated by Willem Buiter of Citibank and Huw Pill of Goldman Sachs, working independently, to be worth between €2-3 trillion, because they will yield more as the economy grows and interest rates return to normal. A Special Purpose Vehicle (SPV) owning the rights could use the ECB to finance the cost of acquiring the bonds without violating Article 123 of the Lisbon Treaty.

Should a country violate the fiscal compact, it would wholly or partly forfeit its reward and be obliged to pay interest on the debt owned by the SPV. That would impose tough fiscal discipline, indeed.

None of this is easy, but the big-picture idea is a good one, which is to tie monetary policy much closer to fiscal policy, so that fiscal policymakers are no longer capable of leaving all the dirty work to central bankers. In this case, if I understand him correctly, Soros is suggesting that the ECB buy up a large chunk of national sovereign debt every year. If any given country is following the EU’s fiscal-compact rules, then the interest on its debt gets rebated to the country; if it isn’t, then the ECB keeps the interest payments for itself.

If you look at what’s happened since the crisis of 2008, elected policymakers have generally turned to central banks and said “hey, we’re bound by all manner of real and imagined constraints, so please can you do what’s necessary in the short term, since we can’t.” This was quite explicit at the height of the crisis, when Hank Paulson felt quite comfortable telling Ben Bernanke what he needed to do. And then, of course, a couple of months later, Barack Obama appointed Tim Geithner, the country’s second most important central banker, as his Treasury secretary. The effect was to ensure that the central bank would always act in line with what the government wanted — and indeed that is exactly what has happened.

But that course of action has unintended consequences. For one thing, it makes the central bank politically unpopular. And for another, it tends to increase imbalances, including income and wealth inequalities, rather than decrease them. The way that El-Erian sees it, monetary policy during a crisis is like a bridge loan: it’s a short-term way of plastering over the gap, while a longer-term fiscal solution is put together. But if that longer-term fiscal solution isn’t put together, then the world’s central banks will just have built a “bridge to nowhere”, and ultimately made things worse rather than better.

El-Erian gives the world’s central banks an either-or choice: if they don’t manage to build a world where they’re rowing with the fiscal-policy tide, they’ll end up finding themselves “having to clean up in the midst of a global recession, forced de-leveraging and disorderly debt deflation”.

I’m not entirely sure I buy all of the analysis here. Central banks’ policies during the financial crisis were necessary, whether or not fiscal authorities end up doing their bit. While a future fiscal recession would be bad, it’s not at all obvious that we would have been better off just letting the world fall off a cliff at the end of 2008.

But it’s certainly true that central banks aren’t in full-on crisis mode right now. The problems they face — none greater than unemployment, in both its long-term and its youth flavors — are not the kind of things which happen suddenly and need to be addressed with massive and instantaneous liquidity programs. Indeed, they’re the kind of things which really ought to be addressed with fiscal policy. Monetary policy is a blunt tool, and, as El-Erian says, central banks can’t engineer “good inflation”, in things like house prices, without “bad inflation”, in things like oil prices.

And so long as central banks continue to push their zero-rate policies alongside unorthodox measures like LTRO and QE, politicians will find the pressure on themselves being lifted, even as the central banks get most of the market attention. As Ryan McCarthy says, the CNBC types are much more obsessed with the Fed than they are with Treasury: fiscal policy simply isn’t seen as economically important in the way that monetary policy is. Not even in an election year.

El-Erian’s main point is a pretty subtle one: bimodal dynamics mean that that we’re now moving towards a topsy-turvy world where good policy can be bad policy. “A good portion of policy making” he says, “and important underpinnings of conventional portfolio management, are based on a traditional bell curve governing the distribution of expected outcomes.” What that means is that if the central bank makes things better, that’s all good.

In a world where the outcome distribution is not a bell curve, however, the right action for a central bank is not nearly as clear-cut. Think of a bimodal distribution with two peaks: a bad one on the left, and a good one on the right. In that world, loose monetary policy will, at the margin, push the outcome to the right: it will make a good outcome better, and a bad outcome less bad. El-Erian’s point is this: what if, at the same time, by taking pressure off politicians, that kind of monetary policy makes a good outcome less likely, and a bad outcome more likely? If loose monetary policy pushes both peaks a little to the right, but at the same time makes the left-hand peak significantly larger than the right-hand peak, it can still be a very bad idea.

It’s an intriguing idea — but it’s also one which is pretty much impossible to model. And yet, the message of INET, or at least one of them, is that economists should be doing exactly that. Central bankers don’t like to try to anticipate the effect that their actions will have on politicians. But if they don’t, the politicians won’t play ball. And if the politicians won’t play ball, there’s nothing the central bankers can do on their own to avert a major fiscal crisis.

The Europe debate

Felix Salmon
Apr 10, 2012 01:00 UTC

Remember the Krugman vs Summers debate last year? That was fun, in its own way. But this year’s Munk Debate looks set to be simply depressing. The invitation has the details: the motion is “be it resolved that the European experiment has failed”. And I’m reasonably confident that the “pro” side — Niall Ferguson and Josef Joffe — is going to win.

That’s partly because Ferguson has the public-speaking chops to dismantle his meeker opponents, Peter Mandelson and Daniel Cohn-Bendit. Ferguson is likely to go strongly for the jugular, while Mandelson and Cohn-Bendit will noodle around ineffectually, hedging their conclusions and sacrificing rhetorical dominance for the sake of intellectual honesty.

You can see this, already, in the invite. Each speaker is introduced with a one-liner; Ferguson says that Europe is conducting “an experiment in the impossible”, while Mandelson says that Europe is, um, “getting there” and that the world is “very impatient”. Cohn-Bendit is weaker still: his quote, “We need a true democratic process for the renewal of Europe, in which the European Parliament has to play a central role,” seems to imply that Europe really is doomed, since there’s no way that the European Parliament is going to play a central role in anything, except perhaps an expenses scandal.

It wasn’t all that long ago that public intellectuals could make a coherent case that European union, both political and monetary, was and would be a great success story. In the wake of Greece’s default, however, and credible beliefs that Portugal is likely to follow suit, disillusionment and pessimism is the order of the day. The era of great European statesmen is over; in their place, we have David Cameron.

I was a believer in the European experiment; indeed, I thought it had a kind of grand historical inevitability to it, and that a strong whole could be made up of vibrant and disparate parts. And from a big-picture historical perspective, Europe is indeed a success: a bloody and war-torn continent has transformed itself into a political union where it’s unthinkable and impossible for one member state to invade another. But if by “the European experiment” we mean the euro, that’s been a disaster, and virtually everybody in Europe would have been better off had it never existed.

In this, curiously, the broad European population was much more prescient than the educated and political elites, who in large part imposed the euro on their unthankful and unwilling countries. Mandelson is a key member of that elite, and he was wrong about the euro and about the advisability of the UK joining it. It’s going to be very hard indeed for him to persuade an audience of Canadians that this time he’s right. Or, for that matter, that they should in any way welcome the prospect of a monetary union with Iceland.

  •