In recent years the RethinkX think tank has been the source of some of the more interesting analysis of the major problems facing the world today (energy, climate, food production)--its address of those issues theoretically rigorous, data-based, bold in its conclusions, and while not slighting the challenges and dangers the world faces, managing to offer intelligent arguments for something other than hopelessness in the face of exceedingly daunting difficulties and grave dangers. Where the climate-energy problem is concerned they see the solution lying in the falling price of renewable-supplied electricity, a pattern RethinkX's authors have compared to the falling price of computing power.
Recently the RethinkX team's Adam Dorr and Tony Seba have produced a report ("The Great Stranding: How Inaccurate Mainstream LCOE Estimates are Creating a Trillion-Dollar Bubble in Conventional Energy Assets") on one of the less obvious consequences of the kind of energy transition they anticipate--namely that new investment in older, conventional electricity generation in coal, gas, nuclear and hydro power will prove to be "stranded," unviable investment. This will be due not to changes in laws forcing "greener" energy on utilities, or government subsidies helping to make it happen, but rather the sheer market forces that have already made renewables cheaper than the older options, such that photovoltaic Solar and onshore Wind, in combination with likewise cheapening Battery backup ("SWB"), are well on their way not only to being the cheapest source of electricity around by 2030, but to providing electricity at "near-zero marginal cost." The result is that selling electricity provided by the other sources at anything approaching break-even cost, never mind a profit, becomes an increasingly rare occurrence, driving down the capacity usage of fossil fuels, nuclear and hydro-based generation to financially unviable levels, with coal's recent past the very near future of the rest. (America's dwindling, aging coal plant fleet, in spite of government solicitousness, and the improvement of the average by the shutting down of the least competitive, averaged a mere two-fifths utilization rate as of 2020, compared with two-thirds in 2010. As a result coal's share of American electricity production fell by half from 40 to just 20 percent. As RethinkX acknowledges cheap fracking-produced gas was the principal cause in coal's decline, but even cheaper SWB will be factoring more heavily in every case in the years to come.)
In spite of this recent history (to say nothing of the pressure on business from civil society to move toward zero-carbon operations) business is continuing to invest trillions in indifference not only to the ecological implications but the dollars-and-cents facts as RethinkX have reported them--implying enormous confidence in the enduring viability of those more conventional sources. One possible conclusion is that they know something the rest of us do not. The other is that they are profoundly misinformed and making a very, very big mistake. It is the latter conclusion that RethinkX draws, holding that the investment is being driven by deeply flawed studies from such institutions as the Energy Information Administration (EIA) that greatly exaggerate the likely capacity usage of all those traditional sources, and thus greatly understate the "Levelized Cost Of Electricity" (LCOE) that takes into account all the associated expenses divided per kilowatt-hour. To cite an obviously glaring example they presume an 85 percent capacity usage rate for coal all the way through 2060, a figure absolutely without basis in recent history--the average capacity usage just 56 percent in 2010-2019, 52 percent in 2015-2019, and 47.5 percent in 2019, the trend fairly consistently one of decline even before the exceptional year of 2020 dragged the figure down so much further in the way noted above. Using more realistic figures, which have electricity providers recouping their costs from far lower capacity usage--far fewer per-kilowatt hours sold--translates to a higher LCOE,with the real per-kilowatt hour LCOE of a coal plant established today not 7.6 cents as the EIA said, but 32.4 cents, over four times as much. Thus does it go with the other sources (gas, nuclear, hydro), with the gap only going up over time (with a coal-fired plant established in 2030 likely to have to sell its electricity at 65 cents per-kilowatt-hour).
Quite blunt about the ill-foundedness of the more commonly touted figures (a "dogma . . . promulgated by a small number of self-appointed authorities within the electric power sector . . . that confirms and amplifies a fixed set of thoughts, beliefs and biases" in favor of conventional energy sources and against renewables), the result has been a "financial bubble" in coal, gas, nuclear and hydro that the RethinkX authors compare to the subprime mortgage bubble. This is not least in regard to the risk that, after having made massive investments that were not just ecologically terrible but quite stupid from a business standpoint, the public will be called on to rescue the parties that made those investments at its expense in the same manner seen time and again these past four decades. (Dorr and Seba, of course, argue that any such bailout should be opposed, and in the meantime everyone doing what they can to minimize the danger of such an outcome--investors steering clear of such investment, and financial lenders and government overseers not enabling those who would make the investment anyway.)
It is another bold prediction from the RethinkX team--the more striking because so few are questioning that massive investment in these older electricity generation technologies we are seeing on financial (or for that matter, any other) grounds. Indeed, it can seem as if the mainstream of the business world, and the media prone to faithfully promulgate its prejudices; and RethinkX; are in two entirely different worlds. Which of these is living in the real one, which in the fantasy? As it happens, the former have been relentlessly biased in favor of fossil fuels and nuclear power, and ferociously hostile to renewables, with all the fury of established businesses (and their allies) looking at a potentially disruptive force--and consistently underestimated, and even strained to deny, the long-term improvement in the economic viability of renewables (something to remember when looking at today's headlines, with their clearly delighted expressions of a "comeback" for fossil fuels and nuclear as, overlooking the way inflation is affecting the price of everything, they gloat about the supposed end of an era of falling renewables prices).
For its part RethinkX has not been unknown to be overoptimistic about the rapidity of dramatic and potentially beneficent technological change as their earlier report on Transportation-as-a-Service makes clear. Still, in the case of self-driving cars RethinkX was wrong about a technology that was as yet poorly understood by its own developers (the potential of today's machine learning, running on today's computers, to acquire an imperfectly specified level of competence at driving), which necessarily made any extrapolations that much more tenuous. By contrast, when it makes comparisons between solar and coal, for example, it is discussing developments that have for the most already happened, not only in their view but the view of other, less audacious, observers as well. (Lazard's report of last October had unsubsidized utility-scale solar running approximately $30-$40 and wind $25-$50 per megawatt-hour, versus $45-$75 for combined cycle gas, $65-$150 for coal, and $130-$200 for nuclear.)
Once again, solar and wind are cheaper than the established sources, a cheapness sufficient that even with current storage costs added in (which may add a $20-$40 per hour additional cost), they are, and are likely to remain, a good deal cheaper than coal or nuclear generally, and even competitive with gas in many cases. Accordingly the question becomes how much more progress one can expect in further lowering solar, wind and battery costs. As it happened the drop in the median price of solar has slowed, gone from perhaps a fall of 13 percent per year in 2010-2015 to a fall of 7 percent per year in 2015-2020. The price of wind power was almost as brisk, some 6 percent a year in that same 5 year period. The fall in the price of energy storage capacity has been more impressive still, an 18 percent a year drop in 2015-2020. Were the next five years to see only half that rate of progress across the board we would still see a one-sixth drop in the price of solar, a one-seventh drop in the price of wind, and a better than one-third drop in the price of batteries, meaning that in 2025 we would start to find ourselves looking at solar and wind with storage prices beginning under the $40 per megawatt-hour mark in today's terms--cheaper than the low end for combined cycle gas. Were the rate of improvement for 2025–2030 to average just half of the already reduced rate of the preceding five years we would at the end of them see solar and wind 20–25 percent cheaper than they are today, and battery storage 50 percent cheaper, giving us battery-equipped solar-and-wind power for as little as $30 per megawatt-hour, and even at the high end of the price range ($60) compete with medium-priced gas at today's, and likely tomorrow's, rates. (And in the view of proponents like RethinkX there is plenty of reason to expect much, much better than that.)
The result is that unless per-kilowatt-hour progress in renewables and batteries abrupt comes to a halt in the next few years, or technologists dealing with the already very mature technologies of coal, gas, Generation II atomic power or hydro pull off a technical miracle (or rather, several of them), it would be very hard to picture RethinkX being very far off the mark in this case about not only solar but the SWB combination beating the competition on price by 2030 as they predict. Accordingly it seems to me RethinkX's prediction deserves far more attention, and far more respectful attention, than it has been accorded to date--and indeed, far more respect than most of the "experts" the media so relentlessly flatters and promotes even as they so consistently show themselves lacking in the competence that business and society alike so sorely need.
Tuesday, February 22, 2022
Thursday, February 10, 2022
The Backlash Against Cellular Agriculture Arrives--in All its Incoherence
I have recently had occasion to remark that technological hype (not just about single products, but the rate of progress generally) tends to rise and fall in cyclical fashion, with boom followed by bust. Back in the mid-'10s we had a big bubble indeed, with huge expectations regarding artificial intelligence, virtual reality, and much else. This has, of course, long since fallen through, though the shift from treating them as all but inevitable in the very near term to sneering at the thought that they would ever appear set in earlier with some elements of the package than others. Certainly self-driving vehicles were just such a case, the dismissals already commonplace years ago. By contrast the sneers at cellular agriculture were slower in coming--but intense negativity has since got the upper hand, with a significant element in this Joe Fassler's article about it over at The Counter, which has seen some drawing comparisons between the cellular agriculture startups and Elizabeth Holmes' Theranos. (Ouch!)
Still, while all this would certainly seem to be part of the downward turn of that familiar cycle the backlashes against particular technologies, just as they tend to have their own particular timing, tend to have their own particular features. Certainly this was the case with self-driving cars, which many influential persons and interests find objectionable for reasons having absolutely nothing whatsoever at all to do with infeasibility or practical shortcomings.
So does it go with cellular agriculture. Certainly there is no arguing the reality of overoptimism, evident in how supposedly by 2022 people would have been buying the stuff off the market for years, whereas in 2022 the promises of the bullish still have it an indefinite way off as those looking for substitutes to conventional meat are instead offered the non-meat of plant-based products that remain unconvincing to the discerning meat-lover (the more frustrating as, in spite of their lower resource inputs, they never seem to offer no relief to the wallets of hard-pressed consumers). Yet there is also the reality of a "culture war" between those who hope for technological solutions to the world's ecological and other problems, and Luddite-Malthusians disdainful of the same; and in particular the culture war within the culture war ongoing between meat-eaters hoping for "meat without guilt" and vegans treating veganism as an end in itself, such that even if we could have sustainable, eco-friendly, totally animal cruelty-free meat they would still refuse it, and carry on with the missionary work encouraging others to do the same.
As is always the case when issues get sucked into kulturkampf the result is a muddle--with perhaps the most obvious aspect of such the reality that cellular agriculture is not solely about meat production--that while meat due to its greater inputs, higher price and other problems has been the more obvious first target for producers cellular agriculture can equally be used to produce "clean veggies" and "clean fruit," which ought to be of concern even to vegans. After all, contrary to what some seem to believe the ecological problems of modern industrial agriculture do not begin and end with meat production. Meanwhile societies everywhere face the hard facts of growing populations, more uncertain weather and a rising risk from pests due to climate change, a looming nitrogen and phosphorus crisis, an ever-more volatile international financial scene ever-ready to wreak havoc with commodity prices, and the disruptions attendant upon that resurgent danger about which far too many people seem to me to be far, far too casual, war. (How much are we hearing about the risk of a major war over Ukraine hitting global food supplies? Not nearly enough.) In cellular agriculture we have a technology with the potential to delink the production of essential foodstuffs from open-field production in particular geographic locales, an object of a significance far beyond "meat-lovers vs. vegans," and that so many seem to think of the issue in the latter terms is yet another depressing reminder of the lousy job the media does in its science and technology reporting, and the equally depressing consistency with which cultural conflicts drag the discourse down, down, down far below the lowest intellectual common denominator.
Still, while all this would certainly seem to be part of the downward turn of that familiar cycle the backlashes against particular technologies, just as they tend to have their own particular timing, tend to have their own particular features. Certainly this was the case with self-driving cars, which many influential persons and interests find objectionable for reasons having absolutely nothing whatsoever at all to do with infeasibility or practical shortcomings.
So does it go with cellular agriculture. Certainly there is no arguing the reality of overoptimism, evident in how supposedly by 2022 people would have been buying the stuff off the market for years, whereas in 2022 the promises of the bullish still have it an indefinite way off as those looking for substitutes to conventional meat are instead offered the non-meat of plant-based products that remain unconvincing to the discerning meat-lover (the more frustrating as, in spite of their lower resource inputs, they never seem to offer no relief to the wallets of hard-pressed consumers). Yet there is also the reality of a "culture war" between those who hope for technological solutions to the world's ecological and other problems, and Luddite-Malthusians disdainful of the same; and in particular the culture war within the culture war ongoing between meat-eaters hoping for "meat without guilt" and vegans treating veganism as an end in itself, such that even if we could have sustainable, eco-friendly, totally animal cruelty-free meat they would still refuse it, and carry on with the missionary work encouraging others to do the same.
As is always the case when issues get sucked into kulturkampf the result is a muddle--with perhaps the most obvious aspect of such the reality that cellular agriculture is not solely about meat production--that while meat due to its greater inputs, higher price and other problems has been the more obvious first target for producers cellular agriculture can equally be used to produce "clean veggies" and "clean fruit," which ought to be of concern even to vegans. After all, contrary to what some seem to believe the ecological problems of modern industrial agriculture do not begin and end with meat production. Meanwhile societies everywhere face the hard facts of growing populations, more uncertain weather and a rising risk from pests due to climate change, a looming nitrogen and phosphorus crisis, an ever-more volatile international financial scene ever-ready to wreak havoc with commodity prices, and the disruptions attendant upon that resurgent danger about which far too many people seem to me to be far, far too casual, war. (How much are we hearing about the risk of a major war over Ukraine hitting global food supplies? Not nearly enough.) In cellular agriculture we have a technology with the potential to delink the production of essential foodstuffs from open-field production in particular geographic locales, an object of a significance far beyond "meat-lovers vs. vegans," and that so many seem to think of the issue in the latter terms is yet another depressing reminder of the lousy job the media does in its science and technology reporting, and the equally depressing consistency with which cultural conflicts drag the discourse down, down, down far below the lowest intellectual common denominator.
Friday, January 28, 2022
Our Automotive Dystopia and the Hope of the Self-Driving Car
About eighty years ago a classic science fiction short story painted a dystopian picture of a nightmare world where cars made possible vastly bigger cities, after which, as the ratio of cars-to-humans approached one-to-two those cities "choked" on those cars, with "[s]eventy million steel juggernauts, operated by imperfect human beings at high speed" proving "more destructive than war" of property and human life, obscene insurance premiums, and of course, the squandering of a finite oil supply.
The solution the story envisioned was a colossal solar-powered public transport system launched as a public works program.
I can imagine that a good many readers are sneering at this story as some "socialist," "hippie" vision. But it was actually the furthest thing from that. The story is "The Roads Must Roll," by Robert Heinlein, published in the science fiction magazine Astounding when John Campbell was running the show--and in its depiction of a power grab by "Functionalists" a (very) thinly veiled right-wing attack on organized labor and Marxism.
The result is that its looking like a leftie vision is a matter of how times have changed, with our attitudes toward cars, oil, solar energy and public works sucked into the "culture wars" rather than treated as objects of rational appraisal--and it has to be acknowledged also, our having come to take what the story presented as a dystopia utterly for granted--so much so that it seems worth discussing the costs of that "way of life." Back in 2010 a study found that car crashes cost the U.S. economy some $1 trillion a year. (Putting it another way, if the cost of those crashes were a whole national economy by itself it would likely be in the top twenty globally--and one could guess that the figure has only gone up in the past decade.) That same year the country saw nearly 4 million injured in such accidents, 33,000 fatally. (More destructive than war, indeed!) Lest it need saying, all this has given us a $300 billion a year car insurance industry whose premiums weigh heavily on the budgets of the motorists that very few can escape becoming, while the problem cars pose from a natural resources standpoint should need no elaboration these days.
Yet anyone who questions that this is the best of all possible worlds is apt to get a hostile reaction, even when no one mentions anything anywhere near so radical as solar-powered moving roadways--as we see in the wildly exaggerated sneering at the prospect of self-driving cars, or the turn to Transportation-as-a-Service that this might make possible, or even the electrification of the automotive fleet (with, par for the course, the media generally treating the weakest arguments on these scores with great respect and their opponents with none). However, I for one am prepared to declare that a world where we move beyond tens--hundreds--of millions of gas-powered steel juggernauts driven at high speed by imperfect humans as our default way of getting about is likely to be a better one. Indeed, I look forward to a day when people look back at our era and shudder at the insanity of a society that actually relied on the alertness, reflexes and judgment of human beings who were so often sick, tired, distracted, irascible, angry or worse to control such juggernauts crowded together on superhighways as a default mode of transport.
The solution the story envisioned was a colossal solar-powered public transport system launched as a public works program.
I can imagine that a good many readers are sneering at this story as some "socialist," "hippie" vision. But it was actually the furthest thing from that. The story is "The Roads Must Roll," by Robert Heinlein, published in the science fiction magazine Astounding when John Campbell was running the show--and in its depiction of a power grab by "Functionalists" a (very) thinly veiled right-wing attack on organized labor and Marxism.
The result is that its looking like a leftie vision is a matter of how times have changed, with our attitudes toward cars, oil, solar energy and public works sucked into the "culture wars" rather than treated as objects of rational appraisal--and it has to be acknowledged also, our having come to take what the story presented as a dystopia utterly for granted--so much so that it seems worth discussing the costs of that "way of life." Back in 2010 a study found that car crashes cost the U.S. economy some $1 trillion a year. (Putting it another way, if the cost of those crashes were a whole national economy by itself it would likely be in the top twenty globally--and one could guess that the figure has only gone up in the past decade.) That same year the country saw nearly 4 million injured in such accidents, 33,000 fatally. (More destructive than war, indeed!) Lest it need saying, all this has given us a $300 billion a year car insurance industry whose premiums weigh heavily on the budgets of the motorists that very few can escape becoming, while the problem cars pose from a natural resources standpoint should need no elaboration these days.
Yet anyone who questions that this is the best of all possible worlds is apt to get a hostile reaction, even when no one mentions anything anywhere near so radical as solar-powered moving roadways--as we see in the wildly exaggerated sneering at the prospect of self-driving cars, or the turn to Transportation-as-a-Service that this might make possible, or even the electrification of the automotive fleet (with, par for the course, the media generally treating the weakest arguments on these scores with great respect and their opponents with none). However, I for one am prepared to declare that a world where we move beyond tens--hundreds--of millions of gas-powered steel juggernauts driven at high speed by imperfect humans as our default way of getting about is likely to be a better one. Indeed, I look forward to a day when people look back at our era and shudder at the insanity of a society that actually relied on the alertness, reflexes and judgment of human beings who were so often sick, tired, distracted, irascible, angry or worse to control such juggernauts crowded together on superhighways as a default mode of transport.
Revisiting George Friedman's The Next Decade
After publishing his geopolitical forecast for the twenty-first century in The Next 100 Years (2009) George Friedman endeavored to provide a more detailed, explanatory and prescriptive discussion of his expectations for the next years, The Next Decade (2011).
This being 2022 we can look back and consider how he did. My thought is that while he made some good points in the book he did not do very well, particularly where his larger and more radical predictions were concerned. Consider the following:
* Friedman predicted that in the wake of the Great Recession the balance of power in economic life would shift back from the private sector to the public, with coming years looking something more like the post-war period. Of course, no such thing happened. The neoliberal model may stand in lower credit with the general public than ever, but it remains the conventional wisdom of business, government, academia, the media--and the few compromises required by policymakers hewing to the line have been slight indeed. (Britain, for example, may have left the EU--but the neoliberal standards of privatization, deregulation, profligacy with corporate welfare and stinginess with the general public, remain the order of the day, as does the foundation of economies on globalization and financialization.)
* Friedman envisioned the U.S. facing a rapprochement between Russia and Germany producing a Paris-Berlin-Moscow axis. Of course, the reality has been quite different.
* Friedman, who has never let go of his once headline-grabbing and since much derided vision of the U.S. and Japan as geopolitical rivals (The Coming War With Japan), or hewing to a pessimistic appraisal of China's prospects, once again predicted China's proving an also-ran in the 2010s, and the U.S. becoming more concerned with checking Japan's power instead. Again, this is not exactly how things have gone.
As readers of this blog know I think it simple-minded to sneer at prediction, and even forecasting, as so many do (with outrageous smugness). After all, as Nicholas Rescher made clear, we have no choice but to constantly make choices based on expectations of future conditions and the future outcomes of our choices, and most of the time we are correct. (For example, when we go to work in the morning we expect that our workplace will be there and be operative, and we are usually right about that.) In fact we take our capacity for correct prediction so much for granted that we are only aware of making predictions when we face the relatively small number of matters where prediction proves trickier. Still, the difficulty does not in itself give us an "out," and we still find ourselves forced to choose, with our only choice making as accurate a prediction as we can, not least by learning from our past mistakes in such situations. And for me that, and not the denigration of prediction, is the reason to revisit Friedman's work and consider where he went wrong--which, it seems to me, was in his simply reasoning from the wrong premises and indulging biases that had proved unhelpful in the past (his superficial grasp of political economy, his attitude toward Germany and China, for example, his endless attempts to rehabilitate his prediction about Japan). Alas, incorrect premises and problematic biases are not at all rare in the business--and unfortunately the media has been more inclined to simple-mindedly make such "experts" into authorities rather than appraise what they have to offer.
This being 2022 we can look back and consider how he did. My thought is that while he made some good points in the book he did not do very well, particularly where his larger and more radical predictions were concerned. Consider the following:
* Friedman predicted that in the wake of the Great Recession the balance of power in economic life would shift back from the private sector to the public, with coming years looking something more like the post-war period. Of course, no such thing happened. The neoliberal model may stand in lower credit with the general public than ever, but it remains the conventional wisdom of business, government, academia, the media--and the few compromises required by policymakers hewing to the line have been slight indeed. (Britain, for example, may have left the EU--but the neoliberal standards of privatization, deregulation, profligacy with corporate welfare and stinginess with the general public, remain the order of the day, as does the foundation of economies on globalization and financialization.)
* Friedman envisioned the U.S. facing a rapprochement between Russia and Germany producing a Paris-Berlin-Moscow axis. Of course, the reality has been quite different.
* Friedman, who has never let go of his once headline-grabbing and since much derided vision of the U.S. and Japan as geopolitical rivals (The Coming War With Japan), or hewing to a pessimistic appraisal of China's prospects, once again predicted China's proving an also-ran in the 2010s, and the U.S. becoming more concerned with checking Japan's power instead. Again, this is not exactly how things have gone.
As readers of this blog know I think it simple-minded to sneer at prediction, and even forecasting, as so many do (with outrageous smugness). After all, as Nicholas Rescher made clear, we have no choice but to constantly make choices based on expectations of future conditions and the future outcomes of our choices, and most of the time we are correct. (For example, when we go to work in the morning we expect that our workplace will be there and be operative, and we are usually right about that.) In fact we take our capacity for correct prediction so much for granted that we are only aware of making predictions when we face the relatively small number of matters where prediction proves trickier. Still, the difficulty does not in itself give us an "out," and we still find ourselves forced to choose, with our only choice making as accurate a prediction as we can, not least by learning from our past mistakes in such situations. And for me that, and not the denigration of prediction, is the reason to revisit Friedman's work and consider where he went wrong--which, it seems to me, was in his simply reasoning from the wrong premises and indulging biases that had proved unhelpful in the past (his superficial grasp of political economy, his attitude toward Germany and China, for example, his endless attempts to rehabilitate his prediction about Japan). Alas, incorrect premises and problematic biases are not at all rare in the business--and unfortunately the media has been more inclined to simple-mindedly make such "experts" into authorities rather than appraise what they have to offer.
Tuesday, January 25, 2022
Carbon Nanotube-Based Microchips: A (Very Short) Primer
As Moore's Law runs its course those looking forward to continued improvements in computing power necessarily turn their attention in other directions--not least, a turn away from silicon to other materials as the material from which we make our integrated circuits. For decades one of the more promising possibilities has been the use of carbon nanotubes.
Carbon nanotubes are one of many materials formed from an arrangement of carbon atoms in hexagonal (six-sided) structures, themselves arranged in a lattice, with every one of the six atoms connected to and forming part of the adjacent structures of the same type. Left as a flat sheet, the resulting material is called graphene. With a nanotube what happens is that the "sheet" is bent, and its opposite, horizontal ends joined, to form a hollow cylinder.
These nanotubes' ultrathin bodies and smooth walls enable charges to flow through them more rapidly, and at a lower supply voltage, than is the case with silicon, making them potentially faster and more energy-efficient. Their smoother, more energy-efficient structures may also make them a more suitable material than silicon for denser "3-D" chip designs--which would not just put more transistors on a single 2-D surface the way conventional chips do, but layer those 2-D surfaces on top of one another to cram more computing power into a given space. And altogether the combination of attributes has led to speculation about their permitting a thousandfold gain in computing speed over silicon, equivalent to about ten more doublings and an additional couple of decades of Moore's Law-like progress, which as a practical matter that would convert today's personal computers into machines comparable to today's supercomputers.
Of course, in considering all this one has to note that there has been a long record of great expectations and great disappointments in regard to the mass-production of carbon nanotubes, not least because of the old "works well enough in the lab but not ready for real life" problem of consistently getting the required quality at a competitive price when producing at the relevant scale. (Back in 2014 IBM said they would have them in 2020. Well, 2020 has come and gone, with the arrival of the technology, like that of so many others, deferred indefinitely into the future.) Still, it is one thing to acknowledge that the technology has been slower to emerge than hoped, another to write it off--and it may well be that the belated arrival of carbon nanotube-based chips, and the boost they deliver to the continued progress of computing power, will be what opens the way to the next great round of advance in the development of an artificial intelligence sector that it may turn out has been held back from realizing its promises only by the limits of the hardware.
Carbon nanotubes are one of many materials formed from an arrangement of carbon atoms in hexagonal (six-sided) structures, themselves arranged in a lattice, with every one of the six atoms connected to and forming part of the adjacent structures of the same type. Left as a flat sheet, the resulting material is called graphene. With a nanotube what happens is that the "sheet" is bent, and its opposite, horizontal ends joined, to form a hollow cylinder.
These nanotubes' ultrathin bodies and smooth walls enable charges to flow through them more rapidly, and at a lower supply voltage, than is the case with silicon, making them potentially faster and more energy-efficient. Their smoother, more energy-efficient structures may also make them a more suitable material than silicon for denser "3-D" chip designs--which would not just put more transistors on a single 2-D surface the way conventional chips do, but layer those 2-D surfaces on top of one another to cram more computing power into a given space. And altogether the combination of attributes has led to speculation about their permitting a thousandfold gain in computing speed over silicon, equivalent to about ten more doublings and an additional couple of decades of Moore's Law-like progress, which as a practical matter that would convert today's personal computers into machines comparable to today's supercomputers.
Of course, in considering all this one has to note that there has been a long record of great expectations and great disappointments in regard to the mass-production of carbon nanotubes, not least because of the old "works well enough in the lab but not ready for real life" problem of consistently getting the required quality at a competitive price when producing at the relevant scale. (Back in 2014 IBM said they would have them in 2020. Well, 2020 has come and gone, with the arrival of the technology, like that of so many others, deferred indefinitely into the future.) Still, it is one thing to acknowledge that the technology has been slower to emerge than hoped, another to write it off--and it may well be that the belated arrival of carbon nanotube-based chips, and the boost they deliver to the continued progress of computing power, will be what opens the way to the next great round of advance in the development of an artificial intelligence sector that it may turn out has been held back from realizing its promises only by the limits of the hardware.
Understanding Moore's Law
Those of us attentive to computing are generally familiar with the existence of something called "Moore's Law." However, really satisfying explanation of what Moore's Law actually is would seem a rarer thing--with one result a great deal of confusion about what it means.
Simply put, Moore's Law has to do with "integrated circuits," or, in more everyday usage, microchips--small wafers ("chips") of semiconducting material, usually silicon, containing an electronic circuit. Within these chips transistors amplify, regulate and switch the electric signals passing through them, enabling them to store and move electronic data. Placing more transistors inside a chip means that more such activity can go on inside it at once, which gives the chip, and the device incorporating it, enabling more "parallelism"--the ability to do more at once, and therefore to work faster. All other things being equal one can only put more transistors on the same-sized chip if the transistors are themselves smaller--which means that the electrons passing through them travel shorter distances, which increases the speed at which the system executes its operations yet again.
Since their invention in the late 1950s microchip manufacturers have steadily increased the number of transistors in their chips, by shrinking transistor size--a process that also caused the cost of each transistor to fall. In 1965 electronics engineer Gordon Moore published a short paper titled "Cramming More Components Into Integrated Circuits" in which he noted that the "density at minimum cost per transistor" doubled every year. He extrapolated from that trend that in the next five years they would have chips with twenty times as many transistors on them, each costing just a tenth of their 1965 price, and that this pattern would continue for "at least ten years."
Moore's prediction (which, it is worth recalling, he never called a "law") was inexactly borne out during those years. He proved somewhat overoptimistic, transistor density not quite doubling annually, and today, in fact, different versions of this "law" get quoted with varying claims about doubling times. (Some say one year, some say eighteen months, some say two years, while claims about the implications for processing power and price also vary.) However, the swift doubling in the number of transistors per chip, and the fall in the price of computing power that went with it, continued for a lot longer than the ten years he suggested, instead going on for a half century past that point. The result is that where an efficiently made chip had fifty transistors on it in 1965, they now contain billions of transistors—all as the low price of these densely transistorized chips means that hundreds of billions of them are manufactured annually, permitting them to be stuffed into just about everything we use.
Nonetheless, Moore's Law has certain in-built limitations. The most significant of these is the physical limit to transistor miniaturization. One cannot make a silicon transistor smaller than a single nanometer (a billionth of a meter, equivalent to the width of a single atom) after all, while even before one gets to that point shrinking size makes transistors so small that the electrons whose movements they are supposed to control actually pass (or "tunnel") through their walls.
Of course, when Moore presented his "Law" the prospects of single atom-wide transistors, or even tunneling, seemed remote in the extreme. Transistors in 1971 were drawn on a ten micrometer (millionth of a meter) scale—ten thousand nanometers in the terms more commonly discussed today. However, by 2017 the transistors in commercially made chips were just a thousandth their earlier size, a mere ten nanometers across. The following year major chipmakers began the mass-production of mere seven nanometer transistors got underway, leaving very little space for further size reductions.
This has led a good many observers to declare that "Moore's Law is dead," or will be before too much longer. The claim is controversial--perhaps more than it ought to be. After all, no one disputes that chip speeds cannot continue to increase on the basis of reducing the size of the transistors on silicon wafers--and that is exactly what Moore's Law was concerned with, not the possibility or impossibility of continued progress in computing power. The result is that those who are convinced that the tendency to the exponential increase of computing power is virtually bound to continue as before might do better to set aside claims for Moore's Law continuing, and instead speak of Ray Kurzweil's "Law of Accelerating Returns."
Simply put, Moore's Law has to do with "integrated circuits," or, in more everyday usage, microchips--small wafers ("chips") of semiconducting material, usually silicon, containing an electronic circuit. Within these chips transistors amplify, regulate and switch the electric signals passing through them, enabling them to store and move electronic data. Placing more transistors inside a chip means that more such activity can go on inside it at once, which gives the chip, and the device incorporating it, enabling more "parallelism"--the ability to do more at once, and therefore to work faster. All other things being equal one can only put more transistors on the same-sized chip if the transistors are themselves smaller--which means that the electrons passing through them travel shorter distances, which increases the speed at which the system executes its operations yet again.
Since their invention in the late 1950s microchip manufacturers have steadily increased the number of transistors in their chips, by shrinking transistor size--a process that also caused the cost of each transistor to fall. In 1965 electronics engineer Gordon Moore published a short paper titled "Cramming More Components Into Integrated Circuits" in which he noted that the "density at minimum cost per transistor" doubled every year. He extrapolated from that trend that in the next five years they would have chips with twenty times as many transistors on them, each costing just a tenth of their 1965 price, and that this pattern would continue for "at least ten years."
Moore's prediction (which, it is worth recalling, he never called a "law") was inexactly borne out during those years. He proved somewhat overoptimistic, transistor density not quite doubling annually, and today, in fact, different versions of this "law" get quoted with varying claims about doubling times. (Some say one year, some say eighteen months, some say two years, while claims about the implications for processing power and price also vary.) However, the swift doubling in the number of transistors per chip, and the fall in the price of computing power that went with it, continued for a lot longer than the ten years he suggested, instead going on for a half century past that point. The result is that where an efficiently made chip had fifty transistors on it in 1965, they now contain billions of transistors—all as the low price of these densely transistorized chips means that hundreds of billions of them are manufactured annually, permitting them to be stuffed into just about everything we use.
Nonetheless, Moore's Law has certain in-built limitations. The most significant of these is the physical limit to transistor miniaturization. One cannot make a silicon transistor smaller than a single nanometer (a billionth of a meter, equivalent to the width of a single atom) after all, while even before one gets to that point shrinking size makes transistors so small that the electrons whose movements they are supposed to control actually pass (or "tunnel") through their walls.
Of course, when Moore presented his "Law" the prospects of single atom-wide transistors, or even tunneling, seemed remote in the extreme. Transistors in 1971 were drawn on a ten micrometer (millionth of a meter) scale—ten thousand nanometers in the terms more commonly discussed today. However, by 2017 the transistors in commercially made chips were just a thousandth their earlier size, a mere ten nanometers across. The following year major chipmakers began the mass-production of mere seven nanometer transistors got underway, leaving very little space for further size reductions.
This has led a good many observers to declare that "Moore's Law is dead," or will be before too much longer. The claim is controversial--perhaps more than it ought to be. After all, no one disputes that chip speeds cannot continue to increase on the basis of reducing the size of the transistors on silicon wafers--and that is exactly what Moore's Law was concerned with, not the possibility or impossibility of continued progress in computing power. The result is that those who are convinced that the tendency to the exponential increase of computing power is virtually bound to continue as before might do better to set aside claims for Moore's Law continuing, and instead speak of Ray Kurzweil's "Law of Accelerating Returns."
How Powerful Would a Genuinely Thinking Computer Have to Be?
Discussing the prospect of a computer matching or exceeding human intelligence we find ourselves forced to consider just how it is that we measure human intelligence. That in itself is an old and difficult problem, reflecting the reality that there remains considerable disagreement about just what precisely human intelligence even is. However, one approach that has been suggested is to consider the human brain as a piece of computer hardware, and attempt to measure its apparent capacity by the yardsticks we commonly apply to computers. Based on that we then identify the minimum hardware performance a computer would have to have in order to display human-like performance.
How do we go about this as a practical matter? By and large it has been standard to measure computing power in terms of the number of calculations a computer can perform per second. Of course, there are a variety of kinds of calculation, but in recent years it has been common to think specifically in terms of "floating-point operations," in contrast with simpler "fixed point" operations. (Adding 1.0 to 2.0 to get 3.0 is a fixed point operation--the decimal in the same place in all three numbers. However, the addition of 1.2570 to 25.4620 to get 26.719 is a floating point operation, in that the decimal point appeared in a different place in each of the two numbers.) Indeed, anyone delving very deeply into the literature on high-end computers quickly encounters the acronym "FLOPS" (short for FLOating-point operations Per Second) and derivatives thereof, such as "teraflop" (a trillion flops per second), a "petaflop" (a quadrillion flops--a thousand trillion flops)" and "exaflop" (a quintillion flops--a thousand petaflops, or a million teraflops).
With computers' performance measured in terms of floating-point operations per second, those speculating about artificial intelligence attempt to equate the human brain's performance with a given number of flops. Among others, Ray Kurzweil published an estimate in his 1999 book The Age of Spiritual Machines, since revised in his 2005 The Singularity is Near. The principle he followed was his taking part of the nervous system, estimating its performance in FLOPS, and then extrapolating from that to the human brain. Working from the estimate that individual synapses are in performance equivalent to a two hundred flop computer, and the human brain contained some hundred trillion synapses, he conservatively estimated a figure of some twenty quadrillion (thousand trillion) floating-point operations per second--twenty petaflops--then suggested that the brain may actually run at about half that speed, ten petaflops sufficing.
In considering this one should note that other analysts have used quite different approaches, from which they produced vastly higher estimates of the brain's performance. This is especially the case when they assume the brain does not produce consciousness at the level of nerves, but rather at the level of quantum phenomena inside the nerves. (Jack Tuszynski and his colleagues suggested that not tens of quadrillions, but tens of trillions of quadrillions, would be required.) Of course such "quantum mind" theories (the best known exponent of whom is probably The Emperor's New Mind author Roger Penrose) are extremely controversial--as yet remaining broadly philosophical rather than scientific in the sense, with as yet no empirical evidence in their favor, and indeed, critics regarding such notions as mystical in a way all too common when people delve into quantum mechanics. Still, the idea that Kurzweil's estimate of just how much computing power a human brain possesses may be too low a couple of orders of magnitude is fairly widespread, popular science articles commonly citing the figure as an exaflop (a thousand petaflops).
Still, it can be said that the most powerful supercomputers have repeatedly attained and increasingly surpassed the level suggested by Kurzweil over the past decade. The Fujitsu "K" supercomputer achieved ten "petaflops" (ten quadrillion "floating-point calculations") per second back in November 2011. It also had a 1.4 petabyte memory, about ten times Kurzweil's estimate of the human brain's memory. Moreover, the Fujitsu K has been exceeded in its turn--by dozens of other supercomputers according to the latest (November 2021) edition of the TOP500 list of the world's fastest systems, in cases by orders of magnitude. At the time of this writing the fastest appears to be yet another Fujitsu machine, the Fujitsu Fukagu, with a performance of 442 petaflops per second--some forty times Kurzweil's estimate of human brain performance. And of course, present computer scientists have set their sights higher than that. Among them is a joint effort by the Department of Energy, Intel and Cray to build the Aurora, which is intended to be an exaflop-level machine--as a matter of course, running a hundred times as many calculations per second as Kurzweil's estimate of the human brain's performance--while even that seems modest next to a report this very day that the I4DI consortium is shooting for a 64 exaflop machine by the end of this very year (equivalent to sixty times those higher estimates of the brain's performance, and six thousand times Kurzweil's estimate).
Reading this one may wonder why Kurzweil's hypothesis about such a computer matching or exceeding the brain's capacity has not already been tested with results pointing one way or the other. The reality is that in practice supercomputers like these, which are as few as they are because they are so hugely expensive to build (the Fukagu's a billion dollar machine) and to run (their voracious energy consumption a constant theme of discussion of such equipment), are normally used by only the biggest-budgeted researchers for the most computationally intense tasks, like simulations of complex physical phenomena, such as the Earth's climate or the cosmos--or code-breaking by intelligence services. They have only rarely been available to artificial intelligence researchers. However, the recent enthusiasm for artificial intelligence research has reportedly meant that artificial intelligence researchers has been cited as a factor in the development of the next round of supercomputers (not least because of the utility of AI in facilitating their work).
Especially with this being the case it seems far from impossible that this will enable it to yield new insights into the subject--just as this past decade it was already the case that our having faster computers available permitted the striking advances in areas like machine learning that we saw this past decade. Indeed, even as the recent excitement over artificial intelligence turns into disappointment with the realization that the most-hyped applications (like Level 5 self-driving) are more remote than certain loud-mouthed hucksters promised, the continued expansion of computing power offers considerable grounds to not write those prospects off just yet.
How do we go about this as a practical matter? By and large it has been standard to measure computing power in terms of the number of calculations a computer can perform per second. Of course, there are a variety of kinds of calculation, but in recent years it has been common to think specifically in terms of "floating-point operations," in contrast with simpler "fixed point" operations. (Adding 1.0 to 2.0 to get 3.0 is a fixed point operation--the decimal in the same place in all three numbers. However, the addition of 1.2570 to 25.4620 to get 26.719 is a floating point operation, in that the decimal point appeared in a different place in each of the two numbers.) Indeed, anyone delving very deeply into the literature on high-end computers quickly encounters the acronym "FLOPS" (short for FLOating-point operations Per Second) and derivatives thereof, such as "teraflop" (a trillion flops per second), a "petaflop" (a quadrillion flops--a thousand trillion flops)" and "exaflop" (a quintillion flops--a thousand petaflops, or a million teraflops).
With computers' performance measured in terms of floating-point operations per second, those speculating about artificial intelligence attempt to equate the human brain's performance with a given number of flops. Among others, Ray Kurzweil published an estimate in his 1999 book The Age of Spiritual Machines, since revised in his 2005 The Singularity is Near. The principle he followed was his taking part of the nervous system, estimating its performance in FLOPS, and then extrapolating from that to the human brain. Working from the estimate that individual synapses are in performance equivalent to a two hundred flop computer, and the human brain contained some hundred trillion synapses, he conservatively estimated a figure of some twenty quadrillion (thousand trillion) floating-point operations per second--twenty petaflops--then suggested that the brain may actually run at about half that speed, ten petaflops sufficing.
In considering this one should note that other analysts have used quite different approaches, from which they produced vastly higher estimates of the brain's performance. This is especially the case when they assume the brain does not produce consciousness at the level of nerves, but rather at the level of quantum phenomena inside the nerves. (Jack Tuszynski and his colleagues suggested that not tens of quadrillions, but tens of trillions of quadrillions, would be required.) Of course such "quantum mind" theories (the best known exponent of whom is probably The Emperor's New Mind author Roger Penrose) are extremely controversial--as yet remaining broadly philosophical rather than scientific in the sense, with as yet no empirical evidence in their favor, and indeed, critics regarding such notions as mystical in a way all too common when people delve into quantum mechanics. Still, the idea that Kurzweil's estimate of just how much computing power a human brain possesses may be too low a couple of orders of magnitude is fairly widespread, popular science articles commonly citing the figure as an exaflop (a thousand petaflops).
Still, it can be said that the most powerful supercomputers have repeatedly attained and increasingly surpassed the level suggested by Kurzweil over the past decade. The Fujitsu "K" supercomputer achieved ten "petaflops" (ten quadrillion "floating-point calculations") per second back in November 2011. It also had a 1.4 petabyte memory, about ten times Kurzweil's estimate of the human brain's memory. Moreover, the Fujitsu K has been exceeded in its turn--by dozens of other supercomputers according to the latest (November 2021) edition of the TOP500 list of the world's fastest systems, in cases by orders of magnitude. At the time of this writing the fastest appears to be yet another Fujitsu machine, the Fujitsu Fukagu, with a performance of 442 petaflops per second--some forty times Kurzweil's estimate of human brain performance. And of course, present computer scientists have set their sights higher than that. Among them is a joint effort by the Department of Energy, Intel and Cray to build the Aurora, which is intended to be an exaflop-level machine--as a matter of course, running a hundred times as many calculations per second as Kurzweil's estimate of the human brain's performance--while even that seems modest next to a report this very day that the I4DI consortium is shooting for a 64 exaflop machine by the end of this very year (equivalent to sixty times those higher estimates of the brain's performance, and six thousand times Kurzweil's estimate).
Reading this one may wonder why Kurzweil's hypothesis about such a computer matching or exceeding the brain's capacity has not already been tested with results pointing one way or the other. The reality is that in practice supercomputers like these, which are as few as they are because they are so hugely expensive to build (the Fukagu's a billion dollar machine) and to run (their voracious energy consumption a constant theme of discussion of such equipment), are normally used by only the biggest-budgeted researchers for the most computationally intense tasks, like simulations of complex physical phenomena, such as the Earth's climate or the cosmos--or code-breaking by intelligence services. They have only rarely been available to artificial intelligence researchers. However, the recent enthusiasm for artificial intelligence research has reportedly meant that artificial intelligence researchers has been cited as a factor in the development of the next round of supercomputers (not least because of the utility of AI in facilitating their work).
Especially with this being the case it seems far from impossible that this will enable it to yield new insights into the subject--just as this past decade it was already the case that our having faster computers available permitted the striking advances in areas like machine learning that we saw this past decade. Indeed, even as the recent excitement over artificial intelligence turns into disappointment with the realization that the most-hyped applications (like Level 5 self-driving) are more remote than certain loud-mouthed hucksters promised, the continued expansion of computing power offers considerable grounds to not write those prospects off just yet.
Tuesday, January 18, 2022
What The Magnificent Ambersons Can Teach Us About Technological Change
I remember reading Booth Tarkington's The Magnificent Ambersons years ago and finding it a rather slight, tepid tale--so much so that I found it hard to understand why Orson Welles, after giving us his ferocious epic Citizen Kane, picked it for a follow-up (and suspected that it was because slight and tepid was what he wanted after the sheer hell he went through making his first movie).
Still, some bits of the novel have stuck in my mind, not least those which had to do with the emergence of the automobile. As men like Eugene Morgan toiled on the vehicles the broader public tended to look at them ironically--an attitude epitomized by the way idiot vulgarians would yell "Git a hoss!" at anybody with a car, and delighting especially in the sight of some motorist stuck repairing a malfunctioning vehicle. However, the technology progressed, and the world changed greatly, leaving the "Git a hoss!"-yelling oafs looking foolish.
Tarkington depicted the shift with some nuance, with this striking me as especially the case in the scenes regarding Aunt Fanny's investment in the headlight manufacturer. She was impressed by a demonstration of the technology, but Morgan, who at this point was growing wealthy from having got into the automotive revolution "on the ground floor" explained that while the headlight in question worked "well enough in the shop" on the road it could only stay lit if a car was going at high speed (twenty-five miles an hour minimum, fifty miles an hour for full illumination)--which meant that the light failed if the motorist drove at all more slowly, and greatly limited the practical usefulness and salability of the technology. Morgan acknowledged that work to improve it continued, but that for the time being she had best eschew putting her money into the company. However, Fanny went ahead and put her money into the company anyway, and ended up broke.
It is as striking a dramatic illustration as I can remember in a major novel of how amid a time of technological flux people go from dismissing a technology altogether to being utterly credulous as a great deal suddenly seems possible--and how at such a moment the word "startup" can seem synonymous with "gold mine." It is a striking illustration, too, of how what works "in the shop" may not be ready for the street just yet, or any time soon, or even ever--with the innovation in question perhaps likely to come out of a different shop, a different startup, than the one that initially caught the eye, because even if it came along later it ended up being the one that made the thing work, or at least cut the deals that got it to market when it became workable.
The more sophisticated technology-watchers, of course, understand this, and indeed NASA developed an excellent system for judging these matters with which I think everyone who cares about these issues should acquaint themselves. Of course, to go by what passes for "science" journalism," which consistently, overwhelmingly, shows and promotes to its audience Fanny's unsophistication rather than Morgan's astuteness, few bother to do so--or in any other way come to understand what Tarkington was able not just to explain but to dramatize in his novel a century ago.
Still, some bits of the novel have stuck in my mind, not least those which had to do with the emergence of the automobile. As men like Eugene Morgan toiled on the vehicles the broader public tended to look at them ironically--an attitude epitomized by the way idiot vulgarians would yell "Git a hoss!" at anybody with a car, and delighting especially in the sight of some motorist stuck repairing a malfunctioning vehicle. However, the technology progressed, and the world changed greatly, leaving the "Git a hoss!"-yelling oafs looking foolish.
Tarkington depicted the shift with some nuance, with this striking me as especially the case in the scenes regarding Aunt Fanny's investment in the headlight manufacturer. She was impressed by a demonstration of the technology, but Morgan, who at this point was growing wealthy from having got into the automotive revolution "on the ground floor" explained that while the headlight in question worked "well enough in the shop" on the road it could only stay lit if a car was going at high speed (twenty-five miles an hour minimum, fifty miles an hour for full illumination)--which meant that the light failed if the motorist drove at all more slowly, and greatly limited the practical usefulness and salability of the technology. Morgan acknowledged that work to improve it continued, but that for the time being she had best eschew putting her money into the company. However, Fanny went ahead and put her money into the company anyway, and ended up broke.
It is as striking a dramatic illustration as I can remember in a major novel of how amid a time of technological flux people go from dismissing a technology altogether to being utterly credulous as a great deal suddenly seems possible--and how at such a moment the word "startup" can seem synonymous with "gold mine." It is a striking illustration, too, of how what works "in the shop" may not be ready for the street just yet, or any time soon, or even ever--with the innovation in question perhaps likely to come out of a different shop, a different startup, than the one that initially caught the eye, because even if it came along later it ended up being the one that made the thing work, or at least cut the deals that got it to market when it became workable.
The more sophisticated technology-watchers, of course, understand this, and indeed NASA developed an excellent system for judging these matters with which I think everyone who cares about these issues should acquaint themselves. Of course, to go by what passes for "science" journalism," which consistently, overwhelmingly, shows and promotes to its audience Fanny's unsophistication rather than Morgan's astuteness, few bother to do so--or in any other way come to understand what Tarkington was able not just to explain but to dramatize in his novel a century ago.
Thursday, January 6, 2022
Putting the Reports About the Thwaites Glacier Into Perspective
Last month reports that the Thwaites glacier in Antarctica is likely to collapse within a number of years (as few as three in some of the reports) made the rounds of the news, along with projections that the event could by itself raise sea levels by two feet, and lead to further collapses producing a ten foot sea level rise, inundating all the world's coastlines.
The tone in which all this was reported, of course, made "could" seem like "will," and does not disabuse anyone not reading closely of the impression that the maximum sea level rise anticipated here will happen instantaneously.
Even worse than most in this regard is the title of the piece in Rolling Stone--"'The Fuse Has Been Blown,' and the Doomsday Glacier Is Coming for Us All."
This flatly tells us not that something potentially very bad may be happening not very far from now, or even that something actually very bad is inevitably happening, but that the worst has already happened.
That is, of course, not actually the case--as the article itself makes clear if one reads it rather than just Retweeting the headline. In spite of Rolling Stone's propensity for "collapse porn" (it was through their pages that I became acquainted with the writing of James Howard Kunstler, whom Leigh Phillips has described as "hav[ing] a veritable hard-on for the end of the world, imagining with relish . . . collapse . . . retreat from modernity and an embrace of the Medieval"), Jeff Goodell's article is considerably better-informed, more factually grounded and intellectually nuanced than the great majority of the other items on the subject I have seen thus far. Goodell (who, by the way, writes that collapse may come inside a decade--rather than the five and even three year periods so many others are talking about) acknowledges that the report's findings are not the surprise much of the media seems to think it is (Goodell himself wrote a noted piece on Thwaites back in 2017), and that there are enormous uncertainties not only regarding when and even if Thwaites will crack up, but what would happen afterward. He notes that this does not completely exclude scenarios where it may not make the (already pretty terrible) picture too much worse. And even the really bad ones (ten feet of sea rise) would likely be a matter of a century.
The result is that the title is a real shame--but all too revealing of how the media has tended to report on this subject (as it does on a great many others), prioritizing shock and fear over comprehension, and getting away with it because of what passes for "reading" these days. Still, even this piece lacked something that I think we should be seeing more of, hearing more of, when discussing matters like the Thwaites glacier and climate change generally--solutions. Obviously reducing the accumulation of greenhouse gases in the atmosphere (decarbonizing our energy-transport system, planting more trees, etc.) is central and indispensable to the ultimate, bottom-line, long-term solution to the larger problem of global warming, and no one serious about the matter suggests not doing so. Yet there has also long been discussing the use of other ameliorative strategies that can help us cope with particular effects of such warming as we seem unlikely to be able to avoid, like saving glaciers through engineering efforts (and indeed, the idea of rescuing Thwaites specifically in this manner specifically is not unprecedented). There is no doubt that the schemes are ambitious, relying on unproven technologies--but it would certainly seem that given the stakes we should be hearing A LOT of calls for programs to develop and deploy anything that will help. However, the championing of such ideas is exactly what we do not see when we look at coverage of the issue.
Of course, it is the media's job to be skeptical--but it goes about that job in awfully "selective" fashion. Consider, for example, the New York Times Magazine's 2019 piece on hypersonic missiles. Hypersonic missiles, certainly, are a radical, far from proven technology--but that does not stop the NYT Magazine piece from being hawkish in the extreme about the claims of such missiles' development being a national priority for the United States. I cannot think of a single occasion when I saw an article in the Times and its associated publications, or any other media outlet of comparable standing, try so hard to sell its readers on the importance of a specific technological program that could help with our environmental problems, or show so much respect for the proponents of such a program, or so uncritically embrace their optimism about the feasibility and value of that technology, as they do in the case of those missiles. Instead they strain for any excuse to dismiss such a project, and close on a note of "Don't get your hopes up." And that says everything about the media's prejudices, not least that preference for fear-mongering and defeatism on the subject--to the very great cost of the dialogue on these matters, and our chances of actually dealing with the problem.
The tone in which all this was reported, of course, made "could" seem like "will," and does not disabuse anyone not reading closely of the impression that the maximum sea level rise anticipated here will happen instantaneously.
Even worse than most in this regard is the title of the piece in Rolling Stone--"'The Fuse Has Been Blown,' and the Doomsday Glacier Is Coming for Us All."
This flatly tells us not that something potentially very bad may be happening not very far from now, or even that something actually very bad is inevitably happening, but that the worst has already happened.
That is, of course, not actually the case--as the article itself makes clear if one reads it rather than just Retweeting the headline. In spite of Rolling Stone's propensity for "collapse porn" (it was through their pages that I became acquainted with the writing of James Howard Kunstler, whom Leigh Phillips has described as "hav[ing] a veritable hard-on for the end of the world, imagining with relish . . . collapse . . . retreat from modernity and an embrace of the Medieval"), Jeff Goodell's article is considerably better-informed, more factually grounded and intellectually nuanced than the great majority of the other items on the subject I have seen thus far. Goodell (who, by the way, writes that collapse may come inside a decade--rather than the five and even three year periods so many others are talking about) acknowledges that the report's findings are not the surprise much of the media seems to think it is (Goodell himself wrote a noted piece on Thwaites back in 2017), and that there are enormous uncertainties not only regarding when and even if Thwaites will crack up, but what would happen afterward. He notes that this does not completely exclude scenarios where it may not make the (already pretty terrible) picture too much worse. And even the really bad ones (ten feet of sea rise) would likely be a matter of a century.
The result is that the title is a real shame--but all too revealing of how the media has tended to report on this subject (as it does on a great many others), prioritizing shock and fear over comprehension, and getting away with it because of what passes for "reading" these days. Still, even this piece lacked something that I think we should be seeing more of, hearing more of, when discussing matters like the Thwaites glacier and climate change generally--solutions. Obviously reducing the accumulation of greenhouse gases in the atmosphere (decarbonizing our energy-transport system, planting more trees, etc.) is central and indispensable to the ultimate, bottom-line, long-term solution to the larger problem of global warming, and no one serious about the matter suggests not doing so. Yet there has also long been discussing the use of other ameliorative strategies that can help us cope with particular effects of such warming as we seem unlikely to be able to avoid, like saving glaciers through engineering efforts (and indeed, the idea of rescuing Thwaites specifically in this manner specifically is not unprecedented). There is no doubt that the schemes are ambitious, relying on unproven technologies--but it would certainly seem that given the stakes we should be hearing A LOT of calls for programs to develop and deploy anything that will help. However, the championing of such ideas is exactly what we do not see when we look at coverage of the issue.
Of course, it is the media's job to be skeptical--but it goes about that job in awfully "selective" fashion. Consider, for example, the New York Times Magazine's 2019 piece on hypersonic missiles. Hypersonic missiles, certainly, are a radical, far from proven technology--but that does not stop the NYT Magazine piece from being hawkish in the extreme about the claims of such missiles' development being a national priority for the United States. I cannot think of a single occasion when I saw an article in the Times and its associated publications, or any other media outlet of comparable standing, try so hard to sell its readers on the importance of a specific technological program that could help with our environmental problems, or show so much respect for the proponents of such a program, or so uncritically embrace their optimism about the feasibility and value of that technology, as they do in the case of those missiles. Instead they strain for any excuse to dismiss such a project, and close on a note of "Don't get your hopes up." And that says everything about the media's prejudices, not least that preference for fear-mongering and defeatism on the subject--to the very great cost of the dialogue on these matters, and our chances of actually dealing with the problem.
Monday, January 3, 2022
I Don't Want to Hear Another Word About Climate Change (Unless You're Going to Tell Us What We're Going To Do About It)
I am not and have never been a climate change denier. I do not dispute that there is an ongoing, anthropogenic and rapidly progressing process of climate change, driven mainly by emissions of GreenHouse Gases (GHGs) like carbon dioxide and methane, taking a large rising toll on the natural environment, and in human life and property (already in 2009 one report estimated 300,000 lives a year--and, given the likelihood of the figure rising rather than falling, implies deaths in the many millions to date). Indeed, I would go further than most in saying that the damage we are seeing indicates that the process has already advanced to an unacceptable degree--that the world is already too hot--and that even were we go to carbon-neutral today the GHGs already in the atmosphere will mean decades more heating, which would very likely have such second-order effects as thawing permafrost likely to intensify the heating that much further, so that even if we did far, far more than we have done to date the survival of human civilization, and perhaps even the human species itself as vast portions of the planet become literally uninhabitable, may be in doubt . . .
Of course, if one recognizes that all this is real then no reasonable person can expect the media to stop talking about it, can they?
No, of course not. But it is also undeniable that what the mainstream media has done is inflict an incessant hard rain of bad news on an already terrified public, while being relentlessly negative about any and every possible way of seriously redressing the problem. It has treated renewable energy (like solar) with disdain, and even as those technologies win victory after victory in the marketplace it still never misses a chance to badmouth them (to the delight of the pro-nuclear trolls, whose activity seems to be way, way up these days--just in case you thought that you'd heard the last of them). It sneers at any talk of a Green New Deal. It falls all over itself trying to criticize even the idea of cellular agriculture, and "Transportation-as-a-Service." It scarcely acknowledges the existence of ideas to save the world's glaciers. And of course, its hostility to geoengineering of any type has been relentless.
In short, after a long period of drawing a false equivalence between acknowledgment of climate change and climate denial it can seem to have shifted (intentionally or unintentionally) to a narrative of climate defeatism, utterly determined to beat down any hope of useful action whatsoever--which leaves us in the same place in the end as that denial to which it was so much an accessory (while, one might add, saddling the powerless with enormous guilt, because somehow not the politicians, not the CEOs, but they, are responsible for it all, especially if they ever ate a burger in their life).
In response I offer a modest proposal. Ordinarily I do not think that it is reasonable to demand that someone pointing out a problem also have a detailed solution in hand. In fact I tend to think of this as a way of suppressing the dialogue over an issue, and thus also the efforts to deal with the problem. However, as we already know how bad things are--are already literally becoming sick over the knowledge of how bad things are as the depressingly, cripplingly, overfamiliar news is pounded into our heads over and over and over again; and as it seems to me that there are a multitude of ideas that could help (and I don't mean the "hairshirt," agonize-over-your-personal-carbon footprint stuff, or even just decarbonizing our electric grid, but also ameliorative stuff like glacier preservation and kelp farming and direct air capture) that ought to be getting far more discussion, and at least some of which seems to me to be worthy of the genuine, massive backing that alone can speed its development and implementation--it is time for the coverage, any coverage we are to take seriously as anything but a promotion of climate defeatism, to start emphasizing what can be done, at length and in detail and as soon as possible, and show the greatest possible rigor in thinking through and explaining those solutions.
This is, of course, not what the media does. It trafficks in fear. And intelligent explanation of problems, never mind solutions, has never been its forte. Yet it is the climate coverage we need to see because we have already long, long passed the point where merely getting people frightened becomes counterproductive.
Of course, if one recognizes that all this is real then no reasonable person can expect the media to stop talking about it, can they?
No, of course not. But it is also undeniable that what the mainstream media has done is inflict an incessant hard rain of bad news on an already terrified public, while being relentlessly negative about any and every possible way of seriously redressing the problem. It has treated renewable energy (like solar) with disdain, and even as those technologies win victory after victory in the marketplace it still never misses a chance to badmouth them (to the delight of the pro-nuclear trolls, whose activity seems to be way, way up these days--just in case you thought that you'd heard the last of them). It sneers at any talk of a Green New Deal. It falls all over itself trying to criticize even the idea of cellular agriculture, and "Transportation-as-a-Service." It scarcely acknowledges the existence of ideas to save the world's glaciers. And of course, its hostility to geoengineering of any type has been relentless.
In short, after a long period of drawing a false equivalence between acknowledgment of climate change and climate denial it can seem to have shifted (intentionally or unintentionally) to a narrative of climate defeatism, utterly determined to beat down any hope of useful action whatsoever--which leaves us in the same place in the end as that denial to which it was so much an accessory (while, one might add, saddling the powerless with enormous guilt, because somehow not the politicians, not the CEOs, but they, are responsible for it all, especially if they ever ate a burger in their life).
In response I offer a modest proposal. Ordinarily I do not think that it is reasonable to demand that someone pointing out a problem also have a detailed solution in hand. In fact I tend to think of this as a way of suppressing the dialogue over an issue, and thus also the efforts to deal with the problem. However, as we already know how bad things are--are already literally becoming sick over the knowledge of how bad things are as the depressingly, cripplingly, overfamiliar news is pounded into our heads over and over and over again; and as it seems to me that there are a multitude of ideas that could help (and I don't mean the "hairshirt," agonize-over-your-personal-carbon footprint stuff, or even just decarbonizing our electric grid, but also ameliorative stuff like glacier preservation and kelp farming and direct air capture) that ought to be getting far more discussion, and at least some of which seems to me to be worthy of the genuine, massive backing that alone can speed its development and implementation--it is time for the coverage, any coverage we are to take seriously as anything but a promotion of climate defeatism, to start emphasizing what can be done, at length and in detail and as soon as possible, and show the greatest possible rigor in thinking through and explaining those solutions.
This is, of course, not what the media does. It trafficks in fear. And intelligent explanation of problems, never mind solutions, has never been its forte. Yet it is the climate coverage we need to see because we have already long, long passed the point where merely getting people frightened becomes counterproductive.
Subscribe to:
Posts (Atom)