Geology

You are currently browsing the archive for the Geology category.

In the library of Earth’s history, there are missing books, and within books there are missing chapters, written in rock that is now gone. John Wesley Powell recorded the greatest example of gone rock in 1869, on his expedition by boat through the Grand Canyon. Floating down the Colorado River, he saw the canyon’s mile-thick layers of reddish sedimentary rock resting on a basement of gray non-sedimentary rock, the layers of which were cocked at an angle from the flatnesses of every layer above. Observing this, he correctly assumed that the upper layers did not continue from the bottom one, because time had clearly passed between when the basement rock was beveled flat, against its own grain, and when the floors of rock above it were successively laid down. He didn’t know how much time had passed between basement and flooring, and could hardly guess.

The answer turned out to be more than a billion years. The walls of the Grand Canyon say nothing about what happened during that time. Geology calls that nothing an unconformity.

In the decades since Powell made his notes, the same gap has been found all over the world and is now called the Great Unconformity. Because of that unconformity, geology knows close to nothing about what happened in the world through stretches of time up to 1.6 billion years long.

All of those absent records end abruptly with the Cambrian Explosion, which began about 541 million years ago. That’s when the Cambrian period arrived and with it an amplitude of history, written in stone.

Many theories attempt to explain what erased such a large span of Earth’s history, but the prevailing guess is perhaps best expressed in “Neoproterozoic glacial origin of the Great Unconformity”, published on the last day of 2018 by nine geologists writing for the National Academy of Sciences. Put simply, they blame snow. Lots of it: enough to turn the planet into one giant snowball, informally called Snowball Earth. A more accurate name for this time would be Glacierball Earth, because glaciers, all formed from accumulated snow, apparently covered most or all of Earth’s land during the Great Unconformity—and most or all of the seas as well. Every continent was a Greenland or an Antarctica.

The relevant fact about glaciers is that they don’t sit still. They push immensities of accumulated ice down on landscapes and then spread sideways, pulverizing and scraping against adjacent landscapes, bulldozing their ways seaward through mountains and across hills and plains. In this manner, glaciers scraped a vastness of geological history off the Earth’s continents and sideways into ocean basins, where plate tectonics could hide the evidence. (A fact little known outside of geology is that nearly all the world’s ocean floors are young: born in spreading centers and killed by subduction under continents or piled up as debris on continental edges here and there. Example: the Bay Area of California is an ocean floor that wasn’t subducted into a trench.) As a result, the stories of Earth’s missing history are partly told by younger rock that remembers only that a layer of moving ice had erased pretty much everything other than a signature on its work.

I bring all this up because I see something analogous to Glacierball Earth happening right now, right here, across our new worldwide digital sphere. A snowstorm of bits is falling on the virtual surface of our virtual sphere, which itself is made of bits even more provisional and temporary than the glaciers that once covered the physical Earth. Nearly all of this digital storm, vivid and present at every moment, is doomed to vanish because it lacks even a glacier’s talent for accumulation.

The World Wide Web is also the World Wide Whiteboard.

Think about it: there is nothing about a bit that lends itself to persistence, other than the media it is written on. Form follows function; and most digital functions, even those we call “storage”, are temporary. The largest commercial facilities for storing digital goods are what we fittingly call “clouds”. By design, these are built to remember no more of what they once contained than does an empty closet. Stop paying for cloud storage, and away goes your stuff, leaving no fossil imprints. Old hard drives, CDs, and DVDs might persist in landfills, but people in the far future may look at a CD or a DVD the way a geologist today looks at Cambrian zircons: as hints of digital activities that may have happened during an interval about which nothing can ever be known. If those fossils speak of what’s happening now at all, it will be of a self-erasing Digital Earth that was born in the late 20th century.

This theory actually comes from my wife, who has long claimed that future historians will look at our digital age as an invisible one because it sucks so royally at archiving itself.

Credit where due: the Internet Archive is doing its best to make sure that some stuff will survive. But what will keep that archive alive, when all the media we have for recalling bits—from spinning platters to solid-state memory—are volatile by nature?

My own future unconformity is announced by the stack of books on my desk, propping up the laptop on which I am writing. Two of those books are self-published compilations of essays I wrote about technology in the mid-1980s, mostly for publications that are long gone. The originals are on floppy disks that can only be read by PCs and apps of that time, some of which are buried in lower strata of boxes in my garage. I just found a floppy with some of those essays. (It’s the one with a blue edge in the wood case near the right end of the photo above.) If those still retain readable files, I am sure there are ways to recover at least the raw ASCII text. But I’m still betting the paper copies of the books under this laptop will live a lot longer than will these floppies or my mothballed PCs, all of which are likely bricked by decades of un-use.

As for other media, the prospect isn’t any better.

At the base of my video collection is a stratum of VHS videotapes, atop of which are strata of MiniDV and Hi8 tapes, and then one of digital stuff burned onto CDs and stored in hard drives, most of which have been disconnected for years. Some of those drives have interfaces and connections (e.g. FireWire) no longer supported by any computers being made today. Although I’ve saved machines to play all of them, none I’ve checked still work. One choked to death on a CD I stuck in it. That was a failure that stopped me from making Christmas presents of family memories recorded on old tapes and DVDs. I meant to renew the project sometime before the following Christmas, but that didn’t happen. Next Christmas? The one after that? I still hope, but the odds are against it.

Then there are my parents’ 8mm and 16mm movies filmed between the 1930s and the 1960s. In 1989, my sister and I had all of those copied over to VHS tape. We then recorded our mother annotating the tapes onto companion cassette tapes while we all watched the show. I still have the original film in a box somewhere, but I haven’t found any of the tapes. Mom died in 2003 at age 90, and her whole generation is now gone.

The base stratum of my audio past is a few dozen open reel tapes recorded in the 1950s and 1960s. Above those are cassette and micro-cassette tapes, plus many Sony MiniDisks recorded in ATRAC, a proprietary compression algorithm now used by nobody, including Sony. Although I do have ways to play some (but not all) of those, I’m cautious about converting any of them to digital formats (Ogg, MPEG, or whatever), because all digital storage media are likely to become obsolete, dead, or both—as will formats, algorithms, and codecs. Already I have dozens of dead external hard drives in boxes and drawers. And, since no commercial cloud service is committed to digital preservation in the absence of payment, my files saved in clouds are sure to be flushed after neither my heirs nor I continue paying for their preservation. I assume my old open reel and cassette tapes are okay, but I can’t tell right now because both my Sony TCWE-475 cassette deck (high end in its day) and my Akai 202D-SS open-reel deck (a quadrophonic model from the early ’70s) are in need of work, since some of their rubber parts have rotted.

The same goes for my photographs. My printed photos—countless thousands of them dating from the late 1800s to 2004—are stored in boxes and albums of photos, negatives and Kodak slide carousels. My digital photos are spread across a mess of duplicated backup drives totaling many terabytes, plus a handful of CDs. About 60,000 photos are exposed to the world on Flickr’s cloud, where I maintain two Pro accounts (here and here) for $50/year apiece. More are in the Berkman Klein Center’s pro account (here) and Linux Journal‘s (here). I doubt any of those will survive after those entities stop getting paid their yearly fees. SmugMug, which now owns Flickr, has said some encouraging things about photos such as mine, all of which are Creative Commons-licensed to encourage re-use. But, as Geoffrey West tells us, companies are mortal. All of them die.

As for my digital works as a whole (or anybody’s), there is great promise in what the Internet Archive and Wikimedia Commons do, but there is no guarantee that either will last for decades more, much less for centuries or millennia. And neither are able to archive everything that matters (much as they might like to).

It should also be sobering to recognize that nobody truly “owns” a domain on the internet. All those “sites” with “domains” at “locations” and “addresses” are rented. We pay a sum to a registrar for the right to use a domain name for a finite period of time. There are no permanent domain names or IP addresses. In the digital world, finitude rules.

So the historic progression I see, and try to illustrate in the photo at the top of this post, is from hard physical records through digital ones we hold for ourselves, and then up into clouds… that go away. Everything digital is snow falling and disappearing on the waters of time.

Will there ever be a way to save for the very long term what we ironically call our digital “assets?” Or is all of it doomed by its own nature to disappear, leaving little more evidence of its passage than a Great Digital Unconformity, when everything was forgotten?

I can’t think of any technical questions more serious than those two.


The original version of this post appeared in the March 2019 issue of Linux Journal.

Source: NASA, ESA, and the Hubble Heritage Team (STScI/AURA)

Pillars of Creation” is a live view of stars forming in a neighboring region of the Milky Way. (Inside the Eagle Nebula, 5,400 to 6,100 light years away.)

The Solar System, formed 4.6 billion years ago. Earth became a planet .46 billion years later. That was 9.247 billion years after the Big Bang, which happened 13.787 billion years ago, meaning that our solar system is more than a third of the age of the Universe. (By the giant impact hypothesis, Earth was enlarged 4.5 billion years ago when it mooshed together with another planet about the size of Mars. That ex-planet is now called Theia, and the leftover material it scattered into space is now the Moon. Another gift to Earth, says the theory, is our water.)

Hydrogen, helium, and other light elements formed with the Big Bang, but the heavy elements were cooked up at various times in the 8 billion year span before our solar system was formed. Some, perhaps, are still cooking.

Life appeared on earth at least 3.5 billion years ago.* Oxygenation sufficient to support life as we know it happened at the start of proterozoic era, about 2.5 billion years ago. The phanerozoic eon, characterized by an abundance of plants and animals, began 0.541 billion years ago and will continue until the Sun gets so large and hot that photosynthesis as we know it becomes impossible. A rough consensus within science is that this will likely happen in just 600 million years, meaning we’re about 80% of our way through the time window for life on Earth.

More perspective: the primary rock formation on which most of Manhattan’s ranking skyscrapers repose—Manhattan Schist—is itself about a half billion years old. (My ass is three floors above some right now.)

In another 4.5 billion years, our galaxy, the Milky Way, will become one with Andromeda, which is currently 2.5 million light-years distant but headed our way on a collision course, looking for now like a thrown frisbee, four moons wide. (It’s actually much larger.) The two will begin merging (not colliding, because nearly all stars are too far apart for that) around 4 billion years from now, and in about 7 billion years will complete a new galaxy resembling an elliptical haze. Here is a video simulation of that future. And here are still panels for the same:

Our Sun will likely be around for all of that future, though by the end it will have become a red giant with a diameter wider than Earth’s orbit, or perhaps will have gone nova, surviving as a white dwarf. (Also—I’m adding this later—Andromeda is weird and scary.)

When that’s done, the Universe will be about 16 billion years old. Or young. In TIME WITHOUT END: PHYSICS AND BIOLOGY IN AN OPEN UNIVERSE, Freeman Dyson gives these estimates for the future age of the Universe:

TABLE I. Summary of time scales.

Closed Universe
Total duration 10^11 yr

Open Universe
Low-mass stars cool off 10^14 yr
Planets detached from stars 10^15 yr
Stars detached from galaxies 10^19 yr
Decay of orbits by gravitational radiation 10^20 yr
Decay of black holes by Hawking process 10^64 yr
Matter liquid at zero temperature 10^65 yr
All matter decays to iron 10^1500 yr
Collapse of ordinary matter to black hole
[alternative (ii)] 10^(10^26) yr
Collapse of stars to neutron stars
or black holes [alternative (iv)] 10^(10^76) yr

So the best guess here is that Universe is about 1% into its lifespan, which has a great many zeros in its number of birthdays. In biological terms, that means it’s not even a baby, or a fetus. It’s more like a zygote, or a blastula.

So maybe… just maybe… the forms of life we know on Earth are just early prototypes of what’s to come in the fullness of time, space, and evolving existence.

Bonus link: Katie Mack, who has forgotten more about all this than I’ll ever know.

[Later…] *Make that 4.1 billion years ago.

[Later again…] Then there is this.

This is the Ostrom Memorial Lecture I gave on 9 October of last year for the Ostrom Workshop at Indiana University. Here is the video. (The intro starts at 8 minutes in, and my part starts just after 11 minutes in.) I usually speak off the cuff, but this time I wrote it out, originally in outline form*, which is germane to my current collaborations with Dave Winer, father of outlining software (and, in related ways, of blogging and podcasting). So here ya go.

Intro

The movie Blade Runner was released in 1982; and was set in a future Los Angeles. Anyone here know when in the future Blade Runner is set? I mean, exactly?

The year was 2019. More precisely, next month: November.

In Blade Runner’s 2019, Los Angeles is a dark and rainy hellscape with buildings the size of mountains, flying cars, and human replicants working on off-world colonies. It also has pay phones and low-def computer screens that are vacuum tubes.

Missing is a communication system that can put everyone in the world at zero distance from everyone else, in disembodied form, at almost no cost—a system that lives on little slabs in people’s pockets and purses, and on laptop computers far more powerful than any computer, of any size, from 1982.

In other words, this communication system—the Internet—was less thinkable in 1982 than flying cars, replicants and off-world colonies. Rewind the world to 1982, and the future Internet would appear a miracle dwarfing the likes of loaves and fish.

In economic terms, the Internet is a common pool resource; but non-rivalrous and non-excludable to such an extreme that to call it a pool or a resource is to insult what makes it common: that it is the simplest possible way for anyone and anything in the world to be present with anyone and anything else in the world, at costs that can round to zero.

As a commons, the Internet encircles every person, every institution, every business, every university, every government, every thing you can name. It is no less exhaustible than presence itself. By nature and design, it can’t be tragic, any more than the Universe can be tragic.

There is also only one of it. As with the universe, it has no other examples.

As a source of abundance, the closest thing to an example the Internet might have is the periodic table. And the Internet might be even more elemental than that: so elemental that it is easy to overlook the simple fact that it is the largest goose ever to lay golden eggs.

It can, however, be misunderstood, and that’s why it’s in trouble.

The trouble it’s in is with human nature: the one that sees more value in the goose’s eggs than in the goose itself.

See, the Internet is designed to support every possible use, every possible institution, and—alas—every possible restriction, which is why enclosure is possible. People, institutions and possibilities of all kinds can be trapped inside enclosures on the Internet. I’ll describe nine of them.

Enclosures

The first enclosure is service provisioning, for example with asymmetric connection speeds. On cable connections you may have up to 400 megabits per second downstream, but still only 10 megabits per second—one fortieth of that—upstream. (By the way this is exactly what Spectrum, formerly Time Warner Cable, provides with its most expensive home service to customers in New York City.)

They do that to maximize consumption while minimizing production by those customers. You can consume all the video you want, and think you’re getting great service. But meanwhile this asymmetrical provisioning prevents production at your end. Want to put out a broadcast or a podcast from your house, to run your own email server, or to store your own video or other personal data in your own personal “cloud”? Forget it.

The Internet was designed to support infinite production by anybody of anything. But cable TV companies don’t want you to have that that power. So you don’t. The home Internet you get from your cable company is nice to have, but it’s not the whole Internet. It’s an enclosed subset of capabilities biased by and for the cable company and large upstream producers of “content.”

So, it’s golden eggs for them, but none for you. Also missing are all the golden eggs you might make possible for those companies as an active producer rather than as a passive consumer.

The second enclosure is through 5G wireless service, currently promoted by phone companies as a new generation of Internet service. The companies deploying 5G promise greater speeds and lower lag times over wireless connections; but is also clear that they want to build in as many choke points as they like, all so you can be billed for as many uses as possible.

You want gaming? Here’s our gaming package. You want cloud storage? Here’s our cloud storage package. Each of these uses will carry terms and conditions that allow some uses and prevent others. Again, this is a phone company enclosure. No cable companies are deploying 5G. They’re fine with their own enclosure.

The third enclosure is government censorship. The most familiar example is China’s. In China’s closed Internet you will find no Google, Facebook, Twitter, Instagram or Reddit. No Pandora, Spotify, Slack or Dropbox. What you will find is pervasive surveillance of everyone and everything—and ranking of people in its Social Credit System.

By March of this year, China had already punished 23 million people with low social credit scores by banning them from traveling. Control of speech has also spread to U.S. companies such as the NBA and ESPN, which are now censoring themselves as well, bowing to the wishes of the Chinese government and its own captive business partners.

The fourth enclosure is the advertising-supported commercial Internet. This is led by Google and Facebook, but also includes all the websites and services that depend on tracking-based advertising. This form of advertising, known as adtech, has in the last decade become pretty much the only kind of advertising online.

Today there are very few major websites left that don’t participate in what Shoshana Zuboff calls surveillance capitalism, and Brett Frischmann and Evan Selinger call, in their book by that title, Re-engineering Humanity. Surveillance of individuals online is now so deep and widespread that nearly every news organization is either unaware of it or afraid to talk about it—in part because the advertising they run is aimed by it.

That’s why you’ll read endless stories about how bad Facebook and Google are, and how awful it is that we’re all being tracked everywhere like marked animals; but almost nothing about how the sites publishing stories about tracking also participate in exactly the same business—and far more surreptitiously. Reporting on their own involvement in the surveillance business is a third rail they won’t grab.

I know of only one magazine that took and shook that third rail, especially in the last year and a half.  That magazine was Linux Journal, where I worked for 24 years and was serving as editor-in-chief when it was killed by its owner in August. At least indirectly that was because we didn’t participate in the surveillance economy.

The fifth enclosure is protectionism. In Europe, for example, your privacy is protected by laws meant to restrict personal data use by companies online. As a result in Europe, you won’t see the Los Angeles Times or the Washington Post in your browsers, because those publishers don’t want to cope with what’s required by the EU’s laws.

While they are partly to blame—because they wish to remain in the reader-tracking business—the laws are themselves terribly flawed—for example by urging every website to put up a “cookie notice” on pages greeting readers. In most cases clicking “accept” to the site’s cookies only gives the site permission to continue doing exactly the kind of tracking the laws are meant to prevent.

So, while the purpose of these laws is to make the Internet safer, in effect they also make its useful space smaller.

The sixth enclosure is what The Guardian calls “digital colonialism.” The biggest example of that is  Facebook.org, originally called “Free Basics” and “Internet.org”

This is a China-like subset of the Internet, offered for free by Facebook in less developed parts of the world. It consists of a fully enclosed Web, only a few dozen sites wide, each hand-picked by Facebook. The rest of the Internet isn’t there.

The seventh enclosure is the forgotten past. Today the World Wide Web, which began as a kind of growing archive—a public set of published goods we could browse as if it were a library—is being lost. Forgotten. That’s because search engines are increasingly biased to index and find pages from the present and recent past, and by following the tracks of monitored browsers. It’s forgetting what’s old. Archival goods are starting to disappear, like snow on the water.

Why? Ask the algorithm.

Of course, you can’t. That brings us to our eighth enclosure: algorithmic opacity.

Consider for a moment how important power plants are, and how carefully governed they are as well. Every solar, wind, nuclear, hydro and fossil fuel power production system in the world is subject to inspection by whole classes of degreed and trained professionals.

There is nothing of the sort for the giant search engine and social networks of the world. Google and Facebook both operate dozens of data centers, each the size of many Walmart stores. Yet the inner workings of those data centers are nearly absent of government oversight.

This owes partly to the speed of change in what these centers do, but more to the simple fact that what they do is unknowable, by design. You can’t look at rows of computers with blinking lights in many acres of racks and have the first idea of what’s going on in there.

I would love to see research, for example, on that last enclosure I listed: on how well search engines continue to index old websites. Or to do anything. The whole business is as opaque as a bowling ball with no holes.

I’m not even sure you can find anyone at Google who can explain exactly why its index does one thing or another, for any one person or another. In fact, I doubt Facebook is capable of explaining why any given individual sees any given ad. They aren’t designed for that. And the algorithm itself isn’t designed to explain itself, perhaps even to the employees responsible for it.

Or so I suppose.

In the interest of moving forward with research on these topics, I invite anyone at Google, Facebook, Bing or Amazon to help researchers at institutions such as the Ostrom Workshop, and to explain exactly what’s going on inside their systems, and to provide testable and verifiable ways to research those goings-on.

The ninth and worst enclosure is the one inside our heads. Because, if we think the Internet is something we use by grace of Apple, Amazon, Facebook, Google and “providers” such as phone and cable companies, we’re only helping all those companies contain the Internet’s usefulness inside their walled gardens.

Not understanding the Internet can result in problems similar to ones we suffer by not understanding common pool resources such as the atmosphere, the oceans, and the Earth itself.

But there is a difference between common pool resources in the natural world, and the uncommon commons we have with the Internet.

See, while we all know that common-pool resources are in fact not limitless—even when they seem that way—we don’t have the same knowledge of the Internet, because its nature as a limitless non-thing is non-obvious.

For example, we know common pool resources in the natural world risk tragic outcomes if our use of them is ungoverned, either by good sense or governance systems with global reach. But we don’t know that the Internet is limitless by design, or that the only thing potentially tragic about it is how we restrict access to it and use of it, by enclosures such as the nine I just listed.

So my thesis here is this: if we can deeply and fully understand what the Internet is, why it is fully important, and why it is in danger of enclosure, we can also understand why, ten years after Lin Ostrom won a Nobel prize for her work on the commons, that work may be exactly what we need to save the Internet as a boundless commons that can support countless others.

The Internet

We’ll begin with what makes the Internet possible: a protocol.

A protocol is a code of etiquette for diplomatic exchanges between computers. A form of handshake.

What the Internet’s protocol does is give all the world’s digital devices and networks a handshake agreement about how to share data between any point A and any point B in the world, across any intermediary networks.

When you send an email, or look at a website, anywhere in the world, the route the shared data takes can run through any number of networks between the two. You might connect from Bloomington to Denver through Chicago, Tokyo and Mexico City. Then, two minutes later, through Toronto and Miami. Some packets within your data flows may also be dropped along the way, but the whole session will flow just fine because the errors get noticed and the data re-sent and re-assembled on the fly.

Oddly, none of this is especially complicated at the technical level, because what I just described is pretty much all the Internet does. It doesn’t concern itself with what’s inside the data traffic it routes, who is at the ends of the connections, or what their purposes are—any more than gravity cares about what it attracts.

Beyond the sunk costs of its physical infrastructure, and the operational costs of keeping the networks themselves standing up, the Internet has no first costs at its protocol level, and it adds no costs along the way. It also has no billing system.

In all these ways the Internet is, literally, neutral. It also doesn’t need regulators or lawmakers to make it neutral. That’s just its nature.

The Internet’s protocol called is called TCP/IP, and by using it, all the networks of the world subordinate their own selfish purposes.

This is what makes the Internet’s protocol generous and supportive to an absolute degree toward every purpose to which it is put. It is a rising tide that lifts all boats.

In retrospect we might say the big networks within the Internet—those run by phone and cable companies, governments and universities—agreed to participate in the Internet because it was so obviously useful that there was no reason not to.

But the rising-tide nature of the Internet was not obvious to all of them at first. In retrospect, they didn’t realize that the Internet was a Trojan Horse, wheeled through their gates by geeks who looked harmless but in fact were bringing the world a technical miracle.

I can support that claim by noting that even though phone and cable companies of the world now make trillions of dollars because of it, they never would have invented it.

Two reasons for that. One is because it was too damn simple. The other is because they would have started with billing. And not just billing you and me. They would have wanted to bill each other, and not use something invented by another company.

A measure of the Internet’s miraculous nature is that actually billing each other would have been so costly and complicated that what they do with each other, to facilitate the movement of data to, from, and across their networks, is called peering. In other words, they charge each other nothing.

Even today it is hard for the world’s phone and cable companies—and even its governments, which have always been partners of a sort—to realize that the Internet became the world-wide way to communicate because it didn’t start with billing.

Again, all TCP/IP says is that this is a way for computers, networks, and everything connected to them, to get along. And it succeeded, producing instant worldwide peace among otherwise competing providers of networks and services. It made every network operator involved win a vast positive-sum game almost none of them knew they were playing. And most of them still don’t.

You know that old joke in which the big fish says to the little fish, “Hi guys, how’s the water?” and one of the little fish says to the other “What’s water?” In 2005, David Foster Wallace gave a legendary commencement address at Kenyon College that I highly recommend, titled “This is water.”

I suspect that, if Wallace were around today, he’d address his point to our digital world.

Human experience

Those of you who already know me are aware that my wife Joyce is as much a companion and collaborator of mine as Vincent Ostrom was of Lin. I bring this up because much of of this talk is hers, including this pair of insights about the Internet: that it has no distance, and also no gravity.

Think about it: when you are on the Internet with another person—for example if you are in a chat or an online conference—there is no functional distance between you and the other person. One of you may be in Chicago and the other in Bangalore. But if the Internet is working, distance is gone. Gravity is also gone. Your face may be right-side-up on the other person’s screen, but it is absent of gravity. The space you both occupy is the other person’s two-dimensional rectangle. Even if we come up with holographic representations of ourselves, we are still incorporeal “on” the Internet. (I say “on” because we need prepositions to make sense of how things are positioned in the world. Yet our limited set of physical-world prepositions—over, under around, through, beside, within and the rest—misdirect our attention away from our disembodied state in the digital one.)

Familiar as that disembodied state may be to all of us by now, it is still new to human experience and inadequately informed by our experience as embodied creatures. It is also hard for us to see both what our limitations are, and how limitless we are at the same time.

Joyce points out that we are also highly adaptive creatures, meaning that eventually we’ll figure out what it means to live where there is no distance or gravity, much as astronauts learn to live as weightless beings in space.

But in the meantime, we’re having a hard time seeing the nature and limits of what’s good and what’s bad in this new environment. And that has to do, at least in part, on forms of enclosure in that world—and how we are exploited within private spaces where we hardly know we are trapped.

In The Medium is the Massage, Marshall McLuhan says every new medium, every new technology, “works us over completely.” Those are his words: works us over completely. Such as now, with digital technology, and the Internet.

I was talking recently with a friend about where our current digital transition ranks among all the other transitions in history that each have a formal cause. Was becoming ditital the biggest thing since the industrial revolution? Since movable type? Writing? Speech?

No, he said. “It’s the biggest thing since oxygenation.”

In case you weren’t there, or weren’t paying attention in geology class, oxygenation happened about 2.5 billion years ago. Which brings us to our next topic:

Institutions

Journalism is just one example of a trusted institution that is highly troubled in the digital world.

It worked fine in a physical world where truth-tellers who dig into topics and reported on them with minimized prejudice were relatively scarce yet easy to find, and to trust. But in a world flooded with information and opinion—a world where everyone can be a reporter, a publisher, a producer, a broadcaster, where the “news cycle” has the lifespan of a joke, and where news and gossip have become almost indistinguishable while being routed algorithmically to amplify prejudice and homophily, journalism has become an anachronism: still important, but all but drowning in a flood of biased “content” paid for by surveillance-led adtech.

People are still hungry for good information, of course, but our appetites are too easily fed by browsing through the surfeit of “content” on the Internet, which we can easily share by text, email or social media. Even if we do the best we can to share trustworthy facts and other substances that sound like truth, we remain suspended in a techno-social environment we mostly generate and re-generate ourselves. Kind of like our ancestral life forms made sense of the seas they oxygenated, long ago.

The academy is another institution that’s troubled in our digital time. After all, education on the Internet is easy to find. Good educational materials are easy to produce and share. For example, take Kahn Academy, which started with one guy tutoring his cousin though online videos.

Authority must still be earned, but there are now countless non-institutional ways to earn it. Credentials still matter, but less than they used to, and not in the same ways. Ad hoc education works in ways that can be cheap or free, while institutions of higher education remain very expensive. What happens when the market for knowledge and know-how starts moving past requirements for advanced degrees that might take students decades of their lives to pay off?

For one example of that risk already at work, take computer programming.

Which do you think matters more to a potential employer of programmers—a degree in computer science or a short but productive track record? For example, by contributing code to the Linux operating system?

To put this in perspective, Linux and operating systems like it are inside nearly every smart thing that connects to the Internet, including TVs, door locks, the world’s search engines, social network, laptops and mobile phones. Nothing could be more essential to computing life.

At the heart of Linux is what’s called the kernel. For code to get into the kernel, it has to pass muster with other programmers who have already proven their worth, and then through testing and debugging. If you’re looking for a terrific programmer, everyone contributing to the Linux kernel is well-proven. And there are thousands of them.

Now here’s the thing. It not only doesn’t matter whether or not those people have degrees in computer science, or even if they’ve had any formal training. What matters, for our purposes here, is that, to a remarkable degree, many of them don’t have either. Or perhaps most of them.

I know a little about this because, in the course of my work at Linux Journal, I would sometimes ask groups of alpha Linux programmers where they learned to code. Almost none told me “school.” Most were self-taught or learned from each other.

My point here is that the degree to which the world’s most essential and consequential operating system depends on the formal education of its makers is roughly zero.

See, the problem for educational institutions in the digital world is that most were built to leverage scarcity: scarce authority, scarce materials, scarce workspace, scarce time, scarce credentials, scarce reputation, scarce anchors of trust. To a highly functional degree we still need and depend on what only educational institutions can provide, but that degree is a lot lower than it used to be, a lot more varied among disciplines, and it risks continuing to decline as time goes on.

It might help at this point to see gravity in some ways as a problem the Internet solves. Because gravity is top-down. It fosters hierarchy and bell curves, sometimes where we need neither.

Absence of gravity instead fosters heterarchy and polycentrism. And, as we know, at the Ostrom Workshop perhaps better than anywhere, commons are good examples of heterarchy and polycentrism at work.

Knowledge Commons

In the first decade of our new millenium, Elinor Ostrom and Charlotte Hess—already operating in our new digital age—extended the commons category to include knowledge, calling it a complex ecosystem that operates as a common: a shared resource subject to social dilemmas.

They looked at ease of access to digital forms of knowledge and easy new ways to store, access and share knowledge as a common. They also looked at the nature of knowledge and its qualities of non-rivalry and non-excludability, which were both unlike what characterizes a natural commons, with its scarcities of rivalrous and excludable goods.

A knowledge commons, they said, is characterized by abundance. This is one way what Yochai Benkler calls Commons Based Peer Production on the Internet is both easy and rampant, giving us, among many other things, both the free software and open source movements in code development and sharing, plus the Internet and the Web.

Commons Based Peer Production also demonstrates how collaboration and non-material incentives can produce better quality products, and less social friction in the course of production.

I’ve given Linux as one example of Commons Based Peer Production. Others are Wikipedia and the Internet Archive. We’re also seeing it within the academy, for example with Indiana University’s own open archives, making research more accessible and scholarship more rich and productive.

Every one of those examples comports with Lin Ostrom’s design principles:

  1. clearly defined group boundaries;
  2. rules governing use of common goods within local needs and conditions;
  3. participation in modifying rules by those affected by the rules;
  4. accessible and low cost ways to resolve disputes;
  5. developing a system, carried out by community members, for monitoring members’ behavior;
  6. graduated sanctions for rule violators;
  7. and governing responsibility in nested tiers from the lowest level up to the entire interconnected system.

But there is also a crisis with Commons Based Peer Production on the Internet today.

Programmers who ten or fifteen years ago would not participate in enclosing their own environments are doing exactly that, for example with 5G, which is designed to put the phone companies in charge of what we can do on the Internet.

The 5G-enclosed Internet might be faster and more handy in many ways, the range of freedoms for each of us there will be bounded by the commercial interests of the phone companies and their partners, and subject to none of Lin’s rules for governing a commons.

Consider this: every one of the nine enclosures I listed at the beginning of this talk are enabled by programmers who either forgot or never learned about the freedom and openness that made the free and open Internet possible. They are employed in the golden egg gathering business—not in one that appreciates the goose that lays those eggs, and which their predecessors gave to us all.

But this isn’t the end of the world. We’re still at the beginning. And a good model for how to begin is—

The physical world

It is significant that all the commons the Ostroms and their colleagues researched in depth were local. Their work established beyond any doubt the importance of local knowledge and local control.

I believe demonstrating this in the digital world is our best chance of saving our digital world from the nine forms of enclosure I listed at the top of this talk.

It’s our best chance because there is no substitute for reality. We may be digital beings now, as well as physical ones. There are great advantages, even in the digital world, to operating in the here-and-now physical world, where all our prepositions still work, and our metaphors still apply.

Back to Joyce again.

In the mid ‘90s, when the Internet was freshly manifest on our home computers, I was mansplaining to Joyce how this Internet thing was finally the global village long promised by tech.

Her response was, “The sweet spot of the Internet is local.” She said that’s because local is where the physical and the virtual intersect. It’s where you can’t fake reality, because you can see and feel and shake hands with it.

She also said the first thing the Internet would obsolesce would be classified ads in newspapers. That’s because the Internet would be a better place than classifieds for parents to find a crib some neighbor down the street might have for sale. Then Craigslist came along and did exactly that.

We had an instructive experience with how the real world and the Internet work together helpfully at the local level about a year and a half ago. That’s when a giant rainstorm fell on the mountains behind Santa Barbara, where we live, and the town next door, called Montecito. This was also right after the Thomas Fire—largest at the time in recorded California history—had burned all the vegetation away, and there was a maximum risk of what geologists call a “debris flow.”

The result was the biggest debris flow in the history of the region: a flash flood of rock and mud that flowed across Montecito like lava from a volcano. Nearly two hundred homes were destroyed, and twenty-three people were killed. Two of them were never found, because it’s hard to find victims buried under what turned out to be at least twenty thousand truckloads of boulders and mud.

Right afterwards, all of Montecito was evacuated, and very little news got out while emergency and rescue workers did their jobs. Our local news media did an excellent job of covering this event as a story. But I also noticed that not much was being said about the geology involved.

So, since I was familiar with debris flows out of the mountains above Los Angeles, where they have infrastructure that’s ready to handle this kind of thing, I put up a post on my blog titled “Making sense of what happened to Montecito.” In that post I shared facts about the geology involved, and also published the only list on the Web of all the addresses of homes that had been destroyed. Visits to my blog jumped from dozens a day to dozens of thousands. Lots of readers also helped improve what I wrote and re-wrote.

All of this happened over the Internet, but it pertained to a real-world local crisis.

Now here’s the thing. What I did there wasn’t writing a story. I didn’t do it for the money, and my blog is a noncommercial one anyway. I did it to help my neighbors. I did it by not being a bystander.

I also did it in the context of a knowledge commons.

Specifically, I was respectful of boundaries of responsibility; notably those of local authorities—rescue workers, law enforcement, reporters from local media, city and county workers preparing reports, and so on. I gave much credit where it was due and didn’t step on the toes of others helping out as well.

An interesting fact about journalism there at the time was the absence of fake news. Sure, there was plenty of fingers pointing in blog comments and in social media. But it was marginalized away from the fact-reporting that mattered most. There was a very productive ecosystem of information, made possible by the Internet in everyone’s midst. And by everyone, I mean lots of very different people.

Humanity

We are learning creatures by nature. We can’t help it. And we don’t learn by freight forwarding

By that, I mean what I am doing here, and what we do with each other when we talk or teach, is not delivering a commodity called information, as if we were forwarding freight. Something much more transformational is taking place, and this is profoundly relevant to the knowledge commons we share.

Consider the word information. It’s a noun derived from the verb to inform, which in turn is derived from the verb to form. When you tell me something I don’t know, you don’t just deliver a sum of information to me. You form me. As a walking sum of all I know, I am changed by that.

This means we are all authors of each other.

In that sense, the word authority belongs to the right we give others to author us: to form us.

Now look at how much more of that can happen on our planet, thanks to the Internet, with its absence of distance and gravity.

And think about how that changes every commons we participate in, as both physical and digital beings. And how much we need guidance to keep from screwing up the commons we have, or forming the ones we don’t, or forming might have in the future—if we don’t screw things up.

A rule in technology is that what can be done will be done—until we find out what shouldn’t be done. Humans have done this with every new technology and practice from speech to stone tools to nuclear power.

We are there now with the Internet. In fact, many of those enclosures I listed are well-intended efforts to limit dangerous uses of the Internet.

And now we are at a point where some of those too are a danger.

What might be the best way to look at the Internet and its uses most sensibly?

I think the answer is governance predicated on the realization that the Internet is perhaps the ultimate commons, and subject to both research and guidance informed by Lin Ostrom’s rules.

And I hope that guides our study.

There is so much to work on: expansion of agency, sensibility around license and copyright, freedom to benefit individuals and society alike, protections that don’t foreclose opportunity, saving journalism, modernizing the academy, creating and sharing wealth without victims, de-financializing our economies… the list is very long. And I look forward to working with many of us here on answers to these and many other questions.

Thank you. 

Sources

Ostrom, Elinor. Governing the Commons. Cambridge University Press, 1990

Ostrom, Elinor and Hess, Charlotte, editors. Understanding Knowledge as a Commons:
From Theory to Practice, MIT Press, 2011
https://mitpress.mit.edu/books/understanding-knowledge-commons
Full text online: https://wtf.tw/ref/hess_ostrom_2007.pdf

Paul D. Aligica and Vlad Tarko, “Polycentricity: From Polanyi to Ostrom, and Beyond” https://asp.mercatus.org/system/files/Polycentricity.pdf

Elinor Ostrom, “Coping With Tragedies of the Commons,” 1998 https://pdfs.semanticscholar.org/7c6e/92906bcf0e590e6541eaa41ad0cd92e13671.pdf

Lee Anne Fennell, “Ostrom’s Law: Property rights in the commons,” March 3, 2011
https://www.thecommonsjournal.org/articles/10.18352/ijc.252/

Christopher W. Savage, “Managing the Ambient Trust Commons: The Economics of Online Consumer Information Privacy.” Stanford Law School, 2019. https://law.stanford.edu/wp-content/uploads/2019/01/Savage_20190129-1.pdf

 

________________

*I wrote it using—or struggling in—the godawful Outline view in Word. Since I succeeded (most don’t, because they can’t or won’t, with good reason), I’ll brag on succeeding at the subhead level:

As I’m writing this, in Febrary, 2020, Dave Winer is working on what he calls writing on rails. That’s what he gave the pre-Internet world with MORE several decades ago, and I’m helping him with now with the Internet-native kind, as a user. He explains that here. (MORE was, for me, like writing on rails. It’ll be great to go back—or forward—to that again.)

Just before it started, the geology meeting at the Santa Barbara Central Library on Thursday looked like this from the front of the room (where I also tweeted the same pano):

Geologist Ed Keller

Our speakers were geology professor Ed Keller of UCSB and Engineering Geologist Larry Gurrola, who also works and studies with Ed. That’s Ed in the shot below.

As a geology freak, I know how easily terms like “debris flow,” “fanglomerate” and “alluvial fan” can clear a room. But this gig was SRO. That’s because around 3:15 in the morning of January 9th, debris flowed out of canyons and deposited fresh fanglomerate across the alluvial fan that comprises most of Montecito, destroying (by my count on the map below) 178 buildings, damaging more than twice that many, and killing 23 people. Two of those—a 2 year old girl and a 17 year old boy—are still interred in the fresh fanglomerate and sought by cadaver dogs.* The whole thing is beyond sad and awful.

The town was evacuated after the disaster so rescue and recovery work could proceed without interference, and infrastructure could be found and repaired: a job that required removing twenty thousand truckloads of mud and rocks. That work continues while evacuation orders are gradually lifted, allowing the town to repopulate itself to the very limited degree it can.

I talked today with a friend whose business is cleaning houses. Besides grieving the dead, some of whom were friends or customers, she reports that the cleaning work is some of the most difficult she has ever faced, even in homes that were spared the mud and rocks. Refrigerators and freezers, sitting closed and without electricity for weeks, reek of death and rot. Other customers won’t be back because their houses are gone.

Highway 101, one of just two freeways connecting Northern and Southern California, runs through town near the coast and more than two miles from the mountain front. Three debris flows converged on the highway and used it as a catch basin, filling its deep parts to the height of at least one bridge before spilling over its far side and continuing to the edge of the sea. It took two weeks of constant excavation and repair work before traffic could move again. Most exits remain closed. Coast Village Road, Montecito’s Main Street, is open for employees of stores there, but little is open for customers yet, since infrastructural graces such as water are not fully restored. (I saw the Honor Bar operating with its own water tank, and a water truck nearby.) Opening Upper Village will take longer. Some landmark institutions, such as San Ysidro Ranch and La Casa Santa Maria, will take years to restore. (From what I gather, San Ysidro Ranch, arguably the nicest hotel in the world, was nearly destroyed. Its website thank firefighters for salvation from the Thomas Fire. But nothing, I gather, could have save it from the huge debris flow wiped out nearly everything on the flanks of San Ysidro Creek. (All the top red dots along San Ysidro Creek in the map below mark lost buildings at the Ranch.)

Here is a map with final damage assessments. I’ve augmented it with labels for the canyons and creeks (with one exception: a parallel creek west of Toro Canyon Creek):

Click on the map for a closer view, or click here to view the original. On that one you can click on every dot and read details about it.

I should pause to note that Montecito is no ordinary town. Demographically, it’s Beverly Hills draped over a prettier landscape and attractive to people who would rather not live in Beverly Hills. (In fact the number of notable persons Wikipedia lists for Montecito outnumbers those it lists for Beverly Hills by a score of 77 to 71.) Culturally, it’s a village. Last Monday in The New Yorker, one of those notable villagers, T.Coraghessan Boyle, unpacked some other differences:

I moved here twenty-five years ago, attracted by the natural beauty and semirural ambience, the short walk to the beach and the Lower Village, and the enveloping views of the Santa Ynez Mountains, which rise abruptly from the coastal plain to hold the community in a stony embrace. We have no sidewalks here, if you except the business districts of the Upper and Lower Villages—if we want sidewalks, we can take the five-minute drive into Santa Barbara or, more ambitiously, fight traffic all the way down the coast to Los Angeles. But we don’t want sidewalks. We want nature, we want dirt, trees, flowers, the chaparral that did its best to green the slopes and declivities of the mountains until last month, when the biggest wildfire in California history reduced it all to ash.

Fire is a prerequisite for debris flows, our geologists explained. So is unusually heavy rain in a steep mountain watershed. There are five named canyons, each its own watershed, above Montecito, as we see on the map above. There are more to the east, above Summerland and Carpinteria, the next two towns down the coast. Those towns also took some damage, though less than Montecito.

Ed Keller put up this slide to explain conditions that trigger debris flows, and how they work:

Ed and Larry were emphatic about this: debris flows are not landslides, nor do many start that way (though one did in Rattlesnake Canyon 1100 years ago). They are also not mudslides, so we should stop calling them that. (Though we won’t.)

Debris flows require sloped soils left bare and hydrophobic—resistant to water—after a recent wildfire has burned off the chaparral that normally (as geologists say) “hairs over” the landscape. For a good look at what soil surfaces look like, and are likely to respond to rain, look at the smooth slopes on the uphill side of 101 east of La Conchita. Notice how the surface is not only a smooth brown or gray, but has a crust on it. In a way, the soil surface has turned to glass. That’s why water runs off of it so rapidly.

Wildfires are common, and chaparral is adapted to them, becoming fuel for the next fire as it regenerates and matures. But rainfalls as intense as this one are not common. In just five minutes alone, more than half an inch of rain fell in the steep and funnel-like watersheds above Montecito. This happens about once every few hundred years, or about as often as a tsunami.

It’s hard to generalize about the combination of factors required, but Ed has worked hard to do that, and this slide of his is one way of illustrating how debris flows happen eventually in places like Montecito and Santa Barbara:

From bottom to top, here’s what it says:

  1. Fires happen almost regularly, spreading most widely where chaparral has matured to become abundant fuel, as the firefighters like to call it.
  2. Flood events are more random, given the relative rarity of rain and even more rare rains of “biblical” volume. But they do happen.
  3. Stream beds in the floors of canyons accumulate rocks and boulders that roll down the gradually eroding slopes over time. The depth of these is expressed as basin instablity. Debris flows clear out the rocks and boulders when a big flood event comes right after a fire and basin becomes stable (relatively rock-free) again.
  4. The sediment yield in a flood (F) is maximum when a debris flow (DF) occurs.
  5. Debris flows tend to happen once every few hundred years. And you’re not going to get the big ones if you don’t have the canyon stream bed full of rocks and boulders.

About this set of debris flows in particular:

  1. Destruction down Oak Creek wasn’t as bad as on Montecito, San Ysidro, Buena Vista and Romero Creeks because the canyon feeding it is smaller.
  2. When debris flows hit an obstruction, such as a bridge, they seek out a new bed to flow on. This is one of the actions that creates an alluvial fan. From the map it appears something like that happened—
    1. Where the flow widened when it hit Olive Mill Road, fanning east of Olive Mill to destroy all three blocks between Olive Mill and Santa Elena Lane before taking the Olive Mill bridge across 101 and down to the Biltmore while also helping other flows fill 101 as well. (See Mac’s comment below, and his link to a top map.)
    2. In the area between Buena Vista Creek and its East Fork, which come off different watersheds
    3. Where a debris flow forked south of Mountain Drive after destroying San Ysidro Ranch, continuing down both Randall and El Bosque Roads.

For those who caught (or are about to catch) Ellen’s Facetime with Oprah visiting neighbors, that happened among the red dots at the bottom end of the upper destruction area along San Ysidro Creek, just south of East Valley Road. Oprah’s own place is in the green area beside it on the left, looking a bit like Versailles. (Credit where due, though: Oprah’s was a good and compassionate report.)

Big question: did these debris flows clear out the canyon floors? We (meaning our geologists, sedimentologists, hydrologists and other specialists) won’t know until they trek back into the canyons to see how it all looks. Meanwhile, we do have clues. For example, here are after-and-before photos of Montecito, shot from space. And here is my close-up of the latter, shot one day after the event, when everything was still bare streambeds in the mountains and fresh muck in town:

See the white lines fanning back into the mountains through the canyons (Cold Spring, San Ysidro, Romero, Toro) above Montecito? Ed explained that these appear to be the washed out beds of creeks feeding into those canyons. Here is his slide showing Cold Spring Creek before and after the event:

Looking back at Ed’s basin threshold graphic above, one might say that there isn’t much sediment left for stream beds to yield, and that those in the floors of the canyons have returned to stability, meaning there’s little debris left to flow.

But that photo was of just one spot. There are many miles of creek beds to examine back in those canyons.

Still, one might hope that Montecito has now had its required 200-year event, and a couple more centuries will pass before we have another one.

Ed and Larry caution against such conclusions, emphasizing that most of Montecito’s and Santa Barbara’s inhabited parts gain their existence, beauty or both by grace of debris flows. If your property features boulders, Ed said, a debris flow put them there, and did that not long ago in geologic time.

For an example of boulders as landscape features, here are some we quarried out of our yard more than a decade ago, when we were building a house dug into a hillside:

This is deep in the heart of Santa Barbara.

The matrix mud we now call soil here is likely a mix of Juncal and Cozy Dell shale, Ed explained. Both are poorly lithified silt and erode easily. The boulders are a mix of Matilija and Coldwater sandstone, which comprise the hardest and most vertical parts of the Santa Ynez mountains. The two are so similar that only a trained eye can tell them apart.

All four of those geological formations were established long after dinosaurs vanished. All also accumulated originally as sediments, mostly on ocean floors, probably not far from the equator.

To illustrate one chapter in the story of how those rocks and sediments got here, UCSB has a terrific animation of how the transverse (east-west) Santa Ynez Mountains came to be where they are. Here are three frames in that movie:

What it shows is how, when the Pacific Plate was grinding its way northwest about eighteen million years ago, a hunk of that plate about a hundred miles long and the shape of a bread loaf broke off. At the top end was the future Malibu hills and at the bottom end was the future Point Conception, then situated south of what’s now Tijuana. The future Santa Barbara was west of the future Newport Beach. Then, when the Malibu end of this loaf got jammed at the future Los Angeles, the bottom end of the loaf swept out, clockwise and intact. At the start it was pointing at 5 o’clock and at the end (which isn’t), it pointed at 9:00. This was, and remains, a sideshow off the main event: the continuing crash of the Pacific Plate and the North American one.

Here is an image that helps, from that same link:

Find more geology, with lots of links, in Making sense of what happened to Montecito. I put that post up on the 15th and have been updating it since then. It’s the most popular post in the history of this blog, which I started in 2007. There are also 58 comments, so far.

I’ll be adding more to this post after I visit as much as I can of Montecito (exclusion zones permitting). Meanwhile, I hope this proves useful. Again, corrections and improvements are invited.

30 January

6 April, 2020
*I was told later, by a rescue worker who was on the case, that it was possible that both victims’ bodies had washed all the way to the ocean, and thus will never be found.

In this Edhat story, Ed Keller visits a recently found prior debris flow. An excerpt:

The mud and boulders from a prehistoric debris flow, the second-to-last major flow in Montecito, have been discovered by a UCSB geologist at the Bonnymede condominiums and Hammond’s Meadow, just east of the Coral Casino.

The flow may have occurred between 1,000 and 2,000 years ago, said Ed Keller, a professor of earth science at the university. He’s calling it the “penultimate event.” It came down a channel of Montecito Creek and was likely larger on that creek than during the disaster of Jan. 9, 2018, Keller said. Of 23 people who perished on Jan. 9, 17 died along Montecito Creek.

The long interval between the two events means that the probability of another catastrophic debris flow occurring in Montecito in the next 1,000 years is very low, Keller said.

“It’s reassuring,” he said, “They’re still pretty rare events, if you consider you need a wildfire first and then an intense rainfall. But smaller debris flows could occur, and you could still get a big flash flood. If people are given a warning to evacuate, they should heed it.”

Montecito is now a quarry with houses in it:

So far twenty dead have been removed. It will take much more time to remove twenty thousand dump truck loads of what geologists call “debris,” just to get down to where civic infrastructure (roads, water, electric, gas) can be fixed. It’s a huge thing.

The big questions:

  1. Did we know a catastrophe this huge was going to happen? (And if so, which among us were the “we” who knew?)
  2. Was there any way to prevent it?

Geologists had their expectations, expressed as degrees of likelihood and detailed on this map by the United States Geological Survey:

That was dated more than a month before huge rains revised to blood-red the colors in the mountains above town. Worries of County Supervisors and other officials were expressed in The Independent on January 3rd and 5th. Edhat also issued warnings on January 5th and 6th.

Edhat’s first report began, “Yesterday, the National Weather Service issued a weather briefing of a potential significant winter storm for Santa Barbara County on January 9-10. With the burn scar created by the Thomas Fire, the threat of flash floods and debris/mud flows is now 10 times greater than before the fire.”

But among those at risk, who knew what a “debris/mud flow” was—especially when nobody had ever seen one of those anywhere around here, even after prior fires?

The first Independent story (on January 3rd) reported, “County water expert Tom Fayram said county workers began clearing the debris basins at San Ysidro and Gobernador canyons ‘as soon as the fire department would let us in.’ It is worth noting, Lewin said, that the Coast Village Road area flooded following the 1971 Romero Fire and the 1964 Coyote Fire. While touring the impact areas in recent days, (Office of Emergency Management Director Robert) Lewin said problems have already occurred. ‘We’re starting to see gravity rock fall, he said. ‘One rock could close a road.'”

The best report I’ve seen about what geologists knew, and expected, is The Independent‘s After the Mudslides, What Does the Next Rain Hold for Montecito?, published four days after the disaster. In that report, Kevin Cooper of the U.S. Forest Service said, “no one alive has probably ever seen one before.” [January 18 update: Nick Welch in The Independent reports, “Last week’s debris flow was hardly Santa Barbara’s first. Jim Stubchaer, then an engineer with County Flood Control, remembers the avalanche of mud that took 250 homes back in November 1964 when heavy rains followed quickly on the heels of the Coyote Fire. He was there in 1969 and 1971 when it happened again.” Here is a long 2009 report on the Coyote Fire in The Independent by Ray Ford, now with Noozhawk. No mention of the homes lost in there. Perhaps Ray can weigh in.]

My point is that debris flows over Montecito ae a sure bet in geologic time, but not in the human one. In the whole history of Montecito and Santa Barbara (of which Montecito is an unincorporated part), there are no recorded debris flows that started on mountain slopes and spread all the way to the sea. But on January 9th we had several debris flows on that scale, originating simultaneously in the canyons feeding Montecito, San Ysidro and Romero Creeks. Those creeks are dry most of the time, and beautiful areas in which to build homes: so beautiful, in fact, that Montecito is the other Beverly Hills. (That’s why all these famous people have called it home.)

One well-studied prehistoric debris flow in Santa Barbara emptied a natural lake that is now Skofield Park,dumping long-gone mud and lots of rocks in Rattlesnake Canyon, leaving its clearest evidence in a charming tree-shaded boulder field next to Mission Creek called Rocky Nook Park.

What geologists at UCSB learned from that flow is detailed in a 2001 report titled UCSB Scientists Study Ancient Debris Flows. It begins, “The next ‘big one’ in Santa Barbara may not be an earthquake but a boulder-carrying flood.” It also says that flood would “most likely occur every few thousand years.”

And we got one in Montecito last Tuesday.

I’ve read somewhere that studies of charcoal from campfires buried in Rocky Nook Park date that debris flow at around 500 years ago. This is a good example of how the geologic present fails to include present human memory. Still, you can get an idea of how big this flow was. Stand in Rattlesnake Canyon downstream from Skofield Park and look at the steep rocky slopes below houses on the south side of the canyon. It isn’t hard to imagine the violence that tore out the smooth hillside that had been there before.

To help a bit more with that exercise, here is a Google Streetview of Scofield Park, looking down at Santa Barbara through Rattlesnake Canyon:

I added the red line to show the approximate height of the natural dam that broke and released that debris flow.

I’ve also learned that the loaf-shaped Riviera landform in Santa Barbara is not a hunk of solid rock, but rather what remains of a giant landslide that slid off the south face of the Santa Ynez Mountains and became free-standing after creeks eroded out the valley behind. I’ve also read that Mission Creek flows westward around the Riviera and behind the Mission because the Riviera itself is also sliding the same direction on its own tectonic sled.

We only see these sleds moving, however, when geologic and human time converge. That happened last Tuesday when rains Kevin Cooper calls “biblical” hit in the darkest hours, saturating the mountain face creek beds that were burned by the Thomas Fire just last month. As a result, debris flows gooped down the canyons and stream valleys below, across Montecito to the sea, depositing lots of geology on top of what was already there.

So in retrospect, those slopes in various colors in the top map above should have been dark red instead. But, to be fair, much of what geology knows is learned the hard way.

Our home, one zip code west of Montecito, is fine. But we can’t count how many people we know who are affected directly. One friend barely escaped. Some victims were friends of friends. Some of the stories are beyond awful.

We all process tragedies like this in the ways we know best, and mine is by reporting on stuff, hopefully in ways others are not, or at least not yet. So I’ll start with this map showing damaged and destroyed buildings along the creeks:

At this writing the map is 70% complete. [January 17 update: 95%.] I’ve clicked on all the red dots (which mark destroyed buildings, most of which are homes), and I’ve copied and pasted the addresses that pop up into the following outline, adding a few links.

Going downstream along Cold Spring Creek, Hot Springs Creek and Montecito Creek (which the others feed), gone are—
  1. 817 Ashley Road
  2. 817 Ashley Road (out building)
  3. 797 Ashley Road
  4. 780 Ashley Road. Amazing architectural treasure that last sold for $12.9 million in ’13.
  5. 809 Ashley Road
  6. 809 Ashley Road (there are two at one address)
  7. 747 Indian Lane
  8. 631 Parra Grande Lane. That’s the mansion where the final scene in Scarface was shot.
  9. 590 Meadowood Lane
  10. 830 Rockbridge Road
  11. 800 Rockbridge Road
  12. 790 Rockbridge Road
  13. 787 Riven Rock Road B
  14. 1261 East Valley Road
  15. 1240 East Valley Road A (mansion)
  16. 1240 East Valley Road B (out building)
  17. 1254 East Valley Drive
  18. 1255 East Valley Road
  19. 1247 East Valley Road A
  20. 1247 East Valley Road B (attached)
  21. 1231 East Valley Road A
  22. 1231 East Valley Road B (detached)
  23. 1231 East Valley Road C (detached)
  24. 1221 East Valley Road A
  25. 1221 East Valley Road B
  26. 369 Hot Springs Road
  27. 341 Hot Springs Road A
  28. 341 Hot Springs Road B
  29. 341 Hot Springs Road C
  30. 355 Hot Springs Road
  31. 335 Hot Springs Road A
  32. 335 Hot Springs Road B
  33. 333 Hot Springs Road (Not marked in final map)
  34. 341 Hot Springs Road A
  35. 341 Hot Springs Road B
  36. 341 Hot Springs Road C
  37. 340 Hot Springs Road
  38. 319 Hot Springs Road
  39. 325 Olive Mill Road
  40. 285 Olive Mill Road
  41. 275 Olive Mill Road
  42. 325 Olive Mill Road
  43. 220 Olive Mill Road
  44. 200 Olive Mill Road
  45. 275 Olive Mill Road
  46. 180 Olive Mill Road
  47. 170 Olive Mill Road
  48. 144 Olive Mill Road
  49. 137 Olive Mill Road
  50. 139 Olive Mill Road
  51. 127 Olive Mill Road
  52. 196 Santa Elena Lane
  53. 192 Santa Elena Lane
  54. 179 Santa Isabel Lane
  55. 175 Santa Elena Lane
  56. 142 Santo Tomas Lane
  57. 82 Olive Mill Road
  58. 1308 Danielson Road
  59. 81 Depot Road
  60. 75 Depot Road
Along Oak Creek—
  1. 601 San Ysidro Road
  2. 560 San Ysidro Road B
Along San Ysidro Creek—
  1. 953 West Park Lane
  2. 941 West Park Lane
  3. 931 West park Lane
  4. 925 West park Lane
  5. 903 West park Lane
  6. 893 West park Lane
  7. 805 W Park Lane
  8. 881 West park Lane
  9. 881 West park Lane (separate building, same address)
  10. 1689 Mountain Drive
  11. 900 San Ysidro Lane C (all the Lane addresses appear to be in San Ysidro Ranch)
  12. 900 San Ysidro Lane Cottage B
  13. 900 San Ysidro Lane Cottage A
  14. 900 San Ysidro Lane Cottage D
  15. 900 San Ysidro Lane E
  16. 900 San Ysidro Lane F
  17. 900 San Ysidro Lane G
  18. 900 San Ysidro Lane H
  19. 900 San Ysidro Lane I
  20. 900 San Ysidro Lane J
  21. 900 San Ysidro Lane K
  22. 900 San Ysidro Lane L
  23. 900 San Ysidro Lane M
  24. 900 San Ysidro Lane N
  25. 900 San Ysidro Lane O
  26. 900 San Ysidro Lane R
  27. 900 San Ysidro Lane S
  28. 900 San Ysidro Lane T
  29. 888 San Ysidro Lane A
  30. 888 San Ysidro Lane B
  31. 888 San Ysidro Lane C
  32. 888 San Ysidro Lane D
  33. 888 San Ysidro Lane E
  34. 888 San Ysidro Lane F
  35. 805 West Park Lane B
  36. 799 East Mountain Drive
  37. 1801 East Mountain Lane
  38. 1807 East Mountain Drive
  39. 771 Via Manana Road
  40. 899 El Bosque Road
  41. 771 Via Manana Road
  42. 898 El Bosque Road
  43. 800 El Bosque Road A (Casa de Maria)
  44. 800 El Bosque Road B (Casa de Maria)
  45. 800 El Bosque Road C (Casa de Maria)
  46. 559 El Bosque Road (This is between Oak Creek and San Ysidro Creek)
  47. 680 Randall Road
  48. 670 Randall Road
  49. 660 Randall Road
  50. 650 Randall Road
  51. 640 Randall Road
  52. 630 Randall Road
  53. 619 Randall Road
  54. 1685 East Valley Road A
  55. 1685 East Valley Road B
  56. 1685 East Valley Road C
  57. 1696 East Valley Road
  58. 1760 Valley Road A
  59. 1725 Valley Road A
  60. 1705 Glenn Oaks Drive A
  61. 1705 Glen Oaks Drive B
  62. 1710 Glen Oaks Drive A
  63. 1790 Glen Oaks Drive A
  64. 1701 Glen Oaks Drive A
  65. 1705 Glen Oaks Drive A
  66. 1705 East Valley Road A
  67. 1705 East Valley Road B
  68. 1705 East Valley Road C
  69. 1780 Glen Oaks Drive N/A
  70. 1780 Glen Oaks Drive (one on top of the other)
  71. 1774 Glen Oaks Drive
  72. 1707 East Valley Road A
  73. 1685 East Valley Road C
  74. 1709 East Valley Road
  75. 1709 East Valley Road B
  76. 1775 Glen Oaks Drive A
  77. 1775 Glen Oaks Drive B
  78. 1779 Glen Oaks Drive A
  79. 1779 Glen Oaks Drive B
  80. 1779 Glen Oaks Drive C
  81. 1781 Glen Oaks Drive A
  82. 1711 East Valley Road (This and what follow are adjacent to Oprah)
  83. 1715 East Valley Road A
  84. 1715 East Valley Road B
  85. 1719 East Valley Road
  86. 1721 East Valley Road A (This might survive. See Dan Seibert’s comment below)
  87. 1721 East Valley Road B (This might survive. See Dan Seibert’s comment below)
  88. 1721 East Valley Road C (This might survive. See Dan Seibert’s comment below)
  89. 1694 San Leandro Lane A
  90. 1694 San Leandro Lane D
  91. 1690 San Leandro Lane C
  92. 1690 San Leandro Lane A
  93. 1694 San Leandro Lane B
  94. 1696 San Leandro Lane
  95. 1710 San Leandro Lane A
  96. 1710 San Leandro Lane B
  97. 190 Tiburon Bay Lane
  98. 193 Tiburon Bay Lane A
  99. 193 Tiburon Bay Lane B
  100. 193 Tiburon Bay Lane C
  101. 197 Tiburon Bay Lane A
Along Buena Vista Creek—
  1. 923 Buena Vista Avenue
  2. 1984 Tollis Avenue A
  3. 1984 Tollis Avenue B
  4. 1984 Tollis Avenue C
  5. 670 Lilac Drive
  6. 658 Lilac Drive
  7. 2075 Alisos Drive (marked earlier, but I don’t see it in the final map)
  8. 627 Oak Grove Lane
Along Romero Creek—
  1. 1000 Romero Canyon Road
  2. 1050 Romero Canyon Road
  3. 860 Romero Canyon Road
  4. 768 Winding Creek Lane
  5. 745 Winding Creek Lane
  6. 744 Winding Creek Lane
  7. 2281 Featherhill Avenue B

Below Toro Canyon—

  1. 876 Toro Canyon Road
  2. 572 Toro Canyon Park Road

Along Arroyo Paredon, between Summerland and Carpinteria, not far east of the Toro Canyon—

  1. 2000 Cravens Lane

Ten flanking Highway 101 by the ocean are marked as damaged, including four on Padero Lane.

When I add those up, I get 142 163* 178† among the destroyed alone.

[* This is on January 17, when the map says it is 95% complete. All the additions appear to be along San Ysidro Creek, especially on San Ysidro Lane, which I believe is mostly in San Ysidro Ranch. Apparently nearly the whole place has been destroyed. Adjectives such as “lovely” fail to describe what it was.]

[† This is on January 18, when the map is complete. I’ll need to go over it again, because there are subtractions as well as additions. Additional note: on March 22, the resident at 809 Ashley Road asked me to make sure that address was also added. There are two homes at that address, both gone.]

Now let’s go back and look more closely at this again from the geological perspective.

What we see is a town revised by nature in full disregard for what was there before—and in full obedience to the pattern of alluvial deposition on the flanks of all fresh mountains that erode down almost as fast as they go up.

This same pattern accounts for much of California, including all of the South Coast and the Los Angeles basin.

To see what I mean, hover your mind above Atlanta and look north at the southern Appalachians. Then dial history back five million years. What you see won’t look much different. Do the same above Los Angeles or San Francisco and nothing will be the same, or even close. Or even there at all.

Five million years is about 1/1000th of Earth’s history. If that history were compressed to a day, California showed up in less than the last forty seconds. In that short time California has formed and re-formed constantly, and is among the most provisional landscapes in the world. All of it is coming up, sliding down, spreading out and rearranging itself, and will continue doing so through all the future that’s worth bothering to foresee. Debris flows are among nature’s most casual methods for revising landscapes. (By the way, I am writing this in a San Marino house that sits atop the Raymond Fault scarp, which on the surface takes the form of a forty-foot hill. The stack of rock strata under the bottom of that hill is displaced 17,000 feet from the identical suite under the base at the top. Many earthquakes produced that displacement, while erosion has buffed 16,960 feet of rock and soil off the top.)

So we might start to look at the Santa Ynez Mountains behind Santa Barbara and Montecito not as a stable land form but rather as a volcano of mud and rock that’s sure to go off every few dozen or hundreds of years—and will possibly deliver a repeat performance if we get more heavy rains and there is plenty of debris left to flow out of mountain areas adjacent to those that flowed on January 9th. If there’s a lot of it, why even bother saving Montecito?

Here’s why:

One enters the Engineering building at the University of Wyoming under that stone plaque, which celebrates what may be our species’ greatest achievement and conceit: controlling nature. (It’s also why geology is starting to call our present epoch the anthropocene.)

This also forecasts exactly what we will do for Montecito. In the long run we’ll lose to nature. But meanwhile we strive on.

In our new strivings, it will help to look toward other places in California that are more experienced with debris flows, because they happen almost constantly there. The largest of these by far is Los Angeles, which has placed catch basins at the mouths of all the large canyons coming out of the San Gabriel Mountains. Most of these dwarf the ones above Montecito. All resemble empty reservoirs. Some are actually quarries for rocks and gravel that roll in constantly from the eroding creek beds above. None are pretty.

To understand the challenge involved, it helps to read John McPhee’s classic book The Control of Nature, which takes its title from the inscription above. Fortunately, you can start right now by reading the first essay in a pair that became the relevant chapter of that book. It’s free on the Web and called Los Angeles Against the Mountains I. Here’s an excerpt:

Debris flows amass in stream valleys and more or less resemble fresh concrete. They consist of water mixed with a good deal of solid material, most of which is above sand size. Some of it is Chevrolet size. Boulders bigger than cars ride long distances in debris flows. Boulders grouped like fish eggs pour downhill in debris flows. The dark material coming toward the Genofiles was not only full of boulders; it was so full of automobiles it was like bread dough mixed with raisins.

The Genofiles were a family that barely survived a debris flow on a slope of Verdugo Mountain, overlooking Los Angeles from Glendale. Here’s another story, about another site not far away:

The snout of the debris flow was twenty feet high, tapering behind. Debris flows sometimes ooze along, and sometimes move as fast as the fastest river rapids. The huge dark snout was moving nearly five hundred feet a minute and the rest of the flow behind was coming twice as fast, making roll waves as it piled forward against itself—this great slug, as geologists would describe it, this discrete slug, this heaving violence of wet cement. Already included in the debris were propane tanks, outbuildings, picnic tables, canyon live oaks, alders, sycamores, cottonwoods, a Lincoln Continental, an Oldsmobile, and countless boulders five feet thick. All this was spread wide a couple of hundred feet, and as the debris flow went through Hidden Springs it tore out more trees, picked up house trailers and more cars and more boulders, and knocked Gabe Hinterberg’s lodge completely off its foundation. Mary and Cal Drake were standing in their living room when a wall came off. “We got outside somehow,” he said later. “I just got away. She was trying to follow me. Evidently, her feet slipped out from under her. She slid right down into the main channel.” The family next door were picked up and pushed against their own ceiling. Two were carried away. Whole houses were torn loose with people inside them. A house was ripped in half. A bridge was obliterated. A large part of town was carried a mile downstream and buried in the reservoir behind Big Tujunga Dam. Thirteen people were part of the debris. Most of the bodies were never found.

This is close to exactly what happened to Montecito in the wee hours of January 9th. (As of March 22, two of the 23 dead still haven’t been recovered, and probably never will be.) (In September 2018 a first responder I talked with said the bodies of a least one the two missing victims, a teenage boy and a toddler, were probably carried to the ocean.)

As of now the 8000-plus residents of Montecito are evacuated and forbidden to return for at least another two weeks—and maybe much longer if officials declare the hills above town ready to flow again.

Highway 101—one of just two major freeways between Southern and Northern California, is closed indefinitely, because it is now itself a stream bed, and re-landscaping the area around it, to get water going where it should, will take some time. So will fixing the road, and perhaps bridges as well.

Meanwhile getting in and out of Santa Barbara from east of Montecito by car requires a detour akin to driving from Manhattan to Queens by way of Vermont. And there have already been accidents, I’ve heard, on highway 166, which is the main detour road. We’ll be taking that detour or one like it on Thursday when we head home via Los Angeles after we fly there from New York, where I’m packing up now.

Expect this post to grow and change.

Bonus links:

Tags: , , , ,

The original version of this ran as a comment under Francine Hardaway‘s Medium post titled Have we progressed at all in the last fifty years?

My short answer is “Yes, but not much, and not evenly.” This is my longer answer.


In your case and mine, it has taken the better part of a century to see how some revolutions take generations to play out. Not only won’t we live to see essential revolutions complete; our children and grandchildren may not either.

Take a topic not on your list: racial equality—or moving past race altogether as a Big Issue. To begin to achieve racial equality in the U.S., we fought the Civil War. The result was various degrees of liberation for the people who had been slaves or already freed in Union states; but apartheid of both the de jure and de facto kind persisted. Jim Crow laws and practices emerged, and in still live on in culture if not in law.

The civil rights movement in the Fifties and Sixties caused positive social, political and other changes. The Civil Rights Act of 1964 especially helped. But the murders of Martin Luther King Jr. and Robert F. Kennedy in 1968 put civil rights almost back where it was before its revolution started. I participated in civil rights activism in Greensboro, North Carolina at the time of both assassinations, and I can’t overstate how deep and defeating our despair felt after both events. And that feeling proved correct.

Small incremental improvements followed over the decades since, but no leaps forward like we had before those murders. (Even the election of Barack Obama failed to change a terribly durable status quo. Backlash against that election is at least partly responsible for Trump and the Republican Congress.)

We are still stuck with inequality for races, religions and so much else. Will we ever get over that? I think we will, inevitably; but only if our species survives.

One collateral victim of those assassinations in the Sixties was the near-end of non-violence as a strategy toward change. Martin Luther King Jr. used it very effectively, and kept the flame alive and well-proven until violence took him out. Martyred though he was, it was not to the cause of nonviolence or pacifism, both of which have been back-burnered for fifty years. We (in the largest sense that includes future generations) may never find out if non-violence can ever succeed—because violence is apparently too deeply ingrained as a human trait.

Back to tech.

I too was, and remain, a cyber-utopian. Or at least a cyber-optimist. But that’s because I see cyber—the digitization and networking of the world—as a fait accompli that offers at least as many opportunities for progress as it does for problems. As Clay Shirky says, a sure sign of a good technology is that one can easily imagine bad uses of it.

What I’m not writing at the moment are my thoughts about why some of those advantaged by power, even in small ways, abuse it so easily. I’m not writing it because I know whatever I say will be praised by some, rebuked by others, and either way will be reduced to simplicities that dismiss whatever subtle and complex points I am trying to make, or questions I am trying to ask. (Because my mind is neither sufficiently informed nor made up.) I also know that, within minutes for most of my piece’s readers, the points it makes will be gone like snow on the water, for such is the nature of writing on the vast sea of almost-nothing that “social” media comprises. And, as of today, all other media repose in the social ones.

Some perspective:

Compared to that, and its effects on the planet, all other concerns shrink to insignificance.

But, as The Onion said a few weeks after 9/11, A Shattered Nation Longs to Care About Stupid Bullshit Again.

Stupid bullshit is what the meteor of humanity hitting the planet cares most about. Always has. Wars have been fought over far less.

The only fully consequential question is how we end the Anthropocene. Or how it ends without us.

Tags:

bigbust

Emanuele Orazio Fenzi, better known as Francesco Franceschi (1843-1924), was an Italian horticulturist responsible for vastly increasing the botanical variety of Santa Barbara (introducing more than 900 species). He was also for awhile the primary landowner on the Riviera, a loaf-shaped hill overlooking the city’s downtown. Most of that hill is now covered with houses, but a large part that isn’t is what remains of the Franceschi estate: 18 acres called Franceschi Park, featuring a crumbling mansion and the bust above, carved into the top of a boulder on the property.

The city doesn’t have much to say about Franceschi, with a website devoted to the park that goes one paragraph deep. This makes sense, because the state of neglect in the park is extreme. I won’t go into details, because they’re well presented all these stories:

Wikipedia, at the top link above, goes deep too. So does this 2002 Pacific Horticulture story, which suggests with this photo—

2002_jas-chamberlin-001-660x896

—that the bust above isn’t a bad likeness.

But that boulder and Franceschi’s head are going to be shards on the road soon if the city, or somebody, doesn’t save it. Simply put, the ground under it is giving way. Take a look. Here’s the bust, on its boulder, a few feet above the ground that has fallen down to Mission Ridge Road below:

fail1

And here you can see the failing slope, and the rubble that has fallen from within it onto the road:

fail2

I shot that a couple days ago, in a break between this winter’s record breaking rainstorms. And here’s a closer look at the slo-mo landslide happening immediately below the sculpture:

fail3Saving Franceschi’s bust is surely an easier job than saving his house. What I’m hoping here is that publishing this blog post will stir up some interest.

BluecutFireTo get away from the heat today—into a little less heat and an excuse to exercise, I drove up to Mt. Wilson, where I visited the Observatory and walked around the antenna farm there. As it happened, the Bluecut Fire was also visiting the same San Gabriel Mountains, a few miles to the east at Cajon Pass. Starting at 10:36 in the morning, it was past 10,000 acres with 0% containment by the time I observed it in the mid to late afternoon.

Here’s a photo set. If anybody wants to use any of them, any way they please, feel free.

The view here is to the east, along the spine of the range, across 10,064-foot (3068m) Mt. San Antonio, also known as Old Baldy. I like to ski there (at Mt. Baldy) in the winter. Nothing like skiing nearly two miles up, looking down on 20 million people enjoying subtropical weather. The lifts are open in the summer (for zip-lining), so you can get up there and watch the fire from a closer (but safe) vantage, I assume. Check first.

 

Save

Save

Save

I’ve been fascinated for years by what comes and goes at the Fort Irwin National Training Center

fortirwin

—in the Mojave Desert, amidst the dark and colorful Calico Mountains of California, situated in the forbidding nowhere that stretches between Barstow and Death Valley.

Here and there, amidst the webwork of trails in the dirt left by tanks, jeeps and other combat vehicles, fake towns and other structures go up and come down. So, for example, here is Etrebat Shar, a fake town in an “artificial Afghanistan” that I shot earlier this month, on June 2:

etrebat-shar1

And here is a broader view across the desert valley east of Fort Irwin itself:

etrebat-shar2

Look to the right of the “town.” See that area where it looks like something got erased? Well, it did. I took the two shots above earlier this month, on June 2. Here’s a shot of the same scene on June 25, 2013:

etrebat-shar3

Not only is the “town” a bit bigger, but there’s this whole other collection of walls and buildings, covering a far larger area, to the right, or east.

I also see in this shot that it was gone on December 8, 2014.

Now I’m fascinated by this town and the erased something-or-other nearby, which I also shot on June 2:

othertown

It appears to be “Medina Wasl,” which Wikipedia says is one of twelve towns built for desert warfare training:

One of the features of the base is the presence of 12 mock “villages” which are used to train troops in Military Operations in Urban Terrain (MOUT) prior to their deployment. The villages mimic real villages and have variety of buildings such as religious sites, hotels, traffic circles, etc. filled with foreign language speaking actors portraying government officials, local police, local military, villagers, street vendors, and insurgents. The largest two are known as Razish and Ujen, the closest located about 30 minutes from the main part of the post. Most of the buildings are created using intermodal containers, stacked to create larger structures, the largest village consists of 585 buildings that can engage an entire brigade combat team into a fight.

Now I’m slowly going through my other shots over the years to see if I can find Razish and Ujen… if they haven’t been erased.

It would be cool to hear from military folk familiar with Fort Irwin, or veterans who have worked or fought mock battles in those towns.

Oil from the Coal Oil Seep Field drifts across Platform Holly, off the shore of UC Santa Barbara.

Oil from the Coal Oil Seep Field drifts across Platform Holly, off the shore of UC Santa Barbara.

Oil in the water is one of the strange graces of life on Califonia’s South Coast.

What we see here is a long slick of oil in the Pacific, drifting across Platform Holly, which taps into the Elwood Oil Field, which is of a piece with the Coal Oil Point Seep Field, all a stone’s throw off Coal Oil Point, better known as UC Santa Barbara.

Wikipedia (at the momentsays this:

The Coal Oil Point seep field offshore from Santa Barbara, California isa petroleum seep area of about three square kilometres, adjacent to the Ellwood Oil Field, and releases about 40 tons of methane per day and about 19 tons of reactive organic gas (ethane, propane, butane and higher hydrocarbons), about twice the hydrocarbon air pollution released by all the cars and trucks in Santa Barbara County in 1990.[1]The liquid petroleum produces a slick that is many kilometres long and when degraded by evaporationand weathering, produces tar balls which wash up on the beaches for miles around.[2]

This seep also releases on the order of 100 to 150 barrels (16 to 24 m3) of liquid petroleum per day.[3] The field produces about 9 cubic meters of natural gas per barrel of petroleum.[2]

Leakage from the natural seeps near Platform Holly, the production platform for the South Ellwood Offshore oilfield, has decreased substantially, probably from the decrease in reservoir pressure due to the oil and gas produced at the platform.[2]

On the day I shot this (February 10), from a plane departing from Santa Barbara for Los Angeles, the quantity of oil in the water looked unusually high to me. But I suppose it varies from day to day.

Interesting fact:

  • Chumash canoes were made planks carved from redwood or pine logs washed ashore after storms, and sealed with asphalt tar from the seeps. There are no redwoods on the South Coast, by the way. The nearest are far up the coast at Big Sur, a couple hundred miles to the northwest. (It is likely that most of the redwood floating into the South Coast came from much farther north, where the Mendicino and Humboldt coastlines are heavily forested with redwood.)
  • National Geographic says that using the tar had the effect of shrinking the size of Chumash heads over many generations.
  • There are also few rocks hard enough to craft into a knife or an ax anywhere near Santa Barbara, or even in the Santa Ynez mountains behind it. All the local rocks are of relatively soft sedimentary kinds. Stones used for tools were mostly obtained by trade with tribes from other regions.

Here’s the whole album of oil seep shots.

« Older entries