Anybody who reads this blog knows I post sporadically at best. I go through some periods of great energy, and then this space can lie fallow for months at a time. Which isn’t to say I don’t frequently have ideas for posts: it’s a more a question of whether the idea that pops into my head is something I can stick the landing on. I have a folder on my desktop full of half-written posts that I’ve either lost the thread on, couldn’t make work to my satisfaction, or simply was distracted by a shiny thing, and by the time I think about returning to the post in progress, it is no longer timely.
This year was interesting: I blogged 35 times, which is not a lot, but then my posts were largely clustered in the first half of the year. I had a fair bit of momentum coming out of 2020, and was propelled through January and February by events (most notably the assault on the Capitol and Biden’s inauguration). I posted eight times in January, which is a lot for me; then my productivity was reduced by two each month following until April (with two posts). May and June were saw five and six posts, respectively, in part because I was being ambitious and attempting to produce several posts on a handful of themes. That tapered off in July … and then nothing until November (once), and a single December post.
I’m never entirely sure why the well goes dry on a fairly regular basis, though as I said, often it’s not so much about the writing as about the finishing (I still have sitting on my desktop a post about galactic empires that I do want to finish). Sometimes it’s reflective of how productive I’m being otherwise, but not always; sometimes my blog is a useful procrastinatory device, something that makes me feel productive when I should be directing my energies elsewhere.
At any rate, I thought I might do something I’ve seen other blogs do, which is a year in review with a list of the best/most read/favourite posts. Given that my readership here is pretty tiny, it would be a bit silly to list my most popular posts. So I’m going with my personal favourites: which is to say, the posts was proudest of, and which I felt managed to get closest to the thoughts that spawned them.
I’m going with my top five, though I’m not ranking them, just listing them in chronological order.
It’s a little odd that, of these top five, two of them deal with the topic of monarchy. This post is about how the American Republican system of government—developed specifically as a revolt against the tyrannical British crown—has ironically ended up imbuing the American chief executive with more king-like qualities than the prime minister in a parliamentary system. This contradiction, I point out, had become all the more glaring in the Age of Trump, whose authoritarian tendencies exacerbated the monarchical elements of the Office of the President.
I will also note that I posted this entry the day before the January 6 insurgency.
As I have noted many times on this blog, I wrote my doctoral dissertation on conspiracy and paranoia in postwar American fiction, film, and popular culture. Through the successive waves of Trutherism, Birtherism, Glenn Beck’s chalkboard rants, and the various paranoid fantasies spun by Tea Partiers, I have thought about dusting off the thesis … and am overcome with the sensation of probing the nerve of a tooth.
QAnon, like everything it does, ratchets that up to eleven. And yet I found myself writing this post, which I think does a pretty good job of breaking down the important elements. And I’ve watched the HBO docuseries Q: Into the Storm and read the book The Storm is Upon Us by Mike Rothschild … and now I find myself writing about it for another project.
It makes me feel like Al Pacino in Godfather III: “Every time I try to get OUT … they keep dragging me back IN.”
My second of two posts on monarch was prompted, perhaps counter-intuitively, by my profound indifference to the various plights of the British royal family. The occasion of this particular bout of indifference was the fallout from Meghan Windsor née Markle’s interview with Oprah Winfrey and the fallout it caused.
But if you’re so indifferent, why go to the effort of writing a blog post, you may ask? Well, imaginary interlocutor, I started pondering precisely why hereditary monarchy has such a powerful hold on the contemporary imagination. And that led me down a rabbit hole of thought that proved quite interesting.
Also, it gave me the opportunity to write the sentence “Piers Morgan was the result of a lab experiment in which a group of scientists got together to create the most perfect distillation of an asshole.” So there’s that, too.
I made two attempts at sustained deep dives into large topics this summer. The first was “History, Memory, and Forgetting,” and the other was a series of posts revisiting the concept of postmodernism (“Remembering Postmodernism”). I have to say, I was very pleased with what I produced on both fronts, and annoyed with myself that the postmodernism series got bogged down and remains unfinished. (It’s for that reason that “Remembering Postmodernism” is not represented here).
To be honest, I think all three of my “History, Memory, and Forgetting” posts deserve a place here, but because they’re sort of a unit, I’ll settle for the one I’m proudest of, which is about how the erosion of memory about the Holocaust—through time, distraction, and the death of survivors—has denuded the historical awareness and created a present situation in which such terms as “Nazi” and “fascist” have lost meaning in the popular imagination.
Early this summer, I was alerted to a backlash against the Tolkien Society’s Summer Seminar—an annual conference in which academics present papers on a theme chosen by the society. Part themes have included Tolkien and Landscape, the Art of Tolkien’s World, and so on. This year? Tolkien and Diversity. Which prompted a not-unpredictable backlash in conservative circles, especially after the paper titles were posted. Though it hardly reached “critical race theory” levels of vitriol, there was an awful lot of angry talk about the “woke mob” coming to tear down J.R.R. Tolkien. Though most of this happened on message boards and social media, it did reach the lofty heights of the National Review.
A few days after I wrote this post, I attended the (virtual) Tolkien Seminar conference and watched every single paper. Guess what? Tolkien’s legacy survived. And guess what else? Everyone presenting at that conference loves Tolkien. No one wanted to tear him down. The very thought would horrify them. The biggest fallacy of “anti-woke” thought—which, really, stretches back to the culture wars of the 90s and Harold Bloom’s castigation of what he called “the school of resentment”—is the idea that people who challenge the traditional canons of art and literature and offer feminist, queer, anti-colonialist, anti-racist, or other such readings are doing so because they hate and resent the genius of such canonical writers as Shakespeare, Milton, or Wordsworth (an irony here being that Tolkien has never been included in “the Western Canon”). While there may be genuinely antagonistic readings of classic authors, most of the time people—like those at the Tolkien seminar—are finding spaces in their work in which they see themselves reflected.
And in the end, it’s a testament to Tolkien’s genius that queer graduate students can find themselves in the work of an ultraconservative Catholic. To those who lambasted the “Tolkien and Diversity” seminar, I ask: how is that a bad thing?
Happy new year, everyone. Did you all have a good New Year’s Eve? Stephanie and I celebrated by ordering pizza and watching Don’t Look Up on Netflix—a film that is at once simultaneously so hilarious and so depressing that I found myself wondering whether watching it on the last day of 2021 was a terrible idea or entirely appropriate … though I suppose it could be both.
I hate New Year’s Eve as a night of celebration, and I always have—even when I was younger and more disposed to party like it’s 1999. A good friend of mine had the perfect summary of why NYE is so terrible. There are two days a year, he said, when you feel the most societal pressure to enjoy yourself: your birthday and NYE. Provided you have friends and/or family who love you, birthdays are great fun because they’re all about you. On New Year’s Eve, by contrast, it’s everybody’s birthday. Which is possibly why the more excessive the celebrations, the more they smack of desperation.
Normally I would have prepared a nice meal as a reluctant nod to mark the day, but I just got back from my parents’ place in Ontario five days ago, which means I’ve been quarantining. Which also means I can’t leave the house until tomorrow and buy groceries, and while we don’t lack for the basics, there isn’t much with which to make anything more then, well, the basics.
So, pizza. Which, given my antipathy to this particular holiday, seemed even more appropriate than anything requiring effort on my part. It might have to become the new custom.
I’ve been seeing a lot of 2021-related memes on social media, most of which involved (1) WTF? And (2) warnings not to jinx the coming year with high expectations, which we did at the start of 2021. But when you think about it, we’ve been exiting the year with a snarl and a backward-facing middle finger since … well, 2016, haven’t we? Which makes many of us want to blame Donald Trump for this series of successively sucky years, but if we haven’t yet collectively understood that he’s not the architect of our societal woes, just the bellwether, then things are gonna keep sliding downhill.
I mean … they probably will anyway, but the big reason for 2021’s unreasonably high expectations was the tacit assumption that with Trump out of office, things would inevitably get better.
And for a time, it looked like they were! And then … well, I was about to say that reality reasserted itself, but that wasn’t the biggest problem with this past year, was it? It would be more accurate to say that unreality reasserted itself. The Big Lie, anti-vaxxers, the hysteria over “critical race theory” and other bogus culture war non-issues, January 6 trutherism, and of course the ongoing state of climate-change denial. Reality has never been the problem, except insofar as accessing it without having to run a gauntlet of disinformation is now more or less impossible.
This profound sense of frustration and disconnect is why Don’t Look Up has landed so hard on its viewers. I laughed throughout the film, because it’s hard not to—it is comedy, based in the wilful obliviousness and ignorance of people so self-interested, so elementally selfish, that they are congenitally incapable of recognizing the very idea of a collective good. Many times through the film, I found myself thinking, “Wow, this is unsubtle.” But then … so is Tucker Carlson. So is QAnon. So is Donald J. Trump, and so are his legions of enablers and imitators. We live, sad to say, in profoundly unsubtle times.
Or as Stephanie put it midway through the movie, “It’s really depressing when there’s no daylight between satire and the thing it’s satirizing.”
There’s a moment that happens on occasion in American football, when a running back, having been handed the ball, breaks through all the defenders and sprints into the open field, with nothing between him and the endzone.
And sometimes, rather than making a touchdown, he gets blindsided by a linebacker he didn’t see coming.
That’s what 2021 has been like, especially these past few months.
I was vaccinated in June, and then again in August. We resumed in-person classes at Memorial University in September. Like many people, I felt a massive sense of relief—the pandemic finally seemed to be heading into the rearview mirror, and things were returning to something resembling normalcy. Except … yeah, not so much. I’m writing this at my parents’ house on Christmas day, but we won’t be properly celebrating the holiday until tomorrow because my niece and nephew were possibly exposed to COVID on a schoolbus, and thus need to quarantine for another day. I will myself have to quarantine for five days on returning to St. John’s.
We should be in the clear, running to the endzone with open field, but we keep getting tackled.
This is what I told all my students earlier this term: I wanted them to know it wasn’t just them, it was a general malaise that I was also experiencing. I told them they weren’t alone in feeling anxious, discomfited, or just generally off. I told them that if it came down to it, if they were struggling to get their classwork done, that there was no shame in dropping my course … or any other course they were taking. I told them that I would do whatever I could to help, provided they kept me in the loop. You don’t need to give me the gory details, I said—just tell me you’re having a rough go of it, and that’s enough. We’ll work something out.
I have to imagine there are those who will say I’m catering to the fragility of the snowflake generation in making such accommodations. Anyone having that thought can fuck right off. What saved me from breakdown these past few months was the fact that I had such extraordinary students: as I said to all my classes at the term’s end, they’re the reason why I have hope for the world. As a GenXer, irony is my default setting; for those inheriting the catastrophes of climate change and resurgent fascism, earnestness is their baseline. What would have been radically uncool in the 1990s is quite possibly what will save us all.
On Christmas day, I find myself thinking about all the best parts of the past year. My students are, as they often are, Exhibit A.
I taught a graduate seminar in the winter term that, in spite of the fact that it happened on Zoom, was hands down my favourite class ever. It was called “The Spectre of Catastrophe,” and looked at 21st-century post-apocalyptic narratives. In spite of the fact that the subject was perhaps a little bit too on the nose for our current situation, it was the most fun I’ve had as a teacher ever (and that’s saying a lot, as my following comments will make clear). The only advantage classes taught over Zoom have over in-person teaching is the comments students can write as the lecture/discussion unfolds. In this particular class, there was always an ongoing conversation scrolling up on the side. I used the ten-minute breaks each hour in our three-hour classes to read through the lively and often hilarious discussions happening in parallel to the actual class.
Getting back into the classroom this past September felt so very, very good. It didn’t hurt that I was teaching a roster of courses that were, for me, an embarrassment of riches: a first-year class called “Imagined Places,” which gave me the (gleeful) chance to teach Tolkien (The Hobbit), Ursula K. Le Guin (A Wizard of Earthsea), Sir Terry Pratchett (Small Gods), Neil Gaiman (The Ocean at the End of the Lane), and Jo Walton (The Just City) (for the record, everybody needs to read Jo Walton). I also taught, for the third time, “Reading The Lord of the Rings.” I had five, count ‘em FIVE, students in LOTR who were also in my fourth-year seminar “The American Weird: Lovecraft, Race, and Genre as Metaphor.” (I always measure my success as an educator by the number of students who take multiple classes with me. I assume it means I’m doing something right, though it’s also possible they’ve just got my number and know what it takes to get a decent grade). Given that the Tolkien and the Lovecraft courses were back-to-back, I had something like an entourage following me from the chemistry building in which I taught LOTR to the Arts buildings that was the home of the weird.
I called my Lovecraft students “weirdos,” because, well, obviously. It only offended them the first time I called them that.
Even given the fact that I was teaching a dream slate of classes, this fall semester was hard. I am congenitally incapable of asking for help, or for that matter recognizing that I need help until after the fact; these past few days of decompressing at my parents’ home have been invaluable in offering relaxation and reflection. I have also realized I need to find my way to a therapist in 2022.
But the moral of the story is that the students I worked with this past year kept me sane, and gave me hope. So on this day of prayer (even for those as atheistic as myself) I am grateful for you all. And given how many of you have signed up for yet more Lockett pontification next term, I can only assume I’m doing something right.
Or you’ve got my number. Either way, it’s all good.
Wow, it’s been a minute. Four months and no posts is … well, as my hiatuses from this blog go, it’s fairly standard.
I’ve been thinking a lot lately about science fiction, which I suppose is hardly unusual for me. To be more specific, I’ve found that my SF thoughts have been squaring up of late with contemporary events and situations, enough so that I’ve been writing them down so as to try to make sense of my inchoate thoughts in a significant way. This post, which is very long, is one of two I’ve been pecking away at for about two weeks.
And again—this is quite long, but I didn’t want to lop it up. I have however divided it into sections, so maybe read it in stages if you don’t feel like slogging through it all at once.
The Books that Shape Us
An internet sage once said: “There are two novels that can change a bookish fourteen-year old’s life: The Lord of the Rings and Atlas Shrugged. One is a childish fantasy that often engenders a lifelong obsession with its unbelievable heroes, leading to an emotionally stunted, socially crippled adulthood, unable to deal with the real world. The other, of course, involves orcs.”
I’ve wondered on occasion about the cult of Ayn Rand among those whom she most venerated, billionaires and multi-millionaires (and the politicians who have built careers catering to the needs of the über-wealthy); I wonder if it was reading Rand’s odes to selfishness as “bookish fourteen-year olds” that set them on the path of rapacious capitalist accumulation, or whether Atlas Shrugged came, post-facto, to provide a useful pseudo-intellectual justification for why they should have vast wealth while 99%+ of people do not.
The question is rhetorical, of course, even perhaps somewhat disingenuous: while I have little doubt that some people did in fact find their life’s mission revealed to them in the pages of Rand’s writing at an early age (Ted Cruz springs to mind as an obvious candidate), the people at the extremes of that opposition are probably few and far between. I know a significant number of people who devoured The Fountainhead and Atlas Shrugged as teens, who later discovered that what they found compelling about the novels when they were young was completely out of step with their adult sensibilities. That of course is not an uncommon experience with certain works of literature more generally: I read The Catcher in the Rye at the age of sixteen; given that Salinger’s novel is scientifically calibrated to resonate with angsty teenage boys, I read it in a single feverish sitting, but doubt I could tolerate Holden Caulfield for more than a few pages today. By contrast, I was a bookish twelve-year old when I read The Lord of the Rings, and it did indeed change my life—it remains a core text in my life, as evidenced by the fact that I am currently teaching my department’s Tolkien course for the third time.1
I recently shared an anecdote from Neil Gaiman with my first-year English students. In a lecture delivered at The Reading Agency in 2013, Gaiman said,
I was in China in 2007, at the first party-approved science fiction and fantasy convention in Chinese history. And at one point I took a top official aside and asked him Why? SF had been disapproved of for a long time. What had changed? It’s simple, he told me. The Chinese were brilliant at making things if other people brought them the plans. But they did not innovate and they did not invent. They did not imagine. So they sent a delegation to the US, to Apple, to Microsoft, to Google, and they asked the people there who were inventing the future about themselves. And they found that all of them had read science fiction when they were boys or girls.
To be clear, reading SF when you’re young doesn’t necessarily translate into becoming a tech genius or innovator (sometimes it means you become an English professor); what Gaiman is advocating in his lecture is reading promiscuously, and especially making available for children as wide a range of reading options as possible: “The simplest way to make sure that we raise literate children is to teach them to read, and to show them that reading is a pleasurable activity. And that means, at its simplest, finding books that they enjoy, giving them access to those books, and letting them read them.” He cites the story from his participation in the SF/F convention in China not to suggest that SF makes you innovative, but that the kind of creative imagination exemplified by SF and a multiplicity of other genres—which facilitates an imaginative engagement with the impossible and the unreal—is a precondition for inventiveness.
As with Rand’s objectivist fiction, however, there’s always a danger in reading stuff prescriptively. I’ve been thinking about this story of Gaiman’s a lot lately, though not so much in terms of the positive sense of it I communicated to my students. I have, rather, been watching as a trio of billionaires seem stuck in what I can only characterize as a sort of arrested development corresponding to certain eras of SF—two of them speaking confidently about colonizing Mars and the third apparently determined to build cyberspace—and none of them seem to have truly understood the SF paradigms they’re trying to make real.
Let’s take them in turn. Both Jeff Bezos and Elon Musk have been quite enthusiastic about the prospect of settling humans on other planets, starting with Mars. (I’m leaving Richard Branson out of this consideration for that reason: his spacefaring ambitions seem more in the line of having the coolest, fastest new vehicle, something made obvious by the sleek aesthetics of Virgin Galactic’s Unity spacecraft vs. Blue Origin’s phallic homage to Dr. Evil’s rocket).
Musk recently tweeted that his ultimate intention is “to get humanity to Mars and preserve the light of consciousness.”
Bezos is also ultimately determined to make it to Mars (though he’s asserted that returning to the moon is a necessary first step), and has reportedly been fascinated by space travel all his life—as valedictorian in high school, he ended his address with the line “Space: the final frontier. Meet me there.” His love of Star Trek, as evidenced in those words, is something he’s frequently expressed, most recently when he shot ninety-year-old William Shatner into space.
It’s hardly unusual for people to love works of art that articulate values at odds with their own—as much as I love LOTR, for example, I’m hardly down with Tolkien’s conservative Catholicism as it emerges in the novel’s pages—but sometimes the dissonance can be jarring. Consider, for example, when former Republican Speaker of the House and Ayn Rand enthusiast Paul Ryan named Rage Against the Machine as his favourite band (which always begs the question: against which machine did you think Zack de la Rocha and Tom Morello were raging, Paul? Spoiler alert, it’s you). It’s unsurprising that a tech guy like Bezos would love Star Trek, but it’s also hard not to see a comparable dissonance between Gene Rodenberry’s utopian, egalitarian vision of the future and the rapacious business ethos embodied by Amazon.
It makes me wonder what the Trek-loving Bezos made of the Star Trek: The Next Generation episode in which the Enterprise comes across a 21st century satellite (somehow dislodged from Earth’s orbit and adrift in deep space) containing cryonically frozen people with terminal illnesses. They are unfrozen and their illnesses cured with 24th century medical technology, which of course had been the point of being frozen, kept in stasis until the day they could be treated. One of the unfrozen people is a stereotypical Wall Street douchenozzle who initially gloats over the success of his plan to cheat death—only possible, he notes, because of his considerable wealth—but is horrified to learn that the long-term investments he’d expected to cash in on would now be nonexistent. Captain Picard tells him, “A lot has changed in the past three hundred years. People are no longer obsessed with the accumulation of things. We have eliminated hunger, want, the need for possessions. We’ve grown out of our infancy” (ep. 1.25, “The Neutral Zone”).
Even in its grittier iterations, Star Trek always retains the germ of Rodenberry’s utopianism, born in the American mid-century before postwar optimism would truly start to fracture under the repeated shocks of political assassinations, the Vietnam War, and the cynical politics of the Nixon administration. “The final frontier” was a specific echo of John F. Kennedy’s “New Frontier” rhetoric; and while there is no lack of cynicism or dystopian predictions in the SF of the 1950s and 60s, the broader trends were more hopeful: one reads the SF of that period and finds, if not always optimism necessarily, then certainly a prevalent sense of the possibilities created by advanced technology … and, as with Star Trek, a not-uncommon faith that such technological advances will be matched by humanity’s moral maturation.
Undoubtedly, apologists for the Musk/Bezos space race would tell me that just because the billionaires are the contemporary embodiment of being “obsessed with the accumulation of things,” it doesn’t obviate the altruism of their visionary space-faring ambitions; that the groundwork laid by these early, seemingly trivial joyrides to the edge of space will be crucial for future larger-scale space travel; that governments have failed in such endeavours, so it falls to the billionaire class to carry humanity into the next stage of our development; almost certainly the Musk/Bezos devotees (especially the Muskovites) would tell me I’m a small-minded person resentful and jealous of their genius as manifested in their vast wealth. Possibly Bezos himself would say something to the effect of how he loves the utopianism of Star Trek and hopes for such a future, but to bring it about he needs to play by the rules of the world as it is today.
It’s possible that some of these things might end up being true; I suppose it’s even possible that my conviction that the very fact of billionaires’ existence in societies where children go hungry is a moral obscenity simply masks my own true desire to accumulate wealth on a cosmic scale. Possible, but unlikely; it strikes me that Bezos and Musk and their cheering sections might have science-fictional enthusiasms about colonizing space, but don’t seem to have paid much attention to the transformations in SF’s preoccupations. While we don’t lack for contemporary SF depicting the spread of humanity through space and the colonization of far-flung planets, the utopian spirit and assumptions of the mid-century has diminished rather dramatically. SF has become darker in its outlook—dystopian figurations are more pervasive, especially those imagining post-apocalyptic futures; increasingly, visions of the future have become tempered by the current environmental realities of the climate crisis and the Anthropocene; and even those imagined futures featuring the colonization of other planets largely eschew such tacit utopian assumptions like those of Star Trek, in which humanity’s expansion into the solar system and beyond entails progressing beyond the factious politics of the present. More and more, space-focused SF grapples specifically with the stubborn realities not just of the vast distances involved in space travel, but of human nature as well (I’m also currently writing a longish post about how, in the present moment, the most unbelievable aspect of much SF is the idea that any given planet could have one-government rule, never mind a galaxy-spanning empire or federation). A good example of this the Expanse novels by James S.A. Corey, which have been adapted into an excellent television series—ironically enough—by Amazon. In the future envisioned by The Expanse, humanity has colonized the Moon, Mars, and a bunch of moons and asteroids collectively known as “the Belt.” One of the things both the novels and the series does well is convey the sense of scale, the enormous distances involved just in our local solar system. Space travel is punishing, inflicting potentially stroke inducing Gs on travelers; people born and raised in the low gravity of Mars or the even lower gravity of the Belt cannot endure the gravity of Earth; and resentments, hatreds, and regional prejudices mark the attitudes of Earther, Martians, and Belters to each other and Earth’s imperial history plays out again in its exploitation of the Belt’s resources. Plus ça change, plus c’est la même chose.
The impossible-to-comprehend vastness of space is central to Kim Stanley Robinson’s Aurora, about a generation ship’s three-century journey to the nearest possibly, hopefully, hospitable planet (spoiler alert: it isn’t). More germane to Musk and Bezos’ Mars ambitions is Andy Weir’s novel The Martian, popularized by the film adaptation starring Matt Damon. When botanist Mark Watney is stranded on Mars, his only way to survive is—in the now-famous line—“to science the shit out of this.” And as he notes, his margin for error is nonexistent: “If the oxygenator breaks down, I’ll suffocate. If the water reclaimer breaks down, I’ll die of thirst. If the hab[itat] breaches, I’ll just kind of implode. If none of those things happen, I’ll eventually run out of food and starve to death.”
The novels of Robinson and Weir are exemplars of “extrapolative” SF, which is to say: SF that is scrupulous in its science, hewing closely to the possible and plausible. What is interesting in the context of what I’m talking about is a shift in focus—The Martian isn’t a trenchant argument against space exploration, but it also eschews an imaginative leap ahead to a Mars in the early stages of terraforming, with safe habitats for thousands of colonists to live, instead focusing on the granular problems involved with even beginning to think about such a project Again, the Musk/Bezos contingent would certainly say that of course we’d have to start small, that it would be harsh and difficult for a long time, but that the billionaire visionaries are the ones laying the groundwork.
To which I say: read the room. Or better yet, read the SF of the past twenty years. Really, just scan the spines in the SF section of a decent bookstore and count the number of novels from the past two decades that are dystopian, post-apocalyptic, or which, like John Scalzi’s recent “Interdependency” trilogy (The Collapsing Empire, The Consuming Fire, The Last Emperox) are essentially allegories about the current climate crisis. Then do a quick calculation and see how much of SF’s current market share comprises such narratives. Because it’s not as though dystopian stories and post-apocalyptic SF are at all new;2 they’re just more prevalent than ever,3 reflecting a zeitgeist more preoccupied with the depredations of modernity and technology than with their capacity for continued expansion.
All of which is by way of saying, if you want the TL;DR: this planet is our home and it would be great to fix the problems we have with it before seriously thinking of colonizing others—not least because the colonization of Mars, even if it proves possible, is the work not of years but centuries, centuries we simply don’t have. The resources Musk and Bezos are bringing to bear on their respective dreams of space travel would be much more profitably devoted to developing, among other things, green energy. Mars ain’t going anywhere.
But then, even billionaires are mortal. It’s possible to see their current competition less as mere dick-measuring (though it is that, to be certain) than as a race to satisfy their dreams of space travel while they still draw breath (in this respect, one wonders if sending ninety-year old William Shatner into space was not just an homage to Bezos’ favourite SF, but a cynical calculation on the part of the fifty-seven-year old billionaire to reassure himself he’s still got at least three decades of space faring years). Elon Musk has said he wants to die on Mars; as with many of his pronouncements, it’s not at all obvious if he’s serious or just taking the piss, but it does cast his ostensible determination, as quoted above, “to get humanity to Mars and preserve the light of consciousness,” in a somewhat more fatalistic light—it seems to suggest that the “light of consciousness” isn’t likely to survive on Earth, and must therefore be seeded elsewhere. It’s difficult not to interpret these words in the context of the broader trend of doomsday prepping among the super-rich that has seen multi-millionaires and billionaires buying remote property on which to build bunkers, and stockpiling food, supplies, guns, and ammunition against the anticipated collapse of civilization. At the time of writing, Elon Musk was accounted the wealthiest man in the world, with his fortune exceeding $300B; it makes a sort of perverse sense that, as status-bunkers go, having one on an entirely different planet would win the game.
I am still stumped, however, by how Musk means “light of consciousness.” Is it him? Does he mean to be found, eons from now, cryonically frozen like the people on The Next Generation by an alien species, the last remaining human? Because it seems to me that the light of human consciousness is best carried into the future by teeming billions living on a planet saved by their collective effort. Sending the light of consciousness to Mars is more akin to the final words of Margaret Atwood’s short fiction “Time Capsule Found on a Dead Planet”: “You who have come here from some distant world, to this dry lakeshore and this cairn, and to this cylinder of brass, in which on the last day of all our recorded days I place our final words: Pray for us, who once, too, thought we could fly.”
… or, from outer space to inner space.
I tried to watch Mark Zuckerberg’s ninety-minute infomercial in which he announced the launch of the “metaverse” and Facebook’s rebranding as “Meta,” and lasted all of about five minutes. I had no expectation of getting through the entire thing—nor of seriously attempting to—but you’d have thought I could have endured more. But apparently not … I didn’t know whether to laugh or cringe, so I did a fair bit of both. It was just so damned earnest, and it baffled me how this putative tech genius could talk so enthusiastically and unironically about this new utopian virtuality, with no acknowledgment of (1) the dystopian shitshow that is the current state of social media generally and Facebook specifically, and (2) the fact that there has yet to be a depiction of virtual reality in fiction that is not even more dystopian than Facebook IRL.
The word “disingenuous” doesn’t even scratch the surface. “Sociopathically delusional” might be closer to the mark.
The next day I subjected my long-suffering first-year students to an impromptu rant about how SF authors in the 80s and 90s had anticipated all of this—specifically, a digital reality embedded in a broader condition in which massive transnational corporations have more power than most nation-states, wealth disparity is a vast and unbridgeable chasm, industry has irrevocably poisoned the environment, and the broader populace is distracted with the soporific of digital entertainment and misinformation—and this was why everybody needed to take more English courses (don’t ever tell me I don’t shill for my own department). I noted that the very term “metaverse” was used by SF writer Neal Stephenson to describe the online virtual environment in his 1992 novel Snow Crash; I also noted that depictions of virtual reality appeared long before computers were ubiquitous, such as in Philip K. Dick’s 1969 novel Ubik, and that William Gibson wrote Neuromancer (1983), which became the definitive cyberpunk novel, on a manual typewriter without ever having owned a personal computer.
And I emphasized the point I made above: that in no instance were any of these depictions anything other than dystopian.
But then, it’s sadly unsurprising that the man who not only invented Facebook, but who for all intents and purposes is Facebook, should forge ahead with the arrogant certainty that he knows better than all the people who envisioned the metaverse long before it was a technical possibility. If Musk and Bezos’ arrested development is stuck sometime in the SF of the 1960s, Zuckerberg’s is in the 80s—but with apparently even less understanding of the material than his billionaire bros.
I could go on with all the elements of cyberpunk Zuckerberg obviously doesn’t get, but I think you get the point … and besides which, my main concern here is with the way in which the choice to rebrand Facebook as “Meta” is perfectly emblematic of Zuckerberg’s apparently congenital inability to engage in self-reflection. It is also, not to put too fine a point on it, what irritates me so profoundly: I have what you might characterize as a professional beef with this rebranding, as it entails the flagrant misuse of a prefix that is crucial to contemporary, and especially postmodernist, literature and culture.
What is “meta-“? Notably, in his infomercial and the letter posted to Facebook announcing the rebranding, Zuckerberg says that “meta” is from the Greek and means “beyond.” This much is true: the second entry in the OED’s definition of meta- is “beyond, above, at a higher level.” Zuckerberg goes on to say that, for him, it “symbolizes that there is always more to build, and there is always a next chapter to the story.” Except that that is entirely not the sense in which meta- is used, not even colloquially. It is not uncommon to hear people say things like “that’s so meta!” or “can we get meta about this?” In both cases, what is meant is not a sense of taking things further, but of reflection and reconsideration. “That’s so meta!” is precisely the kind of thing you might expect to hear from someone watching a movie like Kiss Kiss Bang Bang, a noirish murder mystery that constantly draws attention to the tropes, conventions, and clichés of noirish murder mystery, and which further features Robert Downey Jr.’s character frequently breaking the fourth wall to remind us that, yes, we’re watching a noirish murder mystery movie. The suggestion “can we get meta about this?” usually precedes a discussion not of the issue at hand, but rather things surrounding, informing, framing, or giving rise to the issue at hand.
In other words, as the next line in the OED definition tells us, meta- is typically “prefixed to the name of a subject or discipline to denote another which deals with ulterior issues in the same field, or which raises questions about the nature of the original discipline and its methods, procedures, and assumptions.” Hence, metafiction is fiction about fiction, metacriticism is criticism about criticism, and Kiss Kiss Bang Bang—whose title is taken from a disparaging comment by film critic Pauline Kael about what audiences want from movies—is metacinema (or metanoir, if you want to be more specific). And while the practice of being endlessly self-referential can grow tiresome,5 my point is more that meta- is, literally by definition, introspective.
Something Mark Zuckerberg manifestly is not. To be meta- about “Meta” would not involve plunging forward with a tone-deaf and bloody-minded ninety-minute infomercial promising to double down on Facebook’s putative mission of “bringing people together.” Rather, one might reflect upon the assertion that “We’re still the company that designs technology around people,” and wonder instead if all the internal studies establishing that Instagram is toxic to adolescent girls’ body-image and that Facebook facilitates atrocities and props up dictatorships are perhaps more indicative of technology shaping people’s behaviour. Perhaps all that SF from the 80s and 90s might help in this process?
1. I didn’t read anything by Ayn Rand until I was in grad school. When I finally did read her, it was in part out of curiosity, but more significantly because by then I’d had the basic tenets of objectivism explained to me in some detail; I found the entire philosophy abhorrent, as you might imagine, but also figured that were I ever to find myself in argument with a libertarian about the relative merits of her fiction, I should at least know more precisely what I was talking about. So first I read The Fountainhead, which was a fairly quick read and the more engaging of her two most notable novels. I completely understood its appeal to adolescent readers, especially precocious teens convinced of their own rightness and unsparing in their hatred of anything reeking of compromise and “selling out.” (Thinking about the novel now, it strikes me as possessing a notably modernist sensibility, if not style: architect Howard Roark’s ultimate determination to demolish his building rather than allow “lesser” architectural minds to taint the putative genius of his design reminds me of little else than the unfortunately common tendency among scholars of modernism to express categorical disdain for anything they consider unworthy of study).
Atlas Shrugged was more of a slog, and I often wondered as I read it (in fits and starts over several weeks) at how feverishly its devotees tore through its 1000+ pages, and then obsessively reread it over and over. It’s not that it’s a difficult read—it’s certainly not particularly complex, nor is the prose dense or opaque—I just had difficulty caring about the characters and their stories. I’ve nodded sympathetically many, many times when people have confessed to me that Tolkien’s digressions into the history of Middle-earth lost them, but the purplest of passages in LOTR ain’t got nothin’ on John Galt’s sixty-page-long paean to the moral virtues of selfishness and greed.
A number of years ago, as Breaking Bad was airing its final episodes, I hit on an idea for an article about how that series functioned in part as a sustained and trenchant critique of Ayn Rand. I was fairly deep into drafting it when I hit a wall—specifically, the resigned realization that, if I was going to make a serious run at writing it, I’d have to reread Atlas Shrugged. And, well, that was a bridge too far … though I did write a blog post sketching out the general argument.
2. The critical consensus is that Mary Shelley’s Frankenstein (1818) was the first SF novel, and Frankenstein’s nothing if not a dystopian warning about the unthinking and unethical use of technology. Eight years later she published The Last Man, a post-apocalyptic story set in the midst of a deadly pandemic, so it’s safe to say she was ahead of her time.
3. The corollary to this upswing in post-apocalyptic SF is the proliferation post-apocalyptic narratives that don’t quite classify as SF, with the biggest and most obvious example being the critical mass of zombie apocalypse in fiction, film, television, and games. The corollary of this corollary is the concomitant prevalence of fantasy, with which zombie apocalypse shares DNA insofar as it posits a return to a premodern state of existence. If much SF is dystopian and articulates a bleak outlook, the zombies and the dragons are a kind of nostalgia for a world stripped of the modernity that went so very wrong.
4. Yes, yes, it was Facebook’s “virtual connect” conference … but really, it was an infomercial. A long, tedious, cringeworthy infomercial.
5. I recently watched the new Netflix film Red Notice, which was entirely forgettable, in part because of just how meta- it was. The biggest eye-roll for me was a scene in which the FBI agent and the master thief played, respectively, by Dwayne “The Rock” Johnson and Ryan Reynolds, are searching an arms dealer’s antiquities collection for a certain ancient Egyptian treasure that precipitates the action of the film. “Where is it?” mutters The Rock, to which Reynolds responds, “Check for a box labelled ‘MacGuffin’.”
HOUSEKEEPING NOTE: There has been an unwanted lag in this series of posts, mainly because I’ve been struggling with what was supposed to be part three. Struggling in a good way! The TL;DR on it is that it occurred to me that a consideration of the artistic and literary responses to the two world wars respectively offers a useful insight into key elements of modernism and postmodernism. What initially seemed a straightforward, even simple breakdown has proved (not unpredictably, I now see) a lot more complex but also a lot more interesting. What I ultimately post may end up being the more straightforward version or possibly a two-part, lengthier consideration. One way or another, I’m quite enjoying going down this particular rabbit hole.
So in the meantime, in the interests of not letting this series lag too much, I’m leapfrogging to what was to have been part four.
My doctoral dissertation was on conspiracy theory and paranoia in postmodern American literature and culture. Towards the end of my defense, one of my examiners—a modernism scholar with a Wildean talent for aphoristic wit—asked “Is modernism postmodernism’s conspiracy theory?”
I should pause to note that I thoroughly enjoyed my thesis defense. This was largely because my examiners really liked my thesis, and so instead of being the pressure cooker these academic rites of passage can sometimes be, it was three hours of animated and lively discussion that the defense chair had to bring to a halt with some exasperation when we ran long. For all that, however, it was still three hours of intense scholarly back-and-forth, and so when my examiner asked about modernism as postmodernism’s conspiracy theory, my tank was nearly empty. I found myself wishing—and indeed gave voice to the thought—that the question had been asked in the first hour rather than the third, when I could have done it justice. I still kick myself to this day for not prefacing my response with a Simpsons reference (something the questioner would have appreciated), Reverend Lovejoy’s line “Short answer, yes with an if, long answer, no with a but,” and then getting into the intricacies and implications of the question. As it was, I seem to remember saying something insightful like “Um … sure?”
I’ve thought about that moment many times in the seventeen years (yikes) since I defended, mostly out of fondness for the questioner, who was and remains a friend, as well as annoyance with myself for not giving the question an answer it deserves (now that I think of it, perhaps I should include that in this series). But over the past few years, I’ve thought of it more in the context of how postmodernism itself—and its related but widely misunderstood concept “cultural Marxism”—have come to be treated as essentially conspiratorial in nature.
As I’ve alluded to in previous posts, there is a (mis)understanding of postmodernism among anti-woke culture warriors as something specifically created by Leftists for the purpose of attacking, undermining, and destroying the edifices of Western civilization, as variously manifested in Enlightenment thought, the U.S. constitution, the traditional Western literary canon, manly men, the virtues of European imperialism, and so on. Postmodernity, rather than being the upshot of unchecked corporate capitalism and consumer culture, is seen instead as being the specific invention of resentful and closeted Marxist academics.
Of late, which is to say over the past five years, the most vocal purveyor of this conspiracy theory has been Jordan B. Peterson and his figuration of “postmodern neo-Marxism.” Anyone who knows even the slightest thing about either postmodernism or Marxism understands that, in this formulation, it is the word “neo” doing the heavy lifting—which perhaps betrays at least a slight understanding on Peterson’s part that there can be no such thing as “postmodern Marxism,” as the two terms are very nearly antithetical. Marxism is a modernist philosophy rooted in Enlightenment thought; what I’ve been loosely calling “postmodern thought,” which is to say the loose categorization of theories and philosophy arising largely out of the need to make sense of the postmodern condition, is generally antagonistic to the instrumental reason of the Enlightenment and such totalizing ideologies as Marxism, taking its philosophical leads instead from Friederichs Nietzsche, Edmund Husserl, and Martin Heidegger. So if you’re going to attempt to conflate Marxism with postmodernism, it’s going to have to be very neo- indeed.
Peterson’s basic premise is that the “two architects of the postmodernist movement”—specifically, the French theorists Michel Foucault and Jacques Derrida—were themselves dyed-in-the-wool Marxists; but that when they were making their academic bones in the 1960s, they were faced with the unavoidable failure of Communism as a political force. In one of his ubiquitous YouTube lectures, he declares that “in the late 60s and early 70s, they were avowed Marxists, way, way after anyone with any shred of ethical decency had stopped being Marxist.”
The “postmodernists,” Peterson continues, “knew they were pretty much done with pushing their classic Marxism by the late 60s and the early 70s,” because the evidence of Stalin’s atrocities were by then so unavoidable that carrying on under the Marxist banner was untenable.
This assertion, as it happens, is laughably untrue: the Communist Party of France reliably garnered twenty percent of the legislative vote through the 1960s and 70s. There was no shortage of people, inside the university and out, who were “avowed” Marxists. There was really no reason Derrida and Foucault—who incidentally hated each other, so were hardly co-conspirators—would have been compelled to disguise their Marxist convictions. If indeed they had any: it is an irony that Peterson can make this argument in part because neither Derrida nor Foucault were avowed Marxists. They have, indeed, often been looked upon with suspicion by Marxist scholars, and frequently castigated (Derrida especially) for precisely the reasons I cited above: namely, that they eschewed the principles of Marx’s teleological philosophy and an extrinsic historical order. They were, to coin an expression, a little too “postmodern” for Marxists’ tastes.
Though Peterson is by no means voicing an original idea—the charge that “cultural Marxists” comprise a shadowy cabal of professors seeking to destroy Western civilization was first articulated in the early 1990s—he does imbue his attack with his own uniquely greasy brand of ad hominem logic familiar to anyone who has taken issue with his many transphobic screeds. See if you can spot the code words:
Foucault in particular, who was an outcast and a bitter one, and a suicidal one, and through his entire life did everything he possibly could with his staggering I.Q. to figure out every treacherous way possible to undermine the structure that wouldn’t accept him in all his peculiarity—and it’s no wonder, because there’d be no way of making a structure that could possibly function if it was composed of people as peculiar, bitter, and resentful as Michel Foucault.
Michel Foucault, for those unfamiliar with him, was queer; much of his work was preoccupied with the ways in which people marginalized and stigmatized by their sexuality were policed by society, and the ways in which that policing—that exercise of power—was effected discursively through the designations of mental illness (Madness and Civilization, 1961), the disciplining of society via surveillance (Discipline and Punish, 1975), and the ways in which the categorization of sexual identities exemplifies the normative determination of the self (The History of Sexuality, four volumes, 1976, 1984, 1984, and 2018).
Peterson’s characterization of Foucault is, in this respect, frankly vile—as is his description of Foucault when he first introduces him into his discussion: “A more reprehensible individual you could hardly ever discover, or even dream up, no matter how twisted your imagination.” His repetition of the word “peculiar” is an obvious dog-whistle, and he damns Foucault for being an “outcast,” “bitter,” and “suicidal,” as if Foucault’s “outcast” status as a queer man with a galaxy-sized brain was somehow a character flaw rather than a function of the strictures of a society he understandably took umbrage with. Peterson might be the psychologist here, but I do sense a certain amount of animosity and revulsion that is not entirely directed at Foucault’s philosophy.
One way or another, the charge here is that Foucault and Derrida effectively invented postmodernism as a means of sublimating their doctrinaire Marxism into something more insidious and invidious, which would burrow into university humanities departments like a virus; still speaking of Foucault, Peterson says, “In any case, he did put his brain to work trying to figure out (a) how to resurrect Marxism under a new guise, let’s say, and (b) how to justify the fact that it wasn’t his problem that he was an outsider, it was actually everyone else’s problem.” Some fifty-odd years later, goes the Peterson narrative—which, again, is not specific to him, but prevalent among his fellow-travelers on the so-called “intellectual Dark Web”— we’re dealing with the harvest of what Derrida and Foucault sowed in the form of Black Lives Matter and critical race theory, trans people asserting their right to their preferred pronouns, cancel culture, and the general upending of what cishet white men perceive as the natural order of things.
We can argue over how we got to this moment in history, but the idea that the postmodern condition—or whatever we want to call the present moment—was orchestrated by a handful of resentful French intellectuals should be relegated to the same place of shame as most conspiracy theories. As I’ve been attempting to argue in this series of posts is that while thinkers like Foucault and Derrida have indeed profoundly influenced postmodern thought, they are not—nor are any of their acolytes—responsible for the cultural conditions of postmodernity more broadly. They have, rather, attempted to develop vocabularies that can describe what we’ve come to call postmodernity.
 The substance of Peterson’s conspiracy theory of postmodernism I quote here is from an invited lecture he delivered at the University of Wisconsin in 2017, posted to YouTube (the relevant bit starts at 29:30)
 Though Peterson acknowledges that Foucault and Derrida aren’t the only two masterminds of postmodern thought, they are, to the best of my knowledge, the only two he ever really talks about. I find it odd that, as a professor of psychology and practicing psychologist, he never (again, to the best of my knowledge) ever deals with the work of psycholinguist Jacques Lacan, whose poststructuralist adaptations of Freud are almost as influential as the work of Foucault and Derrida. Given how consistently he gets wrong the basic premises of Foucault and Derrida—but especially Foucault—the absence of Lacan from his diatribes strikes me as further evidence of the poverty of his understanding of the very issues he addresses.
 Derrida never tipped his ideological hand one way or another until his book Specters of Marx (1994), which he wrote in the aftermath of the collapse of the Soviet Union and the putative death of Communism. In this book he does his deconstructivist schtick, playing around with the trope of the “spectre,” talking a lot about the ghost of Hamlet’s father, and sort of admitting “Yeah, I was always a Marxist.” Some Marxist thinkers were overjoyed; many more were decidedly unimpressed, deriding him as a Johnny-come-lately only willing to assume the Marxist mantle as he sat among what he assumed was its ruins. In an essay bitterly titled “Marxism without Marxism,” Terry Eagleton wrote: “it is hard to resist asking, plaintively, where was Jacque Derrida when we needed him, in the long dark night of Reagan-Thatcher,” and continues on to say, “there is something rich … about this sudden dramatic somersault onto a stalled bandwagon. For Specters of Marx doesn’t just want to catch up with Marxism; it wants to outleft it by claiming that deconstruction was all along a radicalized version of the creed.”
 For those familiar with Peterson, this is consonant with his worldview, especially with respect to his anti-transgender animus. He is an unreconstructed Jungian: which is to say, he believes fervently in a sense of the biological imperatives of mythology—that all of our stories and narratives, our societal customs and traditions, are dictated by our most elemental relationships to nature. Hence his weird grafting of pseudo-Darwinian evolutionism onto myths of all stripes, from the Bible to ancient Egypt to the Greeks, and how these innate understandings manifest themselves in the popularity of Disney princesses or the necessary disciplinary presence of the bully Nelson in The Simpsons (seriously). The bottom line is that Peterson’s worldview is predicated on a sense of the innate, immutable nature of gender, gender roles, hierarchies, and the individual as the hero of his own story. Feminism, in this perspective, is a basic betrayal of human nature; people identifying as a gender other than their genitalia dictates? Well, that’s just beyond the pale.
 The opacity of those vocabularies, especially with regards to Derrida and Lacan, is another example of just how complex the postmodern condition is. Old joke from grad school: What do you get when you cross the Godfather with a poststructuralist? Someone who’ll give you an offer you can’t understand.
[PROGRAMMING NOTE: I have not abandoned my “Remembering Postmodernism” series—I’m just having problems putting the next installment into a coherent form that doesn’t run into Tolstoy-length rambling. In the interim, please enjoy this follow-up to my Tolkien and the culture wars post.]
I do so love serendipity. I would go so far as to say it is what most frequently inspires me in my research and writing. Unfortunately, that tends to make my primary research methodology not at all unlike the dog from Up.
This fall, I’ll be teaching a fourth-year seminar on the American Weird. In the original version of the course, which I taught three years ago, we looked at H.P. Lovecraft’s influence on American horror and gothic, and we considered as well how contemporary texts engaged specifically with Lovecraft’s hella racist tendencies—specifically with regard to Victor LaValle’s novella The Ballad of Black Tom, a retelling of Lovecraft’s story “The Horror at Red Hook” (one of his most explicitly racist stories) from the perspective of a Black protagonist, and Matt Ruff’s recently published Lovecraft Country.
Since then, Lovecraft Country was adapted to television by HBO under the imprimatur of Jordan Peele, and the magnificent N.K. Jemisin published an overtly Lovecraftian novel The City We Became. So this time around, we’re going all in on the issue of race and genre in “the American Weird,” not least because the current furor over “critical race theory” in the U.S. makes such a course at least a little timely.
But of course you’ve read the title of my post and might be wondering how this is serendipitous with the Tolkien society’s recent seminar. What on earth does Tolkien have to do with Lovecraft, or questions of race and identity in American literature?
It has to do with the issue of genre and the ways in which genre has come to be both a metaphor and a delineation of community, belonging, and exclusion. Genre has always been about drawing lines and borders: as Neil Gaiman has noted, genre is about letting you know what aisles of the bookstore not to walk down. But in the present moment, we’ve become more alert to this exclusionary function of genre, something that Lovecraft Country—both the novel and the series—took as its principal conceit, taking H.P. Lovecraft’s racist legacy and redeploying it to interrogate the ways in which genre can be opened up to other voices, other identities.
The greatest joy of participating in this Tolkien Society seminar was seeing precisely this dynamic in action—in seeing how an international and diverse cadre of Tolkien scholars found themselves within his works. As I said in my previous post, the backlash against the very idea of a “Tolkien and Diversity” conference employed the rhetoric of invasion and destruction. A post to a Tolkien message board epitomizes quite nicely the zero-sum sensibility I discussed:
The very first paper of the Tolkien Society seminar this weekend was titled “Gondor in Transition: A Brief Introduction to Transgender Realities in The Lord of the Rings.” As you might imagine, the very idea that there could be transgender characters in LOTR has been anathema to the various people slagging the seminar over social media. Tolkien’s world view did not allow for trans identity! has been one of the squawks heard from complainants. Leaving aside the fact that gender fluidity in Norse myth is something Tolkien would have been quite familiar with, this sort of complaint misses the point—which is that Tolkien’s work is voluminous enough to admit a multitude of readings he almost certainly did not have in mind as he composed his mythology.
I wish I could offer a useful precis of the paper, which was quite well done, but I was distracted by Zoom’s chat function and the avalanche of commentary scrolling up the side of my screen as I listened to the presentation—many of the comments being from trans and non-binary Tolkien enthusiasts expressing gratitude for a paper that made them feel seen in the context of Tolkien fandom and scholarship. This was actually the general tone of the two-day seminar: not people who, as the conference’s detractors have charged, look to destroy or desecrate Tolkien, but people who love his mythology and want to see themselves reflected within it. And who in the process demonstrated the capaciousness of Tolkien’s vision—one not limited, as the fellow cited above suggests, to the rigid circumscription of conservative Catholicism.
Genre is an interesting thing—less, from my perspective, for its more esoteric delineations within literary theory and criticism than for its blunter and cruder demarcations in popular culture. This latter understanding is where the real action has been for the past ten or twenty years, as the barriers walling off genre from the realms of art and literature have crumbled—with the advent of prestige television taking on the mob film, the cop procedural, and the western in The Sopranos, The Wire, and Deadwood, respectively; then proceeding to make fantasy and zombie apocalypse respectable with Game of Thrones and The Walking Dead; all while such “literary” authors as Cormac McCarthy, Margaret Atwood, Kazuo Ishiguro, and Colson Whitehead made their own forays into genre with The Road, Oryx and Crake, The Buried Giant, and Zone One, just as “genre” authors like Neil Gaiman, Alan Moore, Alison Bechdel and others attained “literary” reputations.
But as those walls have come down, so too have those genre enthusiasts of years past grown resentful of all the new people moving into their neighbourhoods. As I mentioned in my previous post, both the Gamergaters and Sad Puppies articulated a sense of incursion and loss—women and people of colour and queer folk invading what had previously been predominantly white male spaces. Like those attacking the Tolkien and Diversity seminar, they spoke in apocalyptic terms of things being ruined and destroyed and desecrated—SF/F would never be the same, they bleated, if these people had their way. What will become of our beloved space operas and Joseph Campbell-esque fantasy?
Well, nothing. They will still be there—fantasy in a neo-African setting rather than a neo-European one doesn’t obviate J.R.R. Tolkien or C.S. Lewis or any of the myriad authors they influenced. They’re still there, on your bookshelf or at the library or your local bookstore. However much we might use the metaphor of territory, it’s flawed: in the physical world, territory is finite; in the intellectual world, it’s infinite. Tolkien’s world contains multitudes, and can admit those people who see queer themes and characters and want to interrogate the possible hermaphroditic qualities of dwarves and orcs without displacing all its staunchly conservative readers.
We interrupt your regularly scheduled deep dive into postmodernism for a brief diversion into the latest anti-woke tempest in a teacup.
CORRECTION: when I started writing this post, I meant it as a “brief diversion.” It has grown in the writing and become decidedly non-brief. One might even call it very long. My apologies.
The Tolkien Society was founded in 1969 with J.R.R. Tolkien himself as its president; like most scholarly societies, it has hosted annual conferences at which people present papers exploring aspects of Tolkien’s and related works, and generally mingle to share their passion and enthusiasm for the subject. And like most of these such conferences, each year tends to have a different theme—looking over the society’s web page listing past conferences, we see such previous themes as “21st Century Receptions of Tolkien,” “Adapting Tolkien,” “Tolkien the Pagan?”, “Life, Death, and Immortality,” “Tolkien’s Landscapes,” and so on … you get the idea.
This year’s conference, to be hosted online on July 3-4, is on “Tolkien and Diversity.”
If this blog post was a podcast, I would here insert a sudden needle-scratch sound effect to indicate the sudden, sputtering outrage of all the anti-woke culture warriors who, fresh from their jeremiads about Dr. Seuss and perorations about critical race theory, are now charging that the woke Left has come for Tolkien because these SJWs will not rest until they have destroyed everything that is great and good about white Western civilization. Because, you know, a two-day conference in which people discuss issues of race, gender, and sexuality with regards to Tolkien’s work will destroy The Lord of the Rings.
If that all sounds snarky and hyperbolic, well, I’ll cheerfully concede to the snark, but I’m not overstating the reaction. Survey the Twitter threads on the subject, and you’ll see the phrase “Tolkien spinning in his grave”1 an awful lot, as well as the frequent charge—made with varying degrees of profanity and varying familiarity with grammar—that all of these people who want to talk about race or sexual identity in Tolkien are doing it because they hate Tolkien and seek to destroy this great author and his legacy. Well, you might protest, that’s just Twitter, which is an unmitigated cesspool; and while you wouldn’t be wrong about Twitter, these reactions exist on a continuum with more prominent conservative outrage. The default setting for these reactions is the conviction that people engaging with such “woke” discourse as critical race theory, or queer studies, or feminism for that matter, do so destructively—that their aim is to take something wholesome and good and tear it down, whether that be Western civilization or your favourite genre fiction.
I shouldn’t have to argue that this Tolkien conference, with its grand total of sixteen presenters, isn’t about to do irreparable damage to Tolkien’s standing or his legacy. Nor would reading some or all of the papers ruin LotR or The Hobbit for you. The idea that it might—that this is some form of desecration that indelibly sullies the Great Man’s work—is frankly bizarre. It is, however, a kind of thinking that has come to infect both fandom and politics. Six years ago I wrote a blog post about the Hugo awards and the conservative insurgency calling itself the “Sad Puppies,” who were attempting the hijack the nominating process to exclude authors and works they believed to be “message fiction” (“woke” not yet having entered the lexicon at that point), i.e. SF/F more preoccupied with social justice and identity politics than with good old fashioned space opera and fur-clad barbarians. If I may be so gauche as to quote myself, I argued then what I’m arguing now:
… namely, that the introduction of new voices and new perspectives, some of which you find not to your taste, entails the wholesale destruction of what you love, whether it be gaming or SF/F. Anita Sarkeesian produced a handful of video essays critiquing the representation of women in video games. As such critiques go, they were pretty mild—mainly just taking images from a slew of games and letting them speak for themselves. Given the vitriol with which her videos were met, you’d be forgiven for thinking she’d advocated dictatorial censorship of the gaming industry, incarceration of the game creators, and fines to be levied on those who played them. But of course she didn’t—she just suggested that we be aware of the often unsubtle misogyny of many video games, that perhaps this was something that should be curtailed in the future, and further that the gaming industry would do well to produce more games that female gamers—an ever-growing demographic—would find amenable.
This is the same kind of thinking that had people wailing that the all-women reboot of Ghostbusters somehow retroactively ruined their childhoods, or that The Last Jedi was a thermonuclear bomb detonated at the heart of all of Star Wars.2 It’s the logic that leads people to sign a petition to remake the final season of Game of Thrones with different writers, as if such a thing were even remotely feasible.
By contrast, while I loved Peter Jackson’s LotR trilogy, I thought that his adaptation The Hobbit was an unmitigated trio of shit sandwiches—not least because it displayed precisely none of the respect Jackson had previously shown for the source material. And yet, aside from inevitable bitching about how terrible the films were, when I teach The Hobbit in my first-year class this autumn, I do not expect to be hobbled by the trauma of having seen a work I love treated so terribly. The Hobbit will remain The Hobbit—and nothing Peter Jackson can do on screen or an overeager grad student can do at a conference will change that.
Conservatives love to harp on about the entitled Left and “snowflake” culture, but the current culture war pearl-clutching—when it isn’t merely performative posturing by Republicans looking to gin up their base (though also when it is that)—comprises the very kind of entitlement they ascribe to the people they’re vilifying. When the Sad Puppies try to crowd out authors of colour and queer voices at the Hugos; when Gamergaters attack female gamers; when a Twitter mob drives Leslie Jones off social media in retaliation for her role in the “female” Ghostbusters; when Anita Sarkeesian has to cancel an invited talk due to bomb threats; or when an innocuous conference on diversity in Tolkien—no different from the dozens that have been held on “queering Shakespeare” or masculine fragility in Moby Dick—excites the ire of those declaring that such an abomination somehow desecrates Tolkien’s legacy, what is being articulated is an innate sense of ownership. This is mine, the reactionaries are saying. How dare you trespass on my territory. The zero-sum sensibility is one in which any change, any new voice, any new perspective that is not consonant with the way things have always been, is perceived as a loss. Anything challenging the mythos built up around an idea—be it the acknowledgement of systemic racism in America’s history or misogynist subtext in the Narnia Chronicles—is a betrayal of that mythos.
But we should also talk more specifically about the conference itself, as well as the counter-conference quickly thrown together by “The Society of Tolkien,” a group formed specifically in defiance of the incursion of “wokeness” into Tolkien studies.3
Ideally, I want to put this in perspective, and make clear why those people losing their shit really need to do a deep knee bend and take a fucking breath.
To be completely honest, this entire kerfuffle over “woke” Tolkien might have never showed up on my radar had I not been tagged in a general invitation to join my friend Craig and his partner Dani’s weekly online “office hours.” Craig and Dani are both professors in Philadelphia; they host discussions about a variety of topics, mostly dealing with American conservatism and anti-racism, and they focused this particular discussion on the backlash against the Tolkien and Diversity conference. Among those joining were members of the Tolkien Society, including some people presenting at the upcoming conference.
It was a lovely discussion, and since then I’ve been inducted into the Alliance of Arda—which is to say, I requested to join the Facebook group, and been accepted. There has been, as you could well imagine, a lot of discussion about the “Tolkien and Diversity” conference, particularly in regards to the backlash against it. This blog post came together in part because of my participation in those discussions.
And before I go further, I just need to make something clear: these lovely people who are enthusiastically participating in this conference are not resentful SJWs looking to take Tolkien down a peg or five. They are not haters. They do not want to cancel Tolkien. These are people who love Tolkien and his works. They would be as distressed as their detractors if Tolkien went out of print or was eliminated from curricula.
But they are looking to expand the ways in which we read and understand Tolkien, and in a zero-sum perspective … well, see above.
The original call for papers (CFP) for the conference is prefaced as follows:
While interest in the topic of diversity has steadily grown within Tolkien research, it is now receiving more critical attention than ever before. Spurred by recent interpretations of Tolkien’s creations and the cast list of the upcoming Amazon show The Lord of the Rings, it is crucial we discuss the theme of diversity in relation to Tolkien. How do adaptations of Tolkien’s works (from film and art to music) open a discourse on diversity within Tolkien’s works and his place within modern society? Beyond his secondary-world, diversity further encompasses Tolkien’s readership and how his texts exist within the primary world. Who is reading Tolkien? How is he understood around the globe? How may these new readings enrich current perspectives on Tolkien?
For those unfamiliar with academic conferences, this is pretty standard stuff, and is followed by a non-exhaustive list of possible topics participants might pursue:
Representation in Tolkien’s works (race, gender, sexuality, disability, class, religion, age etc.)
Tolkien’s approach to colonialism and post-colonialism
Adaptations of Tolkien’s works
Diversity and representation in Tolkien academia and readership
Identity within Tolkien’s works
Alterity in Tolkien’s works
Again, this is pretty standard fare for a CFP. Nor is it at all outlandish in terms of the topics being suggested. Speaking as both a professional academic and someone who first read Tolkien thirty-seven years ago, I can think of a half-dozen ways into each of those suggestions without really breaking a sweat.
But of course, the CFP was only the amuse bouche for the backlash. The real meal was made of the paper titles. Here’s a sample:
“Gondor in Transition: A Brief Introduction to Transgender Realities in The Lord of the Rings”
“The Problem of Pain: Portraying Physical Disability in the Fantasy of J. R. R. Tolkien”
“The Problematic Perimeters of Elrond Half-elven and Ronald English-Catholic”
“Pardoning Saruman?: The Queer in Tolkien’s The Lord of the Rings”
“Queer Atheists, Agnostics, and Animists, Oh, My!”
“Stars Less Strange: An Analysis of Fanfiction and Representation within the Tolkien Fan Community”
None of these seem at all beyond the pale to me, but then again I’m a professor and have presented numerous conference papers at numerous conferences, and these kind of titles are pretty standard. I can see where a Lord of the Rings enthusiast who never subjected themselves to grad school might think these titles absurd. True fact: academic paper titles are easily mocked, no matter what discipline, as they always employ a specialized language that tends to be opaque to those not versed in it.4
Case in point: I’m reasonably convinced that when former Globe and Mail contrarian and plagiarist Margaret Wente was at a loss for what aspect of liberal thought she would cherry-pick to vilify, she would comb the list of projects granted fellowships or scholarships by the Social Sciences and Humanities Research Council of Canada (SSHRC), and select the most arcane and convoluted project titles receiving taxpayer dollars and write one of her “And THIS is the state of academia today!” columns.
Emblematic of the Wente treatment was an article in the National Review by Bradley J. Birzer, a history professor and author of J. R. R. Tolkien’s Sanctifying Myth (2002). Writing about the upcoming Tolkien Society conference, he says,
While I have yet to read the papers and know only the titles for reference—some of which are so obscure and obtuse that I remain in a state of some confusion let’s, for a moment, consider “Pardoning Saruman? The Queer in Tolkien’s The Lord of the Rings.” In what way is Saruman, an incarnate Maia angel, sent by the Valar to do good in Middle-earth (Saruman really fails at this), queer? Is he in love with himself? True, with his immense ego, he might very well have been. Is he in love with Orthanc? Perhaps, but there is nothing in the text to support this. Is he in love with Radagast the Brown? No, he considers him a fool. Is he in love with Gandalf the Grey? No, he’s jealous of Gandalf and had been from their first arrival in Middle-earth. Is he in love with his bred Orcs? Wow, this would be twisted. Is he in love with Wormtongue? If so, nothing will come of it, for the lecherous Wormtongue has a leering and creepy gaze only for Eowyn. And, so, I remain baffled by all of this. Nothing about a queer Saruman seems to make sense.
Let’s begin with his admission that he hasn’t read any of the conference papers—which is fair, considering that, as I write this post, the conference is several days away—which means that if any of the presenters are at all like me, they’re just starting to write the papers now. But … do you really want to dive in so deeply based solely on a title? There was a story frequently told of a professor I knew at Western University, my alma mater: in answer to a question he’d asked, a student raised his hand, and said, “Well, I haven’t read the book yet, but …” and then went on confidently for several minutes while the professor nodded along. When the student stopped talking, the professor paused, and said, “Yeah. Read the book.”
Birzen’s choose-your-own-adventure speculation on what the paper might be about is no doubt entertaining to his readers who think, like him, that the whole concept of a queer Saruman is absurd, full stop. But here’s the thing: you don’t know. You don’t know what the paper is going to argue. You don’t even know if, despite the fact that he’s the only named character in the title, the paper has anything to say about Saruman’s queerness. Leaving aside for a moment the fact that considering someone a fool or being jealous of them doesn’t obviate the possibility of sexual desire for them, or the idea that demi-gods are somehow asexual (the Greeks would have something to say about that), perhaps consider that the paper isn’t about specific characters being queer? The pre-colon title, after all, is “Pardoning Saruman.” If you handed me this title in some sort of conference-paper-writing reality TV show, in which I have a few hours to write an essay conforming to the title, my approach would actually be to talk about the mercy both Gandalf and Frodo show Saruman, as something subverting a masculinist approach to punitive practices. Queerness, after all, isn’t merely about sex, but about a constellation of world-views at odds with heteronormativity; that Birzen reduces it to a question of who Saruman wants to fuck reflects the impoverishment of his argument.
But who knows. Will the paper be good? Perhaps, though that actually doesn’t matter much. Academic conferences are not about presenting perfect, ironclad arguments; the best papers are imperfect and lacking, because those are the ones that excite the most discussion. Conference papers tend to be works in progress, embryonic ideas that the presenters throw out into the world to test their viability. At its best, academia is an ongoing argument, always trying out new and weird ideas. Some fall flat; some flourish.
At its worst, academia becomes staid and resistant to change. Which brings me to the “Society of Tolkien” and its counter-conference. Their CFP reads as follows: “When J.R.R. Tolkien created Middle-earth, he filled it with characters, themes, and dangers that leapt from the pages to intrigue, excite, and give hope to his readers. In these sessions, we’ll explore these concepts to celebrate all that makes his works stand the test of time and what we should take from them today.” Or, to translate: no theme at all. Suggested topics include:
Analysis of characters, situations, and linguistics in the books
Military doctrine and tactics portrayed in the books or movies
Themes, lessons, and allegories drawn from or used by Tolkien
Works influenced by Tolkien’s writing
Works which influenced Tolkien’s writing
So, in other words: stuff we’ve seen before numerous times. But what is slightly more interesting is the supplemental list of topics that will not be considered for the conference:
Concepts not included in Tolkien’s writing
The Black Speech of Mordor
Speaking as someone who has attended numerous academic conferences, I can attest to the fact that “general foolishness” is always welcome to break up the monotony. As to “concepts not included in Tolkien’s writing,” that’s a pretty difficult territory to police. As one of my new Tolkien friends points out, the word “queer” appears in both The Hobbit and LotR multiple times—not, perhaps, meant in the way we mean it today, but still—where precisely does one draw the line?
I’m particularly intrigued by the prohibition against “The Black Speech of Mordor.” Because … why? Are they concerned, as were the attendees at the Council of Elrond when Gandalf spoke the rhyme on the One Ring in the language of Mordor, that the very sound of the language is corrupting? Do the people of the “Society of Tolkien” want to prohibit any presentation that might end up being sympathetic to Sauron? Or are they, as I joked in Arda’s comment section, worried that a paper about “Black Speech” could be a Trojan horse sneaking in critical race theory to the conference?
I said it as a joke, but I’m by no means convinced that that was not the reasoning behind the prohibition.
When I was enrolling for my fourth year of undergraduate study, I saw a course in the calendar on Tolkien and C.S. Lewis. I signed up unhesitatingly: it was, as I joked at the time, the one course at the university where I could probably walk into the final exam without ever having been to class and still get an A. I did however have a little bit of anxiety: I had first read LotR when I was twelve, and reread it several times through my teenage years. I did not reread it for the first three and a half years of my undergrad, so inundated was I by all the new texts and authors and ideas that saturated my days. I was concerned, on returning to this novel that had taught me so profoundly how literature can have affect—how it can transform you on a seemingly molecular level—that I might now find LotR deficient or less than it was. Since starting my English degree, I’d had that transformative experience again numerous times: Annie Dillard and Toni Morrison, Gabriel García Márquez and George Eliot, Yeats and Eliot and Adrienne Rich, Salman Rushdie and Isabel Allende, Tom Stoppard and Bertholt Brecht and Ibsen, Woolf and Joyce and Faulkner, to say nothing of a raft of literary theory and criticism from Plato and Aristotle to Derrida and De Man. How would Tolkien measure up? And how, having taken classes on postcolonial literature and women’s literature, would I react to Tolkien’s more problematic depictions of race and gender?
Here’s the thing: I needn’t have worried. Yes, all the problematic stuff was there, but so was the deeper complexity of Tolkien’s world-building, the nuanced friendship of Sam and Frodo, the insoluble problem of Smeagol/Gollum, as well as Tolkien’s bravura descriptive prose that I appreciated in a way I could not when I was a callow teenager. There was also my greater appreciation of Tolkien the person. My professor posed a question to the class: why do you think Tolkien—scholar and translator of Norse and Teutonic sagas featuring such indomitable warriors as Beowulf—why did he, when he wrote his great epic, make diminutive halflings his heroes? Heroes who, she continued, proved doughtier than such Teutonic-adjacent heroes as Boromir?
Could it be, she suggested, that he was annoyed by the deployment of Norse and Teutonic myth by a certain 1930s dictator to bolster the fiction of a master race?
She then read to us Tolkien’s response to a Berlin publishing house interested in publishing a German translation of The Hobbit, but wanted confirmation of Tolkien’s Aryan bona fides before proceeding. Tolkien’s response remains one of the most professorial go fuck yourselves I’ve ever encountered:
Thank you for your letter. I regret that I am not clear as to what you intend by arisch. I am not of Aryan extraction: that is Indo-Iranian; as far as I am aware none of my ancestors spoke Hindustani, Persian, Gypsy, or any related dialects. But if I am to understand that you are enquiring whether I am of Jewish origin, I can only reply that I regret that I appear to have no ancestors of that gifted people. My great-great-grandfather came to England in the eighteenth century from Germany: the main part of my descent is therefore purely English, and I am an English subject — which should be sufficient. I have been accustomed, nonetheless, to regard my German name with pride, and continued to do so throughout the period of the late regrettable war, in which I served in the English army. I cannot, however, forbear to comment that if impertinent and irrelevant inquiries of this sort are to become the rule in matters of literature, then the time is not far distant when a German name will no longer be a source of pride.
Your enquiry is doubtless made in order to comply with the laws of your own country, but that this should be held to apply to the subjects of another state would be improper, even if it had (as it has not) any bearing whatsoever on the merits of my work or its sustainability for publication, of which you appear to have satisfied yourselves without reference to my Abstammung [ancestry].
All of which is by way of saying: Tolkien was a product of his nation and his era. He was a complex person, and would almost certainly blanch at the Tolkien Society’s current roster of conference papers.
But then, he’d probably blanch at many of the ones in years past too.
1. Anyone even halfway familiar with J.R.R. Tolkien would tell you that he most likely started spinning in his grave the moment he was interred and hasn’t paused much since. Some authors are generally indifferent to people’s interpretations of their work and are happy to let fans, academics, and adaptors have their own ideas; others retain a proprietary sense of their work, and will respond to negative reviews or interpretations they think miss the mark of what had been intended, as well as being antagonistic to the idea of someone else adapting their work to stage or screen. Tolkien was quite firmly in the latter category, often engaging in lengthy correspondence with fans in which he corrects their understanding of The Lord of the Rings with his own, especially with regards to allegorical or symbolic readings (with few exceptions, Tolkien adamantly maintained that there was no symbolism or allegory in LotR, and no correspondence to contemporary world-historical events). He was also quite bearish on the idea of adapting The Lord of the Rings to film, even though there was an American project in the works in 1958; in his letters, Tolkien communicates his antipathy quite clearly. One can only imagine what he would have thought of the allusions to Middle-Earth in the music of Rush and Led Zepplin, or Ralph Bakshi’s vaguely hallucinogenic animated film (much beloved by the guardians of “classic” SF/F). Whether he would have appreciated Peter Jackson’s LotR trilogy is less certain, though the dilatory and overstuffed debacle of The Hobbit was surely responsible for seismic events in the vicinity of Tolkien’s grave. Compared to Radagast the Brown’s rabbit sled, King Thranduil’s war moose, Legolas constantly defying the laws of physics, and the dwarves’ battle with Smaug inside the Lonely Mountain (I could go on), a paper titled “’Something Mighty Queer’: Destabilizing Cishetero Amatonormativity in the Works of Tolkien” is small beer.
2. That there is so much hate levied against The Last Jedi, while Revenge of the Sith apparently gets a pass, baffles me.
3. As one commenter observed, the fact that some people split off to form “The Society of Tolkien” is such pure Monty Python that it truly puts the absurdity of all this into context.
4. And it is by no means limited to academia. On the day I wrote this post, I also drove my partner Stephanie, who is a talented guitarist, out to Reed Music in Mount Pearl so she could purchase a Fender Telecaster. Unlike Steph, I am utterly unmusical. As we drove, I idly asked her what the difference was between a Telecaster and a Stratocaster. She then proceeded, very animatedly, to descend into guitarist-speak for the next ten minutes. After a certain point, she looked at me and said, “You didn’t understand any of that, did you?” To which I replied, “No, about ten seconds in, you started to sound like the adults in Charlie Brown specials.” But as I write this, she’s in the next room playing “Sultans of Swing” on her new guitar, and, really, that’s a language I speak just fine.
And we’re back! To reiterate my caveat from my previous post, these discussions of postmodernism are filtered through my lens as an Americanist. Postmodernism is a global phenomenon, and while it is conditioned by American cultural and economic hegemony, its manifestations are various and culturally specific around the world. A significant element of the post-WWII world was the collapse of empires and the fraught process of decolonization, as well as the rise of the neo-imperialism of corporate capitalism. What we’ve come to call postcolonialism isn’t synonymous with postmodernism by any means, but on that Venn diagram there is more than a little overlap.
I reiterate this caveat in part because this post is probably the one in this series most specific to U.S. history. I have thus far offered a broad introduction to postmodernism and offered my specific understanding of what I consider its most basic unifying element; what I want to do here in this post is provide some historical context for the material circumstances that gave rise to the postmodern condition.
Are you sitting comfortably? Then we’ll begin.
Whenever I teach my second-year class on American Literature After 1945, I always begin the semester with a lecture in which I contextualize, historically, where the United States was at the end of the Second World War. And I begin with a question: where do you think the U.S. ranked globally in terms of the size and strength of its military in 1939? My students, to their credit, are always wary at this question—they know I wouldn’t ask it if the answer was obvious. So, obviously not #1 … but they’ve all grown up in a world (as indeed I have) in which American military might is vast, and the Pentagon’s budget is larger, at last accounting, than the next fourteen countries (twelve of which are allies) combined. So some brave soul will usually hazard a guess at … #5? No, I reply. Someone more audacious might suggest #10. But again, no. In 1939, I tell them, the U.S. ranked #19 globally in its military’s size and strength, with an army that was smaller than Portugal’s.
The point of this exercise, as you’ve probably gleaned, is to describe how the U.S. went from being a middling world power to a superpower in the course of six years—and how that fundamentally changed U.S. society and culture, and in the process reshaped the world.
There’s an episode of The West Wing from season three that centers on President Bartlett’s State of the Union Address; at once point speechwriter Sam Seaborn (Rob Lowe) talks about how a president’s words can inspire a nation to great achievements, citing FDR’s SOTU in 1940 in which he predicted that the country would build 50,000 aircraft by the end of the war—which proved incorrect, as the U.S. actually built over 100,000.
While Roosevelt’s shrewd leadership of the U.S. through WWII should not be discounted, Sam’s historical factoid has more to do with propping up the Aaron Sorkin Great Man Theory of History: Oratorical Edition™. But this historical actuality does offer a useful thumbnail sketch of the sheer size and scope of the United States’ industrial capacity when totally mobilized. Consider the haunting images of planes and ships mothballed after the war, all of which had been churned out by American factories and which were now the detritus of a massive war effort.
The U.S. emerged from the devastation of the war with its industrial infrastructure intact, unlike those of its allies and enemies: Britain’s had been badly damaged, France’s suffered from the fierce fighting after the Normandy landings, and both Germany and Japan had basically been pounded flat by Allied bombing; the Soviet Union had had to move its factories far enough east to be out of the range of the Luftwaffe, while also suffering 10 million military and 14 million civilian deaths, more than any other combatant nation by a magnitude (the U.S. by contrast lost 418,500, only 1700 of which were civilians). [EDIT: as an historian friend of mine pointed out, China actually suffered losses comparable to Russia’s between 1937-1945]
All of which meant that the United States had gone from being ranked nineteenth in military strength to number one with a bullet (pun intended), with the world’s only fully functioning industrial base of any note; meanwhile, the returning soldiers benefited from the G.I. Bill passed in 1944, which, among other things, provided for college tuition. The discharged veterans poured into colleges and vocational schools and emerged with degrees that put them in good stead to swell the ranks of white-collar workers and take advantage of other provisions of the G.I. Bill, such as one year of unemployment compensation, low-interest loans to start a business or a farm, and low-cost mortgages. This government investment led to the greatest expansion of middle-class prosperity in history.1
The industry that had been cranked up to eleven in the war did not sit idle; as Americans earned more, they spent more. The Big Three automakers switched from jeeps and tanks to cars for the newly suburban nuclear families. Factories changed over from military production2 to making refrigerators, washing machines, and a host of new appliances that, the rapidly expanding business of commercial advertising declared, all households needed. And somewhere in all of that, television made its presence known as sales of television sets grew exponentially: 7,000 in 1945, 172,000 in 1948, and 5 million in 1950; in 1950, twenty percent of households had a television set; by the end of the decade, ninety percent did.
Meanwhile, the wartime acceleration of technology facilitated the rapid growth of, among other things, air travel: the first commercially successful jetliner, the Boeing 707, took to the skies in 1954.
But, you might ask, what does all this have to do with postmodernity? Well, everything … what I’ve been describing are a series of paradigm shifts in travel and communication technology, as well as the United States’ transition from industrial capitalism to full-bore consumerism. These changes were not just material, but symbolic as well, as the U.S. created of them a counter-narrative to Communism. Consumerism became understood as patriotic: the wealthier Americans grew, the more cars and appliances they bought, the more the American Dream could be held up as the obvious virtuous alternative to dour Soviet unfreedom.
This postwar period that comprises the 1950s and early 60s was the era, in the words of scholar Alan Nadel, of “containment”—a time that was marked, to oversimplify somewhat, by “the general acceptance … of a relatively small set of narratives by a relatively large portion of the population.3 Or to put it even more simply, it was a time of pervasive public trust in American institutions, especially government and the military. Now, to be clear, “pervasive” doesn’t mean “total”—this was also, after all, the time of McCarthyism and the Red Hunts, as well as the first glimmers of cultural dissent that would influence the counterculture of the 1960s (most especially embodied in the Beat Generation), and also the snowballing insurgency of the Civil Rights Movement. But taken overall, the 1950s—for all its nascent and repressed anxieties—embodied a complacency facilitated by prosperity and the higher standard of living prosperity made possible. As Nadel argues in his book Containment Culture, this delimiting of narratives was about containment: containing women in their domestic spaces, containing men in the role of patriarchal breadwinner, containing fathers, mothers, and children in the bubble of the nuclear family, containing Black Americans within the confines of Jim Crow, containing ideological expression within apolitical art forms like abstract expressionism and critical methodologies like the “New Criticism” being practised in English departments across the country, and above all containing the Soviet cultural threat with the power of America’s example and its military threat with the example of America’s power.
It didn’t take long for the 1950s to be nostalgized and mythologized in popular culture: American Graffiti was released in 1973, and Happy Days first aired in 1974. Indeed, the 50s were mythologized in popular culture at the time, with shows like Leave it to Beaver and Father Knows Best embodying the utopian ideals of domesticity and the nuclear family. Even popular stories of youthful rebellion were stripped of any possible political content, perhaps most explicitly in the very title of Rebel Without a Cause (1955), or in Marlon Brando’s iconic line in The Wild One (1953): when asked what he’s rebelling against, he replies “Whaddya got?”
When Donald Trump made “Make America Great Again” his slogan, he was tacitly citing the 1950s as the apogee of America’s greatness.4 This was possibly the canniest of his various simplistic catch phrases, given that 1950s nostalgia has reliably proven attractive at such moments of cultural crisis or ennui as the early 1970s; but like all nostalgic figurations, it leaves a lot out. There were those during the 1950s who labelled the period the “age of anxiety,” though as I’ll delve into more deeply in my next post, it could also have been called the age of avoidance: the rhetoric and discourse of containment sought, among other things, to deflect from such new global realities as the threat of nuclear conflict and the haunting aftermath of the Holocaust.
One way to understanding the emergence of postmodernism—and this is indeed one of the primary arguments of Alan Nadel’s book—is as a product of the breakdown of containment, of the fracturing of a societal consensus in which a relatively large number of people accepted a relatively small number of narratives into a more chaotic and fragmented social order that was deeply suspicious of such narratives. Though this breakdown progressed over a series of shocks to the system—the assassination of JFK, the start of the Vietnam War and the concomitant rise of the anti-war Left, the growth and visibility of the Civil Rights Movement culminating in the traumatic murder of Martin Luther King, Jr., the assassination of Robert Kennedy, all of which unfolded on television screens across the nation—this fracture was seeded in the immediate aftermath of the Second World War. The war spawned a world that was more truly global than at any point in history. The world was simultaneously smaller and larger: smaller because of the instantaneous nature of televisual and telephonic communication, as well as the fact that one could travel around the world in a matter of hours rather than weeks or months; larger because of the ever-increasing volume of information available through new communications technology.
The increasingly efficient and indeed instantaneous transmission of information facilitated the more efficient transmission of capital, which in turn facilitated the growth of transnational corporate capitalism, which by the time we reach the orgy of financial deregulation also known as the Reagan Administration becomes less about manufacturing and industry than the arcane calculus of stock markets. By this time, America’s postwar preeminence as the sole nation with functioning industrial infrastructure had been steadily eclipsed as the rest of the world rebuilt, or, as with the case of China and Southeast Asia and the Indian subcontinent, had begun to establish their own industrial bases—facilitating the departure of blue-collar jobs as multinational corporations offshored their factories to whatever countries offered the lowest wages and best tax breaks.
This confluence of historical, economic, technological, and cultural circumstances created a profound sense of dislocation, a disruption of what Marxist theorist and literary scholar Fredric Jameson calls our “cognitive map.” Jameson is the author of Postmodernism, or The Cultural Logic of Late Capitalism (1991), arguably one of the definitive works theorizing postmodernism (as well as being exhibit A of why “postmodern neo-Marxism” is an incoherent concept, but more on that in a future post). He borrows the concept of cognitive mapping from urban theorist Kevin Lynch’s 1960 book The Image of the City; in Lynch’s telling, cities are comforting or alienating to the extent to which one can have a “cognitive map” of the urban space in which one can picture oneself in relation to the city. Cities with a straightforward layout and visible landmarks (such as Toronto) are comforting; cities that are confusing and repetitive sprawls (such as Los Angeles) are alienating.5 Jameson adapts this concept to culture more broadly: how we see ourselves in relation to what Jameson calls the “totality” of our culture and society is predicated on how much of it we can know. Someone living in a small town with no need to have intercourse with the larger society has a fairly straightforward cognitive map. Someone whose life and livelihood depends on circumstances well outside their control or understanding fares less well when they are negatively impacted by the offshoring of jobs or a financial meltdown that erases their retirement savings. By the same token, the individual in the small town’s ability to maintain their cognitive map will be impacted by financial downturns, or even just by having cable TV and the internet.
We could well substitute “sense of reality” for “cognitive map.” Where we find ourselves—have indeed found ourselves for several decades now—is in a place where reality feels increasingly contingent and unstable. When I wrote my doctoral dissertation on conspiracy theory and paranoia in American postwar culture, my argument basically fell along these lines—that the dislocations of postmodernity make the certainties of conspiracism attractive. I defended that thesis seventeen years ago (egad); as evidenced by the 9/11 Trutherism, anti-Obama birtherism, and the current lunacy of QAnon, conspiracism has only metastasized since then.
I am, to be clear, leaving a lot of stuff out here. But my overarching point is that what we call “postmodernism” is no one thing. It is, rather, the product of all the stuff I’ve described, and more. When the Jordan Petersons of the world accuse postmodernism of positing that any and all interpretations of something are all equally valid, they’re leveling that charge against nerdy academics influenced by a handful of French theorists and philosophers who are not, in fact, saying any such thing; Peterson et al are, though they don’t seem to know it, railing against a cultural condition that has made certitude functionally impossible, and which promulgates an infinitude of interpretations through no fault of philosophers and theorists and critics doing their level best to find the language to describe this condition.
1. To hearken back to my posts on systemic racism, Black Americans largely found themselves shut out of the G.I. Bill’s benefits, both because of explicit restrictions written in by Southern Democrats, but also because almost all banks simply refused to extend the low-interest loans and mortgages to Black veterans. While white America enjoyed the fruits of the postwar boom, accruing wealth that would be passed onto their children and grandchildren, the exclusion of Black America from the 1950s boom continues to contribute to the stark wealth inequality between whites and non-whites in America today.
2. Which is not to say, of course, that U.S. military production slackened much; even as the nation demobilized, it shifted focus to the new enemy, the U.S.S.R., and poured huge amounts of money into R&D for possible new conflicts. Indeed, it can be argued that the recessions of 1947 and 1950, resultant from the downturn of wartime production, were counteracted by the Truman doctrine in 1947—the principle asserting America’s obligation to contain Soviet expansion globally—and the outbreak of the Korean War in 1950. Both events goosed the stock market with the promise of increased military production. By the end of Dwight D. Eisenhower’s presidency in 1961, he had grown so alarmed by the growth of military production and the power shared by the Pentagon and arms manufacturers that the architect of the D-Day landings said, in his valedictory address to the nation:
This conjunction of an immense military establishment and a large arms industry is new in the American experience. The total influence—economic, political, even spiritual—is felt in every city, every State house, every office of the Federal government. We recognize the imperative need for this development. Yet we must not fail to comprehend its grave implications. Our toil, resources and livelihood are all involved; so is the very structure of our society … we must guard against the acquisition of unwarranted influence, whether sought or unsought, by the military-industrial complex. The potential for the disastrous rise of misplaced power exists and will persist.
3. Alan Nadel, Containment Culture (1995) p. 4.
4. The peculiar genius of the slogan “Make America Great Again” is precisely its ahistorical vagueness—playing upon the intuitive contradiction felt by people whose instinct is to think of America as the greatest nation in the world, but who cannot see evidence for that greatness in their day to day lives. Hence, it must be returned to that state of greatness; and though the slogan’s vagueness allows its adherents to imagine for themselves precisely when America was great, the most standard conservative default setting is the 1950s (allowing for the likelihood of the most nativist part of the MAGA crowd pining nostalgically for the 1850s).And indeed, the touchstones of Trump’s 2016 campaign promises—the return of well-paid factory jobs, higher wages for the working class, the tacit and sometimes explicit exclusion of people of colour, and the return to a masculine America in which women know their place—all hearkened back to the era of containment (eliding, of course, the 90% top marginal tax rate and the fact that working-class jobs could support a family because of the pervasiveness of unions). Trump’s promises were all about containment, the boxing-in and walling-off of Black, queer, and women’s voices, and containing America itself from the influx of immigrants, refugees, and Muslims, all of which envisions a very strictly contained understanding of what comprises “real” America.
5. Posthumous apologies to Kevin Lynch for so egregiously oversimplifying his nuanced discussions of urban space.
On rereading my introductory post in this series, I realized I forgot to include an important caveat about the particularity of my perspective. To wit: I am, academically speaking, an Americanist—which is to say my principal research and teaching focus is on contemporary American literature and culture. To that end, my understanding of everything I will be talking about will be filtered through that particular lens of expertise. Postmodernism is, by definition, a global phenomenon, and as I will discuss in my next post, globalism as we know it is, to a large extent, the product of a world war in which the United States emerged as the hegemon of the Western world.
But however much the U.S. put its stamp on what we call postmodernism and postmodernity, there are many iterations specific to other nations, other cultures, other histories—not least of which is the weird Venn diagram of postmodernism and postcolonialism. Or the equally weird 1980s feedback loop between American and Japanese postmodernism.
But for the sake of keeping things simple, though I’ll be gesturing to other such iterations of the postmodern, I’ll mostly be keeping to my wheelhouse. So please chime in with all the stuff I’m leaving out.
OK, so that being said, let’s recap: my introductory post was basically about the complexity of postmodernism and the fact that there is no one definition, and that there is no unifying thread linking all aspects of postmodernism.
So let me suggest own theory of the unifying thread linking all aspects of postmodernism.
Very simply: the postmodern condition is one in which our tacit understanding is that language does not reflect reality, but that language creates reality.
Before I go further, let me stipulate that the phrase “tacit understanding” is doing the heavy lifting in that sentence: as I’ll elaborate below, there is a broad range of responses to my suggestion, not the least of which is outright hostility. An instinctive response from certain quarters to the idea that language “creates” reality would be to see it as confirmation that postmodern thought is all about absolute relativism and the denial of objective truth. As physicist and notorious anti-postmodernist Alan Sokal said in the 1990s, anyone suggesting that gravity was a social construct was welcome to walk out of his third-floor office window.
Except that neither I nor any “postmodern” thinker of any substance is suggesting that objective reality doesn’t exist, or that it only exists as something conjured from words, any more than the postmodern thinkers Sokal mocked believed we could all float up into the air if we just denied the existence of gravity or deleted the word from the dictionary. I find it’s useful in this discussion to make a distinction between “reality” and “actuality,” in which the latter is the world as it is, and the former is how we make it comprehensible to one another.1 No postmodernist thinker not holding court in a 3am dorm room blue with pot smoke seeks to deny the actuality of the world. We all inhabit bodies that feel pain and experience the sensoria of our immediate environments, and if we choose to step out of a third-floor window, it almost certainly isn’t to prove a point of about social constructionism.
In fact, the word and concept of gravity is as good a starting point as any. The word “gravity” comes to us from the Latin gravitas, which means “seriousness” and was one of the ancient Roman virtues. It also could mean dignity or importance, or the moral rigor required to undertake a task of great importance. When you consult the Oxford English Dictionary, the first three entries are (1) “the quality of being grave”; (2) “Grave, weighty, or serious character or nature; importance, seriousness”; (3) “Weighty dignity; reverend seriousness; serious or solemn conduct or demeanour befitting a ceremony, an office, etc.” Only after wading through those three entries do you get to number four, which breaks down gravity’s meanings “in physical senses,” and only at number five to you arrive at “The attractive force by which all bodies tend to move towards the centre of the earth; the degree of intensity with which a body in any given position is affected by this force, measured by the amount of acceleration produced.”
The mathematical expression of Earth’s gravity is 9.6 m/s2, which is your vertical acceleration should you jump out of a plane or Alan Sokal’s office window. A postmodernist consideration of gravity would not be skeptical of the fact that things fall, but rather would be preoccupied with how we understand it, and how the linguistic signifier “gravity” possesses meanings other than its mathematical rendering—pointing to its origins meaning seriousness and “weighty” as a metaphor referring to matters of importance, to the word’s shared root with “grave” and the semantic overlap there. One might also cite the history of how we’ve come to understand gravity, from Aristotle’s assertion that heavy objects fall because their element is of earth, and thus seek their natural state. Though the apple falling on Isaac Newton’s head is almost certainly apocryphal, it makes for a good story, even if his mathematics did not actually do much more than Aristotle to clarify not just how but also why gravity works. That had to wait for Einstein’s theories of relativity and the understanding of gravity as curvatures in space/time shaped by objects with mass. And even now, physicists still scratch their heads over the relationship between gravity and quantum mechanics.
None of which questions the actuality that objects fall to the earth at 9.6 m/s2. What it does do is get us into language games that highlight the contingency of meaning, and a simple trio of facts that I always lead off with in my first-year classes: all language is descriptive; all language is metaphorical; all language is rhetorical. Which is to say, all language seeks to describe the world, it does so invariably through analogy, and it seeks to persuade. What language creates, as I said above, is a shared reality that at its best sharpens and clarifies our understanding of the actuality we all individually inhabit. Also, it’s fun: one of the great pleasures of reading a gifted poet or prose stylist is seeing the ways in which they can make you think of certain things anew by using language in challenging and novel ways. Heh, “novel” ways—see what I did there? By which I mean even the humble pun has the capacity to highlight the slipperiness of our shared vocabularies. “I don’t think you quite grasp the gravity of your situation” is a pun that has been used, among other places, in Star Trek and Doctor Who to refer to the fact that the seriousness of one’s circumstances specifically relates to the imminent danger of falling from a great height. My favourite line from Back to the Future is when Doc Brown, puzzled by Marty McFly’s constant use of the word “Heavy!” finally demands whether something has gone wrong with gravity in the future. And of course, there’s the old chestnut that there is no gravity—the Earth just sucks.
But, I can hear some people protesting, maundering about the various meanings of gravity isn’t what we’re concerned with—what we’re concerned with is postmodernism’s denial of objective truth! Which is a big deal! And yes, it would be a big deal if that were indeed the case. The problem is that the word “truth” entails some significant gradations between straightforward facts in evidence and the capital-T Truths bound up in abstractions like justice, morality, and good and evil. Your average postmodernist has no quibble with facts in evidence, but takes issue with the notion of transcendent truths—such as a concept of absolute justice, or that evil exists outside of our capacity to characterize it semantically. Where people most commonly get postmodernism wrong is in characterizing it as a denial of actuality. One suggestion that has surfaced in a significant number of think-pieces over the past several years, that Donald Trump operates out of the “postmodern playbook,” insofar as he treats reality as fungible and truth as something subject to his own whim, is also a basic misapprehension.2 Postmodernism—or, more accurately, postmodern thought—isn’t about the denial of objective truth or actuality, but the interrogation of the premises and cultural assumptions on which the conceptions of capital-T objective Truths are based.
To return to my earlier assertion and its load-bearing words: when I say that “the postmodern condition is one in which our tacit understanding is that language does not reflect reality, but that language creates reality,” I’m not necessarily asserting that language actually creates reality. (As it happens, I find this understanding completely persuasive, but that’s just me). I grasp why this idea is anathema to many, many people, especially religiously devout people who are deeply invested in the assumption of transcendent Truths that exist beyond language. Relatedly, there is also the very long idealist tradition in Western thought and philosophy that is predicated on the basic idea that there is an objective, external Truth towards which we strive, with language as our principal vehicle in doing so.3 What I’m arguing is that the postmodern condition is one in which the prospect of language creating reality isn’t necessarily something that presents itself as such, but is rather a felt experience presenting in most cases intuitively as suspicion, fear, or just a general anxiety. It is usually not articulated specifically, except by nerdy academics like myself or in angry rejections of postmodern thought by other nerdy academics.
This is what I mean by “tacit understanding”: something bound up in a broader cultural condition in which the critical mass of information and the critical mass of media through which we access information, a global economic system that is bewildering to literally everybody, technology that far outstrips the average individual’s capacity to understand it—and I could go on, but I’ll refer you back to my previous post’s bullet-points—has created a cultural condition in which the language/reality relationship has become, shall we say, unmoored. This unmooring was not the specific creation of such anti-postmodernists’ bêtes noir as Jacques Derrida and Michel Foucault; it is, rather—as I will be delving into in future posts—the elemental lived experience of the cultural condition of postmodernity.
Or to put it more simply: you might reject with all your being the idea that language creates reality, but you live in a world so thoroughly suffused by consumerism, so fractured by the multiplicity of disparate media platforms, so atomized by digital culture, that your lived reality is one in which any access to what you consider objective truth is rendered at best deeply fraught and at worst impossible. Those who wonder at the lunacy of QAnon really need only understand this basic dimension of the postmodern condition: namely, that when people’s sense of reality becomes unmoored they will often latch onto epistemic systems that give them a sense of order (no matter how batshit that system might be). Conspiracy theorizing and conspiracy-based paranoia of course long predate the postmodern era, but they find particularly fertile ground in a situation where reality itself feels contingent and slippery.
By the same token, “the imagination of disaster”—as Susan Sontag called it in her classic 1965 essay—has become pervasive, either imagining extinction-level events (alien invasion, asteroid headed for Earth, etc.) that are ultimately averted after much destruction, and which re-establish a sense of order; or, increasingly, depicting post-apocalyptic scenarios in which civilization has collapsed and survivors navigate a new sparsely populated world. Both are fantasies of return: in the first case to a society in which we can have faith in our institutions, in the latter to a more elemental existence shorn of the distractions and trivialities of postmodern life. Indeed, I would argue (and have argued) that post-apocalyptic shows like The Walking Dead share DNA with fantasy like Game of Thrones, insofar as both zombie apocalypse and fantasy are inherently nostalgic, imagining as they do a return to premodern, pre-industrial worlds in which “objective reality” is boiled down the immediacy of survival, whether in the face of attacking zombies, or the imperative of destroying the One Ring4 … in other words, something elemental and visceral and not subject to the seeming infinitude of mediations and contingencies of meaning manifest in the postmodern condition. As a number of thinkers have suggested, it is easier to imagine the destruction of contemporary civilization than how it might be fixed. Perhaps most notably, Fredric Jameson wryly observed that “It seems to be easier for us today to imagine the thoroughgoing deterioration of the earth and of nature than the breakdown of late capitalism,” and that “perhaps that is due to some weakness in our imaginations.”5
Perhaps it is a weakness of our imaginations, but one that, according to Jameson’s voluminous writings on postmodernism, is entirely understandable. In what is perhaps the definitive study of postmodernism, his 1991 book Postmodernism, or The Cultural Logic of Late Capitalism, Jameson argues that one of postmodernism’s key features is its sheer inescapability: the “prodigious new expansion of multinational capitalism”—which is more or less synonymous with the postmodern condition—“ends up penetrating and colonizing those very precapitalist enclaves (Nature and the Unconscious) which offered extraterritorial and Archimedean footholds for critical effectivity.”6 Or, to translate it into non-Jamesonian English, if you can’t get outside out something—physically, mentally, or otherwise—how can you effect a proper critique? How can you address something with which you are always already complicit?7
Keeping in mind that Jameson made that observation about “Archimedean footholds” in a book published in 1991, I think it’s safe to say that we can remark, from our perspective thirty years later, on how the situation he describes has only expanded by a magnitude with the advent of the internet and the current impossibility—short of decamping to live “off the grid” in the wilderness (itself a disappearing enclave)—of extricating oneself from the digital networks that now penetrate all aspects of life.
So how did we get here? Well, that’s my next post. Stay tuned.
1. We find the same distinction in the philosophy of Immanuel Kant, who distinguished between the noumenon, or “the thing in itself” (das Ding an sich) and the phenomenon, or the thing as it appears to an observer. By the same token, psycholinguist Jacques Lacan—who weirdly is entirely ignored by postmodernism’s detractors (who generally choose to train their fire on Derrida and Foucault)—distinguishes between the “Real” and the “Symbolic.” The Real aligns with actuality, our experience of the world; the Symbolic by contrast is the realm of language, and any translation of the Real into language takes it out of the realm of the Real and into the Symbolic. In other words, your personal experience is specific to you, but is ultimately incommunicable as such—to communicate it to others means translating into the realm of the unreal, i.e. of language and our shared vocabularies.
2. The Trumpian world of “alternative facts” is not a product of the postmodern condition, but is more properly associated with authoritarianism—something Hannah Arendt, in passages much-quoted these past few years, asserted in The Origins of Totalitarianism: “The ideal subject of totalitarian rule is not the convinced Nazi or the convinced Communist,” she writes, “but people for whom the distinction between fact and fiction (i.e., the reality of experience) and the distinction between true and false (i.e., the standards of thought) no longer exist.” She also said that “Before mass leaders seize the power to fit reality to their lies, their propaganda is marked by its extreme contempt for facts as such, for in their opinion fact depends entirely on the power of man who can fabricate it.”
3. As I’ll discuss in a future post, modernist art and literature was largely predicated on the premise that true art was about the accessing of reality—T.S. Eliot called this the “objective correlative,” the idea that the right combination of words and metaphors could conceivably access the fundamental truth of a given emotion. But for the modernists, “objective reality” was vanishingly difficult to touch; its principal revolt was against nineteenth century and Victorian positivism, and its assumption that objective reality could be rendered unproblematically through the practices of realism.
4. There’s an article or possibly an entire book to be written about how the popularity of The Lord of the Rings in America was in part a function of antipathy to the early stirrings of postmodernity—especially considering that Tolkien was especially popular on college campuses in the 1960s, which, when you think about it, is more than a little counter-intuitive, that a deeply conservative, essentially Catholic story would take root amidst leftist radicalism. Even those people amenable to the rise of the New Left and the newly translated writings of Derrida and Foucault et al were inclined to wear buttons declaring “Frodo Lives!”, which suggests that postmodernist theory might have fascinated people, but the lived reality of postmodernity still inspired imaginative escape to worlds not ruled by moral ambiguity and contingent realities.
5. Fredric Jameson, The Cultural Turn: Selected Writings on the Postmodern, 1983-1998. p. 50.
6. Fredric Jameson, Postmodernism, or The Cultural Logic of Late Capitalism. p. 49.
7. Because I don’t want these blog posts to turn into novellas, I will just quickly here note in passing that, within the arguments between those theorizing postmodernism, is this very question. More doctrinaire Marxists like Jameson or scions of the Frankfurt School like Theodor Adorno (more on the Frankfurt School and “cultural Marxism” in a future post) assert that the postmodern condition obviates the possibility of substantive cultural critique. Other critics and theorists see fifth columnists: Linda Hutcheon argues at length that postmodernist art and literature weaponizes irony and parody in what she terms “complicitous critique,” while such thinkers of the Birmingham School as Dick Hebdige and Stuart Hall—who effectively invented what we now call “cultural studies—argued that even within the worst excesses of a consumerist culture industry, artists carve out their own enclaves of resistance.
Before I get under way: why “remembering” postmodernism? Several reasons: back when I was an undergraduate, I came across a book about Canadian art titled Remembering Postmodernism. It struck me on two points: one, that the afterword was written by Linda Hutcheon, whose books and articles on postmodernist fiction had been my gateway drug into the topic; and two, that the title was pleasantly cheeky, considering that as far as I was concerned at the time, postmodernism was an ongoing thing.
Now, I never actually read the book, mainly because postmodernist art was—however interesting I found it—somewhat outside my wheelhouse. But the title stuck with me. Two years ago I taught a fourth-year course on American postmodernist fiction, which I titled “Remembering Postmodernism.” In that instance, the expression was a straightforward acknowledgement of postmodernism as an historical period: we started with Thomas Pynchon’s 1966 novel The Crying of Lot 49 and ended with Monique Truong’s The Book of Salt (2003), and a large part of our discussion of Truong’s novel was whether it was, in fact, postmodernist, or whether it represented whatever the next (as yet unnamed) historical phase was.
Now I’m doing these posts and contemplating whether it’s worthwhile to write a book called The Idiot’s Guide to Postmodernism, for the simple fact that there’s an awful lot of talk about postmodernism and postmodernists these days, and none of it seems to have the slightest clue about it. It is rather a term of disapprobation that seems designed on one hand to vilify contemporary university humanities programs as irredeemably “woke,” and as a shorthand to encompass the varying iterations of “wokeness” on the other.
Hence, as far as I’m concerned, we need to remember what postmodernism was and think about what it might still be. And remembering is a useful term insofar as it means both the calling to mind of things forgotten, as well as the act of re-assemblage.
If I were to ask you what the opposite of “remember” is, you’d most likely say “forget,” for the simple reason that that is entirely correct. But in a more strictly linguistic and semantic way, the opposite of “remember” is “dismember.” Sometimes when we remember something, it’s a simple matter of that something simply springing into our mind—where we left our car keys, for instance, or the fact that today is someone’s birthday. The more laborious process of remembering, however, is one of re-membering, of finding those sundered scraps of the past and putting them together like Isis reassembling Osiris’ dismembered body.
And lest you think this post is just an excuse for me to nerd out over semantics (which, to be fair, it totally is—more of that to come), the fallibilities of memory and their relationship to how we conceive of history are quite germane to postmodern thought.
So, I want to do a deep dive on postmodernism over a series of posts. And to be clear, by “deep dive” I actually mean more of a SparkNotes-type run-through of its history, its basic premises, and what it means in the grander scheme of things. I’ve been studying and writing on postmodernism for the better part of my academic career, starting with my undergrad years, and there are libraries-worth of scholarship on the subject. So I’m hardly going to do much more in a handful of blog posts than offer the general contours.
Why, then, do I want to even bother? Two reasons: first, as I posted previously, because I want to use my blog this summer to work through my confused thoughts on a variety of issues. And second, because nobody honking off about it in the present moment—outside of those who have actually considered postmodernism within the confines of academe—seems to have any bloody clue what they’re talking about. As a case in point, I recently read an otherwise interesting article in The Bulwark, an online publication by anti-Trump conservative thinkers, on anti-democratic intellectuals of the new Right who opine that the tenets of “classical liberalism” in fact contain the seeds of tyranny. The article in question was a reasoned and persuasive defense of Enlightenment thought and its influence on the United States’ founding fathers. But then there was this paragraph:
Contemporary “political correctness” or “wokeness” comes from Marx and Nietzsche by way of the Postmodernists, not from John Locke or the Founding Fathers. A serious person would feel the need to at least attempt to trace some of that intellectual history and confront the ideological differences.
Yes, a serious person would feel the need to at least attempt to trace some of that intellectual history and confront the ideological differences—which the author of this article obviously did not do when throwing Marx, Nietzsche, and postmodernism into the same bucket and tacitly ascribing a causal line of influence. He is smart enough at least to cite Marx and Nietzsche—which is more than most people invoking the dreaded specter of postmodernism do—and obviously knows enough to understand that there is a relationship between those two thinkers and some aspects of postmodernism. In other words, there is a very general sense in which he isn’t wrong, but any number of specific senses in which everything about that statement betrays a profound ignorance of all three of its subjects—starting with the basic fact that Marxism and postmodernism are, if not strictly speaking antithetical, then extremely antagonistic.1
But what is postmodernism? Well, let’s start with the pervasive misapprehension that it constitutes a sort of absolute relativism—a contradiction in terms, yes, but one that ostensibly obviates the possibility of objective meaning. After all, if everything is relative to everything else, where is the capital-T Truth?
As with the paragraph I quoted above, this understanding gets some things right, but remains a basic misapprehension for the simple reason that there is no singular “postmodernism.” The rejection of absolutes, the fundamental skepticism about grand narratives, and the understanding of power not as something external to ourselves but bound up in the flux of discourse and language, are all key aspects of what I’ll be calling “postmodern thought.” Postmodern thought is what Jordan Peterson is referring to when he asserts that, for postmodernists, all interpretations of anything at all are equally valid. I will have occasion in future posts to explain why this is completely wrong (which doesn’t differentiate it from most of his assertions), but for my first few posts I want to distinguish between the various intellectual strands of thinking that emerged from postmodernism, and postmodernism as the cultural condition that gave rise to postmodern thought.
Because the one thing I want to clarify, even if I don’t manage to make anything else clear in these posts, is that postmodernism—or more specifically postmodernity—is above all other things a cultural condition emerging from a confluence of historical circumstances and contexts. The tacit understanding of postmodernism when deployed as a term of disapprobation is that it is a pernicious relativist mode of thinking that exerts a profound and deleterious influence on, well, everything; in the crudest and most conspiratorial version of this thinking, such contemporary social movements that enrage conservatives like feminism, trans rights, Black Lives Matter (and the current bête noir, “critical race theory”) , and a host of other “woke” causes, can be traced back to a handful of French intellectuals in the late 1960s who invented postmodernism as a means of pursuing Marxism and the destruction of Western civilization by other means.2 And while the loose assemblage of schools of thought I’m calling “postmodern” has undoubtedly shaped contemporary attitudes—sometimes positively, sometimes not—my larger argument is that they have emerged in response to cultural conditions; what postmodernism’s detractors call “postmodernism” has usually been less a tool of cultural change than a series of attempts to find language to describe the dramatic cultural transformations of the post-WWII landscape. The cultural changes in question, from intersectional understandings of identity to the normalization of gay marriage, might have been facilitated in part by aspects of postmodern thought, but have been far more facilitated by the technological and economic transformations of postmodernity.
Which, once again, brings us back to the basic question of just what postmodernism is. As I say above, it is a lot of things, including but not limited to:
The cultural logic of late capitalism.
The breakdown of faith in such societal grand narratives as religion, governance, justice, science, etc.
A set of aesthetic practices in art, fiction, film, and architecture (among others) that reflect and articulate such breakdowns.
The rise of multinational corporate capital, and its transformation of the imperialist project from a national, colonial one to the subjugation of national interests to the global market.
The ascendancy of neoliberal free-market fundamentalism.
The snowballing of technology, especially communication technology—from television to the internet to social media—and the concomitant erosion of traditional informational gatekeepers (e.g. legacy media).
The inescapability of consumer culture and the culture industry.
A set of philosophical, theoretical, and critical attempts to adequately describe all of the above.
So … that clears things up, right? Just kidding—of course it doesn’t. But hopefully that starts to communicate the complexity of the subject.
I do want to make clear that this series of posts is not meant as an apologia for postmodernity, postmodernism, or postmodern thought. I have for many years studied postmodernism and written about it and taught it in university classrooms, but I would not call myself a postmodernist (though others, mostly the people who don’t really understand it, certainly would). My own thinking and philosophical inclinations bend toward pragmatism (the philosophical kind) and a sort of small-h humanism; my lodestars in this respect are Richard Rorty and Terry Pratchett. I passionately love some aspects of postmodern culture, mostly its aesthetic incarnations—the fiction of Toni Morrison and Don DeLillo, for example, or films like Blade Runner—and I find the theoretical armature defining postmodernism, whether it’s celebratory like that of Linda Hutcheon or Brian McHale, or damning like that of Fredric Jameson, fascinating; but huge swaths of what I’m calling postmodernity are pernicious and harmful (corporate capitalism and social media among them).
This series of posts is rather a quixotic attempt to re-inject some nuance into what has become an impoverished and overdetermined understanding of a bewilderingly complex concept. More importantly, it is my own way of working through my own thoughts on the matter. So bear with me.
1. There is also the rather amusing thought of just how disgusted Nietzsche would be with “woke” sentiments.
2. More on this in my future post on “The Conspiracy Theory of Postmodernism.”