Posts Tagged ‘History’

Douglas Murray wrote a book called The Strange Death of Europe (2017), which is behind a long list of books as yet unread by me. Doubtful I will ever get to is. My familiarity with his thesis stems from his many appearances in interviews and webcasts describing the book. Murray continues a long line of declinists (mentioned here) prophesying and/or chronicling the long, slow demise of Europe, and in an only slightly wider sense, the West. Analyses of the causes and manifestations of decline range all over the map but typically include a combination of incompetence, exhaustion, and inevitability. For Murray, the strangeness is that it appears self-inflicted and openly desired out of some misplaced sense of guilt and shame following an entirely atypical (for the West) moral reckoning, perhaps for having been the dominant global culture for roughly five hundred years. This blog already says plenty about disaster, destruction, and doom, so let me instead pose a different question: Is the cultural inheritance of the West at all worth honoring and preserving? Won’t be an easy answer.

The question is partly prompted by the breathless announcement last fall in the alumni magazine of one of my alma maters of a newly created position (a further example of needless administrative bloat): Associate Dean of Equity and Inclusion (among other flowery academic titles attached to the position). For reasons unknown, the full sequence Diversity, Equity, and Inclusion (a/k/a DIE — yes, the initialism is purposely disordered) was avoided, but as with many trendy turns of phrase, precision hardly matters since the omitted word is inferred. Let’s first consider the term cultural inheritance. If one takes a warts-and-all approach, the full spectrum from glory to atrocity accumulated throughout history informs what descends from our forebears in the form of inheritance. Indeed, highs and lows are scarcely separable. Whereas individual memory tends to select mostly good parts, cultural memory records everything — albeit with interpretive bias — so long as one has the inclination to consult history books.

If one takes a charitable view of history, admirable events, practices, norms, and artifacts become the celebratory focus while nasty parts may be acknowledged but are nonetheless elided, forgotten, or shoved forcibly down the memory hole. But if one takes a pessimistic view, the very worst parts represent an irredeemable, permanent stain on everything else (the aforementioned moral reckoning). Recent arguments (revisionist history say many) that the founding of the United States of America, a political entity, should be redated to 1619 to coincide with the establishment of slavery rather than 1776 with the Declaration of Independence is an example. Uncharacteristically perhaps, I take the charitable view while the trés chic view in academe (echoed in corporate life and far-left government policy) is to condemn the past for failing to embody Woke standards of the present. Multiple influential segments of Western culture have thus succumbed to ideological possession as demonstrated by self-denunciation, self-flagellation, and complete loss of faith in Western institutions because of, well, slavery, genocide, colonialism, warfare, torture, racism, child labor, human trafficking, and other abominations, almost none not all of which are relegated to the past.

(more…)

Even without being a historian (you or me), it’s easy to recognize seminal figures in U.S. history who have articulated the basic ideology behind what has grown to be a maniacal notion of what a world power can and should be. For instance, not very long after the American Revolution and the purported end of the Colonial Era, President James Monroe established the Monroe Doctrine, claiming the entire Western Hemisphere as being within America’s sphere of influence and warning others across the Atlantic not to intervene. Later in the 19th century, Abraham Lincoln responded to the Southern Secession by launching the American Civil War, establishing that no state could leave the Union. A period of isolationism followed, broken when the U.S. joined WWI (unclear to me why the U.S. fought that war). Woodrow Wilson laid out the principles of liberal internationalism in 1917:

The American military, the president told a joint session of Congress, was a force that could be used to make the world “safe for democracy” … Wilson’s doctrine was informed by two main ideas: first, the Progressive Era fantasy that modern technologies and techniques — especially those borrowed from the social sciences — could enable the rational management of foreign affairs, and second, the notion that “a partnership of democratic nations” was the surest way to establish a “steadfast concert for peace.”

from “Empire Burlesque” by Daniel Bessner (Harper’s Magazine, July 2022)

.
Note that that bit of rhetoric, “safe for democracy,” has been trotted out for over a century now yet shows no sign of losing its mojo. It helps, of course, that no one really knows what democracy is anymore. The public is subjected to relentless narrative spin and propaganda, bread and circuses, and inferior to nonexistent education that muddies the concept beyond recognition. Ten months prior to the U.S. entry into the next world war, influential American magazine publisher (Time, Life, Fortune, Sports Illustrated) Henry Luce added further justification for growing U.S. geopolitical ambitions:

… the Japanese attacked Pearl Harbor, and the United States, which had already been aiding the Allies, officially entered the war. Over the next four years, a broad swath of the foreign policy elite arrived at Luce’s conclusion [from just before the war]: the only way to guarantee the world’s safety was for the United States to dominate it. By the war’s end, Americans had accepted this righteous duty, of becoming, in Luce’s words, “the powerhouse … lifting the life of mankind from the level of the beasts to what the Psalmist called a little lower than the angels.”

from “Empire Burlesque” by Daniel Bessner (Harper’s Magazine, July 2022)

.
There has since been no going back, only solidification and strengthening of what is called The American Century (thanks again to Luce) but really represents the spread of a global empire. So much for the end of colonialism, now pursued primarily through other means but still reverting to overt militarism whenever and wherever necessary. Just like civilizations, empires have come and gone throughout human history with power centers shifting somewhat reliably if unpredictably. The American Empire will undoubtedly join others in the dustbin of history no matter whether anyone survives the 21st century to survey the wreckage. Moreover, the illusion that The American Century can be extended is handily dispelled by the Macrofutilist, noting that corporations are leading the charge into the abyss:

Humans have no agency in this world dominated, at every institution and at every level of those institutions, by corporations and states that function as corporations. Under the rubric of the corporation, every public good or resource is under relentless exploitation, subject only to the fictional “control” by political or legal structures. Bolstered by near-total capture of every ancillary human social event or condition, corporations are wonderfully positioned to lead humanity off its cliff of resource degradation and impending scarcity … The horror is so monumental, so vast in its iniquity, so above any moderation, so all-consuming in its reach, so supreme in its command, that the subject of corporate ownership of the means of species destruction risks becoming boring. Who has the right to speak of “resistance” or “change” or “giving back” when all forms of social control are under obdurate corporate ownership?

from Corporations Are the Perfect Vehicle to Drive Humanity to Its Self-Extinction

.
Although it’s impossible to establish beyond reasonable doubt who’s actually driving the bus — corporations, the military-industrial complex (those two form a tautology by now), elected members of government, the Deep State, or some other nefarious cabal — it’s probably fair to say that members of each group have taken into their hearts the desire for full-spectrum dominance. That term originally meant complete military control of a theater of war. However, as its very name frankly admits, activities of the Select Subcommittee on the Weaponization of the Federal Government signal a new style of Hobbesian war of all against all has begun. Indeed, what I used to call creeping fascism no longer needs the modifier creeping. The end game may have finally arrived, the evidence being everywhere if one has the fortitude to look.

Buzzwords circulating heavily in the public sphere these days include equality, equity, inclusion, representation, diversity, pluralism, multiculturalism, and privilege. How they are defined, understood, and implemented are contentious issues that never seem to resolve. Indeed, looking back on decade after decade of activism undertaken to address various social scourges, limited progress has been made, which amounts to tinkering around the edges. The underlying bigotry (opinion, motivation, activity) has never really been eradicated, though it may be diminished somewhat through attrition. Sexism and racism in particular (classics in an expanding universe of -isms) continue to rage despite surface features taking on salutary aspects. Continued activism use the buzzwords above (and others) as bludgeons to win rhetorical battles — frequently attacks on language itself through neologism, redefinition, and reclamation — without really addressing the stains on our souls that perpetuate problematic thinking (as though anyone ever had a lock on RightThink). Violence (application of actual force, not just mean or emphatic words) is the tool of choice deployed by those convinced their agenda is more important than general societal health and wellbeing. Is violence sometimes appropriate in pursuit of social justice? Yes, in some circumstances, probably so. Is this a call for violent protest? No.

Buzzwords stand in for what we may think we want as a society. But there’s built-in tension between competition and cooperation, or alternatively, individual and society (see the start of this multipart blog post) and all the social units that nest between. Each has its own desires (known in politics by the quaint term special interests), which don’t usually combine to form just societies despite mythology to that effect (e.g., the invisible hand). Rather, competition results in winners and losers even when the playing field is fair, which isn’t often. To address what is sometimes understood (or misunderstood, hard to know) as structural inequity, activists advocate privileging disenfranchised groups. Competence and merit are often sacrificed in the process. Various forms of criminal and sociopathic maneuvering also keep a sizeable portion of the population (the disenfranchised) in a perpetual and unnecessary state of desperation. That’s the class struggle.

So here’s my beef: if privilege (earned or unearned) is categorically bad because it’s been concentrated in a narrow class (who then position themselves to retain and/or grow it), why do activists seek to redistribute privilege by bestowing it on the downtrodden? Isn’t that a recipe for destroying ambition? If the game instead becomes about deploying one or more identifiers of oppression to claim privilege rather than working diligently to achieve a legitimate goal, such as acquiring skill or understanding, why bother to try hard? Shortcuts magically appear and inherent laziness is incentivized. Result: the emergent dynamic flattens valuable, nay necessary, competence hierarchies. In it’s communist formulation, social justice is achieved by making everyone equally precarious and miserable. Socialism fares somewhat better. Ideologues throughout history have wrecked societies (and their members) by redistributing or demolishing privilege forcibly while hypocritically retaining privilege for themselves. Ideology never seems to work out as theorized, though the current state of affairs (radical inequality) is arguably no more just.

More to unpack in further installments.

What if everyone showed up to an event with an expectation of using all their tech and gadgets to facilitate the group objective only to discover that nothing worked? You go to a fireworks display but the fireworks won’t ignite. You go to a concert but the instruments and voices make no sound. You go to a sporting event but none of the athletes can move. Does everyone just go home? Come back another day to try again? Or better yet, you climb into your car to go somewhere but it won’t start. Does everyone just stay home and the event never happens?

Those questions relate to a new “soft” weapons system called Scorpius (no link). The device or devices are said to disrupt and disable enemy tech by issuing a narrowly focused electromagnetic beam. (Gawd, just call it a raygun or phaser. No embarrassment over on-the-nose naming of other things.) Does the beam fry the electronics of its target, like a targeted Carrington event, or just simply scramble the signals, making the tech inoperable? Can tech be hardened against attack, such as being encased in a Faraday cage? Wouldn’t such isolation itself make tech nonfunctional, since electronic communications between locations is the essence of modern devices, especially for targeting and telemetry? These are a few more idle questions (unanswered, since announcements of new weaponry I consulted read like advertising copy) about this latest advance (euphemism alert) in the arms race. Calling a device that can knock a plane (um, warplane) out of the sky (crashing somewhere, obviously) “soft protection” because the mechanism is a beam rather than a missile rather obfuscates the point. Sure, ground-based technologies might be potentially disabled without damage, but would that require continuous beam-based defense?

I recall an old Star Trek episode, the one with the Gorn, where omnipotent aliens disabled all weapons systems of two spaceships postured for battle by superheating controls, making them too hot to handle. Guess no one thought of oven mitts or pencils to push the “Fire!” buttons. (Audiences were meant to think, considering Star Trek was a thinking person’s entertainment, but not too much.) Instead of mass carnage, the two captains were transported to the surface of a nearby planet to battle by proxy (human vs. reptile). In quintessential Star Trek fashion — imagining a hopeful future despite militaristic trappings — the human captain won not only the physical battle but the moral battle (with self) by refusing to dispatch the reptile captain after he/it was disabled. The episode posed interesting questions so long as no one searched in the weeds for plausibility.

We’re confronted now, and again, with many of these same questions, some technical, some strategic, but more importantly, others moral and ethical. Thousands of years of (human) history have already demonstrated the folly of war (but don’t forget profitability). It’s a perennial problem, and from my vantage point, combatants on all sides are no closer to Trekkie moral victory now than in the 1960s. For instance, the U.S. and its allies are responsible for millions of deaths in Iraq, Afghanistan, Syria, and elsewhere just in the last two decades. Go back further in time and imperial designs look more and more like sustained extermination campaigns. But hey, we came to play, and any strategic advantage must be developed and exploited, moral quandaries notwithstanding.

It’s worth pointing out that in the Gorn episode, the captains were deprived of their weapons and resorted to brute force before the human improvised a projectile weapon out of materials handily strewn about, suggesting perhaps that intelligence is the most deadly weapon. Turns out to have been just another arms race.

A quick search revealed that over 15 years of blog posts, the word macrohistory has been used only once. On reflection, macrohistory is something in which I’ve been involved for some time — mostly as a dilettante. Several book reviews and three book-blogging series (one complete, two either on hiatus or fully abandoned) concern macrohistory, and my own several multi-part blogs connect disparate dots over broader topics (if not quite history in the narrow sense). My ambition, as with macrohistory, is to tease out better (if only slightly) understandings of ourselves (since humans and human culture are obviously the most captivating thing evar). Scientists direct similar fascination to the inner workings of nonhuman systems — or at least larger systems in which humans are embedded. Thus, macrohistory can be distinguished from natural history by their objects of study. Relatedly, World-Systems Theory associated with Immanuel Wallerstein and The Fourth Turning (1997 book by William Strauss and Neil Howe) take similarly broad perspectives and attempt to identify historical dynamics and patterns not readily apparent. Other examples undoubtedly exist.

This is all preliminary to discussing a rather depressing article from the December 2020 issue of Harper’s Magazine: Rana Dasgupta’s disquieting (ahem) essay “The Silenced Majority” (probably behind a paywall). The subtitle poses the question, “Can America still afford democracy?” This innocuous line begs the question whether the U.S. (America and the United States of America [and its initialisms U.S. and U.S.A.] being sloppily equivalent almost everywhere, whereas useful distinctions describe the United Kingdom, Great Britain, and England) actually has or practices democracy anymore, to which many would answer flatly “nope.” The essay is an impressive exercise, short of book length, in macrohistory, though it’s limited to Western cultures, which is often the case with history told from inside the bubble. Indeed, if (as the aphorism goes) history is written/told primarily by the victors, one might expect to hear only of an ongoing series of victories and triumphs with all the setbacks, losses, and discontinuities excised like some censored curated Twitter or Facebook Meta discussion. One might also wonder how that same history reads when told from the perspective of non-Western countries, especially those in transitional regions such as Poland, Ukraine, Turkey, and Iran or those with histories long predating the rise of the West roughly 500 years ago, i.e., China, Japan, Egypt, and the lost cultures of Central America. Resentments of the Islamic world, having been eclipsed by the West, are a case in point. My grasp of world history is insufficient to entertain those perspectives. I note, however, that with globalism, the histories of all regions of the world are now intimately interconnected even while perspectives differ.

Dasgupta describes foundational Enlightenment innovations that animate Western thinking, even though the ideas are often poorly contextualized or understood. To wit:

In the seventeenth century, England was an emerging superpower. Supremacy would come from its invention of a world principle of property. This principle was developed following contact with the Americas, where it became possible to conjure vast new English properties “out of nothing”—in a way that was impracticable, for instance, in the militarized, mercantile societies of India. Such properties were created by a legal definition of ownership designed so that it could be applied only to the invaders. “As much land as a man tills, plants, improves, cultivates, and can use the product of,” John Locke wrote in 1689, “so much is his property.” When combined with other new legal categories such as “the savage” and “the state of nature,” this principle of property engendered societies such as Carolina, where Locke’s patron, the first earl of Shaftesbury, was a lord proprietor.

Obvious, isn’t it, that by imposing the notion of private property on indigenous inhabitants of North America, colonialists established ownership rights over territories where none had previously existed? Many consider that straightforward theft (again, begging the question) or at least fencing the commons. (Attempts to do the same in the open oceans and in space [orbit] will pick up as technology allows, I surmise.) In addition, extension of property ownership to human trafficking, i.e., slavery and its analogues still practiced today, has an exceptionally long history and was imported to the Americas, though the indigenous population proved to be poor candidates for subjugation. Accordingly, others were brought to North America in slave trade that extended across four centuries.

Dasgupta goes on:

From their pitiless opposition to the will of the people, we might imagine that British elites were dogmatic and reactionary. (Period dramas depicting stuck-up aristocrats scandalized by eccentricity and innovation flatter this version of history.) The truth is that they were open-minded radicals. They had no sentimentality about the political order, cutting the head off one king and sending another into exile. They could invent financial and legal structures (such as the Bank of England, founded in 1694) capable of releasing unprecedented market energies. Even their decision to exploit American land with African labor demonstrated their world-bending pursuit of wealth. Their mines and plantations would eventually supply the capital for the first industrial revolution. They loved fashion and technology, they believed in rationality, progress, and transparency. They were the “founding fathers” of our modern world.

And yet they presided over a political system as brutal as it was exclusive. Why? The answer is simple. They could not afford democracy, but also, crucially, they did not need it. [emphasis in original]

So much for the awe and sacred respect in which Enlightenment philosophers and the Founders are held — or used to be. Statues of these dudes (always dudes, natch) are being pulled down all the time. Moreover, association of liberal democracy with the 17th century is a fundamental mistake, though neoliberalism (another poorly defined and understood term) aims to shift backwards to a former or hybrid state of human affairs some are beginning to call digital feudalism.

The article goes on to discuss the balancing act and deals struck over the course of centuries to maintain economic and political control by the ownership class. It wasn’t until the 1930s and the postwar economic boom in the U.S. that democracy as commonly understood took root significantly. The labor movement in particular was instrumental in forcing FDR’s New Deal social programs, even though populism and socialism as political movements had been successfully beaten back. Interestingly, the hallowed American nuclear family (limited in its scope racially), an ahistorical formation that enjoyed a roughly 30-year heyday from 1945 to 1975, coincides with the rise of the American middle class and now-aged democratic institutions. They’re all connected with widely distributed wealth and prosperity. But after the oil crisis and stagflation of the middle 1970s, gains enjoyed by the middle class have steadily eroded and/or been actively beaten back (again!) so that dominant themes today are austerity imposed on the masses and inequality coughing up hundy-billionaires with increasing frequency. Estimates are that 30-40% of the American citizenry lives in poverty, bumping up against failed state territory. Inequality has returned to Gilded Age levels if not exceeded them. Dasgupta fails to cite perhaps the major underlying cause of this shift away from affordable democracy, back toward the brutal, world principal of property: falling EROI. Cheap foreign labor, productivity gains, and creation of a giant debtor society have simply not offset the disappearance of cheap energy.

Dasgupta’s further discussion of an emerging two-tier economy along with the Silicon Valley technocracy follows, but I’ll stop short here and encourage readers instead to investigate and think for themselves. Lots of guides and analyses help to illuminate the macrohistory, though I find the conclusions awful in their import. Dasgupta drives home the prognosis:

The neoliberal revolution aimed to restore the supremacy of capital after its twentieth-century subjugation by nation-states, and it has succeeded to an astonishing degree. As states compete and collude with gargantuan new private powers, a new political world arises. The principle of labor, which dominated the twentieth century—producing the industrious, democratic society we have come to regard, erroneously, as the norm—is once again being supplanted by a principle of property, the implications and consequences of which we know only too well from our history books.

A listicle called “10 Things We Have Learned During the Covid Coup,” supporting text abbreviated ruthlessly:

1. Our political system is hopelessly corrupt …

2. Democracy is a sham. It has been a sham for a very long time …

3. The system will stop at nothing to hold on to its power …

4. So-called radical movements are usually nothing of the sort …

5. Any “dissident” voice you have ever heard of through corporate media is probably a fake …

6. Most people in our society are cowards …

7. The mainstream media is nothing but a propaganda machine for the system …

8. Police are not servants of the public but servants of a powerful and extremely wealthy minority …

9. Scientists cannot be trusted …

10. Progress is a misleading illusion …

Let’s Be Evil, pt. 05

Posted: May 12, 2021 in Culture, History, Outrage, Politics, War
Tags:

Does this miserable joke meme

inform the following image?

Time moves on yet the story remains stubbornly the same. Like the United States before it (and others elsewhere), Israel is carrying out an extermination campaign — with the aid of the U.S. empire. There’s something uniquely despicable about being unrepentant winners in the unabated practice of colonialism.

Returning to the subject of this post, I asserted that the modern era frustrates a deep, human yearning for meaning. As a result, the Medieval Period, and to a lesser degree, life on the highroad, became narrative fixations. Had I time to investigate further, I would read C.S. Lewis’ The Discarded Image (1964), but my reading list is already overfull. Nonetheless, I found an executive summary of how Lewis describes the Medieval approach to history and education:

Medieval historians varied in that some of them were more scientific, but most historians tried to create a “picture of the past.” This “picture” was not necessarily based in fact and was meant more to entertain curiosity than to seriously inform. Educated people in medieval times, however, had a high standard for education composed of The Seven Liberal Arts of grammar, dialectic, rhetoric, arithmetic, music, geometry, and astronomy.

In the last chapter, Lewis summarizes the influence of the Medieval Model. In general, the model was widely accepted, meaning that most people of the time conformed to the same way of thinking. The model, he reiterates, satisfied imagination and curiosity, but was not necessarily accurate or factual, specifically when analyzed by modern thinkers.

Aside. Regular readers of The Spiral Staircase may also recognize how consciousness informs this blog post. Historical psychology offers a glimpse into worldviews of bygone eras, with the Medieval Period perhaps being the easiest to excavate contemplate due to proximity. Few storytellers (cinema or literature) attempt to depict what the world was truly like in the past (best as we can know) but instead resort to an ahistorical modern gloss on how men and women thought and behaved. One notable exception may be the 1986 film The Name of the Rose, which depicts the emerging rational mind in stark conflict with the cloistered Medieval mind. Sword-and-sandal epics set in ancient Rome and Greece get things even worse.

(more…)

Unlike turtles, humans do not have protective shells into which we can withdraw when danger presents. Nor can we lift off, fly away, and elude danger the way birds do. These days, we’re sorely beset by an invisible pandemic spread by exposure to carriers (read: other people) and so asked or forced to submit to being locked down and socially distanced. Thus, we are withdrawn into the protective shell of the home in cycles of varying intensity and obeisance to maintain health and safety. Yet life goes on, and with it, numerous physical requirements (ignoring psychological needs) that can’t be met virtually demand we venture out into the public sphere to gather resources, risking exposure to the scourge. Accordingly, the conduct of business has adapted to enable folks to remain in the protective shells of their vehicles, taking delivery through the car window and rarely if ever entering a brick-and-mortar establishment except in defiance or at the option of acceptable risk. In effect, we’re being driven into our cars ever more, and the vehicle is readily understood as a proxy for its inhabitant(s). Take note of pictures of people in bread lines during the Great Depression having been replaced by pictures of cars lined up for miles during the pandemic to get packaged meals from charitable organizations.

Reflecting on this aspect of modern life, I realized that it’s not exactly novel. The widespread adoption of the individual vehicle in the 1940s and 50s, as distinguished from mass transit, and the construction of the interstate highway system promised (and delivered) flexibility and freedom of tremendous appeal. While the shift into cars (along with air travel) doomed now moribund passenger rail (except intracity in the few American cities with effective rail systems), it enabled the buildout of suburbs and exurbs now recognized as urban sprawl. And like all those packages now clogging delivery systems as we shift even more heavily during the holiday season to online shopping, a loss of efficiency was inevitable. All those individual cars and boxes create congestion that cry out for solutions.

Among the solutions (really a nonsolution) were the first drive-through banks of the 1970s. Is doing one’s banking without leaving the vehicle’s protective shell really an efficiency? Or is it merely an early acknowledgement and enabling of antisocial individualism? Pneumatic tubes that permitted drive-through banking did not speed up transactions appreciably, but the novel mechanism undoubtedly reinforced the psychological attachment Americans felt with their cars. That growing attachment was already apparent in the 1950s, with two bits of Americana from that decade still resonating: the drive-in theater and the drive-in restaurant. The drive-in theater was a low-fidelity efficiency and alternative to the grand movie houses built in the 1920s and 30s seating a few thousand people in one cavernous space. (A different sort of efficiency enabling choice later transformed most cinema establishments into multiplexes able to show 8–10 titles instead of one, handily diminishing audiences of thousands to hundreds or even tens and robbing the group experience of much of its inherent power. Now that premium streaming content is delivered to screens at home and we are disallowed assembly into large audiences, we have instead become something far more inert — viewers — with fully anticipatable degradation of the entertainment experience notwithstanding the handsome technologies found within the comforts of the home.) I’ve heard that drive-ins are experiencing a renaissance of sorts in 2020, with Walmart parking lots converted into showplaces, at least temporarily, to resemble (poorly) group experience and social cohesion connection. The drive-in restaurant of the 1950s, with their iconic carhops (sometimes on roller skates), is a further example of enabling car culture to proliferate. Never mind that eating in the car is actually kinda sad and maybe a little disgusting as odors and refuse collect in that confined space. One might suspect that drive-ins were directed toward teenyboppers and cruisers of the 1950s exploring newfound freedom, mobility, and the illusion of privacy in their cars, parked in neat rows at drive-ins (and Lookout Points for smooch sessions) all across the country. However, my childhood memory was that it was also a family affair.

Inevitably, fast food restaurants followed the banks in the 1970s and quickly established drive-through lanes, reinforcing the degradation of the food experience into mere feeding (often on one’s lonesome) rather than dining in community. Curiously, the pandemic has made every restaurant still operating, even the upscale ones, a drive-through and forced those with and without dedicated drive-through lanes to bring back the anachronistic carhop to serve the congestion. A trip to a local burger joint in Chicago last week revealed 40+ cars in queue and a dozen or so carhops on the exterior directing traffic and making deliveries through the car window (briefly penetrating the protective shell) so that no one would have to enter the building and expose oneself to virus carriers. I’ve yet to see a 2020 carhop wearing roller skates (now roller blades) or a poodle skirt.

Such arrangements are probably effective at minimizing pandemic risk and have become one of several new normals (discussion of political dysfunction deferred). Who can say how long they will persist? Still, it’s strange to observe the psychology of our response, even if only superficially and preliminarily. Car culture has been a curious phenomenon since at least the middle of the 20th century. New dynamics reinforcing our commitment to cars are surprising, perhaps, but a little unsurprising, too, considering how we made ourselves so dependent on them as the foundation of personal transportation infrastructure. As a doomer, I had rather expected that Peak Oil occurring around 2006 or so would spell the gradual (or sudden) end of happy motoring as prices at the pump, refusal to elevate standard fuel efficiency above 50 mph, and climbing average cost of new vehicles placed individual options beyond the reach of average folks. However, I’ve been genuinely surprised by fuel costs sinking to new lows (below the cost of production, even bizarrely inverting to the point that producers paid buyers to take inventory) and continued attempts to engineer (only partially) around the limitations of Peak Oil, if not indeed Peak Energy. I continue to believe these are mirages, like the record-setting bull market of 2020 occurring in the midst of simultaneous economic, social, and health crises.

Medieval Pilgrimage

Posted: November 16, 2020 in Artistry, Culture, Music
Tags: , , ,

Listening to the recording shown at left, my mind drifted to various cinematic treatments of Medievalism, including The Lord of the Rings, Game of Thrones, The Chronicles of Narnia, and too many others to cite. Other associations also came tumbling out of memory, including my review of The Hobbit (the book, not the movie, though I reviewed both) and a previous blog post called “What’s Missing.” That post was a rumination on community and meaning lost in modern technocratic societies. In light of fetishization of the Medieval Period, including for example the popularity of Renaissance Faires, there seems to be more to say about what’s missing.

The Llibre Vermell de Montserrat (English: Red Book of Montserrat), known as such because of its 19th-century binding and its being held the Monastery of Montserrat in Catalonia (a region of Spain), is a collection of devotional texts also containing late Medieval songs. The Wikipedia article indicates that the monastery also holds the shrine of the Virgin of Montserrat, a major site of pilgrimage at the time the Red Book was compiled. Accordingly, its songs and dances were probably intended for pilgrims to the shrine and were part of a well-developed oral folk tradition. The 14th-century manuscript does not identify authors or composers. Furthermore, it predates modern musical notation, so performances and recordings today are reconstructions.

The music on the recording fuses sacred and secular (folk) elements and strongly suggests communal participation. In contrast, the modern concert hall has become the scene of rigid propriety. Audience members try to sit in stone silence (notwithstanding inevitable cell phone interruptions) while performers demonstrate their, um, emotionless professionalism. Live concerts of popular musics (multiple genres) instead feature audiences dancing and singing along, creating an organic experience that transforms the concertgoer into a participant situated in the middle of the flow rather than at the distant receiving end. Middle ground, such as when symphony orchestras perform film or video game music, often draws untutored audiences who may want to participate and in doing so frankly offend others trained to be still.

Is there cultural connection between pilgrimages, processions, and parades? The first is plainly religious is motivation, such as visits to Catholic shrines, the Wailing Wall in Jerusalem, or Mecca. Processions are more ceremonial and may not be religious in orientation. A wedding procession is a good example. Parades are more nearly civil in character, unless one insists on nationalism (e.g., Independence Day in the U.S., Bastille Day in France, Victory Day in Russia) being civil religions. The sense of devotion, sacrifice, and hardship associated with pilgrimage, historical or modern, contrasts with the party atmosphere of a parade, where Carnival, Mardi Gras, and Día de Muertos in particular invite licentious participation. Typical U.S. holiday parades (e.g., Independence Day, Thanksgiving) feature spectators arrayed lazily along the streets. There is even a subgenre of march form (used in band concerts) called a “patrol” that employs a broad crescendo-diminuendo (getting louder then fading away) to depict a military column as it marches by.

I suspect that modern processions and parades are weak echos of pilgrimage, a gradual transformation of one thing into something else. Yet the call of the open road (a/k/a wanderlust) resurfaces periodically even when not specifically religious in motivation. The great westward migration of Europeans to North American and then Americans across the untamed frontiers attests to that venturing spirit. In literature, Jack London’s memoir The Road (1907) describes the hobo life hopping trains in the 1890s, while Jack Kerouac’s On the Road (1957) tells of traveling across America by car. Another expression of wanderlust was penned by forgotten American poet Vachel Lindsay in his self-published War Bulletin #3 (1909):

Let us enter the great offices and shut the desk lids and cut the telephone wires. Let us see that the skyscrapers are empty and locked, and the keys thrown into the river. Let us break up the cities. Let us send men on a great migration: set free, purged of the commerce-made manners and fat prosperity of America; ragged with the beggar’s pride, starving with the crusader’s fervor. Better to die of plague on the highroad seeing the angels, than live on iron streets playing checkers with dollars ever and ever.

Lindsay invites his readers to embrace a life better lived traversing the landscape in a voyage of self-discovery. His version lacks the religious orientation of pilgrimage, but like the Medieval cultures depicted in film and music from the period, possesses tremendous appeal for modern Westerners starved of meaning that arises naturally out of tradition.