(Roughly) Daily

Archive for August 2020

“Doomsday is quite within our reach, if we will only stretch for it”*…

 

komet-001-930x598

 

Ah, to have planned for a comet apocalypse in the Belle Epoque! Looks like a wild time, according to all the 19th and 20th century postcards with folks soaring through the skies, bum over noggin, towards a starry death. Out of all the (ultimately) anti-climactic close comet calls, the 1910 approach of Halley’s Comet takes the cake. Not only did it lead to the first ever photograph of a comet, and the actual gathering of “spectroscopic ” data, but because the anticipation of its arrival caused a media frenzy. Stores started selling “anti-comet pills.” Papers put out ads for escape submarine rentals. One cult considered sacrificing a virgin…

It’s hard for us science laypeople to fathom the workings of space, let alone its errant pebbles. Although, Halley’s never been just a random fly bye event; by 1705, its orbital period of 75 or so years was confirmed by the English astronomer, Edmond Halley, bringing us both the comet’s eponymous name and a new tradition: once, maybe even twice in a lifetime, the comet’s 24-million-mile long tail would become visible to the naked eye. Fun fact: he might not have had his breakthrough, if he hadn’t consulted a friend and fellow scholar named Isaac Newton.

But people started to wonder – were there consequences of such seemingly close contact? One account of its 1835 passing describes a “vapour trail” in the sky…

When Halley was next slated to return, it was at the beginning of a new century in 1910, when advancements in media and technology had radically accelerated the circulation of people and ideas, and major breakthroughs were being made in the automobile and radio industries; the world had seen the first airplanes and photographs – including the photographs of astronomical objects. That meant it was finally, hopefully, the moment to capture an image of the comet when it neared earth in one of its shortest return cycles yet, a mere 74 years. It had one man in particular, a French scientist named Camille Flammarion, feeling rather worried. Flammarion was a prominent, and above all colourful presence in the astronomy scene. He ran the journal L’Astronomie, as well as his own private, castle-like observatory in Juvisy-sur-Org, France, which you can still visit today…

As an author, he penned both scientific essays and science fiction with a talent for poetic turns of a phrase. Readers loved it; critics tended to roll their eyes at his tendency for sensationalism. “This end of the world will occur without noise, without revolution, without cataclysm,” he wrote in L’atmosphère : météorologie populaire in 1888, “Just as a tree loses leaves in the autumn wind, so the earth will see in succession the falling and perishing of all its children, and in this eternal winter, which will envelop it from then on, she can no longer hope for either a new sun or a new spring. She will purge herself of the history of the worlds.” Yikes.

The incoming of Halley’s comet, he said, contained a poisonous cyanogen gas that “would impregnate the atmosphere and possibly snuff out all life on the planet.” When The New York Times ran a story on his assertion, the fear amplified on a global scale in the tabloids. One science writer, Matt Simon, said folks were so frightened, they began sealing up the keyholes of their houses to “keep the poison out of their homes.”… Comet pills, comet shelters, comet soap, and even submarine rentals became the norm for doomsday preppers… Even fashion took a turn. It wasn’t uncommon to find both amateur and professional-grade comet buttons, broaches and jewellery…

“I came in with Halley’s comet in 1835,” Mark Twain famously wrote in 1909, decreeing that he fully expected “to go out with it in 1910,” which, to his credit, he did. Was Twain’s death a feat of willpower, coincidence, or fate? Who knows. But it was one in a series of events that used the comet as a launching pad for a kind of self-fulfilling prophecy. Some cited the event as the cause of death of King Edward VII. The civil unrest surrounding the comet even helped spark China’s Xinhai Revolution in 1911, which effectively ended the last dynasty…

Halleys_comet_1910-930x1035

Halley in April 1910, from Harvard’s Southern Hemisphere Station, taken with an 8-inch Bache Doublet

 

It’s the end of the world as they knew it: “Doomsday Prepping in the Belle Epoque.”

* Loudon Wainwright III

###

As we eye the sky, we might send far-seeing birthday greetings to James Ferguson; he was born on this date in 1797.  Working at the U.S. Naval Observatory with a 9.6 inch refractor telescope, made the first discovery of an asteroid from North America (31 Euphrosyne).

150px-James_Ferguson_(astronomer) source

 

Written by (Roughly) Daily

August 31, 2020 at 1:01 am

“Biology gives you a brain. Life turns it into a mind.”*…

 

Mind

 

I have a confession. As a physicist and psychiatrist, I find it difficult to engage with conversations about consciousness. My biggest gripe is that the philosophers and cognitive scientists who tend to pose the questions often assume that the mind is a thing, whose existence can be identified by the attributes it has or the purposes it fulfils.

But in physics, it’s dangerous to assume that things ‘exist’ in any conventional sense. Instead, the deeper question is: what sorts of processes give rise to the notion (or illusion) that something exists? For example, Isaac Newton explained the physical world in terms of massive bodies that respond to forces. However, with the advent of quantum physics, the real question turned out to be the very nature and meaning of the measurements upon which the notions of mass and force depend – a question that’s still debated today.

As a consequence, I’m compelled to treat consciousness as a process to be understood, not as a thing to be defined. Simply put, my argument is that consciousness is nothing more and nothing less than a natural process such as evolution or the weather. My favourite trick to illustrate the notion of consciousness as a process is to replace the word ‘consciousness’ with ‘evolution’ – and see if the question still makes sense. For example, the question What is consciousness for? becomes What is evolution for? Scientifically speaking, of course, we know that evolution is not for anything. It doesn’t perform a function or have reasons for doing what it does – it’s an unfolding process that can be understood only on its own terms. Since we are all the product of evolution, the same would seem to hold for consciousness and the self.

My view on consciousness resonates with that of the philosopher Daniel Dennett, who has spent his career trying to understand the origin of the mind. Dennett is concerned with how mindless, mere ‘causes’ (A leads to B) can give rise to the species of mindful ‘reasons’ as we know them (A happens so that B can happen). Dennett’s solution is what he calls ‘Darwin’s dangerous idea’: the insight that it’s possible to have design in the absence of a designer, competence in the absence of comprehension, and reasons (or ‘free-floating rationales’) in the absence of reasoners. A population of beetles that has outstripped another has probably done so for some ‘reason’ we can identify – a favourable mutation which produces a more camouflaging colour, for example. ‘Natural selection is thus an automatic reason-finder, which “discovers” and “endorses” and “focuses” reasons over many generations,’ Dennett writes in From Bacteria to Bach and Back: The Evolution of Minds (2017). ‘The scare quotes are to remind us that natural selection doesn’t have a mind, doesn’t itself have reasons, but is nevertheless competent to perform this “task” of design refinement.’

I hope to show you that nature can drum up reasons without actually having them for herself. In what follows, I’m going to argue that things don’t exist for reasons, but certain processes can nonetheless be cast as engaged in reasoning…

Distinguished neuroscientist and psychiatrist Karl Friston argues that the special trick of consciousness is being able to project action and time into a range of possible futures: “The mathematics of mind-time.”

See also: “How the Brain Creates a Timeline of the Past” (source of the image above).

* Jeffrey Eugenides, Middlesex

###

As we get our minds around our minds, we might spare a thought for Oliver Wolf Sacks; he died on this date in 2015.  A neurologist, naturalist, historian of science, and author, he had an active clinical practice, but is more widely-remembered for his writing, mostly case studies from his clinical experience and memoir in which which he treats himself as a clinical subject.  Awakenings, for example, recounted his experience treating post-encephalitic patients with a new drug (levodopa); it was a best seller that served as the basis of a BBC Discovery documentary and was adapted into a feature film.  Widely honored for his prolific work, Sacks was the recipient of the 2001 Lewis Thomas Prize for excellence in scientific writing.

250px-9.13.09OliverSacksByLuigiNovi source

 

Written by (Roughly) Daily

August 30, 2020 at 1:01 am

“Imperial is lit, but Metric is liter”…*

 

barleycorn

 

The English-speaking world has been famously (and, many argue, problematically) slow to switch to the metric system of measurement.  One of the reasons is the way in which traditional “imperial” measures are baked into our understanding of products and services we use every day.

Consider the barleycorn, which is still used as the basis of shoe sizes in English-speaking countries.

In ancient Rome, the inch (which was one twelfth of a foot) measured the width across the (interphalangeal) joint of the thumb. By the 7th century in England, the barleycorn became a standard measurement with three ears of corn, laid end to end, equalling one inch. It took until the thirteenth century before the inch was officially sanctioned. Under pressure, Edward II (r. 1307-27) eventually succumbed to appeals from scholars and tradesmen to issue a decree to standardise measurement (Ledger, 1985).

Henceforth an English inch was the distance measured across three barleycorns. Thirty nine (39) barleycorns laid end to end became a foot, and 117 laid end to end became a yard. Whilst the barleycorn decree of Edward II had nothing to do with shoe sizes per se, many shoemakers began to use shoe sticks. Tradesmen had traditionally used the hand span method of measurement, which preferred the quarter of an inch unit, but after the introduction of the barleycorn measure, many began to adopt the third of an inch unit. With 39 barley corns approximating the length of a normal foot this was graded Size 13 and became the largest shoe size. Other sizes were graded down by 1/3 rd of an inch or one barleycorn…  [source]

The barleycorn is but one of the old English measures that. more or less obviously, still shape our encounters with and experience of the world:

406px-English_Length_Units_Graph.svg

Forgotten, but not gone: the barleycorn.

* bad joke

###

As we muse on measurement, we might send inventive birthday greetings to Charles Franklin Kettering; he was born on this date in 1876.  An engineer, businessman, and inventor (the holder of 186 patents), he worked at National Cash Register (where he created the first electric cash register with an electric motor that opened the drawer), co-founded DELCO, and was head of research at General Motors from 1920 to 1947.  He invented the key-operated self-starting motor and developed several new engine types, quick-drying lacquer finishes, anti-knock fuels, and variable-speed transmissions.  In association with the DuPont Chemical Company, he was also responsible for the invention of Freon refrigerant for refrigeration and air conditioning systems.  While working with the Dayton-Wright Company he developed the “Bug” aerial torpedo, considered the world’s first aerial missile.  In 1927, he founded the Kettering Foundation, a non-partisan research foundation devoted to answering the question: “what does it take for democracy to work as it should?”

220px-Time-magazine-cover-charles-kettering source

 

Written by (Roughly) Daily

August 29, 2020 at 1:01 am

“All poets write bad poetry. Bad poets publish them, good poets burn them.”*…

 

Thunderstorm with the Death of Amelia 1784 by William Williams active 1758-1797

Thunderstorm with the Death of Amelia, by William Williams, 1784. Photograph © Tate (CC-BY-NC-ND 3.0).

 

Readers may recall an earlier nod to William Topaz McGonagall, widely considered to be the worst published poet in British history.  McGonagall, best known for his widely-excoriated verse recounting of “The Tay Bridge Disaster,” distributed his poems, often about momentous events, on handbills and performed them publicly (often, it is reported, to cat calls and thrown food).  And he collected his verse into volumes including Poetic Gems, More Poetic Gems, Still More Poetic Gems, Further Poetic Gems, and Yet Further Poetic Gems.  Imagine your correspondent’s surprise and delight to find a learned appreciation of McGonagall’s place in poetic history:

Not unjustly, McGonagall is rarely mentioned without an epithet: some version of “the worst poet in the English language.” And by any reasonable account, any judgment based on the most universally shared values of poetics, prosody, and taste, there is little to admire in McGonagall. The rest of his corpus shares—replicates, really—the faults of “The Tay Bridge Disaster”: its lapses into bathos, its involuted syntactical structures, its rhymes so slanted as to be more or less horizontal.

There have been worse poets, of course, and as such it would be more accurate to describe McGonagall as the worst famous poet in the English language, a testament in part to the man’s powers of self-promotion and the caprices of literary history. But McGonagall’s notoriety still owes much to the singularly strange power of his own badness. There’s something, I think, in poems like “The Tay Bridge Disaster”—as well as McGonagall’s many poems on his great themes of death and destruction—that is worth examining; something that might redeem him, ever so slightly, from the annals of amusing semi-obscurity; something unsettling about his ostensibly blinkered artistic vision that might help to account for why he lingers as the patron saint of misbegotten verse…

On William Topaz McGonagall, the worst famous poet in the English language: “The Disaster Poet.”

(Readers will find a selection of McGonagall’s poems here.)

* Umberto Eco

###

As we bathe in bathos, we might spare a thought for the decidedly more-accomplished poet (and playwright, artist, biologist, theoretical physicist, and philosopher) Johann Wolfgang von Goethe; he died on this date in 1832.  Probably best remembered these days for Faust, he was “the master spirit of the German people,” and, after Napoleon, the leading figure of his age.

 source

 

Written by (Roughly) Daily

August 28, 2020 at 1:01 am

“When you’re accustomed to wealth, you don’t show it, right? That’s why the white kids in school could wear bummy sneakers”*…

 

Chucks

The Converse Rubber Shoe Company’s non-skid All Star sneakers, from 1923

 

It was not until the 1920s that industrialization made sneakers widely available and affordable. Once an emblem of privileged leisure on the tennis court, the canvas-and-rubber high-top adapted to the new, egalitarian team sport of basketball. The Converse Rubber Shoe Company—founded in 1908 as a producer of galoshes—introduced its first basketball shoe, the All Star, in 1917. In a stroke of marketing genius, Converse enlisted basketball coaches and players as brand ambassadors, including Chuck Taylor, the first athlete to have a sneaker named after him.

Politics, however, fueled the rise of sneakers as much as athletics. As [Elizabeth Semmelhack, curator of the traveling exhibtion, Out of the Box: The Rise of Sneaker Culture] explained, “the fragile peace of World War I increased interest in physical culture, which became linked to rising nationalism and eugenics. Countries encouraged their citizens to exercise not just for physical perfection but to prepare for the next war. It’s ironic that the sneaker became one of the most democratized forms of footwear at the height of fascism.” Mass exercise rallies were features of fascist life in Germany, Japan, and Italy. But sneakers could also represent resistance. Jesse Owens’ dominance at the 1936 Berlin Olympic Games stung the event’s Nazi hosts even more because he trained in German-made Dassler running shoes. (The company was later split between the two Dassler brothers, who renamed their shares Puma and Adidas).

When the U.S. government rationed rubber during World War II, sneakers were exempted following widespread protests. The practical, inexpensive, and casual shoe had become central to American identity, on and off the playing field. The growing influence of television in the 1950s created two new cultural archetypes: the celebrity athlete and the teenager. James Dean effectively rebranded Chuck Taylors as the footwear of choice for young rebels without a cause.

Sneakers became footnotes in the history of the Civil Rights movement. In 1965, I Spy was the first weekly TV drama to feature a black actor—Bill Cosby—in a lead role. His character, a fun-loving CIA agent going undercover as a tennis coach, habitually wore white Adidas sneakers, easily identifiable by their prominent trio of stripes. This updated gumshoe alluded to the “sneaky” origins of sneakers, while also serving as shorthand for new-school cool. Sneakers played a more explicit part at the 1968 Olympic Games in Mexico City, where American gold medalist sprinter Tommie Smith and his bronze medal-winning teammate, John Carlos, removed their Puma Suedes and mounted the medal podium in their stocking feet, to symbolize African-American poverty, their heads lowered and black-gloved fists raised in a Black Power salute. The ensuing controversy didn’t hurt the success of the Suede, still in production today.

Around the same time, the jogging craze necessitated low-rise, high-tech footwear that bore little resemblance to the familiar canvas-and-rubber basketball high-top. But these state-of-the-art shoes weren’t made for running alone; they were colorful, covetable fashion statements. In 1977, Vogue declared that “real runner’s sneakers” had become status symbols, worn by famous non-athletes like Farrah Fawcett and Mick Jagger. Instead of one pair of sneakers, people needed a whole wardrobe of them, custom-made for different activities—or genders. Sneaker companies embraced women’s liberation as a promotional ploy, advertising shoes specifically designed for female bodies and lifestyles.

As the suburbs became overrun with joggers, America’s cities saw a rise in basketball players, particularly New York, where a bold new style of play transformed the game into a spectacle of masculine swagger. Like break dancing, schoolyard basketball ritualized a competitive physicality, which bled into mainstream (white) culture. “In the 1970s, New Yorkers in the basketball and hip-hop community changed the perception of sneakers from sports equipment to tools for cultural expression,” the sneaker historian Bobbito Garcia explains in the Out of the Box catalogue. “The progenitors of sneaker culture were predominantly … kids of color who grew up in a depressed economic era.” The 2015 documentary Fresh Dressed highlighted the prominent role of sneakers in the history of black urban culture—and its appropriation by whites.

The humble canvas sneaker, since the ’60s supplanted in the sports world by more ergonomic designs in futuristic materials, found new life as an everyday shoe. Over the next few decades, canvas sneakers came to embody youthful rebellion as much as athleticism. Beatniks, rockers, and skateboarders adopted them because they were cheap, anonymous, and authentic—not necessarily because they were comfortable or cool. Converse, Keds, and Vans got their street cred not from sports stars, but from the Ramones, Sid Vicious, and Kurt Cobain. (In 2008, Converse angered Nirvana fans by issuing special-edition high-tops tastelessly covered with sketches and scribbles from the late frontman’s diary.) The All Star, formerly available only in black or white, suddenly appeared in a rainbow of fashion colors.

The ascent of aerobics in the early ’80s left Nike, known for its jogging shoes, struggling to adjust. In February 1984, the company reported its first-ever quarterly loss, but that same year Nike signed basketball rookie Michael Jordan to an endorsement deal—arguably the birth of modern sneaker culture. Jordan wore his signature Air Jordans in NBA games, in defiance of league rules. Nike happily paid his $5,000-per-game fine, while airing ads declaring: “The NBA can’t keep you from wearing them.” And so when the first Air Jordans hit stores in 1985, the sneakers carried with them a distinct whiff of sticking it to The Man, despite their $65 price tag. But not everyone wanted to be like Mike. As Jordan grew rich off of his Nike partnership, he was accused of staying silent on political issues affecting the African American community. “Republicans buy sneakers, too,” he allegedly responded.

The growing popularity of sneakers on both sides of the political divide set the stage for a raging culture war over the shoes’ ties to criminality, or lack thereof. In “My Adidas” (1986)—one of many hip-hop sneaker shout-outs—Run-DMC defended their laceless Adidas Superstars against sneakers’ thuggish image as “felon shoes,” rapping: “I wore my sneakers, but I’m not a sneak.” (The band was rewarded with an Adidas endorsement deal, a first for a musical group.)

But Nike’s all-white Air Force 1 sneaker, released in the same year as “My Adidas,” may have merited the name of “felon shoes.” Having enough money to step out in “fresh”—i.e., pristine and unscuffed—Air Force 1s became a point of pride among street drug dealers. “Like the complicated icon of the cowboy, the drug dealer was also a symbol of rugged individualism whose fashion was hypermasculine and easily marketed … in ways that capitalized on both its American-ness and its exoticism simultaneously,” Semmelhack writes in the exhibition catalogue. The AF1, far from a public-relations disaster, became an instant classic…

Athletic shoes, a $65 billion global market, are about much more than athletics—conveying ideas about national identity, class, race, and other forms of social meaning: “Sneakers Have Always Been Political Shoes.”

See also “A Brief History of America’s Obsession With Sneakers,” “The Psychology of Sneaker Collecting,” “What’s driving retail’s sneaker obsession?,” and “The History of Sneaker Culture.”

[TotH to friend SL for pointing me in this direction]

* “When you’re accustomed to wealth, you don’t show it, right? That’s why the white kids in school could wear bummy sneakers; it’s almost like, ‘Don’t show wealth – that’s crass.’… As kids we didn’t complain about being poor; we talked about how rich we were going to be and made moves to get the lifestyle we aspired to by any means we could. And as soon as we had a little money, we were eager to show it…”   – Jay-Z

###

As we tuck into our Chucks, we might recall that it was on this date in 1975 that Austrian tennis player Onny Parun defeated two-time Grand Slam singles champion Stan Smith in the first U.S. Open match ever played at night.  Today Smith is probably better-known for the line of Adidas sneakers named for him.

stan smith source