Posts Tagged ‘movies’
“In comics at their best, words and pictures are like partners in a dance, and each one takes turns leading”*…

In his new book, Lost Literacies: Experiments in the 19th Century US Comic Strip, literary historian Alex Beringer demonstrates how the birth of the genre of printed comic long preceded the Sunday Funny Pages. He elaborates in conversation with Tim Brinkhof, who introduces the colloquy…
Most people consider the introduction of the Funny Pages in the late nineteenth century as the birthday of the “modern” American comic strip. Alex Beringer is not most people.
A literary historian and professor of English at the University of Montevallo, Beringer dates the history of comics earlier, to roughly the mid-1800s, a period of prolific and uninhibited experimentation. He came to this understanding by piecing together the medium’s fractured archaeological record, diving through myriad online resources and archives. In the middle of the nineteenth century, New York-based artists followed the lead of their French and Swiss colleagues, particularly Rodolphe Töpffer, the “Father of the Comic Strip,” exchanging single-image political cartoons and caricatures for multi-panel sequences that, many believe, for the first time enabled them to play around with characterization, worldbuilding, and—well—storytelling.
Coming decades before the standardization of speech bubbles and panel borders, these early American comics seem to have little in common with their modern, more streamlined counterparts; they featured sudden and purposefully jarring jump cuts reminiscent of the yet-to-be-invented film montage or musical notes instead of text. One comic artist tells a story through shadows behind the curtains of a window; another, with hieroglyphs the reader must decipher with the help of a legend.
“The audience for this first wave of US comic strips was strikingly sophisticated in its reception of this material,” Beringer writes in Lost Literacies: Experiments in the Nineteenth-Century US Comic Strip, which chronicles this oft-forgotten renaissance. Out from the Ohio State University Press, the book is one of hundreds of titles included in JSTOR’s Path to Open program, making scholarly books accessible online to wide audiences (read chapter four here, free of charge).
“The sense of flux—the idea that the visual language could turn on a dime—was often precisely the appeal,” Beringer observes in his chronicle of this oft-forgotten renaissance.
Foretelling the philosopher Martin Heidegger’s assertion that drawing is in itself a “form of knowing,” early comic strip artists and their consumers treated the medium as a philosophical exercise; Beringer quotes the observation by media scholars Hilary Chute and Patrick Jagoda that comics “enable an intense focus on how complexly woven stories unfold across time and space and, particularly, how these involve the reader…to generate meaning through interacting with, or themselves shaping, spatiotemporal form.”
While some early American artists blatantly plagiarized illustrations and formats that originated in France and Switzerland, others used them as a springboard, giving European drawings a decidedly American twist. For example, where Töpffer’s character Monsieur Vieux Bois (“Mr. Oldbuck”) satirized the European bourgeoisie, comics featuring his Yankee doppelganger, Jeremiah Oldpot (artist unknown), a New York tin merchant who leaves his family to prospect gold in California, often hinge on what Beringer defines as the contradiction between his “romantic view of himself as a rugged frontiersman and his attachment to consumer goods.”
Beringer discusses this and other critical facets of this period in comics history…
Read on for their fascinating exchange: “Lost Literacies Strips Down the Dawn of Comics,” from @jstordaily.bsky.social.
* Scott McCloud, in his wonderful Understanding Comics: The Invisible Art
###
As we tell and show, we might ponder where all of this has led, recalling that it was on this date in 2007 that the then-latest entry in a comic-born franchise dropped: TMNT, the first animated entry in the Teenage Mutant Ninja Turtles film series, was released. The film (which was entirely computer animated), is set after the final defeat of their arch-enemy, the Shredder; the four Turtles — Leonardo, Raphael, Donatello, and Michelangelo (voiced respectively by James Arnold Taylor, Nolan North, Mitchell Whitfield, and Mikey Kelley) — having grown apart, reunite and overcome their faults to save the world from evil ancient creatures. It also features the voices of Chris Evans, Sarah Michelle Gellar, Mako, Kevin Smith, Patrick Stewart, and Ziyi Zhang, with narration by Laurence Fishburne.
TMNT ranked number one at the box office on its opening weekend, beating 300 (the top film of the previous two weeks), The Last Mimzy, Shooter, Pride, The Hills Have Eyes 2, and Reign Over Me, grossing $25.45 million over the weekend of March 23–25, 2007. That said, the film grossed (only) $95.8 million million worldwide, including $54 million domestically during its 91-day run in the 3,120 North American theaters… as the Rotten Tomatoes consensus read: “TMNT’s art direction is splendid, but the plot is non-existent and the dialogue lacks the irony and goofy wit of the earlier Ninja Turtles movies.”
“Mr. Hackett turned the corner and saw, in the failing light, at some distance, his seat”*…
Michael Wolf is an award-winning and widely-exhibited photographer famous for his documentation of big city architecture and life around the world, but especially in Hong Kong… Consider this series…
Much more at “Informal Seating Arrangements in Hong Kong” and more of Wolf’s other wonderful work on his site.
* Samuel Beckett, Watt
###
As we grab a chair, we might keep our focus on Hong Kong: it was on this date in 1978 that Snake in Eagle’s Shadow was released. A Hong Kong martial arts action comedy film, it was the debut of director Yuen Woo-ping, and the breakthrough outing for its stars, Jackie Chan, Hwang Jang-lee, and (Yuen Woo-ping’s real life father) Yuen Siu-tien.
The film is the story of Chien Fu (Jackie Chan), an orphan who is bullied at a kung fu school, but meets an old beggar, Pai Cheng-tien (Yuen Siu-tien), who becomes his sifu (teacher) and trains him in Snake Kung Fu. The film established Chan’s slapstick kung fu comedy style– which he further developed with Drunken Master, also directed by Yuen Woo-ping, released in the same year, and also starring Jackie Chan, Hwang Jang-lee and Yuen Siu-tien. Snake in Eagle’s Shadow (and Drunken Master) established the basic plot structure used in many, many martial arts films internationally since then.
“[They] would think that the truth is nothing but the shadows cast by the artifacts.”*…
How do AI models “understand” and represent reality? Is the inside of a vision model at all like a language model? As Ben Brubaker reports, researchers argue that as the models grow more powerful, they may be converging toward a singular “Platonic” way to represent the world…
Read a story about dogs, and you may remember it the next time you see one bounding through a park. That’s only possible because you have a unified concept of “dog” that isn’t tied to words or images alone. Bulldog or border collie, barking or getting its belly rubbed, a dog can be many things while still remaining a dog.
Artificial intelligence systems aren’t always so lucky. These systems learn by ingesting vast troves of data in a process called training. Often, that data is all of the same type — text for language models, images for computer vision systems, and more exotic kinds of data for systems designed to predict the odor of molecules or the structure of proteins. So to what extent do language models and vision models have a shared understanding of dogs?
Researchers investigate such questions by peering inside AI systems and studying how they represent scenes and sentences. A growing body of research has found that different AI models can develop similar representations, even if they’re trained using different datasets or entirely different data types. What’s more, a few studies have suggested that those representations are growing more similar as models grow more capable. In a 2024 paper, four AI researchers at the Massachusetts Institute of Technology argued that these hints of convergence are no fluke. Their idea, dubbed the Platonic representation hypothesis, has inspired a lively debate among researchers and a slew of follow-up work.
The team’s hypothesis gets its name from a 2,400-year-old allegory by the Greek philosopher Plato. In it, prisoners trapped inside a cave perceive the world only through shadows cast by outside objects. Plato maintained that we’re all like those unfortunate prisoners. The objects we encounter in everyday life, in his view, are pale shadows of ideal “forms” that reside in some transcendent realm beyond the reach of the senses.
The Platonic representation hypothesis is less abstract. In this version of the metaphor, what’s outside the cave is the real world, and it casts machine-readable shadows in the form of streams of data. AI models are the prisoners. The MIT team’s claim is that very different models, exposed only to the data streams, are beginning to converge on a shared “Platonic representation” of the world behind the data.
“Why do the language model and the vision model align? Because they’re both shadows of the same world,” said Phillip Isola, the senior author of the paper.
Not everyone is convinced. One of the main points of contention involves which representations to focus on. You can’t inspect a language model’s internal representation of every conceivable sentence, or a vision model’s representation of every image. So how do you decide which ones are, well, representative? Where do you look for the representations, and how do you compare them across very different models? It’s unlikely that researchers will reach a consensus on the Platonic representation hypothesis anytime soon, but that doesn’t bother Isola.
“Half the community says this is obvious, and the other half says this is obviously wrong,” he said. “We were happy with that response.”…
Read on: “Distinct AI Models Seem To Converge On How They Encode Reality,” from @quantamagazine.bsky.social.
Bracket with: “AGI is here (and I feel fine),” from Robin Sloan and “We Need to Talk About How We Talk About ‘AI’,” from Emily Bender and Nanna Inie.
* from Socrates “Allegory of the Cave,” in Plato’s Republic (Book VII)
###
As we interrogate ideas and Ideas, we might recall that it was on this date that the fictional HAL 9000 computer became operational, according to Arthur C. Clarke’s 2001: A Space Odyssey., in which the artificially-intelligent computer states: “I am a HAL 9000 computer, Production Number 3. I became operational at the HAL Plant in Urbana, Illinois, on January 12, 1997.” (Kubrik’s 1968 movie adaptation put his birthdate in 1992.)












You must be logged in to post a comment.