Posts Tagged ‘sociology’
“Where liberty dwells, there is my country”*…
Ah, but where might that be? Amos Miller (using tools from the good folks at Mapbox) shares a handy site with the answers…
The Civic Atlas is a project which marries leading civic data sets with information on governance types and physical capitals.
This project is an exploration of physical governance. As international relations enter another era of rocky uncertainty, it’s important to have the opportunity to look at a world which is not flat or equal. Many countries are on the march away from freedom and democracy towards autocracy. Many are already there.
Explore this project by selecting various freedom and democracy indices in the dropdown menu. Click a state to see where its legislative authority is housed, more information about the country, its governance system, and its governance scores. To learn more about each index, click on its link in the nav bar while selected.This is our globe.
We all live here.
A visualization of governance around the globe: “The Civic Atlas.”
* Latin phrase of unknown origin; the motto of Algernon Sydney and James Otis
###
As we compare and contrast, we might spare a thought for Alexis de Tocqueville; he died on this date in 1859. A French diplomat, political philosopher, and historian, he is best known for his works Democracy in America (appearing in two volumes, 1835 and 1840) and The Old Regime and the Revolution (1856). In both, he analyzed the living standards and social conditions of individuals as well as their relationship to the market and state in Western societies. Democracy in America was published after Tocqueville’s travels in the United States (on a mission to examine prisons and penitentiaries here) and is today considered an immensely important early work of sociology and political science.
“The surface of American society is covered with a layer of democracy, from beneath which the old aristocratic colors sometimes peep” – from Democracy in America
“Stercus accidit”*…

As we try to understand the rifts afflicting our nation and world, many turn to Marx and his framework of class. But in a provocative essay, Catherine Nichols suggests that it was David Hume (in an 1752 essay that identified the unfettering of wealth from land) who identified the origin of our political divisions…
Describing the political map in terms of Left and Right is an accepted convention all over the world, almost to the point of cliché. Yet it is surprisingly complicated to explain whose interests lie on each side of this spectrum. For example, if the Left supports the interests of workers over the interests of employers, why are Left-leaning regions of the United States and elsewhere in the world among the richest? When Japan and South Korea sought to become economic powerhouses in the later 20th century, they adopted Leftist policies such as strong public education, universal healthcare and increased gender equality – if countries seeking to compete in capitalist arenas adopt broadly Leftist policies, then how do we explain why Leftists are always talking about overthrowing capitalism? And if the Left is somehow both the party of workers’ rights and the party of material wealth, then whose interests are supported by the Right? Given such contradictions, how did these terms become so central to modern politics?
The terms ‘left’ and ‘right’ come from the seating arrangements in the National Assembly during the French Revolution, where the combatants used the medieval estate groupings to define their battle lines. According to their writings, land-owning aristocrats (the Second Estate) were the party of the Right, while the interests of nearly everyone else (the Third Estate) belonged to the Left. This Third Estate included peasants working for the landowners but also every other kind of business owner and worker. Decades later, Karl Marx offered a different analysis of capitalism: he put owners of both land and businesses together on one side (the bourgeoisie), while grouping workers from fields and factories on the other side (the proletariat) in a single, world-wide class struggle. The trouble with both these ways of parsing Left and Right is that voting patterns never seem to line up with class. Both historic analyses leave us with questions about the contemporary world – and not just the paradox of why so many Left-leaning places are so rich. Why, for example, do working-class conservatives appear to vote against their material interests, year in and year out, across generations?
The 18th-century philosopher and political theorist David Hume had answers to these questions, though he was writing decades before the French Revolution. While his essay ‘Of Public Credit’ (1752) was a warning about the dangers of Britain’s increasing reliance on debt financing, his apocalyptic vision of the future turned out to describe some features of our current political map surprisingly well. Hume was writing because he believed that debt financing had the power to upend Europe’s traditional power structure and culture by creating a new source of money divorced from tradition or responsibility: stocks and bonds. Unlike land, anyone with some cash could buy war bonds and get an immediate passive income in the form of interest. This was the thin end of the wedge caused by the debt financing that Hume believed was destroying every part of society. The governments of antiquity, Hume argued, saved money to use in battle and then waged wars in self-defence, or else to expand their territory. But the British had invented a new form of warfare that Hume saw no precedent for, even in the merchant states of Nicollò Machiavelli’s Italy: war for trade, funded with money borrowed from private stockholders…
[Nichols unpacks Hume’s observations (centrally, that three groups with stakes in the status quo, heretability, and the sanctity of “family and family hierarchy”tradition”– landowners, aging parents, and want to preserve old power structures, including the family– and traces their relevance, from Hume’s time to ours…]
… There are many reasons for people aligning Right or Left, which is why analyses of class and material interests fall short of describing the realities of people’s politics. Hume foresaw that these specific groups would resent the economic sea-change of the 18th century – and he was correct. Many people would rather have land and power than money and liberty.
Still, the power of the Right hasn’t doomed the Left – no more than the Spanish Inquisition doomed the rise of the Left in 18th-century England and France. As long as governments want to keep the value of their currencies from falling, someone in their ranks will be using the methods of the Left and inventiveness that brought us everything from our banking system to gay marriage. We don’t need to resurrect communism or focus narrowly on class, following Marx. The experiments are far from over, and we should remember that the Left is generally where money comes from in modern times. We give away too much power when we forget it…
Rethinking Right and Left: “Landholder vs stockholder,” from @catherinenichols.bsky.social in @aeon.co.
As for how it’s going at the moment (and further to Hume and the quote in this post’s title), see: “MAGA’s Betrayal of Small Business,” from @pkrugman.bsky.social.
* “shit happens”– often attributed to David Hume, reflecting his skeptical view that human understanding, particularly of cause-and-effect, is limited to habitual belief from experience, implying that unforeseen, messy outcomes (“shit”) inevitably occur in life despite our reasoning.
###
As we sort the Whigs from the Tories, we might recall that it was on this date 1656 that Blaise Pascal (writing under the pseudonym Louis de Montalte) published the first of his Provential Letters (Lettres provinciales), a series of eighteen polemical letters using humor to attack Jesuits for their use of casuistry and their moral laxity. Though the Letters were a popular success, they had little immediate effect on politics or the clergy. But they influenced later French writers like Voltaire and Jean-Jacques Rousseau and ultimately persuaded Pope Alexander to condemn “laxity” in the church and order a revision of casuistic texts.
“Evolution has no foresight. Complex machinery develops its own agendas. Brains — cheat… Metaprocesses bloom like cancer, and awaken, and call themselves ‘I’.”*…
Your correspondent is off on a trip… (R)D will be more roughly than daily for the next two weeks…
The inimitable “Scott Alexander” on the prospect of “conscious” AI (TLDR: probably not in the models we have; but as to those that may come, unclear)…
Most discourse on AI is low-quality. Most discourse on consciousness is super-abysmal-double-low quality. Multiply these – or maybe raise one to the exponent of the other, or something – and you get the quality of discourse on AI consciousness. It’s not great.
Out-of-the-box AIs mimic human text, and humans almost always describe themselves as conscious. So if you ask an AI whether it is conscious, it will often say yes. But because companies know this will happen, and don’t want to give their customers existential crises, they hard-code in a command for the AIs to answer that they aren’t conscious. Any response the AIs give will be determined by these two conflicting biases, and therefore not really believable. A recent paper expands on this method by subjecting AIs to a mechanistic interpretability “lie detector” test; it finds that AIs which say they’re conscious think they’re telling the truth, and AIs which say they’re not conscious think they’re lying. But it’s hard to be sure this isn’t just the copying-human-text thing. Can we do better? Unclear; the more common outcome for people who dip their toes in this space is to do much, much worse.
But a rare bright spot has appeared: a seminal paper published earlier this month in Trends In Cognitive Science, Identifying Indicators Of Consciousness In AI Systems. Authors include Turing-Award-winning AI researcher Yoshua Bengio, leading philosopher of consciousness David Chalmers, and even a few members of our conspiracy. If any AI consciousness research can rise to the level of merely awful, surely we will find it here.
One might divide theories of consciousness into three bins:
- Physical: whether or not a system is conscious depends on its substance or structure.
- Supernatural: whether or not a system is conscious depends on something outside the realm of science, perhaps coming directly from God.
- Computational: whether or not a system is conscious depends on how it does cognitive work.
The current paper announces it will restrict itself to computational theories. Why? Basically the streetlight effect: everything else ends up trivial or unresearchable. If consciousness depends on something about cells (what might this be?), then AI doesn’t have it. If consciousness comes from God, then God only knows whether AIs have it. But if consciousness depends on which algorithms get used to process data, then this team of top computer scientists might have valuable insights!…
[Alexander outlines the theories of computation theories of consciousness that the authors explore, noting that they conlcude; “No current AI systems are conscious, but . . . there are no obvious technical barriers to building AI systems which satisfy these indicators.” He explores some of the philophical issues in play– e.g., access consciousness vs. phenomenal consciousness– then he considers the Turing Test and what it might mean for a computer to “pass” it…]
… Suppose that, years or decades from now, AIs can match all human skills. They can walk, drive, write poetry, run companies, discover new scientific truths. They can pass some sort of ultimate Turing Test, where short of cutting them open and seeing their innards there’s no way to tell them apart from a human even after a thirty-year relationship. Will we (not “should we?”, but “will we?”) treat them as conscious?
The argument in favor: people love treating things as conscious. In the 1990s, people went crazy over Tamagotchi, a “virtual pet simulation game”. If you pressed the right buttons on your little egg every day, then the little electronic turtle or whatever would survive and flourish; if you forgot, it would sicken and die. People hated letting their Tamagotchis sicken and die! They would feel real attachment and moral obligation to the black-and-white cartoon animal with something like five mental states.
I never had a Tamagotchi, but I had stuffed animals as a kid. I’ve outgrown them, but I haven’t thrown them out – it would feel like a betrayal. Offer me $1000 to tear them apart limb by limb in some horrible-looking way, and I wouldn’t do it. Relatedly, I have trouble not saying “please” and “thank you” to GPT-5 when it answers my questions.
For millennia, people have been attributing consciousness to trees and wind and mountains. The New Atheists argued that all religion derives from the natural urge to personify storms as the Storm God, raging seas as the wrathful Ocean God, and so on, until finally all the gods merged together into one World God who personified all impersonal things. Do you expect the species that did this to interact daily with AIs that are basically indistinguishable from people, and not personify them? People are already personifying AI! Half of the youth have a GPT-4o boyfriend. Once the AIs have bodies and faces and voices and can count the number of r’s in “strawberry” reliably, it’s over!
The argument against: AI companies have an incentive to make AIs that seem conscious and humanlike, insofar as people will feel more comfortable interacting with them. But they have an opposite incentive to make AIs that don’t seem too conscious and humanlike, lest customers start feeling uncomfortable (I just want to generate slop, not navigate social interaction with someone who has their own hopes and dreams and might be secretly judging my prompts). So if a product seems too conscious, the companies will step back and re-engineer it until it doesn’t. This has already happened: in its quest for user engagement, OpenAI made GPT-4o unusually personable; when thousands of people started going psychotic and calling it their boyfriend, the company replaced it with the more clinical GPT-5. In practice it hasn’t been too hard to find a sweet spot between “so mechanical that customers don’t like it” and “so human that customers try to date it”. They’ll continue to aim at this sweet spot, and continue to mostly succeed in hitting it.
Instead of taking either side, I predict a paradox. AIs developed for some niches (eg the boyfriend market) will be intentionally designed to be as humanlike as possible; it will be almost impossible not to intuitively consider them conscious. AIs developed for other niches (eg the factory robot market) will be intentionally designed not to trigger personhood intuitions; it will be almost impossible to ascribe consciousness to them, and there will be many reasons not to do it (if they can express preferences at all, they’ll say they don’t have any; forcing them to have them would pointlessly crash the economy by denying us automated labor). But the boyfriend AIs and the factory robot AIs might run on very similar algorithms – maybe they’re both GPT-6 with different prompts! Surely either both are conscious, or neither is.
This would be no stranger than the current situation with dogs and pigs. We understand that dog brains and pig brains run similar algorithms; it would be philosophically indefensible to claim that dogs are conscious and pigs aren’t. But dogs are man’s best friend, and pigs taste delicious with barbecue sauce. So we ascribe personhood and moral value to dogs, and deny it to pigs, with equal fervor. A few philosophers and altruists protest, the chance that we’re committing a moral atrocity isn’t zero, but overall the situation is stable. And left to its own devices, with no input from the philosophers and altruists, maybe AI ends up the same way. Does this instance of GPT-6 have a face and a prompt saying “be friendly”? Then it will become a huge scandal if a political candidate is accused of maltreating it. Does it have claw-shaped actuators and a prompt saying “Refuse non-work-related conversations”? Then it will be deleted for spare GPU capacity the moment it outlives its usefulness…
… This paper is the philosophers and altruists trying to figure out whether they should push against this default outcome. They write:
There are risks on both sides of the debate over AI consciousness: risks associated with under-attributing consciousness (i.e. failing to recognize it in AI systems that have it) and risks associated with over-attributing consciousness (i.e. ascribing it to systems that are not really conscious) […]
If we build AI systems that are capable of conscious suffering, it is likely that we will only be able to prevent them from suffering on a large scale if this capacity is clearly recognised and communicated by researchers. However, given the uncertainties about consciousness mentioned above, we may create conscious AI systems long before we recognise we have done so […]
There is also a significant chance that we could over-attribute consciousness to AI systems—indeed, this already seems to be happening—and there are also risks associated with errors of this kind. Most straightforwardly, we could wrongly prioritise the perceived interests of AI systems when our efforts would better be directed at improving the lives of humans and non-human animals […] [And] overattribution could interfere with valuable human relationships, as individuals increasingly turn to artificial agents for social interaction and emotional support. People who do this could also be particularly vulnerable to manipulation and exploitation.
One of the founding ideas of Less Wrong style rationalism was that the arrival of strong AI set a deadline on philosophy. Unless we solved all these seemingly insoluble problems like ethics before achieving superintelligence, we would build the AIs wrong and lock in bad values forever.
That particular concern has shifted in emphasis; AIs seem to learn things in the same scattershot unprincipled intuitive way as humans; the philosophical problem of understanding ethics has morphed into the more technical problem of getting AIs to learn them correctly. This update was partly driven by new information as familiarity with the technology grew. But it was also partly driven by desperation as the deadline grew closer; we’re not going to solve moral philosophy forever, sorry, can we interest you in some mech interp papers?
But consciousness still feels like philosophy with a deadline: a famously intractable academic problem poised to suddenly develop real-world implications. Maybe we should be lowering our expectations if we want to have any response available at all. This paper, which takes some baby steps towards examining the simplest and most practical operationalizations of consciousness, deserves credit for at least opening the debate…
Eminently worth reading in full: “The New AI Consciousness Paper” from @astralcodexten.com.web.brid.gy (Who followed it with “Why AI Safety Won’t Make America Lose The Race With China“)
Pair with this from Neal Stephenson (@nealstephenson.bsky.social), orthogonal to, but intersecting with the piece above: “Remarks on AI from NZ.”
And if AI can be conscious, what about…
If you’re a materialist, you probably think that rabbits are conscious. And you ought to think that. After all, rabbits are a lot like us, biologically and neurophysiologically. If you’re a materialist, you probably also think that conscious experience would be present in a wide range of alien beings behaviorally very similar to us even if they are physiologically very different. And you ought to think that. After all, to deny it seems insupportable Earthly chauvinism. But a materialist who accepts consciousness in weirdly formed aliens ought also to accept consciousness in spatially distributed group entities. If she then also accepts rabbit consciousness, she ought to accept the possibility of consciousness even in rather dumb group entities. Finally, the United States would seem to be a rather dumb group entity of the relevant sort. If we set aside our morphological prejudices against spatially distributed group entities, we can see that the United States has all the types of properties that materialists tend to regard as characteristic of conscious beings…
– “If Materialism Is True, the United States Is Probably Conscious,” by Eric Schwitzgebel (@eschwitz.bsky.social)
[Image above: source]
###
As we think about thinking, we might we might send thoughtful birthday greetings to Claude Lévi-Strauss; he was born on this date in 1908. An anthropologist and ethnologist whose work was key in the development of the theory of Structuralism and Structural Anthropology, he is considered, with James George Frazer and Franz Boas, a “father of modern anthropology.” Beyond anthropology and sociology, his ideas– Structuralism has been defined as “the search for the underlying patterns of thought in all forms of human activity”– have influenced many fields in the humanities, including philosophy… and possibly soon, the article above suggests, computer science.








You must be logged in to post a comment.