(Roughly) Daily

Posts Tagged ‘Jack Kilby

“Life is more fun if you play games”*…

A scanned image of the title page of a scientific paper titled 'The Unsuccessful Self-Treatment of a Case of Writer's Block' by Dennis Upper, published in the Journal of Applied Behavior Analysis in 1974.

David Freidman ponders scientific satire…

I first encountered the scientific paper simply titled “Strapless Evening Gowns” four years ago, when I was flipping through a collection of magazines that once belonged to the cybernetics pioneer Norbert Wiener.

Among his magazines was the May 1960 issue of Voo Doo magazine, which was MIT’s “only intentionally humorous campus publication” going all the way back to 1919.

I got a chuckle out of this article, which attempted to semi-seriously analyze what exactly keeps a strapless dress from falling down:

A page from the May 1960 issue of Voo Doo magazine featuring an article titled 'Strapless Evening Gowns,' discussing the structural analysis of strapless dresses with diagrams and equations.

Scientists have a long history of amusing themselves with humor. In addition to Voo Doo, other science humor magazines include the Annals of Improbable Research, the Journal of Irreproducible Results, and the Worm Runner’s Digest which included both satirical and serious scientific papers, much to the confusion of their readers – a problem eventually solved by printing the satirical articles upside down.

And then there are the quasi-serious scientific studies meant to be amusing, such as this study on the effectiveness of tin foil hats in protecting you from government surveillance (spoiler: tin foil hats can actually amplify certain radio frequencies, so the authors speculate that the government has been behind promotion of tin foil hats all along).

And back in 1974, the Journal of Applied Behavior Analysis published a paper by clinical psychologist Dennis Upper called “The Unsuccessful Self-Treatment Of A Case Of Writer’s Block” [pictured at the top]. And lest you question the veracity of the author’s finding, I should note that the author’s failure to treat his writer’s block has been successfully replicated

Read on for more on the engineering of the formal dress, both the social (largely sexist) context and the (interestingly meaninful) scientific content– and the art it has inspired: “Science And The Strapless Evening Gown” from @ironicsans.com.

More seriously: “a Nature analysis signals the beginnings of a US science brain drain“: “Researchers in the United States are seeking career opportunities abroad as President Donald Trump’s administration slashes science funding and workforce numbers, finds an analysis of Nature’s jobs-board data…”

* Roald Dahl

###

As we play, we might recall that it was on this date in 1961 that Robert Noyce was issued patent number 2981877 for his “semiconductor device-and-lead structure,” the first patent for what would come to be known as the integrated circuit.  In fact another engineer, Jack Kilby, had separately and essentially simultaneously developed the same technology. Ineligible (as a new Texas Instruments empoyee) for a vacation over the summer of 1959, he gave himself the “assignment” of creating “a body of semiconductor material … wherein all the components of the electronic circuit are completely integrated.”

Kilby’s design was rooted in germanium; Noyce’s in silicon and had filed a few months earlier than Noyce. But Kilby’s invention was not a true monolithic integrated circuit chip since it had external gold-wire connections, which would have made it difficult to mass-produce– an obstacle Noyce overcame. Still, Kilby’s contribution was recognized in 2000 when he was Awarded the Nobel Prize– in which Noyce, who had died in 1990, did not share.

A historical black-and-white photo of Robert Noyce holding a semiconductor device design while posed in front of a wooden backdrop.
Noyce with his “motherboard” (source)

Written by (Roughly) Daily

April 25, 2025 at 1:00 am

“Give credit where credit is due”*…

In the early 19th century, a young woman revolutionized the lumber business…

As a young woman, Tabitha Babbitt was a weaver in Harvard, Massachusetts. She used to watch the workers at the local sawmill.  Observing them use the difficult two-man whipsaw, she noticed that half of their motion was wasted. It had two handles which two men would pull from side to side. However, the saw only cut the wood when it was being pulled forward. This meant the second or reverse pull was fairly useless other than to get the saw back to starting position which was a waste of energy.  Tabitha proposed creating a round blade to increase efficiency. Eventually she came up with a prototype, attaching a circular blade to her spinning wheel, using the pedal of her wheel to power it. As the blade spun, no movement was wasted. The circular saw was connected to a water-powered machine to reduce the effort to cut lumber, meaning that wood could be cut faster with half the manpower. The first circular saw she allegedly made is in Albany, New York State USA. A larger version of her design was later installed in the sawmill.

But – Tabitha was a member of the Shakers, a Christian sect founded circa 1747 in England who had emigrated and settled in revolutionary colonial America.  Their core beliefs centred round a perfect society, created through communal living, gender and racial equality, pacifism, confession of sin, celibacy and separation from the world.  As such, they valued hard manual work, a simple lifestyle, and thrived on the forestry industry.

However, their beliefs prohibited any member applying for patents as they believed intellectual properties should be shared by the community with no restriction. Because she did not patent it (and according to wiki the reference to her invention exists only in Shaker lore), there is controversy over whether she was the first true inventor of the circular saw. 

Two French men patented the circular saw in USA after reading about her saw in Shaker papers. One of the patentees, Stephen Miller argues she wasn’t the first inventor based on the date she joined the sect. He contended that it was invented at Mount Lebanon Shaker Village by Amos Bishop or Benjamin Bruce in 1793 – or not by a Shaker at all.

Samuel Miller obtained a patent in UK for a saw windmill which supposedly used a form of a circular saw in 1777 though the type of saw is only mentioned in passing, making it seem as though it was not his invention. Walter Taylor a few years later in same area of the United Kingdom seemed to have types of circular saws at his sawmill but in fact he only ever received patents for improvements to blockmaking.

However, it appears Babbitt’s circular saw design was much larger than other circular saw mechanisms and enough modifications were made to differentiate her invention from the rest. Her basic design was also the one that soon was copied at various American sawmills, popularising the use of circular saw in mills.

Tabitha was also credited with improving the spinning wheel head, inventing a process to manufacture false teeth, and inventing a process for manufacturing the then semi-revolutionary type of nail known as “cut nails” which replaced forged nails, a claim to fame she shares with a few other inventors including famed inventor Eli Whitney…

An unsung hero: “Tabitha Babbitt” from the Mills Archive, via Mathew Ingram‘s When the Going Gets Weird (also source of the image above).

* Attributed to Samuel Adams, who used the phrase in a late 18th century letter

###

As we investigate innovation, we might recall that it was on this date in 1959 that Jack Kilby of Texas Instruments filed the first patent for an integrated circuit (U.S. Patent 3,138,743).  In mid-1958, as a newly employed engineer at Texas Instruments, Kilby didn’t yet have the right to a summer vacation.  So he spent the summer working on the problem in circuit design known as the “tyranny of numbers” (how to add more and more components, all soldered to all of the others, to improve performance).  He finally came to the conclusion that manufacturing the circuit components en masse in a single piece of semiconductor material could provide a solution. On September 12, he presented his findings to the management: a piece of germanium with an oscilloscope attached. Kilby pressed a switch, and the oscilloscope showed a continuous sine wave– proving that his integrated circuit worked and thus that he had solved the problem. 

Kilby is generally credited as co-inventor of the integrated circuit, along with Robert Noyce (who independently made a similar circuit a few months later).  Kilby has been honored in many ways for his breakthrough, probably most augustly with the 2000 Nobel Prize in Physics.

Kilby’s first integrated circuit

“Artificial intelligence is growing up fast”*…

A simple prototype system sidesteps the computing bottleneck in tuning– teaching– artificial intelligence algorithms…

A simple electrical circuit [pictured above] has learned to recognize flowers based on their petal size. That may seem trivial compared with artificial intelligence (AI) systems that recognize faces in a crowd, transcribe spoken words into text, and perform other astounding feats. However, the tiny circuit outshines conventional machine learning systems in one key way: It teaches itself without any help from a computer—akin to a living brain. The result demonstrates one way to avoid the massive amount of computation typically required to tune an AI system, an issue that could become more of a roadblock as such programs grow increasingly complex.

“It’s a proof of principle,” says Samuel Dillavou, a physicist at the University of Pennsylvania who presented the work here this week at the annual March meeting of the American Physical Society. “We are learning something about learning.”…

More at “Simple electrical circuit learns on its own—with no help from a computer, from @ScienceMagazine.

* Diane Ackerman

###

As we brace ourselves (and lest we doubt the big things can grow from humble beginnings like these), we might recall that it was on this date in 1959 that Texas Instruments (TI) demonstrated the first working integrated circuit (IC), which had been invented by Jack Kilby. Kilby created the device to prove that resistors and capacitors could exist on the same piece of semiconductor material. His circuit consisted of a sliver of germanium with five components linked by wires. It was Fairchild’s Robert Noyce, however, who filed for a patent within months of Kilby and who made the IC a commercially-viable technology. Both men are credited as co-inventors of the IC. (Kilby won the Nobel Prize for his work in 2000; Noyce, who died in 1990, did not share.)

Kilby and his first IC (source)

“Do you think that the soul first shows itself by a gnashing of teeth?”*…

In January 2020, as a new plague began to upend life on Earth, a small research team from Tufts, the University of Vermont, and Harvard announced that they, too, had turned life on Earth upside-down. Their discovery wasn’t quite so dramatic at first glance. Any regular person peering through a microscope at their creation would see little more than a few globs of dirty pond water in a petri dish. But those globs were alive; in fact, they were alive in a way that nothing has ever been alive before, in an uncharted space between biology and technology. They called them Xenobots, the world’s first living robot—the world’s first programmable organism.

Xenobot: Xeno as in Xenopus laevis, a voracious frog native to the wetlands of Sub-Saharan Africa; bot, of course, as in robot. It’s an unconventional name for an unconventional organism, so novel that even its makers struggle to conceptualize it. “The terminology that has served us well for many years is just not any good anymore,” concedes Michael Levin, the team’s iconoclastic biologist. His collaborator Josh Bongard, a computer scientist and robotics expert, has called Xenobots “novel living machines.” Sam Kriegman—the team’s postdoc—prefers the term “Computer Designed Organism,” although he’s been trying on “living deepfake” for size recently.

And they’re all right, in a way. Xenobots are deepfakes in the sense that they aren’t what they seem. They’re robots in the sense that they’re autonomous, programmable agents. They’re Computer Designed in the sense that their morphology—the form their tiny bodies take—was designed by an evolutionary computer algorithm in Bongard’s UVM lab. They’re living in the sense that they’re made of embryonic frog cells, and they’re machines in the sense that humans are machines: biological mechanisms made up of constituent parts.

Xenobots are the first living creatures whose immediate evolution occurred inside a computer and not in the biosphere. The result is a simple organism. Xenobots have no brains; the shape of their bodies is what determines how they behave. And yet, Levin and Bongard do not fully understand why Xenobots behave the way they do. “What you’re seeing de novois a completely novel creature with new proto-cognitive capacities, preferences, capabilities, IQ,” Levin explains. “All of those things appear out of nowhere.” Sometimes a Xenobot will head in one direction and then abruptly double back, as though changing its mind. What force guides such behaviors? Can a frog’s cells, in some way, think? Xenobots seem to have “nano free-will,” Levin jokes.

And this is where the can of worms—or tadpoles, maybe—pops open…

The word “robot” recently celebrated its centennial. It comes from the Czech playwright Karel Čapek’s 1920 play “Rossum’s Universal Robots,” about a worker uprising in a robot factory. Čapek’s robots are biological, the result of a vaguely alchemical process involving “albumen” with a “raging thirst for life.” Our conception of a robot as being something metallic, with clanging gears and servo-motors, is more recent baggage, a consequence of the science-fiction stories and films of the mid-twentieth century. In order to understand what Xenobots might mean for our future, we’ll have to divest ourselves from the idea that a robot—or any kind of autonomous being—can be wholly defined by its materiality…

As Norbert Weiner, the father of cybernetics, observed: “Let us remember that the automatic machine is the precise economic equivalent of slave labor. Any labor which competes with slave labor must accept the economic consequences of slave labor.”

Claire Evans (@TheUniverse) explains how “Xenobots may change how we think about intelligence.”

For apposite background, see also “The Link Between Bioelectricity and Consciousness.”

Karel Čapek, R.U.R. (Rossumovi Univerzální Roboti, or in English, Rossum’s Universal Robots)

###

As we ponder life itself, we might recall (with, perhaps, a touch of nostalgia) that it was on this date in 1959 that Texas Instruments (TI) demonstrated the first working integrated circuit (IC), which had been invented by Jack Kilby. Kilby created the device to prove that resistors and capacitors could exist on the same piece of semiconductor material. His circuit consisted of a sliver of germanium with five components linked by wires. It was Fairchild’s Robert Noyce, however, who filed for a patent within months of Kilby and who made the IC a commercially-viable technology. Both men are credited as co-inventors of the IC. (Kilby won the Nobel Prize for his work in 2000; Noyce, who died in 1990, did not share.)

source

Written by (Roughly) Daily

March 24, 2021 at 1:01 am

“Please cut off a nanosecond and send it over to me”*…

960px-Commodore_Grace_M._Hopper,_USN_(covered) “Amazing  Grace” Hopper, seminal computer scientist and Rear Admiral in the U.S. Navy, explains a “nanosecond”… * Grace Hopper

###

As we celebrate clarity, we might recall that it was on this date in 1961 that Robert Noyce (then at Fairchild Semiconductor) was issued patent number 2981877 for his “semiconductor device-and-lead structure,” the first patent for what would come to be known as the integrated circuit.  In fact another engineer, Jack Kilby, had separately and essentially simultaneously developed the same technology (Kilby’s design was rooted in germanium; Noyce’s in silicon) and had filed a few months earlier than Noyce… a fact that was recognized in 2000 when Kilby was Awarded the Nobel Prize– in which Noyce, who had died in 1990, did not share.

Noyce (left) and Kilby (right)

 source

Written by (Roughly) Daily

April 25, 2019 at 1:01 am