Bits of Books - Books by Title
Samuel Morse, trying to invent a way to transmit data, started with idea of sending numbers, with each number representing a word in a look up book. But after wasting a lot of time compiling an early version, he realized it would become too cumbersome (although this turned out to be the only practicable way to transmit Chinese language telegraphy). Plan B was using dots and dashes to represent letters. But in those days nobody had studied letter frequency, so Morse resorted to visiting local printer to look at the letters they used. Found they had 12000 E's, 9000 T's and only 100 Z's, so he arranged his alphabet accordingly, with the most common letters represented by shorter sequences. They had originally assigned dash-dash-dot to represent T, but changed it to a single dash, thus saving telegraph operators billions of key taps in years to come. A little quantitative research beat uninformed intuition.
More books on Inventions
African languages are tonal, so a word lisaka could mean a puddle, a promise or a poison, depending on which syllables stressed. So when sending messages by drumming, lot of potential ambiguity. To get round that, added a lot of what seemed like redundancy: so 'the moon' is rendered as 'the moon looks down on the earth'. The extra drumbeats provide context. As message starts, each ambiguous word begins as a cloud of possible interpretations, then the alternative possibilities evaporate.
Redundancy makes it possible for us to understand conversations at a noisy party. The lack of context causes the confusions and misunderstandings that arise from text messages.
Irony that in Africa the drums have almost died out as modern generations adopt mobile phones.
The puzzle of information packing—how to cram knowledge into the tiniest space possible—has fueled technological development at least since the emergence of Chinese letters 8,000 years ago. In his new book, The Information, journalist James Gleick argues that information is “the blood and the fuel, the vital principle” of our lives. Delving deep into the history behind today’s data-driven world, Gleick explores the mysterious drumming language of the African talking drum, whose irregular rhythms carried messages through the jungles of the Congo. He considers musical compositions like Johann Sebastian Bach’s 18th-century “Well-Tempered Clavier” as data streams that could capture sounds as varied as wind, cricket chirps, or the clatter of a horse-drawn cart. But for Gleick the pivotal moment initiating our data-drenched era came in 1948, when mathematician Claude Shannon conceived of the bit as a unit of information. Shannon’s work propelled us headlong into the flood of blogs, emails, tweets, and news updates that shape our lives today.
In 1948 the Bell Telephone Laboratories announced the invention of a tiny electronic semiconductor, “an amazingly simple device” that could do anything a vacuum tube could do and more efficiently. It was a crystalline sliver, so small that 100 would fit in the palm of a hand. In May scientists formed a committee to come up with a name. Transistor won out. “It may have far-reaching significance in electronics and electrical communication,” Bell Labs declared in a press release, and for once the reality surpassed the hype. The transistor sparked the revolution in electronics, setting the technology on its path of miniaturization and ubiquity. But it was only the second-most significant development of that year. The transistor was only hardware.
An invention even more profound and more fundamental came in a monograph spread across 79 pages of The Bell System Technical
Journal in July and October. No one bothered with a press release. It carried a title both simple and grand—“A Mathematical Theory of Communication”—and the message was hard to summarize. But it was a fulcrum around which the world began to turn. Like the transistor, this development also involved a neologism: the word bit, chosen
in this case not by a committee but by the lone author, a 32-year-old named Claude Shannon. The bit now joined the inch, the pound, the quart, and the minute as a determinate quantity—a fundamental unit of measure.
But measuring what? “A unit for measuring information,” Shannon wrote, as though there were such a thing, measurable and quantifiable information.
In 1949, when Claude Shannon took a sheet of paper and penciled his outline of the measures of information, the scale went from tens of bits to hundreds to thousands, millions, billions, and trillions. The transistor was one year old and Moore’s law yet to be conceived. At the top of his information pyramid was Shannon’s estimate for the Library of Congress—100 trillion bits, 1014. He was about right, but the pyramid was growing.
After bits came kilobits, naturally enough. After all, engineers had coined the word kilobuck—“a scientist’s idea of a short way to say ‘a thousand dollars,’ ” The New York Times helpfully explained in 1951. The measures of information climbed up an exponential scale, as the realization dawned in the 1960s that everything to do with information would now grow exponentially. That idea was casually expressed by Gordon Moore, who had been an undergraduate studying chemistry when Shannon jotted his note and found his way to electronic engineering and the development of integrated circuits. In 1965, three years before he founded the Intel Corporation, Moore was merely, modestly suggesting that within a decade, by 1975, we would be able to combine as many as 65,000 transistors on a single wafer of silicon. He predicted a doubling every year or two—a doubling of the number of components that could be packed on a chip, but then also, as it turned out, the doubling of all kinds of memory capacity and processing speed, a halving of size and cost, seemingly without end.
Everything may be recorded and preserved, at least potentially: every musical performance; every crime in a shop, elevator, or city street; every tsunami on the remotest shore.
Kilobits could be used to express speed of transmission as well as quantity of storage. As of 1972 businesses could lease high-speed lines carrying data as fast as 240 kilobits per second. Following the lead of IBM, whose hardware typically processed information in chunks of eight bits, engineers soon adopted the modern and slightly whimsical unit, the byte. Bits and bytes. A kilobyte, then, represented 8,000 bits; a megabyte (following hard upon), 8 million. In the order of things as worked out by international standards committees, mega- led to giga-, tera-, peta-, and exa-, drawn from Greek, though with less and less linguistic fidelity. That was enough, for everything measured, until 1991, when the need was seen for the zettabyte (1,000,000,000,000,000,000,000) and the inadvertently comic-sounding yottabyte (1,000,000,000,000,000,000,000,000). In this climb up the exponential ladder, information left other gauges behind. Money, for example, is scarce by comparison.
After kilobucks, there were megabucks and gigabucks, and people can joke about inflation leading to terabucks, but all the wealth amassed
by all the generations of humanity does not amount to a petabuck.
The 1970s were the decade of megabytes. In the summer of 1970, IBM introduced two new computer models with more memory than ever before: the Model 155, with 768,000 bytes of memory, and the larger Model 165, with a full megabyte, in a large cabinet. One of these room-filling mainframes could be purchased for $4,674,160.
By 1982 Prime Computer was marketing a megabyte of memory on a single circuit board, for $36,000. When the publishers of the Oxford English Dictionary began digitizing its contents in 1987 (120 typists; an IBM mainframe), they estimated its size at a gigabyte. A gigabyte also encompasses the entire human genome. A thousand of those would fill a terabyte. A terabyte was the amount of disk storage Larry Page and Sergey Brin managed to patch together with the help of $15,000 spread across their personal credit cards in 1998, when they were Stanford graduate students building a search-engine prototype, which they first called BackRub and then renamed Google. A terabyte is how much data a typical analog television station broadcasts daily, and it was the size of the United States government’s database of patent and trademark records when it went online in 1998. By 2010 one could buy a terabyte disk drive for a hundred dollars and hold it in the palm of one hand.
The books in the Library of Congress represent about 10 terabytes (as Shannon guessed), and the number is many times greater when images and recorded music are counted. The library now archives websites; by April 2011 it had collected 160 terabytes’ worth.
As the train hurtled onward, its passengers sometimes felt the pace foreshortening their sense of their own history. The computer scientist Jaron Lanier describes the feeling this way: “It’s as if you kneel to plant the seed of a tree and it grows so fast that it swallows your whole town before you can even rise to your feet.”
A more familiar metaphor is the cloud. All that information—all that information capacity—looms over us, not quite visible, not quite tangible, but awfully real; amorphous, spectral; hovering nearby, yet not situated in any one place. Heaven must once have felt this way to the faithful. People talk about shifting their lives to the cloud—their informational lives, at least. You may store photographs in the cloud; Google will manage your business in the cloud; Google is putting all the world’s books into the cloud; email passes to and from the cloud and never really leaves the cloud. All traditional ideas of privacy, based on doors and locks, physical remoteness and invisibility, are upended in the cloud.
Money lives in the cloud; the old forms are vestigial tokens of knowledge about who owns what, who owes what. To the 21st century these will be seen as anachronisms, quaint or even absurd: bullion carried from shore to shore in fragile ships, subject to the tariffs of pirates and the god Poseidon; metal coins tossed from moving cars into baskets at highway tollbooths and thereafter trucked about (now the history of your automobile is in the cloud); paper checks torn from pads and signed in ink; tickets for trains, performances, air travel, or anything at all, printed on weighty perforated paper with watermarks, holograms, or fluorescent fibers; and, soon enough, all forms of cash. The economy of the world is transacted in the cloud.
Its physical aspect could not be less cloudlike. Server farms proliferate in unmarked brick buildings and steel complexes, with smoked windows or no windows, miles of hollow floors, diesel generators, cooling towers, seven-foot intake fans, and aluminum chimney stacks. This hidden infrastructure grows in a symbiotic relationship with the electrical infrastructure it increasingly resembles. There are information switchers, control centers, and substations. They are clustered and distributed.
These are the wheelworks; the cloud is their avatar. The information produced and consumed by humankind used to vanish—that was the norm, the default. The sights, the sounds, the songs, the spoken word just melted away. Marks on stone, parchment, and paper were the special case. It did not occur to Sophocles’ audiences that it would be sad for his plays to be lost; they enjoyed the show. Now expectations have inverted. Everything may be recorded and preserved, at least potentially: every musical performance; every crime in a shop, elevator, or city street; every volcano or tsunami on the remotest shore; every card played or piece moved in an online game; every rugby scrum and cricket match.
Having a camera at hand is normal, not exceptional; something like 500 billion images were captured in 2010. YouTube was streaming more than a billion videos a day. Most of this is haphazard and unorganized, but there are extreme cases. The computer pioneer Gordon Bell, at Microsoft Research in his seventies, began recording every moment of his day, every conversation, message, document, a megabyte per hour or a gigabyte per month, wearing around his neck what he called a “SenseCam” to create what he called a “LifeLog.” Where does it end? Not with the Library of Congress.
It is finally natural—even inevitable—to ask how much information is in the universe. It is the consequence of Charles Babbage and Edgar Allan Poe saying, “No thought can perish.” Seth Lloyd does the math. He is a moonfaced, bespectacled quantum engineer at MIT, a theorist and designer of quantum computers. The universe, by existing, registers information, he says. By evolving in time, it processes information. How much? To figure that out, Lloyd takes into account how fast this “computer” works and how long it has been working. Considering the fundamental limit on speed and on memory space, along with the speed of light and the age of the universe since the Big Bang, Lloyd calculates
that the universe can have performed something on the order of 10120 “ops” in its entire history. Considering “every degree of freedom of every particle in the universe,” it could now hold something like 1090 bits. And counting.
Books by Title
Books by Author
Books by Topic
Bits of Books To Impress