2 results back to index
In the Plex: How Google Thinks, Works, and Shapes Our Lives by Steven Levy
23andMe, AltaVista, Anne Wojcicki, Apple's 1984 Super Bowl advert, autonomous vehicles, book scanning, Brewster Kahle, Burning Man, business process, clean water, cloud computing, crowdsourcing, Dean Kamen, discounted cash flows, don't be evil, Donald Knuth, Douglas Engelbart, Douglas Engelbart, El Camino Real, fault tolerance, Firefox, Gerard Salton, Gerard Salton, Google bus, Google Chrome, Google Earth, Googley, HyperCard, hypertext link, IBM and the Holocaust, informal economy, information retrieval, Internet Archive, Jeff Bezos, John Markoff, Kevin Kelly, Mark Zuckerberg, Menlo Park, one-China policy, optical character recognition, PageRank, Paul Buchheit, Potemkin village, prediction markets, recommendation engine, risk tolerance, Rubik’s Cube, Sand Hill Road, Saturday Night Live, search inside the book, second-price auction, selection bias, Silicon Valley, skunkworks, Skype, slashdot, social graph, social software, social web, spectrum auction, speech recognition, statistical model, Steve Ballmer, Steve Jobs, Steven Levy, Ted Nelson, telemarketer, trade route, traveling salesman, turn-by-turn navigation, Vannevar Bush, web application, WikiLeaks, Y Combinator
When DEC opened it to outsiders on December 15, 1995, nearly 300,000 people tried it out. They were dazzled. AltaVista’s actual search quality techniques—what determined the ranking of results—were based on traditional information retrieval (IR) algorithms. Many of those algorithms arose from the work of one man, a refugee from Nazi Germany named Gerard Salton, who had come to America, got a PhD at Harvard, and moved to Cornell University, where he cofounded its computer science department. Searching through databases using the same commands you’d use with a human—“natural language” became the term of art—was Salton’s specialty. During the 1960s, Salton developed a system that was to become a model for information retrieval.
Born in the Indian state of Uttar Pradesh, in the foothills of the Himalayas, Singhal had arrived in the United States in 1992 to pursue a master’s degree in computer science at the University of Minnesota. He’d become fascinated with the field then known as information retrieval and was desperate to study with its pioneering innovator, Gerard Salton. “I only applied to one grad school, and it was Cornell,” he says. “And I wrote in my statement of purpose that if I was ever going to get a PhD, it’s with Gerry Salton. Otherwise, I didn’t think a PhD was worth it.” He became Salton’s assistant, got his PhD at Cornell, and eventually wound up at AT&T Labs.
The Art of Computer Programming: Fundamental Algorithms by Donald E. Knuth
discrete time, distributed generation, Donald Knuth, fear of failure, Fermat's Last Theorem, Gerard Salton, Isaac Newton, Jacquard loom, Jacquard loom, John von Neumann, linear programming, linked data, Menlo Park, probability theory / Blaise Pascal / Pierre de Fermat, Richard Feynman, sorting algorithm, stochastic process, Turing machine