Coders at Work: Reflections on the Craft of Programming


Peter Seibel - 2009
    As the words "at work" suggest, Peter Seibel focuses on how his interviewees tackle the day–to–day work of programming, while revealing much more, like how they became great programmers, how they recognize programming talent in others, and what kinds of problems they find most interesting. Hundreds of people have suggested names of programmers to interview on the Coders at Work web site: http://www.codersatwork.com. The complete list was 284 names. Having digested everyone’s feedback, we selected 16 folks who’ve been kind enough to agree to be interviewed:- Frances Allen: Pioneer in optimizing compilers, first woman to win the Turing Award (2006) and first female IBM fellow- Joe Armstrong: Inventor of Erlang- Joshua Bloch: Author of the Java collections framework, now at Google- Bernie Cosell: One of the main software guys behind the original ARPANET IMPs and a master debugger- Douglas Crockford: JSON founder, JavaScript architect at Yahoo!- L. Peter Deutsch: Author of Ghostscript, implementer of Smalltalk-80 at Xerox PARC and Lisp 1.5 on PDP-1- Brendan Eich: Inventor of JavaScript, CTO of the Mozilla Corporation - Brad Fitzpatrick: Writer of LiveJournal, OpenID, memcached, and Perlbal - Dan Ingalls: Smalltalk implementor and designer- Simon Peyton Jones: Coinventor of Haskell and lead designer of Glasgow Haskell Compiler- Donald Knuth: Author of The Art of Computer Programming and creator of TeX- Peter Norvig: Director of Research at Google and author of the standard text on AI- Guy Steele: Coinventor of Scheme and part of the Common Lisp Gang of Five, currently working on Fortress- Ken Thompson: Inventor of UNIX- Jamie Zawinski: Author of XEmacs and early Netscape/Mozilla hackerWhat you’ll learn:How the best programmers in the world do their jobWho is this book for?Programmers interested in the point of view of leaders in the field. Programmers looking for approaches that work for some of these outstanding programmers.

The Computer and the Brain


John von Neumann - 1958
    This work represents the views of a mathematician on the analogies between computing machines and the living human brain.

Pattern Classification


David G. Stork - 1973
    Now with the second edition, readers will find information on key new topics such as neural networks and statistical pattern recognition, the theory of machine learning, and the theory of invariances. Also included are worked examples, comparisons between different methods, extensive graphics, expanded exercises and computer project topics.An Instructor's Manual presenting detailed solutions to all the problems in the book is available from the Wiley editorial department.

The Formula: How Algorithms Solve all our Problems … and Create More


Luke Dormehl - 2014
    What if everything in life could be reduced to a simple formula? What if numbers were able to tell us which partners we were best matched with – not just in terms of attractiveness, but for a long-term committed marriage? Or if they could say which films would be the biggest hits at the box office, and what changes could be made to those films to make them even more successful? Or even who out of us is likely to commit certain crimes, and when? This may sound like the world of science-fiction, but in fact it is just the tip of the iceberg in a world that is increasingly ruled by complex algorithms and neural networks.In The Formula, Luke Dormehl takes you inside the world of numbers, asking how we came to believe in the all-conquering power of algorithms; introducing the mathematicians, artificial intelligence experts and Silicon Valley entrepreneurs who are shaping this brave new world, and ultimately asking how we survive in an era where numbers can sometimes seem to create as many problems as they solve.

Machine Learning for Absolute Beginners


Oliver Theobald - 2017
    The manner in which computers are now able to mimic human thinking is rapidly exceeding human capabilities in everything from chess to picking the winner of a song contest. In the age of machine learning, computers do not strictly need to receive an ‘input command’ to perform a task, but rather ‘input data’. From the input of data they are able to form their own decisions and take actions virtually as a human would. But as a machine, can consider many more scenarios and execute calculations to solve complex problems. This is the element that excites companies and budding machine learning engineers the most. The ability to solve complex problems never before attempted. This is also perhaps one reason why you are looking at purchasing this book, to gain a beginner's introduction to machine learning. This book provides a plain English introduction to the following topics: - Artificial Intelligence - Big Data - Downloading Free Datasets - Regression - Support Vector Machine Algorithms - Deep Learning/Neural Networks - Data Reduction - Clustering - Association Analysis - Decision Trees - Recommenders - Machine Learning Careers This book has recently been updated following feedback from readers. Version II now includes: - New Chapter: Decision Trees - Cleanup of minor errors

The Fourth Paradigm: Data-Intensive Scientific Discovery


Tony Hey - 2009
    Increasingly, scientific breakthroughs will be powered by advanced computing capabilities that help researchers manipulate and explore massive datasets. The speed at which any given scientific discipline advances will depend on how well its researchers collaborate with one another, and with technologists, in areas of eScience such as databases, workflow management, visualization, and cloud-computing technologies. This collection of essays expands on the vision of pioneering computer scientist Jim Gray for a new, fourth paradigm of discovery based on data-intensive science and offers insights into how it can be fully realized.

Time Reborn: From the Crisis in Physics to the Future of the Universe


Lee Smolin - 2013
    You experience it passing every day when you watch clocks tick, bread toast, and children grow. But most physicists see things differently, from Newton to Einstein to today’s quantum theorists. For them, time isn’t real. You may think you experience time passing, but they say it’s just an illusion.Lee Smolin, author of the controversial bestseller The Trouble with Physics, argues this limited notion of time is holding physics back. It’s time for a major revolution in scientific thought. The reality of time could be the key to the next big breakthrough in theoretical physics.What if the laws of physics themselves were not timeless? What if they could evolve? Time Reborn offers a radical new approach to cosmology that embraces the reality of time and opens up a whole new universe of possibilties. There are few ideas that, like our notion of time, shape our thinking about literally everything, with major implications for physics and beyond—from climate change to the economic crisis. Smolin explains in lively and lucid prose how the true nature of time impacts our world.

From Mathematics to Generic Programming


Alexander A. Stepanov - 2014
    If you're a reasonably proficient programmer who can think logically, you have all the background you'll need. Stepanov and Rose introduce the relevant abstract algebra and number theory with exceptional clarity. They carefully explain the problems mathematicians first needed to solve, and then show how these mathematical solutions translate to generic programming and the creation of more effective and elegant code. To demonstrate the crucial role these mathematical principles play in many modern applications, the authors show how to use these results and generalized algorithms to implement a real-world public-key cryptosystem. As you read this book, you'll master the thought processes necessary for effective programming and learn how to generalize narrowly conceived algorithms to widen their usefulness without losing efficiency. You'll also gain deep insight into the value of mathematics to programming--insight that will prove invaluable no matter what programming languages and paradigms you use. You will learn aboutHow to generalize a four thousand-year-old algorithm, demonstrating indispensable lessons about clarity and efficiencyAncient paradoxes, beautiful theorems, and the productive tension between continuous and discreteA simple algorithm for finding greatest common divisor (GCD) and modern abstractions that build on itPowerful mathematical approaches to abstractionHow abstract algebra provides the idea at the heart of generic programmingAxioms, proofs, theories, and models: using mathematical techniques to organize knowledge about your algorithms and data structuresSurprising subtleties of simple programming tasks and what you can learn from themHow practical implementations can exploit theoretical knowledge

The Nothing That Is: A Natural History of Zero


Robert M. Kaplan - 1999
    As we enter the year 2000, zero is once again making its presence felt. Nothing itself, it makes possible a myriad of calculations. Indeed, without zero mathematicsas we know it would not exist. And without mathematics our understanding of the universe would be vastly impoverished. But where did this nothing, this hollow circle, come from? Who created it? And what, exactly, does it mean? Robert Kaplan's The Nothing That Is: A Natural History of Zero begins as a mystery story, taking us back to Sumerian times, and then to Greece and India, piecing together the way the idea of a symbol for nothing evolved. Kaplan shows us just how handicapped our ancestors were in trying to figurelarge sums without the aid of the zero. (Try multiplying CLXIV by XXIV). Remarkably, even the Greeks, mathematically brilliant as they were, didn't have a zero--or did they? We follow the trail to the East where, a millennium or two ago, Indian mathematicians took another crucial step. By treatingzero for the first time like any other number, instead of a unique symbol, they allowed huge new leaps forward in computation, and also in our understanding of how mathematics itself works. In the Middle Ages, this mathematical knowledge swept across western Europe via Arab traders. At first it was called dangerous Saracen magic and considered the Devil's work, but it wasn't long before merchants and bankers saw how handy this magic was, and used it to develop tools likedouble-entry bookkeeping. Zero quickly became an essential part of increasingly sophisticated equations, and with the invention of calculus, one could say it was a linchpin of the scientific revolution. And now even deeper layers of this thing that is nothing are coming to light: our computers speakonly in zeros and ones, and modern mathematics shows that zero alone can be made to generate everything.Robert Kaplan serves up all this history with immense zest and humor; his writing is full of anecdotes and asides, and quotations from Shakespeare to Wallace Stevens extend the book's context far beyond the scope of scientific specialists. For Kaplan, the history of zero is a lens for looking notonly into the evolution of mathematics but into very nature of human thought. He points out how the history of mathematics is a process of recursive abstraction: how once a symbol is created to represent an idea, that symbol itself gives rise to new operations that in turn lead to new ideas. Thebeauty of mathematics is that even though we invent it, we seem to be discovering something that already exists.The joy of that discovery shines from Kaplan's pages, as he ranges from Archimedes to Einstein, making fascinating connections between mathematical insights from every age and culture. A tour de force of science history, The Nothing That Is takes us through the hollow circle that leads to infinity.

Fluent Python: Clear, Concise, and Effective Programming


Luciano Ramalho - 2015
    With this hands-on guide, you'll learn how to write effective, idiomatic Python code by leveraging its best and possibly most neglected features. Author Luciano Ramalho takes you through Python's core language features and libraries, and shows you how to make your code shorter, faster, and more readable at the same time.Many experienced programmers try to bend Python to fit patterns they learned from other languages, and never discover Python features outside of their experience. With this book, those Python programmers will thoroughly learn how to become proficient in Python 3.This book covers:Python data model: understand how special methods are the key to the consistent behavior of objectsData structures: take full advantage of built-in types, and understand the text vs bytes duality in the Unicode ageFunctions as objects: view Python functions as first-class objects, and understand how this affects popular design patternsObject-oriented idioms: build classes by learning about references, mutability, interfaces, operator overloading, and multiple inheritanceControl flow: leverage context managers, generators, coroutines, and concurrency with the concurrent.futures and asyncio packagesMetaprogramming: understand how properties, attribute descriptors, class decorators, and metaclasses work"

The Essential Turing: Seminal Writings in Computing, Logic, Philosophy, Artificial Intelligence, and Artificial Life Plus the Secrets of Enigma


Alan Turing - 2004
    In 1935, aged 22, he developed the mathematical theory upon which all subsequent stored-program digital computers are modeled.At the outbreak of hostilities with Germany in September 1939, he joined the Government Codebreaking team at Bletchley Park, Buckinghamshire and played a crucial role in deciphering Engima, the code used by the German armed forces to protect their radio communications. Turing's work on the versionof Enigma used by the German navy was vital to the battle for supremacy in the North Atlantic. He also contributed to the attack on the cyphers known as Fish, which were used by the German High Command for the encryption of signals during the latter part of the war. His contribution helped toshorten the war in Europe by an estimated two years.After the war, his theoretical work led to the development of Britain's first computers at the National Physical Laboratory and the Royal Society Computing Machine Laboratory at Manchester University.Turing was also a founding father of modern cognitive science, theorizing that the cortex at birth is an unorganized machine which through training becomes organized into a universal machine or something like it. He went on to develop the use of computers to model biological growth, launchingthe discipline now referred to as Artificial Life.The papers in this book are the key works for understanding Turing's phenomenal contribution across all these fields. The collection includes Turing's declassified wartime Treatise on the Enigma; letters from Turing to Churchill and to codebreakers; lectures, papers, and broadcasts which opened upthe concept of AI and its implications; and the paper which formed the genesis of the investigation of Artifical Life.

Cybernetics: or the Control and Communication in the Animal and the Machine


Norbert Wiener - 1948
    It is a ‘ must’ book for those in every branch of science . . . in addition, economists, politicians, statesmen, and businessmen cannot afford to overlook cybernetics and its tremendous, even terrifying implications. "It is a beautifully written book, lucid, direct, and despite its complexity, as readable by the layman as the trained scientist." -- John B. Thurston, "The Saturday Review of Literature" Acclaimed one of the "seminal books . . . comparable in ultimate importance to . . . Galileo or Malthus or Rousseau or Mill," "Cybernetics" was judged by twenty-seven historians, economists, educators, and philosophers to be one of those books published during the "past four decades", which may have a substantial impact on public thought and action in the years ahead." -- Saturday Review

Secrets of Mental Math: The Mathemagician's Guide to Lightning Calculation and Amazing Math Tricks


Arthur T. Benjamin - 1993
    Get ready to amaze your friends—and yourself—with incredible calculations you never thought you could master, as renowned “mathemagician” Arthur Benjamin shares his techniques for lightning-quick calculations and amazing number tricks. This book will teach you to do math in your head faster than you ever thought possible, dramatically improve your memory for numbers, and—maybe for the first time—make mathematics fun.Yes, even you can learn to do seemingly complex equations in your head; all you need to learn are a few tricks. You’ll be able to quickly multiply and divide triple digits, compute with fractions, and determine squares, cubes, and roots without blinking an eye. No matter what your age or current math ability, Secrets of Mental Math will allow you to perform fantastic feats of the mind effortlessly. This is the math they never taught you in school.Also available as an eBook

Introduction to Information Retrieval


Christopher D. Manning - 2008
    Written from a computer science perspective by three leading experts in the field, it gives an up-to-date treatment of all aspects of the design and implementation of systems for gathering, indexing, and searching documents; methods for evaluating systems; and an introduction to the use of machine learning methods on text collections. All the important ideas are explained using examples and figures, making it perfect for introductory courses in information retrieval for advanced undergraduates and graduate students in computer science. Based on feedback from extensive classroom experience, the book has been carefully structured in order to make teaching more natural and effective. Although originally designed as the primary text for a graduate or advanced undergraduate course in information retrieval, the book will also create a buzz for researchers and professionals alike.

One, Two, Three...Infinity: Facts and Speculations of Science


George Gamow - 1947
    . . full of intellectual treats and tricks, of whimsy and deep scientific philosophy. It is highbrow entertainment at its best, a teasing challenge to all who aspire to think about the universe." — New York Herald TribuneOne of the world's foremost nuclear physicists (celebrated for his theory of radioactive decay, among other accomplishments), George Gamow possessed the unique ability of making the world of science accessible to the general reader.He brings that ability to bear in this delightful expedition through the problems, pleasures, and puzzles of modern science. Among the topics scrutinized with the author's celebrated good humor and pedagogical prowess are the macrocosm and the microcosm, theory of numbers, relativity of space and time, entropy, genes, atomic structure, nuclear fission, and the origin of the solar system.In the pages of this book readers grapple with such crucial matters as whether it is possible to bend space, why a rocket shrinks, the "end of the world problem," excursions into the fourth dimension, and a host of other tantalizing topics for the scientifically curious. Brimming with amusing anecdotes and provocative problems, One Two Three . . . Infinity also includes over 120 delightful pen-and-ink illustrations by the author, adding another dimension of good-natured charm to these wide-ranging explorations.Whatever your level of scientific expertise, chances are you'll derive a great deal of pleasure, stimulation, and information from this unusual and imaginative book. It belongs in the library of anyone curious about the wonders of the scientific universe. "In One Two Three . . . Infinity, as in his other books, George Gamow succeeds where others fail because of his remarkable ability to combine technical accuracy, choice of material, dignity of expression, and readability." — Saturday Review of Literature