Surfing Through Hyperspace: Understanding Higher Universes in Six Easy Lessons


Clifford A. Pickover - 1999
    With this astonishing guidebook, Surfing Through Hyperspace, you need not be a mathematician or an astrophysicist to explore the all-but-unfathomable concepts of hyperspace and higher-dimensional geometry.No subject in mathematics has intrigued both children and adults as much as the idea of a fourth dimension. Philosophers and parapsychologists have meditated on this mysterious space that no one can point to but may be all around us. Yet this extra dimension has a very real, practical value to mathematicians and physicists who use it every day in their calculations. In the tradtion of Flatland, and with an infectious enthusiasm, Clifford Pickover tackles the problems inherent in our 3-D brains trying to visualize a 4-D world, muses on the religious implications of the existence of higher-dimensional consciousness, and urges all curious readers to venture into the unexplored territory lying beyond the prison of the obvious. Pickover alternates sections that explain the science of hyperspace with sections that dramatize mind-expanding concepts through a fictional dialogue between two futuristic FBI agents who dabble in the fourth dimension as a matter of national security. This highly accessible and entertaining approach turns an intimidating subject into a scientific game open to all dreamers.Surfing Through Hyperspace concludes with a number of puzzles, computer experiments and formulas for further exploration, inviting readers to extend their minds across this inexhaustibly intriguing scientific terrain.

Statistics: A Very Short Introduction


David J. Hand - 2008
    From randomized clinical trials in medical research, to statistical models of risk in banking and hedge fund industries, to the statistical tools used to probe vast astronomical databases, the field of statistics has become centrally important to how we understand our world. But the discipline underlying all these is not the dull statistics of the popular imagination. Long gone are the days of manual arithmetic manipulation. Nowadays statistics is a dynamic discipline, revolutionized by the computer, which uses advanced software tools to probe numerical data, seeking structures, patterns, and relationships. This Very Short Introduction sets the study of statistics in context, describing its history and giving examples of its impact, summarizes methods of gathering and evaluating data, and explains the role played by the science of chance, of probability, in statistical methods. The book also explores deep philosophical issues of induction--how we use statistics to discern the true nature of reality from the limited observations we necessarily must make.About the Series: Combining authority with wit, accessibility, and style, Very Short Introductions offer an introduction to some of life's most interesting topics. Written by experts for the newcomer, they demonstrate the finest contemporary thinking about the central problems and issues in hundreds of key topics, from philosophy to Freud, quantum theory to Islam.

Introduction to Graph Theory


Douglas B. West - 1995
    Verification that algorithms work is emphasized more than their complexity. An effective use of examples, and huge number of interesting exercises, demonstrate the topics of trees and distance, matchings and factors, connectivity and paths, graph coloring, edges and cycles, and planar graphs. For those who need to learn to make coherent arguments in the fields of mathematics and computer science.

A Short Account of the History of Mathematics


W.W. Rouse Ball - 1900
    From the early Greek influences to the Middle Ages and the Renaissance to the end of the 19th century, trace the fascinating foundation of mathematics as it developed through the ages. Aristotle, Galileo, Kepler, Newton: you know the names. Now here's what they really did, and the effect their discoveries had on our culture, all explained in a way the layperson can understand. Begin with the basis of arithmetic (Plato and the introduction of geometry), and discover why the use of Arabic numerals was critical to the development of both commerce and science. The development of calculus made space travel a reality, while the abacus prefigured the computer. The greats examined in depth include Leonardo da Vinci, a brilliant mathematician as well as artist; Pascal, who laid out the theory of probabilities; and Fermat, whose intriguing theory has only recently been solved.

Introductory Statistics


Neil A. Weiss - 1987
    This book develops statistical thinking over rote drill and practice. The Nature of Statistics; Organizing Data; Descriptive Measures; Probability Concepts; Discrete Random Variables; The Normal Distribution; The Sampling Distribution of the Sample Menu; Confidence Intervals for One Population Mean; Hypothesis Tests for One Population Mean; Inferences for Two Population Means; Inferences for Population Standard Deviations; Inferences for Population Proportions; Chi-Square Procedures; Descriptive Methods in Regression and Correlation; Inferential Methods in Regression and Correlation; Analysis of Variance (ANOVA) For all readers interested in Introductory Statistics.

The Elements of Statistical Learning: Data Mining, Inference, and Prediction


Trevor Hastie - 2001
    With it has come vast amounts of data in a variety of fields such as medicine, biology, finance, and marketing. The challenge of understanding these data has led to the development of new tools in the field of statistics, and spawned new areas such as data mining, machine learning, and bioinformatics. Many of these tools have common underpinnings but are often expressed with different terminology. This book describes the important ideas in these areas in a common conceptual framework. While the approach is statistical, the emphasis is on concepts rather than mathematics. Many examples are given, with a liberal use of color graphics. It should be a valuable resource for statisticians and anyone interested in data mining in science or industry. The book's coverage is broad, from supervised learning (prediction) to unsupervised learning. The many topics include neural networks, support vector machines, classification trees and boosting—the first comprehensive treatment of this topic in any book. Trevor Hastie, Robert Tibshirani, and Jerome Friedman are professors of statistics at Stanford University. They are prominent researchers in this area: Hastie and Tibshirani developed generalized additive models and wrote a popular book of that title. Hastie wrote much of the statistical modeling software in S-PLUS and invented principal curves and surfaces. Tibshirani proposed the Lasso and is co-author of the very successful An Introduction to the Bootstrap. Friedman is the co-inventor of many data-mining tools including CART, MARS, and projection pursuit.

The Miracle (Storycuts)


Frederick Forsyth - 2000
    When they are waylaid they end up in a quiet courtyard being told a mesmerising story by a strange young man. It is an unforgettable tale that begins with the persecution of a young nun in the turbulent days of the sixteenth century and culminates in the bitter German retreat from Italy. Part of the Storycuts series, this short story was previously published in the collection The Veteran.

Grokking Algorithms An Illustrated Guide For Programmers and Other Curious People


Aditya Y. Bhargava - 2015
    The algorithms you'll use most often as a programmer have already been discovered, tested, and proven. If you want to take a hard pass on Knuth's brilliant but impenetrable theories and the dense multi-page proofs you'll find in most textbooks, this is the book for you. This fully-illustrated and engaging guide makes it easy for you to learn how to use algorithms effectively in your own programs.Grokking Algorithms is a disarming take on a core computer science topic. In it, you'll learn how to apply common algorithms to the practical problems you face in day-to-day life as a programmer. You'll start with problems like sorting and searching. As you build up your skills in thinking algorithmically, you'll tackle more complex concerns such as data compression or artificial intelligence. Whether you're writing business software, video games, mobile apps, or system utilities, you'll learn algorithmic techniques for solving problems that you thought were out of your grasp. For example, you'll be able to:Write a spell checker using graph algorithmsUnderstand how data compression works using Huffman codingIdentify problems that take too long to solve with naive algorithms, and attack them with algorithms that give you an approximate answer insteadEach carefully-presented example includes helpful diagrams and fully-annotated code samples in Python. By the end of this book, you will know some of the most widely applicable algorithms as well as how and when to use them.

Applied Cryptography: Protocols, Algorithms, and Source Code in C


Bruce Schneier - 1993
    … The book the National Security Agency wanted never to be published." –Wired Magazine "…monumental… fascinating… comprehensive… the definitive work on cryptography for computer programmers…" –Dr. Dobb's Journal"…easily ranks as one of the most authoritative in its field." —PC Magazine"…the bible of code hackers." –The Millennium Whole Earth CatalogThis new edition of the cryptography classic provides you with a comprehensive survey of modern cryptography. The book details how programmers and electronic communications professionals can use cryptography—the technique of enciphering and deciphering messages-to maintain the privacy of computer data. It describes dozens of cryptography algorithms, gives practical advice on how to implement them into cryptographic software, and shows how they can be used to solve security problems. Covering the latest developments in practical cryptographic techniques, this new edition shows programmers who design computer applications, networks, and storage systems how they can build security into their software and systems. What's new in the Second Edition? * New information on the Clipper Chip, including ways to defeat the key escrow mechanism * New encryption algorithms, including algorithms from the former Soviet Union and South Africa, and the RC4 stream cipher * The latest protocols for digital signatures, authentication, secure elections, digital cash, and more * More detailed information on key management and cryptographic implementations

Statistics for Dummies


Deborah J. Rumsey - 2003
    . ." and "The data bear this out. . . ." But the field of statistics is not just about data. Statistics is the entire process involved in gathering evidence to answer questions about the world, in cases where that evidence happens to be numerical data. Statistics For Dummies is for everyone who wants to sort through and evaluate the incredible amount of statistical information that comes to them on a daily basis. (You know the stuff: charts, graphs, tables, as well as headlines that talk about the results of the latest poll, survey, experiment, or other scientific study.) This book arms you with the ability to decipher and make important decisions about statistical results, being ever aware of the ways in which people can mislead you with statistics. Get the inside scoop on number-crunching nuances, plus insight into how you canDetermine the odds Calculate a standard score Find the margin of error Recognize the impact of polls Establish criteria for a good survey Make informed decisions about experiments This down-to-earth reference is chock-full of real examples from real sources that are relevant to your everyday life: from the latest medical breakthroughs, crime studies, and population trends to surveys on Internet dating, cell phone use, and the worst cars of the millennium. Statistics For Dummies departs from traditional statistics texts, references, supplement books, and study guides in the following ways:Practical and intuitive explanations of statistical concepts, ideas, techniques, formulas, and calculations. Clear and concise step-by-step procedures that intuitively explain how to work through statistics problems. Upfront and honest answers to your questions like, "What does this really mean?" and "When and how I will ever use this?" Chances are, Statistics For Dummies will be your No. 1 resource for discovering how numerical data figures into your corner of the universe.

Concrete Mathematics: A Foundation for Computer Science


Ronald L. Graham - 1988
    "More concretely," the authors explain, "it is the controlled manipulation of mathematical formulas, using a collection of techniques for solving problems."

Introducing Mathematics: A Graphic Guide


Ziauddin Sardar - 1995
    It affects us all. We depend on it in our daily lives, and yet many of the tools of mathematics, such as geometry, algebra and trigonometry, are descended from ancient or non-Western civilizations.Introducing Mathematics traces the story of mathematics from the ancient world to modern times, describing the great discoveries and providing an accessible introduction to such topics as number-systems, geometry and algebra, the calculus, the theory of the infinite, statistical reasoning and chaos theory. It shows how the history of mathematics has seen progress and paradox go hand in hand - and how this is still happening today.

Information Theory, Inference and Learning Algorithms


David J.C. MacKay - 2002
    These topics lie at the heart of many exciting areas of contemporary science and engineering - communication, signal processing, data mining, machine learning, pattern recognition, computational neuroscience, bioinformatics, and cryptography. This textbook introduces theory in tandem with applications. Information theory is taught alongside practical communication systems, such as arithmetic coding for data compression and sparse-graph codes for error-correction. A toolbox of inference techniques, including message-passing algorithms, Monte Carlo methods, and variational approximations, are developed alongside applications of these tools to clustering, convolutional codes, independent component analysis, and neural networks. The final part of the book describes the state of the art in error-correcting codes, including low-density parity-check codes, turbo codes, and digital fountain codes -- the twenty-first century standards for satellite communications, disk drives, and data broadcast. Richly illustrated, filled with worked examples and over 400 exercises, some with detailed solutions, David MacKay's groundbreaking book is ideal for self-learning and for undergraduate or graduate courses. Interludes on crosswords, evolution, and sex provide entertainment along the way. In sum, this is a textbook on information, communication, and coding for a new generation of students, and an unparalleled entry point into these subjects for professionals in areas as diverse as computational biology, financial engineering, and machine learning.

Structure and Interpretation of Computer Programs


Harold Abelson - 1984
    This long-awaited revision contains changes throughout the text. There are new implementations of most of the major programming systems in the book, including the interpreters and compilers, and the authors have incorporated many small changes that reflect their experience teaching the course at MIT since the first edition was published. A new theme has been introduced that emphasizes the central role played by different approaches to dealing with time in computational models: objects with state, concurrent programming, functional programming and lazy evaluation, and nondeterministic programming. There are new example sections on higher-order procedures in graphics and on applications of stream processing in numerical programming, and many new exercises. In addition, all the programs have been reworked to run in any Scheme implementation that adheres to the IEEE standard.

Uncharted: Big Data and an Emerging Science of Human History


Erez Aiden - 2013
    Gigabytes, exabytes (that’s one quintillion bytes) of data are sitting on servers across the world. So how can we start to access this explosion of information, this “big data,” and what can it tell us?   Erez Aiden and Jean-Baptiste Michel are two young scientists at Harvard who started to ask those questions. They teamed up with Google to create the Ngram Viewer, a Web-based tool that can chart words throughout the massive Google Books archive, sifting through billions of words to find fascinating cultural trends. On the day that the Ngram Viewer debuted in 2010, more than one million queries were run through it.   On the front lines of Big Data, Aiden and Michel realized that this big dataset—the Google Books archive that contains remarkable information on the human experience—had huge implications for looking at our shared human history. The tool they developed to delve into the data has enabled researchers to track how our language has evolved over time, how art has been censored, how fame can grow and fade, how nations trend toward war. How we remember and how we forget. And ultimately, how Big Data is changing the game for the sciences, humanities, politics, business, and our culture.