Concrete Mathematics: A Foundation for Computer Science


Ronald L. Graham - 1988
    "More concretely," the authors explain, "it is the controlled manipulation of mathematical formulas, using a collection of techniques for solving problems."

Programming Collective Intelligence: Building Smart Web 2.0 Applications


Toby Segaran - 2002
    With the sophisticated algorithms in this book, you can write smart programs to access interesting datasets from other web sites, collect data from users of your own applications, and analyze and understand the data once you've found it.Programming Collective Intelligence takes you into the world of machine learning and statistics, and explains how to draw conclusions about user experience, marketing, personal tastes, and human behavior in general -- all from information that you and others collect every day. Each algorithm is described clearly and concisely with code that can immediately be used on your web site, blog, Wiki, or specialized application. This book explains:Collaborative filtering techniques that enable online retailers to recommend products or media Methods of clustering to detect groups of similar items in a large dataset Search engine features -- crawlers, indexers, query engines, and the PageRank algorithm Optimization algorithms that search millions of possible solutions to a problem and choose the best one Bayesian filtering, used in spam filters for classifying documents based on word types and other features Using decision trees not only to make predictions, but to model the way decisions are made Predicting numerical values rather than classifications to build price models Support vector machines to match people in online dating sites Non-negative matrix factorization to find the independent features in a dataset Evolving intelligence for problem solving -- how a computer develops its skill by improving its own code the more it plays a game Each chapter includes exercises for extending the algorithms to make them more powerful. Go beyond simple database-backed applications and put the wealth of Internet data to work for you. "Bravo! I cannot think of a better way for a developer to first learn these algorithms and methods, nor can I think of a better way for me (an old AI dog) to reinvigorate my knowledge of the details."-- Dan Russell, Google "Toby's book does a great job of breaking down the complex subject matter of machine-learning algorithms into practical, easy-to-understand examples that can be directly applied to analysis of social interaction across the Web today. If I had this book two years ago, it would have saved precious time going down some fruitless paths."-- Tim Wolters, CTO, Collective Intellect

R Cookbook: Proven Recipes for Data Analysis, Statistics, and Graphics


Paul Teetor - 2011
    The R language provides everything you need to do statistical work, but its structure can be difficult to master. This collection of concise, task-oriented recipes makes you productive with R immediately, with solutions ranging from basic tasks to input and output, general statistics, graphics, and linear regression.Each recipe addresses a specific problem, with a discussion that explains the solution and offers insight into how it works. If you're a beginner, R Cookbook will help get you started. If you're an experienced data programmer, it will jog your memory and expand your horizons. You'll get the job done faster and learn more about R in the process.Create vectors, handle variables, and perform other basic functionsInput and output dataTackle data structures such as matrices, lists, factors, and data framesWork with probability, probability distributions, and random variablesCalculate statistics and confidence intervals, and perform statistical testsCreate a variety of graphic displaysBuild statistical models with linear regressions and analysis of variance (ANOVA)Explore advanced statistical techniques, such as finding clusters in your dataWonderfully readable, R Cookbook serves not only as a solutions manual of sorts, but as a truly enjoyable way to explore the R language--one practical example at a time.--Jeffrey Ryan, software consultant and R package author

Principles of Quantum Mechanics


Ramamurti Shankar - 1980
    The postulates of quantum mechanics and the mathematical underpinnings are discussed in a clear, succinct manner." (American Scientist)"No matter how gently one introduces students to the concept of Dirac's bras and kets, many are turned off. Shankar attacks the problem head-on in the first chapter, and in a very informal style suggests that there is nothing to be frightened of." (Physics Bulletin)Reviews of the Second Edition:"This massive text of 700 and odd pages has indeed an excellent get-up, is very verbal and expressive, and has extensively worked out calculational details---all just right for a first course. The style is conversational, more like a corridor talk or lecture notes, though arranged as a text. ... It would be particularly useful to beginning students and those in allied areas like quantum chemistry." (Mathematical Reviews)R. Shankar has introduced major additions and updated key presentations in this second edition of Principles of Quantum Mechanics. New features of this innovative text include an entirely rewritten mathematical introduction, a discussion of Time-reversal invariance, and extensive coverage of a variety of path integrals and their applications. Additional highlights include:- Clear, accessible treatment of underlying mathematics- A review of Newtonian, Lagrangian, and Hamiltonian mechanics- Student understanding of quantum theory is enhanced by separate treatment of mathematical theorems and physical postulates- Unsurpassed coverage of path integrals and their relevance in contemporary physicsThe requisite text for advanced undergraduate- and graduate-level students, Principles of Quantum Mechanics, Second Edition is fully referenced and is supported by many exercises and solutions. The book's self-contained chapters also make it suitable for independent study as well as for courses in applied disciplines.

The Linux Command Line Beginner's Guide


Jonathan Moeller - 2012
    ABOUT THE AUTHORStanding over six feet tall, Jonathan Moeller has the piercing blue eyes of a Conan of Cimmeria, the bronze-colored hair a Visigothic warrior-king, and the stern visage of a captain of men, none of which are useful in his career as a computer repairman, alas.He has written the "Demonsouled" trilogy of sword-and-sorcery novels, and continues to write the "Ghosts" sequence about assassin and spy Caina Amalas, the "Computer Beginner's Guide" series of computer books, and numerous other works.

Epidemiology: An Introduction


Kenneth J. Rothman - 2002
    These areas of knowledge have converged into a modern theory of epidemiology that has been slow to penetrate into textbooks, particularly at the introductory level. Epidemiology: An Introduction closes the gap. It begins with a brief, lucid discussion of causal thinking and causal inference and then takes the reader through the elements of epidemiology, focusing on the measures of disease occurrence and causal effects. With these building blocks in place, the reader learns how to design, analyze and interpret problems that epidemiologists face, including confounding, the role of chance, and the exploration of interactions. All these topics are layered on the foundation of basic principles presented in simple language, with numerous examples and questions for further thought.

Data Science for Business: What you need to know about data mining and data-analytic thinking


Foster Provost - 2013
    This guide also helps you understand the many data-mining techniques in use today.Based on an MBA course Provost has taught at New York University over the past ten years, Data Science for Business provides examples of real-world business problems to illustrate these principles. You’ll not only learn how to improve communication between business stakeholders and data scientists, but also how participate intelligently in your company’s data science projects. You’ll also discover how to think data-analytically, and fully appreciate how data science methods can support business decision-making.Understand how data science fits in your organization—and how you can use it for competitive advantageTreat data as a business asset that requires careful investment if you’re to gain real valueApproach business problems data-analytically, using the data-mining process to gather good data in the most appropriate wayLearn general concepts for actually extracting knowledge from dataApply data science principles when interviewing data science job candidates

R in a Nutshell: A Desktop Quick Reference


Joseph Adler - 2009
    R in a Nutshell provides a quick and practical way to learn this increasingly popular open source language and environment. You'll not only learn how to program in R, but also how to find the right user-contributed R packages for statistical modeling, visualization, and bioinformatics.The author introduces you to the R environment, including the R graphical user interface and console, and takes you through the fundamentals of the object-oriented R language. Then, through a variety of practical examples from medicine, business, and sports, you'll learn how you can use this remarkable tool to solve your own data analysis problems.Understand the basics of the language, including the nature of R objectsLearn how to write R functions and build your own packagesWork with data through visualization, statistical analysis, and other methodsExplore the wealth of packages contributed by the R communityBecome familiar with the lattice graphics package for high-level data visualizationLearn about bioinformatics packages provided by Bioconductor"I am excited about this book. R in a Nutshell is a great introduction to R, as well as a comprehensive reference for using R in data analytics and visualization. Adler provides 'real world' examples, practical advice, and scripts, making it accessible to anyone working with data, not just professional statisticians."

Principles of Statistics


M.G. Bulmer - 1979
    There are equally many advanced textbooks which delve into the far reaches of statistical theory, while bypassing practical applications. But between these two approaches is an unfilled gap, in which theory and practice merge at an intermediate level. Professor M. G. Bulmer's Principles of Statistics, originally published in 1965, was created to fill that need. The new, corrected Dover edition of Principles of Statistics makes this invaluable mid-level text available once again for the classroom or for self-study.Principles of Statistics was created primarily for the student of natural sciences, the social scientist, the undergraduate mathematics student, or anyone familiar with the basics of mathematical language. It assumes no previous knowledge of statistics or probability; nor is extensive mathematical knowledge necessary beyond a familiarity with the fundamentals of differential and integral calculus. (The calculus is used primarily for ease of notation; skill in the techniques of integration is not necessary in order to understand the text.)Professor Bulmer devotes the first chapters to a concise, admirably clear description of basic terminology and fundamental statistical theory: abstract concepts of probability and their applications in dice games, Mendelian heredity, etc.; definitions and examples of discrete and continuous random variables; multivariate distributions and the descriptive tools used to delineate them; expected values; etc. The book then moves quickly to more advanced levels, as Professor Bulmer describes important distributions (binomial, Poisson, exponential, normal, etc.), tests of significance, statistical inference, point estimation, regression, and correlation. Dozens of exercises and problems appear at the end of various chapters, with answers provided at the back of the book. Also included are a number of statistical tables and selected references.

Systems Engineering and Analysis


Benjamin S. Blanchard - 1981
    Each

Deep Learning


Ian Goodfellow - 2016
    Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning.The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models.Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors.

The Demon in the Machine: How Hidden Webs of Information Are Solving the Mystery of Life


Paul C.W. Davies - 2019
    if you want to understand how the concept of life is changing, read this' Professor Andrew Briggs, University of OxfordWhen Darwin set out to explain the origin of species, he made no attempt to answer the deeper question: what is life? For generations, scientists have struggled to make sense of this fundamental question. Life really does look like magic: even a humble bacterium accomplishes things so dazzling that no human engineer can match it. And yet, huge advances in molecular biology over the past few decades have served only to deepen the mystery. So can life be explained by known physics and chemistry, or do we need something fundamentally new?In this penetrating and wide-ranging new analysis, world-renowned physicist and science communicator Paul Davies searches for answers in a field so new and fast-moving that it lacks a name, a domain where computing, chemistry, quantum physics and nanotechnology intersect. At the heart of these diverse fields, Davies explains, is the concept of information: a quantity with the power to unify biology with physics, transform technology and medicine, and even to illuminate the age-old question of whether we are alone in the universe.From life's murky origins to the microscopic engines that run the cells of our bodies, The Demon in the Machine is a breath-taking journey across the landscape of physics, biology, logic and computing. Weaving together cancer and consciousness, two-headed worms and bird navigation, Davies reveals how biological organisms garner and process information to conjure order out of chaos, opening a window on the secret of life itself.

Microdosing Psychedelics: A Practical Guide to Upgrade Your Life


Paul Austin - 2018
    Weaving together anecdotal narratives and preliminary research on the use of small amounts of psychedelics, Microdosing Psychedelics is oriented toward anyone interested in microdosing to improve their general well-being, particularly those in leadership and creative positions. Over the last few years, microdosing has entered the cultural zeitgeist. From bio-hackers in Silicon Valley seeking to improve their focus, to writers and artists looking for a creative edge, to suburban moms struggling with depression, to everyday people who want to achieve emotional balance, microdosing has led the way in mainstreaming and normalizing psychedelics. In the most comprehensive book of its kind, author, public speaker, and microdosing expert Paul Austin provides a thorough overview of how just about anyone can benefit from an intentional, structured microdosing protocol. In Microdosing Psychedelics, Austin provides readers with everything they need to know about microdosing, including: • What exactly is microdosing and how did it become a “thing?” • How to set up a weekly schedule if you want to experiment with microdosing • Three ways microdosing can benefit your professional career • The safest and easiest ways to obtain psychedelics if you wish to experiment with microdosing

Applied Predictive Modeling


Max Kuhn - 2013
    Non- mathematical readers will appreciate the intuitive explanations of the techniques while an emphasis on problem-solving with real data across a wide variety of applications will aid practitioners who wish to extend their expertise. Readers should have knowledge of basic statistical ideas, such as correlation and linear regression analysis. While the text is biased against complex equations, a mathematical background is needed for advanced topics. Dr. Kuhn is a Director of Non-Clinical Statistics at Pfizer Global R&D in Groton Connecticut. He has been applying predictive models in the pharmaceutical and diagnostic industries for over 15 years and is the author of a number of R packages. Dr. Johnson has more than a decade of statistical consulting and predictive modeling experience in pharmaceutical research and development. He is a co-founder of Arbor Analytics, a firm specializing in predictive modeling and is a former Director of Statistics at Pfizer Global R&D. His scholarly work centers on the application and development of statistical methodology and learning algorithms. Applied Predictive Modeling covers the overall predictive modeling process, beginning with the crucial steps of data preprocessing, data splitting and foundations of model tuning. The text then provides intuitive explanations of numerous common and modern regression and classification techniques, always with an emphasis on illustrating and solving real data problems. Addressing practical concerns extends beyond model fitting to topics such as handling class imbalance, selecting predictors, and pinpointing causes of poor model performance-all of which are problems that occur frequently in practice. The text illustrates all parts of the modeling process through many hands-on, real-life examples. And every chapter contains extensive R code f

Machine Learning


Tom M. Mitchell - 1986
    Mitchell covers the field of machine learning, the study of algorithms that allow computer programs to automatically improve through experience and that automatically infer general laws from specific data.