The Calculus Direct


John Weiss - 2009
    The calculus is not a hard subject and I prove this through an easy to read and obvious approach spanning only 100 pages. I have written this book with the following type of student in mind; the non-traditional student returning to college after a long break, a notoriously weak student in math who just needs to get past calculus to obtain a degree, and the garage tinkerer who wishes to understand a little more about the technical subjects. This book is meant to address the many fundamental thought-blocks that keep the average 'mathaphobe' (or just an interested person who doesn't have the time to enroll in a course) from excelling in mathematics in a clear and concise manner. It is my sincerest hope that this book helps you with your needs.Show more Show less

Storytelling with Data: A Data Visualization Guide for Business Professionals


Cole Nussbaumer Knaflic - 2015
    You'll discover the power of storytelling and the way to make data a pivotal point in your story. The lessons in this illuminative text are grounded in theory, but made accessible through numerous real-world examples--ready for immediate application to your next graph or presentation.Storytelling is not an inherent skill, especially when it comes to data visualization, and the tools at our disposal don't make it any easier. This book demonstrates how to go beyond conventional tools to reach the root of your data, and how to use your data to create an engaging, informative, compelling story. Specifically, you'll learn how to:Understand the importance of context and audience Determine the appropriate type of graph for your situation Recognize and eliminate the clutter clouding your information Direct your audience's attention to the most important parts of your data Think like a designer and utilize concepts of design in data visualization Leverage the power of storytelling to help your message resonate with your audience Together, the lessons in this book will help you turn your data into high impact visual stories that stick with your audience. Rid your world of ineffective graphs, one exploding 3D pie chart at a time. There is a story in your data--Storytelling with Data will give you the skills and power to tell it!

Why Does E=mc²? (And Why Should We Care?)


Brian Cox - 2009
    Breaking down the symbols themselves, they pose a series of questions: What is energy? What is mass? What has the speed of light got to do with energy and mass? In answering these questions, they take us to the site of one of the largest scientific experiments ever conducted. Lying beneath the city of Geneva, straddling the Franco-Swiss boarder, is a 27 km particle accelerator, known as the Large Hadron Collider. Using this gigantic machine—which can recreate conditions in the early Universe fractions of a second after the Big Bang—Cox and Forshaw will describe the current theory behind the origin of mass.Alongside questions of energy and mass, they will consider the third, and perhaps, most intriguing element of the equation: 'c' - or the speed of light. Why is it that the speed of light is the exchange rate? Answering this question is at the heart of the investigation as the authors demonstrate how, in order to truly understand why E=mc2, we first must understand why we must move forward in time and not backwards and how objects in our 3-dimensional world actually move in 4-dimensional space-time. In other words, how the very fabric of our world is constructed. A collaboration between two of the youngest professors in the UK, Why Does E=mc2? promises to be one of the most exciting and accessible explanations of the theory of relativity in recent years.

Think Complexity: Complexity Science and Computational Modeling


Allen B. Downey - 2009
    Whether you’re an intermediate-level Python programmer or a student of computational modeling, you’ll delve into examples of complex systems through a series of exercises, case studies, and easy-to-understand explanations.You’ll work with graphs, algorithm analysis, scale-free networks, and cellular automata, using advanced features that make Python such a powerful language. Ideal as a text for courses on Python programming and algorithms, Think Complexity will also help self-learners gain valuable experience with topics and ideas they might not encounter otherwise.Work with NumPy arrays and SciPy methods, basic signal processing and Fast Fourier Transform, and hash tablesStudy abstract models of complex physical systems, including power laws, fractals and pink noise, and Turing machinesGet starter code and solutions to help you re-implement and extend original experiments in complexityExplore the philosophy of science, including the nature of scientific laws, theory choice, realism and instrumentalism, and other topicsExamine case studies of complex systems submitted by students and readers

The Evolution of Cooperation


Robert Axelrod - 1984
    Widely praised and much-discussed, this classic book explores how cooperation can emerge in a world of self-seeking egoists—whether superpowers, businesses, or individuals—when there is no central authority to police their actions. The problem of cooperation is central to many different fields. Robert Axelrod recounts the famous computer tournaments in which the “cooperative” program Tit for Tat recorded its stunning victories, explains its application to a broad spectrum of subjects, and suggests how readers can both apply cooperative principles to their own lives and teach cooperative principles to others.

Data Analysis with Open Source Tools: A Hands-On Guide for Programmers and Data Scientists


Philipp K. Janert - 2010
    With this insightful book, intermediate to experienced programmers interested in data analysis will learn techniques for working with data in a business environment. You'll learn how to look at data to discover what it contains, how to capture those ideas in conceptual models, and then feed your understanding back into the organization through business plans, metrics dashboards, and other applications.Along the way, you'll experiment with concepts through hands-on workshops at the end of each chapter. Above all, you'll learn how to think about the results you want to achieve -- rather than rely on tools to think for you.Use graphics to describe data with one, two, or dozens of variablesDevelop conceptual models using back-of-the-envelope calculations, as well asscaling and probability argumentsMine data with computationally intensive methods such as simulation and clusteringMake your conclusions understandable through reports, dashboards, and other metrics programsUnderstand financial calculations, including the time-value of moneyUse dimensionality reduction techniques or predictive analytics to conquer challenging data analysis situationsBecome familiar with different open source programming environments for data analysisFinally, a concise reference for understanding how to conquer piles of data.--Austin King, Senior Web Developer, MozillaAn indispensable text for aspiring data scientists.--Michael E. Driscoll, CEO/Founder, Dataspora

Linear Algebra With Applications


Steven J. Leon - 1980
    Each chapter contains integrated worked examples and chapter tests. This edition has the ancillary ATLAST computer exercise guide and new MATLAB and Maple guides.

The Math Book: From Pythagoras to the 57th Dimension, 250 Milestones in the History of Mathematics


Clifford A. Pickover - 2009
    Beginning millions of years ago with ancient “ant odometers” and moving through time to our modern-day quest for new dimensions, it covers 250 milestones in mathematical history. Among the numerous delights readers will learn about as they dip into this inviting anthology: cicada-generated prime numbers, magic squares from centuries ago, the discovery of pi and calculus, and the butterfly effect. Each topic gets a lavishly illustrated spread with stunning color art, along with formulas and concepts, fascinating facts about scientists’ lives, and real-world applications of the theorems.

Feynman Lectures On Computation


Richard P. Feynman - 1996
    Feynman gave his famous course on computation at the California Institute of Technology, he asked Tony Hey to adapt his lecture notes into a book. Although led by Feynman, the course also featured, as occasional guest speakers, some of the most brilliant men in science at that time, including Marvin Minsky, Charles Bennett, and John Hopfield. Although the lectures are now thirteen years old, most of the material is timeless and presents a “Feynmanesque” overview of many standard and some not-so-standard topics in computer science such as reversible logic gates and quantum computers.

Probabilistic Graphical Models: Principles and Techniques


Daphne Koller - 2009
    The framework of probabilistic graphical models, presented in this book, provides a general approach for this task. The approach is model-based, allowing interpretable models to be constructed and then manipulated by reasoning algorithms. These models can also be learned automatically from data, allowing the approach to be used in cases where manually constructing a model is difficult or even impossible. Because uncertainty is an inescapable aspect of most real-world applications, the book focuses on probabilistic models, which make the uncertainty explicit and provide models that are more faithful to reality.Probabilistic Graphical Models discusses a variety of models, spanning Bayesian networks, undirected Markov networks, discrete and continuous models, and extensions to deal with dynamical systems and relational data. For each class of models, the text describes the three fundamental cornerstones: representation, inference, and learning, presenting both basic concepts and advanced techniques. Finally, the book considers the use of the proposed framework for causal reasoning and decision making under uncertainty. The main text in each chapter provides the detailed technical development of the key ideas. Most chapters also include boxes with additional material: skill boxes, which describe techniques; case study boxes, which discuss empirical cases related to the approach described in the text, including applications in computer vision, robotics, natural language understanding, and computational biology; and concept boxes, which present significant concepts drawn from the material in the chapter. Instructors (and readers) can group chapters in various combinations, from core topics to more technically advanced material, to suit their particular needs.

Convex Optimization


Stephen Boyd - 2004
    A comprehensive introduction to the subject, this book shows in detail how such problems can be solved numerically with great efficiency. The focus is on recognizing convex optimization problems and then finding the most appropriate technique for solving them. The text contains many worked examples and homework exercises and will appeal to students, researchers and practitioners in fields such as engineering, computer science, mathematics, statistics, finance, and economics.

Machine Learning: A Probabilistic Perspective


Kevin P. Murphy - 2012
    Machine learning provides these, developing methods that can automatically detect patterns in data and then use the uncovered patterns to predict future data. This textbook offers a comprehensive and self-contained introduction to the field of machine learning, based on a unified, probabilistic approach.The coverage combines breadth and depth, offering necessary background material on such topics as probability, optimization, and linear algebra as well as discussion of recent developments in the field, including conditional random fields, L1 regularization, and deep learning. The book is written in an informal, accessible style, complete with pseudo-code for the most important algorithms. All topics are copiously illustrated with color images and worked examples drawn from such application domains as biology, text processing, computer vision, and robotics. Rather than providing a cookbook of different heuristic methods, the book stresses a principled model-based approach, often using the language of graphical models to specify models in a concise and intuitive way. Almost all the models described have been implemented in a MATLAB software package—PMTK (probabilistic modeling toolkit)—that is freely available online. The book is suitable for upper-level undergraduates with an introductory-level college math background and beginning graduate students.

In Pursuit of the Unknown: 17 Equations That Changed the World


Ian Stewart - 2012
    We often overlook the historical link between mathematics and technological advances, says Stewart—but this connection is integral to any complete understanding of human history.Equations are modeled on the patterns we find in the world around us, says Stewart, and it is through equations that we are able to make sense of, and in turn influence, our world. Stewart locates the origins of each equation he presents—from Pythagoras's Theorem to Newton's Law of Gravity to Einstein's Theory of Relativity—within a particular historical moment, elucidating the development of mathematical and philosophical thought necessary for each equation's discovery. None of these equations emerged in a vacuum, Stewart shows; each drew, in some way, on past equations and the thinking of the day. In turn, all of these equations paved the way for major developments in mathematics, science, philosophy, and technology. Without logarithms (invented in the early 17th century by John Napier and improved by Henry Briggs), scientists would not have been able to calculate the movement of the planets, and mathematicians would not have been able to develop fractal geometry. The Wave Equation is one of the most important equations in physics, and is crucial for engineers studying the vibrations in vehicles and the response of buildings to earthquakes. And the equation at the heart of Information Theory, devised by Claude Shannon, is the basis of digital communication today.An approachable and informative guide to the equations upon which nearly every aspect of scientific and mathematical understanding depends, In Pursuit of the Unknown is also a reminder that equations have profoundly influenced our thinking and continue to make possible many of the advances that we take for granted.

Statistical Inference


George Casella - 2001
    Starting from the basics of probability, the authors develop the theory of statistical inference using techniques, definitions, and concepts that are statistical and are natural extensions and consequences of previous concepts. This book can be used for readers who have a solid mathematics background. It can also be used in a way that stresses the more practical uses of statistical theory, being more concerned with understanding basic statistical concepts and deriving reasonable statistical procedures for a variety of situations, and less concerned with formal optimality investigations.

Classical Mechanics


John R. Taylor - 2003
    John Taylor has brought to his new book, Classical Mechanics, all of the clarity and insight that made his introduction to Error Analysis a best-selling text.