Book picks similar to
Probability Theory: A Comprehensive Course by Achim Klenke
math
mathematics
probability
statistics
A Mathematician's Lament: How School Cheats Us Out of Our Most Fascinating and Imaginative Art Form
Paul Lockhart - 2009
Witty and accessible, Paul Lockhart’s controversial approach will provoke spirited debate among educators and parents alike and it will alter the way we think about math forever.Paul Lockhart, has taught mathematics at Brown University and UC Santa Cruz. Since 2000, he has dedicated himself to K-12 level students at St. Ann’s School in Brooklyn, New York.
Chaos and Fractals: New Frontiers of Science
Heinz-Otto Peitgen - 1992
At the time we were hoping that our approach of writing a book which would be both accessible without mathematical sophistication and portray these exiting new fields in an authentic manner would find an audience. Now we know it did. We know from many reviews and personal letters that the book is used in a wide range of ways: researchers use it to acquaint themselves, teachers use it in college and university courses, students use it for background reading, and there is also a substantial audience of lay people who just want to know what chaos and fractals are about. Every book that is somewhat technical in nature is likely to have a number of misprints and errors in its first edition. Some of these were caught and brought to our attention by our readers. One of them, Hermann Flaschka, deserves to be thanked in particular for his suggestions and improvements. This second edition has several changes. We have taken out the two appendices from the firstedition. At the time of the first edition Yuval Fishers contribution, which we published as an appendix was probably the first complete expository account on fractal image compression. Meanwhile, Yuvals book Fractal Image Compression: Theory and Application appeared and is now the publication to refer to.
Gamma: Exploring Euler's Constant
Julian Havil - 2003
Following closely behind is y, or gamma, a constant that arises in many mathematical areas yet maintains a profound sense of mystery. In a tantalizing blend of history and mathematics, Julian Havil takes the reader on a journey through logarithms and the harmonic series, the two defining elements of gamma, toward the first account of gamma's place in mathematics. Introduced by the Swiss mathematician Leonhard Euler (1707-1783), who figures prominently in this book, gamma is defined as the limit of the sum of 1 + 1/2 + 1/3 + . . . Up to 1/n, minus the natural logarithm of n--the numerical value being 0.5772156. . . . But unlike its more celebrated colleagues π and e, the exact nature of gamma remains a mystery--we don't even know if gamma can be expressed as a fraction. Among the numerous topics that arise during this historical odyssey into fundamental mathematical ideas are the Prime Number Theorem and the most important open problem in mathematics today--the Riemann Hypothesis (though no proof of either is offered!). Sure to be popular with not only students and instructors but all math aficionados, Gamma takes us through countries, centuries, lives, and works, unfolding along the way the stories of some remarkable mathematics from some remarkable mathematicians.-- "Notices of the American Mathematical Society"
Why Does E=mc²? (And Why Should We Care?)
Brian Cox - 2009
Breaking down the symbols themselves, they pose a series of questions: What is energy? What is mass? What has the speed of light got to do with energy and mass? In answering these questions, they take us to the site of one of the largest scientific experiments ever conducted. Lying beneath the city of Geneva, straddling the Franco-Swiss boarder, is a 27 km particle accelerator, known as the Large Hadron Collider. Using this gigantic machine—which can recreate conditions in the early Universe fractions of a second after the Big Bang—Cox and Forshaw will describe the current theory behind the origin of mass.Alongside questions of energy and mass, they will consider the third, and perhaps, most intriguing element of the equation: 'c' - or the speed of light. Why is it that the speed of light is the exchange rate? Answering this question is at the heart of the investigation as the authors demonstrate how, in order to truly understand why E=mc2, we first must understand why we must move forward in time and not backwards and how objects in our 3-dimensional world actually move in 4-dimensional space-time. In other words, how the very fabric of our world is constructed. A collaboration between two of the youngest professors in the UK, Why Does E=mc2? promises to be one of the most exciting and accessible explanations of the theory of relativity in recent years.
Networks: An Introduction
M.E.J. Newman - 2010
The rise of the Internet and the wide availability of inexpensive computers have made it possible to gather and analyze network data on a large scale, and the development of a variety of new theoretical tools has allowed us to extract new knowledge from many different kinds of networks.The study of networks is broadly interdisciplinary and important developments have occurred in many fields, including mathematics, physics, computer and information sciences, biology, and the social sciences. This book brings together for the first time the most important breakthroughs in each of these fields and presents them in a coherent fashion, highlighting the strong interconnections between work in different areas.Subjects covered include the measurement and structure of networks in many branches of science, methods for analyzing network data, including methods developed in physics, statistics, and sociology, the fundamentals of graph theory, computer algorithms, and spectral methods, mathematical models of networks, including random graph models and generative models, and theories of dynamical processes taking place on networks.
The Cult of Statistical Significance: How the Standard Error Costs Us Jobs, Justice, and Lives
Stephen Thomas Ziliak - 2008
If it takes a book to get it across, I hope this book will do it. It ought to.”—Thomas Schelling, Distinguished University Professor, School of Public Policy, University of Maryland, and 2005 Nobel Prize Laureate in Economics “With humor, insight, piercing logic and a nod to history, Ziliak and McCloskey show how economists—and other scientists—suffer from a mass delusion about statistical analysis. The quest for statistical significance that pervades science today is a deeply flawed substitute for thoughtful analysis. . . . Yet few participants in the scientific bureaucracy have been willing to admit what Ziliak and McCloskey make clear: the emperor has no clothes.”—Kenneth Rothman, Professor of Epidemiology, Boston University School of Health The Cult of Statistical Significance shows, field by field, how “statistical significance,” a technique that dominates many sciences, has been a huge mistake. The authors find that researchers in a broad spectrum of fields, from agronomy to zoology, employ “testing” that doesn’t test and “estimating” that doesn’t estimate. The facts will startle the outside reader: how could a group of brilliant scientists wander so far from scientific magnitudes? This study will encourage scientists who want to know how to get the statistical sciences back on track and fulfill their quantitative promise. The book shows for the first time how wide the disaster is, and how bad for science, and it traces the problem to its historical, sociological, and philosophical roots. Stephen T. Ziliak is the author or editor of many articles and two books. He currently lives in Chicago, where he is Professor of Economics at Roosevelt University. Deirdre N. McCloskey, Distinguished Professor of Economics, History, English, and Communication at the University of Illinois at Chicago, is the author of twenty books and three hundred scholarly articles. She has held Guggenheim and National Humanities Fellowships. She is best known for How to Be Human* Though an Economist (University of Michigan Press, 2000) and her most recent book, The Bourgeois Virtues: Ethics for an Age of Commerce (2006).
The Analysis of Biological Data
Michael C. Whitlock - 2008
To reach this unique audience, Whitlock and Schluter motivate learning with interesting biological and medical examples; they emphasize intuitive understanding; and they focus on real data. The book covers basic topics in introductory statistics, including graphs, confidence intervals, hypothesis testing, comparison of means, regression, and designing experiments. It also introduces the principles behind such modern topics as likelihood, linear models, meta-analysis and computer-intensive methods. Instructors and students consistently praise the book's clear and engaging writing, strong visualization techniques, and its variety of fascinating and relevant biological examples.
Book of Proof
Richard Hammack - 2009
It is a bridge from the computational courses (such as calculus or differential equations) that students typically encounter in their first year of college to a more abstract outlook. It lays a foundation for more theoretical courses such as topology, analysis and abstract algebra. Although it may be more meaningful to the student who has had some calculus, there is really no prerequisite other than a measure of mathematical maturity. Topics include sets, logic, counting, methods of conditional and non-conditional proof, disproof, induction, relations, functions and infinite cardinality.
Doing Math with Python
Amit Saha - 2015
Python is easy to learn, and it's perfect for exploring topics like statistics, geometry, probability, and calculus. You’ll learn to write programs to find derivatives, solve equations graphically, manipulate algebraic expressions, even examine projectile motion.Rather than crank through tedious calculations by hand, you'll learn how to use Python functions and modules to handle the number crunching while you focus on the principles behind the math. Exercises throughout teach fundamental programming concepts, like using functions, handling user input, and reading and manipulating data. As you learn to think computationally, you'll discover new ways to explore and think about math, and gain valuable programming skills that you can use to continue your study of math and computer science.If you’re interested in math but have yet to dip into programming, you’ll find that Python makes it easy to go deeper into the subject—let Python handle the tedious work while you spend more time on the math.
Applied Multivariate Statistical Analysis
Richard A. Johnson - 1982
of Wisconsin-Madison) and Wichern (Texas A&M U.) present the newest edition of this college text on the statistical methods for describing and analyzing multivariate data, designed for students who have taken two or more statistics courses. The fifth edition includes the addition of seve
The Data Detective: Ten Easy Rules to Make Sense of Statistics
Tim Harford - 2020
That’s a mistake, Tim Harford says in The Data Detective. We shouldn’t be suspicious of statistics—we need to understand what they mean and how they can improve our lives: they are, at heart, human behavior seen through the prism of numbers and are often “the only way of grasping much of what is going on around us.” If we can toss aside our fears and learn to approach them clearly—understanding how our own preconceptions lead us astray—statistics can point to ways we can live better and work smarter.As “perhaps the best popular economics writer in the world” (New Statesman), Tim Harford is an expert at taking complicated ideas and untangling them for millions of readers. In The Data Detective, he uses new research in science and psychology to set out ten strategies for using statistics to erase our biases and replace them with new ideas that use virtues like patience, curiosity, and good sense to better understand ourselves and the world. As a result, The Data Detective is a big-idea book about statistics and human behavior that is fresh, unexpected, and insightful.
Introduction to Linear Algebra
Gilbert Strang - 1993
Topics covered include matrix multiplication, row reduction, matrix inverse, orthogonality and computation. The self-teaching book is loaded with examples and graphics and provides a wide array of probing problems, accompanying solutions, and a glossary. Chapter 1: Introduction to Vectors; Chapter 2: Solving Linear Equations; Chapter 3: Vector Spaces and Subspaces; Chapter 4: Orthogonality; Chapter 5: Determinants; Chapter 6: Eigenvalues and Eigenvectors; Chapter 7: Linear Transformations; Chapter 8: Applications; Chapter 9: Numerical Linear Algebra; Chapter 10: Complex Vectors and Matrices; Solutions to Selected Exercises; Final Exam. Matrix Factorizations. Conceptual Questions for Review. Glossary: A Dictionary for Linear Algebra Index Teaching Codes Linear Algebra in a Nutshell.
The Improbability Principle: Why Coincidences, Miracles, and Rare Events Happen Every Day
David J. Hand - 2014
Hand argues that extraordinarily rare events are anything but. In fact, they’re commonplace. Not only that, we should all expect to experience a miracle roughly once every month. But Hand is no believer in superstitions, prophecies, or the paranormal. His definition of “miracle” is thoroughly rational. No mystical or supernatural explanation is necessary to understand why someone is lucky enough to win the lottery twice, or is destined to be hit by lightning three times and still survive. All we need, Hand argues, is a firm grounding in a powerful set of laws: the laws of inevitability, of truly large numbers, of selection, of the probability lever, and of near enough. Together, these constitute Hand’s groundbreaking Improbability Principle. And together, they explain why we should not be so surprised to bump into a friend in a foreign country, or to come across the same unfamiliar word four times in one day. Hand wrestles with seemingly less explicable questions as well: what the Bible and Shakespeare have in common, why financial crashes are par for the course, and why lightning does strike the same place (and the same person) twice. Along the way, he teaches us how to use the Improbability Principle in our own lives—including how to cash in at a casino and how to recognize when a medicine is truly effective. An irresistible adventure into the laws behind “chance” moments and a trusty guide for understanding the world and universe we live in, The Improbability Principle will transform how you think about serendipity and luck, whether it’s in the world of business and finance or you’re merely sitting in your backyard, tossing a ball into the air and wondering where it will land.
Time Series Analysis
James Douglas Hamilton - 1994
This book synthesizes these recent advances and makes them accessible to first-year graduate students. James Hamilton provides the first adequate text-book treatments of important innovations such as vector autoregressions, generalized method of moments, the economic and statistical consequences of unit roots, time-varying variances, and nonlinear time series models. In addition, he presents basic tools for analyzing dynamic systems (including linear representations, autocovariance generating functions, spectral analysis, and the Kalman filter) in a way that integrates economic theory with the practical difficulties of analyzing and interpreting real-world data. Time Series Analysis fills an important need for a textbook that integrates economic theory, econometrics, and new results.The book is intended to provide students and researchers with a self-contained survey of time series analysis. It starts from first principles and should be readily accessible to any beginning graduate student, while it is also intended to serve as a reference book for researchers.-- "Journal of Economics"
