Book picks similar to
Bayesian Inference in Statistical Analysis by George E.P. Box
mathematics
math-statistics
8107
statistics-probability
The Cartoon Guide to Statistics
Larry Gonick - 1993
Never again will you order the Poisson Distribution in a French restaurant!This updated version features all new material.
Mathematical Analysis
Tom M. Apostol - 1957
It provides a transition from elementary calculus to advanced courses in real and complex function theory and introduces the reader to some of the abstract thinking that pervades modern analysis.
Applied Multivariate Statistical Analysis
Richard A. Johnson - 1982
of Wisconsin-Madison) and Wichern (Texas A&M U.) present the newest edition of this college text on the statistical methods for describing and analyzing multivariate data, designed for students who have taken two or more statistics courses. The fifth edition includes the addition of seve
Hands-On Machine Learning with Scikit-Learn and TensorFlow
Aurélien Géron - 2017
Now that machine learning is thriving, even programmers who know close to nothing about this technology can use simple, efficient tools to implement programs capable of learning from data. This practical book shows you how.By using concrete examples, minimal theory, and two production-ready Python frameworks—Scikit-Learn and TensorFlow—author Aurélien Géron helps you gain an intuitive understanding of the concepts and tools for building intelligent systems. You’ll learn how to use a range of techniques, starting with simple Linear Regression and progressing to Deep Neural Networks. If you have some programming experience and you’re ready to code a machine learning project, this guide is for you.This hands-on book shows you how to use:Scikit-Learn, an accessible framework that implements many algorithms efficiently and serves as a great machine learning entry pointTensorFlow, a more complex library for distributed numerical computation, ideal for training and running very large neural networksPractical code examples that you can apply without learning excessive machine learning theory or algorithm details
Information Theory, Inference and Learning Algorithms
David J.C. MacKay - 2002
These topics lie at the heart of many exciting areas of contemporary science and engineering - communication, signal processing, data mining, machine learning, pattern recognition, computational neuroscience, bioinformatics, and cryptography. This textbook introduces theory in tandem with applications. Information theory is taught alongside practical communication systems, such as arithmetic coding for data compression and sparse-graph codes for error-correction. A toolbox of inference techniques, including message-passing algorithms, Monte Carlo methods, and variational approximations, are developed alongside applications of these tools to clustering, convolutional codes, independent component analysis, and neural networks. The final part of the book describes the state of the art in error-correcting codes, including low-density parity-check codes, turbo codes, and digital fountain codes -- the twenty-first century standards for satellite communications, disk drives, and data broadcast. Richly illustrated, filled with worked examples and over 400 exercises, some with detailed solutions, David MacKay's groundbreaking book is ideal for self-learning and for undergraduate or graduate courses. Interludes on crosswords, evolution, and sex provide entertainment along the way. In sum, this is a textbook on information, communication, and coding for a new generation of students, and an unparalleled entry point into these subjects for professionals in areas as diverse as computational biology, financial engineering, and machine learning.
Statistical Methods for the Social Sciences
Alan Agresti - 1986
No previous knowledge of statistics is assumed, and mathematical background is assumed to be minimal (lowest-level high-school algebra). This text may be used in a one or two course sequence. Such sequences are commonly required of social science graduate students in sociology, political science, and psychology. Students in geography, anthropology, journalism, and speech also are sometimes required to take at least one statistics course.
The Visual Display of Quantitative Information
Edward R. Tufte - 1983
Theory and practice in the design of data graphics, 250 illustrations of the best (and a few of the worst) statistical graphics, with detailed analysis of how to display data for precise, effective, quick analysis. Design of the high-resolution displays, small multiples. Editing and improving graphics. The data-ink ratio. Time-series, relational graphics, data maps, multivariate designs. Detection of graphical deception: design variation vs. data variation. Sources of deception. Aesthetics and data graphical displays. This is the second edition of The Visual Display of Quantitative Information. Recently published, this new edition provides excellent color reproductions of the many graphics of William Playfair, adds color to other images, and includes all the changes and corrections accumulated during 17 printings of the first edition.
Grokking Algorithms An Illustrated Guide For Programmers and Other Curious People
Aditya Y. Bhargava - 2015
The algorithms you'll use most often as a programmer have already been discovered, tested, and proven. If you want to take a hard pass on Knuth's brilliant but impenetrable theories and the dense multi-page proofs you'll find in most textbooks, this is the book for you. This fully-illustrated and engaging guide makes it easy for you to learn how to use algorithms effectively in your own programs.Grokking Algorithms is a disarming take on a core computer science topic. In it, you'll learn how to apply common algorithms to the practical problems you face in day-to-day life as a programmer. You'll start with problems like sorting and searching. As you build up your skills in thinking algorithmically, you'll tackle more complex concerns such as data compression or artificial intelligence. Whether you're writing business software, video games, mobile apps, or system utilities, you'll learn algorithmic techniques for solving problems that you thought were out of your grasp. For example, you'll be able to:Write a spell checker using graph algorithmsUnderstand how data compression works using Huffman codingIdentify problems that take too long to solve with naive algorithms, and attack them with algorithms that give you an approximate answer insteadEach carefully-presented example includes helpful diagrams and fully-annotated code samples in Python. By the end of this book, you will know some of the most widely applicable algorithms as well as how and when to use them.
Statistics for People Who (Think They) Hate Statistics
Neil J. Salkind - 2000
The book begins with an introduction to the language of statistics and then covers descriptive statistics and inferential statistics. Throughout, the author offers readers:- Difficulty Rating Index for each chapter′s material- Tips for doing and thinking about a statistical technique- Top tens for everything from the best ways to create a graph to the most effective techniques for data collection- Steps that break techniques down into a clear sequence of procedures- SPSS tips for executing each major statistical technique- Practice exercises at the end of each chapter, followed by worked out solutions.The book concludes with a statistical software sampler and a description of the best Internet sites for statistical information and data resources. Readers also have access to a website for downloading data that they can use to practice additional exercises from the book. Students and researchers will appreciate the book′s unhurried pace and thorough, friendly presentation.
Algebra
Israel M. Gelfand - 1992
This is a very old science and its gems have lost their charm for us through everyday use. We have tried in this book to refresh them for you. The main part of the book is made up of problems. The best way to deal with them is: Solve the problem by yourself - compare your solution with the solution in the book (if it exists) - go to the next problem. However, if you have difficulties solving a problem (and some of them are quite difficult), you may read the hint or start to read the solution. If there is no solution in the book for some problem, you may skip it (it is not heavily used in the sequel) and return to it later. The book is divided into sections devoted to different topics. Some of them are very short, others are rather long. Of course, you know arithmetic pretty well. However, we shall go through it once more, starting with easy things. 2 Exchange of terms in addition Let's add 3 and 5: 3+5=8. And now change the order: 5+3=8. We get the same result. Adding three apples to five apples is the same as adding five apples to three - apples do not disappear and we get eight of them in both cases. 3 Exchange of terms in multiplication Multiplication has a similar property. But let us first agree on notation.
Deep Learning
Ian Goodfellow - 2016
Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning.The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models.Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors.
Signals and Systems
A. Anand Kumar - 2011
Written with student centred, pedagogically driven approach, the text provides a self-contained introduction to the theory of signals and systems. This book looks at the concepts of systems, and also examines signals and the way that signals interact with physical systems. It covers topics ranging from basic signals and systems to signal analysis, properties of continuous-time Fourier transforms including Fourier transforms of standard signals, signal transmission through linear systems, relation between convolution and correlation of signals, sampling theorems and techniques, and transform analysis of LTI systems. All the solved and unsolved problems in this book are designed to illustrate the topics in a clear way.
Machine Learning
Tom M. Mitchell - 1986
Mitchell covers the field of machine learning, the study of algorithms that allow computer programs to automatically improve through experience and that automatically infer general laws from specific data.
The Model Thinker: What You Need to Know to Make Data Work for You
Scott E. Page - 2018
But as anyone who has ever opened up a spreadsheet packed with seemingly infinite lines of data knows, numbers aren't enough: we need to know how to make those numbers talk. In The Model Thinker, social scientist Scott E. Page shows us the mathematical, statistical, and computational models—from linear regression to random walks and far beyond—that can turn anyone into a genius. At the core of the book is Page's "many-model paradigm," which shows the reader how to apply multiple models to organize the data, leading to wiser choices, more accurate predictions, and more robust designs. The Model Thinker provides a toolkit for business people, students, scientists, pollsters, and bloggers to make them better, clearer thinkers, able to leverage data and information to their advantage.
Introduction to Probability Models
Sheldon M. Ross - 1972
This updated edition of Ross's classic bestseller provides an introduction to elementary probability theory and stochastic processes, and shows how probability theory can be applied to the study of phenomena in fields such as engineering, computer science, management science, the physical and social sciences, and operations research. With the addition of several new sections relating to actuaries, this text is highly recommended by the Society of Actuaries.This book now contains a new section on compound random variables that can be used to establish a recursive formula for computing probability mass functions for a variety of common compounding distributions; a new section on hiddden Markov chains, including the forward and backward approaches for computing the joint probability mass function of the signals, as well as the Viterbi algorithm for determining the most likely sequence of states; and a simplified approach for analyzing nonhomogeneous Poisson processes. There are also additional results on queues relating to the conditional distribution of the number found by an M/M/1 arrival who spends a time t in the system; inspection paradox for M/M/1 queues; and M/G/1 queue with server breakdown. Furthermore, the book includes new examples and exercises, along with compulsory material for new Exam 3 of the Society of Actuaries.This book is essential reading for professionals and students in actuarial science, engineering, operations research, and other fields in applied probability.