Book picks similar to
Control and Optimization Methods for Electric Smart Grids by Aranya Chakrabortty
physics
smart-grid
statistics
applied-math
Introduction to Graph Theory
Douglas B. West - 1995
Verification that algorithms work is emphasized more than their complexity. An effective use of examples, and huge number of interesting exercises, demonstrate the topics of trees and distance, matchings and factors, connectivity and paths, graph coloring, edges and cycles, and planar graphs. For those who need to learn to make coherent arguments in the fields of mathematics and computer science.
Information Theory: A Tutorial Introduction
James V. Stone - 2015
In this richly illustrated book, accessible examples are used to show how information theory can be understood in terms of everyday games like '20 Questions', and the simple MatLab programs provided give hands-on experience of information theory in action. Written in a tutorial style, with a comprehensive glossary, this text represents an ideal primer for novices who wish to become familiar with the basic principles of information theory.Download chapter 1 from http://jim-stone.staff.shef.ac.uk/Boo...
Networks: A Very Short Introduction
Guido Caldarelli - 2012
It is impossible to understand the spread of an epidemic, a computer virus, large-scale blackouts, or massive extinctions without taking into account the network structure that underlies all these phenomena. In this Very Short Introduction, Guido Caldarelli and Michele Catanzaro discuss the nature and variety of networks, using everyday examples from society, technology, nature, and history to explain and understand the science of network theory. They show the ubiquitous role of networks; how networks self-organize; why the rich get richer; and how networks can spontaneously collapse. They conclude by highlighting how the findings of complex network theory have very wide and important applications in genetics, ecology, communications, economics, and sociology.
Deep Learning
Ian Goodfellow - 2016
Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning.The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models.Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors.
Foundations of Statistical Natural Language Processing
Christopher D. Manning - 1999
This foundational text is the first comprehensive introduction to statistical natural language processing (NLP) to appear. The book contains all the theory and algorithms needed for building NLP tools. It provides broad but rigorous coverage of mathematical and linguistic foundations, as well as detailed discussion of statistical methods, allowing students and researchers to construct their own implementations. The book covers collocation finding, word sense disambiguation, probabilistic parsing, information retrieval, and other applications.
Computer Age Statistical Inference: Algorithms, Evidence, and Data Science
Bradley Efron - 2016
'Big data', 'data science', and 'machine learning' have become familiar terms in the news, as statistical methods are brought to bear upon the enormous data sets of modern science and commerce. How did we get here? And where are we going? This book takes us on an exhilarating journey through the revolution in data analysis following the introduction of electronic computation in the 1950s. Beginning with classical inferential theories - Bayesian, frequentist, Fisherian - individual chapters take up a series of influential topics: survival analysis, logistic regression, empirical Bayes, the jackknife and bootstrap, random forests, neural networks, Markov chain Monte Carlo, inference after model selection, and dozens more. The distinctly modern approach integrates methodology and algorithms with statistical inference. The book ends with speculation on the future direction of statistics and data science.
The Demon in the Machine: How Hidden Webs of Information Are Solving the Mystery of Life
Paul C.W. Davies - 2019
if you want to understand how the concept of life is changing, read this' Professor Andrew Briggs, University of OxfordWhen Darwin set out to explain the origin of species, he made no attempt to answer the deeper question: what is life? For generations, scientists have struggled to make sense of this fundamental question. Life really does look like magic: even a humble bacterium accomplishes things so dazzling that no human engineer can match it. And yet, huge advances in molecular biology over the past few decades have served only to deepen the mystery. So can life be explained by known physics and chemistry, or do we need something fundamentally new?In this penetrating and wide-ranging new analysis, world-renowned physicist and science communicator Paul Davies searches for answers in a field so new and fast-moving that it lacks a name, a domain where computing, chemistry, quantum physics and nanotechnology intersect. At the heart of these diverse fields, Davies explains, is the concept of information: a quantity with the power to unify biology with physics, transform technology and medicine, and even to illuminate the age-old question of whether we are alone in the universe.From life's murky origins to the microscopic engines that run the cells of our bodies, The Demon in the Machine is a breath-taking journey across the landscape of physics, biology, logic and computing. Weaving together cancer and consciousness, two-headed worms and bird navigation, Davies reveals how biological organisms garner and process information to conjure order out of chaos, opening a window on the secret of life itself.
Mathematics: The Loss of Certainty
Morris Kline - 1980
Mathematics: The Loss of Certainty refutes that myth.
Once Upon A Number: The Hidden Mathematical Logic Of Stories
John Allen Paulos - 1998
Good stories are full of life: they engage our emotions and have subtlety and nuance, but they lack rigor and the truths they tell are elusive and subject to debate. As ways of understanding the world around us, numbers and stories seem almost completely incompatible. Once Upon a Number shows that stories and numbers aren't as different as you might imagine, and in fact they have surprising and fascinating connections. The concepts of logic and probability both grew out of intuitive ideas about how certain situations would play out. Now, logicians are inventing ways to deal with real world situations by mathematical means -- by acknowledging, for instance, that items that are mathematically interchangeable may not be interchangeable in a story. And complexity theory looks at both number strings and narrative strings in remarkably similar terms. Throughout, renowned author John Paulos mixes numbers and narratives in his own delightful style. Along with lucid accounts of cutting-edge information theory we get hilarious anecdotes and jokes; instructions for running a truly impressive pyramid scam; a freewheeling conversation between Groucho Marx and Bertrand Russell (while they're stuck in an elevator together); explanations of why the statistical evidence against OJ Simpson was overwhelming beyond doubt and how the Unabomber's thinking shows signs of mathematical training; and dozens of other treats. This is another winner from America's favorite mathematician.
Libellus de Numeros
Jim West - 2014
With a cruel council leading the only safe city of it's kind in this world, she will have to prove her worth to stay as well as help this city as it is the target for two evil wizards who seek to destroy the city and it's ruling council.Will the council's mighty army of guardians be enough to repel the onslaught of the two wizards' wrath?To help the city and also get back home, she will need the help of the greatest mathematician of all time, Archimedes. In a world where math is magic, Alex wishes she paid more attention in math class.
MAKE: Electronics: Learning Through Discovery
Charles Platt - 2008
I also love the sense of humor. It's very good at disarming the fear. And it's gorgeous. I'll be recommending this book highly." --Tom Igoe, author of Physical Computing and Making Things TalkWant to learn the fundamentals of electronics in a fun, hands-on way? With Make: Electronics, you'll start working on real projects as soon as you crack open the book. Explore all of the key components and essential principles through a series of fascinating experiments. You'll build the circuits first, then learn the theory behind them!Build working devices, from simple to complex You'll start with the basics and then move on to more complicated projects. Go from switching circuits to integrated circuits, and from simple alarms to programmable microcontrollers. Step-by-step instructions and more than 500 full-color photographs and illustrations will help you use -- and understand -- electronics concepts and techniques.Discover by breaking things: experiment with components and learn from failureSet up a tricked-out project space: make a work area at home, equipped with the tools and parts you'll needLearn about key electronic components and their functions within a circuitCreate an intrusion alarm, holiday lights, wearable electronic jewelry, audio processors, a reflex tester, and a combination lockBuild an autonomous robot cart that can sense its environment and avoid obstaclesGet clear, easy-to-understand explanations of what you're doing and why
Head First Data Analysis: A Learner's Guide to Big Numbers, Statistics, and Good Decisions
Michael G. Milton - 2009
If your job requires you to manage and analyze all kinds of data, turn to Head First Data Analysis, where you'll quickly learn how to collect and organize data, sort the distractions from the truth, find meaningful patterns, draw conclusions, predict the future, and present your findings to others. Whether you're a product developer researching the market viability of a new product or service, a marketing manager gauging or predicting the effectiveness of a campaign, a salesperson who needs data to support product presentations, or a lone entrepreneur responsible for all of these data-intensive functions and more, the unique approach in Head First Data Analysis is by far the most efficient way to learn what you need to know to convert raw data into a vital business tool. You'll learn how to:Determine which data sources to use for collecting information Assess data quality and distinguish signal from noise Build basic data models to illuminate patterns, and assimilate new information into the models Cope with ambiguous information Design experiments to test hypotheses and draw conclusions Use segmentation to organize your data within discrete market groups Visualize data distributions to reveal new relationships and persuade others Predict the future with sampling and probability models Clean your data to make it useful Communicate the results of your analysis to your audience Using the latest research in cognitive science and learning theory to craft a multi-sensory learning experience, Head First Data Analysis uses a visually rich format designed for the way your brain works, not a text-heavy approach that puts you to sleep.
Calculated Risks: How to Know When Numbers Deceive You
Gerd Gigerenzer - 2002
G. Wells predicted that statistical thinking would be as necessary for citizenship in a technological world as the ability to read and write. But in the twenty-first century, we are often overwhelmed by a baffling array of percentages and probabilities as we try to navigate in a world dominated by statistics. Cognitive scientist Gerd Gigerenzer says that because we haven't learned statistical thinking, we don't understand risk and uncertainty. In order to assess risk -- everything from the risk of an automobile accident to the certainty or uncertainty of some common medical screening tests -- we need a basic understanding of statistics.Astonishingly, doctors and lawyers don't understand risk any better than anyone else. Gigerenzer reports a study in which doctors were told the results of breast cancer screenings and then were asked to explain the risks of contracting breast cancer to a woman who received a positive result from a screening. The actual risk was small because the test gives many false positives. But nearly every physician in the study overstated the risk. Yet many people will have to make important health decisions based on such information and the interpretation of that information by their doctors.Gigerenzer explains that a major obstacle to our understanding of numbers is that we live with an illusion of certainty. Many of us believe that HIV tests, DNA fingerprinting, and the growing number of genetic tests are absolutely certain. But even DNA evidence can produce spurious matches. We cling to our illusion of certainty because the medical industry, insurance companies, investment advisers, and election campaigns have become purveyors of certainty, marketing it like a commodity.To avoid confusion, says Gigerenzer, we should rely on more understandable representations of risk, such as absolute risks. For example, it is said that a mammography screening reduces the risk of breast cancer by 25 percent. But in absolute risks, that means that out of every 1,000 women who do not participate in screening, 4 will die; while out of 1,000 women who do, 3 will die. A 25 percent risk reduction sounds much more significant than a benefit that 1 out of 1,000 women will reap.This eye-opening book explains how we can overcome our ignorance of numbers and better understand the risks we may be taking with our money, our health, and our lives.
The Advent of the Algorithm: The 300-Year Journey from an Idea to the Computer
David Berlinski - 2000
A basic idea that proved elusive for hundreds of years and bent the minds of the greatest thinkers in the world, the algorithm is what made the modern world possible. Without the algorithm, there would have been no computer, no Internet, no virtual reality, no e-mail, or any other technological advance that we rely on every day.In The Advent of the Algorithm, David Berlinski combines science, history, and math to explain and explore the intriguing story of how the algorithm was finally discovered by a succession of mathematicians and logicians, and how this paved the way for the digital age. Beginning with Leibniz and culminating in the middle of the twentieth century with the groundbreaking work of Gödel and Turing, The Advent of the Algorithm is an epic tale told with clarity and imaginative brilliance.
Bayesian Data Analysis
Andrew Gelman - 1995
Its world-class authors provide guidance on all aspects of Bayesian data analysis and include examples of real statistical analyses, based on their own research, that demonstrate how to solve complicated problems. Changes in the new edition include:Stronger focus on MCMC Revision of the computational advice in Part III New chapters on nonlinear models and decision analysis Several additional applied examples from the authors' recent research Additional chapters on current models for Bayesian data analysis such as nonlinear models, generalized linear mixed models, and more Reorganization of chapters 6 and 7 on model checking and data collectionBayesian computation is currently at a stage where there are many reasonable ways to compute any given posterior distribution. However, the best approach is not always clear ahead of time. Reflecting this, the new edition offers a more pluralistic presentation, giving advice on performing computations from many perspectives while making clear the importance of being aware that there are different ways to implement any given iterative simulation computation. The new approach, additional examples, and updated information make Bayesian Data Analysis an excellent introductory text and a reference that working scientists will use throughout their professional life.