The Non-Designer's Design Book


Robin P. Williams - 2003
    Not to worry: This book is the one place you can turn to find quick, non-intimidating, excellent design help. In The Non-Designer's Design Book, 2nd Edition, best-selling author Robin Williams turns her attention to the basic principles of good design and typography. All you have to do is follow her clearly explained concepts, and you'll begin producing more sophisticated, professional, and interesting pages immediately. Humor-infused, jargon-free prose interspersed with design exercises, quizzes, illustrations, and dozens of examples make learning a snap—which is just what audiences have come to expect from this best-selling author.

Thinking in Systems: A Primer


Donella H. Meadows - 2008
    Edited by the Sustainability Institute’s Diana Wright, this essential primer brings systems thinking out of the realm of computers and equations and into the tangible world, showing readers how to develop the systems-thinking skills that thought leaders across the globe consider critical for 21st-century life.Some of the biggest problems facing the world—war, hunger, poverty, and environmental degradation—are essentially system failures. They cannot be solved by fixing one piece in isolation from the others, because even seemingly minor details have enormous power to undermine the best efforts of too-narrow thinking.While readers will learn the conceptual tools and methods of systems thinking, the heart of the book is grander than methodology. Donella Meadows was known as much for nurturing positive outcomes as she was for delving into the science behind global dilemmas. She reminds readers to pay attention to what is important, not just what is quantifiable, to stay humble, and to stay a learner.In a world growing ever more complicated, crowded, and interdependent, Thinking in Systems helps readers avoid confusion and helplessness, the first step toward finding proactive and effective solutions.

Linear Algebra Done Right


Sheldon Axler - 1995
    The novel approach taken here banishes determinants to the end of the book and focuses on the central goal of linear algebra: understanding the structure of linear operators on vector spaces. The author has taken unusual care to motivate concepts and to simplify proofs. For example, the book presents - without having defined determinants - a clean proof that every linear operator on a finite-dimensional complex vector space (or an odd-dimensional real vector space) has an eigenvalue. A variety of interesting exercises in each chapter helps students understand and manipulate the objects of linear algebra. This second edition includes a new section on orthogonal projections and minimization problems. The sections on self-adjoint operators, normal operators, and the spectral theorem have been rewritten. New examples and new exercises have been added, several proofs have been simplified, and hundreds of minor improvements have been made throughout the text.

Deep Learning


Ian Goodfellow - 2016
    Because the computer gathers knowledge from experience, there is no need for a human computer operator to formally specify all the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones; a graph of these hierarchies would be many layers deep. This book introduces a broad range of topics in deep learning.The text offers mathematical and conceptual background, covering relevant concepts in linear algebra, probability theory and information theory, numerical computation, and machine learning. It describes deep learning techniques used by practitioners in industry, including deep feedforward networks, regularization, optimization algorithms, convolutional networks, sequence modeling, and practical methodology; and it surveys such applications as natural language processing, speech recognition, computer vision, online recommendation systems, bioinformatics, and videogames. Finally, the book offers research perspectives, covering such theoretical topics as linear factor models, autoencoders, representation learning, structured probabilistic models, Monte Carlo methods, the partition function, approximate inference, and deep generative models.Deep Learning can be used by undergraduate or graduate students planning careers in either industry or research, and by software engineers who want to begin using deep learning in their products or platforms. A website offers supplementary material for both readers and instructors.

Probability And Statistics For Engineering And The Sciences


Jay L. Devore - 1982
    In this book, a wealth of exercises are provided throughout each section, designed to reinforce learning and the logical comprehension of topics. The use of real data is incorporated much more extensively than in any other book on the market. Consist of strong coverage of computer-based methods, especially in the coverage of analysis of variance and regression. This text stresses mastery of methods most often used in medical research, with specific reference to actual medical literature and actual medical research. The approach minimizes mathematical formulation, yet gives complete explanations of all important concepts. Every new concept is systematically developed through completely worked-out examples from current medical research problems. Computer output is used to illustrate concepts when appropriate.

Contemporary Linguistics: An Introduction


William D. O'Grady - 1987
    Meticulously prepared, it is one of the most comprehensive, authoritative, up-to-date introductory resources on the market. The book’s extensive examples and exercises help students master the course material, and its lucid writing style makes complex concepts easy to understand.

Elementary Linear Algebra with Applications


Howard Anton - 1973
    It proceeds from familiar concepts to the unfamiliar, from the concrete to the abstract. Readers consistently praise this outstanding text for its expository style and clarity of presentation. The applications version features a wide variety of interesting, contemporary applications. Clear, accessible, step-by-step explanations make the material crystal clear. Established the intricate thread of relationships between systems of equations, matrices, determinants, vectors, linear transformations and eigenvalues.

A Guide to the Project Management Body of Knowledge (PMBOK® Guide)


Project Management Institute - 1995
    This internationally recognized standard provides the essential tools to practice project management and deliver organizational results.

The C Programming Language


Brian W. Kernighan - 1978
    It is the definitive reference guide, now in a second edition. Although the first edition was written in 1978, it continues to be a worldwide best-seller. This second edition brings the classic original up to date to include the ANSI standard. From the Preface: We have tried to retain the brevity of the first edition. C is not a big language, and it is not well served by a big book. We have improved the exposition of critical features, such as pointers, that are central to C programming. We have refined the original examples, and have added new examples in several chapters. For instance, the treatment of complicated declarations is augmented by programs that convert declarations into words and vice versa. As before, all examples have been tested directly from the text, which is in machine-readable form. As we said in the first preface to the first edition, C "wears well as one's experience with it grows." With a decade more experience, we still feel that way. We hope that this book will help you to learn C and use it well.

Diagnostic and Statistical Manual of Mental Disorders


American Psychiatric Association - 2013
    Their dedication and hard work have yielded an authoritative volume that defines and classifies mental disorders in order to improve diagnoses, treatment, and research.The criteria are concise and explicit, intended to facilitate an objective assessment of symptom presentations in a variety of clinical settings -- inpatient, outpatient, partial hospital, consultation-liaison, clinical, private practice, and primary care. New features and enhancements make DSM-5 easier to use across all settings:- The chapter organization reflects a lifespan approach, with disorders typically diagnosed in childhood (such as neurodevelopmental disorders) at the beginning of the manual, and those more typical of older adults (such as neurocognitive disorders) placed at the end. Also included are age-related factors specific to diagnosis. - The latest findings in neuroimaging and genetics have been integrated into each disorder along with gender and cultural considerations.- The revised organizational structure recognizes symptoms that span multiple diagnostic categories, providing new clinical insight in diagnosis. - Specific criteria have been streamlined, consolidated, or clarified to be consistent with clinical practice (including the consolidation of autism disorder, Asperger's syndrome, and pervasive developmental disorder into autism spectrum disorder; the streamlined classification of bipolar and depressive disorders; the restructuring of substance use disorders for consistency and clarity; and the enhanced specificity for major and mild neurocognitive disorders).- Dimensional assessments for research and validation of clinical results have been provided.- Both ICD-9-CM and ICD-10-CM codes are included for each disorder, and the organizational structure is consistent with the new ICD-11 in development.The Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition, is the most comprehensive, current, and critical resource for clinical practice available to today's mental health clinicians and researchers of all orientations. The information contained in the manual is also valuable to other physicians and health professionals, including psychologists, counselors, nurses, and occupational and rehabilitation therapists, as well as social workers and forensic and legal specialists.

Inorganic Chemistry


D.F. Shriver - 1990
    The bestselling textbook inorganic chemistry text on the market covers both theoretical and descriptive aspects of the subject, and emphasizes experimental methods, industrial applications, and modern topics.

The Structure of Scientific Revolutions


Thomas S. Kuhn - 1962
    The Structure of Scientific Revolutions is that kind of book. When it was first published in 1962, it was a landmark event in the history and philosophy of science. Fifty years later, it still has many lessons to teach. With The Structure of Scientific Revolutions, Kuhn challenged long-standing linear notions of scientific progress, arguing that transformative ideas don’t arise from the day-to-day, gradual process of experimentation and data accumulation but that the revolutions in science, those breakthrough moments that disrupt accepted thinking and offer unanticipated ideas, occur outside of “normal science,” as he called it. Though Kuhn was writing when physics ruled the sciences, his ideas on how scientific revolutions bring order to the anomalies that amass over time in research experiments are still instructive in our biotech age. This new edition of Kuhn’s essential work in the history of science includes an insightful introduction by Ian Hacking, which clarifies terms popularized by Kuhn, including paradigm and incommensurability, and applies Kuhn’s ideas to the science of today. Usefully keyed to the separate sections of the book, Hacking’s introduction provides important background information as well as a contemporary context.  Newly designed, with an expanded index, this edition will be eagerly welcomed by the next generation of readers seeking to understand the history of our perspectives on science.

Linear Algebra and Its Applications [with CD-ROM]


David C. Lay - 1993
    

Causality: Models, Reasoning, and Inference


Judea Pearl - 2000
    It shows how causality has grown from a nebulous concept into a mathematical theory with significant applications in the fields of statistics, artificial intelligence, philosophy, cognitive science, and the health and social sciences. Pearl presents a unified account of the probabilistic, manipulative, counterfactual and structural approaches to causation, and devises simple mathematical tools for analyzing the relationships between causal connections, statistical associations, actions and observations. The book will open the way for including causal analysis in the standard curriculum of statistics, artifical intelligence, business, epidemiology, social science and economics. Students in these areas will find natural models, simple identification procedures, and precise mathematical definitions of causal concepts that traditional texts have tended to evade or make unduly complicated. This book will be of interest to professionals and students in a wide variety of fields. Anyone who wishes to elucidate meaningful relationships from data, predict effects of actions and policies, assess explanations of reported events, or form theories of causal understanding and causal speech will find this book stimulating and invaluable. Professor of Computer Science at the UCLA, Judea Pearl is the winner of the 2008 Benjamin Franklin Award in Computers and Cognitive Science.

Information Theory, Inference and Learning Algorithms


David J.C. MacKay - 2002
    These topics lie at the heart of many exciting areas of contemporary science and engineering - communication, signal processing, data mining, machine learning, pattern recognition, computational neuroscience, bioinformatics, and cryptography. This textbook introduces theory in tandem with applications. Information theory is taught alongside practical communication systems, such as arithmetic coding for data compression and sparse-graph codes for error-correction. A toolbox of inference techniques, including message-passing algorithms, Monte Carlo methods, and variational approximations, are developed alongside applications of these tools to clustering, convolutional codes, independent component analysis, and neural networks. The final part of the book describes the state of the art in error-correcting codes, including low-density parity-check codes, turbo codes, and digital fountain codes -- the twenty-first century standards for satellite communications, disk drives, and data broadcast. Richly illustrated, filled with worked examples and over 400 exercises, some with detailed solutions, David MacKay's groundbreaking book is ideal for self-learning and for undergraduate or graduate courses. Interludes on crosswords, evolution, and sex provide entertainment along the way. In sum, this is a textbook on information, communication, and coding for a new generation of students, and an unparalleled entry point into these subjects for professionals in areas as diverse as computational biology, financial engineering, and machine learning.