Introduction to Graph Theory


Douglas B. West - 1995
    Verification that algorithms work is emphasized more than their complexity. An effective use of examples, and huge number of interesting exercises, demonstrate the topics of trees and distance, matchings and factors, connectivity and paths, graph coloring, edges and cycles, and planar graphs. For those who need to learn to make coherent arguments in the fields of mathematics and computer science.

Diagnostic and Statistical Manual of Mental Disorders DSM-IV-TR


American Psychiatric Association - 1952
    Updated information is included about the associated features, culture, age & gender features, prevalence, course & familial pattern of mental disorders. The DSM-IV-TR(R) brings this essential diagnostic tool to-date, to promote effective diagnosis, treatment & quality of care. One can get all the essential diagnostic information from the DSM-IV(R) along with important updates not in the '94 edition. Benefit from new research into Schizophrenia, Asperger's Disorder & other conditions. Utilize additional information about the epidemiology & other facets of DSM conditions. Update ICD-9-CM codes implemented since 1994 including Conduct Disorder, Dementia, Somatoform Disorders.Use of the manual DSM-IV-TR classification Multiaxial assessment Disorders usually 1st diagnosed in infancy, childhood or adolescenceDelirium, dementia & amnestic & other cognitive disordersMental disorders due to a general medical condition Substance-related disorders Schizophrenia & other psychotic disordersMood disorders Anxiety disordersSomatoform disordersFactitious disordersDissociative disordersSexual & gender identity disordersEating disorders Sleep disorders Impulse-control disorders not elsewhere classifiedAdjustment disordersPersonality disordersOther conditions that may be a focus of clinical attentionAdditional codes Appendix A: Decision trees for differential diagnosis Appendix B: Criteria sets & axes provided for further study Appendix C: Glossary of technical terms Appendix D: Highlights of changes in DSM-IV text revisionAppendix E: Alphabetical listing of DSM-IV-TR diagnoses & codesAppendix F: Numerical listing of DSM-IV-TR diagnoses & codesAppendix G: ICD-9-CM codes for selected general medical conditions & medication-induced disordersAppendix H: DSM-IV classification with ICD-10 codesAppendix I: Outline for cultural formulation & glossary of culture-bound syndromesAppendix J: DSM-IV contributorsAppendix K: DSM-IV text revision advisers

Modern Operating Systems


Andrew S. Tanenbaum - 1992
    What makes an operating system modern? According to author Andrew Tanenbaum, it is the awareness of high-demand computer applications--primarily in the areas of multimedia, parallel and distributed computing, and security. The development of faster and more advanced hardware has driven progress in software, including enhancements to the operating system. It is one thing to run an old operating system on current hardware, and another to effectively leverage current hardware to best serve modern software applications. If you don't believe it, install Windows 3.0 on a modern PC and try surfing the Internet or burning a CD. Readers familiar with Tanenbaum's previous text, Operating Systems, know the author is a great proponent of simple design and hands-on experimentation. His earlier book came bundled with the source code for an operating system called Minux, a simple variant of Unix and the platform used by Linus Torvalds to develop Linux. Although this book does not come with any source code, he illustrates many of his points with code fragments (C, usually with Unix system calls). The first half of Modern Operating Systems focuses on traditional operating systems concepts: processes, deadlocks, memory management, I/O, and file systems. There is nothing groundbreaking in these early chapters, but all topics are well covered, each including sections on current research and a set of student problems. It is enlightening to read Tanenbaum's explanations of the design decisions made by past operating systems gurus, including his view that additional research on the problem of deadlocks is impractical except for "keeping otherwise unemployed graph theorists off the streets." It is the second half of the book that differentiates itself from older operating systems texts. Here, each chapter describes an element of what constitutes a modern operating system--awareness of multimedia applications, multiple processors, computer networks, and a high level of security. The chapter on multimedia functionality focuses on such features as handling massive files and providing video-on-demand. Included in the discussion on multiprocessor platforms are clustered computers and distributed computing. Finally, the importance of security is discussed--a lively enumeration of the scores of ways operating systems can be vulnerable to attack, from password security to computer viruses and Internet worms. Included at the end of the book are case studies of two popular operating systems: Unix/Linux and Windows 2000. There is a bias toward the Unix/Linux approach, not surprising given the author's experience and academic bent, but this bias does not detract from Tanenbaum's analysis. Both operating systems are dissected, describing how each implements processes, file systems, memory management, and other operating system fundamentals. Tanenbaum's mantra is simple, accessible operating system design. Given that modern operating systems have extensive features, he is forced to reconcile physical size with simplicity. Toward this end, he makes frequent references to the Frederick Brooks classic The Mythical Man-Month for wisdom on managing large, complex software development projects. He finds both Windows 2000 and Unix/Linux guilty of being too complicated--with a particular skewering of Windows 2000 and its "mammoth Win32 API." A primary culprit is the attempt to make operating systems more "user-friendly," which Tanenbaum views as an excuse for bloated code. The solution is to have smart people, the smallest possible team, and well-defined interactions between various operating systems components. Future operating system design will benefit if the advice in this book is taken to heart. --Pete Ostenson

Machine Learning


Tom M. Mitchell - 1986
    Mitchell covers the field of machine learning, the study of algorithms that allow computer programs to automatically improve through experience and that automatically infer general laws from specific data.

A Manual for Writers of Research Papers, Theses, and Dissertations: Chicago Style for Students and Researchers


Kate L. Turabian - 1955
    Bellow. Strauss. Friedman. The University of Chicago has been the home of some of the most important thinkers of the modern age. But perhaps no name has been spoken with more respect than Turabian. The dissertation secretary at Chicago for decades, Kate Turabian literally wrote the book on the successful completion and submission of the student paper. Her Manual for Writers of Research Papers, Theses, and Dissertations, created from her years of experience with research projects across all fields, has sold more than seven million copies since it was first published in 1937.Now, with this seventh edition, Turabian’s Manual has undergone its most extensive revision, ensuring that it will remain the most valuable handbook for writers at every level—from first-year undergraduates, to dissertation writers apprehensively submitting final manuscripts, to senior scholars who may be old hands at research and writing but less familiar with new media citation styles. Gregory G. Colomb, Joseph M. Williams, and the late Wayne C. Booth—the gifted team behind The Craft of Research—and the University of Chicago Press Editorial Staff combined their wide-ranging expertise to remake this classic resource. They preserve Turabian’s clear and practical advice while fully embracing the new modes of research, writing, and source citation brought about by the age of the Internet.Booth, Colomb, and Williams significantly expand the scope of previous editions by creating a guide, generous in length and tone, to the art of research and writing. Growing out of the authors’ best-selling Craft of Research, this new section provides students with an overview of every step of the research and writing process, from formulating the right questions to reading critically to building arguments and revising drafts. This leads naturally to the second part of the Manual for Writers, which offers an authoritative overview of citation practices in scholarly writing, as well as detailed information on the two main citation styles (“notes-bibliography” and “author-date”). This section has been fully revised to reflect the recommendations of the fifteenth edition of The Chicago Manual of Style and to present an expanded array of source types and updated examples, including guidance on citing electronic sources.The final section of the book treats issues of style—the details that go into making a strong paper. Here writers will find advice on a wide range of topics, including punctuation, table formatting, and use of quotations. The appendix draws together everything writers need to know about formatting research papers, theses, and dissertations and preparing them for submission. This material has been thoroughly vetted by dissertation officials at colleges and universities across the country.This seventh edition of Turabian’s Manual for Writers of Research Papers, Theses, and Dissertations is a classic reference revised for a new age. It is tailored to a new generation of writers using tools its original author could not have imagined—while retaining the clarity and authority that generations of scholars have come to associate with the name Turabian.

The Elements of Statistical Learning: Data Mining, Inference, and Prediction


Trevor Hastie - 2001
    With it has come vast amounts of data in a variety of fields such as medicine, biology, finance, and marketing. The challenge of understanding these data has led to the development of new tools in the field of statistics, and spawned new areas such as data mining, machine learning, and bioinformatics. Many of these tools have common underpinnings but are often expressed with different terminology. This book describes the important ideas in these areas in a common conceptual framework. While the approach is statistical, the emphasis is on concepts rather than mathematics. Many examples are given, with a liberal use of color graphics. It should be a valuable resource for statisticians and anyone interested in data mining in science or industry. The book's coverage is broad, from supervised learning (prediction) to unsupervised learning. The many topics include neural networks, support vector machines, classification trees and boosting—the first comprehensive treatment of this topic in any book. Trevor Hastie, Robert Tibshirani, and Jerome Friedman are professors of statistics at Stanford University. They are prominent researchers in this area: Hastie and Tibshirani developed generalized additive models and wrote a popular book of that title. Hastie wrote much of the statistical modeling software in S-PLUS and invented principal curves and surfaces. Tibshirani proposed the Lasso and is co-author of the very successful An Introduction to the Bootstrap. Friedman is the co-inventor of many data-mining tools including CART, MARS, and projection pursuit.

Statistics Done Wrong: The Woefully Complete Guide


Alex Reinhart - 2013
    Politicians and marketers present shoddy evidence for dubious claims all the time. But smart people make mistakes too, and when it comes to statistics, plenty of otherwise great scientists--yes, even those published in peer-reviewed journals--are doing statistics wrong."Statistics Done Wrong" comes to the rescue with cautionary tales of all-too-common statistical fallacies. It'll help you see where and why researchers often go wrong and teach you the best practices for avoiding their mistakes.In this book, you'll learn: - Why "statistically significant" doesn't necessarily imply practical significance- Ideas behind hypothesis testing and regression analysis, and common misinterpretations of those ideas- How and how not to ask questions, design experiments, and work with data- Why many studies have too little data to detect what they're looking for-and, surprisingly, why this means published results are often overestimates- Why false positives are much more common than "significant at the 5% level" would suggestBy walking through colorful examples of statistics gone awry, the book offers approachable lessons on proper methodology, and each chapter ends with pro tips for practicing scientists and statisticians. No matter what your level of experience, "Statistics Done Wrong" will teach you how to be a better analyst, data scientist, or researcher.

Language, Proof and Logic: Text and CD


Jon Barwise - 1999
    The unique on-line grading services instantly grades solutions to hundred of computer exercises. It is specially devised to be used by philosophy instructors in a way that is useful to undergraduates of philosophy, computer science, mathematics, and linguistics.The book is a completely rewritten and much improved version of The Language of First-order Logic. Introductory material is presented in a more systematic and accessible fashion. Advanced chapters include proofs of soundness and completeness for propositional and predicate logic, as well as an accessible sketch of Godel's first incompleteness theorem. The book is appropriate for a wide range of courses, from first logic courses for undergraduates (philosophy, mathematics, and computer science) to a first graduate logic course.The package includes four pieces of software:Tarski's World 5.0, a new version of the popular program that teaches the basic first-order language and its semantics; Fitch, a natural deduction proof environment for giving and checking first-order proofs;Boole, a program that facilitates the construction and checking of truth tables and related notions (tautology, tautological consequence, etc.);Submit, a program that allows students to submit exercises done with the above programs to the Grade Grinder, the automatic grading service.Grade reports are returned to the student and, if requested, to the student's instructor, eliminating the need for tedious checking of homework. All programs are available for Windows, Macintosh and Linux systems.Instructors do not need to use the programs themselves in order to be able to take advantage of their pedagogical value. More about the software can be found at lpl.stanford.edu.The price of a new text/software package includes one Registration ID, which must be used each time work is submitted to the grading service. Once activated, the Registration ID is not transferable.

e: the Story of a Number


Eli Maor - 1993
    Louis are all intimately connected with the mysterious number e. In this informal and engaging history, Eli Maor portrays the curious characters and the elegant mathematics that lie behind the number. Designed for a reader with only a modest mathematical background, this biography brings out the central importance of e to mathematics and illuminates a golden era in the age of science.

Electronic Principles


Albert Paul Malvino - 1979
    It's been updated to keep coverage in step with the fast-changing world of electronics. Yet, it retains Malvino's clear writing style, supported throughout by abundant illustrations and examples.

Problem Solving with C++: The Object of Programming


Walter J. Savitch - 1995
    It introduces the use of classes; shows how to write ADTs that maximize the perfomance of C++ in creating reusable code; and provides coverage of all important OO functions, including inheritance, polymorphism and encapsulation.

Fundamentals of Astrodynamics


Roger R. Bate - 1971
    Air Force Academy and designed as a first course emphasizes the universal variable formulation. Develops the basic two-body and n-body equations of motion; orbit determination; classical orbital elements, coordinate transformations; differential correction; more. Includes specialized applications to lunar and interplanetary flight, example problems, exercises. 1971 edition.

Introduction to Mathematical Thinking


Keith Devlin - 2012
    This is not the same as “doing math.” The latter usually involves the application of formulas, procedures, and symbolic manipulations; mathematical thinking is a powerful way of thinking about things in the world -- logically, analytically, quantitatively, and with precision. It is not a natural way of thinking, but it can be learned. Mathematicians, scientists, and engineers need to “do math,” and it takes many years of college-level education to learn all that is required. Mathematical thinking is valuable to everyone, and can be mastered in about six weeks by anyone who has completed high school mathematics. Mathematical thinking does not have to be about mathematics at all, but parts of mathematics provide the ideal target domain to learn how to think that way, and that is the approach taken by this short but valuable book. The book is written primarily for first and second year students of science, technology, engineering, and mathematics (STEM) at colleges and universities, and for high school students intending to study a STEM subject at university. Many students encounter difficulty going from high school math to college-level mathematics. Even if they did well at math in school, most are knocked off course for a while by the shift in emphasis, from the K-12 focus on mastering procedures to the “mathematical thinking” characteristic of much university mathematics. Though the majority survive the transition, many do not. To help them make the shift, colleges and universities often have a “transition course.” This book could serve as a textbook or a supplementary source for such a course. Because of the widespread applicability of mathematical thinking, however, the book has been kept short and written in an engaging style, to make it accessible to anyone who seeks to extend and improve their analytic thinking skills. Going beyond a basic grasp of analytic thinking that everyone can benefit from, the STEM student who truly masters mathematical thinking will find that college-level mathematics goes from being confusing, frustrating, and at times seemingly impossible, to making sense and being hard but doable. Dr. Keith Devlin is a professional mathematician at Stanford University and the author of 31 previous books and over 80 research papers. His books have earned him many awards, including the Pythagoras Prize, the Carl Sagan Award, and the Joint Policy Board for Mathematics Communications Award. He is known to millions of NPR listeners as “the Math Guy” on Weekend Edition with Scott Simon. He writes a popular monthly blog “Devlin’s Angle” for the Mathematical Association of America, another blog under the name “profkeithdevlin”, and also blogs on various topics for the Huffington Post.

Think Python


Allen B. Downey - 2002
    It covers the basics of computer programming, including variables and values, functions, conditionals and control flow, program development and debugging. Later chapters cover basic algorithms and data structures.

Analytical Mechanics


Grant R. Fowles - 1970
    This book includes discussions which aid in student understanding of theoretical material through the use of specific cases. It is suitable for undergraduate Mechanics course.