Software-Kanban beruht auf der japanischen Produktionsmethode Kanban, die zuerst erfolgreich in der Automobilfertigung bei Toyota zum Einsatz kam, jetzt aber umfassend für die spezifischen Bedürfnisse der Softwareentwicklung angepasst wurde. Software-Kanban lässt sich evolutionär in kleinen Schritten einführen und führt schnell zu kürzeren Durchlaufzeiten, besserer Qualität, gleichmäßiger Arbeitsbelastung und höherer Kundenzufriedenheit. So können Softwareentwicklungsprozesse kontinuierlich verbessert werden, denn die Veränderung erfolgt sanft, vorsichtig und fließend. In diesem Buch hat David J. Anderson, der ´´Vater´´ von Software-Kanban, seine Erfahrungen mit dem Einsatz dieser evolutionären Methode aus verschiedenen großen Softwareprojekten (Microsoft, Motorola) zusammengetragen. Es behandelt alle grundlegenden Themen rund um Software-Kanban. Illustriert wird das Buch durch verschiedene reale Praxisbeispiele, in denen der Autor Software-Kanban erfolgreich angewendet hat.
Data stewards in business and IT are the backbone of a successful data governance implementation because they do the work to make a company´s data trusted, dependable, and high quality. Data Stewardship explains everything you need to know to successfully implement the stewardship portion of data governance, including how to organize, train, and work with data stewards, get high-quality business definitions and other metadata, and perform the day-to-day tasks using a minimum of the steward´s time and effort. David Plotkin has loaded this book with practical advice on stewardship so you can get right to work, have early successes, and measure and communicate those successes, gaining more support for this critical effort. Provides clear and concise practical advice on implementing and running data stewardship, including guidelines on how to organize based on company structure, business functions, and data ownership Shows how to gain support for your stewardship effort, maintain that support over the long-term, and measure the success of the data stewardship effort and report back to management Includes detailed lists of responsibilities for each type of data steward and strategies to help the Data Governance Program Office work effectively with the data stewards
We´re losing hundreds of billions of dollars a year on broken software, and great new ideas such as agile development and Scrum don´t always pay off. But there´s hope. The nine software development practices in Beyond Legacy Code are designed to solve the problems facing our industry.
In this practical guide, machine learning and security specialists Clarence Chio and David Freeman provide a framework for discussing the marriage of these two fields, as well as a toolkit of machine-learning algorithms that you can apply to an array of security problems.
Mit der deutschen Übersetzung zur fünfter Auflage des amerikanischen Klassikers Computer Organization and Design - The Hardware/Software Interface ist das Standardwerk zur Rechnerorganisation wieder auf dem neusten Stand - David A. Patterson und John L. Hennessy gewähren die gewohnten Einblicke in das Zusammenwirken von Hard- und Software, Leistungseinschätzungen und zahlreicher Rechnerkonzepte in einer Tiefe, die zusammen mit klarer Didaktik und einer eher lockeren Sprache den Erfolg dieses weltweit anerkannten Standardwerks begründen. Patterson und Hennessy achten darauf, nicht nur auf das ´´Wie´´ der dargestellten Konzepte, sondern auch auf ihr ´´Warum´´ einzugehen und zeigen damit Gründe für Veränderungen und neue Entwicklungen auf. Jedes der Kapitel steht für einen deutlich umrissenen Teilbereich der Rechnerorganisation und ist jeweils gleich aufgebaut: Eine Einleitung, gefolgt von immer tiefgreifenderen Grundkonzepten mit steigernder Komplexität. Darauf eine aktuelle Fallstudie, ´´Fallstricke und Fehlschlüsse´´, Zusammenfassung und Schlussbetrachtung, historische Perspektiven und Literaturhinweise sowie Aufgaben. In der neuen Auflage sind die Inhalte in den Kapiteln 1-5 an vielen Stellen punktuell verbessert und aktualisiert, mit der Vorstellung neuerer Prozessoren worden, und der Kapitel 6... from Client to Cloud wurde stark überarbeitetUmfangreiches Zusatzmaterial (Werkzeuge mit Tutorien etc.) stehtOnline zur Verfügung.
Continuous Delivery shows how tocreate fully automated, repeatable, and reliable processes for rapidly moving changes through build, deploy, test, and release. Using these techniques, software organizations are getting critical fixes and other new releases into production in hours - sometimes even minutes - evenin large projects with complex code bases. Jez Humble and David Farley begin by presenting the high-level principles and practices required to succeed with regular, repeatable, low-risk releases. Next, they introduce the ´´deployment pipeline,´´ an automated process for managing all changes, from check-in to release. Finally, they discuss the ´´ecosystem´´ needed to support deployment pipelines, from infrastructure to data management and governance. The authors introduce many state-of-the-art techniques, including in-production monitoring and tracing, dependency management, and the use of virtualization. For each, they review key issues, demonstrate how to mitigate risks, and identify best practices. Coverage includes · Overcoming ´´anti-patterns´´ that slow down releases and reduce quality · Automating all facets of configuration management and testing · Implementing deployment pipelines at team and organizational levels · Scripting highly-effective automated build and deployment processes · Triggering automated processes whenever a change is made · Automating acceptance testing, from analysis to implementation · Testing capacity and other non-functional requirements · Utilizing continuous deployment, rollbacks, and zero-downtime releases · Managing infrastructure, data, components, dependencies, and versions · Navigating risk management, compliance, and other obstacles Whether you´re a developer, architect, tester, or manager, this book will help you move from idea to release faster than ever - so you can deliver far more value, far more rapidly. Product Description Winner of the 2011 Jolt Excellence Award! Getting software released to users is often a painful, risky, and time-consuming process. This groundbreaking new book sets out the principles and technical practices that enable rapid, incremental delivery of high quality, valuable new functionality to users. Through automation of the build, deployment, and testing process, and improved collaboration between developers, testers, and operations, delivery teams can get changes released in a matter of hours- sometimes even minutes-no matter what the size of a project or the complexity of its code base. Jez Humble and David Farley begin by presenting the foundations of a rapid, reliable, low-risk delivery process. Next, they introduce the ´´deployment pipeline,´´ an automated process for managing all changes, from check-in to release. Finally, they discuss the ´´ecosystem´´ needed to support continuous delivery, from infrastructure, data and configuration management to governance. The authors introduce state-of-the-art techniques, including automated infrastructure management and data migration, and the use of virtualization. For each, they review key issues, identify best practices, and demonstrate how to mitigate risks. Coverage includes . Automating all facets of building, integrating, testing, and deploying software . Implementing deployment pipelines at team and organizational levels . Improving collaboration between developers, testers, and operations . Developing features incrementally on large and distributed teams . Implementing an effective configuration management strategy . Automating acceptance testing, from analysis to implementation . Testing capacity and other non-functional requirements . Implementing continuous deployment and zero-downtime releases . Managing infrastructure, data, components and dependencies . Navigating risk management, compliance, and auditing Whether you´re a developer, systems administrator, tester, or manager, this book will help your organization move from idea to release faster than ever-so you can deliver value to your business rapidly and reliably. Backcover Getting software released to users is often a painful, risky, and time-consuming process. This groundbreaking new book sets out the principles and technical practices that enable rapid, incremental delivery of high quality, valuable new functionality to users. Through automation of the build, deployment, and testing process, and improved collaboration between developers, testers, and operations, delivery teams can get changes released in a matter of hours-
Computer Architecture: A Quantitative Approach, Sixth Edition has been considered essential reading by instructors, students and practitioners of computer design for over 20 years. The sixth edition of this classic textbook from Hennessy and Patterson, winners of the 2017 ACM A.M. Turing Award recognizing contributions of lasting and major technical importance to the computing field, is fully revised with the latest developments in processor and system architecture. The text now features examples from the RISC-V (RISC Five) instruction set architecture, a modern RISC instruction set developed and designed to be a free and openly adoptable standard. It also includes a new chapter on domain-specific architectures and an updated chapter on warehouse-scale computing that features the first public information on Google´s newest WSC. True to its original mission of demystifying computer architecture, this edition continues the longstanding tradition of focusing on areas where the most exciting computing innovation is happening, while always keeping an emphasis on good engineering design. Winner of a 2019 Textbook Excellence Award (Texty) from the Textbook and Academic Authors Association Includes a new chapter on domain-specific architectures, explaining how they are the only path forward for improved performance and energy efficiency given the end of Moore´s Law and Dennard scaling Features the first publication of several DSAs from industry Features extensive updates to the chapter on warehouse-scale computing, with the first public information on the newest Google WSC Offers updates to other chapters including new material dealing with the use of stacked DRAM; data on the performance of new NVIDIA Pascal GPU vs. new AVX-512 Intel Skylake CPU; and extensive additions to content covering multicore architecture and organization Includes ´´Putting It All Together´´ sections near the end of every chapter, providing real-world technology examples that demonstrate the principles covered in each chapter Includes review appendices in the printed text and additional reference appendices available online Includes updated and improved case studies and exercises ACM named John L. Hennessy and David A. Patterson, recipients of the 2017 ACM A.M. Turing Award for pioneering a systematic, quantitative approach to the design and evaluation of computer architectures with enduring impact on the microprocessor industry
Artificial intelligence, including machine learning, has emerged as a transformational science and engineering discipline. Artificial Intelligence: Foundations of Computational Agents presents AI using a coherent framework to study the design of intelligent computational agents. By showing how the basic approaches fit into a multidimensional design space, readers learn the fundamentals without losing sight of the bigger picture. The new edition also features expanded coverage on machine learning material, as well as on the social and ethical consequences of AI and ML. The book balances theory and experiment, showing how to link them together, and develops the science of AI together with its engineering applications. Although structured as an undergraduate and graduate textbook, the book´s straightforward, self-contained style will also appeal to an audience of professionals, researchers, and independent learners. The second edition is well-supported by strong pedagogical features and online resources to enhance student comprehension.