Nicht aus der Schweiz? Besuchen Sie lehmanns.de
The Mikado Method - Ola Ellnestam, Daniel Brolund

The Mikado Method

Buch | Softcover
217 Seiten
2014
Manning Publications (Verlag)
978-1-61729-121-0 (ISBN)
CHF 61,30 inkl. MwSt
  • Understand your technical debt
  • Surface the dependencies in legacy systems
  • Isolate and resolve core concerns while creating minimal disruption
  • Create a roadmap for your changes

»The Mikado Method« is a book written by the creators of this process.

It describes a pragmatic, straightforward, and empirical method to plan and perform non-trivial technical improvements on an existing software system. The method has simple rules, but the applicability is vast.

As you read, you'll practice a step-by-step system for identifying the scope and nature of your technical debt, mapping the key dependencies, and determining the safest way to approach the »Mikado«—your goal.
Technical debt is best understood as the work remaining before a job can be considered complete. It is one of the worst bottlenecks when it comes to creating value with software and the ability to deal with it is often the difference between a great business and closing the shop.
The Mikado Method is a process which systematically eliminates technical debt. It gets its name from a simple game known as »pick-up sticks.«

Starting with a jumbled pile of sticks, the goal is to remove the Mikado, or Emperor, stick without disturbing the others. Players carefully remove sticks one at a time, leaving the rest of the heap intact, slowly exposing the Mikado. The game is a great metaphor for eliminating technical debt-carefully extracting each intertwined dependency until successfully resolving the central issue.

»The Mikado Method« describes a pragmatic and straightforward method to plan and perform non-trivial technical improvements on an existing software system. Step by step, readers will identify the scope and nature of technical debt, map the key dependencies, and determine the safest way to approach the »Mikado«-the goal.

A natural byproduct of this process is the Mikado Graph, a minimalistic, relevant, just-in-time roadmap and information radiator that reflects deep understanding of how the system works.
  • Written by the creators of the Mikado Method
  • Step-by-step guide with hands on examples
  • Points out sources of technical debts


This book is for software developers with a basic understanding of Java or familiarity with a C-like language. No previous experience with the Mikado Method is necessary.

The Mikado Method offers a way to improve software systems without breaking them or going through long periods of having a half functioning system.

Ola Ellnestam is a coach and mentor for both business and technical teams. He loves to combine technology, people and business. He has developed complex computer systems within health care, defense, and on-line banking. He knows that software must be easy to use, extend, and deploy in order to be worth developing. More than anything else, he likes to share his discoveries and knowledge with others because he believes that this is how new knowledge and insight is created.

Daniel Brolund is a software developer that always sees things to improve-to the joy and grief of his fellow workers. He has successfully worked with global web sites deployed on hundreds of servers, desktop applications for just a few users, and on-line gaming applications, just to mention a few.

foreword
preface
acknowledgments
about this book
about the cover illustration
about the authors
Part 1 The basics of the Mikado Method

1 Meet the Mikado Method
1.1 What is the Mikado Method?
1.2 The Mikado Method recipe
1.3 Summary
2 Hello, Mikado Method!
2.1 Your first Mikado Graph
2.2 A slightly more complicated change
2.3 Making the code configurable
2.4 Summary
3 Goals, graphs, and guidelines
3.1 The Mikado Method guidelines
3.2 The Mikado Graph
3.3 Tips for using the Mikado Method
3.4 Relation to other thinking models
3.5 Summary
4 Organizing your work
4.1 The scope of a change
4.2 How to approach a change
4.3 Summary

Part 2 Principles and patterns for improving software

5 Breaking up a monolith
5.1 The code of an online loan institute
5.2 Beheading the beast
5.3 Getting to the true leaves of the graph
5.4 Summary
6 Emergent design
6.1 Designing classes
6.2 Designing packages
6.3 Summary
7 Common restructuring patterns
7.1 Graph patterns
7.2 Scattered-code patterns
7.3 Code tricks
7.4 Summary

appendix A Technical debt
appendix B Setting the stage for improvements
appendix C Dealing with dynamically typed languages
index

Most software discussions, books, and articles seem to assume that development work starts with an empty codebase. The Agile literature usually presumes that the detailed learning about what is needed to solve a problem for a stakeholder community can be discovered iteratively and incrementally. This might be a valid assumption if one is starting fresh, but I seldom encounter teams with the opportunity of starting from scratch; there are always constraints. Most investment in software today involves modifications or extensions to existing applications and environments. Thus, in addition to discovering and implementing solutions to our organizational problems/opportunities, we have the constraint of fitting into the environment created by earlier development teams. We need to discover and codify not just the applicable domain knowledge, policies, and organizational goals that drive our current development work, but also to understand how the changes we make affect the existing application environment. Everyone who has done this kind of work knows that the information we need is only weakly represented in the documentation left behind by preceding teams. One has to look at the source code to get reliable information about the constraints it imposes on additions and changes. If the source code includes an effective set of automated tests, we’re in luck because tests illustrate the behavior required of the code by previous implementers to solve previous problems. If these tests pass when executed, we know that the code behaves as they expected. More often, automated tests were never created, and we’re left with only the source code itself. The question then becomes how to learn what we need to know to avoid breaking previous work. The barrier is sheer complexity. Analysis of the preexisting code has proven to be a weak tool for tackling this complexity. Much of the code we have to confront no longer communicates effectively to us the nuances we need to understand to avoid breaking it as we implement our changes, and it usually ends up taking far more time than we can afford. Therefore, after we exhaust our patience with analysis, we’re forced to go ahead and take a chance at breaking the fragile existing code. What we do next is critical. “Traditional” approaches have ended with a long, tedious, unpredictable, test-and-fix period that often consumes the second and sometimes third 90% of our project schedule. Agile and lean thinking have taught us to test immediately, integrate continuously, and fix every problem as soon as we discover it. This works well with well- structured code that’s weakly dependent on its environment—provided we have an effective automated test suite. The teams who preceded us, however, have seldom had the perspective, skills, or inclination to leave us with such a foundation, because they were driven mostly by the need to get-it-done (project scope focus) even at the expense of keep-it-clean (longer-term product lifecycle focus). So we find ourselves in a hole! The first rule of holes is this: “When you find yourself in a hole, stop digging!” Agile thinking suggests that for all the code you add to an application, you should use your refactoring, clean coding, and TDD practices to ensure that you don’t make the situation worse. You’ll probably have to add some higher-level automated functional tests to be able to do even this safely. Even then, complexity easily becomes overwhelming, and quality, speed, and cost all suffer. Managing this unavoidable complexity is what this book is about. Rather than permitting complexity to cascade, The Mikado Method enables you to discover what you need to know and address it in manageable pieces by maintaining your focus on always having a known good code base. Ultimately, this is the deepest core of Agile—always ensure that the work you have done so far is correct. While the form of the Mikado Method is simple, it’s tied closely to and greatly helps with the application of principles and practices we’ve learned over the last decade. It addresses individual, pair, and team-wide practices to reliably tackle changes in small steps. Ubiquitous language and the SOLID and DRY principles characteristic of clean coding practices guide our Mikado Goals and the trees of prerequisites that structure our work. Mikado Graphs help keep our packaging, dependency, TDD, and refactoring work properly focused. The initial focus of the Mikado Method is effectively dealing with a messy reality. The long-term goal, however, is to understand the forces that drive organizations to create bad code in the first place so we can avoid creating more code that our successors will have to struggle with unhappily. When we’re confronted with the normal levels of complexity in today’s software, we can’t keep everything we need in our heads at once. People have discovered many ways of effectively working together in complex tasks, starting with the scientific method and extending to the Theory of Constraints, empirical control systems, and pull/flow systems. The Mikado Method’s approach works with all these strategies to help us stay in control. Tom Poppendieck

We once worked on an application that was used to configure the behavior of a specialized electronics module in an expensive hardware product. Our application was a part of the delivery of that product. One day, news reached our software development team that Sales had unexpectedly landed a contract with a new client. At first we didn’t know what to make of it, but in time we learned that the new client needed a version of our application that was almost identical to the one we had. This was fortunate. But after further investigation, we realized that there were certain parts of the application code that under no circumstances could be shared between the new client and our old ones. We had to come up with a solution to this problem, or we couldn’t deliver to the new client at all. It finally boiled down to two possible solutions. We could duplicate the whole application, or we could restructure first and break out the parts that couldn’t be shared. Our team argued for well over a week before we reached consensus. We decided that the best solution was the restructuring approach: restructure first, and then add the new functionality. This shouldn’t take more than a month, or maybe two, we thought to ourselves. We started working, and guess what? The work was more complex than we could imagine! The approach we chose was supposed to be simple, and we thought we could get away with moving some classes to new packages, fixing the compilation errors, and then adding the new functionality. But as we tried to split the application and separate sensitive information from the more general parts, we ran into trouble. We had anticipated some trouble, but not that much. After every change we made, we got about 20 compiler errors, and as we tried to fix them, a myriad of new errors came up. When we were just about done with the new errors, the compiler moved ahead a bit, only to reveal another round of errors. Fixing those errors led to more restructurings and of course more errors. More than two weeks went by, and we had well over a thousand files checked out and more than two hundred compiler errors. By that time, the high-level system tests, our indicator of a healthy application, had long since been put out of play because we couldn’t even compile the application. To make matters worse, the rest of the team was moving along at their regular pace, and our efforts at keeping the code in sync became more and more difficult. It became obvious that we couldn’t keep up with the rest of the team, and that the complexity of the change was more than overwhelming. We were trying to behead a Software Hydra—for every head we cut off, two more grew out. The Software Hydra—for every head we cut off, two more grew out. We still remember the day when we swallowed our pride and made the tough decision to revert everything we had done so far. Only later did it dawn on us that reverting the code was a defining moment in our software careers. Apart from realizing that it’s never too late to turn back, we had also discovered how to approach our particular problem, and other problems like it. Our insight was that in order to get the system in shape, we had to do almost all of the things we had already done, but we needed to do them in the opposite order. You might think this sounds crazy, but the key was to take the errors and turn them into prerequisites. Brownfield development You’ve likely worked with systems like the one we described—software hydras that aren’t easy to behead. For us, this is more or less the norm. We’ve had to clean up lots of messy code, refactor smaller and larger parts of systems, and restructure complex architecture. We’ve even tried rewriting code from scratch, also known as greenfield development. But we’ve noticed that the field doesn’t stay green very long. We’ve long since faced it; we’re stuck with brownfield development, whether we like it or not. We, the developers, hold the fate of the code in our hands, and we’re the only ones who have the power to improve it. It’s our responsibility to keep the code clean and fit for its purpose. This means we have to be able to add code and improve our own code and the code of others. If computer programs don’t improve, they’re doomed to a slow death. How do we approach large improvement efforts? How do we deliver while the pressure is high, because we must deploy functionality once a month, or even more frequently than that? How do we deal with huge systems that are hard to comprehend, and whose complexity makes it hard to keep all the necessary details in our heads at once? We morph systems into a new shape! The goal: morphing a system Most organizations that deliver software don’t enjoy halting the development of new functionality. The stakeholders want to see value coming out as money is put in. As a result, software developers have to be able to cope with the current situation as future problems are encountered, solved, and sometimes avoided entirely. If we want to be successful software developers, we need to learn how to morph an existing system into a desired new shape. When we say morph, we mean perform the necessary actions and steps that take us from a system of one shape to another without breaking things in between. The transitions between the different states need to be smooth and possibly add value themselves. What complicates the process of morphing software even more is the fact that it’s more or less the norm to develop software as a team. Software is very seldom developed by a single individual, and working in groups makes distributing the work possible. But teamwork also adds another dimension of complexity, and it doesn’t take long to realize that the need to communicate necessary improvements within a team is essential for any real progress. The Mikado Method The Mikado Method is our guide to morphing. It helps us visualize, plan, and perform business-value-focused improvements over several iterations and increments of work, without ever having a broken codebase. It enhances communication, collaboration, and learning in software development teams. It also helps individuals and programming pairs stay on track while doing their day-to-day work. The framework that the method provides can help the whole team morph any system into the desired new shape. The method itself is straightforward and simple: after you’ve decided on a goal, you start to build a model of required changes and dependencies by doing quick experiments. We try to accomplish the goal in a straightforward fashion. When we can’t, we create a graph of prerequisite nodes. We use the information from our quick experiments to determine whether a change can or can’t be implemented right away. If a problem occurs, we roll back to the original state. By solving the prerequisites first, we work our way backwards, and the graph we build guides us toward our goal. The name of the method The Mikado Method gets its name from the Mikado game. Mikado is a pick-up sticks game originating in Europe. In 1936 it was brought from Hungary to the USA and was mostly called pick-up sticks. This term is not very specific in respect to existing stick game variations. Probably the “Mikado” name was not used because it was a brand name of a game producer. The game got its name from the highest scoring (blue) stick “Mikado” (Emperor of Japan). The Buddhistic Chien Tung also contains a stick called “emperor.” (Wikipedia, February 2011, http://en.wikipedia.org/wiki/Mikado_(game)) To start the game, you drop the bunch of straw-sized sticks onto a surface, creating a pile. The goal is to pick up as many sticks as possible to score points, and preferably the Mikado stick, because it’s the highest-scoring stick. If you’re unable to pick up a stick without moving another one, the stick is put back and the turn goes to the next player. As long as you can pick up sticks without moving other sticks, your turn continues. The trick is to pick up the easy sticks first—the sticks that have no other sticks on top of them. Eventually, by using that strategy, a player can pick up the Mikado stick and likely win the game. The similarities between the Mikado game and restructuring software are striking. There are loads of dependencies and structural complexities in code that have to be taken into account. These dependencies have to be navigated and changed with care until you have an opportunity to do something without breaking stuff. Characteristics These are some characteristics of the Mikado Method: It fits nicely in an incremental process. It’s very lightweight (pen and paper, or whiteboard). It increases the visibility of the work. It provides stability to the codebase while you’re changing it. It supports continuous deployments by finding a nondestructive change path. It improves communication between people. It enhances learning. It aids reflection on the work done. It leverages different competencies, abilities, and knowledge. It helps collaboration within a team. It scales by enabling distribution of the workload over the team. It’s easy to use. A thinking tool When we introduce the Mikado Method to people who develop software, we often see how they change the way they look at, approach, and talk about large structural improvements. Their perspective shifts from a very analytical view of problems to a more practical approach, where the focus is on removing the minimum number of obstacles at a time in order to achieve real results. The effect is shorter, focused conversations that are about finding the changes that can be made without breaking the code, instead of rigorous analyzing and guessing. They get more insight on how to deal with the unnecessary complexity of a software system, and the method serves as a thinking tool when they solve difficult problems. Even though we’re becoming increasingly seasoned developers ourselves, we still find it challenging to decide where to start digging and where to take the code. The Mikado Method not only helps us find a starting point, it also shows us where to go and tells us when we’re done. We hope you’ll find the Mikado Method useful, just as we have.

Vorwort Tom Poppendieck
Zusatzinfo figures
Verlagsort New York
Sprache englisch
Maße 189 x 231 mm
Gewicht 406 g
Einbandart kartoniert
Themenwelt Informatik Software Entwicklung Software Projektmanagement
ISBN-10 1-61729-121-8 / 1617291218
ISBN-13 978-1-61729-121-0 / 9781617291210
Zustand Neuware
Haben Sie eine Frage zum Produkt?
Mehr entdecken
aus dem Bereich
Unternehmensweite IT-Planung und zentrale IT-Steuerung in der Praxis

von Ernst Tiemeyer

Buch (2022)
Hanser, Carl (Verlag)
CHF 83,95