When I first started learning to program at school, my teacher said to me, “if you are doing the same thing manually more than three times, you are probably doing it wrong.”” That has stuck with me. Another way of looking at it is that the best developers are lazy developers. They don’t want to build a container by hand, manually deploy a new version multiple times a day, or spend ages trawling through logs to debug errors. The most productive developers consider this work “toil” and seek to eliminate it and get back to building applications.
This is where DevOps comes from, a set of practices that combines software development (Dev) and operations (Ops) to speed up the development and increase software quality. As you will see in later chapters, this includes:
Continuous integration
Automatically building and testing changes in the factory (Chapter 12)
Continuous delivery
Automatically making the output of the factory ready for deployment to the citadel (Chapter 11)
Monitoring and logging
Exposing what is happening inside the citadel in the observatory (Chapter 13) to make correcting problems straightforward
Infrastructure as code
The mechanism for building the infrastructure of the factory and the citadel in a reproducible way
Managed services
Using services managed by the cloud provider in the citadel in preference to self-management
Ultimately, when working with cloud native apps, there is the approach of “you build it, you run it.” As it is the developers that are on the hook if the applications fail at 2 a.m., there is a great incentive to make sure that doesn’t happen. If it does, there is the incentive that it be easy to fix so the developer can get back to bed. DevOps is operations performed by a developer who prefers to be writing code and getting a good night’s sleep.
Google has developed a particular implementation of DevOps called site reliability engineering (SRE) that it uses for running production systems. This is explained in Site Reliability Engineering by Betsy Beyer et al. (O’Reilly). You will see how SRE principles permeate into Google Cloud in Chapter 13.
Summary
Successfully developing cloud native applications relies on a comprehensive understanding of various tools, techniques, and technologies. While each of these subjects merits detailed study, here are the primary points to remember:
- Cloud native applications tend to be made from autonomous components communicating by messages, not by sharing state.
- The 12 factors are useful recommendations, but remember that they were designed for earlier platforms.
- Microservices are one destination but don’t have to be the starting point.
- Event-driven and API-driven architecture are both useful approaches.
- Domain-driven design helps identify the boundaries between microservices.
- Event Storming helps map out communication in event-driven systems.
- Containerizing components provide flexibility and portability.
- Kubernetes is just one way of running containers.
- Patterns exist to help address common problems and mitigate the risk of vendor lock-in.
- Extreme programming techniques help sustain fast feedback, especially test-driven development.
- Testing increases quality and provides the confidence needed to keep going fast.
- Automation saves time, provides consistency, and lets developers focus on solving business problems.
One crucial thing to remember is that everything comes with trade-offs. There aren’t any definitive rules dictating what a cloud native application should be. Instead, there are proven principles that generally provide good results. Cloud native doesn’t impose the use of Kubernetes, microservices, or even containers. Situations may arise where these principles don’t apply, and it’s these shades of gray that underline the importance of knowing the options available and making informed decisions based on your specific circumstances.
In Part II, you’ll begin your exploration of Google Cloud, focusing on how it can be leveraged to build cloud native applications.