The principle of port binding is a fundamental aspect of how services communicate within and outside of an application. It allows a service to become accessible by attaching itself to a specific port number on a host machine. This mechanism ensures that different services can coexist on the same host machine without interfering with each other. For instance, it’s quite common for a service with an HTTP endpoint to be exposed on port 8080. As you delve deeper into the intricacies of Google Cloud in subsequent sections, you will discover how much of this port binding management is seamlessly handled by the platform itself.
VIII. Concurrency: Scale Out via the Process Model
Traditional applications often achieve scaling through a vertical approach, which involves bolstering a single process with additional resources like CPU and memory. Given the availability of multiple CPUs, this necessitates concurrent threads to effectively utilize the added resources, consequently increasing the complexity. It’s noteworthy that Heroku, originally designed for Ruby applications, adhered to this principle likely because concurrency wasn’t a forte of Ruby at that time.
This principle of concurrency advocates the more efficient alternative of horizontal scaling, which involves adding more, smaller instances of the service rather than augmenting the size of existing instances. This approach optimizes the usage of the underlying infrastructure without escalating the complexity. An additional advantage of this strategy is the enhancement of application reliability through multiple instances, thereby reinforcing its place as a key principle in the cloud native paradigm.
IX. Disposability: Maximize Robustness with Fast Startup and Graceful Shutdown
Cloud native applications are designed with the expectation of potential failure of the underlying infrastructure. Consequently, at any given moment, an instance may need to be terminated and restarted. This principle highlights the importance of treating instances as disposable, emphasizing their ability to shut down gracefully—avoiding data loss, for example. It’s equally crucial for new instances to be capable of starting up swiftly, thus minimizing periods of unavailability. This focus on disposability not only maximizes the robustness of the system but also prepares it to handle unexpected disruptions or demand fluctuations efficiently.
X. Dev/Prod Parity: Keep Development, Staging, and Production as Similar as Possible
The dev/prod parity principle is another step to solving the “but it runs on my machine problem.” The idea behind it is that the development, testing, and production environment should be as similar as possible to avoid unexpected behavior when an application moves to a new environment. This includes having the same operating system, dependencies, services, and configuration wherever possible. As I will discuss later in Chapter 12, containers and infrastructure as code (IaC) do a lot to help with this.
XI. Logs: Treat Logs as Event Streams
In a conventional application operating on a single machine, logs can be written to a local file for subsequent examination. However, in a cloud native system with numerous instances of components distributed across multiple machines, managing an assortment of log files on various machines is hardly practical. Rather than dealing with disparate log files, logs should be treated as event streams that are directed toward a centralized logging solution. This method enables efficient processing, correlation, and searchability of logs, thereby improving the overall visibility of system activities and aiding in debugging. In Chapter 13, I’ll delve into how services like Google Cloud offer outstanding features to facilitate these log management practices.
XII. Admin Processes: Run Admin/Management Tasks as One-Off Processes
Admin and management tasks, such as database migrations, data imports, or system updates, are often required in systems and are typically run on a schedule or on demand. This principle advises isolating these types of tasks from the main application execution. By doing so, these tasks can be run independently, reducing potential side effects and simplifying their maintenance.
When making architectural decisions, it’s beneficial to refer back to the 12 factors. These principles should be viewed as guidelines rather than stringent rules. However, if you choose to deviate from them, it’s important to understand why.