<img src="https://ws.zoominfo.com/pixel/JHVDdRXH2uangmUMQBZd" width="1" height="1" style="display: none;">

Magalix Introduces the Cloud-Native Application Policy Pack

Exit icon Learn More
The Sidecar Pattern
DevOps Containers Tutorials Kubernetes Design Patterns Cloud Native Docker services Microservices kubernetesio Patterns Deployments Log shipping side car log aggregation
The Sidecar Pattern

Do One Thing, Do it Best

The basic idea on which UNIX was designed is to not have a complex monolithic tool do everything. Instead, UNIX makes use of small, pluggable components whereby their usage separately is not of great use. But when combined, they can perform powerful operations.

Let’s take the ps command as an example; ps on its own displays the currently running processes on your UNIX/Linux box. It has a decent number of flags that allows you to display many aspects of the processes. For example:

  • The user that started a process
  • How much CPU each running process is using
  • What the command used to start the process is and a lot more.

The ps command does an excellent job displaying information about the running processes. However, there isn’t any ps flag that filters its output. The lack of this functionality is not a missing feature in the tool; this is intentional.

There is another tool that does an excellent job filtering text fed into it; grep. So, using the pipe (|) character, you can filter the output of ps to show only the SSH processes running on your system like this: ps -ef | grep -i ssh. The ps tool is concerned with displaying each and every possible aspect of running processes. The grep command is concerned with offering the ability of filtering text, any text in many different ways. Do one thing and do it best is how UNIX was built.

Because of both UNIX power and simplicity, this principle was used in many other domains in addition to operating systems. In Kubernetes, for example, each container should do only one job and do it well. You might want to ask that what if the container’s job requires extra procedures to aid it or enhance it?, there is nothing to worry about because the same way we piped the output of the ps command to grep, we can use another container sitting beside the main one in the same Pod. That second container carries out the auxiliary logic needed by the first container to function correctly. That second container is commonly known as the Sidecar.

What Does a Sidecar Container Do?

A Pod is the basic atomic unit of deployment in Kubernetes. Typically, a Pod contains a single container. However, multiple containers can be placed in the same Pod. All containers running on the same Pod share the same volume and network interface of the Pod. Actually, the Pod itself is a container that executes the pause command. Its sole purpose is to hold the network interfaces and the Linux namespaces needed to run other containers.

The Sidecar Pattern

A Sidecar container is a second container added to the Pod definition. Why it must be placed in the same Pod is that it needs to use the same resources being used by the main container. Let’s have an example to demonstrate the use cases of this pattern.

Working in production with Kubernetes? Trying to deploy more frequently but can't keep your team focused?

Learn how Magalix can help you move fast while adopting cloud-native best practices.

Learn More


Scenario: Log-Shipping Sidecar

The Sidecar Pattern

In this scenario, we have a web server container running the nginx image. The access and error logs produced by the web server are not critical enough to be placed on a persistent volume. However, developers need access to the last 24 hours of logs so they can trace issues and bugs. Therefore, we need to ship the access and error logs for the web server to a log-aggregation service. Following the separation of concerns principle, we implement the Sidecar pattern by deploying a second container that ships the error and access logs from nginx. Nginx does one thing, and it does it well; serving web pages. The second container also specializes in its task; shipping logs. Since containers are running on the same Pod, we can use a shared emptyDir volume to read and write logs. The definition file for such a Pod may look as follows:

apiVersion: v1
kind: Pod
  name: webserver
    - name: shared-logs
      emptyDir: {}

    - name: nginx
      image: nginx
        - name: shared-logs
          mountPath: /var/log/nginx

    - name: sidecar-container
      image: busybox
      command: ["sh","-c","while true; do cat /var/log/nginx/access.log /var/log/nginx/error.log; sleep 30; done"]
        - name: shared-logs
          mountPath: /var/log/nginx

The above definition is a standard Kubernetes Pod definition except that it deploys two containers to the same Pod. The sidecar container conventionally comes second in the definition so that when you issue the kubectl execute command, you target the main container by default.

The main container is an nginx container that’s instructed to store its logs on a volume mounted on /var/log/nginx. Mounting a volume at that location prevents Nginx from outputting its log data to the standard output and forces it to write them to access.log and error.log files.

Side Note on log Aggregation

Notice that the default behavior of the Nginx image is to store its logs to the standard output to be picked by Dockers’ log collector. Docker stores those logs under /var/lib/docker/containers/container-ID/container-ID-json.log on the host machine. With more than one container (from different Pods) running on the same host and using the same location for storing their logs, you can use a DaemonSet to deploy a log-collector container like Filebeat or Logstash to collect those logs and send them to a log-aggregator like ElasticSearch. You’ll need to mount /var/lib/docker/containers as a hostPath volume to the DaemonSet Pod to give the log-collector container access to the logs.

The sidecar container runs with the nginx container on the same Pod. This enables the sidecar container to access the same volume as the web server. In the above example, we used the cat command to simulate sending the log data to a log aggregator every thirty seconds.

Magalix Trial


  • The Sidecar Pattern entails designing modular containers. A modular container can be plugged in more than one place with minimal changes.
  • A sidecar is just a container that runs on the same Pod as the application container, because it shares the same volume and network as the main container, it can “help” or enhance how the application operates.
  • Common examples of sidecar containers are log shippers, log watchers, monitoring agents among others.
  • Having a separate container for auxiliary tasks gives you access to health checks, automatic restart and other functionality offered by Kubernetes for containers.
  • When following the Sidecar Pattern, you should consider making a small sidecar container that does not consume many resources. If you find that the sidecar container logic is getting more complex and/or becoming more tightly coupled with the main application container, it may better be integrated with the main application’s code instead. Remember, the strong point of a sidecar container lies in its ability to be small and pluggable.

Comments and Responses

Related Articles

7 Notable and Costly Security Breaches

Learn some notable security breaches that happened a few years ago, the root causes, and how Magalix can help protect your Kubernetes infrastructure

Read more
Security Context Settings Help Mitigate Kubernetes Risk

Kubernetes isn't secure by default and is attacked relentlessly. But security context settings help DevOps teams better secure their pods and containers.

Read more
DevOps Policy as Code
Cloud Data Asset Management and Tagging Cloud Resources

Learn how cloud data asset management enables organizations to manage, optimize and secure their cloud data assets and resource tagging is a key part of it

Read more