OpenShift Unleashed

Explore the future of containerized applications with Red Hat OpenShift Container Platform: A guide to understanding the basics, nuances of the platform.

May 21, 2024 | 8 min read
avatar
Akshay PK
Lead Principal Systems Engineer - CS Expert Services
avatar
Sriram Seshadri
Principal System Engineer - Customer Success
avatar
Sri Balaji
System Engineer - Customer Success

Introduction

OpenShift Container Platform is a cloud-based Kubernetes container platform. The foundation of OpenShift Container Platform is based on Kubernetes and therefore shares the same technology. It is designed to allow applications and the data centers that support them to expand from just a few machines and applications to thousands of machines that serve millions of clients.

OpenShift Container Platform enables you to do the following:

  • Provide developers and IT organizations with cloud application platforms that can be used for deploying applications on secure and scalable resources.

  • Require minimal configuration and management overhead.

  • Bring the Kubernetes platform to customer data centers and cloud.

  • Meet security, privacy, compliance, and governance requirements.

With its foundation in Kubernetes, OpenShift Container Platform incorporates the same technology that serves as the engine for massive telecommunications, streaming video, gaming, banking, and other applications. Its implementation in open Red Hat technologies lets you extend your containerized applications beyond a single cloud to on-premise and multi-cloud environments.

For more understanding of the RedHat Openshift Container Platform please refer to the following link : OpenShift Container Platform overview | Getting started | OpenShift Container Platform 4.15

Kubernetes Overview

Problem Statement

The below diagram explains as to how traditionally, applications were deployed on top of a single operating system. With virtualization, you can split the physical host into several virtual hosts. Working on virtual instances on shared resources is not optimal for efficiency and scalability. Because a virtual machine (VM) consumes as many resources as a physical machine, providing resources to a VM such as CPU, RAM, and storage can be expensive. Also, you might see your application degrading in performance due to virtual instance usage on shared resources.


The Solution

To solve this problem, you can use containerization technologies that segregate applications in a containerized environment. Similar to a VM, a container has its own filesystem, vCPU, memory, process space, dependencies, and more. Containers are decoupled from the underlying infrastructure, and are portable across clouds and OS distributions. Containers are inherently much lighter than a fully featured OS, and are lightweight isolated processes that run on the operating system kernel.

To further more read about the components of Kubernetes : Kubernetes overview | Getting started | OpenShift Container Platform 4.15

Openshift Architecture Overview & Product Architecture

The following pages help explain what the architecture overview is and the product architecture of the RedHat Openshift Container Platform :

Architecture overview | Architecture | OpenShift Container Platform 4.15

Product architecture | Architecture | OpenShift Container Platform 4.15

On a very high level this following diagram present in the pages above describes the components of RedHat Openshift Container Platform.

Deployment Topology

Let’s consider that all the necessary infrastructure components are present for the deployment of microservices and development is in full swing with Secure SDLC practices, deployment on RedHat Openshift Container Platform would look something very similar to deploying microservices on a cloud based platform such as AWS, Azure or GCP with certain nuances specific to RedHat Openshift Container Platform. The below diagram depicts an example for the same :

In any traditional Kubernetes deployment, the microservices would be deployed to a namespace whereas in Openshift, they are called as Openshift projects.

In the above diagram, GitHub and JFrog Artifactory are two tools that are used for source code management and artifact/binary management which are two specific examples of tools used. For more details on the same :

GitHub: Let’s build from here

Artifactory - Universal Artifact Management

From an architecture standpoint, Openshift uses various operators to automate the creation, configuration, and management of instances of Kubernetes-native applications. To learn more about the same : What are Red Hat OpenShift Operators?

In the above diagram one such operator is mentioned which helps facilitate deployments to the Openshift project or even across projects or even across clusters : https://www.redhat.com/en/technologies/cloud-computing/openshift/gitops

About OpenShift GitOps | Understanding OpenShift GitOps | Red Hat OpenShift GitOps 1.12

The above operator comes bundled with all the niceties of ArgoCD which is one of the most widely used tool in current times.

Nuances - RedHat Openshift Container Platform

This section describes the nuances and differences of working with the RedHat Openshift Container Platform and the deployment aspects of microservices.

Tightened Security

By default, the Openshift container platform comes with security context constraints that’s described here : Managing security context constraints | Authentication and authorization | OpenShift Container Platform 4.15

The tightened security of the RedHat Openshift Container Platform under the restricted SCC : Managing SCCs in OpenShift because of the way UID’s are allocated during the creation of an Openshift project and the way a UID is allocated from that range when a pod is created which is described in this page : A Guide to OpenShift and UIDs

Kubernetes Ingress vs Openshift Route

This page even though written 6 years back is still a great article to learn about the differences between a Kubernetes Ingress and an Openshift Route : Kubernetes Ingress vs OpenShift Route

In most projects that run Openshift the recommended object to use for routing is an Openshift route as it has more features than that of a Kubernetes Ingress which is described in the above page.

Multiple ways to create a route through configuration where one such example is described below where a route template manifest is added to the templates/ folder of a helm chart with a condition to enable or disable the creation of the same.


{{- if .Values.route.enabled }}
apiVersion: route.openshift.io/v1
kind: Route
metadata:
  name: {{ .Values.route.name }}
  labels:
    app: {{ .Values.app.name }}
spec:
  host: {{ .Values.route.host }}
  to:
    kind: Service
    name: {{ .Values.service.name }}
  port:
    targetPort: {{ .Values.route.targetPort }}
  tls:
    termination: {{ .Values.route.tls.termination }}
    insecureEdgeTerminationPolicy: {{ .Values.route.tls.insecureEdgeTerminationPolicy }}
  wildcardPolicy: {{ .Values.route.wildcardPolicy }}
{{- end }}

With this template, route can be created to expose a service to the outside of a cluster, simply add this to the helm chart’s values file.

route:
  enabled: true
  name: #<name of the route>
  host: #<domain name of the route>
  targetPort: 8080
  tls:
    termination: edge
    insecureEdgeTerminationPolicy: Redirect
  wildcardPolicy: None

Block Exposed URLs

Some exposed URLs that have to be blocked from public internet access and to only be accessible from the private network. These URLs can be blocked from public internet access directly on the Openshift route or any layer of networking above the Openshift route such as a WAF as a recommended practice. The following snippet shows how to do IP-based restriction on an Openshift route with a path to ensure it’s blocked from public access but can still be accessed from the private network. This configuration is however recommended to do on a WAF layer.

haproxy.router.openshift.io/ip_whitelist: '<IPv4 address>'

Rate Limiting

Rate limiting restricts the number of requests allowed from each originating IP address within a specific time frame.

On the Openshift route layer, this can also be done through a set of annotations for TCP and HTTP traffic which is described here : Route configuration - Configuring Routes | Networking | OpenShift Container Platform 4.15

In simple terms the following configuration should be more than sufficient to enable rate limiting for a particular route and path. The configuration can be adjusted based on client requirements. This can also be tested if it works using JMeter and similar tools by sending concurrent requests. This configuration is however recommended to do on a WAF layer.

haproxy.router.openshift.io/rate-limit-connections: 'true'
haproxy.router.openshift.io/rate-limit-connections.rate-http: '500' # Values are dependent on customer usecase

Cache-Control Header

The Openshift infrastructure layer from 4.14 version of Openshift onwards, the response header could be set on the Openshift route itself as part of it’s configuration : Route configuration - Configuring Routes | Networking | OpenShift Container Platform 4.15

An example snippet of the same under the spec section:

spec:
  httpHeaders:
    actions:
      response:
      - name: Cache-Control
        action:
          type: Set
          set:
            value: no-cache,no-store

This article from OWASP describes what the SameSite attribute for a Cookie is all about and what are it’s possible values SameSite | OWASP Foundation

As part of Openshift Route configuration the SameSite cookie attribute to set it to Strict, Lax or None can be handled as part of the annotations section. An example snippet of the same :

router.openshift.io/cookie-same-site: 'Strict'

Service Monitors

The RedHat Openshift Container Platform comes bundled with a monitoring stack that uses Promethues which is explained in a lot of detail here : Monitoring overview - Monitoring | Observability | OpenShift Container Platform 4.15

The below diagram is taken from the above page to describe all components of the monitoring stack.

All the default platform monitoring components are present in the openshift-monitoring project.

To scrape metrics of a user defined application one can use ServiceMonitor object described here : Managing metrics - Monitoring | Observability | OpenShift Container Platform 4.15

This allows to gather metrics related to the application such as JVM, GC etc based on a label that’s assigned to each deployment of a microservice when deployed to the cluster. In the following ServiceMonitor object example a label of app.example.com/monitoring: "true" is added while deploying the microservice to the cluster based on which applications are selected to be scraped for metrics.

apiVersion: monitoring.coreos.com/v1
kind: ServiceMonitor
metadata:
  labels:
    k8s-app: my-apps
  name: my-apps-sm
  namespace: #specify openshift project
spec:
  endpoints:
  - interval: 10s
    port: http
    scheme: http
    path: /actuator/prometheus
  selector:
    matchLabels:
      app.example.com/monitoring: "true"

If the actuator endpoints for scraping are protected behind authentication, the ServiceMonitor object also supports basic authentication with username/password as per the below example under the spec.endpoints section where the username/password can be referred from a Kubernetes Secret.

endpoints:
  - basicAuth:
      password:
        name: <client>-<env>-monitoring-secret
        key: password
      username:
        name: <client>-<env>-monitoring-secret
        key: username
    interval: 10s
    port: http
    scheme: http
    path: /actuator/prometheus

Conclusion

RedHat OpenShift Container Platform is a powerful solution that simplifies the complexities of containerization and Kubernetes orchestration, offering automated operations, integrated CI/CD pipelines, and multi-cloud support. By mastering its architecture and core components, and understanding nuances such as networking, monitoring and security, enterprises can significantly enhance security, developer productivity, and operational efficiency. OpenShift's enterprise-grade features ensure scalability, reliability, and flexibility, making it an ideal choice for accelerating digital transformation and delivering innovative solutions effectively.

References

Red Hat OpenShift Container Platform
OpenShift GitOps
GitHub: Let’s build from here
Software Supply Chain Platform for DevOps & Security | JFrog
Argo CD - Declarative GitOps CD for Kubernetes

Continue the Journey: Explore Related Posts

Cisco Appdynamics June 25, 2024 | 7 min read

Explore the Future of Application Performance Management with Cisco AppDynamics.

avatar
Akshay PK
avatar
Sriram Seshadri
avatar
Sri Balaji
Read more
A New Chapter: Exploring Azure Resources November 21, 2023 | 4 min read

Azure’s limitless potential sets the stage for a promising journey, where opportunities abound on the horizon.

avatar
Sri Balaji
avatar
Ashish Arora
Read more