Skip to content

Latest commit

 

History

History
172 lines (116 loc) · 7.55 KB

README.md

File metadata and controls

172 lines (116 loc) · 7.55 KB

OpenShift Deployment and Pipeline

HETS (Hired Equipment Tracking System) makes use of BCDK and pipline-cli to manage OpenShift deployments. These tools enable a Github Pull Request (PR) based CI/CD pipeline. Once the Jenkins instance is configured and deployed, it will monitor Github repository for pull requests and start a new build based on that.

This document will not go into the details of how to use BCDK and pipeline-cli. For documentation on those, you can refer to the following links

HETS is using the official BCDevOps Backup Container for backing up DB. Below is README file from BCDevOps backup-container

Prerequisites

  • Admin access to OpenShift namespaces, preferably using the standard BC Gov setup of tools, dev, test and prod namespaces
  • Github personal access token for an account that has contributor access to your repository. It needs to include repo:status, repo_deployment, public_repo and admin:repo_hook permissions

Github Personal Access Token

The Github Personal Access Token is used to give Jenkins access to monitor your repository for changes and create the necessary webhooks. It is recommended to use a shared team Github account for this.

The access token must have repo:status, repo_deployment, public_repo and admin:repo_hook permissions for the pipeline to work properly.

Please refer to Github's guide for more details.

Pipeline Setup

This section will describe the necessary steps required to configure the pipeline to run in your OpenShift environment.

Create Secret Objects

There are a few secret objects that must be created manually in the dev, test, and prod namespaces. These are required for the pods to function properly.

You can use the following templates to create the secret objects. Make sure to replace the parameter variables with actual values encoded to base64.

Configure Pipeline-cli

The .jenkins and .pipeline directories contain the scripts genereated by the BCDK tool in order to use pipeline-cli against your OpenShift namespaces. The scripts in the .jenkins and .pipeline directories in this repository are configured to run against the official SBI OpenShift namespaces (e0cee6-).

You will need to configure the scripts in the .jenkins and .pipeline directories to work with your OpenShift namespaces. There are two ways to do this:

  1. Update the namespace values in both lib/config.js file and create the Jenkins secret objects maually.
  2. Re-run the BCDK tool to setup your namespaces and pipeline-cli automatically. However you will need to merge the generated scripts with the scripts in this repository.

Update Manually

This is easier method to adapt the pipeline scripts to your namespaces. The scripts need to be updated are

The values to modify are the namespace and host values. Update them to match your OpenShift namespaces.

The last step is to create the Jenkins secret objects in the tools namespace. Use your Github username and personal access token.

apiVersion: template.openshift.io/v1
kind: Template
objects:
  - apiVersion: v1
    kind: Secret
    metadata:
      annotations: null
      name: template.${NAME}-slave-user
    stringData:
      metadata.name: template.${NAME}-slave-user
      username: jenkins-slave
  - apiVersion: v1
    kind: Secret
    metadata:
      annotations: null
      name: template.${NAME}-github
    stringData:
      metadata.name: template.${NAME}-github
      username: ${GH_USERNAME}
      password: ${GH_ACCESS_TOKEN}
parameters:
  - description: A name used for all objects
    displayName: Name
    name: NAME
    required: true
    value: jenkins
  - name: GH_USERNAME
    required: true
  - description: GitHub Personal Access Token
    name: GH_ACCESS_TOKEN
    required: true

Update using BCDK

Refer to the official BCDK documentation on how to use this method. After you have ran the BCDK generators, you will need to merge the generated scripts and the existing scripts in this repository.

Deploy Jenkins

The Jenkins instance is used to run the pipeline automatically when pull requests are detected. For that to happen you will need to build and deploy Jenkins to the tools namespace

IMPORTANT! Make sure the jenkins-prod service account from tools namespace have admin access to the dev, test, and prod namespaces. This allows Jenkins to deploy to the namespaces and perform cleanups.

Use the following steps to deploy Jenkins. This assumes you have already logged in

# Switch to tools namespace
oc project <namespace>-tools

# Change working directory to .jenkins/.pipeline
cd .jenkins/.pipeline

# Install required NPM modules
npm install

# Build Jenkins from local code package
# --dev-mode=true enables uploading local source code as an archive
# dev-mode is useful for building changes without pushing to remote first
# --pr=0 indicates pull request 0, generally used for dev-mode builds
npm run build -- --pr=0 --dev-mode=true

# Wait for builds to finish and deploy
npm run deploy -- --pr=0 --env=prod

# Optionally you can deploy as dev first
# npm run deploy -- --pr=0 --env=dev

Once the prod Jenkins instance is running, you should see new webhooks created in your repository settings. Jenkins is now ready to monitor pull requests.

Pull Request Build and Deploy

Once Jenkins is properly configured and deployed, it is ready to monitor pull requests.

Every pull request made to your repository will trigger a new build and create a standalone deployment in the dev namespace. This allows you to test new features independantly of other features.

If configured properly, Jenkins will also automatically clean up the environments when a pull request is merged or closed.

Note! Pull requests to a branch other than master will cause the pipeline to end after deploying the dev namespace. Pull requests to master will cause Jenkins to execute all the stages in the Jenkinsfile unless manually terminated.

Manually Build and Deploy

Use the following steps to manually build and deploy. This assumes you have already logged in.

# Change directory to .pipeline
cd .pipeline

# Installed the required NPM packages
npm install

# Create build for a particular PR on Github
npm run build -- --pr=<pr#>

# Deploy to dev, test, or prod
npm run deploy -- --pr=<pr#> --env=<env>

You can also build from your local source code using the --dev-mode=true flag. This will upload your local source repository instead of cloning the repository from Github. You may need to delete the client/node_modules directory before using dev-mode. The node_modules directory contains too many files and will often cause the build to fail.

dev-mode is useful for testing the pipeline without pushing any changes to the remote repository.

npm run build --pr=0 --dev-mode=true

You can clean up builds and deployments using

npm run clean -- --pr=<pr#> ---env=<env>

# Alternative you can use --env=all if you have the transient option configured properly
npm run clean -- --pr=<pr#> --env=all