- testing MPL pipeline with the current LTS Jenkins every night
- testing MPL pipeline with the current Latest Jenkins every night
Shared jenkins library with modular structure allow to write a simple pipeline modules, test it properly and use in any kind of pipelines.
- Provide core to support flexible modular structure
- Enable support for declarative & scripted pipelines
- Ensure that nested libraries could reuse the MPL features
- Allow module overrides from the project or nested libraries
- Prepare set of simple & minimal basic pipelines
- Make it easy to create tests for modules & pipelines
This readme contains mostly technical information, if you need some overview - please check the next resources:
- Grid Dynamics Blog
- Overview & demo videos:
You also can check MPL Wiki to find additional info.
- Jenkins >= 2.74 LTS
- workflow-cps >= 2.44
- workflow-job >= 2.15
- workflow-cps-global-lib >= 2.8
Go to: Manage Jenkins --> Configure System --> Global Pipeline Libraries:
- Name:
mpl
- Default Version:
master
- Load implicitly:
false
- Allow default version to be overridden:
true
- Include @Library changes in job recent changes:
true
- Retrieval method:
Modern SCM
- Source Code Management:
Git
- Project Repository:
https://github.com/griddynamics/mpl.git
- Behaviors:
Discover branches
,Discover tags
- Default Version:
You can use MPL in 3 different ways - it's 3 layers of the library:
- Pipeline - Standard pipelines implemented for you. Just use them for your project or as example for your custom pipeline.
- Modules - Common steps you can find in each pipeline. Prepare your pipeline structure and just use the required modules.
- Features - Some useful libraries that you can use. Prepare your own pipeline with custom modules and use just features from the library.
Just two lines to use default Master pipeline in your project Jenkinsfile or in the Jenkins Pipeline Script:
@Library('mpl') _
MPLPipeline {}
You can find pipelines in the MPL library interfaces: {MPL}/vars/MPL{Name}Pipeline.groovy
files.
There is 2 ways to slightly reconfigure the provided standard pipeline:
- Provide some configuration when you using pipeline in your jenkinsfile (see examples)
- Override pipeline modules in the project (see
Creating / Overriding steps modules
section)
If both ways are not helping - just create your own pipeline definition in the Jenkinsfile and use MPL modules or create your own nested library with your standard pipelines.
Usually configuration is initialized in the pipeline - it's calling MPLPipelineConfig
interface with arguments:
body
- Map/Closure with configuration from the Jenkinsfiledefaults
- pipeline default values (could be overridden by Jenkinsfile)overrides
- pipeline hard values (could not be overridden by Jenkinsfile)
After that pipeline defining MPL object and use it's common functions to define the pipeline itself. Pipeline is calling MPLModule
that calling the required module logic.
In the module we have a common predefined variables (like default steps
, env
, params
, currentBuild
...) and a new variable contains the pipeline/module configs: CFG
.
It's a special MPLConfig
object that defines interface to get and set the properties. It's promising a number of things:
- Get value will never throw exception
- Unable to change values of
CFG
through get or clone - It's not related to the parent module or pipeline
CFG
- any changes could be only forwarded to the children module - You can get raw config Map/List - but they will act like a normal Map/List (could cause exceptions)
- Set of the
CFG
value could cause exceptions in certain circumstances:- set improper List index (non-positive integer)
- set sublevels for defined non-collections
MPLConfig
object is not the usual Map
- you have to use it with known key you defined before.
That was introduced to make sure you will see what kind of configuration is used or required on each
stage. So you will need to use it as def a = CFG.really_needed_data
instead of just def a = CFG
or for( a in CFG ) println(a)
which will tell you much more useful information and simplify debug.
Use of the CFG
object is quite simple. Imagine we have the next pipeline configuration:
[
agent_label: '',
val1: 4,
val2: [
val_nested: 'value',
val_list: [1,2,3,4],
],
]
-
Get value of specific property:
CFG.val1
==CFG.'val1'
==4
CFG.'val2.val_nested'
=='value'
CFG.'val2.not_exist'
==null
CFG.'val2.not_exist' ?: 'default'
=='default'
CFG.'val2.val_list.2'
==3
-
Get raw Map or List:
CFG.'val2.val_list'
==[1,2,3,4]
CFG.val2
==[val_nested:'value',val_list:[1,2,3,4]]
-
Set value of a specific property:
CFG.val1 = 3
;CFG.val1
==3
CFG.'val2.val_nested' = 555
;CFG.val2.val_nested
==555
-
Create new properties:
CFG.val4 = 'newval'
;CFG.val4
=='newval'
CFG.'val2.val_list.5' = 333
;CFG.'val2.val_list'
==[1,2,3,4,null,333]
-
Replace entire Map or List:
CFG.'val2.val_list' = null
;CFG.val2
==[val_nested:'value',val_list:null]
CFG.val2 = [new_key:[4,3,2,1]]
;CFG.val2
==[new_key:[4,3,2,1]]
So you got the point - hopefully this will be helpful and will allow you to create the great interfaces for your modules.
MPLModule
step running the specified module with CFG
configuration and returns OUT
configuration.
OUT
is always empty when a module just started and could be modified inside the module. So:
- you can set some variable like "Module/SomeModule.groovy":
OUT.'artifact.version' = 1
- and use it in parent module "Module/Module.groovy" as:
def version = MPLModule('Some Module').'artifact.version' echo "${version}" OUT.'artifact_info.artifact.version' = version
To modify the pipeline config with the module output - just use MPLPipelineConfigMerge
step - we
recommend to use it only in the pipeline step specification to concentrate any pipeline-related
changes in the pipeline definition itself.
You can use MPLPipelineConfigMerge
in the pipeline like this - the logic will put artifact
key
with value [version: 1]
in the global configuration and you will be able to use
`CFG.'artifact.version' in the following modules:
pipeline {
...
steps {
MPLPipelineConfigMerge(MPLModule('Some Module').artifact_info)
}
...
MPL is mostly modules with logic. Basic features:
- Simple groovy sandbox step files with all the pipeline features
- Could be used in declarative & scripted pipelines
- Override system with loop protection and simple nesting
In fact modules could be loaded from a number of places:
{ProjectRepo}/.jenkins/modules/{Stage}/{Name}{Stage}.groovy
- custom modules for the project{Library}/{SpecifiedModulesLoadPath}/modules/{Stage}/{Name}{Stage}.groovy
- custom modules load path that added while init{Library}/resources/com/griddynamics/devops/mpl/modules/{Stage}/{Name}{Stage}.groovy
- library modules for everyone
If you will override module Build in your project repo, it will be used first. If you will try to require Build module from this overridden Build module - original library module will be used.
Check the usage examples & library modules to get some clue about the nesting system.
If your project is special - you can override or provide aditional modules just for the project.
What do you need to do:
- Create a step file:
{ProjectRepo}/.jenkins/modules/{Stage}/{Name}{Stage}.groovy
(name could be empty) - Fill the step with your required logic: you can use
CFG
config object & MPL functions inside the steps definition - Use this step in your custom pipeline (or, if it's override, in standard pipeline) via MPLModule method.
- Provided custom modules will be available to use after the checkout of your project source only
For example: "Maven Build" steps have path modules/Build/MavenBuild.groovy
and placed in the library - feel free to check it out.
MPL supports 2 useful poststep interfaces which allow you to store all the logic of module in the same file.
All the poststeps will be executed in LIFO order and all the exceptions will be collected & displayed in the logs.
Allow to set some actions that need to be executed right after the current module (doesn't matter it fails or not).
Could be useful when you need to collect reports or clean stage agent before it will be killed.
If module post step fails - it's fatal for the module, so the pipeline will fail (unlike general poststeps). All the poststeps for the module will be executed and errors will be printed, but module will fail.
{NestedLibModules}/Build/MavenBuild.groovy
:MPLModulePostStep { junit 'target/junitReport/*.xml' } // Could fail but our poststep will be executed MPLModule('Maven Build', CFG)
General poststep interface usually used in the pipelines. Requires 2 calls - first one to define poststep in a module and second one to execute it and usually placed in the pipeline post actions.
When error occurs during poststeps execution - it will be printed in the log, but status of pipeline will not be affected.
{NestedLibModules}/Deploy/OpenshiftDeploy.groovy
:
MPLPostStep('always') {
echo "OpenShift Deploy Decomission poststep"
}
echo 'Executing Openshift Deploy process'
{NestedLib}/var/CustomPipeline.groovy
:
def call(body) {
// ...
pipeline {
// ...
stages {
// ...
stage( 'Openshift Deploy' ) {
steps {
MPLModule()
}
}
// ...
}
post {
always {
MPLPostStepsRun('always')
}
success {
MPLPostStepsRun('success')
}
failure {
MPLPostStepsRun('failure')
}
}
}
}
To make sure that some of your stages will be executed for sure - you can add a list of modules that could be overrided on the project side.
Just make sure, that you executing function MPLEnforce
and provide list of modules that could be overriden in your pipeline script:
- Jenkins job script:
@Library('mpl') _ // Only 'Build Maven' & 'Deploy' modules could be overriden on the project side MPLEnforce(['Build Maven', 'Deploy']) // ... Your enforced pipeline
Notices:
- The function
MPLEnforce
could be executed only once, after that it will ignore any further executions. - This trick is really working only if you controlling the job pipeline scripts, with Jenkinsfile it's not so secure.
MPL supporting the nested libraries to simplify work for a big teams which would like to use MPL but with some modifications.
Basically you just need to provide your vars
interfaces and specify the mpl library to use it:
{NestedLib}/vars/MPLInit.groovy
:def call() { // Using the latest release MPL and adding the custom path to find modules library('mpl') MPLModulesPath('com/yourcompany/mpl') }
{NestedLib}/vars/NestedPipeline.groovy
:def call(body) { MPLInit() // Init the MPL library // ... Specifying the configs / pipelines and using modules etc. }
And after that and configuring the library for your jenkins (just put it after the mpl
config) - you can use it in the project's Jenkinsfile (see examples).
Also you can override resources of the MPL library - but it's forever. You can't use MPL resource anymore if you will override it in the nested library.
You can cover the nested library with tests as well as MPL library - please check the nested library example on wiki page.
Jenkins shared libraries is just a repositories connected to the Jenkins Master instance. So you can use any branch/tag from the MPL or nested lib repo.
- branch:18.04 - released version of the library. Can be used to pin the version.
- branch:master - most fresh functions (and bugs) are here. You can use it for testing purposes.
- branch:TICKET-1234 - feature branch, could be used for testing purposes.
Please check the wiki page to see some MPL examples: MPL Wiki
If we fine with standard pipeline, but need to slightly modify options.
{ProjectRepo}/Jenkinsfile
:@Library('mpl') _ // Use default master pipeline MPLPipeline { // Pipeline configuration override here // Example: (check available options in the pipeline definition) agent_label = 'LS' // Set agent label modules.Build.tool_version = 'Maven 2' // Change tool for build stage modules.Test = null // Disable Test stage }
We fine with standard pipeline, but would like to use different deploy stage.
{ProjectRepo}/Jenkinsfile
:@Library('mpl') _ // Use default master pipeline MPLPipeline {}
{ProjectRepo}/.jenkins/modules/Deploy/Deploy.groovy
:// Any step could be here, config modification, etc. echo "Let's begin the deployment process!" // Run original deployment from the library MPLModule('Deploy', CFG) echo "Deployment process completed!"
{ProjectRepo}/Jenkinsfile
:@Library('mpl') _ pipeline { // Declarative pipeline agent { label 'LS' } stages { stage( 'Build' ) { parallel { // Parallel build for 2 subprojects stage( 'Build Project A' ) { steps { dir( 'subProjectA' ) { MPLModule('Maven Build', [ // Using common Maven Build with specified configs tool_version: 'Maven 2' ]) } } } stage( 'Build Project B' ) { steps { dir( 'subProjectB' ) { // Custom build process (it's better to put it into the project custom module) sh 'gradle build' sh 'cp -a target my_data' } } } } } } }
{ProjectRepo}/Jenkinsfile
:@Library('nested-mpl') _ NestedPipeline { // config here }
We should be ensure that modules will not be broken accidentally, so tests is mandatory.
MPL supports MPLModule testing via slightly modified JenkinsPipelineUnit.
You can find module tests (as well as modified base test classes & overridden requirements classes) in the test
directory.
To run tests just execute mvn clean test
- and it will compile the classes & execute tests for the modules.
MPL provided pipelines should be really simple, but could be improved with new best practices - so changes are always welcome.
If you have some interesting module - you for sure can prepare changes for existing module or a new module, write tests, create the pull-request, describe it - and after that this module could be approved to be included n the base library. If not - you always could use (or create) your own nested library to share this fancy module across your projects.