Skip to content

Latest commit

 

History

History
151 lines (125 loc) · 5.82 KB

README.md

File metadata and controls

151 lines (125 loc) · 5.82 KB

Ceph-CI

CEPH-CI is a framework tightly coupled with CentralCI and Redhat Builds for testing Ceph downstream builds with CentralCI and Jenkins.

It uses a modified version of Mita to create/destroy Ceph resources dynamically

Getting Started

Prerequisites

  1. Python 3.7

Installing

It is recommended that you use a python virtual environment to install the necessary dependencies and execute cephci.

  1. Setup a python 3.7 virtual environment. This is actually quite easy to do now.
    • python3.7 -m venv <path/to/venv>
    • source <path/to/venv>/bin/activate
  2. Install requirements with pip install -r requirements.txt

Initial Setup

Configure your cephci.yaml file:

This file is used to allow configuration around a number of things within cephci. The template can be found at the top level of the repository, cephci.yaml.template. The required keys are in the template. Values are placeholders and should be replaced by legitimate values. Values for report portal or polarion are only required if you plan on posting to that particular service.

Move a copy of the template to your user directory and edit it from there with the proper values.

cp cephci.yaml.template ~/.cephci.yaml

Tests

CentralCI Authentication

CentralCI auth files are kept in the osp directory.

The osp-cred.yaml file has OpenStack credentials details to create/destroy resources. For local cephci runs, you will want to replace the username/password with your own OpenStack credentials.

Cluster Configuration

Cluster configuration files are kept in a directory under conf for each ceph version. For jewel, configs are under conf/jewel. For luminous, configs are under conf/luminous.

The conf files describes the test bed configuration. The image-name inside globals: define what image is used to clone ceph-nodes( mon, osd, mds etc), The role maps to ceph role that the node will take and osd generally attach 3 additional volumes with disk-size as specified in config.

Inventory Files

Inventory files are kept under conf/inventory, and are used to specify which operating system to be used for the cluster resources.

Test Suites

All test suite configurations are found inside the suites directory.

There are various suites that are mapped to versions of Ceph under test

suites/jewel/sanity_ceph_ansible will be valid for 2.0 builds
suites/luminous/sanity_ceph_ansible will be valid for 3.0 builds

The tests inside the suites are described in yaml format

tests:
   - test:
      name: ceph deploy
      module: test_ceph_deploy.py
      config:
        base_url: 'http://download-node-02.eng.bos.redhat.com/rcm-guest/ceph-drops/auto/ceph-1.3-rhel-7-compose/RHCEPH-1.3-RHEL-7-20161010.t.0/'
        installer_url: 
      desc: test cluster setup using ceph-deploy
      destroy-cluster: False
      abort-on-fail: True
      
   - test:
      name: rados workunit
      module: test_workunit.py
      config:
            test_name: rados/test_python.sh
            branch: hammer
      desc: Test rados python api

The above snippet describes two tests and the module is the name of the python script that is executed to verify the test, every module can take a config dict that is passed to it from the run wrapper, The run wrapper executes the tests serially found in the suites. The test scripts are location in the tests folder.

Usage

run.py is the main script for ceph-ci. You can view the full usage details by passing in the --help argument.

python run.py --help

Required Arguments

There are a few arguments that are required for cephci execution:

  • --rhbuild <build_version>
  • --osp-cred <cred_file>
  • --global-conf <conf_file>
  • --inventory <inventory_file>
  • --suite <suite_file>

Useful Arguments

Some non-required arguments that we end up using a lot:

  • --log level <level> - set the log level that is output to stdout.

Examples

Ceph ansible install suite:

python run.py --rhbuild 3.2 --global-conf conf/luminous/upgrade.yaml --osp-cred osp/osp-cred-ci-2.yaml
--inventory conf/inventory/rhel-7.6-server-x86_64-released.yaml --suite suites/luminous/sanity_ceph_ansible.yaml
--log-level info

Upgrade suite:

python run.py --rhbuild 3.2 --global-conf conf/luminous/upgrade.yaml --osp-cred osp/osp-cred-ci-2.yaml
--inventory conf/inventory/rhel-7.6-server-x86_64-released.yaml --suite suites/luminous/upgrades.yaml
--log-level info

Containerized upgrade suite:

python run.py --rhbuild 3.2 --global-conf conf/luminous/upgrade.yaml --osp-cred osp/osp-cred-ci-2.yaml
--inventory conf/inventory/rhel-7.6-server-x86_64-released.yaml --suite suites/luminous/upgrades_containerized.yaml
--log-level info --ignore-latest-container --insecure-registry --skip-version-compare

Manual cluster cleanup

Ceph-CI also has the ability to manually clean up cluster nodes if anything was left behind during a test run. All you need to provide is your osp credentials and the instances name for the cluster.

python run.py --osp-cred <cred_file> --cleanup <instances_name>

Results

In order to post results properly or receive results emails you must first configure your ~/.cephci.yaml file. Please see the Initial Setup section of the readme if you haven't done that.

Polarion

Results are posted to polarion if the --post-results argument is passed to run.py. When this argument is used, any tests that have a polarion-id configured in the suite will have it's result posted to polarion.

Report Portal

Results are posted to report portal if the --report-portal argument is passed to run.py.

Email

A result email is automatically sent to the address configured in your ~/.cephci.yaml file. In addition to personally configured emails, if the --post-results or --report-portal arguments are passed to run.py an email will also be sent to [email protected].