Cilium E2E Upgrade (ci-e2e-upgrade) #116
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Cilium E2E Upgrade (ci-e2e-upgrade) | |
# Any change in triggers needs to be reflected in the concurrency group. | |
on: | |
workflow_dispatch: | |
inputs: | |
PR-number: | |
description: "Pull request number." | |
required: true | |
context-ref: | |
description: "Context in which the workflow runs. If PR is from a fork, will be the PR target branch (general case). If PR is NOT from a fork, will be the PR branch itself (this allows committers to test changes to workflows directly from PRs)." | |
required: true | |
SHA: | |
description: "SHA under test (head of the PR branch)." | |
required: true | |
extra-args: | |
description: "[JSON object] Arbitrary arguments passed from the trigger comment via regex capture group. Parse with 'fromJson(inputs.extra-args).argName' in workflow." | |
required: false | |
default: '{}' | |
push: | |
branches: | |
- 'renovate/main-**' | |
# Run every 8 hours | |
schedule: | |
- cron: '0 5/8 * * *' | |
# By specifying the access of one of the scopes, all of those that are not | |
# specified are set to 'none'. | |
permissions: | |
# To read actions state with catchpoint/workflow-telemetry-action | |
actions: read | |
# To be able to access the repository with actions/checkout | |
contents: read | |
# To allow retrieving information from the PR API | |
pull-requests: read | |
# To be able to set commit status | |
statuses: write | |
concurrency: | |
# Structure: | |
# - Workflow name | |
# - Event type | |
# - A unique identifier depending on event type: | |
# - schedule: SHA | |
# - workflow_dispatch: PR number | |
# | |
# This structure ensures a unique concurrency group name is generated for each | |
# type of testing, such that re-runs will cancel the previous run. | |
group: | | |
${{ github.workflow }} | |
${{ github.event_name }} | |
${{ | |
(github.event_name == 'push' && github.sha) || | |
(github.event_name == 'schedule' && github.sha) || | |
(github.event_name == 'workflow_dispatch' && github.event.inputs.PR-number) | |
}} | |
cancel-in-progress: true | |
env: | |
test_concurrency: 5 | |
jobs: | |
echo-inputs: | |
if: ${{ github.event_name == 'workflow_dispatch' }} | |
name: Echo Workflow Dispatch Inputs | |
runs-on: ubuntu-24.04 | |
steps: | |
- name: Echo Workflow Dispatch Inputs | |
run: | | |
echo '${{ tojson(inputs) }}' | |
commit-status-start: | |
name: Commit Status Start | |
runs-on: ubuntu-latest | |
steps: | |
- name: Set initial commit status | |
uses: myrotvorets/set-commit-status-action@3730c0a348a2ace3c110851bed53331bc6406e9f # v2.0.1 | |
with: | |
sha: ${{ inputs.SHA || github.sha }} | |
wait-for-images: | |
name: Wait for images | |
runs-on: ubuntu-latest | |
timeout-minutes: 30 | |
steps: | |
- name: Checkout context ref (trusted) | |
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 | |
with: | |
ref: ${{ inputs.context-ref || github.sha }} | |
persist-credentials: false | |
- name: Wait for images | |
uses: ./.github/actions/wait-for-images | |
with: | |
SHA: ${{ inputs.SHA }} | |
setup-and-test: | |
needs: [wait-for-images] | |
runs-on: ${{ vars.GH_RUNNER_EXTRA_POWER_UBUNTU_LATEST || 'ubuntu-latest' }} | |
name: 'Setup & Test' | |
env: | |
job_name: 'Setup & Test' | |
strategy: | |
fail-fast: false | |
max-parallel: 25 | |
matrix: | |
include: | |
- name: '1' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'rhel8.6-20241031.113911' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'vxlan' | |
host-fw: 'true' | |
- name: '2' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.4-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'disabled' | |
host-fw: 'true' | |
- name: '3' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.10-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'disabled' | |
endpoint-routes: 'true' | |
- name: '4' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.10-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'vxlan' | |
lb-mode: 'snat' | |
endpoint-routes: 'true' | |
egress-gateway: 'true' | |
- name: '5' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.15-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'disabled' | |
lb-mode: 'dsr' | |
endpoint-routes: 'true' | |
egress-gateway: 'true' | |
host-fw: 'false' # enabling breaks downgrading (missed tail calls) | |
- name: '6' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '6.1-20241104.124130' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
tunnel: 'vxlan' | |
lb-mode: 'snat' | |
egress-gateway: 'true' | |
host-fw: 'true' | |
lb-acceleration: 'testing-only' | |
ingress-controller: 'true' | |
bgp-control-plane: 'true' | |
- name: '7' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-next-20241105.013305' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'disabled' | |
lb-mode: 'snat' | |
egress-gateway: 'true' | |
lb-acceleration: 'testing-only' | |
ingress-controller: 'true' | |
local-redirect-policy: 'true' | |
node-local-dns: 'true' | |
- name: '8' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-next-20241105.013305' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'geneve' | |
endpoint-routes: 'true' | |
misc: 'socketLB.enabled=false,nodePort.enabled=true,bpf.masquerade=true' | |
local-redirect-policy: 'true' | |
node-local-dns: 'true' | |
- name: '9' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.10-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'vxlan' | |
encryption: 'wireguard' | |
encryption-node: 'false' | |
lb-mode: 'snat' | |
endpoint-routes: 'true' | |
egress-gateway: 'true' | |
- name: '10' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.15-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'disabled' | |
encryption: 'wireguard' | |
encryption-node: 'false' | |
encryption-strict-mode: 'true' | |
misc: 'socketLB.enabled=true' | |
local-redirect-policy: 'true' | |
node-local-dns: 'true' | |
- name: '11' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '6.1-20241104.124130' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'vxlan' | |
encryption: 'wireguard' | |
encryption-node: 'true' | |
lb-mode: 'snat' | |
egress-gateway: 'true' | |
ingress-controller: 'true' | |
misc: 'socketLB.hostNamespaceOnly=true' | |
local-redirect-policy: 'true' | |
node-local-dns: 'true' | |
- name: '12' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-next-20241105.013305' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'disabled' | |
encryption: 'wireguard' | |
encryption-node: 'true' | |
encryption-strict-mode: 'true' | |
lb-mode: 'snat' | |
egress-gateway: 'true' | |
ingress-controller: 'true' | |
- name: '13' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'rhel8.6-20241031.113911' | |
kube-proxy: 'iptables' | |
kpr: 'false' | |
tunnel: 'vxlan' | |
misc: 'policyCIDRMatchMode=nodes' | |
- name: '14' | |
# Switch to 5.15 until https://github.com/cilium/cilium/issues/27642 | |
# has been resolved. https://github.com/cilium/cilium/pull/30837#issuecomment-1960897445 | |
# explains why 5.4 might cause north-south-loadbalancing tests to | |
# fail. | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.15-20241104.124130' | |
kube-proxy: 'iptables' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'vxlan' | |
lb-mode: 'snat' | |
egress-gateway: 'true' | |
lb-acceleration: 'testing-only' | |
ingress-controller: 'true' | |
- name: '15' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-next-20241105.013305' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'disabled' | |
ingress-controller: 'true' | |
# Disable bpf.tproxy=true until https://github.com/cilium/cilium/issues/31918 | |
# has been resolved. | |
misc: 'bpfClockProbe=false,cni.uninstall=false' | |
ciliumendpointslice: 'true' | |
- name: '16' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: '5.15-20241104.124130' | |
kube-proxy: 'none' | |
kpr: 'true' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
tunnel: 'vxlan' | |
lb-mode: 'snat' | |
encryption: 'wireguard' | |
encryption-node: 'false' | |
host-fw: 'true' | |
ciliumendpointslice: 'true' | |
ingress-controller: 'true' | |
- name: '17' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit,bpf.masquerade=true,enableIPv4BIGTCP=true,enableIPv6BIGTCP=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
ipv6: 'true' | |
tunnel: 'disabled' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '18' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit-l2,bpf.masquerade=true,enableIPv4BIGTCP=true,enableIPv6BIGTCP=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
ipv6: 'true' | |
tunnel: 'disabled' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '19' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit,bpf.masquerade=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
tunnel: 'vxlan' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '20' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit-l2,bpf.masquerade=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
tunnel: 'vxlan' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '21' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit,bpf.masquerade=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
tunnel: 'geneve' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '22' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit-l2,bpf.masquerade=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
tunnel: 'geneve' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
- name: '23' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-20241105.013305' | |
misc: 'bpf.datapathMode=netkit,bpf.masquerade=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
ipv6: 'true' | |
tunnel: 'disabled' | |
devices: '{eth0,eth1}' | |
secondary-network: 'true' | |
ingress-controller: 'true' | |
host-fw: 'true' | |
skip-upgrade: 'true' | |
- name: '24' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-net-20241105.013305' | |
misc: 'bpf.datapathMode=netkit,bpf.masquerade=true,enableIPv4BIGTCP=true,enableIPv6BIGTCP=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
ipv6: 'true' | |
tunnel: 'disabled' | |
ciliumendpointslice: 'true' | |
endpoint-routes: 'true' | |
skip-upgrade: 'true' | |
- name: '25' | |
# renovate: datasource=docker depName=quay.io/lvh-images/kind | |
kernel: 'bpf-net-20241105.013305' | |
misc: 'bpf.datapathMode=netkit-l2,bpf.masquerade=true,enableIPv4BIGTCP=true,enableIPv6BIGTCP=true' | |
kube-proxy: 'none' | |
kpr: 'true' | |
ipv6: 'true' | |
tunnel: 'disabled' | |
ciliumendpointslice: 'true' | |
endpoint-routes: 'true' | |
skip-upgrade: 'true' | |
# Example of a feature that is being introduced, and we want to test | |
# it without performing an upgrade, we use skip-upgrade: 'true' | |
# - name: '23' | |
# # renovate: datasource=docker depName=quay.io/lvh-images/kind | |
# kernel: 'bpf-20241105.013305' | |
# misc: 'bpf.datapathMode=netkit-l2,bpf.masquerade=true' | |
# skip-upgrade: 'true' | |
timeout-minutes: 55 | |
steps: | |
- name: Collect Workflow Telemetry | |
uses: catchpoint/workflow-telemetry-action@94c3c3d9567a0205de6da68a76c428ce4e769af1 # v2.0.0 | |
with: | |
comment_on_pr: false | |
- name: Checkout context ref (trusted) | |
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 | |
with: | |
ref: ${{ inputs.context-ref || github.sha }} | |
persist-credentials: false | |
- name: Cleanup Disk space in runner | |
if: runner.name == 'ubuntu-latest' | |
uses: ./.github/actions/disk-cleanup | |
- name: Set Environment Variables | |
uses: ./.github/actions/set-env-variables | |
- name: Set up job variables | |
id: vars | |
run: | | |
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then | |
SHA="${{ inputs.SHA }}" | |
else | |
SHA="${{ github.sha }}" | |
fi | |
echo sha=${SHA} >> $GITHUB_OUTPUT | |
CILIUM_DOWNGRADE_VERSION=$(contrib/scripts/print-downgrade-version.sh stable) | |
echo downgrade_version=${CILIUM_DOWNGRADE_VERSION} >> $GITHUB_OUTPUT | |
- name: Derive stable Cilium installation config | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
id: cilium-stable-config | |
uses: ./.github/actions/cilium-config | |
with: | |
image-tag: ${{ steps.vars.outputs.downgrade_version }} | |
chart-dir: './untrusted/cilium-downgrade/install/kubernetes/cilium/' | |
tunnel: ${{ matrix.tunnel }} | |
devices: ${{ matrix.devices }} | |
endpoint-routes: ${{ matrix.endpoint-routes }} | |
ipv6: ${{ matrix.ipv6 }} | |
kpr: ${{ matrix.kpr }} | |
lb-mode: ${{ matrix.lb-mode }} | |
lb-acceleration: ${{ matrix.lb-acceleration }} | |
encryption: ${{ matrix.encryption }} | |
encryption-node: ${{ matrix.encryption-node }} | |
encryption-strict-mode: ${{ matrix.encryption-strict-mode }} | |
egress-gateway: ${{ matrix.egress-gateway }} | |
host-fw: ${{ matrix.host-fw }} | |
mutual-auth: false | |
ingress-controller: ${{ matrix.ingress-controller }} | |
misc: ${{ matrix.misc || 'bpfClockProbe=false,cni.uninstall=false' }} | |
ciliumendpointslice: ${{ matrix.ciliumendpointslice }} | |
local-redirect-policy: ${{ matrix.local-redirect-policy }} | |
bgp-control-plane: ${{ matrix.bgp-control-plane }} | |
- name: Derive newest Cilium installation config | |
id: cilium-newest-config | |
uses: ./.github/actions/cilium-config | |
with: | |
image-tag: ${{ steps.vars.outputs.sha }} | |
chart-dir: './untrusted/cilium-newest/install/kubernetes/cilium' | |
tunnel: ${{ matrix.tunnel }} | |
devices: ${{ matrix.devices }} | |
endpoint-routes: ${{ matrix.endpoint-routes }} | |
ipv6: ${{ matrix.ipv6 }} | |
kpr: ${{ matrix.kpr }} | |
lb-mode: ${{ matrix.lb-mode }} | |
lb-acceleration: ${{ matrix.lb-acceleration }} | |
encryption: ${{ matrix.encryption }} | |
encryption-node: ${{ matrix.encryption-node }} | |
encryption-strict-mode: ${{ matrix.encryption-strict-mode }} | |
egress-gateway: ${{ matrix.egress-gateway }} | |
host-fw: ${{ matrix.host-fw }} | |
mutual-auth: false | |
ingress-controller: ${{ matrix.ingress-controller }} | |
misc: ${{ matrix.misc || 'bpfClockProbe=false,cni.uninstall=false' }} | |
ciliumendpointslice: ${{ matrix.ciliumendpointslice }} | |
local-redirect-policy: ${{ matrix.local-redirect-policy }} | |
bgp-control-plane: ${{ matrix.bgp-control-plane }} | |
- name: Set Kind params | |
id: kind-params | |
shell: bash | |
run: | | |
IP_FAM="dual" | |
if [ "${{ matrix.ipv6 }}" == "false" ]; then | |
IP_FAM="ipv4" | |
fi | |
echo params="--xdp --secondary-network \"\" 3 \"\" \"\" ${{ matrix.kube-proxy }} $IP_FAM" >> $GITHUB_OUTPUT | |
- name: Provision K8s on LVH VM | |
uses: ./.github/actions/lvh-kind | |
with: | |
test-name: e2e-conformance | |
kernel: ${{ matrix.kernel }} | |
kind-params: "${{ steps.kind-params.outputs.params }}" | |
kind-image: ${{ env.KIND_K8S_IMAGE }} | |
- name: Install Cilium CLI | |
uses: cilium/cilium-cli@3286926bbf80fdd0103a372256459e577224f9f6 # v0.16.20 | |
with: | |
skip-build: ${{ env.CILIUM_CLI_SKIP_BUILD }} | |
image-repo: ${{ env.CILIUM_CLI_IMAGE_REPO }} | |
image-tag: ${{ steps.vars.outputs.sha }} | |
# Warning: since this is a privileged workflow, subsequent workflow job | |
# steps must take care not to execute untrusted code. | |
- name: Checkout pull request branch (NOT TRUSTED) | |
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 | |
with: | |
ref: ${{ steps.vars.outputs.sha }} | |
persist-credentials: false | |
path: untrusted/cilium-newest | |
sparse-checkout: | | |
install/kubernetes/cilium | |
examples | |
- name: Checkout ${{ steps.vars.outputs.downgrade_version }} branch to get the Helm chart | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 | |
with: | |
ref: ${{ steps.vars.outputs.downgrade_version }} | |
persist-credentials: false | |
path: untrusted/cilium-downgrade | |
sparse-checkout: | | |
install/kubernetes/cilium | |
- name: Install Cilium ${{ matrix.skip-upgrade == 'true' && 'from main' || steps.vars.outputs.downgrade_version }} | |
shell: bash | |
run: | | |
kubectl patch node kind-worker3 --type=json -p='[{"op":"add","path":"/metadata/labels/cilium.io~1no-schedule","value":"true"}]' | |
if ${{ matrix.skip-upgrade != 'true' }}; then | |
cilium install ${{ steps.cilium-stable-config.outputs.config }} | |
else | |
cilium install ${{ steps.cilium-newest-config.outputs.config }} | |
fi | |
cilium status --wait | |
kubectl get pods --all-namespaces -o wide | |
kubectl -n kube-system exec daemonset/cilium -- cilium status | |
mkdir -p cilium-junits | |
- name: Install node local DNS | |
if: ${{ matrix.node-local-dns == 'true' }} | |
shell: bash | |
run: | | |
kubedns=$(kubectl get svc kube-dns -n kube-system -o jsonpath={.spec.clusterIP}) && sed -i "s/__PILLAR__DNS__SERVER__/$kubedns/g;" untrusted/cilium-newest/examples/kubernetes-local-redirect/node-local-dns.yaml | |
sed -i "s/__PILLAR__UPSTREAM__SERVERS__/1.1.1.1/g;" untrusted/cilium-newest/examples/kubernetes-local-redirect/node-local-dns.yaml | |
kubectl apply -k untrusted/cilium-newest/examples/kubernetes-local-redirect | |
kubectl rollout status -n kube-system ds/node-local-dns | |
- name: Start conn-disrupt-test | |
shell: bash | |
run: | | |
# Create pods which establish long lived connections. It will be used by | |
# subsequent connectivity tests with --include-conn-disrupt-test to catch any | |
# interruption in such flows. | |
cilium connectivity test --include-conn-disrupt-test --conn-disrupt-test-setup \ | |
--conn-disrupt-test-restarts-path "./cilium-conn-disrupt-restarts" \ | |
--conn-disrupt-dispatch-interval 0ms | |
- name: Upgrade Cilium | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
shell: bash | |
run: | | |
cilium upgrade \ | |
${{ steps.cilium-newest-config.outputs.config }} | |
cilium status --wait --wait-duration=10m | |
kubectl get pods --all-namespaces -o wide | |
kubectl -n kube-system exec daemonset/cilium -- cilium status | |
- name: Setup Cilium CLI flags | |
id: cli-flags | |
run: | | |
EXTRA=() | |
if [ "${{ matrix.secondary-network }}" = "true" ]; then | |
EXTRA+=("\"--secondary-network-iface=eth1\"") | |
fi | |
if [ "${{ matrix.encryption-strict-mode }}" = "true" ]; then | |
# "Test Cilium after upgrade" ran strict-mode-encryption test which caused temporary drops. | |
EXTRA+=("\"--expected-drop-reasons=+Traffic is unencrypted\"") | |
fi | |
echo flags="--include-unsafe-tests \ | |
--collect-sysdump-on-failure \ | |
--flush-ct \ | |
--sysdump-hubble-flows-count=1000000 \ | |
--sysdump-hubble-flows-timeout=5m \ | |
--sysdump-output-filename \"cilium-sysdump-${{ matrix.name }}-<ts>\" \ | |
--junit-file \"cilium-junits/${{ env.job_name }} (${{ matrix.name }}).xml\" \ | |
--junit-property github_job_step=\"Run tests upgrade 2 (${{ matrix.name }})\" \ | |
${EXTRA[@]}" >> $GITHUB_OUTPUT | |
- name: Test Cilium ${{ matrix.skip-upgrade != 'true' && 'after upgrade' }} | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--include-conn-disrupt-test \ | |
--test "no-interrupted-connections" \ | |
--conn-disrupt-test-restarts-path "./cilium-conn-disrupt-restarts" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Run sequential Cilium tests | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--test "seq-.*" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Run concurrent Cilium tests | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--test-concurrency=${{ env.test_concurrency }} \ | |
--test "!seq-.*" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Setup conn dirstupt tests after a flush-ct Cilium tests | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--include-conn-disrupt-test \ | |
--conn-disrupt-test-setup \ | |
--conn-disrupt-test-restarts-path "./cilium-conn-disrupt-restarts" \ | |
--conn-disrupt-dispatch-interval 0ms | |
- name: Downgrade Cilium ${{ steps.vars.outputs.downgrade_version }} | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
shell: bash | |
run: | | |
cilium upgrade \ | |
${{ steps.cilium-stable-config.outputs.config }} | |
cilium status --wait --wait-duration=10m | |
kubectl get pods --all-namespaces -o wide | |
kubectl -n kube-system exec daemonset/cilium -- cilium status | |
- name: Test Cilium after downgrade to ${{ steps.vars.outputs.downgrade_version }} | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--include-conn-disrupt-test \ | |
--test "no-interrupted-connections" \ | |
--conn-disrupt-test-restarts-path "./cilium-conn-disrupt-restarts" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Run sequential Cilium tests | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--test "seq-.*" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Run concurrent Cilium tests | |
if: ${{ matrix.skip-upgrade != 'true' }} | |
shell: bash | |
run: | | |
cilium connectivity test \ | |
--test-concurrency=${{ env.test_concurrency }} \ | |
--test "!seq-.*" \ | |
${{ steps.cli-flags.outputs.flags }} | |
- name: Fetch artifacts | |
if: ${{ !success() }} | |
shell: bash | |
run: | | |
kubectl get pods --all-namespaces -o wide | |
cilium status | |
mkdir -p cilium-sysdumps | |
cilium sysdump --output-filename cilium-sysdump-${{ matrix.name }}-final | |
- name: Upload artifacts | |
if: ${{ !success() }} | |
uses: actions/upload-artifact@b4b15b8c7c6ac21ea08fcf65892d2ee8f75cf882 # v4.4.3 | |
with: | |
name: cilium-sysdumps-${{ matrix.name }} | |
path: cilium-sysdump-*.zip | |
- name: Upload JUnits [junit] | |
if: ${{ always() }} | |
uses: actions/upload-artifact@b4b15b8c7c6ac21ea08fcf65892d2ee8f75cf882 # v4.4.3 | |
with: | |
name: cilium-junits-${{ matrix.name }} | |
path: cilium-junits/*.xml | |
- name: Publish Test Results As GitHub Summary | |
if: ${{ always() }} | |
uses: aanm/junit2md@332ebf0fddd34e91b03a832cfafaa826306558f9 # v0.0.3 | |
with: | |
junit-directory: "cilium-junits" | |
merge-upload: | |
if: ${{ always() }} | |
name: Merge and Upload Artifacts | |
runs-on: ubuntu-latest | |
needs: setup-and-test | |
steps: | |
- name: Checkout context ref (trusted) | |
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 | |
with: | |
ref: ${{ inputs.context-ref || github.sha }} | |
persist-credentials: false | |
- name: Merge Sysdumps | |
uses: ./.github/actions/merge-artifacts | |
with: | |
name: cilium-sysdumps | |
pattern: cilium-sysdumps-* | |
token: ${{ secrets.GITHUB_TOKEN }} | |
- name: Merge JUnits | |
uses: ./.github/actions/merge-artifacts | |
with: | |
name: cilium-junits | |
pattern: cilium-junits-* | |
token: ${{ secrets.GITHUB_TOKEN }} | |
commit-status-final: | |
if: ${{ always() }} | |
name: Commit Status Final | |
needs: setup-and-test | |
runs-on: ubuntu-latest | |
steps: | |
- name: Set final commit status | |
uses: myrotvorets/set-commit-status-action@3730c0a348a2ace3c110851bed53331bc6406e9f # v2.0.1 | |
with: | |
sha: ${{ inputs.SHA || github.sha }} | |
status: ${{ needs.setup-and-test.result }} |