From e3df3a9f5d77e6d6d81fded9bb130622e7a226b6 Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Mon, 22 May 2023 12:14:21 -0400 Subject: [PATCH] cuda/pytorch matrix builds --- .github/workflows/main.yml | 26 ++++++++++++++++++++++---- 1 file changed, 22 insertions(+), 4 deletions(-) diff --git a/.github/workflows/main.yml b/.github/workflows/main.yml index 8919a8825..6e51fef3c 100644 --- a/.github/workflows/main.yml +++ b/.github/workflows/main.yml @@ -10,6 +10,15 @@ jobs: build-axolotl: if: github.repository_owner == 'OpenAccess-AI-Collective' # this job needs to be run on self-hosted GPU runners... + strategy: + matrix: + include: + - cuda: cu118 + cuda_version: 11.8.0 + pytorch: 2.0.0 + - cuda: cu117 + cuda_version: 11.7.0 + pytorch: 1.13.1 runs-on: self-hosted steps: - name: Checkout @@ -31,10 +40,10 @@ jobs: with: context: . build-args: | - BASE_TAG=${{ github.ref_name }}-base + BASE_TAG=${{ github.ref_name }}-base-${{ matrix.cuda }}-${{ matrix.pytorch }} file: ./docker/Dockerfile push: ${{ github.event_name != 'pull_request' }} - tags: ${{ steps.metadata.outputs.tags }} + tags: ${{ steps.metadata.outputs.tags }}-${{ matrix.cuda }}-${{ matrix.pytorch }} labels: ${{ steps.metadata.outputs.labels }} cache-from: type=gha cache-to: type=gha,mode=max @@ -42,6 +51,15 @@ jobs: needs: build-axolotl if: github.repository_owner == 'OpenAccess-AI-Collective' # this job needs to be run on self-hosted GPU runners... + strategy: + matrix: + include: + - cuda: cu118 + cuda_version: 11.8.0 + pytorch: 2.0.0 + - cuda: cu117 + cuda_version: 11.7.0 + pytorch: 1.13.1 runs-on: self-hosted steps: - name: Checkout @@ -63,10 +81,10 @@ jobs: with: context: . build-args: | - BASE_TAG=${{ github.ref_name }} + BASE_TAG=${{ github.ref_name }}-${{ matrix.cuda }}-${{ matrix.pytorch }} file: ./docker/Dockerfile-runpod push: ${{ github.event_name != 'pull_request' }} - tags: ${{ steps.metadata.outputs.tags }} + tags: ${{ steps.metadata.outputs.tags }}-${{ matrix.cuda }}-${{ matrix.pytorch }} labels: ${{ steps.metadata.outputs.labels }} cache-from: type=gha cache-to: type=gha,mode=max