diff --git a/.github/workflows/debug-dind.yml b/.github/workflows/debug-dind.yml index b282d24f..e13915a0 100644 --- a/.github/workflows/debug-dind.yml +++ b/.github/workflows/debug-dind.yml @@ -10,7 +10,7 @@ jobs: group: gcp-ct5lp-hightpu-8t container: image: us-central1-docker.pkg.dev/tpu-pytorch-releases/docker/xla:r2.4.0_3.10_tpuvm - options: --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} -v /mnt/hf_cache:/mnt/hf_cache -e PJRT_DEVICE=TPU + options: --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} --network test-network -v /mnt/hf_cache:/mnt/hf_cache -e PJRT_DEVICE=TPU steps: - name: Checkout code @@ -33,7 +33,7 @@ jobs: run: | # Run the whoami container with environment variables # @pauline adding --ipc host, --privileged and/or --network host did not help me. I also tried to use different ports 80, 8080, 5001 - docker run ${{ vars.V5_LITEPOD_8_ENV}} -d --name network-test -p 5001:80 traefik/whoami + docker run ${{ vars.V5_LITEPOD_8_ENV}} -d --name network-test --network test-network -p 5001:80 traefik/whoami # Wait for container to be ready sleep 5 diff --git a/.github/workflows/test-pytorch-xla-tpu-tgi-integration.yml b/.github/workflows/test-pytorch-xla-tpu-tgi-integration.yml index 71875310..af30bc36 100644 --- a/.github/workflows/test-pytorch-xla-tpu-tgi-integration.yml +++ b/.github/workflows/test-pytorch-xla-tpu-tgi-integration.yml @@ -1,7 +1,7 @@ name: Optimum TPU / Test TGI on TPU / Integration Tests on: - # push: + push: pull_request: branches: [ main ] paths: @@ -20,7 +20,7 @@ jobs: container: image: us-central1-docker.pkg.dev/tpu-pytorch-releases/docker/xla:r2.4.0_3.10_tpuvm - options: --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} -v /mnt/hf_cache:/mnt/hf_cache -e PJRT_DEVICE=TPU + options: --shm-size "16gb" --ipc host --privileged ${{ vars.V5_LITEPOD_8_ENV}} --network test-network -v /mnt/hf_cache:/mnt/hf_cache -e PJRT_DEVICE=TPU env: PJRT_DEVICE: TPU @@ -49,6 +49,7 @@ jobs: run: | # Start docker container in background docker run -d --name tgi-tests-gpt2 \ + --network test-network \ -e LOG_LEVEL=info,text_generation_router,text_generation_launcher=debug \ -e HF_HUB_ENABLE_HF_TRANSFER=0 \ -e MAX_BATCH_SIZE=4 \