Try better ROCm configuration

This commit is contained in:
doctorpangloss 2024-09-12 17:45:17 -07:00
parent 3548fcd3c2
commit 5d1eda7832

View File

@ -53,7 +53,7 @@ jobs:
- labels: [self-hosted, Linux, X64, cpu] - labels: [self-hosted, Linux, X64, cpu]
container: "ubuntu:22.04" container: "ubuntu:22.04"
- labels: [self-hosted, Linux, X64, rocm-7600-8gb] - labels: [self-hosted, Linux, X64, rocm-7600-8gb]
container: "rocm/pytorch:rocm6.1.3_ubuntu22.04_py3.10_pytorch_release-2.1.2" container: "rocm/pytorch:rocm6.2_ubuntu22.04_py3.10_pytorch_release_2.3.0"
- labels: [self-hosted, Linux, X64, cuda-3060-12gb] - labels: [self-hosted, Linux, X64, cuda-3060-12gb]
container: "nvcr.io/nvidia/pytorch:24.03-py3" container: "nvcr.io/nvidia/pytorch:24.03-py3"
steps: steps:
@ -73,7 +73,8 @@ jobs:
pip install --break-system-packages --no-build-isolation .[dev] || pip install --no-build-isolation .[dev] pip install --break-system-packages --no-build-isolation .[dev] || pip install --no-build-isolation .[dev]
- name: Run tests - name: Run tests
run: | run: |
export HSA_OVERRIDE_GFX_VERSION=11.0.0 export HSA_OVERRIDE_GFX_VERSION=11.0.2
export HIP_VISIBLE_DEVICES=0
export PYTORCH_HIP_ALLOC_CONF=expandable_segments:True export PYTORCH_HIP_ALLOC_CONF=expandable_segments:True
export NUMBA_THREADING_LAYER=omp export NUMBA_THREADING_LAYER=omp
export AMD_SERIALIZE_KERNEL=1 export AMD_SERIALIZE_KERNEL=1