We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 2b89299 commit 9f70085Copy full SHA for 9f70085
.github/workflows/flash_attention.yml
@@ -45,6 +45,7 @@ jobs:
45
--ulimit stack=67108864 \
46
--tty \
47
--detach \
48
+ --privileged \
49
--security-opt seccomp=unconfined \
50
--shm-size=4g \
51
-v "${GITHUB_WORKSPACE}:/tmp/workspace" \
@@ -54,13 +55,12 @@ jobs:
54
55
56
# Build and run FlashAttention CuTe DSL
57
docker exec -t "${container_name}" bash -c "
- set -x
58
+ nvidia-smi
59
+
60
pushd fa4
61
python setup.py install
62
pip install -e flash_attn/cute/
63
- nvidia-smi
-
64
echo '<h1>B200' >> /tmp/workspace/fa4_output.txt
65
nvidia-smi -q -d POWER | grep 'Current Power Limit' | head -1 | cut -d : -f 2 >> /tmp/workspace/fa4_output.txt
66
echo '</h1>' >> /tmp/workspace/fa4_output.txt
0 commit comments