Skip to content

Commit eeb7a79

Browse files
authored
Bump torch version to 2.7 (#1832)
1 parent a51c96c commit eeb7a79

18 files changed

+45
-45
lines changed

.github/workflows/docker.yaml

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -17,11 +17,11 @@ jobs:
1717
strategy:
1818
matrix:
1919
include:
20-
- name: "2.6.0_cu124"
21-
base_image: mosaicml/pytorch:2.6.0_cu124-python3.12-ubuntu22.04
20+
- name: "2.7.0_cu128"
21+
base_image: mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04
2222
dep_groups: "[all]"
23-
- name: "2.6.0_cu124_aws"
24-
base_image: mosaicml/pytorch:2.6.0_cu124-python3.12-ubuntu22.04-aws
23+
- name: "2.7.0_cu128_aws"
24+
base_image: mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04-aws
2525
dep_groups: "[all]"
2626
steps:
2727

.github/workflows/pr-cpu.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,9 +21,9 @@ jobs:
2121
strategy:
2222
matrix:
2323
include:
24-
- name: "cpu-2.6.0"
24+
- name: "cpu-2.7.0"
2525
pip_deps: "[all-cpu]"
26-
container: mosaicml/pytorch:2.6.0_cpu-python3.12-ubuntu22.04
26+
container: mosaicml/pytorch:2.7.0_cpu-python3.12-ubuntu22.04
2727
markers: "not gpu"
2828
pytest_command: "coverage run -m pytest"
2929
steps:

.github/workflows/pr-gpu.yaml

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@ jobs:
2222
fail-fast: false
2323
matrix:
2424
include:
25-
- name: "gpu-2.6.0-1"
26-
container: mosaicml/llm-foundry:2.6.0_cu124-latest
25+
- name: "gpu-2.7.0-1"
26+
container: mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04
2727
markers: "gpu"
2828
pip_deps: "[all]"
2929
pytest_command: "coverage run -m pytest"
@@ -51,8 +51,8 @@ jobs:
5151
fail-fast: false
5252
matrix:
5353
include:
54-
- name: "gpu-2.6.0-2"
55-
container: mosaicml/llm-foundry:2.6.0_cu124-latest
54+
- name: "gpu-2.7.0-2"
55+
container: mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04
5656
markers: "gpu"
5757
pip_deps: "[all]"
5858
pytest_command: "coverage run -m pytest"
@@ -80,8 +80,8 @@ jobs:
8080
fail-fast: false
8181
matrix:
8282
include:
83-
- name: "gpu-2.6.0-4"
84-
container: mosaicml/llm-foundry:2.6.0_cu124-latest
83+
- name: "gpu-2.7.0-4"
84+
container: mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04
8585
markers: "gpu"
8686
pip_deps: "[all]"
8787
pytest_command: "coverage run -m pytest"

.github/workflows/release.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@ jobs:
9393
${{ env.AWS_DOCKER_TAG }}
9494
${{ env.AWS_LATEST_TAG }}
9595
build-args: |
96-
BASE_IMAGE=mosaicml/pytorch:2.6.0_cu124-python3.12-ubuntu22.04-aws
96+
BASE_IMAGE=mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04-aws
9797
BRANCH_NAME=${{ env.BRANCH_NAME }}
9898
DEP_GROUPS=[all]
9999
KEEP_FOUNDRY=true
@@ -108,7 +108,7 @@ jobs:
108108
${{ env.DOCKER_TAG }}
109109
${{ env.LATEST_TAG }}
110110
build-args: |
111-
BASE_IMAGE=mosaicml/pytorch:2.6.0_cu124-python3.12-ubuntu22.04
111+
BASE_IMAGE=mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04
112112
BRANCH_NAME=${{ env.BRANCH_NAME }}
113113
DEP_GROUPS=[all]
114114
KEEP_FOUNDRY=true

README.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -113,24 +113,24 @@ If you have success/failure using LLM Foundry on other systems, please let us kn
113113

114114
| Device | Torch Version | Cuda Version | Status |
115115
| -------------- | ------------- | ------------ | ---------------------------- |
116-
| A100-40GB/80GB | 2.6.0 | 12.4 | :white_check_mark: Supported |
117-
| H100-80GB | 2.6.0 | 12.4 | :white_check_mark: Supported |
116+
| A100-40GB/80GB | 2.7.0 | 12.8 | :white_check_mark: Supported |
117+
| H100-80GB | 2.7.0 | 12.8 | :white_check_mark: Supported |
118118

119119
## MosaicML Docker Images
120120
We highly recommend using our prebuilt Docker images. You can find them here: https://hub.docker.com/orgs/mosaicml/repositories.
121121

122122
The `mosaicml/pytorch` images are pinned to specific PyTorch and CUDA versions, and are stable and rarely updated.
123123

124124
The `mosaicml/llm-foundry` images are built with new tags upon every commit to the `main` branch.
125-
You can select a specific commit hash such as `mosaicml/llm-foundry:2.6.0_cu124-9867a7b` or take the latest one using `mosaicml/llm-foundry:2.6.0_cu124-latest`.
125+
You can select a specific commit hash such as `mosaicml/llm-foundry:2.7.0_cu128-9867a7b` or take the latest one using `mosaicml/llm-foundry:2.7.0_cu128-latest`.
126126

127127
**Please Note:** The `mosaicml/llm-foundry` images do not come with the `llm-foundry` package preinstalled, just the dependencies. You will still need to `pip install llm-foundry` either from PyPi or from source.
128128

129129
| Docker Image | Torch Version | Cuda Version | LLM Foundry dependencies installed? |
130130
| ------------------------------------------------------ | ------------- | ----------------- | ----------------------------------- |
131-
| `mosaicml/pytorch:2.6.0_cu124-python3.12-ubuntu22.04` | 2.6.0 | 12.4 (Infiniband) | No |
132-
| `mosaicml/llm-foundry:2.6.0_cu124-latest` | 2.6.0 | 12.4 (Infiniband) | Yes |
133-
| `mosaicml/llm-foundry:2.6.0_cu124_aws-latest` | 2.6.0 | 12.4 (EFA) | Yes |
131+
| `mosaicml/pytorch:2.7.0_cu128-python3.12-ubuntu22.04` | 2.7.0 | 12.8 (Infiniband) | No |
132+
| `mosaicml/llm-foundry:2.7.0_cu128-latest` | 2.7.0 | 12.8 (Infiniband) | Yes |
133+
| `mosaicml/llm-foundry:2.7.0_cu128_aws-latest` | 2.7.0 | 12.8 (EFA) | Yes |
134134

135135

136136
# Installation

llmfoundry/_version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,4 +2,4 @@
22
# SPDX-License-Identifier: Apache-2.0
33
"""The LLM Foundry Version."""
44

5-
__version__ = '0.21.0.dev0'
5+
__version__ = '0.22.0.dev0'

mcli/mcli-1b-eval.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
integrations:
22
- integration_type: git_repo
33
git_repo: mosaicml/llm-foundry
4-
git_branch: v0.20.0
4+
git_branch: v0.21.0
55
# git_commit: # OR use your commit hash
66
pip_install: .[gpu]
77
ssh_clone: false # Should be true if using a private repo
88

99
command: |
1010
cd llm-foundry/scripts/
1111
composer eval/eval.py /mnt/config/parameters.yaml
12-
image: mosaicml/llm-foundry:2.6.0_cu124-latest
12+
image: mosaicml/llm-foundry:2.7.0_cu128-latest
1313
name: mpt-1b-eval
1414

1515
compute:

mcli/mcli-1b-max-seq-len-8k.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
integrations:
22
- integration_type: git_repo
33
git_repo: mosaicml/llm-foundry
4-
git_branch: v0.20.0
4+
git_branch: v0.21.0
55
# git_commit: # OR use your commit hash
66
pip_install: .[gpu]
77
ssh_clone: false # Should be true if using a private repo
@@ -17,7 +17,7 @@ command: |
1717
--out_root ./my-copy-c4 --splits train_small val_small \
1818
--concat_tokens 8192 --tokenizer EleutherAI/gpt-neox-20b --eos_text '<|endoftext|>'
1919
composer train/train.py /mnt/config/parameters.yaml
20-
image: mosaicml/llm-foundry:2.6.0_cu124-latest
20+
image: mosaicml/llm-foundry:2.7.0_cu128-latest
2121
name: mpt-1b-ctx-8k-gpus-8
2222

2323
compute:

mcli/mcli-1b.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
integrations:
22
- integration_type: git_repo
33
git_repo: mosaicml/llm-foundry
4-
git_branch: v0.20.0
4+
git_branch: v0.21.0
55
# git_commit: # OR use your commit hash
66
pip_install: .[gpu]
77
ssh_clone: false # Should be true if using a private repo
@@ -21,7 +21,7 @@ command: |
2121
eval_loader.dataset.split=val_small \
2222
max_duration=100ba \
2323
eval_interval=0
24-
image: mosaicml/llm-foundry:2.6.0_cu124-latest
24+
image: mosaicml/llm-foundry:2.7.0_cu128-latest
2525
name: mpt-1b-gpus-8
2626

2727
compute:

mcli/mcli-benchmark-mpt.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,12 +6,12 @@ compute:
66
# cluster: TODO # Name of the cluster to use for this run
77
# gpu_type: a100_80gb # Type of GPU to use. We use a100_80gb in our experiments
88

9-
image: mosaicml/llm-foundry:2.6.0_cu124-latest
9+
image: mosaicml/llm-foundry:2.7.0_cu128-latest
1010

1111
integrations:
1212
- integration_type: git_repo
1313
git_repo: mosaicml/llm-foundry
14-
git_branch: v0.20.0
14+
git_branch: v0.21.0
1515
# git_commit: # OR use your commit hash
1616
pip_install: .[gpu]
1717

0 commit comments

Comments
 (0)