Shortcuts

torchtune CLI

This page is the documentation for using the torchtune CLI - a convenient way to download models, find and copy relevant recipes/configs, and run recipes. It is automatically available when you install torchtune.

Getting started

The --help option will show all the possible commands available through the torchtune CLI, with a short description of each.

$ tune --help
usage: tune [-h] {download,ls,cp,run,validate} ...

Welcome to the torchtune CLI!

options:
-h, --help            show this help message and exit

subcommands:
  {download,ls,cp,run,validate}
    download            Download a model from the Hugging Face Hub.
    ls                  List all built-in recipes and configs
    ...

The --help option is convenient for getting more details about any command. You can use it anytime to list all available options and their details. For example, tune download --help provides more information on how to download files using the CLI.

Download a model

The tune download <path> command downloads any model from the Hugging Face Hub.

--output-dir

Directory in which to save the model.

--output-dir-use-symlinks

To be used with output-dir. If set to ‘auto’, the cache directory will be used and the file will be either duplicated or symlinked to the local directory depending on its size. It set to True, a symlink will be created, no matter the file size. If set to False, the file will either be duplicated from cache (if already exists) or downloaded from the Hub and not cached.

--hf-token

Hugging Face API token. Needed for gated models like Llama.

--ignore-patterns

If provided, files matching any of the patterns are not downloaded. Defaults to ignoring safetensors files to avoid downloading duplicate weights.

$ tune download meta-llama/Meta-Llama-3-8B-Instruct
Successfully downloaded model repo and wrote to the following locations:
./model/config.json
./model/README.md
./model/model-00001-of-00002.bin
...

Download a gated model

A lot of recent large pretrained models released from organizations like Meta or MistralAI require you to agree to the usage terms and conditions before you are allowed to download their model. If this is the case, you can specify a Hugging Face access token.

You can find the access token here.

$ tune download meta-llama/Meta-Llama-3-8B-Instruct --hf-token <TOKEN>
Successfully downloaded model repo and wrote to the following locations:
./model/config.json
./model/README.md
./model/model-00001-of-00002.bin
...

Note

If you’d prefer, you can also use huggingface-cli login to permanently login to the Hugging Face Hub on your machine. The tune download command will pull the access token from your environment.

Specify model files you don’t want to download

Some checkpoint directories can be very large and it can eat up a lot of bandwith and local storage to download the all of the files every time, even if you might not need a lot of them. This is especially common when the same checkpoint exists in different formats. You can specify patterns to ignore to prevent downloading files with matching names. By default we ignore safetensor files, but if you want to include all files you can pass in an empty string.

$ tune download meta-llama/Meta-Llama-3-8B-Instruct --hf-token <TOKEN> --ignore-patterns ""
Successfully downloaded model repo and wrote to the following locations:
./model/config.json
./model/README.md
./model/model-00001-of-00030.safetensors
...

Note

Just because a model can be downloaded does not mean that it will work OOTB with torchtune’s built-in recipes or configs. For a list of supported model families and architectures, see models.

List built-in recipes and configs

The tune ls command lists out all the built-in recipes and configs within torchtune.

$ tune ls
RECIPE                                   CONFIG
full_finetune_single_device              llama2/7B_full_low_memory
                                         code_llama2/7B_full_low_memory
                                         llama3/8B_full_single_device
                                         mistral/7B_full_low_memory
                                         phi3/mini_full_low_memory
full_finetune_distributed                llama2/7B_full
                                         llama2/13B_full
                                         llama3/8B_full
                                         llama3/70B_full
...

Copy a built-in recipe or config

The tune cp <recipe|config> <path> command copies built-in recipes and configs to a provided location. This allows you to make a local copy of a library recipe or config to edit directly for yourself.

-n, --no-clobber

Do not overwrite destination if it already exists

--make-parents

Create parent directories for destination if they do not exist. If not set to True, will error if parent directories do not exist

$ tune cp lora_finetune_distributed .
Copied file to ./lora_finetune_distributed.py

Run a recipe

The tune run <recipe> --config <config> is a wrapper around torchrun. tune run allows you to specify a built-in recipe or config by name, or by path to use your local recipes/configs.

To run a tune recipe

tune run lora_finetune_single_device --config llama3/8B_lora_single_device

Specifying distributed (torchrun) arguments

tune run supports launching distributed runs by passing through arguments preceding the recipe directly to torchrun. This follows the pattern used by torchrun of specifying distributed and host machine flags before the script (recipe). For a full list of available flags for distributed setup, see the torchrun docs.

Some common flags:

--nproc-per-node

Number of workers per node; supported values: [auto, cpu, gpu, int].

--nnodes

Number of nodes, or the range of nodes in form <minimum_nodes>:<maximum_nodes>.

--max-restarts

Maximum number of worker group restarts before failing.

--rdzv-backend

Rendezvous backend.

--rdzv-endpoint

Rendezvous backend endpoint; usually in form <host>:<port>.

tune run --nnodes=1 --nproc-per-node=4 lora_finetune_distributed --config llama3/8B_lora

Note

If no arguments are provided before the recipe, tune will bypass torchrun and launch directly with python. This can simplify running and debugging recipes when distributed isn’t needed. If you want to launch with torchrun, but use only a single device, you can specify tune run --nnodes=1 --nproc-per-node=1 <recipe> --config <config>.

Running a custom (local) recipe and config

To use tune run with your own local recipes and configs, simply pass in a file path instead of a name to the run command. You can mix and match a custom recipe with a torchtune config or vice versa or you can use both custom configs and recipes.

tune run my/fancy_lora.py --config my/configs/8B_fancy_lora.yaml

Overriding the config

You can override existing parameters from the command line using a key=value format. Let’s say you want to set the number of training epochs to 1. Further information on config overrides can be found here.

tune run <RECIPE> --config <CONFIG> epochs=1

Validate a config

The tune validate <config> command will validate that your config is formatted properly.

# If you've copied over a built-in config and want to validate custom changes
$ tune validate my_configs/llama3/8B_full.yaml
Config is well-formed!

Docs

Access comprehensive developer documentation for PyTorch

View Docs

Tutorials

Get in-depth tutorials for beginners and advanced developers

View Tutorials

Resources

Find development resources and get your questions answered

View Resources