Troubleshooting
Use this page to diagnose and resolve common issues when installing and running RapidFire AI.
Note
RapidFire AI requires Python 3.12+. Verify your shell’s python3
is 3.12 before creating/activating the venv.
Quick diagnostics
If you encounter any error, run the doctor command to get a complete diagnostic report (Python env, relevant packages, GPU/CUDA, and key environment variables):
rapidfireai doctor
Hugging Face permission errors (login not picked up)
Run the Hugging Face login from the SAME virtual environment where you installed RapidFire AI.
Activate your venv and log in:
source .venv/bin/activate
pip install huggingface-hub
huggingface-cli login
huggingface-cli whoami # Prints the HF account/orgs for the credentials this venv sees
Using Jupyter notebooks:
If you logged in while a notebook was already running, restart the notebook kernel so it picks up the new Hugging Face credentials.
Ensure the notebook uses the same venv kernel.
Port conflicts (services already running)
If you encounter port conflicts, you can kill existing processes.
lsof -t -i:5002 | xargs kill -9 # mlflow
lsof -t -i:8080 | xargs kill -9 # dispatcher
lsof -t -i:3000 | xargs kill -9 # frontend server
Select specific GPU(s) to use
Set the CUDA_VISIBLE_DEVICES
environment variable BEFORE running rapidfireai start
to control which GPU(s) RapidFire can see and use.
export CUDA_VISIBLE_DEVICES=2 # use GPU index 2 only
rapidfireai start
Multiple GPUs (example: GPUs 0 and 2):
export CUDA_VISIBLE_DEVICES=0,2
rapidfireai start
From a Python script (set before importing/starting RapidFire):
import os
os.environ["CUDA_VISIBLE_DEVICES"] = "2"
# then start your RapidFire workflow
See also
For known limitations and workarounds, see Known Issues.
If you are just getting started, follow the Walkthrough.