Instructions to use nvidia/Alpamayo-R1-10B with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use nvidia/Alpamayo-R1-10B with Transformers:
# Load model directly from transformers import AlpamayoR1 model = AlpamayoR1.from_pretrained("nvidia/Alpamayo-R1-10B", dtype="auto") - Notebooks
- Google Colab
- Kaggle
download the model parameter files locally by myself
#7
by wangxue9 - opened
My network is restricted and I cannot connect to Hugging Face. If I want to download the model parameter files locally by myself, which files should I download?
https://huggingface.co/nvidia/Alpamayo-R1-10B/tree/main
Hi, thanks for your question.
You would need all the weight files (.safetensors), the model.safetensors.index.json file to map tensor keys to tensors, and the model config.json (which are close to all the files in this model repo).
Closing as completed
BorisIvanovic changed discussion status to closed