FAQs
General Questions
Q: What is the difference between Model Repositories and Model Endpoints?
A: Model Repositories are storage systems for model files, while Model Endpoints are running services that serve inference requests. Repositories store models; Endpoints use models from repositories to serve API requests.
Q: Can I use Model Repositories without Model Endpoints?
A: Yes! You can use Model Repositories independently for storage, versioning, and sharing models. They're also useful with Instances.
Q: What file formats are supported?
A: Model Repositories support any file format. There's no restriction - you can store PyTorch models, TensorFlow models, ONNX, TensorRT engines, or any custom format.
Storage Questions
Q: What storage options are available?
A: Three options:
- New EOS Bucket - TIR provisions a new bucket automatically
- Existing EOS Bucket - Link to an existing EOS bucket
- External EOS Bucket - Connect to external S3-compatible storage
Q: Can I change the storage type after creation?
A: No, storage type cannot be changed after creation. You'll need to create a new repository with the desired storage type.
Q: How do I access files in a Model Repository?
A: Multiple ways:
- Dashboard: File Browser tab for visual browsing
- CLI: Minio CLI (
mc) ors3cmd - SDK: Python SDK or other language SDKs
- API: S3-compatible REST API
Q: Can I encrypt my models?
A: Yes! EOS encryption is available. Enable encryption when creating a new EOS bucket.
Encryption must be enabled at bucket creation.
Versioning Questions
Q: How do I version my models?
A: Use folder structures:
/v1,/v2for numbered versions/production,/stagingfor environments/experiments/exp-001for experimental versions
Q: Can I delete specific versions?
A: Yes, you can delete specific folders/files using CLI or SDK. Be careful - deletion is permanent.
Q: How many versions should I keep?
A: Best practice: Keep 2-3 previous versions for rollback. Archive or delete older versions to save storage costs.
Integration Questions
Q: How do I link a Model Repository to a Model Endpoint?
A: You can link a repository in two ways.
- From Model Repository: Use the Deploy Model option in the repository table; you are taken to the Model Endpoint flow, where you select the framework (e.g., vLLM, SGLang, NVIDIA Triton) and link the repository to the endpoint.
- From Model Endpoints: Create a Model Endpoint, select Link with Model Repository, choose your repository, and specify the model path (e.g.,
/or/v1). In both cases, the endpoint automatically downloads model files from the repository during startup.
Troubleshooting
Q: Upload is slow. How can I speed it up?
A: Try the following:
- Use parallel uploads:
mc cp --parallel - Ensure good network connectivity
- Compress files before upload
- Use SDK for programmatic uploads (may be faster)
Q: Model Endpoint can't find my model files. What's wrong?
A: Check:
- Model path in endpoint configuration matches folder structure
- Files are actually uploaded (verify via File Browser)
- Bucket name is correct
- Check endpoint logs for specific errors