Skip to main content
Models power your AI applications. Add a model to a project to deploy it, then make inference requests.

Deploy a model

Add a model to your project, then deploy it:
adaptive.models.add_to_project(model="llama-3.1-8b-instruct")
adaptive.models.deploy(model="llama-3.1-8b-instruct", wait=True)

# Or use attach() to do both in one call
adaptive.models.attach(model="llama-3.1-8b-instruct", wait=True)
ParameterTypeRequiredDescription
modelstrYesModel key from the registry
waitboolNoBlock until model is online (default: False)
make_defaultboolNoSet as default model for the project
The model becomes available within a few minutes. Adaptive supports most transformer-based models including Llama, Qwen, Gemma, Mistral, and DeepSeek. See Integrations for proprietary models.Multimodal models (suffixed -multimodal) include a vision encoder that is automatically spawned alongside the decoder. The vision encoder is frozen during training. See Multimodal StringThread for supported models and image handling.See SDK Reference for all model methods.