Principle:Openai Openai python Fine Tuned Model Usage
| Knowledge Sources | |
|---|---|
| Domains | Fine_Tuning, Model_Deployment |
| Last Updated | 2026-02-15 00:00 GMT |
Overview
A model deployment pattern for using fine-tuned models in production by referencing the custom model identifier in standard API calls.
Description
After a fine-tuning job succeeds, the resulting model is available via its unique identifier (format: ft:model:org:suffix:id). This model can be used in any standard API call (chat completions, responses) by passing it as the model parameter. Models can also be deleted when no longer needed to free resources and stop incurring storage costs.
Usage
Use this principle after a fine-tuning job completes successfully. Retrieve the model name from job.fine_tuned_model and use it in chat completions or responses API calls. Delete the model when it is no longer needed.
Theoretical Basis
# Use fine-tuned model (same API, different model name)
response = chat_completions(
model="ft:gpt-4o-mini-2024-07-18:org:suffix:id",
messages=[...]
)
# Delete when no longer needed
delete_model(model="ft:gpt-4o-mini-2024-07-18:org:suffix:id")