docker model run
| 内容説明 | Run a model and interact with it using a submitted prompt or chat mode |
|---|---|
| 利用方法 | docker model run MODEL [PROMPT] |
内容説明
When you run a model, Docker calls an inference server API endpoint hosted by the Model Runner through Docker Desktop. The model stays in memory until another model is requested, or until a pre-defined inactivity timeout is reached (currently 5 minutes).
You do not have to use Docker model run before interacting with a specific model from a host process or from within a container. Model Runner transparently loads the requested model on-demand, assuming it has been pulled and is locally available.
You can also use chat mode in the Docker Desktop Dashboard when you select the model in the Models tab.
オプション
| オプション | デフォルト | 内容説明 |
|---|---|---|
--color | auto | Use colored output (auto|yes|no) |
--debug | Enable debug logging | |
--ignore-runtime-memory-check | Do not block pull if estimated runtime memory for model exceeds system resources. |
利用例
One-time prompt
docker model run ai/smollm2 "Hi"
Output:
Hello! How can I assist you today?
Interactive chat
docker model run ai/smollm2
Output:
Interactive chat mode started. Type '/bye' to exit.
> Hi
Hi there! It's SmolLM, AI assistant. How can I help you today?
> /bye
Chat session ended.