You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The `scaleway_inference_model` data source allows you to retrieve information about an inference model available in the Scaleway Inference API, either by providing the model's `name` or its `model_id`.
9
+
10
+
## Example Usage
11
+
12
+
### Basic
13
+
14
+
```hcl
15
+
data "scaleway_inference_model" "my_model" {
16
+
name = "meta/llama-3.1-8b-instruct:fp8"
17
+
}
18
+
```
19
+
20
+
## Argument Reference
21
+
22
+
You must provide either name or model_id, but not both.
23
+
24
+
-`name` (Optional, Conflicts with model_id) The fully qualified name of the model to look up (e.g., "meta/llama-3.1-8b-instruct:fp8"). The provider will search for a model with an exact name match in the selected region and project.
25
+
-`model_id` (Optional, Conflicts with name) The ID of the model to retrieve. Must be a valid UUID with locality (i.e., Scaleway's zoned UUID format).
26
+
-`project_id` (Optional) The project ID to use when listing models. If not provided, the provider default project is used.
27
+
-`region` (Optional) The region where the model is hosted. If not set, the provider default region is used.
28
+
29
+
## Attributes Reference
30
+
31
+
In addition to the input arguments above, the following attributes are exported:
32
+
33
+
-`id` - The unique identifier of the model.
34
+
-`tags` - Tags associated with the model.
35
+
-`status` - The current status of the model (e.g., ready, error, etc.).
36
+
-`description` - A textual description of the model (if available).
37
+
-`has_eula` - Whether the model requires end-user license agreement acceptance before use.
38
+
-`parameter_size_bits` - Size, in bits, of the model parameters.
39
+
-`size_bytes` - Total size, in bytes, of the model archive.
40
+
-`nodes_support` - List of supported node types and their quantization options. Each entry contains:
41
+
- `node_type_name` - The type of node supported.
42
+
- `quantization` - A list of supported quantization options, including:
43
+
- `quantization_bits` - Number of bits used for quantization (e.g., 8, 16).
44
+
- `allowed` - Whether this quantization is allowed.
45
+
- `max_context_size` - Maximum context length supported by this quantization.
-`model_name` - (Required) The model name to use for the deployment. Model names can be found in Console or using Scaleway's CLI (`scw inference model list`)
33
+
-`model_id` - (Required) The model id used for the deployment.
30
34
-`node_type` - (Required) The node type to use for the deployment. Node types can be found using Scaleway's CLI (`scw inference node-type list`)
31
35
-`name` - (Optional) The deployment name.
32
36
-`accept_eula` - (Optional) Some models (e.g Meta Llama) require end-user license agreements. Set `true` to accept.
The scaleway_inference_model resource allows you to upload and manage inference models in the Scaleway Inference ecosystem. Once registered, a model can be used in any scaleway_inference_deployment resource.
-`name` - (Required) The name of the model. This must be unique within the project.
47
+
-`url` - (Required) The HTTPS source URL from which the model will be downloaded. This is typically a Hugging Face repository URL (e.g., https://huggingface.co/agentica-org/DeepCoder-14B-Preview). The URL must be publicly accessible or require valid credentials via `secret`
48
+
-`secret` - (Optional, Sensitive) Authentication token used to pull the model from a private or gated URL (e.g., a Hugging Face access token with read permission).
49
+
-`region` - (Defaults to [provider](../index.md#region)`region`) The [region](../guides/regions_and_zones.md#regions) in which the deployment is created.
50
+
-`project_id` - (Defaults to [provider](../index.md#project_id)`project_id`) The ID of the project the deployment is associated with.
51
+
52
+
## Attributes Reference
53
+
54
+
In addition to all arguments above, the following attributes are exported:
55
+
56
+
-`id` - The unique identifier of the model.
57
+
-`tags` - Tags associated with the model.
58
+
-`status` - The current status of the model (e.g., ready, error, etc.).
59
+
-`description` - A textual description of the model (if available).
60
+
-`has_eula` - Whether the model requires end-user license agreement acceptance before use.
61
+
-`parameter_size_bits` - Size, in bits, of the model parameters.
62
+
-`size_bytes` - Total size, in bytes, of the model archive.
63
+
-`nodes_support` - List of supported node types and their quantization options. Each entry contains:
64
+
- `node_type_name` - The type of node supported.
65
+
- `quantization` - A list of supported quantization options, including:
66
+
- `quantization_bits` - Number of bits used for quantization (e.g., 8, 16).
67
+
- `allowed` - Whether this quantization is allowed.
68
+
- `max_context_size` - Maximum context length supported by this quantization.
69
+
70
+
## Import
71
+
72
+
Models can be imported using, `{region}/{id}`, as shown below:
0 commit comments