@@ -91,6 +91,7 @@ google/gemma-3-27b-it:bf16
91
91
```
92
92
| Attribute | Value |
93
93
| -----------| -------|
94
+ | Supports parallel tool calling | No |
94
95
| Supported image formats | PNG, JPEG, WEBP, and non-animated GIFs |
95
96
| Maximum image resolution (pixels) | 896x896 |
96
97
| Token dimension (pixels)| 56x56 |
@@ -103,6 +104,7 @@ This model was optimized to have a dense knowledge and faster tokens throughput
103
104
104
105
| Attribute | Value |
105
106
| -----------| -------|
107
+ | Supports parallel tool calling | Yes |
106
108
| Supported images formats | PNG, JPEG, WEBP, and non-animated GIFs |
107
109
| Maximum image resolution (pixels) | 1540x1540 |
108
110
| Token dimension (pixels)| 28x28 |
@@ -123,6 +125,7 @@ It can analyze images and offer insights from visual content alongside text.
123
125
124
126
| Attribute | Value |
125
127
| -----------| -------|
128
+ | Supports parallel tool calling | Yes |
126
129
| Supported images formats | PNG, JPEG, WEBP, and non-animated GIFs |
127
130
| Maximum image resolution (pixels) | 1024x1024 |
128
131
| Token dimension (pixels)| 16x16 |
@@ -148,6 +151,10 @@ allenai/molmo-72b-0924:fp8
148
151
Released December 6, 2024, Meta’s Llama 3.3 70b is a fine-tune of the [ Llama 3.1 70b] ( /managed-inference/reference-content/model-catalog/#llama-31-70b-instruct ) model.
149
152
This model is still text-only (text in/text out). However, Llama 3.3 was designed to approach the performance of Llama 3.1 405B on some applications.
150
153
154
+ | Attribute | Value |
155
+ | -----------| -------|
156
+ | Supports parallel tool calling | Yes |
157
+
151
158
#### Model name
152
159
```
153
160
meta/llama-3.3-70b-instruct:fp8
@@ -158,6 +165,10 @@ meta/llama-3.3-70b-instruct:bf16
158
165
Released July 23, 2024, Meta’s Llama 3.1 is an iteration of the open-access Llama family.
159
166
Llama 3.1 was designed to match the best proprietary models and outperform many of the available open source on common industry benchmarks.
160
167
168
+ | Attribute | Value |
169
+ | -----------| -------|
170
+ | Supports parallel tool calling | Yes |
171
+
161
172
#### Model names
162
173
```
163
174
meta/llama-3.1-70b-instruct:fp8
@@ -168,6 +179,10 @@ meta/llama-3.1-70b-instruct:bf16
168
179
Released July 23, 2024, Meta’s Llama 3.1 is an iteration of the open-access Llama family.
169
180
Llama 3.1 was designed to match the best proprietary models and outperform many of the available open source on common industry benchmarks.
170
181
182
+ | Attribute | Value |
183
+ | -----------| -------|
184
+ | Supports parallel tool calling | Yes |
185
+
171
186
#### Model names
172
187
```
173
188
meta/llama-3.1-8b-instruct:fp8
@@ -197,6 +212,10 @@ nvidia/llama-3.1-nemotron-70b-instruct:fp8
197
212
Released January 21, 2025, Deepseek’s R1 Distilled Llama 70B is a distilled version of the Llama model family based on Deepseek R1.
198
213
DeepSeek R1 Distill Llama 70B is designed to improve the performance of Llama models on reasoning use cases such as mathematics and coding tasks.
199
214
215
+ | Attribute | Value |
216
+ | -----------| -------|
217
+ | Supports parallel tool calling | No |
218
+
200
219
#### Model name
201
220
```
202
221
deepseek/deepseek-r1-distill-llama-70b:fp8
@@ -247,6 +266,10 @@ Mistral Nemo is a state-of-the-art transformer model of 12B parameters, built by
247
266
This model is open-weight and distributed under the Apache 2.0 license.
248
267
It was trained on a large proportion of multilingual and code data.
249
268
269
+ | Attribute | Value |
270
+ | -----------| -------|
271
+ | Supports parallel tool calling | Yes |
272
+
250
273
#### Model name
251
274
```
252
275
mistral/mistral-nemo-instruct-2407:fp8
@@ -302,6 +325,10 @@ kyutai/moshika-0.1-8b:fp8
302
325
Qwen2.5-coder is your intelligent programming assistant familiar with more than 40 programming languages.
303
326
With Qwen2.5-coder deployed at Scaleway, your company can benefit from code generation, AI-assisted code repair, and code reasoning.
304
327
328
+ | Attribute | Value |
329
+ | -----------| -------|
330
+ | Supports parallel tool calling | No |
331
+
305
332
#### Model name
306
333
```
307
334
qwen/qwen2.5-coder-32b-instruct:int8
0 commit comments