You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This PR updates models to v0.2.39 compiled with
mlc-ai/binary-mlc-llm-libs#123
The main change is the new MLC-LLM runtime, which supports grammar (i.e.
json mode) for Llama3.
- Hence we now read in field `tokenizer_info` (or deprecated
`token_table_postproc_method`) from `mlc-chat-config.json` when post
processing token table for Grammar
- If neither is available, we use the default `byte_fallback`
New prebuilt models introduced:
- Phi3-mini-4k
- Hermes-2-Pro-Llama-3-8B
- Qwen1.5-1.8B
- StableLM-2-zephyr_1.6B
Updates on examples:
- json-mode and json-schema now use Llama3 to demonstrate
- Function calling inside json-schema now uses `Hermes-2-Pro-Llama-3-8B`
instead of `Hermes-2-Pro-Mistral`
@@ -184,12 +184,12 @@ async function functionCallingExample() {
184
184
{
185
185
role: "system",
186
186
content: `You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: <tools> ${JSON.stringify(
187
-
tools
187
+
tools,
188
188
)} </tools>. Do not stop calling functions until the task has been accomplished or you've reached max iteration of 10.
189
189
Calling multiple functions at once can overload the system and increase cost so call one function at a time please.
190
190
If you plan to continue with analysis, always call another function.
191
191
Return a valid json object (using double quotes) in the following schema: ${JSON.stringify(
0 commit comments