Ollama

Feature Available
Tools No
Multimodal No

We also support the Ollama inference server. Spin up a model with

ollama run mistral

Then specify the endpoints like so:

MODELS=`[
  {
    "name": "Ollama Mistral",
    "chatPromptTemplate": "<s>{{#each messages}}{

{#ifUser}

}[INST] {

{#if @first}

}{

{#if @root.preprompt}

}{{@root.preprompt}}\n{

{/if}

}{

{/if}

} {{content}} [/INST]{{/ifUser}}{

{#ifAssistant}

}{{content}}</s> {{/ifAssistant}}{{/each}}",
    "parameters": {
      "temperature": 0.1,
      "top_p": 0.95,
      "repetition_penalty": 1.2,
      "top_k": 50,
      "truncate": 3072,
      "max_new_tokens": 1024,
      "stop": ["</s>"]
    },
    "endpoints": [
      {
        "type": "ollama",
        "url" : "http://127.0.0.1:11434",
        "ollamaName" : "mistral"
      }
    ]
  }
]`
< > Update on GitHub