Support for Ollama context window configuration
This commit is contained in:
parent
73b5e8832e
commit
c0705d1d9e
10 changed files with 170 additions and 54 deletions
|
|
@ -6,8 +6,8 @@ export const PROVIDER_INFO = {
|
|||
key: 'ollama',
|
||||
displayName: 'Ollama',
|
||||
};
|
||||
import { ChatOllama } from '@langchain/community/chat_models/ollama';
|
||||
import { OllamaEmbeddings } from '@langchain/community/embeddings/ollama';
|
||||
import { ChatOllama } from '@langchain/ollama';
|
||||
import { OllamaEmbeddings } from '@langchain/ollama';
|
||||
|
||||
export const loadOllamaChatModels = async () => {
|
||||
const ollamaApiEndpoint = getOllamaApiEndpoint();
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue