Support for Ollama context window configuration
This commit is contained in:
parent
73b5e8832e
commit
c0705d1d9e
10 changed files with 170 additions and 54 deletions
|
|
@ -6,8 +6,8 @@ export const PROVIDER_INFO = {
|
|||
key: 'ollama',
|
||||
displayName: 'Ollama',
|
||||
};
|
||||
import { ChatOllama } from '@langchain/community/chat_models/ollama';
|
||||
import { OllamaEmbeddings } from '@langchain/community/embeddings/ollama';
|
||||
import { ChatOllama } from '@langchain/ollama';
|
||||
import { OllamaEmbeddings } from '@langchain/ollama';
|
||||
|
||||
export const loadOllamaChatModels = async () => {
|
||||
const ollamaApiEndpoint = getOllamaApiEndpoint();
|
||||
|
|
|
|||
|
|
@ -239,7 +239,6 @@ class MetaSearchAgent implements MetaSearchAgentType {
|
|||
optimizationMode: 'speed' | 'balanced' | 'quality',
|
||||
systemInstructions: string,
|
||||
) {
|
||||
console.log("optimization mode", optimizationMode);
|
||||
return RunnableSequence.from([
|
||||
RunnableMap.from({
|
||||
systemInstructions: () => systemInstructions,
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue