forked from AI_team/Philosophy-RAG-demo
Fix a small bug using local LLM's
This commit is contained in:
parent
450a00e6fe
commit
976ccd155f
@ -7,7 +7,7 @@ from langchain_core.embeddings import Embeddings
|
|||||||
from langchain_core.language_models.chat_models import BaseChatModel
|
from langchain_core.language_models.chat_models import BaseChatModel
|
||||||
from langchain_google_vertexai import VertexAIEmbeddings
|
from langchain_google_vertexai import VertexAIEmbeddings
|
||||||
from langchain_huggingface import HuggingFaceEmbeddings
|
from langchain_huggingface import HuggingFaceEmbeddings
|
||||||
from langchain_ollama import OllamaLLM
|
from langchain_ollama import ChatOllama
|
||||||
from langchain_openai import AzureChatOpenAI, AzureOpenAIEmbeddings
|
from langchain_openai import AzureChatOpenAI, AzureOpenAIEmbeddings
|
||||||
from langchain_openai import OpenAIEmbeddings
|
from langchain_openai import OpenAIEmbeddings
|
||||||
|
|
||||||
@ -41,7 +41,7 @@ def get_chat_model(backend_type: BackendType) -> BaseChatModel:
|
|||||||
return init_chat_model(model=os.environ["AWS_CHAT_MODEL"], model_provider="bedrock_converse")
|
return init_chat_model(model=os.environ["AWS_CHAT_MODEL"], model_provider="bedrock_converse")
|
||||||
|
|
||||||
if backend_type == BackendType.local:
|
if backend_type == BackendType.local:
|
||||||
return OllamaLLM(model=os.environ["LOCAL_CHAT_MODEL"])
|
return ChatOllama(model=os.environ["LOCAL_CHAT_MODEL"])
|
||||||
|
|
||||||
raise ValueError(f"Unknown backend type: {backend_type}")
|
raise ValueError(f"Unknown backend type: {backend_type}")
|
||||||
|
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user