47 lines
1.9 KiB
Python
47 lines
1.9 KiB
Python
|
|
from app.core.config import settings
|
||
|
|
from langchain_openai import OpenAIEmbeddings
|
||
|
|
from langchain_ollama import OllamaEmbeddings
|
||
|
|
# If you plan on adding other embeddings, import them here
|
||
|
|
# from some_other_module import AnotherEmbeddingClass
|
||
|
|
|
||
|
|
|
||
|
|
class EmbeddingsFactory:
|
||
|
|
@staticmethod
|
||
|
|
def create():
|
||
|
|
"""
|
||
|
|
Factory method to create an embeddings instance based on .env config.
|
||
|
|
"""
|
||
|
|
# Suppose your .env has a value like EMBEDDINGS_PROVIDER=openai
|
||
|
|
embeddings_provider = settings.EMBEDDINGS_PROVIDER.lower()
|
||
|
|
|
||
|
|
if embeddings_provider == "openai":
|
||
|
|
return OpenAIEmbeddings(
|
||
|
|
openai_api_key=settings.OPENAI_API_KEY,
|
||
|
|
openai_api_base=settings.OPENAI_API_BASE,
|
||
|
|
model=settings.OPENAI_EMBEDDINGS_MODEL
|
||
|
|
)
|
||
|
|
elif embeddings_provider == "dashscope":
|
||
|
|
return OpenAIEmbeddings(
|
||
|
|
openai_api_key=settings.DASH_SCOPE_API_KEY,
|
||
|
|
openai_api_base=settings.DASH_SCOPE_API_BASE,
|
||
|
|
model=settings.DASH_SCOPE_EMBEDDINGS_MODEL,
|
||
|
|
# DashScope OpenAI-compatible embedding expects string input,
|
||
|
|
# while LangChain's len-safe path may send token ids.
|
||
|
|
check_embedding_ctx_length=False,
|
||
|
|
tiktoken_enabled=False,
|
||
|
|
skip_empty=True,
|
||
|
|
# DashScope embedding API supports at most 10 inputs per batch.
|
||
|
|
chunk_size=10,
|
||
|
|
)
|
||
|
|
elif embeddings_provider == "ollama":
|
||
|
|
return OllamaEmbeddings(
|
||
|
|
model=settings.OLLAMA_EMBEDDINGS_MODEL,
|
||
|
|
base_url=settings.OLLAMA_API_BASE
|
||
|
|
)
|
||
|
|
|
||
|
|
# Extend with other providers:
|
||
|
|
# elif embeddings_provider == "another_provider":
|
||
|
|
# return AnotherEmbeddingClass(...)
|
||
|
|
else:
|
||
|
|
raise ValueError(f"Unsupported embeddings provider: {embeddings_provider}")
|