diff --git a/flowsettings.py b/flowsettings.py index e4b95b9..0e149c3 100644 --- a/flowsettings.py +++ b/flowsettings.py @@ -176,14 +176,14 @@ if config("LOCAL_MODEL", default=""): } # additional LLM configurations -KH_LLMS["claude"] = { - "spec": { - "__type__": "kotaemon.llms.chats.LCAnthropicChat", - "model_name": "claude-3-5-sonnet-20240620", - "api_key": "your-key", - }, - "default": False, -} +# KH_LLMS["claude"] = { +# "spec": { +# "__type__": "kotaemon.llms.chats.LCAnthropicChat", +# "model_name": "claude-3-5-sonnet-20240620", +# "api_key": "your-key", +# }, +# "default": False, +# } # KH_LLMS["gemini"] = { # "spec": { # "__type__": "kotaemon.llms.chats.LCGeminiChat", diff --git a/libs/kotaemon/pyproject.toml b/libs/kotaemon/pyproject.toml index 37ce1dd..e3299b7 100644 --- a/libs/kotaemon/pyproject.toml +++ b/libs/kotaemon/pyproject.toml @@ -30,7 +30,6 @@ dependencies = [ "langchain>=0.1.16,<0.2.0", "langchain-community>=0.0.34,<0.1.0", "langchain-openai>=0.1.4,<0.2.0", - "langchain-anthropic", "llama-hub>=0.0.79,<0.1.0", "llama-index>=0.10.40,<0.11.0", "llama-index-vector-stores-chroma>=0.1.9",