mirror of
https://github.com/Mintplex-Labs/anything-llm
synced 2026-04-25 17:15:37 +02:00
* feat: add new model provider: Novita AI * feat: finished novita AI * fix: code lint * remove unneeded logging * add back log for novita stream not self closing * Clarify ENV vars for LLM/embedder seperation for future Patch ENV check for workspace/agent provider --------- Co-authored-by: Jason <ggbbddjm@gmail.com> Co-authored-by: shatfield4 <seanhatfield5@gmail.com>
370 lines
13 KiB
JavaScript
370 lines
13 KiB
JavaScript
import { MagnifyingGlass } from "@phosphor-icons/react";
|
|
import { useEffect, useState, useRef } from "react";
|
|
import OpenAiLogo from "@/media/llmprovider/openai.png";
|
|
import GenericOpenAiLogo from "@/media/llmprovider/generic-openai.png";
|
|
import AzureOpenAiLogo from "@/media/llmprovider/azure.png";
|
|
import AnthropicLogo from "@/media/llmprovider/anthropic.png";
|
|
import GeminiLogo from "@/media/llmprovider/gemini.png";
|
|
import OllamaLogo from "@/media/llmprovider/ollama.png";
|
|
import LMStudioLogo from "@/media/llmprovider/lmstudio.png";
|
|
import LocalAiLogo from "@/media/llmprovider/localai.png";
|
|
import TogetherAILogo from "@/media/llmprovider/togetherai.png";
|
|
import FireworksAILogo from "@/media/llmprovider/fireworksai.jpeg";
|
|
import AnythingLLMIcon from "@/media/logo/anything-llm-icon.png";
|
|
import MistralLogo from "@/media/llmprovider/mistral.jpeg";
|
|
import HuggingFaceLogo from "@/media/llmprovider/huggingface.png";
|
|
import PerplexityLogo from "@/media/llmprovider/perplexity.png";
|
|
import OpenRouterLogo from "@/media/llmprovider/openrouter.jpeg";
|
|
import GroqLogo from "@/media/llmprovider/groq.png";
|
|
import KoboldCPPLogo from "@/media/llmprovider/koboldcpp.png";
|
|
import TextGenWebUILogo from "@/media/llmprovider/text-generation-webui.png";
|
|
import LiteLLMLogo from "@/media/llmprovider/litellm.png";
|
|
import AWSBedrockLogo from "@/media/llmprovider/bedrock.png";
|
|
import DeepSeekLogo from "@/media/llmprovider/deepseek.png";
|
|
import APIPieLogo from "@/media/llmprovider/apipie.png";
|
|
import NovitaLogo from "@/media/llmprovider/novita.png";
|
|
import XAILogo from "@/media/llmprovider/xai.png";
|
|
|
|
import CohereLogo from "@/media/llmprovider/cohere.png";
|
|
import OpenAiOptions from "@/components/LLMSelection/OpenAiOptions";
|
|
import GenericOpenAiOptions from "@/components/LLMSelection/GenericOpenAiOptions";
|
|
import AzureAiOptions from "@/components/LLMSelection/AzureAiOptions";
|
|
import AnthropicAiOptions from "@/components/LLMSelection/AnthropicAiOptions";
|
|
import LMStudioOptions from "@/components/LLMSelection/LMStudioOptions";
|
|
import LocalAiOptions from "@/components/LLMSelection/LocalAiOptions";
|
|
import NativeLLMOptions from "@/components/LLMSelection/NativeLLMOptions";
|
|
import GeminiLLMOptions from "@/components/LLMSelection/GeminiLLMOptions";
|
|
import OllamaLLMOptions from "@/components/LLMSelection/OllamaLLMOptions";
|
|
import MistralOptions from "@/components/LLMSelection/MistralOptions";
|
|
import HuggingFaceOptions from "@/components/LLMSelection/HuggingFaceOptions";
|
|
import TogetherAiOptions from "@/components/LLMSelection/TogetherAiOptions";
|
|
import FireworksAiOptions from "@/components/LLMSelection/FireworksAiOptions";
|
|
import PerplexityOptions from "@/components/LLMSelection/PerplexityOptions";
|
|
import OpenRouterOptions from "@/components/LLMSelection/OpenRouterOptions";
|
|
import GroqAiOptions from "@/components/LLMSelection/GroqAiOptions";
|
|
import CohereAiOptions from "@/components/LLMSelection/CohereAiOptions";
|
|
import KoboldCPPOptions from "@/components/LLMSelection/KoboldCPPOptions";
|
|
import TextGenWebUIOptions from "@/components/LLMSelection/TextGenWebUIOptions";
|
|
import LiteLLMOptions from "@/components/LLMSelection/LiteLLMOptions";
|
|
import AWSBedrockLLMOptions from "@/components/LLMSelection/AwsBedrockLLMOptions";
|
|
import DeepSeekOptions from "@/components/LLMSelection/DeepSeekOptions";
|
|
import ApiPieLLMOptions from "@/components/LLMSelection/ApiPieOptions";
|
|
import NovitaLLMOptions from "@/components/LLMSelection/NovitaLLMOptions";
|
|
import XAILLMOptions from "@/components/LLMSelection/XAiLLMOptions";
|
|
|
|
import LLMItem from "@/components/LLMSelection/LLMItem";
|
|
import System from "@/models/system";
|
|
import paths from "@/utils/paths";
|
|
import showToast from "@/utils/toast";
|
|
import { useNavigate } from "react-router-dom";
|
|
|
|
const TITLE = "LLM Preference";
|
|
const DESCRIPTION =
|
|
"AnythingLLM can work with many LLM providers. This will be the service which handles chatting.";
|
|
|
|
const LLMS = [
|
|
{
|
|
name: "OpenAI",
|
|
value: "openai",
|
|
logo: OpenAiLogo,
|
|
options: (settings) => <OpenAiOptions settings={settings} />,
|
|
description: "The standard option for most non-commercial use.",
|
|
},
|
|
{
|
|
name: "Azure OpenAI",
|
|
value: "azure",
|
|
logo: AzureOpenAiLogo,
|
|
options: (settings) => <AzureAiOptions settings={settings} />,
|
|
description: "The enterprise option of OpenAI hosted on Azure services.",
|
|
},
|
|
{
|
|
name: "Anthropic",
|
|
value: "anthropic",
|
|
logo: AnthropicLogo,
|
|
options: (settings) => <AnthropicAiOptions settings={settings} />,
|
|
description: "A friendly AI Assistant hosted by Anthropic.",
|
|
},
|
|
{
|
|
name: "Gemini",
|
|
value: "gemini",
|
|
logo: GeminiLogo,
|
|
options: (settings) => <GeminiLLMOptions settings={settings} />,
|
|
description: "Google's largest and most capable AI model",
|
|
},
|
|
{
|
|
name: "HuggingFace",
|
|
value: "huggingface",
|
|
logo: HuggingFaceLogo,
|
|
options: (settings) => <HuggingFaceOptions settings={settings} />,
|
|
description:
|
|
"Access 150,000+ open-source LLMs and the world's AI community",
|
|
},
|
|
{
|
|
name: "Ollama",
|
|
value: "ollama",
|
|
logo: OllamaLogo,
|
|
options: (settings) => <OllamaLLMOptions settings={settings} />,
|
|
description: "Run LLMs locally on your own machine.",
|
|
},
|
|
{
|
|
name: "Novita AI",
|
|
value: "novita",
|
|
logo: NovitaLogo,
|
|
options: (settings) => <NovitaLLMOptions settings={settings} />,
|
|
description:
|
|
"Reliable, Scalable, and Cost-Effective for LLMs from Novita AI",
|
|
},
|
|
{
|
|
name: "LM Studio",
|
|
value: "lmstudio",
|
|
logo: LMStudioLogo,
|
|
options: (settings) => <LMStudioOptions settings={settings} />,
|
|
description:
|
|
"Discover, download, and run thousands of cutting edge LLMs in a few clicks.",
|
|
},
|
|
{
|
|
name: "Local AI",
|
|
value: "localai",
|
|
logo: LocalAiLogo,
|
|
options: (settings) => <LocalAiOptions settings={settings} />,
|
|
description: "Run LLMs locally on your own machine.",
|
|
},
|
|
{
|
|
name: "KoboldCPP",
|
|
value: "koboldcpp",
|
|
logo: KoboldCPPLogo,
|
|
options: (settings) => <KoboldCPPOptions settings={settings} />,
|
|
description: "Run local LLMs using koboldcpp.",
|
|
},
|
|
{
|
|
name: "Oobabooga Web UI",
|
|
value: "textgenwebui",
|
|
logo: TextGenWebUILogo,
|
|
options: (settings) => <TextGenWebUIOptions settings={settings} />,
|
|
description: "Run local LLMs using Oobabooga's Text Generation Web UI.",
|
|
},
|
|
{
|
|
name: "Together AI",
|
|
value: "togetherai",
|
|
logo: TogetherAILogo,
|
|
options: (settings) => <TogetherAiOptions settings={settings} />,
|
|
description: "Run open source models from Together AI.",
|
|
},
|
|
{
|
|
name: "Fireworks AI",
|
|
value: "fireworksai",
|
|
logo: FireworksAILogo,
|
|
options: (settings) => <FireworksAiOptions settings={settings} />,
|
|
description:
|
|
"The fastest and most efficient inference engine to build production-ready, compound AI systems.",
|
|
},
|
|
{
|
|
name: "Mistral",
|
|
value: "mistral",
|
|
logo: MistralLogo,
|
|
options: (settings) => <MistralOptions settings={settings} />,
|
|
description: "Run open source models from Mistral AI.",
|
|
},
|
|
{
|
|
name: "Perplexity AI",
|
|
value: "perplexity",
|
|
logo: PerplexityLogo,
|
|
options: (settings) => <PerplexityOptions settings={settings} />,
|
|
description:
|
|
"Run powerful and internet-connected models hosted by Perplexity AI.",
|
|
},
|
|
{
|
|
name: "OpenRouter",
|
|
value: "openrouter",
|
|
logo: OpenRouterLogo,
|
|
options: (settings) => <OpenRouterOptions settings={settings} />,
|
|
description: "A unified interface for LLMs.",
|
|
},
|
|
{
|
|
name: "Groq",
|
|
value: "groq",
|
|
logo: GroqLogo,
|
|
options: (settings) => <GroqAiOptions settings={settings} />,
|
|
description:
|
|
"The fastest LLM inferencing available for real-time AI applications.",
|
|
},
|
|
{
|
|
name: "Cohere",
|
|
value: "cohere",
|
|
logo: CohereLogo,
|
|
options: (settings) => <CohereAiOptions settings={settings} />,
|
|
description: "Run Cohere's powerful Command models.",
|
|
},
|
|
{
|
|
name: "LiteLLM",
|
|
value: "litellm",
|
|
logo: LiteLLMLogo,
|
|
options: (settings) => <LiteLLMOptions settings={settings} />,
|
|
description: "Run LiteLLM's OpenAI compatible proxy for various LLMs.",
|
|
},
|
|
{
|
|
name: "DeepSeek",
|
|
value: "deepseek",
|
|
logo: DeepSeekLogo,
|
|
options: (settings) => <DeepSeekOptions settings={settings} />,
|
|
description: "Run DeepSeek's powerful LLMs.",
|
|
},
|
|
{
|
|
name: "APIpie",
|
|
value: "apipie",
|
|
logo: APIPieLogo,
|
|
options: (settings) => <ApiPieLLMOptions settings={settings} />,
|
|
description: "A unified API of AI services from leading providers",
|
|
},
|
|
{
|
|
name: "Generic OpenAI",
|
|
value: "generic-openai",
|
|
logo: GenericOpenAiLogo,
|
|
options: (settings) => <GenericOpenAiOptions settings={settings} />,
|
|
description:
|
|
"Connect to any OpenAi-compatible service via a custom configuration",
|
|
},
|
|
{
|
|
name: "AWS Bedrock",
|
|
value: "bedrock",
|
|
logo: AWSBedrockLogo,
|
|
options: (settings) => <AWSBedrockLLMOptions settings={settings} />,
|
|
description: "Run powerful foundation models privately with AWS Bedrock.",
|
|
},
|
|
{
|
|
name: "xAI",
|
|
value: "xai",
|
|
logo: XAILogo,
|
|
options: (settings) => <XAILLMOptions settings={settings} />,
|
|
description: "Run xAI's powerful LLMs like Grok-2 and more.",
|
|
},
|
|
{
|
|
name: "Native",
|
|
value: "native",
|
|
logo: AnythingLLMIcon,
|
|
options: (settings) => <NativeLLMOptions settings={settings} />,
|
|
description:
|
|
"Use a downloaded custom Llama model for chatting on this AnythingLLM instance.",
|
|
},
|
|
];
|
|
|
|
export default function LLMPreference({
|
|
setHeader,
|
|
setForwardBtn,
|
|
setBackBtn,
|
|
}) {
|
|
const [searchQuery, setSearchQuery] = useState("");
|
|
const [filteredLLMs, setFilteredLLMs] = useState([]);
|
|
const [selectedLLM, setSelectedLLM] = useState(null);
|
|
const [settings, setSettings] = useState(null);
|
|
const formRef = useRef(null);
|
|
const hiddenSubmitButtonRef = useRef(null);
|
|
const isHosted = window.location.hostname.includes("useanything.com");
|
|
const navigate = useNavigate();
|
|
|
|
useEffect(() => {
|
|
async function fetchKeys() {
|
|
const _settings = await System.keys();
|
|
setSettings(_settings);
|
|
setSelectedLLM(_settings?.LLMProvider || "openai");
|
|
}
|
|
fetchKeys();
|
|
}, []);
|
|
|
|
function handleForward() {
|
|
if (hiddenSubmitButtonRef.current) {
|
|
hiddenSubmitButtonRef.current.click();
|
|
}
|
|
}
|
|
|
|
function handleBack() {
|
|
navigate(paths.onboarding.home());
|
|
}
|
|
|
|
const handleSubmit = async (e) => {
|
|
e.preventDefault();
|
|
const form = e.target;
|
|
const data = {};
|
|
const formData = new FormData(form);
|
|
data.LLMProvider = selectedLLM;
|
|
// Default to AnythingLLM embedder and LanceDB
|
|
data.EmbeddingEngine = "native";
|
|
data.VectorDB = "lancedb";
|
|
for (var [key, value] of formData.entries()) data[key] = value;
|
|
|
|
const { error } = await System.updateSystem(data);
|
|
if (error) {
|
|
showToast(`Failed to save LLM settings: ${error}`, "error");
|
|
return;
|
|
}
|
|
navigate(paths.onboarding.userSetup());
|
|
};
|
|
|
|
useEffect(() => {
|
|
setHeader({ title: TITLE, description: DESCRIPTION });
|
|
setForwardBtn({ showing: true, disabled: false, onClick: handleForward });
|
|
setBackBtn({ showing: true, disabled: false, onClick: handleBack });
|
|
}, []);
|
|
|
|
useEffect(() => {
|
|
const filtered = LLMS.filter((llm) =>
|
|
llm.name.toLowerCase().includes(searchQuery.toLowerCase())
|
|
);
|
|
setFilteredLLMs(filtered);
|
|
}, [searchQuery, selectedLLM]);
|
|
|
|
return (
|
|
<div>
|
|
<form ref={formRef} onSubmit={handleSubmit} className="w-full">
|
|
<div className="w-full relative border-slate-300/40 shadow border-2 rounded-lg text-white">
|
|
<div className="w-full p-4 absolute top-0 rounded-t-lg backdrop-blur-sm">
|
|
<div className="w-full flex items-center sticky top-0">
|
|
<MagnifyingGlass
|
|
size={16}
|
|
weight="bold"
|
|
className="absolute left-4 z-30 text-white"
|
|
/>
|
|
<input
|
|
type="text"
|
|
placeholder="Search LLM providers"
|
|
className="bg-zinc-600 z-20 pl-10 h-[38px] rounded-full w-full px-4 py-1 text-sm border-2 border-slate-300/40 outline-none focus:outline-primary-button active:outline-primary-button outline-none text-white"
|
|
onChange={(e) => setSearchQuery(e.target.value)}
|
|
autoComplete="off"
|
|
onKeyDown={(e) => {
|
|
if (e.key === "Enter") e.preventDefault();
|
|
}}
|
|
/>
|
|
</div>
|
|
</div>
|
|
<div className="px-4 pt-[70px] flex flex-col gap-y-1 max-h-[390px] overflow-y-auto no-scroll pb-4">
|
|
{filteredLLMs.map((llm) => {
|
|
if (llm.value === "native" && isHosted) return null;
|
|
return (
|
|
<LLMItem
|
|
key={llm.name}
|
|
name={llm.name}
|
|
value={llm.value}
|
|
image={llm.logo}
|
|
description={llm.description}
|
|
checked={selectedLLM === llm.value}
|
|
onClick={() => setSelectedLLM(llm.value)}
|
|
/>
|
|
);
|
|
})}
|
|
</div>
|
|
</div>
|
|
<div className="mt-4 flex flex-col gap-y-1">
|
|
{selectedLLM &&
|
|
LLMS.find((llm) => llm.value === selectedLLM)?.options(settings)}
|
|
</div>
|
|
<button
|
|
type="submit"
|
|
ref={hiddenSubmitButtonRef}
|
|
hidden
|
|
aria-hidden="true"
|
|
></button>
|
|
</form>
|
|
</div>
|
|
);
|
|
}
|