Files
anything-llm/server/utils/agents/aibitat/providers/deepseek.js
Timothy Carambat 7ead0449f2 Fix deepseek v4 reasoning inject thoughts (#5527)
* Fix deepseek v4 injectReasoning error
resolves #5509

* comment
2026-04-24 15:40:05 -07:00

224 lines
5.9 KiB
JavaScript

const OpenAI = require("openai");
const Provider = require("./ai-provider.js");
const InheritMultiple = require("./helpers/classes.js");
const UnTooled = require("./helpers/untooled.js");
const { tooledStream, tooledComplete } = require("./helpers/tooled.js");
const { RetryError } = require("../error.js");
const { toValidNumber } = require("../../../http/index.js");
class DeepSeekProvider extends InheritMultiple([Provider, UnTooled]) {
model;
constructor(config = {}) {
super();
const { model = "deepseek-chat" } = config;
const client = new OpenAI({
baseURL: "https://api.deepseek.com/v1",
apiKey: process.env.DEEPSEEK_API_KEY ?? null,
maxRetries: 3,
});
this._client = client;
this.model = model;
this.verbose = true;
this.maxTokens = process.env.DEEPSEEK_MAX_TOKENS
? toValidNumber(process.env.DEEPSEEK_MAX_TOKENS, 1024)
: 1024;
}
get client() {
return this._client;
}
get supportsAgentStreaming() {
return true;
}
/**
* All current DeepSeek models (deepseek-chat and deepseek-reasoner)
* support native OpenAI-compatible tool calling.
* @returns {boolean}
*/
supportsNativeToolCalling() {
return true;
}
/**
* DeepSeek models do not support vision/image inputs.
* Strip attachments from messages to prevent API errors.
* @param {Object} message - Message with potential attachments
* @returns {Object} Message without attachments
*/
formatMessageWithAttachments(message) {
const { attachments: _, ...rest } = message;
return rest;
}
/**
* Check if the model is a thinking model
* because we need to inject reasoning content into the messages
* or else the DeepSeek API will return an error for specific models.
* There is no official way to predetect if a model will require this
* so we have to hardcode the list of thinking models.
* @returns {boolean}
*/
get #isThinkingModel() {
return [
"deepseek-reasoner",
"deepseek-v4-flash",
"deepseek-v4-pro",
].includes(this.model);
}
get #tooledOptions() {
return {
provider: this,
...(this.#isThinkingModel ? { injectReasoningContent: true } : {}),
};
}
async #handleFunctionCallChat({ messages = [] }) {
return await this.client.chat.completions
.create({
model: this.model,
messages,
max_tokens: this.maxTokens,
})
.then((result) => {
if (!result.hasOwnProperty("choices"))
throw new Error("DeepSeek chat: No results!");
if (result.choices.length === 0)
throw new Error("DeepSeek chat: No results length!");
return result.choices[0].message.content;
})
.catch((_) => {
return null;
});
}
async #handleFunctionCallStream({ messages = [] }) {
return await this.client.chat.completions.create({
model: this.model,
stream: true,
messages,
});
}
/**
* Strip attachments from all messages since DeepSeek doesn't support vision.
* @param {Array} messages - Array of messages
* @returns {Array} Messages with attachments removed
*/
#stripAttachments(messages) {
let hasAttachments = false;
const stripped = messages.map((msg) => {
if (msg.attachments && msg.attachments.length > 0) {
hasAttachments = true;
const { attachments: _, ...rest } = msg;
return rest;
}
return msg;
});
if (hasAttachments) {
this.providerLog(
"DeepSeek does not support vision - stripped image attachments from messages."
);
}
return stripped;
}
async stream(messages, functions = [], eventHandler = null) {
const useNative = functions.length > 0 && this.supportsNativeToolCalling();
const cleanedMessages = this.#stripAttachments(messages);
if (!useNative) {
return await UnTooled.prototype.stream.call(
this,
cleanedMessages,
functions,
this.#handleFunctionCallStream.bind(this),
eventHandler
);
}
this.providerLog(
"Provider.stream (tooled) - will process this chat completion."
);
try {
return await tooledStream(
this.client,
this.model,
cleanedMessages,
functions,
eventHandler,
this.#tooledOptions
);
} catch (error) {
console.error(error.message, error);
if (error instanceof OpenAI.AuthenticationError) throw error;
if (
error instanceof OpenAI.RateLimitError ||
error instanceof OpenAI.InternalServerError ||
error instanceof OpenAI.APIError
) {
throw new RetryError(error.message);
}
throw error;
}
}
async complete(messages, functions = []) {
const useNative = functions.length > 0 && this.supportsNativeToolCalling();
const cleanedMessages = this.#stripAttachments(messages);
if (!useNative) {
return await UnTooled.prototype.complete.call(
this,
cleanedMessages,
functions,
this.#handleFunctionCallChat.bind(this)
);
}
try {
const result = await tooledComplete(
this.client,
this.model,
cleanedMessages,
functions,
this.getCost.bind(this),
this.#tooledOptions
);
if (result.retryWithError) {
return this.complete([...messages, result.retryWithError], functions);
}
return result;
} catch (error) {
if (error instanceof OpenAI.AuthenticationError) throw error;
if (
error instanceof OpenAI.RateLimitError ||
error instanceof OpenAI.InternalServerError ||
error instanceof OpenAI.APIError
) {
throw new RetryError(error.message);
}
throw error;
}
}
/**
* Get the cost of the completion.
*
* @param _usage The completion to get the cost for.
* @returns The cost of the completion.
*/
getCost(_usage) {
return 0;
}
}
module.exports = DeepSeekProvider;