Skip to content

Commit

Permalink
Format
Browse files Browse the repository at this point in the history
  • Loading branch information
jacoblee93 committed Jan 27, 2025
1 parent 2fde952 commit f05ca0d
Show file tree
Hide file tree
Showing 3 changed files with 11 additions and 15 deletions.
10 changes: 0 additions & 10 deletions docs/core_docs/docs/integrations/chat/deepseek.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -193,16 +193,6 @@
")"
]
},
{
"cell_type": "markdown",
"id": "d1ee55bc-ffc8-4cfa-801c-993953a08cfd",
"metadata": {},
"source": [
"## TODO: Any functionality specific to this model provider\n",
"\n",
"E.g. creating/using finetuned models via this provider. Delete if not relevant."
]
},
{
"cell_type": "markdown",
"id": "3a5bb5ca-c3ae-4a58-be67-2cd18574b9a3",
Expand Down
12 changes: 8 additions & 4 deletions libs/langchain-deepseek/src/chat_models.ts
Original file line number Diff line number Diff line change
@@ -1,5 +1,10 @@
import { getEnvironmentVariable } from "@langchain/core/utils/env";
import { ChatOpenAI, ChatOpenAICallOptions, ChatOpenAIFields, OpenAIClient } from "@langchain/openai";
import {
ChatOpenAI,
ChatOpenAICallOptions,
ChatOpenAIFields,
OpenAIClient,
} from "@langchain/openai";

export interface ChatDeepseekCallOptions extends ChatOpenAICallOptions {
headers?: Record<string, string>;
Expand Down Expand Up @@ -317,8 +322,7 @@ export class ChatDeepseek extends ChatOpenAI<ChatDeepseekCallOptions> {
rawResponse,
defaultRole
);
messageChunk.additional_kwargs.reasoning_content =
delta.reasoning_content;
messageChunk.additional_kwargs.reasoning_content = delta.reasoning_content;
return messageChunk;
}

Expand All @@ -336,4 +340,4 @@ export class ChatDeepseek extends ChatOpenAI<ChatDeepseekCallOptions> {
).reasoning_content;
return langChainMessage;
}
}
}
4 changes: 3 additions & 1 deletion libs/langchain-deepseek/src/tests/chat_models.int.test.ts
Original file line number Diff line number Diff line change
Expand Up @@ -11,5 +11,7 @@ test("Can send deepseek-reasoner requests", async () => {
// Models also accept a list of chat messages or a formatted prompt
const result = await llm.invoke(input);
console.log(result);
expect((result.additional_kwargs.reasoning_content as any).length).toBeGreaterThan(10);
expect(
(result.additional_kwargs.reasoning_content as any).length
).toBeGreaterThan(10);
});

0 comments on commit f05ca0d

Please sign in to comment.