From a24c96368ca3982f77d22721f7dff94033520840 Mon Sep 17 00:00:00 2001 From: bracesproul Date: Tue, 28 Jan 2025 17:27:39 -0800 Subject: [PATCH] cr --- libs/langchain-openai/src/chat_models.ts | 1 - libs/langchain-openai/src/tests/chat_models.int.test.ts | 4 ++-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/libs/langchain-openai/src/chat_models.ts b/libs/langchain-openai/src/chat_models.ts index 21bce43decac..80deeff037c6 100644 --- a/libs/langchain-openai/src/chat_models.ts +++ b/libs/langchain-openai/src/chat_models.ts @@ -1451,7 +1451,6 @@ export class ChatOpenAI< const messagesMapped: OpenAICompletionParam[] = _convertMessagesToOpenAIParams(messages, this.model); - console.log("GENERATING!!", params.stream, this.streaming); if (params.stream) { const stream = this._streamResponseChunks(messages, options, runManager); const finalChunks: Record = {}; diff --git a/libs/langchain-openai/src/tests/chat_models.int.test.ts b/libs/langchain-openai/src/tests/chat_models.int.test.ts index 7abd7d3bb7a8..3f2e90d29c73 100644 --- a/libs/langchain-openai/src/tests/chat_models.int.test.ts +++ b/libs/langchain-openai/src/tests/chat_models.int.test.ts @@ -1286,7 +1286,7 @@ test.skip("Allow overriding", async () => { } }); -test.only("Streaming with o1 will yield at least one chunk with content", async () => { +test("Streaming with o1 will yield at least one chunk with content", async () => { const model = new ChatOpenAI({ model: "o1", disableStreaming: false, @@ -1307,5 +1307,5 @@ test.only("Streaming with o1 will yield at least one chunk with content", async } expect(content.length).toBeGreaterThan(10); - expect(numStreamChunks).toBe(1) + expect(numStreamChunks).toBe(1); });