diff --git a/libs/langchain-openai/src/llms.ts b/libs/langchain-openai/src/llms.ts index e29f1a3bcdb4..af83f681c319 100644 --- a/libs/langchain-openai/src/llms.ts +++ b/libs/langchain-openai/src/llms.ts @@ -157,7 +157,6 @@ export class OpenAI getEnvironmentVariable("OPENAI_ORGANIZATION"); this.model = fields?.model ?? fields?.modelName ?? this.model; - console.log(this.model); if ( (this.model?.startsWith("gpt-3.5-turbo") || this.model?.startsWith("gpt-4") || diff --git a/libs/langchain-openai/src/tests/azure/chat_models.int.test.ts b/libs/langchain-openai/src/tests/azure/chat_models.int.test.ts index 1cbbe902ca9a..b4f94ff5aeb7 100644 --- a/libs/langchain-openai/src/tests/azure/chat_models.int.test.ts +++ b/libs/langchain-openai/src/tests/azure/chat_models.int.test.ts @@ -28,24 +28,6 @@ import { AzureChatOpenAI } from "../../azure/chat_models.js"; // Save the original value of the 'LANGCHAIN_CALLBACKS_BACKGROUND' environment variable const originalBackground = process.env.LANGCHAIN_CALLBACKS_BACKGROUND; -beforeAll(() => { - if (!process.env.AZURE_OPENAI_API_KEY) { - process.env.AZURE_OPENAI_API_KEY = process.env.TEST_AZURE_OPENAI_API_KEY; - } - if (!process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME) { - process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME = - process.env.TEST_AZURE_OPENAI_API_DEPLOYMENT_NAME; - } - if (!process.env.AZURE_OPENAI_BASE_PATH) { - process.env.AZURE_OPENAI_BASE_PATH = - process.env.TEST_AZURE_OPENAI_BASE_PATH; - } - if (!process.env.AZURE_OPENAI_API_VERSION) { - process.env.AZURE_OPENAI_API_VERSION = - process.env.TEST_AZURE_OPENAI_API_VERSION; - } -}); - test("Test Azure ChatOpenAI call method", async () => { const chat = new AzureChatOpenAI({ modelName: "gpt-3.5-turbo", diff --git a/libs/langchain-openai/src/tests/azure/chat_models.standard.int.test.ts b/libs/langchain-openai/src/tests/azure/chat_models.standard.int.test.ts index 64052685d6c2..718263172aed 100644 --- a/libs/langchain-openai/src/tests/azure/chat_models.standard.int.test.ts +++ b/libs/langchain-openai/src/tests/azure/chat_models.standard.int.test.ts @@ -1,5 +1,5 @@ /* eslint-disable no-process-env */ -import { test, expect, beforeAll, afterAll } from "@jest/globals"; +import { test, expect, afterAll } from "@jest/globals"; import { ChatModelIntegrationTests } from "@langchain/standard-tests"; import { AIMessageChunk } from "@langchain/core/messages"; import { AzureChatOpenAI } from "../../azure/chat_models.js"; @@ -7,30 +7,6 @@ import { ChatOpenAICallOptions } from "../../chat_models.js"; let openAIAPIKey: string | undefined; -beforeAll(() => { - if (process.env.OPENAI_API_KEY) { - openAIAPIKey = process.env.OPENAI_API_KEY; - process.env.OPENAI_API_KEY = ""; - } - - if (!process.env.AZURE_OPENAI_API_KEY) { - process.env.AZURE_OPENAI_API_KEY = process.env.TEST_AZURE_OPENAI_API_KEY; - } - if (!process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME) { - process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME = - process.env.TEST_AZURE_OPENAI_API_DEPLOYMENT_NAME ?? - process.env.AZURE_OPENAI_CHAT_DEPLOYMENT_NAME; - } - if (!process.env.AZURE_OPENAI_BASE_PATH) { - process.env.AZURE_OPENAI_BASE_PATH = - process.env.TEST_AZURE_OPENAI_BASE_PATH; - } - if (!process.env.AZURE_OPENAI_API_VERSION) { - process.env.AZURE_OPENAI_API_VERSION = - process.env.TEST_AZURE_OPENAI_API_VERSION; - } -}); - afterAll(() => { if (openAIAPIKey) { process.env.OPENAI_API_KEY = openAIAPIKey; diff --git a/libs/langchain-openai/src/tests/azure/embeddings.int.test.ts b/libs/langchain-openai/src/tests/azure/embeddings.int.test.ts index 634cca967d74..4f848e47d61e 100644 --- a/libs/langchain-openai/src/tests/azure/embeddings.int.test.ts +++ b/libs/langchain-openai/src/tests/azure/embeddings.int.test.ts @@ -2,24 +2,6 @@ import { test, expect } from "@jest/globals"; import { AzureOpenAIEmbeddings as OpenAIEmbeddings } from "../../azure/embeddings.js"; -beforeAll(() => { - if (!process.env.AZURE_OPENAI_API_KEY) { - process.env.AZURE_OPENAI_API_KEY = process.env.TEST_AZURE_OPENAI_API_KEY; - } - if (!process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME) { - process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME = - process.env.TEST_AZURE_OPENAI_API_DEPLOYMENT_NAME; - } - if (!process.env.AZURE_OPENAI_BASE_PATH) { - process.env.AZURE_OPENAI_BASE_PATH = - process.env.TEST_AZURE_OPENAI_BASE_PATH; - } - if (!process.env.AZURE_OPENAI_API_VERSION) { - process.env.AZURE_OPENAI_API_VERSION = - process.env.TEST_AZURE_OPENAI_API_VERSION; - } -}); - test("Test AzureOpenAIEmbeddings.embedQuery", async () => { const embeddings = new OpenAIEmbeddings(); const res = await embeddings.embedQuery("Hello world"); diff --git a/libs/langchain-openai/src/tests/azure/llms.int.test.ts b/libs/langchain-openai/src/tests/azure/llms.int.test.ts index 5ce8369aad1c..412ab907b3a7 100644 --- a/libs/langchain-openai/src/tests/azure/llms.int.test.ts +++ b/libs/langchain-openai/src/tests/azure/llms.int.test.ts @@ -15,24 +15,6 @@ import { AzureOpenAI } from "../../azure/llms.js"; // Save the original value of the 'LANGCHAIN_CALLBACKS_BACKGROUND' environment variable const originalBackground = process.env.LANGCHAIN_CALLBACKS_BACKGROUND; -beforeAll(() => { - if (!process.env.AZURE_OPENAI_API_KEY) { - process.env.AZURE_OPENAI_API_KEY = process.env.TEST_AZURE_OPENAI_API_KEY; - } - if (!process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME) { - process.env.AZURE_OPENAI_API_DEPLOYMENT_NAME = - process.env.TEST_AZURE_OPENAI_API_DEPLOYMENT_NAME; - } - if (!process.env.AZURE_OPENAI_BASE_PATH) { - process.env.AZURE_OPENAI_BASE_PATH = - process.env.TEST_AZURE_OPENAI_BASE_PATH; - } - if (!process.env.AZURE_OPENAI_API_VERSION) { - process.env.AZURE_OPENAI_API_VERSION = - process.env.TEST_AZURE_OPENAI_API_VERSION; - } -}); - test("Test Azure OpenAI invoke", async () => { const model = new AzureOpenAI({ maxTokens: 5, @@ -262,7 +244,7 @@ test("Test Azure OpenAI in streaming mode with multiple prompts", async () => { const model = new AzureOpenAI({ maxTokens: 5, - modelName: "gpt-3.5-turbo", + modelName: "gpt-3.5-turbo-instruct", streaming: true, n: 1, callbacks: CallbackManager.fromHandlers({ diff --git a/libs/langchain-openai/src/tests/llms.int.test.ts b/libs/langchain-openai/src/tests/llms.int.test.ts index 3699d18a8d91..0bafa5b34721 100644 --- a/libs/langchain-openai/src/tests/llms.int.test.ts +++ b/libs/langchain-openai/src/tests/llms.int.test.ts @@ -239,7 +239,7 @@ test("Test OpenAIChat in streaming mode with multiple prompts", async () => { const model = new OpenAI({ maxTokens: 5, - modelName: "gpt-3.5-turbo", + modelName: "gpt-3.5-turbo-instruct", streaming: true, n: 1, callbacks: CallbackManager.fromHandlers({