diff --git a/examples/src/models/llm/llm_streaming_stream_method.ts b/examples/src/models/llm/llm_streaming_stream_method.ts index 7dcd9bd872e7..2dec6326cbfb 100644 --- a/examples/src/models/llm/llm_streaming_stream_method.ts +++ b/examples/src/models/llm/llm_streaming_stream_method.ts @@ -3,6 +3,7 @@ import { OpenAI } from "langchain/llms/openai"; // To enable streaming, we pass in `streaming: true` to the LLM constructor. const model = new OpenAI({ maxTokens: 25, + streaming: true, }); const stream = await model.stream("Tell me a joke.");