mirror of
https://github.com/stackblitz/bolt.new
synced 2024-11-27 22:42:21 +00:00
52 lines
1.7 KiB
TypeScript
52 lines
1.7 KiB
TypeScript
import { type ActionFunctionArgs } from '@remix-run/cloudflare';
|
|
import { StreamingTextResponse } from 'ai';
|
|
import { MAX_RESPONSE_SEGMENTS, MAX_TOKENS } from '~/lib/.server/llm/constants';
|
|
import { CONTINUE_PROMPT } from '~/lib/.server/llm/prompts';
|
|
import { streamText, type Messages, type StreamingOptions } from '~/lib/.server/llm/stream-text';
|
|
import SwitchableStream from '~/lib/.server/llm/switchable-stream';
|
|
|
|
export async function action({ context, request }: ActionFunctionArgs) {
|
|
const { messages } = await request.json<{ messages: Messages }>();
|
|
|
|
const stream = new SwitchableStream();
|
|
|
|
try {
|
|
const options: StreamingOptions = {
|
|
toolChoice: 'none',
|
|
onFinish: async ({ text: content, finishReason }) => {
|
|
if (finishReason !== 'length') {
|
|
return stream.close();
|
|
}
|
|
|
|
if (stream.switches >= MAX_RESPONSE_SEGMENTS) {
|
|
throw Error('Cannot continue message: Maximum segments reached');
|
|
}
|
|
|
|
const switchesLeft = MAX_RESPONSE_SEGMENTS - stream.switches;
|
|
|
|
console.log(`Reached max token limit (${MAX_TOKENS}): Continuing message (${switchesLeft} switches left)`);
|
|
|
|
messages.push({ role: 'assistant', content });
|
|
messages.push({ role: 'user', content: CONTINUE_PROMPT });
|
|
|
|
const result = await streamText(messages, context.cloudflare.env, options);
|
|
|
|
return stream.switchSource(result.toAIStream());
|
|
},
|
|
};
|
|
|
|
const result = await streamText(messages, context.cloudflare.env, options);
|
|
|
|
stream.switchSource(result.toAIStream());
|
|
|
|
return new StreamingTextResponse(stream.readable);
|
|
} catch (error) {
|
|
console.log(error);
|
|
|
|
throw new Response(null, {
|
|
status: 500,
|
|
statusText: 'Internal Server Error',
|
|
});
|
|
}
|
|
}
|