This commit is contained in:
Anirban Kar 2024-12-06 16:58:04 +05:30
parent 5ead47992d
commit 7efad13284
3 changed files with 60 additions and 50 deletions

View File

@ -58,7 +58,12 @@ function extractPropertiesFromMessage(message: Message): { model: string; provid
return { model, provider, content: cleanedContent };
}
export async function streamText(messages: Messages, env: Env, options?: StreamingOptions,apiKeys?: Record<string, string>) {
export async function streamText(
messages: Messages,
env: Env,
options?: StreamingOptions,
apiKeys?: Record<string, string>,
) {
let currentModel = DEFAULT_MODEL;
let currentProvider = DEFAULT_PROVIDER.name;
const MODEL_LIST = await getModelList(apiKeys || {});
@ -69,6 +74,7 @@ export async function streamText(messages: Messages, env: Env, options?: Streami
if (MODEL_LIST.find((m) => m.name === model)) {
currentModel = model;
}
currentProvider = provider;
return { ...message, content };

View File

@ -29,7 +29,7 @@ function parseCookies(cookieHeader:string) {
}
async function chatAction({ context, request }: ActionFunctionArgs) {
const { messages, model } = await request.json<{
const { messages } = await request.json<{
messages: Messages;
model: string;
}>();
@ -37,7 +37,7 @@ async function chatAction({ context, request }: ActionFunctionArgs) {
const cookieHeader = request.headers.get('Cookie');
// Parse the cookie's value (returns an object or null if no cookie exists)
const apiKeys = JSON.parse(parseCookies(cookieHeader||"").apiKeys || '{}');
const apiKeys = JSON.parse(parseCookies(cookieHeader || '').apiKeys || '{}');
const stream = new SwitchableStream();

View File

@ -295,7 +295,6 @@ const staticModels: ModelInfo[] = PROVIDER_LIST.map((p) => p.staticModels).flat(
export let MODEL_LIST: ModelInfo[] = [...staticModels];
export async function getModelList(apiKeys: Record<string, string>) {
MODEL_LIST = [
...(
@ -312,13 +311,14 @@ export async function getModelList(apiKeys: Record<string, string>) {
async function getTogetherModels(apiKeys?: Record<string, string>): Promise<ModelInfo[]> {
try {
let baseUrl = import.meta.env.TOGETHER_API_BASE_URL || '';
let provider='Together'
const baseUrl = import.meta.env.TOGETHER_API_BASE_URL || '';
const provider = 'Together';
if (!baseUrl) {
return [];
}
let apiKey = import.meta.env.OPENAI_LIKE_API_KEY ?? ''
let apiKey = import.meta.env.OPENAI_LIKE_API_KEY ?? '';
if (apiKeys && apiKeys[provider]) {
apiKey = apiKeys[provider];
@ -334,13 +334,14 @@ async function getTogetherModels(apiKeys?: Record<string, string>): Promise<Mode
},
});
const res = (await response.json()) as any;
let data: any[] = (res || []).filter((model: any) => model.type=='chat')
const data: any[] = (res || []).filter((model: any) => model.type == 'chat');
return data.map((m: any) => ({
name: m.id,
label: `${m.display_name} - in:$${(m.pricing.input).toFixed(
label: `${m.display_name} - in:$${m.pricing.input.toFixed(
2,
)} out:$${(m.pricing.output).toFixed(2)} - context ${Math.floor(m.context_length / 1000)}k`,
provider: provider,
)} out:$${m.pricing.output.toFixed(2)} - context ${Math.floor(m.context_length / 1000)}k`,
provider,
maxTokenAllowed: 8000,
}));
} catch (e) {
@ -349,7 +350,6 @@ async function getTogetherModels(apiKeys?: Record<string, string>): Promise<Mode
}
}
const getOllamaBaseUrl = () => {
const defaultBaseUrl = import.meta.env.OLLAMA_API_BASE_URL || 'http://localhost:11434';
@ -396,11 +396,13 @@ async function getOpenAILikeModels(): Promise<ModelInfo[]> {
if (!baseUrl) {
return [];
}
let apiKey = import.meta.env.OPENAI_LIKE_API_KEY ?? '';
let apikeys = JSON.parse(Cookies.get('apiKeys')||'{}')
if (apikeys && apikeys['OpenAILike']){
apiKey = apikeys['OpenAILike'];
const apikeys = JSON.parse(Cookies.get('apiKeys') || '{}');
if (apikeys && apikeys.OpenAILike) {
apiKey = apikeys.OpenAILike;
}
const response = await fetch(`${baseUrl}/models`, {
@ -458,6 +460,7 @@ async function getLMStudioModels(): Promise<ModelInfo[]> {
if (typeof window === 'undefined') {
return [];
}
try {
const baseUrl = import.meta.env.LMSTUDIO_API_BASE_URL || 'http://localhost:1234';
const response = await fetch(`${baseUrl}/v1/models`);
@ -476,6 +479,7 @@ async function getLMStudioModels(): Promise<ModelInfo[]> {
async function initializeModelList(): Promise<ModelInfo[]> {
let apiKeys: Record<string, string> = {};
try {
const storedApiKeys = Cookies.get('apiKeys');
@ -486,9 +490,8 @@ async function initializeModelList(): Promise<ModelInfo[]> {
apiKeys = parsedKeys;
}
}
} catch (error) {
} catch (error: any) {
console.warn(`Failed to fetch apikeys from cookies:${error?.message}`);
}
MODEL_LIST = [
...(
@ -500,6 +503,7 @@ async function initializeModelList(): Promise<ModelInfo[]> {
).flat(),
...staticModels,
];
return MODEL_LIST;
}