mirror of
https://github.com/YuuKi-OS/Yuuki-chat.git
synced 2026-02-18 22:01:09 +00:00
yuuki-deidad
This commit is contained in:
189
app/api/chat/route.ts
Normal file
189
app/api/chat/route.ts
Normal file
@@ -0,0 +1,189 @@
|
||||
import { NextRequest, NextResponse } from "next/server";
|
||||
|
||||
const HF_MODELS: Record<string, string> = {
|
||||
"yuuki-v0.1":
|
||||
"https://api-inference.huggingface.co/models/YuuKi-OS/Yuuki-v0.1",
|
||||
"yuuki-3.7":
|
||||
"https://api-inference.huggingface.co/models/YuuKi-OS/Yuuki-3.7",
|
||||
"yuuki-best":
|
||||
"https://api-inference.huggingface.co/models/YuuKi-OS/Yuuki-best",
|
||||
};
|
||||
|
||||
const YUUKI_API_MODELS: Record<string, string> = {
|
||||
"yuuki-v0.1": "yuuki-v0.1",
|
||||
"yuuki-3.7": "yuuki-3.7",
|
||||
"yuuki-best": "yuuki-best",
|
||||
};
|
||||
|
||||
/**
|
||||
* Calls the Yuuki API (yuuki-api.vercel.app) with a yk- token.
|
||||
* This is an OpenAI-compatible endpoint.
|
||||
*/
|
||||
async function callYuukiApi(
|
||||
token: string,
|
||||
model: string,
|
||||
messages: { role: string; content: string }[]
|
||||
) {
|
||||
const modelId = YUUKI_API_MODELS[model] || "yuuki-best";
|
||||
|
||||
const response = await fetch("https://yuuki-api.vercel.app/api/chat", {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
Authorization: `Bearer ${token}`,
|
||||
},
|
||||
body: JSON.stringify({
|
||||
model: modelId,
|
||||
messages,
|
||||
max_tokens: 1024,
|
||||
temperature: 0.7,
|
||||
}),
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
const errorText = await response.text();
|
||||
throw new Error(
|
||||
`Yuuki API error (${response.status}): ${errorText.slice(0, 200)}`
|
||||
);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
const content =
|
||||
data.choices?.[0]?.message?.content || data.content || "No response";
|
||||
return { content, id: data.id || `chatcmpl-${Date.now()}`, model: modelId };
|
||||
}
|
||||
|
||||
/**
|
||||
* Calls HuggingFace Inference API directly with an hf_ token.
|
||||
*/
|
||||
async function callHuggingFace(
|
||||
token: string,
|
||||
model: string,
|
||||
messages: { role: string; content: string }[]
|
||||
) {
|
||||
const modelUrl = HF_MODELS[model] || HF_MODELS["yuuki-best"];
|
||||
|
||||
const prompt =
|
||||
messages
|
||||
.map((m) => {
|
||||
if (m.role === "system") return `System: ${m.content}`;
|
||||
if (m.role === "user") return `User: ${m.content}`;
|
||||
if (m.role === "assistant") return `Assistant: ${m.content}`;
|
||||
return m.content;
|
||||
})
|
||||
.join("\n") + "\nAssistant:";
|
||||
|
||||
const response = await fetch(modelUrl, {
|
||||
method: "POST",
|
||||
headers: {
|
||||
Authorization: `Bearer ${token}`,
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: JSON.stringify({
|
||||
inputs: prompt,
|
||||
parameters: {
|
||||
max_new_tokens: 1024,
|
||||
temperature: 0.7,
|
||||
top_p: 0.9,
|
||||
repetition_penalty: 1.1,
|
||||
return_full_text: false,
|
||||
},
|
||||
}),
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
const errorText = await response.text();
|
||||
throw new Error(
|
||||
`HuggingFace error (${response.status}): ${errorText.slice(0, 200)}`
|
||||
);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
let generatedText = "";
|
||||
|
||||
if (Array.isArray(data) && data[0]?.generated_text) {
|
||||
generatedText = data[0].generated_text.trim();
|
||||
} else if (typeof data === "string") {
|
||||
generatedText = data.trim();
|
||||
} else if (data?.generated_text) {
|
||||
generatedText = data.generated_text.trim();
|
||||
} else {
|
||||
generatedText = JSON.stringify(data);
|
||||
}
|
||||
|
||||
// Clean up artifacts
|
||||
const cutoffs = ["User:", "System:", "\nUser", "\nSystem"];
|
||||
for (const cutoff of cutoffs) {
|
||||
const idx = generatedText.indexOf(cutoff);
|
||||
if (idx > 0) generatedText = generatedText.substring(0, idx).trim();
|
||||
}
|
||||
|
||||
return {
|
||||
content:
|
||||
generatedText ||
|
||||
"I received your message but couldn't generate a response. Please try again.",
|
||||
id: `chatcmpl-${Date.now()}`,
|
||||
model,
|
||||
};
|
||||
}
|
||||
|
||||
export async function POST(req: NextRequest) {
|
||||
try {
|
||||
const body = await req.json();
|
||||
const { messages, model, token, tokenSource } = body;
|
||||
|
||||
if (!messages || !Array.isArray(messages)) {
|
||||
return NextResponse.json(
|
||||
{ error: "messages is required" },
|
||||
{ status: 400 }
|
||||
);
|
||||
}
|
||||
|
||||
const modelKey = model || "yuuki-best";
|
||||
if (!HF_MODELS[modelKey]) {
|
||||
return NextResponse.json({ error: "Invalid model" }, { status: 400 });
|
||||
}
|
||||
|
||||
let result;
|
||||
|
||||
if (tokenSource === "demo") {
|
||||
// Demo mode: use server-side HF_DEMO_TOKEN directly against HuggingFace
|
||||
const demoToken = process.env.HF_DEMO_TOKEN;
|
||||
if (!demoToken) {
|
||||
return NextResponse.json(
|
||||
{ error: "Demo token not configured on server" },
|
||||
{ status: 500 }
|
||||
);
|
||||
}
|
||||
result = await callHuggingFace(demoToken, modelKey, messages);
|
||||
} else if (tokenSource === "yuuki-api") {
|
||||
// Yuuki API: yk- tokens go to yuuki-api.vercel.app
|
||||
if (!token) {
|
||||
return NextResponse.json(
|
||||
{ error: "No API token provided" },
|
||||
{ status: 401 }
|
||||
);
|
||||
}
|
||||
result = await callYuukiApi(token, modelKey, messages);
|
||||
} else if (tokenSource === "huggingface") {
|
||||
// HuggingFace: hf_ tokens go directly to HF Inference API
|
||||
if (!token) {
|
||||
return NextResponse.json(
|
||||
{ error: "No API token provided" },
|
||||
{ status: 401 }
|
||||
);
|
||||
}
|
||||
result = await callHuggingFace(token, modelKey, messages);
|
||||
} else {
|
||||
return NextResponse.json(
|
||||
{ error: "Invalid token source" },
|
||||
{ status: 400 }
|
||||
);
|
||||
}
|
||||
|
||||
return NextResponse.json(result);
|
||||
} catch (error) {
|
||||
const message = error instanceof Error ? error.message : "Unknown error";
|
||||
return NextResponse.json({ error: message }, { status: 500 });
|
||||
}
|
||||
}
|
||||
54
app/api/research/route.ts
Normal file
54
app/api/research/route.ts
Normal file
@@ -0,0 +1,54 @@
|
||||
import { NextRequest, NextResponse } from "next/server";
|
||||
|
||||
export async function POST(req: NextRequest) {
|
||||
try {
|
||||
const { query } = await req.json();
|
||||
|
||||
if (!query) {
|
||||
return NextResponse.json({ error: "query is required" }, { status: 400 });
|
||||
}
|
||||
|
||||
const apiKey = process.env.TAVILY_API_KEY;
|
||||
if (!apiKey) {
|
||||
return NextResponse.json({ error: "Tavily API key not configured" }, { status: 500 });
|
||||
}
|
||||
|
||||
const response = await fetch("https://api.tavily.com/search", {
|
||||
method: "POST",
|
||||
headers: { "Content-Type": "application/json" },
|
||||
body: JSON.stringify({
|
||||
api_key: apiKey,
|
||||
query,
|
||||
search_depth: "advanced",
|
||||
include_answer: true,
|
||||
include_raw_content: false,
|
||||
max_results: 5,
|
||||
}),
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
const text = await response.text();
|
||||
return NextResponse.json(
|
||||
{ error: `Tavily error: ${response.status}`, details: text },
|
||||
{ status: response.status }
|
||||
);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
|
||||
return NextResponse.json({
|
||||
answer: data.answer || "",
|
||||
results: (data.results || []).map(
|
||||
(r: { title: string; url: string; content: string; score: number }) => ({
|
||||
title: r.title,
|
||||
url: r.url,
|
||||
content: r.content,
|
||||
score: r.score,
|
||||
})
|
||||
),
|
||||
});
|
||||
} catch (error) {
|
||||
const message = error instanceof Error ? error.message : "Unknown error";
|
||||
return NextResponse.json({ error: message }, { status: 500 });
|
||||
}
|
||||
}
|
||||
62
app/api/youtube/route.ts
Normal file
62
app/api/youtube/route.ts
Normal file
@@ -0,0 +1,62 @@
|
||||
import { NextRequest, NextResponse } from "next/server";
|
||||
|
||||
export async function POST(req: NextRequest) {
|
||||
try {
|
||||
const { query } = await req.json();
|
||||
|
||||
if (!query) {
|
||||
return NextResponse.json({ error: "query is required" }, { status: 400 });
|
||||
}
|
||||
|
||||
const apiKey = process.env.YOUTUBE_API_KEY;
|
||||
if (!apiKey) {
|
||||
return NextResponse.json({ error: "YouTube API key not configured" }, { status: 500 });
|
||||
}
|
||||
|
||||
const params = new URLSearchParams({
|
||||
part: "snippet",
|
||||
q: query,
|
||||
key: apiKey,
|
||||
maxResults: "5",
|
||||
type: "video",
|
||||
});
|
||||
|
||||
const response = await fetch(
|
||||
`https://www.googleapis.com/youtube/v3/search?${params.toString()}`
|
||||
);
|
||||
|
||||
if (!response.ok) {
|
||||
const text = await response.text();
|
||||
return NextResponse.json(
|
||||
{ error: `YouTube error: ${response.status}`, details: text },
|
||||
{ status: response.status }
|
||||
);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
|
||||
const videos = (data.items || []).map(
|
||||
(item: {
|
||||
id: { videoId: string };
|
||||
snippet: {
|
||||
title: string;
|
||||
description: string;
|
||||
channelTitle: string;
|
||||
thumbnails: { medium: { url: string } };
|
||||
};
|
||||
}) => ({
|
||||
id: item.id.videoId,
|
||||
title: item.snippet.title,
|
||||
description: item.snippet.description,
|
||||
channel: item.snippet.channelTitle,
|
||||
thumbnail: item.snippet.thumbnails?.medium?.url || "",
|
||||
url: `https://youtube.com/watch?v=${item.id.videoId}`,
|
||||
})
|
||||
);
|
||||
|
||||
return NextResponse.json({ videos });
|
||||
} catch (error) {
|
||||
const message = error instanceof Error ? error.message : "Unknown error";
|
||||
return NextResponse.json({ error: message }, { status: 500 });
|
||||
}
|
||||
}
|
||||
Reference in New Issue
Block a user