2024-06-30 20:45:06 +05:30
|
|
|
import { ChatChromeAI } from "./ChatChromeAi"
|
|
|
|
|
import { ChatOllama } from "./ChatOllama"
|
|
|
|
|
|
|
|
|
|
export const pageAssistModel = async ({
|
|
|
|
|
model,
|
|
|
|
|
baseUrl,
|
|
|
|
|
keepAlive,
|
|
|
|
|
temperature,
|
|
|
|
|
topK,
|
|
|
|
|
topP,
|
|
|
|
|
numCtx,
|
2024-08-20 16:11:50 +05:30
|
|
|
seed,
|
|
|
|
|
numGpu
|
2024-06-30 20:45:06 +05:30
|
|
|
}: {
|
|
|
|
|
model: string
|
|
|
|
|
baseUrl: string
|
2024-08-05 00:49:27 +05:30
|
|
|
keepAlive?: string
|
|
|
|
|
temperature?: number
|
|
|
|
|
topK?: number
|
|
|
|
|
topP?: number
|
|
|
|
|
numCtx?: number
|
|
|
|
|
seed?: number
|
2024-08-20 16:11:50 +05:30
|
|
|
numGpu?: number
|
2024-06-30 20:45:06 +05:30
|
|
|
}) => {
|
|
|
|
|
switch (model) {
|
|
|
|
|
case "chrome::gemini-nano::page-assist":
|
|
|
|
|
return new ChatChromeAI({
|
|
|
|
|
temperature,
|
|
|
|
|
topK
|
|
|
|
|
})
|
|
|
|
|
default:
|
|
|
|
|
return new ChatOllama({
|
|
|
|
|
baseUrl,
|
|
|
|
|
keepAlive,
|
|
|
|
|
temperature,
|
|
|
|
|
topK,
|
|
|
|
|
topP,
|
|
|
|
|
numCtx,
|
|
|
|
|
seed,
|
2024-08-20 16:11:50 +05:30
|
|
|
model,
|
|
|
|
|
numGpu
|
2024-06-30 20:45:06 +05:30
|
|
|
})
|
|
|
|
|
}
|
|
|
|
|
}
|