Updated LLM config and added read_webpage
All checks were successful
Publish Library / Build NPM Project (push) Successful in 46s
Publish Library / Tag Version (push) Successful in 6s

This commit is contained in:
2026-02-01 13:16:08 -05:00
parent 28904cddbe
commit 7b57a0ded1
11 changed files with 840 additions and 384 deletions

View File

@@ -1,5 +1,5 @@
import {OpenAI as openAI} from 'openai';
import {findByProp, objectMap, JSONSanitize, JSONAttemptParse} from '@ztimson/utils';
import {findByProp, objectMap, JSONSanitize, JSONAttemptParse, clean} from '@ztimson/utils';
import {AbortablePromise, Ai} from './ai.ts';
import {LLMMessage, LLMRequest} from './llm.ts';
import {LLMProvider} from './provider.ts';
@@ -7,9 +7,12 @@ import {LLMProvider} from './provider.ts';
export class OpenAi extends LLMProvider {
client!: openAI;
constructor(public readonly ai: Ai, public readonly apiToken: string, public model: string) {
constructor(public readonly ai: Ai, public readonly host: string | null, public readonly token: string, public model: string) {
super();
this.client = new openAI({apiKey: apiToken});
this.client = new openAI(clean({
baseURL: host,
apiKey: token
}));
}
private toStandard(history: any[]): LLMMessage[] {
@@ -64,16 +67,17 @@ export class OpenAi extends LLMProvider {
ask(message: string, options: LLMRequest = {}): AbortablePromise<LLMMessage[]> {
const controller = new AbortController();
const response = new Promise<any>(async (res, rej) => {
let history = this.fromStandard([...options.history || [], {role: 'user', content: message, timestamp: Date.now()}]);
let history = [...options.history || [], {role: 'user', content: message, timestamp: Date.now()}];
if(options.compress) history = await this.ai.language.compressHistory(<any>history, options.compress.max, options.compress.min, options);
history = this.fromStandard(<any>history);
const tools = options.tools || this.ai.options.tools || [];
const tools = options.tools || this.ai.options.llm?.tools || [];
const requestParams: any = {
model: options.model || this.model,
messages: history,
stream: !!options.stream,
max_tokens: options.max_tokens || this.ai.options.max_tokens || 4096,
temperature: options.temperature || this.ai.options.temperature || 0.7,
max_tokens: options.max_tokens || this.ai.options.llm?.max_tokens || 4096,
temperature: options.temperature || this.ai.options.llm?.temperature || 0.7,
tools: tools.map(t => ({
type: 'function',
function: {