From 905a4a873684bf8f342f6c690ab943545de16639 Mon Sep 17 00:00:00 2001 From: Reza Rahemtola Date: Thu, 4 Jul 2024 13:26:10 +0200 Subject: [PATCH] fix: Uniformize persona role property --- package.json | 2 +- src/inference.ts | 16 ++++++++-------- src/types.ts | 4 ++-- 3 files changed, 11 insertions(+), 11 deletions(-) diff --git a/package.json b/package.json index 5c8fea0..afc17f8 100644 --- a/package.json +++ b/package.json @@ -1,6 +1,6 @@ { "name": "@libertai/libertai-js", - "version": "0.0.6", + "version": "0.0.7", "description": "In-browser SDK for interacting with LibertAI Decentralized AI Network", "keywords": [], "type": "module", diff --git a/src/inference.ts b/src/inference.ts index b743dfb..e4fa6a7 100644 --- a/src/inference.ts +++ b/src/inference.ts @@ -1,6 +1,6 @@ import axios from 'axios'; -import { Persona, Model, Message } from './types.js'; +import { Message, Model, Persona } from './types.js'; import { calculateTokenLength } from './utils.js'; // Simple wrapper class around basic AI inference @@ -69,9 +69,9 @@ export class LlamaCppApiEngine { } let fullResults = compoundedResult + lastResult; - for (let i = 0; i < stop_sequences.length; i++) { - fullResults = fullResults.split(`\n${stop_sequences[i]}`).join('|||||'); - fullResults = fullResults.split(`${stop_sequences[i]}`).join('|||||'); + for (const element of stop_sequences) { + fullResults = fullResults.split(`\n${element}`).join('|||||'); + fullResults = fullResults.split(`${element}`).join('|||||'); } const results = fullResults.split('|||||'); @@ -152,11 +152,11 @@ export class LlamaCppApiEngine { targetUser = messages[messages.length - 1].role; } - // Set {{char}} based on persona.name + // Set {{char}} based on persona.role // Set {{user}} based on targetUser // Set {{model}} based on model.name let description = persona.description; - description = description.replace(/\{\{char\}\}/g, persona.name); + description = description.replace(/\{\{char\}\}/g, persona.role); description = description.replace(/\{\{user\}\}/g, targetUser); description = description.replace(/\{\{model\}\}/g, model.name); @@ -170,9 +170,9 @@ export class LlamaCppApiEngine { // Determine how many tokens we have left usedTokens = calculateTokenLength(systemPrompt); - // Iterate over messagse in reverse order + // Iterate over messages in reverse order // to generate the chat log - let chatLog = `${promptFormat.userPrepend}${persona.name.toLowerCase()}${promptFormat.userAppend}`; + let chatLog = `${promptFormat.userPrepend}${persona.role.toLowerCase()}${promptFormat.userAppend}`; for (let i = messages.length - 1; i >= 0; i--) { const message = messages[i]; let messageLog = ''; diff --git a/src/types.ts b/src/types.ts index 13eb7b6..227a59d 100644 --- a/src/types.ts +++ b/src/types.ts @@ -63,8 +63,8 @@ export interface PromptFormat { // and extensible persona format of our own // https://github.com/TavernAI/TavernAI/tree/main export interface Persona { - // Persona name - name: string; + // Persona role + role: string; // Persona description description: string;