From effe423e05dfc9af96ade9cc36888fe25a8b93a8 Mon Sep 17 00:00:00 2001 From: Clay Ferguson Date: Fri, 20 Dec 2024 20:42:12 -0600 Subject: [PATCH] removed deprecated Perplexity AI models --- QuantaAI/quanta_ai.py | 10 ---------- src/main/java/quanta/model/client/AIModel.java | 2 -- src/main/resources/public/src/AIUtil.ts | 14 -------------- src/main/resources/public/src/JavaIntf.ts | 2 -- 4 files changed, 28 deletions(-) diff --git a/QuantaAI/quanta_ai.py b/QuantaAI/quanta_ai.py index 32fe5d578..09aa377a1 100644 --- a/QuantaAI/quanta_ai.py +++ b/QuantaAI/quanta_ai.py @@ -28,8 +28,6 @@ OPENAI_MODEL_COMPLETION_O1_PREVIEW = "o1-preview" OPENAI_MODEL_COMPLETION_O1_MINI = "o1-mini" PPLX_MODEL_COMPLETION_ONLINE = "llama-3.1-sonar-huge-128k-online" -PPLX_MODEL_COMPLETION_LLAMA3 = "llama-3.1-70b-instruct" -PPLX_MODEL_COMPLETION_CHAT = "llama-3.1-sonar-large-128k-chat" GEMINI_MODEL_COMPLETION_CHAT = "gemini-1.5-pro" GEMINI_FLASH_MODEL_COMPLETION_CHAT = "gemini-1.5-flash" XAI_MODEL_COMPLETION_CHAT = "grok-beta" @@ -273,14 +271,6 @@ def calculate_cost(input_tokens, output_tokens, model) -> float: input_ppm = 3.0 output_ppm = 15.0 - elif model == PPLX_MODEL_COMPLETION_CHAT: - input_ppm = 1.0 - output_ppm = 1.0 - - elif model == PPLX_MODEL_COMPLETION_LLAMA3: - input_ppm = 1.0 - output_ppm = 1.0 - elif model == PPLX_MODEL_COMPLETION_ONLINE: input_ppm = 1.0 output_ppm = 1.0 diff --git a/src/main/java/quanta/model/client/AIModel.java b/src/main/java/quanta/model/client/AIModel.java index 16f149533..dd93f2b51 100644 --- a/src/main/java/quanta/model/client/AIModel.java +++ b/src/main/java/quanta/model/client/AIModel.java @@ -13,10 +13,8 @@ public enum AIModel { OPENAI_O1_PREVIEW("openAi_o1Preview", "openai", "o1-preview", "OpenAI: ChatGPT o1-preview", "O1P", true, 128_000), // OPENAI_O1_MINI("openAi_o1Mini", "openai", "o1-mini", "OpenAI: ChatGPT o1-mini", "O1M", true, 128_000), // - PPLX_CHAT("pplxAi", "perplexity", "llama-3.1-sonar-large-128k-chat", "Perplexity: Basic", "PPB", true, 127_000), // PPLX_ONLINE("pplxAi_online", "perplexity", "llama-3.1-sonar-huge-128k-online", "Perplexity: Recent News Aware", "PPN", false, 127_000), // - PPLX_LLAMA3("llama3", "perplexity", "llama-3.1-70b-instruct", "Meta: Llama 3", "PPL", true, 131_000), // ANTH("anthAi", "anthropic", "claude-3-opus-20240229", "Anthropic: Claude 3 Opus", "ACL", true, 200_000), // diff --git a/src/main/resources/public/src/AIUtil.ts b/src/main/resources/public/src/AIUtil.ts index c17f49070..a905bc816 100755 --- a/src/main/resources/public/src/AIUtil.ts +++ b/src/main/resources/public/src/AIUtil.ts @@ -49,15 +49,6 @@ export class AIUtil { }); } - if (S.quanta.config.usePplxAi) { - this.aiServices.push(// - { - name: J.AIModel.PPLX_LLAMA3, - description: "Meta: Llama 3", - longDescription: "Meta's Open Source Llama 3 (131K context, max $1/megatoken)" - }); - } - if (S.quanta.config.useAnthAi) { this.aiServices.push(// { @@ -74,11 +65,6 @@ export class AIUtil { if (S.quanta.config.usePplxAi) { this.aiServices.push(// - { - name: J.AIModel.PPLX_CHAT, - description: "Perplexity: Basic", - longDescription: "Perplexity's best high-end powerful general-purpose AI (127K context, max $1/megatoken)" - }, { name: J.AIModel.PPLX_ONLINE, description: "Perplexity: Recent News Aware", diff --git a/src/main/resources/public/src/JavaIntf.ts b/src/main/resources/public/src/JavaIntf.ts index 8ef1e7381..f1c40b9c4 100644 --- a/src/main/resources/public/src/JavaIntf.ts +++ b/src/main/resources/public/src/JavaIntf.ts @@ -1101,9 +1101,7 @@ export const enum AIModel { OPENAI_MINI = "openAiMini", OPENAI_O1_PREVIEW = "openAi_o1Preview", OPENAI_O1_MINI = "openAi_o1Mini", - PPLX_CHAT = "pplxAi", PPLX_ONLINE = "pplxAi_online", - PPLX_LLAMA3 = "llama3", ANTH = "anthAi", ANTH_SONNET = "anthAi_sonnet", GEMINI = "geminiAi",