Skip to content

Commit

Permalink
Merge pull request #216 from lmnr-ai/dev
Browse files Browse the repository at this point in the history
update models in pipeline templates to 4o family (#215)
  • Loading branch information
dinmukhamedm authored Nov 15, 2024
2 parents f3a0450 + 89498e7 commit b30e993
Showing 1 changed file with 16 additions and 16 deletions.
32 changes: 16 additions & 16 deletions frontend/lib/db/seed.json
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@
"id": "<node_0_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{instruction}}",
"stream": false,
Expand Down Expand Up @@ -184,7 +184,7 @@
"id": "<node_0_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{instruction}}",
"stream": false,
Expand Down Expand Up @@ -277,7 +277,7 @@
"id": "<node_2_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [
{
"id": "23fce76d-2061-4fea-b5cb-c307d7eab811",
Expand Down Expand Up @@ -393,7 +393,7 @@
"id": "<node_2_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [
{
"id": "23fce76d-2061-4fea-b5cb-c307d7eab811",
Expand Down Expand Up @@ -701,7 +701,7 @@
"id": "<node_1_id>",
"name": "candidate_1",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand All @@ -727,7 +727,7 @@
"id": "<node_2_id>",
"name": "candidate_3",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand All @@ -753,7 +753,7 @@
"id": "<node_3_id>",
"name": "Evaluator",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "You will be given a prompt and 3 outputs of Large Language Models in response to this prompt. Your goal is to produce strengths and weaknesses of each response.\n\n---\n\nPrompt:\n{{prompt}}\n\nCandidate 1:\n{{cand_1}}\n\nCandidate 2:\n{{cand_2}}\n\nCandidate 3:\n{{cand_3}}",
"stream": false,
Expand Down Expand Up @@ -797,7 +797,7 @@
"id": "<node_4_id>",
"name": "final_LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "You will be given a prompt, 3 candidate responses and strengths and weaknesses of each of responses. Your goal is to produce the best overall response which incorporates strengths and weaknesses of each candidate.\n\nPay close attention to highlighted strengths and weaknesses. Solve weaknesses and keep strengths.\n\nprompt:\n{{prompt}}\n\n\u0421andidate 1:\n{{cand_1}}\n\n\u0421andidate 2:\n{{cand_2}}\n\n\u0421andidate 3:\n{{cand_3}}\n\n---\nStrengths and Weaknesses:\n{{reasoning}}",
"stream": false,
Expand Down Expand Up @@ -847,7 +847,7 @@
"id": "<node_5_id>",
"name": "candidate_2",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand Down Expand Up @@ -1080,7 +1080,7 @@
"id": "<node_1_id>",
"name": "candidate_1",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand Down Expand Up @@ -1120,7 +1120,7 @@
"id": "<node_5_id>",
"name": "candidate_2",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand Down Expand Up @@ -1160,7 +1160,7 @@
"id": "<node_2_id>",
"name": "candidate_3",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "{{prompt}}",
"stream": false,
Expand Down Expand Up @@ -1200,7 +1200,7 @@
"id": "<node_3_id>",
"name": "Evaluator",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "You will be given a prompt and 3 outputs of Large Language Models in response to this prompt. Your goal is to produce strengths and weaknesses of each response.\n\n---\n\nPrompt:\n{{prompt}}\n\nCandidate 1:\n{{cand_1}}\n\nCandidate 2:\n{{cand_2}}\n\nCandidate 3:\n{{cand_3}}",
"stream": false,
Expand Down Expand Up @@ -1287,7 +1287,7 @@
"id": "<node_4_id>",
"name": "final_LLM",
"type": "LLM",
"model": "openai:gpt-3.5-turbo",
"model": "openai:gpt-4o-mini",
"inputs": [],
"prompt": "You will be given a prompt, 3 candidate responses and strengths and weaknesses of each of responses. Your goal is to produce the best overall response which incorporates strengths and weaknesses of each candidate.\n\nPay close attention to highlighted strengths and weaknesses. Solve weaknesses and keep strengths.\n\nprompt:\n{{prompt}}\n\n\u0421andidate 1:\n{{cand_1}}\n\n\u0421andidate 2:\n{{cand_2}}\n\n\u0421andidate 3:\n{{cand_3}}\n\n---\nStrengths and Weaknesses:\n{{reasoning}}",
"stream": false,
Expand Down Expand Up @@ -1783,7 +1783,7 @@
"id": "<node_2_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-4-turbo",
"model": "openai:gpt-4o",
"inputs": [],
"prompt": "You are given a question, context and answer. Your goal is to evaluate on the scale from 1 to 5 how relevant is the answer to the question and whether answer is grounded in the context. The score represents the fraction of claims in the output that are relevant to the question over total claims in the output. A high score means most retrieved content directly answers the question. A low score indicates excess or irrelevant information.\n\nIf answer seems correct, but it does not have any relevance to the context, grade it with a low score\n\nIf the context does not contain information related to the question, and the answer correctly acknowledges it, then give higher score.\n\nQuestion:\n{{question}}\n\nAnswer:\n{{answer}}\n\nContext:\n{{context}}",
"stream": false,
Expand Down Expand Up @@ -2016,7 +2016,7 @@
"id": "<node_2_id>",
"name": "LLM",
"type": "LLM",
"model": "openai:gpt-4-turbo",
"model": "openai:gpt-4o",
"inputs": [],
"prompt": "You are given a question, context and answer. Your goal is to evaluate on the scale from 1 to 5 how relevant is the answer to the question and whether answer is grounded in the context. The score represents the fraction of claims in the output that are relevant to the question over total claims in the output. A high score means most retrieved content directly answers the question. A low score indicates excess or irrelevant information.\n\nIf answer seems correct, but it does not have any relevance to the context, grade it with a low score\n\nIf the context does not contain information related to the question, and the answer correctly acknowledges it, then give higher score.\n\nQuestion:\n{{question}}\n\nAnswer:\n{{answer}}\n\nContext:\n{{context}}",
"stream": false,
Expand Down

0 comments on commit b30e993

Please sign in to comment.