Galileo
Search
⌃K

Prompt Sweeps

Execute prompts in bulk for faster experimentation
In Galileo, you can execute multiple prompt runs using what we call "Prompt Sweeps".
A sweep allows you to execute, in bulk, multiple LLM end-to-end runs with different compbinations of - prompt templates, models, data and hyperparameters such as temperature.
Prompt Sweeps for bulk execution and auto tuning
import promptquality as pq
from promptquality import Scorers
from promptquality import SupportedModels
​
models = [
SupportedModels.text_davinci_3,
SupportedModels.chat_gpt_16k,
SupportedModels.gpt_4
]
​
templates = [
""" Given the following context, please answer the question.
Context: {context}
Question: {question}
Your answer: """,
""" You are a helpful assistant. Given the following context,
please answer the question.
----
Context: {context}
----
Question: {question}
----
Your answer:
""",
""" You are a helpful assistant. Given the following context,
please answer the question. Provide an accurate and factual answer.
----
Context: {context}
----
Question: {question}
----
Your answer: """,
""" You are a helpful assistant. Given the following context,
please answer the question. Provide an accurate and factual answer.
If the question is about science, religion or politics, say "I don't
have enough information to answer that question based on the given context."
----
Context: {context}
----
Question: {question}
----
Your answer: """]
​
metrics = [
Scorers.groundedness,
Scorers.latency,
# Uncertainty, BLEU and ROUGE are automatically included
]
​
pq.run_sweep(project_name='my_project_name',
templates=templates,
dataset='my_dataset.csv',
scorers=metrics,
model_aliases=models,
execute=True)
See the PromptQuality Python Library Docs for more information.