Galileo Evaluate enables you to evaluate and optimize your prompts with out-of-the-box Guardrail metrics.

  1. Pip Install promptquality and create runs in your Python notebook.

  2. Next, you execute promptquality.run() like shown below.


    import promptquality as pq

    pq.login({YOUR_GALILEO_URL})

    template = "Explain {topic} to me like I'm a 5 year old"

    data = {"topic": ["Quantum Physics", "Politics", "Large Language Models"]}

    pq.run(project_name='my_first_project',
           template=template,
           dataset=data,
           settings=pq.Settings(model_alias='ChatGPT (16K context)',
                                temperature=0.8,
                                max_tokens=400))

The code snippet above uses ChatGPT API endpoint from OpenAI. Want to use other models (Azure OpenAI, Cohere, Anthropic, Mistral, etc)? Check out the integration page here.