Skip to content
Discussion options

You must be logged in to vote

No, you cannot directly use an LLM-as-a-Judge evaluator you set up in Langfuse as your my_eval_fn in the code example you provided. The my_eval_fn function in your script is meant for custom, local evaluation logic—such as comparing outputs or running your own checks—executed in your Python environment(1).

LLM-as-a-Judge evaluators are managed and executed within Langfuse itself. To use these, you should configure them in the Langfuse UI. When you run a Dataset Run (experiment) from the UI, Langfuse will automatically apply any selected LLM-as-a-Judge evaluators to the outputs of your application, scoring them based on the evaluator's prompts and logic(2)(3). This process is separate from…

Replies: 6 comments 5 replies

Comment options

You must be logged in to vote
1 reply
@erantr1
Comment options

Answer selected by jannikmaierhoefer
Comment options

You must be logged in to vote
1 reply
@erantr1
Comment options

Comment options

You must be logged in to vote
1 reply
@MFerraz
Comment options

Comment options

You must be logged in to vote
1 reply
@MFerraz
Comment options

Comment options

You must be logged in to vote
1 reply
@MFerraz
Comment options

Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
2 participants