Detect faults in LLM output
callback
is provided, the evaluation is passed to that function, and inference is not affected.
If no callback
is provided, and stream
is true
, then the evaluation is available on the last chunk.
If no callback
is provided and stream
is false
or none
, then the evaluation can be found on the completion response.