Skip to content

Commit 1ce9592

Browse files
committed
[buffbench] add more logging to eval errors
1 parent 7433060 commit 1ce9592

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

evals/git-evals/run-git-evals.ts

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@ import path from 'path'
44

55
import { disableLiveUserInputCheck } from '@codebuff/backend/live-user-inputs'
66
import { promptAiSdkStructured } from '@codebuff/backend/llm-apis/vercel-ai-sdk/ai-sdk'
7+
import { errorToObject } from '@codebuff/common/util/object'
78
import { withTimeout } from '@codebuff/common/util/promise'
89
import { generateCompactId } from '@codebuff/common/util/string'
910
import { cloneDeep } from 'lodash'
@@ -247,7 +248,7 @@ Explain your reasoning in detail.`,
247248
return {
248249
...evalRun,
249250
judging_results: {
250-
analysis: 'Judging failed due to error',
251+
analysis: `Judging failed due to error:\n${judgingError instanceof Error ? errorToObject(judgingError) : JSON.stringify(judgingError)}`,
251252
strengths: [],
252253
weaknesses: ['Judging process encountered an error'],
253254
metrics: {

0 commit comments

Comments
 (0)