mirror of
https://github.com/voson-wang/toon.git
synced 2026-01-29 15:24:10 +08:00
test: add LLM retrieval accuracy tests
This commit is contained in:
39
benchmarks/src/constants.ts
Normal file
39
benchmarks/src/constants.ts
Normal file
@@ -0,0 +1,39 @@
|
||||
import process from 'node:process'
|
||||
import * as url from 'node:url'
|
||||
|
||||
export const ROOT_DIR: string = url.fileURLToPath(new URL('../../', import.meta.url))
|
||||
export const BENCHMARKS_DIR: string = url.fileURLToPath(new URL('../', import.meta.url))
|
||||
|
||||
/**
|
||||
* Benchmark execution configuration
|
||||
*/
|
||||
|
||||
/**
|
||||
* Enable dry run mode for quick testing with limited AI requests
|
||||
*
|
||||
* @remarks
|
||||
* Set via environment variable: `DRY_RUN=true`
|
||||
*/
|
||||
export const DRY_RUN: boolean = process.env.DRY_RUN === 'true'
|
||||
|
||||
/**
|
||||
* Limits applied when DRY_RUN is enabled
|
||||
*/
|
||||
export const DRY_RUN_LIMITS = {
|
||||
/** Maximum number of questions to evaluate */
|
||||
maxQuestions: 10,
|
||||
/** Maximum number of formats to test */
|
||||
maxFormats: undefined as number | undefined,
|
||||
/** Models to use in dry run */
|
||||
allowedModels: [] as string[],
|
||||
}
|
||||
|
||||
/**
|
||||
* Default concurrency for parallel evaluations
|
||||
*/
|
||||
export const DEFAULT_CONCURRENCY = 20
|
||||
|
||||
/**
|
||||
* Delay between API requests to avoid rate limiting (in milliseconds)
|
||||
*/
|
||||
export const RATE_LIMIT_DELAY_MS = 100
|
||||
Reference in New Issue
Block a user