Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- # Comment to be included in the beginning of the final report.
- comment = ""
- [server]
- url = "http://localhost:11434/v1"
- api_key = "api key"
- model = "Qwen2.5-32B-Instruct-32k-Q3_K_S:latest"
- timeout = 600.0
- [inference]
- # Ssettings below are from evaluate_from_local.py for VLLM on TIGER-AI-Lab/MMLU-Pro
- temperature = 0.0
- top_p = 1.0 # not specified but default for VLLM
- max_tokens = 2048
- # The variable {subject} will be replaced with appropriate value in runtime.
- system_prompt = "The following are multiple choice questions (with answers) about {subject}. Think step by step and then finish your answer with \"the answer is (X)\" where X is the correct letter choice."
- # "multi_chat" inserts COT examples into multi-turn messages. Use for instruct/chat models.
- # "no_chat" uses v1/completion api. Use for non-instruct/chat model.
- # "single_chat" (from the script for GPT-4O) inserts all the COT examples and question into a single message. Not recommended, use only for legacy compatibility.
- style = "multi_chat"
- [test]
- categories = ['biology', 'business', 'chemistry', 'computer science', 'economics', 'engineering', 'health', 'history', 'law', 'math', 'philosophy', 'physics', 'psychology', 'other']
- parallel = 1
- [log]
- # Verbosity between 0-2
- verbosity = 0
- # If true, logs exact prompt sent to the model in the test result files.
- log_prompt = true
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement