-
Notifications
You must be signed in to change notification settings - Fork 0
/
test_config.yml
45 lines (43 loc) · 1.5 KB
/
test_config.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
loader:
class_path: smileval.loaders.mcq.MCQQuestionLoader
init_args:
input_file_path: test5.csv
use_example: false
use_shuffle: true
model:
# class_path: smileval.models.ollama.OllamaChatCompletionModel
class_path: smileval.models.openai.OpenAIChatCompletionModel
# class_path: smileval.models.openai.OpenAITextCompletionModel
init_args:
# name: "phi3:3.8b-mini-instruct-4k-fp16" #"doesntmatterforllamacppserver"
# name: "llama3.1:8b"
name: "hf:NousResearch/Nous-Hermes-2-Yi-34B"
# tokenizer: https://huggingface.co/microsoft/Phi-3-medium-128k-instruct/blob/main/tokenizer_config.json
# name: "openchat/openchat-7b:free" # :8b-instruct-fp16
# ensures we don't artifically run out of context
# spoof_api_name: "gpt-4-32k"
# you may find this useful to comment out to use env variables
# api_key: sk-example
# use this if non-openai st andard api json body parameters are supported in the POST request, (hint: llama.cpp or ollama)
extended: true
# base_url: "https://openrouter.ai/api/v1"
namespace: mcq_v5
id: glhfchat-Nous-Hermes-2-Yi-34B
seed: 69424269
parallel: 1
sleep: 0
options:
class_path: smileval.models.ChatCompletionOptions
dict_kwargs:
use_system_prompt_workaround: false
max_tokens: 1024 # too much model yapping is also bad
# mirostat: 2
#stop_tokens:
# - "\n"
# phi moment
# - "<|end|>"
# - "<|eot_id|>"
# command R broken in llama.cpp moment
# also yi be like
#stop_tokens:
# - "<|im_end|>"