-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathconfig.yml
More file actions
80 lines (65 loc) · 2.57 KB
/
config.yml
File metadata and controls
80 lines (65 loc) · 2.57 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
# FactCheck Configuration File
# Dataset configuration
dataset:
name: "YAGO" # Options: DBpedia, YAGO, FactBench
# Method configuration
method:
name: "GIV-F" # Options: DKA, GIV-Z, GIV-F, RAG
# LLM configuration
llm:
# Available models:
# Commercial: gpt-4o-mini and all models available here: https://learn.microsoft.com/en-us/azure/ai-services/openai/concepts/models?tabs=global-standard%2Cstandard-chat-completions
# Open Source: Qwen2.5:7B, Llama3.1:8B, Gemma2:9B, Mistral:7B, QWQ:32B and all models available here: https://ollama.com/library
mode: "open_source" # Options: commercial, open_source
model: "gemma2:9b"
# Model parameters
parameters:
temperature: 0.75
top_p: 0.9
max_tokens: 512
# majority vote
majority_vote:
mode: 'commercial' # Options: commercial, open_source
final_tie_breaker: 'most_consistent' # Options: least_consistent, most_consistent, Null (for commercial)
num_votes: 3 # Number of votes for each model
llms:
- "mistral:7B"
- "qwen2.5:7B"
- "llama3.1:7B"
- "gemma2:9B"
higher_parameter_model:
qwen2.5:7b: 'qwen2.5:7b'
mistral:7b: 'mistral:7b'
llama3.1:7b: 'llama3.1:latest'
gemma2:9b: 'gemma2:9b'
commercial_model:
- "gpt-4o-mini"
- "gpt-4o"
rag:
embedding_model: 'BAAI/bge-small-en-v1.5'
chunking_strategy: 'sliding_window' # Options: fixed_size, sliding_window, small2big
chunk_size: 512 # needed for fixed_size
chunks_small2big: [128, 256, 512] # needed for small2big
window_size: 3 # suggest: choose a number between 2 and 10
similarity_cutoff: 0.3
top_k: 6
OpenAI: # OpenAI API configuration, get your API key from azure portal
azure_endpoint: "" # Your Azure OpenAI endpoint, e.g., https://<your-resource-name>.openai.azure.com/
api_key: "" # Your OpenAI API key,
api_version: "" # API version, e.g., 2023-05-15
# Evaluation configuration
evaluation:
metrics: # Evaluation metrics
accuracy: 'balanced' # Options: balanced, normal -- suggest: balanced
f1_score: "macro" # Options: micro, macro, weighted -- suggest: macro
openlit: true # Whether to use OpenLit for token and real-cost calculation. should be setup by yourself -> https://github.com/openlit/openlit -- recommend docker
# Knowledge Graph configuration
knowledge_graph:
kg_ids: ['5', 'correct_death_00040', '0'] # Comma-separated list of IDs, e.g., ["id1", "id2", "id3"]
# Output configuration
output:
directory: "./results"
# Logging configuration
logging:
level: "INFO" # Options: DEBUG, INFO, WARNING, ERROR, CRITICAL
format: "%(asctime)s - %(name)s - %(levelname)s - %(message)s"