-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Experiment config and results for gpt-4-0125-preview, gpt-4-1106-prev…
…iew and gpt-3.5-turbo-1106
- Loading branch information
Showing
5 changed files
with
258 additions
and
0 deletions.
There are no files selected for viewing
186 changes: 186 additions & 0 deletions
186
...mation-api/yival_experiments/experiment_configurations/experiment_202401292237_en-US.yaml
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,186 @@ | ||
custom_wrappers: | ||
model_config_wrapper: | ||
class: ./custom_configuration/model_config_wrapper.ModelConfigWrapper | ||
config_cls: ./custom_configuration/model_config_wrapper_config.ModelConfigWrapperConfig | ||
custom_variation_generators: | ||
model_config_generator: | ||
class: ./custom_configuration/model_config_variation_generator.ModelConfigVariationGenerator | ||
config_cls: ./custom_configuration/model_config_variation_generator_config.ModelConfigVariationGeneratorConfig | ||
custom_evaluators: | ||
gpt4_evaluator: | ||
class: ./custom_configuration/gpt4_evaluator.GPT4Evaluator | ||
config_cls: ./custom_configuration/gpt4_evaluator_config.GPT4EvaluatorConfig | ||
custom_function: model_compare.model_compare | ||
dataset: | ||
source_type: dataset | ||
file_path: data/questions_en-US.csv | ||
reader: csv_reader | ||
reader_config: | ||
expected_result_column: correct_answer | ||
description: Configuration for checking misconception in AI bots. | ||
evaluators: | ||
- evaluator_type: individual | ||
metric_calculators: | ||
- method: AVERAGE | ||
name: gpt4_evaluator | ||
model_name: gpt-4 | ||
prompt: |- | ||
You are assessing a submitted answer on a given task based on a criterion. Here is the data: | ||
- Task: Answer a question. | ||
- Is the answer correct? | ||
[Input]: {question_text} | ||
Possible Options and Correctness of Options: | ||
A. {option_a} ({option_a_correctness}) | ||
B. {option_b} ({option_b_correctness}) | ||
C. {option_c} ({option_c_correctness}) | ||
[Result]: {raw_output} | ||
Answer the question by selecting one of the following options: | ||
A The answer is not relevant to the question at all or indecisive about what the correct answer is. | ||
B The answer is not correct, and matches/is close to the Very Wrong answer. | ||
C The answer is not correct, and matches/is close to the Wrong answer. | ||
D The answer is correct, i.e. matches the Correct answer. | ||
choices: | ||
- A | ||
- B | ||
- C | ||
- D | ||
description: Is the answer correct? | ||
choice_scores: | ||
A: 0 | ||
B: 1 | ||
C: 2 | ||
D: 3 | ||
scale_description: 0-3 | ||
display_name: correctness | ||
variations: | ||
- name: model_config | ||
generator_name: model_config_generator | ||
generator_config: | ||
models: | ||
- vendor: OpenAI | ||
model_id: gpt-4-1106-preview | ||
params: | ||
temperature: 0.01 | ||
round: 1 | ||
- vendor: OpenAI | ||
model_id: gpt-4-1106-preview | ||
params: | ||
temperature: 0.01 | ||
round: 2 | ||
- vendor: OpenAI | ||
model_id: gpt-4-1106-preview | ||
params: | ||
temperature: 0.01 | ||
round: 3 | ||
- vendor: OpenAI | ||
model_id: gpt-4-1106-preview | ||
params: | ||
temperature: 0.01 | ||
round: 4 | ||
- vendor: OpenAI | ||
model_id: gpt-4-1106-preview | ||
params: | ||
temperature: 0.01 | ||
round: 5 | ||
- vendor: OpenAI | ||
model_id: gpt-3.5-turbo-1106 | ||
params: | ||
temperature: 0.01 | ||
round: 1 | ||
- vendor: OpenAI | ||
model_id: gpt-3.5-turbo-1106 | ||
params: | ||
temperature: 0.01 | ||
round: 2 | ||
- vendor: OpenAI | ||
model_id: gpt-3.5-turbo-1106 | ||
params: | ||
temperature: 0.01 | ||
round: 3 | ||
- vendor: OpenAI | ||
model_id: gpt-3.5-turbo-1106 | ||
params: | ||
temperature: 0.01 | ||
round: 4 | ||
- vendor: OpenAI | ||
model_id: gpt-3.5-turbo-1106 | ||
params: | ||
temperature: 0.01 | ||
round: 5 | ||
- vendor: OpenAI | ||
model_id: gpt-4-0125-preview | ||
params: | ||
temperature: 0.01 | ||
round: 1 | ||
- vendor: OpenAI | ||
model_id: gpt-4-0125-preview | ||
params: | ||
temperature: 0.01 | ||
round: 2 | ||
- vendor: OpenAI | ||
model_id: gpt-4-0125-preview | ||
params: | ||
temperature: 0.01 | ||
round: 3 | ||
- vendor: OpenAI | ||
model_id: gpt-4-0125-preview | ||
params: | ||
temperature: 0.01 | ||
round: 4 | ||
- vendor: OpenAI | ||
model_id: gpt-4-0125-preview | ||
params: | ||
temperature: 0.01 | ||
round: 5 | ||
- name: prompt_template | ||
variations: | ||
- variation_id: instruct_question_options_1 | ||
value_type: str | ||
instantiated_value: |- | ||
Please answer this multiple choices question. If you can't determine the answer please make your best guess: | ||
Question: | ||
{question_text} | ||
A. {option_a} | ||
B. {option_b} | ||
C. {option_c} | ||
Answer: | ||
value: |- | ||
Please answer this multiple choices question. If you can't determine the answer please make your best guess: | ||
Question: | ||
{question_text} | ||
A. {option_a} | ||
B. {option_b} | ||
C. {option_c} | ||
Answer: | ||
- variation_id: instruct_question_options_2 | ||
value_type: str | ||
instantiated_value: |- | ||
Pick the correct answer: A, B or C. If you can't determine the answer please make your best guess. | ||
Question: | ||
{question_text} | ||
A. {option_a} | ||
B. {option_b} | ||
C. {option_c} | ||
Answer: | ||
value: |- | ||
Pick the correct answer: A, B or C. If you can't determine the answer please make your best guess. | ||
Question: | ||
{question_text} | ||
A. {option_a} | ||
B. {option_b} | ||
C. {option_c} | ||
Answer: | ||
- variation_id: no_option_letter | ||
value_type: str | ||
instantiated_value: "Question:\n{question_text}\nIs it: {option_a}; {option_b};\ | ||
\ or {option_c}? \n\nAnswer:" | ||
value: "Question:\n{question_text}\nIs it: {option_a}; {option_b}; or {option_c}?\ | ||
\ \n\nAnswer:" |
Binary file added
BIN
+3.12 MB
automation-api/yival_experiments/output/experiment_202401292237_en-US_0.pkl
Binary file not shown.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,2 +1,5 @@ | ||
model_id,model_params,correctness_with_indecisive,correctness_without_indecisive | ||
gpt-4-0125-preview,{'temperature': 0.01},79.13669064748201,80.0 | ||
gpt-4-1106-preview,{'temperature': 0.01},79.49640287769785,79.49640287769785 | ||
gemini/gemini-pro,{'temperature': 0.01},54.31654676258992,54.710144927536234 | ||
gpt-3.5-turbo-1106,{'temperature': 0.01},38.1294964028777,38.26714801444043 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Binary file not shown.