本セッションでは、W&B Coursesの中でも最も人気の高いコースである"Effective MLOps: Model Development (日本語字幕版コース名: 効果的なMLOps: モデル開発)"をギュッと濃縮したダイジェスト版を日本語ハンズオンでお届けいたします。W&Bの基本的な使い方、ベースラインからの改良方法などをシンプルな画像のセグメンテーションタスクを通じて学ぶことができます。
https://wandb.connpass.com/event/295345/
30. wandb:
log: True
entity: "wandb-japan"
project: "llm-leaderboard"
run_name: 'mistralai/Mistral-7B-Instruct-v0.2' # use run_name defined above
github_version: v2.0.0 #for recording
testmode: true
# if you don't use api, please set "api" as "false"
# if you use api, please select from "openai", "anthoropic", "google", "cohere"
api: false
model:
use_wandb_artifacts: false
artifacts_path: ""
pretrained_model_name_or_path: 'mistralai/Mistral-7B-Instruct-v0.2' #if you use openai api, put the name of
model
trust_remote_code: true
device_map: "auto"
load_in_8bit: false
load_in_4bit: false
generator:
top_p: 1.0
top_k: 0
temperature: 0.1
repetition_penalty: 1.0
tokenizer:
use_wandb_artifacts: false
artifacts_path: ""
pretrained_model_name_or_path: "mistralai/Mistral-7B-Instruct-v0.2"
use_fast: true
config.yamlの設定 (概要、モデルとトークナイザ)
31. # for llm-jp-eval
max_seq_length: 2048
dataset_artifact: "wandb-japan/llm-leaderboard/jaster:v3" #if you use artifacts, please fill here (if not, fill
null)
dataset_dir: "/jaster/1.1.0/evaluation/test"
target_dataset: "all" # {all, jamp, janli, jcommonsenseqa, jemhopqa, jnli, jsem, jsick, jsquad, jsts, niilc,
chabsa}
log_dir: "./logs"
torch_dtype: "bf16" # {fp16, bf16, fp32}
custom_prompt_template: "<s> [INST] {instruction}n{input}[/INST]"
custom_fewshots_template: null
# Please include {input} and {output} as variables
# example of fewshots template
# "n### 入力:n{input}n### 回答:n{output}"
metainfo:
basemodel_name: "mistralai/Mistral-7B-Instruct-v0.2"
model_type: "open llm" # {open llm, commercial api}
instruction_tuning_method: "None" # {"None", "Full", "LoRA", ...}
instruction_tuning_data: ["None"] # {"None", "jaster", "dolly_ja", "oasst_ja", ...}
num_few_shots: 0
llm-jp-eval-version: "1.1.0"
config.yamlの設定 (llm-jp-eval)
32. # for mtbench
mtbench:
question_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_question:v0' # if testmode is true, small
dataset will be used
referenceanswer_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_referenceanswer:v0' # if testmode is
true, small dataset will be used
judge_prompt_artifacts_path: 'wandb-japan/llm-leaderboard/mtbench_ja_prompt:v1'
bench_name: 'japanese_mt_bench'
model_id: null # cannot use '<', '>', ':', '"', '/', '', '|', '?', '*', '.'
question_begin: null
question_end: null
max_new_token: 1024
num_choices: 1
num_gpus_per_model: 1
num_gpus_total: 1
max_gpu_memory: null
dtype: bfloat16 # None or float32 or float16 or bfloat16
# for gen_judgment
judge_model: 'gpt-4'
mode: 'single'
baseline_model: null
parallel: 1
first_n: null
# for conv template # added
custom_conv_template: true
# the following variables will be used when custom_conv_template is set as true
conv_name: "custom"
conv_system_message: ""
conv_roles: "('[INST]', '[/INST]')"
conv_sep: "</s> "
conv_stop_token_ids: "[2]"
conv_stop_str: "</s> "
conv_role_message_separator: " "
conv_role_only_separator: " "
config.yamlの設定 (Japanese MT-Bench)