Skip to content
Snippets Groups Projects
Commit be1e8306 authored by Benoit Favre's avatar Benoit Favre
Browse files

add llama models

parent bec285c0
No related branches found
No related tags found
No related merge requests found
Showing
with 148 additions and 0 deletions
{
"base_model_name_or_path": "decapoda-research/llama-13b-hf",
"bias": "none",
"enable_lora": null,
"fan_in_fan_out": false,
"inference_mode": true,
"lora_alpha": 16,
"lora_dropout": 0.05,
"merge_weights": false,
"modules_to_save": null,
"peft_type": "LORA",
"r": 4,
"target_modules": [
"q_proj",
"v_proj"
],
"task_type": "CAUSAL_LM"
}
\ No newline at end of file
File added
{
"backbone": "decapoda-research/llama-13b-hf",
"micro_batch_size": 12,
"batch_size": 24,
"epochs": 1,
"learning_rate": 0.0003,
"cutoff_len": 256,
"warmup_ratio": 0.05,
"llama_variant": 13,
"is_int8": true,
"lora_r": 4,
"lora_alpha": 16,
"lora_dropout": 0.05,
"prompt": "Ceci est une question de QCM de l'examen de pharmacie. R\u00e9ponds avec la ou les lettres correspondant \u00e0 la bonne r\u00e9ponse.\n\n%s",
"output_path": "models/deft_llama-13b-hf_lora_80c5a900-3771-415e-941f-052fe1417ce9"
}
\ No newline at end of file
{
"base_model_name_or_path": "decapoda-research/llama-30b-hf",
"bias": "none",
"enable_lora": null,
"fan_in_fan_out": false,
"inference_mode": true,
"lora_alpha": 16,
"lora_dropout": 0.05,
"merge_weights": false,
"modules_to_save": null,
"peft_type": "LORA",
"r": 4,
"target_modules": [
"q_proj",
"v_proj"
],
"task_type": "CAUSAL_LM"
}
\ No newline at end of file
File added
{
"backbone": "decapoda-research/llama-30b-hf",
"micro_batch_size": 6,
"batch_size": 24,
"epochs": 1,
"learning_rate": 0.0003,
"cutoff_len": 256,
"warmup_ratio": 0.05,
"llama_variant": 30,
"is_int8": true,
"lora_r": 4,
"lora_alpha": 16,
"lora_dropout": 0.05,
"prompt": "Ceci est une question de QCM de l'examen de pharmacie. R\u00e9ponds avec la ou les lettres correspondant \u00e0 la bonne r\u00e9ponse.\n\n%s",
"output_path": "models/deft_llama-30b-hf_lora_64b3606a-ffa2-44b7-90df-fa942884e8cb"
}
\ No newline at end of file
{
"base_model_name_or_path": "decapoda-research/llama-65b-hf",
"bias": "none",
"enable_lora": null,
"fan_in_fan_out": false,
"inference_mode": true,
"lora_alpha": 16,
"lora_dropout": 0.05,
"merge_weights": false,
"modules_to_save": null,
"peft_type": "LORA",
"r": 4,
"target_modules": [
"q_proj",
"v_proj"
],
"task_type": "CAUSAL_LM"
}
\ No newline at end of file
File added
{
"backbone": "decapoda-research/llama-65b-hf",
"micro_batch_size": 1,
"batch_size": 24,
"epochs": 1,
"learning_rate": 0.0003,
"cutoff_len": 256,
"warmup_ratio": 0.05,
"llama_variant": 65,
"is_int8": true,
"lora_r": 4,
"lora_alpha": 16,
"lora_dropout": 0.05,
"prompt": "Ceci est une question de QCM de l'examen de pharmacie. R\u00e9ponds avec la ou les lettres correspondant \u00e0 la bonne r\u00e9ponse.\n\n%s",
"output_path": "models/deft_llama-65b-hf_lora_98075de5-9200-4d66-ab35-61ca2a380692"
}
\ No newline at end of file
{
"base_model_name_or_path": "decapoda-research/llama-7b-hf",
"bias": "none",
"enable_lora": null,
"fan_in_fan_out": false,
"inference_mode": true,
"lora_alpha": 16,
"lora_dropout": 0.05,
"merge_weights": false,
"modules_to_save": null,
"peft_type": "LORA",
"r": 4,
"target_modules": [
"q_proj",
"v_proj"
],
"task_type": "CAUSAL_LM"
}
\ No newline at end of file
File added
{
"backbone": "decapoda-research/llama-7b-hf",
"micro_batch_size": 24,
"batch_size": 24,
"epochs": 1,
"learning_rate": 0.0003,
"cutoff_len": 256,
"warmup_ratio": 0.05,
"llama_variant": 7,
"is_int8": true,
"lora_r": 4,
"lora_alpha": 16,
"lora_dropout": 0.05,
"prompt": "Ceci est une question de QCM de l'examen de pharmacie. R\u00e9ponds avec la ou les lettres correspondant \u00e0 la bonne r\u00e9ponse.\n\n%s",
"output_path": "models/deft_llama-7b-hf_lora_df327e03-1db8-4136-8cd0-af2776db875f"
}
\ No newline at end of file
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment