| | --- |
| | license: apache-2.0 |
| | datasets: |
| | - vicgalle/alpaca-gpt4 |
| | language: |
| | - en |
| | --- |
| | This is a finetuned version microsoft phi - 2. This is finetuned using the alpaca dataset. |
| |
|
| |
|
| | ## Training arguments |
| |
|
| | num_train_epochs=1 |
| |
|
| | per_device_train_batch_size=1 |
| |
|
| | gradient_accumulation_steps=8 |
| |
|
| | optim="paged_adamw_32bit" |
| |
|
| | logging_steps=25 |
| | |
| | learning_rate=2e-4 |
| |
|
| | weight_decay=0.001 |
| | |
| | max_grad_norm=0.3 |
| | |
| | warmup_ratio=0.03 |
| |
|
| | lr_scheduler_type="cosine" |
| |
|
| | ## Prompt format: |
| |
|
| | Use the below format |
| |
|
| | ``` |
| | Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction:{prompt} ### Response:{generated_output} |
| | ``` |
| |
|