Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama factory llama2-13b and mistral-7b pipeline #93

Merged
merged 21 commits into from
Nov 12, 2023

Conversation

Jasonqi146
Copy link
Member

@Jasonqi146 Jasonqi146 commented Nov 8, 2023

Closes #

πŸ“‘ Description

Trying to fix llama-factory convergence by replicating fastchat

βœ… Checks

  • My pull request adheres to the code style of this project
  • My code requires changes to the documentation
  • I have updated the documentation as required
  • All the tests have passed
  • Branch name follows type/descript (e.g. feature/add-llm-agents)
  • Ready for code review

β„Ή Additional Information

@Jasonqi146 Jasonqi146 changed the title Feature/llama factory llama2 pipeline llama factory llama2-13b and mistral-7b pipeline Nov 12, 2023
@Jasonqi146 Jasonqi146 merged commit cb93c7c into main Nov 12, 2023
3 checks passed
ruiyiw pushed a commit that referenced this pull request Nov 13, 2023
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
ruiyiw pushed a commit that referenced this pull request Nov 14, 2023
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Nov 16, 2023
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Nov 17, 2023
lwaekfjlk pushed a commit that referenced this pull request Nov 17, 2023
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 13, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
Signed-off-by: Haofei Yu <1125027232@qq.com>
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
Signed-off-by: Haofei Yu <1125027232@qq.com>
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral

(cherry picked from commit 0c53e37)
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
lwaekfjlk pushed a commit that referenced this pull request Mar 14, 2024
* added llama-factory under llm_rl

* added sft training bash

* added datasets from llama-factory; will delete later

* finished llama-2-13b train and inference

* fixed minor errors

* changed config

* added deepspeed config

* added more training config to train bash

* adding fix for wandb tags and distributed ranks

* added fastchat data to replicate training for 2k

* tyring to replicate fastchat as close as possible

* before merging

* changed finetue scripts for better performance

* added new data

* example bash

* example bash for mistral
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant