From d610507348a5c7038b0d66cbaa89358d096731e3 Mon Sep 17 00:00:00 2001 From: teowu <2711123841@qq.com> Date: Sun, 29 Sep 2024 11:36:59 +0800 Subject: [PATCH] add training recipe & related LMMs --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 0b2fcf6..7a2dc92 100644 --- a/README.md +++ b/README.md @@ -132,6 +132,8 @@ For training, please refer to the [Q-Align](https://github.com/Q-Future/Q-Align) ```shell #!/bin/bash +# Use 8 GPUs to replicate the training + LOAD='MAGAer13/mplug-owl2-llama2-7b' echo 'Converting data format...' @@ -149,11 +151,11 @@ deepspeed --master_port 25801 q_align/train/train_mem.py \ --image_aspect_ratio pad \ --group_by_modality_length True \ --bf16 True \ - --output_dir ./qinstruct_v0.3 \ + --output_dir ./coinstruct_replicated \ --num_train_epochs 1 \ --per_device_train_batch_size 24 \ --per_device_eval_batch_size 4 \ - --gradient_accumulation_steps 2 \ + --gradient_accumulation_steps 1 \ --evaluation_strategy "no" \ --save_strategy "steps" \ --save_steps 1100 \