We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llamafactory
FORCE_TORCHRUN=0 CUDA_VISIBLE_DEVICES=1 llamafactory-cli train ../qwen_pretrain.yaml
qwen_pretrain.yaml
### model model_name_or_path: /home/s-duy20/qwen trust_remote_code: true ### method stage: pt do_train: true finetuning_type: lora lora_target: all ### dataset dataset: pretrain cutoff_len: 500 max_samples: 127 overwrite_cache: true # preprocessing_num_workers: 1 ### output output_dir: /home/s-duy20/saves/qwen/lora/pretrain logging_steps: 10 save_steps: 500 plot_loss: true overwrite_output_dir: true ### train per_device_train_batch_size: 1 gradient_accumulation_steps: 8 learning_rate: 5.0e-5 num_train_epochs: 3.0 lr_scheduler_type: cosine warmup_ratio: 0.1 bf16: true # ddp_timeout: 180000000 lora_rank: 7 ### eval val_size: 0.05 per_device_eval_batch_size: 1 eval_strategy: steps eval_steps: 500
[INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file vocab.json [INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file merges.txt [INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file tokenizer.json [INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file added_tokens.json [INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file special_tokens_map.json [INFO|tokenization_utils_base.py:2204] 2025-01-17 11:01:20,810 >> loading file tokenizer_config.json [INFO|tokenization_utils_base.py:2470] 2025-01-17 11:01:21,138 >> Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. [INFO|2025-01-17 11:01:21] llamafactory.data.loader:157 >> Loading dataset pretrain.json... Generating train split: 0 examples [00:00, ? examples/s]Killed
从日志中可以看到无任何报错信息,直接被Killed了,请问这是怎么回事?怎么解决?
No response
The text was updated successfully, but these errors were encountered:
观察下内存吧,只能是内存问题
Sorry, something went wrong.
hiyouga
No branches or pull requests
Reminder
System Info
llamafactory-cli env
llamafactory
version: 0.9.2.dev0Reproduction
qwen_pretrain.yaml
中的配置为:RAM内存空间剩余有100GB左右(足够大),显存约12GB,磁盘空间足够大。
从日志中可以看到无任何报错信息,直接被Killed了,请问这是怎么回事?怎么解决?
Others
No response
The text was updated successfully, but these errors were encountered: