Loading finetune_hoc.lsf +7 −7 Original line number Diff line number Diff line #!/bin/bash #BSUB -nnodes 2 #BSUB -W 0:30 #BSUB -W 2:00 #BSUB -P med106 #BSUB -alloc_flags "smt4 nvme" #BSUB -J hoc #BSUB -o hoc.%J #BSUB -J hoc_FULL #BSUB -o hoc_FULL.%J #BSUB -q batch set +x Loading @@ -27,15 +27,15 @@ echo $nnodes #export TRAIN_DATA=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/picodata/train.tsv #TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata/train.tsv" TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata_small" TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata" #export TRAIN_DATA=picodata/train.tsv #export VALID_DATA=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/picodata/dev.tsv #VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata/dev.tsv" VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata_small" VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata" #export VALID_DATA=picodata/dev.tsv export VOCAB_FILE=/gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt export CHECKPOINT_PATH=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/chkpt_222 export CHECKPOINT_PATH=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/finetune-HOC_BIG export PRETRAINED_CHECKPOINT=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/chkptt jsrun --smpiargs="-disable_gpu_hooks" -n $nnodes -r 1 -g 6 -a 6 -c 42 python tasks/main.py \ Loading @@ -60,7 +60,7 @@ jsrun --smpiargs="-disable_gpu_hooks" -n $nnodes -r 1 -g 6 -a 6 -c 42 python tas --eval-iters 10 \ --weight-decay 1e-2 \ --tokenizer-type BertWordPieceLowerCase \ --epochs 1 \ --epochs 5 \ --micro-batch-size 4 \ --lr 0.0001 \ --lr-warmup-fraction 0.06 \ Loading preprocess_data/preprocess.lsf +2 −2 Original line number Diff line number Diff line Loading @@ -15,6 +15,6 @@ conda deactivate conda activate /gpfs/alpine/med106/world-shared/irl1/rhel8/mytorch export OMP_NUM_THREADS=1 export PYTHONPATH=$PYTHONPATH:/gpfs/alpine/med106/world-shared/irl1/rhel8/Megatron-LM export PYTHONPATH=$PYTHONPATH:/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron jsrun python ../tools/preprocess_data.py --input /gpfs/alpine/med106/world-shared/irl1/rhel8/Megatron-LM/preprocess_data/pm014.json --output-prefix pm014 --vocab /gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt --dataset-impl mmap --tokenizer-type BertWordPieceLowerCase --split-sentences --json-keys abstract --workers 42 jsrun python ../tools/preprocess_data.py --input /gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/preprocess_data/pm000.json --output-prefix pm000 --vocab /gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt --dataset-impl mmap --tokenizer-type BertWordPieceLowerCase --split-sentences --json-keys abstract --workers 42 pretrain_pubmed.lsf +123 −120 File changed.Preview size limit exceeded, changes collapsed. Show changes Loading
finetune_hoc.lsf +7 −7 Original line number Diff line number Diff line #!/bin/bash #BSUB -nnodes 2 #BSUB -W 0:30 #BSUB -W 2:00 #BSUB -P med106 #BSUB -alloc_flags "smt4 nvme" #BSUB -J hoc #BSUB -o hoc.%J #BSUB -J hoc_FULL #BSUB -o hoc_FULL.%J #BSUB -q batch set +x Loading @@ -27,15 +27,15 @@ echo $nnodes #export TRAIN_DATA=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/picodata/train.tsv #TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata/train.tsv" TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata_small" TRAIN_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata" #export TRAIN_DATA=picodata/train.tsv #export VALID_DATA=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/picodata/dev.tsv #VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata/dev.tsv" VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata_small" VALID_DATA="/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/hocdata" #export VALID_DATA=picodata/dev.tsv export VOCAB_FILE=/gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt export CHECKPOINT_PATH=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/chkpt_222 export CHECKPOINT_PATH=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/finetune-HOC_BIG export PRETRAINED_CHECKPOINT=/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/chkptt jsrun --smpiargs="-disable_gpu_hooks" -n $nnodes -r 1 -g 6 -a 6 -c 42 python tasks/main.py \ Loading @@ -60,7 +60,7 @@ jsrun --smpiargs="-disable_gpu_hooks" -n $nnodes -r 1 -g 6 -a 6 -c 42 python tas --eval-iters 10 \ --weight-decay 1e-2 \ --tokenizer-type BertWordPieceLowerCase \ --epochs 1 \ --epochs 5 \ --micro-batch-size 4 \ --lr 0.0001 \ --lr-warmup-fraction 0.06 \ Loading
preprocess_data/preprocess.lsf +2 −2 Original line number Diff line number Diff line Loading @@ -15,6 +15,6 @@ conda deactivate conda activate /gpfs/alpine/med106/world-shared/irl1/rhel8/mytorch export OMP_NUM_THREADS=1 export PYTHONPATH=$PYTHONPATH:/gpfs/alpine/med106/world-shared/irl1/rhel8/Megatron-LM export PYTHONPATH=$PYTHONPATH:/gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron jsrun python ../tools/preprocess_data.py --input /gpfs/alpine/med106/world-shared/irl1/rhel8/Megatron-LM/preprocess_data/pm014.json --output-prefix pm014 --vocab /gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt --dataset-impl mmap --tokenizer-type BertWordPieceLowerCase --split-sentences --json-keys abstract --workers 42 jsrun python ../tools/preprocess_data.py --input /gpfs/alpine/med106/world-shared/irl1/rhel8/fork-megatron/preprocess_data/pm000.json --output-prefix pm000 --vocab /gpfs/alpine/world-shared/med106/g8o/pubmed_bert-vocab.txt --dataset-impl mmap --tokenizer-type BertWordPieceLowerCase --split-sentences --json-keys abstract --workers 42
pretrain_pubmed.lsf +123 −120 File changed.Preview size limit exceeded, changes collapsed. Show changes