| #!/bin/sh |
| |
| |
| |
| |
| |
| |
|
|
| QUERY_PATH="query_list.txt" |
|
|
| query_name=$(sed -n "${SLURM_ARRAY_TASK_ID}p" $QUERY_PATH) |
|
|
| echo "processing ${query_name}" |
|
|
| DATASET="/scratch/ssd004/datasets/cellxgene/scb_strict/${query_name}/all_counts" |
| VOCAB_PATH="/scratch/ssd004/datasets/cellxgene/scFormer/scformer/tokenizer/default_census_vocab.json" |
| bash ~/.bashrc |
|
|
| NPROC=$SLURM_GPUS_ON_NODE |
| JOB_NAME="cellxgene_census_${QUERY_NAME}" |
| LOG_INTERVAL=2000 |
| VALID_SIZE_OR_RATIO=0.03 |
| MAX_LENGTH=1200 |
| per_proc_batch_size=32 |
| LAYERS=12 |
| MODEL_SCALE=8 |
| SAVE_DIR="/scratch/ssd004/datasets/cellxgene/profile_tmp" |
| |
| alias python_=~/.cache/pypoetry/virtualenvs/scformer-9yG_XnDJ-py3.9/bin/python |
| python_ -c "import torch; print(torch.version.cuda)" |
| python_ process_allcounts.py \ |
| --data-source $DATASET \ |
| --save-dir ${SAVE_DIR}/${JOB_NAME}-$(date +%b%d-%H-%M-%Y) \ |
| --vocab-path ${VOCAB_PATH} \ |
| --valid-size-or-ratio $VALID_SIZE_OR_RATIO \ |
| --max-seq-len $MAX_LENGTH \ |
| --batch-size $per_proc_batch_size \ |
| --eval-batch-size $(($per_proc_batch_size * 2)) \ |
| --nlayers $LAYERS \ |
| --nheads 8 \ |
| --embsize $((MODEL_SCALE * 64)) \ |
| --d-hid $((MODEL_SCALE * 64)) \ |
| --grad-accu-steps $((128 / $per_proc_batch_size)) \ |
| --epochs 2 \ |
| --lr 0.0001 \ |
| --warmup-ratio-or-step 10000 \ |
| --log-interval $LOG_INTERVAL \ |
| --save-interval $(($LOG_INTERVAL * 3)) \ |
| --trunc-by-sample \ |
| --no-cls \ |
| --no-cce \ |
| --fp16 | |
| awk '{ print strftime("[%Y-%m-%d %H:%M:%S]"), $0; fflush(); }' |
|
|