Apply flash-attention-3 and pre-load all models (no dynamic reloading) (#10) 361cf94 verified littlebird13 multimodalart HF Staff commited on Jan 29