Spaces:
Running on Zero
Running on Zero
test2
#27
by
MILOPS - opened
- env.py +0 -5
- requirements.txt +2 -3
env.py
CHANGED
|
@@ -30,8 +30,6 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 30 |
'BlueDancer/Artisanica_XL',
|
| 31 |
'neta-art/neta-noob-1.0',
|
| 32 |
'OnomaAIResearch/Illustrious-xl-early-release-v0',
|
| 33 |
-
'martineux/bismuth7-xl',
|
| 34 |
-
'martineux/perfectdeliberate8',
|
| 35 |
'Raelina/Rae-Diffusion-XL-V2',
|
| 36 |
'Raelina/Raemu-XL-V4',
|
| 37 |
'Raelina/Raemu-XL-V5',
|
|
@@ -47,9 +45,6 @@ LOAD_DIFFUSERS_FORMAT_MODEL = [
|
|
| 47 |
'Raelina/Raehoshi-illust-XL-7',
|
| 48 |
'Raelina/Raehoshi-illust-XL-7.1',
|
| 49 |
'Raelina/Raehoshi-illust-XL-8',
|
| 50 |
-
'Raelina/Raehoshi-illust-XL-8.1',
|
| 51 |
-
'Raelina/Raehoshi-illust-XL-9',
|
| 52 |
-
'Raelina/Raehoshi-illust-vpred',
|
| 53 |
'camenduru/FLUX.1-dev-diffusers',
|
| 54 |
'black-forest-labs/FLUX.1-schnell',
|
| 55 |
'sayakpaul/FLUX.1-merged',
|
|
|
|
| 30 |
'BlueDancer/Artisanica_XL',
|
| 31 |
'neta-art/neta-noob-1.0',
|
| 32 |
'OnomaAIResearch/Illustrious-xl-early-release-v0',
|
|
|
|
|
|
|
| 33 |
'Raelina/Rae-Diffusion-XL-V2',
|
| 34 |
'Raelina/Raemu-XL-V4',
|
| 35 |
'Raelina/Raemu-XL-V5',
|
|
|
|
| 45 |
'Raelina/Raehoshi-illust-XL-7',
|
| 46 |
'Raelina/Raehoshi-illust-XL-7.1',
|
| 47 |
'Raelina/Raehoshi-illust-XL-8',
|
|
|
|
|
|
|
|
|
|
| 48 |
'camenduru/FLUX.1-dev-diffusers',
|
| 49 |
'black-forest-labs/FLUX.1-schnell',
|
| 50 |
'sayakpaul/FLUX.1-merged',
|
requirements.txt
CHANGED
|
@@ -3,7 +3,7 @@ diffusers
|
|
| 3 |
transformers
|
| 4 |
accelerate
|
| 5 |
huggingface_hub
|
| 6 |
-
|
| 7 |
hf_xet
|
| 8 |
torch==2.5.1
|
| 9 |
torchvision
|
|
@@ -19,6 +19,5 @@ rapidfuzz
|
|
| 19 |
sentencepiece
|
| 20 |
unidecode
|
| 21 |
matplotlib-inline
|
| 22 |
-
mediapipe==0.10.5
|
| 23 |
https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
|
| 24 |
-
|
|
|
|
| 3 |
transformers
|
| 4 |
accelerate
|
| 5 |
huggingface_hub
|
| 6 |
+
hf_transfer
|
| 7 |
hf_xet
|
| 8 |
torch==2.5.1
|
| 9 |
torchvision
|
|
|
|
| 19 |
sentencepiece
|
| 20 |
unidecode
|
| 21 |
matplotlib-inline
|
|
|
|
| 22 |
https://github.com/mjun0812/flash-attention-prebuild-wheels/releases/download/v0.4.11/flash_attn-2.8.3+cu124torch2.5-cp310-cp310-linux_x86_64.whl
|
| 23 |
+
pydantic==2.10.6
|