fix: loosen tokenizers pin to resolve dependency conflict with transformers

#1
by vawsgit - opened
Files changed (1) hide show
  1. requirements.txt +1 -1
requirements.txt CHANGED
@@ -14,7 +14,7 @@ pydantic>=2.0.0
14
  # Note: flash-attn must be installed separately - see README for instructions
15
  torch>=2.7.0
16
  transformers>=4.48.0
17
- tokenizers==0.20.3
18
  addict>=2.4.0
19
  easydict>=1.10
20
  einops>=0.6.0
 
14
  # Note: flash-attn must be installed separately - see README for instructions
15
  torch>=2.7.0
16
  transformers>=4.48.0
17
+ tokenizers>=0.20.3
18
  addict>=2.4.0
19
  easydict>=1.10
20
  einops>=0.6.0