Spaces:
Build error
Build error
Fetching metadata from the HF Docker repository...
Ctrl+K
- PY3 Upload 76 files
- 6.15 kB Upload 76 files
- 8.57 kB Upload 76 files
- czech.pickle1.27 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object"
xetUpload 76 files - danish.pickle1.26 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "__builtin__.object",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktParameters",
- "copy_reg._reconstructor"
xetUpload 76 files - dutch.pickle743 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.int",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor"
xetUpload 76 files - english.pickle433 kB
Detected Pickle imports (9)
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken"
xetUpload 76 files - estonian.pickle1.6 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters"
xetUpload 76 files - finnish.pickle1.95 MB
Detected Pickle imports (9)
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "__builtin__.int"
xetUpload 76 files - french.pickle583 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object"
xetUpload 76 files - german.pickle1.53 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.int"
xetUpload 76 files - greek.pickle1.95 MB
Detected Pickle imports (9)
- "__builtin__.object",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int"
xetUpload 76 files - italian.pickle658 kB
Detected Pickle imports (9)
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object"
xetUpload 76 files - malayalam.pickle221 kB
Detected Pickle imports (7)
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.int",
- "__builtin__.set",
- "collections.defaultdict"
xetUpload 76 files - norwegian.pickle1.26 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.object",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "copy_reg._reconstructor"
xetUpload 76 files - polish.pickle2.04 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "collections.defaultdict",
- "copy_reg._reconstructor"
xetUpload 76 files - portuguese.pickle649 kB
Detected Pickle imports (9)
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer"
xetUpload 76 files - russian.pickle33 kB
Detected Pickle imports (7)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.long",
- "nltk.tokenize.punkt.PunktToken"
xetUpload 76 files - slovene.pickle833 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "__builtin__.object"
xetUpload 76 files - spanish.pickle598 kB
Detected Pickle imports (9)
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars"
xetUpload 76 files - swedish.pickle1.03 MB
Detected Pickle imports (9)
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "copy_reg._reconstructor",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken"
xetUpload 76 files - turkish.pickle1.23 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "copy_reg._reconstructor",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "__builtin__.object",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars"
xetUpload 76 files