| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| from typing import TYPE_CHECKING |
|
|
| from transformers.utils import ( |
| OptionalDependencyNotAvailable, |
| _LazyModule, |
| is_sentencepiece_available, |
| is_tokenizers_available, |
| is_torch_available, |
| ) |
|
|
|
|
| _import_structure = { |
| "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], |
| } |
|
|
| try: |
| if not is_sentencepiece_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| _import_structure["tokenization_llama"] = ["LlamaTokenizer"] |
|
|
| try: |
| if not is_tokenizers_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| _import_structure["tokenization_llama_fast"] = ["LlamaTokenizerFast"] |
|
|
| try: |
| if not is_torch_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| _import_structure["modeling_llama"] = [ |
| "LlamaForCausalLM", |
| "LlamaModel", |
| "LlamaPreTrainedModel", |
| "LlamaForSequenceClassification", |
| ] |
|
|
|
|
| if TYPE_CHECKING: |
| from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig |
|
|
| try: |
| if not is_sentencepiece_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| from .tokenization_llama import LlamaTokenizer |
|
|
| try: |
| if not is_tokenizers_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| from .tokenization_llama_fast import LlamaTokenizerFast |
|
|
| try: |
| if not is_torch_available(): |
| raise OptionalDependencyNotAvailable() |
| except OptionalDependencyNotAvailable: |
| pass |
| else: |
| from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel |
|
|
|
|
| else: |
| import sys |
|
|
| sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) |
|
|