| [tool.poetry] |
| name = "llama-cpp-scripts" |
| version = "0.0.0" |
| description = "Scripts that ship with llama.cpp" |
| authors = ["GGML <ggml@ggml.ai>"] |
| readme = "README.md" |
| homepage = "https://ggml.ai" |
| repository = "https://github.com/ggerganov/llama.cpp" |
| keywords = ["ggml", "gguf", "llama.cpp"] |
| packages = [{ include = "*.py", from = "." }] |
| classifiers = [ |
| "Programming Language :: Python :: 3", |
| "License :: OSI Approved :: MIT License", |
| "Operating System :: OS Independent", |
| ] |
|
|
| [tool.poetry.dependencies] |
| python = ">=3.9" |
| numpy = "^1.25.0" |
| sentencepiece = ">=0.1.98,<=0.2.0" |
| transformers = ">=4.35.2,<5.0.0" |
| protobuf = ">=4.21.0,<5.0.0" |
| gguf = { path = "./gguf-py" } |
| torch = { version = "^2.2.0", source = "pytorch" } |
|
|
| [tool.poetry.dev-dependencies] |
| pytest = "^5.2" |
|
|
|
|
| |
| |
| [[tool.poetry.source]] |
| name = "pytorch" |
| url = "https://download.pytorch.org/whl/cpu" |
| priority = "explicit" |
|
|
| [build-system] |
| requires = ["poetry-core>=1.0.0"] |
| build-backend = "poetry.core.masonry.api" |
|
|
| [tool.poetry.scripts] |
| llama-convert-hf-to-gguf = "convert_hf_to_gguf:main" |
| llama-convert-llama-ggml-to-gguf = "convert_llama_ggml_to_gguf:main" |
| llama-ggml-vk-generate-shaders = "ggml_vk_generate_shaders:main" |
|
|