Update README.md
Browse files
README.md
CHANGED
|
@@ -8,18 +8,35 @@ tags:
|
|
| 8 |
- python-3.13
|
| 9 |
---
|
| 10 |
|
| 11 |
-
# Llama-CPP-Python Pre-built Wheels (
|
| 12 |
|
| 13 |
-
|
| 14 |
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
-
|
| 18 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
-
## 🛠 Usage in HF Spaces
|
| 21 |
|
| 22 |
-
|
| 23 |
-
Add this to your Dockerfile to install the wheel instantly:
|
| 24 |
-
```dockerfile
|
| 25 |
-
RUN pip install [https://huggingface.co/Jameson040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/Jameson040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
|
|
|
|
| 8 |
- python-3.13
|
| 9 |
---
|
| 10 |
|
| 11 |
+
# 🦙 Llama-CPP-Python Pre-built Wheels (Python 3.13)
|
| 12 |
|
| 13 |
+
### The solution for Hugging Face "Build Timeout" errors on the Free CPU Tier.
|
| 14 |
|
| 15 |
+
If you are using **Python 3.13** on a Hugging Face Free Space, compiling `llama-cpp-python` from source usually crashes or times out. This repository provides pre-compiled **manylinux** wheels that install in seconds.
|
| 16 |
+
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
## 🚀 Why use these wheels?
|
| 20 |
+
|
| 21 |
+
* **No Compilation:** Skips the 15+ minute build process.
|
| 22 |
+
* **Python 3.13 Support:** Specifically built for the latest Python version.
|
| 23 |
+
* **Generic CPU Optimization:** Compiled with `GGML_NATIVE=OFF`. This ensures the model runs on HF's shared CPUs without "Illegal Instruction" or "Core Dump" errors.
|
| 24 |
+
* **Lightweight:** Only ~4.3 MB compared to the massive overhead of building from source.
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
## 🛠️ How to use in your HF Space
|
| 29 |
+
|
| 30 |
+
### Option A: Using `requirements.txt`
|
| 31 |
+
Simply paste this direct link into your `requirements.txt` file:
|
| 32 |
+
|
| 33 |
+
```text
|
| 34 |
+
[https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
Option B: Using a Dockerfile
|
| 38 |
+
If you are using a custom Docker setup, add this line:
|
| 39 |
+
RUN pip install [https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl](https://huggingface.co/James040/llama-cpp-python-wheels/resolve/main/llama_cpp_python-0.3.16-cp313-cp313-linux_x86_64.whl)
|
| 40 |
|
|
|
|
| 41 |
|
| 42 |
+
📦 Build SpecificationsThese wheels were built using a high-performance automated pipeline on GitHub.SpecificationValuePython Version3.13PlatformLinux x86_64 (Manylinux)Build FlagsGGML_NATIVE=OFF, GGML_BLAS=OFFBuild SourceJameson040/my_lama-wheels
|
|
|
|
|
|
|
|
|