Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
# Model card metadata following Hugging Face specification:
|
| 3 |
+
# https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
|
| 4 |
+
# Documentation: https://huggingface.co/docs/hub/model-cards
|
| 5 |
+
license: mit
|
| 6 |
+
tags:
|
| 7 |
+
- llama-cpp-python
|
| 8 |
+
- cuda
|
| 9 |
+
- nvidia
|
| 10 |
+
- blackwell
|
| 11 |
+
- windows
|
| 12 |
+
- prebuilt-wheels
|
| 13 |
+
- python
|
| 14 |
+
- machine-learning
|
| 15 |
+
- large-language-models
|
| 16 |
+
- gpu-acceleration
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
# llama-cpp-python 0.3.9 Prebuilt Wheel with CUDA Support for Windows
|
| 20 |
+
|
| 21 |
+
This repository provides a prebuilt Python wheel for **llama-cpp-python** (version 0.3.9) with NVIDIA CUDA support, optimized for Windows 10/11 (x64) systems. This wheel enables GPU-accelerated inference for large language models (LLMs) using the `llama.cpp` library, simplifying setup by eliminating the need to compile from source. The wheel is compatible with Python 3.10 and supports NVIDIA GPUs, including the latest Blackwell architecture.
|
| 22 |
+
|
| 23 |
+
## Available Wheel
|
| 24 |
+
- `llama_cpp_python-0.3.9-cp310-cp310-win_amd64.whl` (Python 3.10, CUDA 12.8)
|
| 25 |
+
|
| 26 |
+
## Compatibility
|
| 27 |
+
The prebuilt wheels are designed for NVIDIA Blackwell GPUs but have been tested and confirmed compatible with previous-generation NVIDIA GPUs, including:
|
| 28 |
+
- NVIDIA RTX 5090
|
| 29 |
+
- NVIDIA RTX 3090
|
| 30 |
+
|
| 31 |
+
## Installation
|
| 32 |
+
To install the wheel, use the following command in your Python 3.10 environment:
|
| 33 |
+
|
| 34 |
+
```bash
|
| 35 |
+
pip install llama_cpp_python-0.3.9-cp310-cp310-win_amd64.whl
|