chore: format

This commit is contained in:
Timothy Jaeryang Baek 2025-01-28 19:57:21 -08:00
parent b803e507d5
commit 40c5cc4b0f

View File

@ -2,8 +2,6 @@
`llama-cpp-runner` is the ultimate Python library for running [llama.cpp](https://github.com/ggerganov/llama.cpp) with zero hassle. It automates the process of downloading prebuilt binaries from the upstream repo, keeping you always **up to date** with the latest developments. All while requiring no complicated setups—everything works **out-of-the-box**. `llama-cpp-runner` is the ultimate Python library for running [llama.cpp](https://github.com/ggerganov/llama.cpp) with zero hassle. It automates the process of downloading prebuilt binaries from the upstream repo, keeping you always **up to date** with the latest developments. All while requiring no complicated setups—everything works **out-of-the-box**.
---
## Key Features 🌟 ## Key Features 🌟
1. **Always Up-to-Date**: Automatically fetches the latest prebuilt binaries from the upstream llama.cpp GitHub repo. No need to worry about staying current. 1. **Always Up-to-Date**: Automatically fetches the latest prebuilt binaries from the upstream llama.cpp GitHub repo. No need to worry about staying current.
@ -12,7 +10,6 @@
4. **Built-in HTTP Server**: Automatically spins up a server for chat interactions and manages idle timeouts to save resources. 4. **Built-in HTTP Server**: Automatically spins up a server for chat interactions and manages idle timeouts to save resources.
5. **Cross-Platform Support**: Works on **Windows**, **Linux**, and **macOS** with automatic detection for AVX/AVX2/AVX512/ARM architectures. 5. **Cross-Platform Support**: Works on **Windows**, **Linux**, and **macOS** with automatic detection for AVX/AVX2/AVX512/ARM architectures.
---
## Why Use `llama-cpp-runner`? ## Why Use `llama-cpp-runner`?
@ -20,8 +17,6 @@
- **Streamlined Model Serving**: Effortlessly manage multiple models and serve them with an integrated HTTP server. - **Streamlined Model Serving**: Effortlessly manage multiple models and serve them with an integrated HTTP server.
- **Fast Integration**: Use prebuilt binaries from upstream so you can spend more time building and less time troubleshooting. - **Fast Integration**: Use prebuilt binaries from upstream so you can spend more time building and less time troubleshooting.
---
## Installation 🚀 ## Installation 🚀
Installing `llama-cpp-runner` is quick and easy! Just use pip: Installing `llama-cpp-runner` is quick and easy! Just use pip:
@ -30,8 +25,6 @@ Installing `llama-cpp-runner` is quick and easy! Just use pip:
pip install llama-cpp-runner pip install llama-cpp-runner
``` ```
---
## Usage 📖 ## Usage 📖
### Initialize the Runner ### Initialize the Runner
@ -58,38 +51,28 @@ response = llama_runner.chat_completion({
print(response) print(response)
``` ```
---
## How It Works 🛠️ ## How It Works 🛠️
1. Automatically detects your system architecture (e.g., AVX, AVX2, ARM) and platform. 1. Automatically detects your system architecture (e.g., AVX, AVX2, ARM) and platform.
2. Downloads and extracts the prebuilt llama.cpp binaries from the official repo. 2. Downloads and extracts the prebuilt llama.cpp binaries from the official repo.
3. Spins up a lightweight HTTP server for chat interactions. 3. Spins up a lightweight HTTP server for chat interactions.
---
## Advantages 👍 ## Advantages 👍
- **Hassle-Free**: No need to compile binaries or manage system-specific dependencies. - **Hassle-Free**: No need to compile binaries or manage system-specific dependencies.
- **Latest Features, Always**: Stay up to date with llama.cpps improvements with every release. - **Latest Features, Always**: Stay up to date with llama.cpps improvements with every release.
- **Optimized for Your System**: Automatically fetches the best binary for your architecture. - **Optimized for Your System**: Automatically fetches the best binary for your architecture.
---
## Supported Platforms 🖥️ ## Supported Platforms 🖥️
- Windows - Windows
- macOS - macOS
- Linux - Linux
---
## Contributing 💻 ## Contributing 💻
Wed love your contributions! Bug reports, feature requests, and pull requests are all welcome. Wed love your contributions! Bug reports, feature requests, and pull requests are all welcome.
---
## License 📜 ## License 📜
This library is open-source and distributed under the MIT license. This library is open-source and distributed under the MIT license.