mirror of
https://github.com/dogkeeper886/ollama37.git
synced 2025-12-10 15:57:04 +00:00
llama server wrapper
This commit is contained in:
34
server/README.md
Normal file
34
server/README.md
Normal file
@@ -0,0 +1,34 @@
|
||||
# Server
|
||||
|
||||
🙊
|
||||
|
||||
## Installation
|
||||
|
||||
If using Apple silicon, you need a Python version that supports arm64:
|
||||
|
||||
```bash
|
||||
wget https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-MacOSX-arm64.sh
|
||||
bash Miniforge3-MacOSX-arm64.sh
|
||||
```
|
||||
|
||||
Get the dependencies:
|
||||
|
||||
```bash
|
||||
pip install llama-cpp-python
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
## Running
|
||||
|
||||
Put your model in `models/` and run:
|
||||
|
||||
```bash
|
||||
python server.py
|
||||
```
|
||||
|
||||
## API
|
||||
|
||||
### `POST /generate`
|
||||
|
||||
model: `string` - The name of the model to use in the `models` folder.
|
||||
prompt: `string` - The prompt to use.
|
||||
Reference in New Issue
Block a user