43 lines
866 B
Plaintext
43 lines
866 B
Plaintext
|
Metadata-Version: 2.1
|
||
|
Name: yura
|
||
|
Version: 14.3.7
|
||
|
Summary: Yura async AI client
|
||
|
Author: retoor
|
||
|
Author-email: retoor@retoor.io
|
||
|
License: MIT
|
||
|
Requires-Python: >=3.7
|
||
|
Description-Content-Type: text/markdown
|
||
|
Requires-Dist: websockets
|
||
|
|
||
|
# Yura LLM Client for Katya server
|
||
|
|
||
|
Part of project with as target replacing the native ollama protocol. This protocol supports streaming and is usable trough https and it is possible to directly attach a web client to the backend.
|
||
|
|
||
|
## Install
|
||
|
```bash
|
||
|
pip install -e .
|
||
|
```
|
||
|
|
||
|
## Build
|
||
|
```bash
|
||
|
make build
|
||
|
```
|
||
|
|
||
|
## Command line usage
|
||
|
```bash
|
||
|
yura ws://[host]:[port]/[path]/
|
||
|
```
|
||
|
|
||
|
## Python
|
||
|
```python
|
||
|
import asyncio
|
||
|
from yura.client import AsyncClient
|
||
|
|
||
|
async def communicate():
|
||
|
client = AsyncClient("ws://[host]:[port]/[path]/")
|
||
|
async for response in client.chat("Your prompt"):
|
||
|
print(response)
|
||
|
|
||
|
asyncio.run(communicate())
|
||
|
```
|