# Free GPU's for everyone! Run Ollama on Google Colab by running only this script by running the curl a few lines below.
# Execute this script by running the following command in Jupyther Notebook (Google Collab):
# !curl -fsSL https://molodetz.nl/retoor/uberlama/raw/branch/main/ollama-colab-v2.sh | sh
# Install Ollama.
curl -fsSL https://ollama.com/install.sh | sh
# Start Ollama on the background.
nohup ollama serve > ollama.log 2>&1 &
# Wait a second.
sleep 2
# Your model.
ollama pull qwen2.5-coder:14b
# Install the ollama molodetz client to publish your model to uberlama.
pip install aiohttp
wget https://retoor.molodetz.nl/retoor/uberlama/raw/branch/main/client.py
nohup python client.py > client.log 2>&1 &
# Keep process active. Display output in dem Jupyter Notebook.
tail -f *.log
# You can use the original Ollama / any OpenAI client (JS / Python whatever).
# All you have to do is change server to https://ollama.molodetz.nl.
# And configure it to use the model name above ofcourse!
#
# Test your model quickly by changing model name in the curl command below.
# curl https://ollama.molodetz.nl/v1/completions \
# -H "Content-Type: application/json" \
# -d '{
# "model": "qwen2.5:0.5b",
# "prompt": "What is Molodetz?"
# }'
#
# Retoor.