Self-Hosting

How Ollama Handles Parallel Requests

How Ollama Handles Parallel Requests

Configuring ollama for parallel requests executions.

When the Ollama server receives two requests at the same time, its behavior depends on its configuration and available system resources.

Testing Deepseek-R1 on Ollama

Testing Deepseek-R1 on Ollama

Comparing two deepseek-r1 models to two base ones

DeepSeek’s first-generation of reasoning models with comparable performance to OpenAI-o1, including six dense models distilled from DeepSeek-R1 based on Llama and Qwen.

PyInstaller Quickstart

PyInstaller Quickstart

Combining python program into single file

Use PyInstaller to package Python programs as standalone executables. It works on Linux, Windows, and Mac.

Not the most efficient way, but usable.