Writing effective prompts for LLMs
Requires some experimenting but
Still there are some common approaches how to write good prompts so LLM would not get confused trying to understand what you wand from it.
Requires some experimenting but
Still there are some common approaches how to write good prompts so LLM would not get confused trying to understand what you wand from it.
8 llama3 (Meta+) and 5 phi3 (Microsoft) LLM versions
Testing how models with different number of parameters and quantization are behaving.
Ollama LLM model files take a lot of space
After installing ollama better to reconfigure ollama to store them in new place right away. So after we pull a new model, it doesn’t get downloaded to the old location.
Let's test the LLMs' speed on GPU vs CPU
Comparing prediction speed of several versions of LLMs: llama3 (Meta/Facebook), phi3 (Microsoft), gemma (Google), mistral(open source) on CPU and GPU.
Let's test logical fallacy detection quality of different LLMs
Here I am comparing several LLM versions: Llama3 (Meta), Phi3 (Microsoft), Gemma (Google), Mistral Nemo(Mistral AI) and Qwen(Alibaba).