How to Use Local AI with Obsidian: Smart Notes with Ollama

A practical guide to connecting local LLMs to Obsidian for AI-powered note-taking: setting up Smart Connections with nomic-embed-text embeddings to search and chat across your entire vault, using the Ollama plugin for in-note summarisation and writing assistance, the BMO Chatbot for a persistent sidebar assistant, practical workflows for daily note processing, research synthesis, idea generation, and meeting note cleanup, and context window configuration for long notes.

How to Use torch.profiler to Find Training Bottlenecks

A practical guide to using torch.profiler for GPU bottleneck analysis: profiler setup and schedule configuration, reading TensorBoard traces, identifying DataLoader gaps and CPU-GPU sync stalls, memory profiling, CUDA kernel occupancy, distributed training profiling, and a step-by-step workflow for translating profiler output into concrete training optimizations.

How to Use Ollama Vision Models for Local Image Analysis

A practical guide to running vision models locally with Ollama: available models including LLaVA, moondream2, Qwen2.5-VL and Gemma 3, using vision models from the CLI with image flags, calling them via the Ollama Python library and the OpenAI-compatible API with base64-encoded images, batch processing a folder of images, and practical use cases including screenshot OCR, chart analysis, document data extraction, and inventory cataloguing.

How to Run Ollama with Docker and Docker Compose

A complete guide to running Ollama in Docker: CPU-only and NVIDIA GPU passthrough with the NVIDIA Container Toolkit, AMD ROCm setup, pulling models into the running container, a full Docker Compose stack combining Ollama with Open WebUI, environment variables for parallel requests and model keep-alive, persisting models across redeployments with named volumes, and when to use Docker versus native Ollama installation.

Msty: The Local LLM App That Lets You Compare Models Side by Side

A complete guide to Msty, the local LLM desktop app built for multi-model comparison: connecting to existing Ollama models, running the same prompt against multiple models simultaneously, attaching document Knowledge collections to conversations, conversation branching for exploring different approaches, the prompts library for reusable system prompts, integrating cloud providers like OpenAI and Anthropic alongside local models, and how Msty compares to Jan, LM Studio, and Open WebUI.

How to Set Up Continue in VS Code with Ollama for Local AI Coding

A complete setup guide for the Continue VS Code extension with Ollama: installing Continue and configuring config.json with chat and autocomplete models, enabling tab completions with fill-in-the-middle using Qwen2.5-Coder, indexing your codebase with nomic-embed-text embeddings for @codebase search, using built-in slash commands for editing, testing and commenting code, creating custom slash commands, switching between multiple local and cloud models, JetBrains setup, and troubleshooting connection issues.

How to Add OpenTelemetry Tracing to LLM Applications

A practical guide to instrumenting LLM applications with OpenTelemetry: setting up the SDK and OTLP exporter, creating spans around LLM calls with GenAI semantic conventions, tracing multi-step RAG pipelines with nested spans, what attributes to record, backend options from Grafana Tempo to Honeycomb and Langfuse, and when auto-instrumentation is enough vs when to add manual spans.