WeichertLabs
Run Local AI with Ollama and WebUI
Setting up Ollama and Open WebUI on Windows 11 gives you access to a powerful, local AI assistant — similar to ChatGPT, but fully offline and private. In this guide, you’ll learn how to install Ollama, download AI models like LLaMA3, and run a local web interface using Docker.
This will allow you to:
- Run AI prompts locally without cloud dependencies
- Analyze scan results from Kali (WSL2)
- Create reports based on your own scripts and findings
- Build a fully offline, AI-powered pentest lab
Once completed, you’ll have a secure and fast AI tool available at any time — perfect for cybersecurity labs, automation, or just experimenting with local LLMs.
Please note: All guides and scripts are provided for educational purposes. Always review and understand any code before running it – especially with administrative privileges. Your system, your responsibility.
Use at your own risk: While every effort is made to ensure accuracy, I cannot take responsibility for issues caused by applying tutorials or scripts. Test in a safe environment before using in production.

Step 1 – Create the Virtual Machine
Screenshots & Code Snippets
Run the Proxmox Post-Install Script (Video Demo)
Video description