If you’ve ever wanted a single repo that lets you flip a switch between a locally‑hosted LLM, a hosted OpenAI model, and a Python‑based evaluation suite—while keeping every call visible in a unified observability stack—this is it.

Fabian G. Williams
Principal Product Manager, Microsoft Subscribe to my YouTube.
Mastering Llama 3.3 – A Deep Dive into Running Local LLMs
I compared Llama 3.3 with older models like 3.170B and Phi 3, tested Microsoft Graph API integrations using OpenAPI specs, and explored function calling quirks. The results? Not all models are created equal!