Rick W / Tuesday, March 17, 2026 / Categories: Artificial Intelligence llama.cpp: Fast Local LLM Inference, Hardware Choices & Tuning Deploy Public MCP servers as an API endpoint and integrate its tools into LLM workflows using function calling. Previous Article Setting Up a Google Colab AI-Assisted Coding Environment That Actually Works Next Article How Walmart Is Reengineering AI Delivery Speed - with David Glick of Walmart Print 2 Tags: LLM