Tag AI

Building a Local GenAI Service with Ollama, Mistral, and Go

Running Large Language Models (LLMs) locally provides data privacy, zero latency costs, and full control over your inference environment. This guide demonstrates how to containerize Ollama, automate the Mistral model download, and expose it through an Nginx reverse proxy to…

Building Custom AI Applications with LLMs and RAG

Introduction: The Power Duo – LLMs and RAG Large Language Models (LLMs) like GPT-4, Llama 2, or Mistral have demonstrated incredible capabilities in understanding and generating human-like text. They are trained on vast amounts of data, enabling them to perform…