Building Your Own AI Agents

2025-01-24

All articles from this blog can ONLY be redistributed on an Attribution-NonCommercial-NoDerivs basis. Please credit the source, thank you.
Twitter:@kelvinshen
Blog:Kelvin Shen's Blog

Large language models (LLMs) are powerful, but relying on cloud-based access has drawbacks: cost and privacy. Running LLMs locally gives you control, speed, data security, and the freedom to explore a wider range of models. This article records some of my learning journey.

Hardware

To run AI models locally, you need sufficient computing power. While it is possible to run LLM models on CPUs. Both GPU processing power and graphic memory capacity are key factors for optimal speed.

Currently (Jan 20215), Nvidia GPUs have been dominating the market. The newer and the more memory, the better.

In fact, RTX 4090 and newer models are so good that they are banned to be exported to China.

Nvidia GPUs are popular, but Apple’s M1/M2/M3/M4 chips with integrated GPUs are a strong alternative. However, sometimes it is not available to environments inside docker containers.

If you are looking for a powerful AI laptop, MSI Creator 16 AI Studio Laptop seems to be a good option.

Hosting Options

For hosting AI models locally, several options are available:

  • Anything LLM: Supports running LLM models locally and offers agents.
  • LM Studio: Provides an easier-to-use interface for running models locally.
  • Ollama: Allows you to run models locally with a great degree of flexibility and its target audiance are developers or advanced users.
  • n8n Self-hosted AI Start Kit: An open-source template that quickly sets up a local AI environment in docker containers.

Leave a Reply

comments powered by Disqus


Table of Content