Skip to content

Latest commit

 

History

History
36 lines (26 loc) · 1.67 KB

File metadata and controls

36 lines (26 loc) · 1.67 KB

Hugging_Face_LLM_2025

Experimenting with: Tiny Large Language Models.

Setup and Use

docker compose up
  1. http://localhost:8000/public/index.html
  2. http://localhost:8000/api/llm?prompt=abcdefghijklmnopqrstuvwxyz

Warning

  1. This is a simple example (NOT Production-worthy) and basic safeguards (like input field validation, param sanitization, and the like) are mostly omitted here.
    • There's not a tremendous lot one can do within the context of this simple demo to break things/be malicious.
    • But the design pattern should be avoided in Production as is (without the addition of typical Production security mechanisms)!
  2. The build can take upwards of 30 minutes due to large (many GB) LLM Models.

Notes

  1. Hugging Face will download and cache models--arnir0--Tiny-LLM into root/.cache/huggingface/hub.
  2. Also, Docker Desktop 4.40 now supports Docker Model Runner.
    • The following commands will launch of Dockerized LLM Model (somewhat similar to this one):
      • docker model pull ai/deepseek-r1-distill-llama
      • docker model run ai/deepseek-r1-distill-llama
      • Docker Model Runner supports Hugging Face Models.

Resources and Links

  1. https://huggingface.co/arnir0/Tiny-LLM
  2. https://huggingface.co/learn/llm-course/