← All features

Local LLM Answers

Ask questions, get cited answers from your own files

Type a question and GNO gives you an answer grounded in your documents, with citations pointing to the exact source. Runs entirely on your machine — no cloud APIs, no data sharing. Start with a lightweight model (~1GB) or move up for higher quality.

Use cases
Private local RAG over notes, docs, and code
Remote-GPU setups that still keep the corpus local
Teams needing grounded answers rather than generic chat completions

What it gives you

  • Answers come from your documents, not the internet
  • Citations link to the exact source paragraph
  • 100% local — no cloud APIs, no API keys
  • Optional GPU offload if you have a server on your network
  • Multiple quality levels: fast and lightweight to slow and thorough
  • Different models per collection for tuned accuracy

Try it yourself

Representative commands and entry points. Full reference lives in the documentation.

gno ask 'your question' --answer
gno models use balanced
gno models pull

Keep reading

Related features and docs.