Skip to content

A private, local chat interface for running LLM models. Provide the model path and enjoy a refreshing chat interface with zero setup friction.

Notifications You must be signed in to change notification settings

Debo0013/Web_llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

3 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Local LLM Chat Web App

A private, local chat interface for running GGUF models (like Gemma) using FastAPI and llama-cpp-python.

πŸš€ Setup

  1. Install Requests:

    pip install -r requirements.txt
  2. Download Model: Ensure you have your GGUF model file.

    • Default path configured in server.py: /Users/user_name/ml_models/gemma_models/gemma3_4b_it/gemma-3-4b-it-Q4_K_M.gguf
    • Edit MODEL_PATH in server.py if your model is elsewhere.

πŸƒβ€β™‚οΈ Running the Server

Start the chatbot server:

python3 server.py

πŸ”§ Troubleshooting

"Failed to fetch" or Network Errors

If you are using Brave Browser or have strict ad-blockers:

  1. Disable Shields: Click the Brave Lion icon (or ad-blocker icon) in your address bar.
  2. Allow Connection: Toggle shields/blocking DOWN for localhost.
  3. Reload: Press Cmd+Shift+R to hard refresh.

Reason: Privacy browsers often block connections to local ports (like 8080) if they suspect it's a tracking script.

"Conversation roles must alternate"

If you see server errors about roles:

  • The server now auto-magically merges "system" prompts into the first user message to keep the model happy. No action needed!

About

A private, local chat interface for running LLM models. Provide the model path and enjoy a refreshing chat interface with zero setup friction.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published