Damus
iefan πŸ•ŠοΈ profile picture
iefan πŸ•ŠοΈ
@iefan πŸ•ŠοΈ
Building a private AI assistant for truly private chat and personal data processing (e.g., notes, PDFs).

Planning to create a PWA web app that allows users to run any LLM model, including the latest LLaMA 3.2-1B, locally on any device or OS, even offline.

Users will simply select the model, and the app will automatically download, cache, and store it for future use. The next time, loading will be much faster.

The MLC-LLM library makes implementation straightforward, supporting all platforms and leveraging WebGPU for maximum performance.

Coming soon, meanwhile check there library repo. They have lots of examples.

MLC library repo: https://llm.mlc.ai/docs/deploy/webllm.html
44❀️11πŸ‘€2πŸ€™2β™₯1❀️1πŸ‘1
KernelKind · 73w
what’s the advantage of this over ollama?
S!ayer · 73w
nice to see Nextweb Chat in the wild
yavin5 · 72w
This is badass! I do already use Ollama but that's clearly a different use case than an in-browser LLM. Kudos for the excellent work!