Page 1
Page 1
Started By
Message

Anyone built their own offline LLM AI?

Posted on 4/17/26 at 10:13 pm
Posted by HailToTheChiz
Back in Auburn
Member since Aug 2010
54671 posts
Posted on 4/17/26 at 10:13 pm
I'm talking something for business purposes for emails, calendar, doc review etc

I'm working on the idea of having an offline LLM AI that runs on its own laptop. Any work done would need to be based on files loaded to it from USB drives

Anyone tackled this yet?
Posted by LemmyLives
Texas
Member since Mar 2019
15817 posts
Posted on 4/18/26 at 9:08 am to
Posted by HailToTheChiz
Back in Auburn
Member since Aug 2010
54671 posts
Posted on 4/18/26 at 12:28 pm to
Posted by j1897
Member since Nov 2011
4587 posts
Posted on 4/18/26 at 12:43 pm to
There are tons you can run locally, i run them inside n8n for workflow automation.


Creating your "own" will require a few million GPUs.
Posted by HailToTheChiz
Back in Auburn
Member since Aug 2010
54671 posts
Posted on 4/18/26 at 12:48 pm to
Yeah i don't necessarily want to "create" my own. Bad word to use.

Honestly not even sure how i would fully implement but wanted to start some research. May just need to dabble
Posted by LSshoe
Burrowing through a pile o MikePoop
Member since Jan 2008
4571 posts
Posted on 4/18/26 at 2:11 pm to
My old gaming computer that has a gpu runs ollama, I have a couple other older pcs that run proxmox including a VM that runs docker containers. Among those are openwebui which gives you a sort of private gpt interface, especially when you use ollama on the backend. N8n uses ollama to do some automation and any other service that can use llm things I point to ollama too.
Posted by Brisketeer
Texas
Member since Aug 2013
1681 posts
Posted on 4/18/26 at 8:23 pm to
Google released Gemma 4 open models recently.
Posted by LSshoe
Burrowing through a pile o MikePoop
Member since Jan 2008
4571 posts
Posted on 4/19/26 at 7:40 am to
I have to run the smallest versions of those because i have no vram but the gemma4 models have been the best that I've used locally (on my hardware)
Posted by jivy26
Member since Nov 2008
2862 posts
Posted on 4/19/26 at 10:26 am to
I use ollama and pulled gemma4. It's running on my gaming laptop with no issues at all. Depending on your goal, you'll need to create an app layer to help dial in responses from the model. I've been working on mine for a week and a half and it's 65% like gpt or Claude premium versions.

Just takes a bit of time to refine answers and such. But Gemma 4 is highly recommended, as it has a low resource requirement.
This post was edited on 4/19/26 at 10:29 am
Posted by jivy26
Member since Nov 2008
2862 posts
Posted on 4/19/26 at 4:40 pm to
Also if you dont want to make your own agent, you could use Hermes; though I have not tried it
first pageprev pagePage 1 of 1Next pagelast page
refresh

Back to top
logoFollow TigerDroppings for LSU Football News
Follow us on X, Facebook and Instagram to get the latest updates on LSU Football and Recruiting.

FacebookXInstagram