Re: AI LLM Artificial intelligence infrastructure
By: phigan to bbsing on Tue Mar 10 2026 14:01:15
Hi phigan,
Re: AI LLM Artificial intelligence infrastructure
By: bbsing to MRO on Mon Mar 09 2026 09:23 am
Trying to get a system capable of running LLM models equals lots of cash, an I'm on the fence if I should get into it, so I'm wondering if anyone out her
I've got a system set up with ollama and 24gb vram. I can show you what a
What models are you liking, and what sizes are you getting good TPS?
What do you think is good TPS?
Have you tried Deepseek-v3?
vllm?
hosted model can do via chat or something. It's really not very good.
A cloud-hosted one might do better, but the risk is still extremely high of it doing things you don't want.
13 years ago messing things up was my concern with scripting aspects of sys admin work. I worked with a Jr admin, and he was really liking powershell, but at the time was new to it. My worry was scripts have a great way have doing things quickly but also screwing things up quickly. He was a great team member and his scripting use inspired me to adopt his ideas. He left for a different job closer to home. I miss working with him. Now most of my work is via scripting because I have too much to do and without scripts it takes too long. I really get to know individual systems better with full hands on vs scripts, and I like knowing each system, its behavior, but efficeincy demands quickery.
I've been contemplating building an system for LLM personal lab work on infrastructure, with the idea of discriptive text for total orchistration, including builds, configuration, and management on windows/linux/bsd type systems. I'm having a hell of a time getting costs down. 24vgb is expensive, about as expensive as 32vgb rtx 5090.
I've looked at so many builds with rtx 3060, rtx 4090, rtx 5090. 3060's are gone, 4090's are way overpriced, and if they burn up, are likely not going to be the card to replace a broken device. rtx 5080s are around but the older cards are just as good, but their way over MSRP. Its been pretty dang frustrating.
Back in 2000, I was facing a similar issue. I was priced out of hardware and OS due to cost, so I couldn't build a lab to learn on. Now I'm in the same place. I don't really want to spend $8K on a build that is obsolete in two years. I also don't want to miss out on all the hype/fun/understanding.
I was hoping distributive computing would have been the thing instead of giant LLM type shtuff.
---
þ Synchronet þ Lunar Outpost BBS