Before reading your comment, I knew IN MY SOUL, that this was from the shittier part of Florida
- 0 Posts
- 20 Comments
-
You can look at manufacturers info pages and see what they support. Intel integrated chips usually list the capabilities and you’ll want to double check with your mini PC or motherboard manufacturer to make sure they support it too. I think any i5+ from the past 5 years with integrated graphics should be able to play/decode 4k media (someone correct me if this sounds crazy). Fornsure my core ultra 265. As far as codec support, I’m not familiar with the compatibilities but I’m sure everything CAN be played on recentish hardware. Encoding is out of my weelhouse.
-
I’ve used HDMI 2.1 hdr 4k120 on Linux with Nvidia, AMD and Integrated Intel. AMD will be the best experience especially on cards from the past 5 years. Nvidia, with proprietary drivers, on 3000 series or newer should be good for a few more years. I heard 2000 series will be dropped from support soonish m. Intel HDMI 2.1 is a pain on linux and I’ve only been able to get HDR 4k120 using a special DP to HDMI cable.
-
afk_strats@lemmy.worldto
Lemmy Shitpost@lemmy.world•Negotiating with other family membersEnglish
5·2 months ago
afk_strats@lemmy.worldto
Open Source@lemmy.ml•Notepad++ hijacked by state-sponsored hackersEnglish
9·2 months agoNotepad++ works fine on Wine on Mac and Linux. After being away from it from awhile, I realized I don’t need it anymore. I would often use the column edit mode and recorded macros, but I just bash script those now. I guess I’m a different person now?!?
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What's an objectively terrible movie that you love anyway?English
4·2 months agoProud
afk_strats@lemmy.worldto
Technology@lemmy.world•Researchers figured out how to run a 120-billion parameter model across four regular desktop PCsEnglish
2·3 months agoI still think AI is mostly a toy and a corporate inflation device. There are valid use cases but I don’t think that’s the majority of the bubble
- For my personal use, I used it to learn how models work from a compute perspective. I’ve been interested and involved with natural language processing and sentiment analysis since before LLMs became a thing. Modern models are an evolution of that.
- A small, consumer grade model like GPT-oss-20 is around 13GB and can run on a single mid-grade consumer GPU and maybe some RAM. It’s capable of parsing text and summarizing, troubleshooting computer issues, and some basic coding or code review for personal use. I built some bash and home assistant automatons for myself using these models as crutches. Also, there is software that can index text locally to help you have conversations with large documents. I use this with documentation for my music keyboard which is a nightmare to program and with complex APIs.
- A mid-size model like Nemotron3 30B is around 20GB can run on a larger consumer card (like my 7900xtx with 24 gb of VRAM, or 2 5060tis with 16gb of vRAM each) and will have vaguely the same usability as the small commercial models, like Gemini Flash, or Claude Haiku. These can write better, more complex code. I also use these to help me organize personal notes. I dump everything in my brain to text and have the model give it structure.
- A large model like GLM4.7 is around 150GB can do all the things ChatGPT or Gemini Pro can do, given web access and a pretty wrapper. This requires big RAM and some patience or a lot of VRAM. There is software designed to run these larger models in RAM faster, namely ik_llama but, at this scale, you’re throwing money at AI.
I played around with image creation and there isn’t anything there other than a toy for me. I take pictures with a camera.
afk_strats@lemmy.worldto
Technology@lemmy.world•Researchers figured out how to run a 120-billion parameter model across four regular desktop PCsEnglish
71·3 months agoI think you’re missing the point or not understanding.
Let me see if I can clarify
What you’re talking about is just running a model on consumer hardware with a GUI
The article talks about running models on consumer hardware. I am making the point that this is not a new concept. The GUI is optional but, as I mentioned, llama.cpp and other open source tools provide an OpenAI-compatible api just like the product described in the article.
We’ve been running models for a decade like that.
No. LLMs, as we know them, aren’t that old, were a harder to run and required some coding knowledge and environment setup until 3ish years ago, give or take when these more polished tools started coming out.
Llama is just a simplified framework for end users using LLMs.
Ollama matches that description. Llama is a model family from Facebook. Llama.cpp, which is what I was talking about, is an inference and quantization tool suite made for efficient deployment on a variety of hardware including consumer hardware.
The article is essentially describing a map reduce system over a number of machines for model workloads, meaning it’s batching the token work, distributing it up amongst a cluster, then combining the results into a coherent response.
Map reduce, in very simplified terms, means spreading out compute work to highly pararelized compute workers. This is, conceptually, how all LLMs are run at scale. You can’t map reduce or parallelize LLMs any more than they already are. The article doent imply map reduce other than taking about using multiple computers.
They aren’t talking about just running models as you’re describing.
They don’t talk about how the models are run in the article. But I know a tiny bit about how they’re run. LLMs require very simple and consistent math computations on extremely large matrixes of numbers. The bottleneck is almost always data transfer, not compute. Basically, every LLM deployment tool is already tries to use as much parallelism as possible while reducing data transfer as much as possible.
The article talks about gpt-oss120, so were aren’t talking about novel approaches to how the data is laid out or how the models are used. We’re talking about tranformer models and how they’re huge and require a lot of data transfer. So, the preference is try to keep your model on the fastest-transfer part of your machine. On consumer hardware, which was the key point of the article, you are best off keeping your model in your GPU’s memory. If you can’t, you’ll run into bottlenecks with PCIe, RAM and network transfer speed. But consumers don’t have GPUs with 63+ GB of VRAM, which is how big GPT-OSS 120b is, so they MUST contend with these speed bottlenecks. This article doesn’t address that. That’s what I’m talking about.
afk_strats@lemmy.worldto
Technology@lemmy.world•Researchers figured out how to run a 120-billion parameter model across four regular desktop PCsEnglish
192·3 months agoThis is basically meaningless. You can already run gpt-OSS 120 across consumer grade machines. In fact, I’ve done it with open source software with a proper open source licence, offline, at my house. It’s called llama.cpp and it is one of the most popular projects on GitHub. It’s the basis of ollama which Facebook coopted and is the engine for LMStudio, a popular LLM app.
The only thing you need is around 64 gigs of free RAM and you can serve gpt-oss120 as an OpenAI-like api endpoint. VRAM is preferred but llama.cpp can run in system RAM or on top of multiple different GPU addressing technologies. It has a built-in server which allows it to pool resources from multiple machines…
I bet you could even do it over a series of high-ram phones in a network.
So I ask is this novel or is it an advertisement packaged as a press release?
- Cream Theater
- System of a Town
- Go:jira
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What is it about technology that fascinates you?English
8·3 months agoIf you know the right incantatuons, you can make sand do the things you want. IM A FUCKING WIZARD, HARRY
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What YouTube channel to you has degraded in time?English
9·4 months agoStopped watching around COVID but before that he was a great teacher about how home objects were created. It really made me understand how everyday objects are made and why they’re made a certain way. There’s a niche there of an experienced manufacturing engineer to teach the masses
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What YouTube channel to you has degraded in time?English
38·4 months agoThis is VERY recent but NetworkChuck. He’s such a great speaker and his videos were good intros into tech (Linux, networking, etc.) but with practical applicability. I liked that he skewed family-friendly and family positive, and some of the solutions he presented were from his perspective as a father. He was never shy about him being religious and being religiously involved. Imo 100% wholesome.
This year, several weeks of his videos were paid sponsor content for very corporate Cisco BS. Then, he comes back from that and it’s AI stuff that’s kind of a strech for a typical home user… Then “I’m going to end my videos in a prayer”. I unsubscribed right there.
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What YouTube channel to you has degraded in time?English
4·4 months agoContrapoints along the same lines?
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•What YouTube channel to you has degraded in time?English
1·4 months agoRespectfully disagree. I was a fan from the first Honda Accord video and was a pattreon supporter when he opened it. I still watch watch every video.
It’s been so long. I can’t expect the same content as 10/15 years ago and I can’t exext the guys to be the same people. But the groove is there. The humor is there. It’s still raunchy and the furry stuff is different and fresh. I love RCR and I think the content still peaking.
afk_strats@lemmy.worldto
Technology@lemmy.world•Introducing SlopStop: Community-driven AI slop detection in Kagi SearchEnglish
211·5 months agoCan we make an extension for Firefox and call it Sloppy-Stoppy?
afk_strats@lemmy.worldto
Linux@lemmy.ml•Tired of reinstalling every time you tweak your system? Build a resilient Linux desktop with Btrfs, LUKS and borg in one afternoonEnglish
2·5 months agoThis is really neat. Thank you. I would love a script or a more newb-friendly guide, not just for me, but for a lot of other users.
Can I make a suggestion? Post your script on github or similar with a proper (open) liscence so people can make suggestions or versions they find useful.
afk_strats@lemmy.worldto
Selfhosted@lemmy.world•Stop cramming everything onto one Pi: treat your home lab like a tiny ISP - hardware, stack, backups and an update planEnglish
31·5 months agoI’ve been on the internet a long time and this made me say “what the fuck” out loud
Edit: not sure whether I should ask what this all is or if ibshpuld complement you on your “output”
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•'Read' and its past tense are spelled the same. How should they be spelled?English
1·6 months agoThe Chaos by Gerard Nolst Trenité (1922)
https://ncf.idallen.com/english.html
Dearest creature in creation
Studying English pronunciation,
I will teach you in my verse
Sounds like corpse, corps, horse and worse.I will keep you, Susy, busy,
Make your head with heat grow dizzy;
Tear in eye, your dress you’ll tear;
Queer, fair seer, hear my prayer.Pray, console your loving poet,
Make my coat look new, dear, sew it!
Just compare heart, hear and heard,
Dies and diet, lord and word.
…Very long. Highly recommended
afk_strats@lemmy.worldto
Ask Lemmy@lemmy.world•Programmers of Lemmy, what are your interviewing horror stories?English
1·1 year agoEdit: this is from the perspective of a technical interviewer.
I’ve done around 200 or so technical interviews for mostly senior data engineering roles. I’ve seen every version of made up code, terrible implementation suggestion and dozens of folks with 5+ years of experience and couldn’t wrote a JOIN to save their lives.
The there were a couple where the resume was obviously made up because they couldn’t back up a single point and they just did not know a thing about data. They would usually talk in circles about buzzwords and Excel jaron. “They big data’d the data lake warehouse pivot hadoop in Azure Redshift.” Sure, ya did, buddy.
Yes, they were “pre-screened”. This was one of the BIG tech companies.
“unlimited PTO”
*looks inside
”4 weeks of PTO unless you have VP approval except you’ll never get it”