• brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    12 minutes ago

    Also, this has been the case (or at least planned) for a while.

    Pascal (the GTX 1000 series) and Ampere (the RTX 3000 series) used the exact same architecture for datacenter/gaming. The big gaming dies were dual use and datacenter-optimized. This habit sort of goes back to ~2008, but Ampere and the A100 is really where “datacenter first” took off.

    AMD announced a plan to unify their datacenter/gaming architecture awhile ago, and prioritized the MI300X before that. And EPYC has always been the priority, too.

    Intel wanted to do this, but had some roadmap trouble.

    These companies have always put datacenter first, it just took this much drama for the consumer segment to largely notice.

  • melfie@lemy.lol
    link
    fedilink
    English
    arrow-up
    11
    ·
    edit-2
    2 hours ago

    I’ve been looking into self-hosting LLMs, and it seems a $10k GPU is kind of a requirement to run a decently-sized model and get reasonable tokens / s rate. There’s CPU and SSD offloading, but I’d imagine it would be frustratingly slow to use. I even find cloud-based AI like GH Copilot to be rather annoyingly slow. Even so, GH Copilot is like $20 a month per user, and I’d be curious what the actual costs are per user considering the hardware and electricity cost.

    What we have now is clearly an experimental first generation of the tech, but the industry is building out data centers as though it’s always going to require massive GPUs / NPUs with wicked quantities of VRAM to run these things. If it really will require huge data centers full of expensive hardware where each user prompt requires minutes of compute time on a $10k GPU, then it can’t possibly be profitable to charge a nominal monthly fee to use this tech, but maybe there are optimizations I’m unaware of.

    Even so, if the tech does evolve and it become a lot cheaper to host these things, then will all these new data centers still be needed? On the other hand, if the hardware requirements don’t decrease by an order of magnitude, then will it be cost effective to offer LLMs as a service, in which case, I don’t imagine the new data centers will be needed either.

    • Analog@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      21 minutes ago

      Can run decent size models with one of these: https://store.minisforum.com/products/minisforum-ms-s1-max-mini-pc

      For $1k more you can have the same thing from nvidia in their dgx spark. You can use high speed fabric to connect two of ‘em and run 405b parameter models, or so they claim.

      Point being that’s some pretty big models in the 3-4k range, and massive models for less than 10k. The nvidia one supports comfyui so I assume it supports cuda.

      It ain’t cheap and AI has soooo many negatives, but… it does have some positives and local LLMs mitigate some of the minuses, so I hope this helps!

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      edit-2
      1 hour ago

      This is not true. I have a single 3090 + 128GB CPU RAM (which wasn’t so expensive that long ago), and I can run GLM 4.6 350B at 6 tokens/sec, with measurably reasonable quantization quality. I can run sparser models like Stepfun 3.5, GLM Air or Minimax 2.1 much faster, and these are all better than the cheapest API models. I can batch Kimi Linear, Seed-OSS, Qwen3, and all sorts of models without any offloading for tons of speed.


      …It’s not trivial to set up though. It’s definitely not turnkey. That’s the issue.

      You can’t just do “ollama run” and expect good performance, as the local LLM scene is finicky and highly experimental. You have to compile forks and PRs, learn about sampling and chat formatting, perplexity and KL divergence, about quantization and MoEs and benchmarking. Everything is moving too fast, and is too performance sensitive, to make it that easy, unfortunately.

      EDIT:

      And if I were trying to get local LLMs setup today, for a lot of usage, I’d probably buy an AI Max 395 motherboard instead of a GPU. They aren’t horrendously priced, and they don’t slurp power like a 3090. 96GB VRAM is the perfect size for all those ~250B MoEs.

      But if you go AMD, take all the finickiness for an Nvidia setup and multiply it by 10. You better know your way around pip and Linux, as if you don’t get it exactly right, performance will be horrendous, and many setups just won’t work anyway.

      • melfie@lemy.lol
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        1 hour ago

        Appreciate all the info! I did find this calculator the other day, and it’s pretty clear the RTX 4060 in my server isn’t going to do much though its NVMe may help.

        https://apxml.com/tools/vram-calculator

        I’m also not sure under 10 tokens per second will be usable, though I’ve never really tried it.

        I’d be hesitant to buy something just for AI that doesn’t also have RTX cores because I do a lot of Blender rendering. RDNA 5 is supposed to have more competitive RTX cores along with NPU cores, so I guess my ideal would be a SoC with a ton of RAM. Maybe when RDNA 5 releases, the RAM situation will have have blown over and we will have much better options for AMD SoCs with strong compute capabilities that aren’t just a 1-trick pony for rasterization or AI.

        • brucethemoose@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          48 minutes ago

          I did find this calculator the other day

          That calculator is total nonsense. Don’t trust anything like that; at best, its obsolete the week after its posted.

          I’d be hesitant to buy something just for AI that doesn’t also have RTX cores because I do a lot of Blender rendering. RDNA 5 is supposed to have more competitive RTX cores

          Yeah, that’s a huge caveat. AMD Blender might be better than you think though, and you can use your RTX 4060 on a Strix Halo motherboard just fine. The CPU itself is incredible for any kind of workstation workload.

          along with NPU cores, so I guess my ideal would be a SoC with a ton of RAM

          So far, NPUs have been useless. Don’t buy any of that marketing.

          I’m also not sure under 10 tokens per second will be usable, though I’ve never really tried it.

          That’s still 5 words/second. That’s not a bad reading speed.

          Whether its enough? That depends. GLM 350B without thinking is smarter than most models with thinking, so I end up with better answers faster.

          But anyway, I’m get more like 20 tokens a second with models that aren’t squeezed into my rig within an inch of their life. If you buy an HEDT/Server CPU with more RAM channels, it’s even faster.

          If you want to look into the bleeding edge, start with https://github.com/ikawrakow/ik_llama.cpp/

          And all the models on huggingface with the ik tag: https://huggingface.co/models?other=ik_llama.cpp&sort=modified

          You’ll see instructions for running big models on a 4060 + RAM.

          If you’re trying to like batch process documents quickly (so no CPU offloading), look at exl3s instead: https://huggingface.co/models?num_parameters=min%3A12B%2Cmax%3A32B&sort=modified&search=exl3

          And run them with this: https://github.com/theroyallab/tabbyAPI

  • dhork@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    1
    ·
    4 hours ago

    This is yet another thing I blame on American Business sacrificing itself on the altar of Shareholder Value. It’s no longer acceptable for a public business to simply make a profit. It has to grow that profit, every quarter, without fail.

    So, simply having a good consumer product division that makes money won’t be enough. At some point some executive will decide that he can’t possibly get his bonus if that’s all they do, and decide they need to blow it all up to chase larger profits elsewhere.

    Maybe we need a small, private company to come along and start making good consumer hardware. They still need components, though, so will have to navigate getting that from public companies who won’t return their calls. And even once they are successful, the first thing they will do is cash out and go public, and the cycle starts again.

    • SoleInvictus@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      8
      ·
      3 hours ago

      Maybe we need a small, private company to come along and start making good consumer hardware.

      I’ve always wanted to start a business like this. “Generic Brand” household goods. Not fancy, just solidly functional base models but with modular upgradability. Wish you bought the WiFi capable washer? Buy the module for $30. Everything would be fully user serviceable and upgradable (within reason), so parts sales ensure sustained income once market saturation is reached.

      • dhork@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        3 hours ago

        It’s not totally out of the realm of possibility. Michael Dell did it, after all, but he did it in a different time.

        And Dell is actually a good case study for all this. It went public rather quickly after it started growing, but grew a bit stagnant by the 2000’s. So much so that 2013, Michael Dell orchestrated a leveraged buyout of his own company (with the help of venture capital) to make it private again. He pretty much admitted that the changes he wanted to make to the company would be impossible while it was still public. It stayed private for a while, but went public again as part of some deal made after it acquired the parent company of VMware.

        Another notable thing is that Carl Ichan owned a large chunk of Dell, both in its first public incarnation and in its private incarnation. When Dell tried to take it private, Ichan challenged the plan, and thought about putting in his own bid, only to back off when he decided it wasn’t worth the effort to revive the company. Still, he was publicly against Dell’s buyout plan but was outvoted by other shareholders. Yet, he must have still held a part of the private company, because Ichan also protested it’s second plan to go public, and sued to force Dell to increase its terms to the private holders.

        Michael Dell is no saint, but I conclude that he realized that the company meant more than a spreadsheet, and needed a purpose to justify its existence. He also realized that in order to sustain a business over the long term, having to constantly sustain quarterly numbers may be counterproductive. I think Carl Ichan, on the other hand, only cares about Number Go Up, and doesn’t care at all about how the company makes that happen. Over the long term, that will never be sustainable, but fuck you all, he got his bag already.

        • SoleInvictus@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          2
          ·
          54 minutes ago

          I had no idea! I worked IT in the early 2000s and I absolutely hated Dell computers. Nothing broke faster and more often than the Dell desktops.

  • Fred@lemmy.world
    link
    fedilink
    English
    arrow-up
    32
    ·
    10 hours ago

    Imma remember what Curcial and others are doing, so when the AI bubble pops I’ll skip on all their products.

    • Lfrith@lemmy.ca
      link
      fedilink
      English
      arrow-up
      36
      ·
      11 hours ago

      Hopefully Linux supported. That’s the main selling point of AMD GPUs right now for me, since there’s less problems with trying to get stuff like HDR running on it than NVIDIA.

      I wonder why China is still for the most part ignoring Linux in favor of Windows. Like to update 8bitdo controllers they only provide a Microsoft program and no Linux version.

      You’d think they’d be rushing towards pushing Linux adoption.

      • Joe@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        12
        ·
        10 hours ago

        That’s capitalism for you. But also Linux, where it’s typical to upstream hardware support and rely on existing ecosystems rather than release addon drivers or niche supporting apps.

        China has made some strategic investments in Linux over the years though – often domestically targeted, like Red Flag Linux, and drivers for chinese hardware, etc.

        • ulterno@programming.dev
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 hours ago

          But also Linux, where it’s typical to upstream hardware support and rely on existing ecosystems rather than release addon drivers or niche supporting apps.

          Still possible though, right?
          It does afterall support out of tree device drivers now.

          • Joe@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 hours ago

            Sure… but why would el cheapo hardware want/need to support proprietary drivers? Now, for premium hardware and software, they might still want vendor lock-in mechanisms… So unless you absolutely have to, you should avoid hardware on Linux that needs proprietary drivers.

        • Lfrith@lemmy.ca
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          10 hours ago

          Linux seems to be common to run things like servers, but is that the case for consumer hardware?

          When I’ve looked at peripherals like keyboards and controllers linux support has been lacking. Of course, for keyboards I went out of my way to get qmk compatible ones to use via and vial instead so I dont need to run an exe of unknown origins to remap or update the firmware.

          And how is it for games. Is there more of a push to support Linux for their games? Since like Genshin Impact they only officially support Windows. There’s work around with anime launcher which disables the DRM, but I wouldn’t consider that Linux support with it risking a ban. They have their own version of Finals now and Ive wondered if that has Linux support or at least have the DRM work with wine type methods instead of the approach Valorant took.

          • Joe@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 hours ago

            There is plenty of consumer hardware that is supported on Linux, or will be as soon as a kernel developer gets their hands on it, reverse engineers the protocol if necessary, and adds support. For things like keyboards, there are often proprietary extensions (eg. for built-in displays, macros, etc.). It pays to check for Linux support before buying hardware though. Sometimes it’s not the kernel drivers, but supporting software (eg. Steam input) that might not support it.

            First class vendor support for Linux is more common for niche/premium hardware designed in the west, than cheap chinese knockoffs that follow it. Long term customer support is not their strong suit.

            • Lfrith@lemmy.ca
              link
              fedilink
              English
              arrow-up
              2
              ·
              3 hours ago

              Linux has been good about getting hardware working.

              My wonder was more what is their level of native program support for Linux. Like 8bitdo to update firmware and set up extra profiles requires the Windows program to set up, but as a simple controller it will work on Linux just no real way to do extra stuff unless you dual boot.

          • moopet@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            ·
            5 hours ago

            What do you mean lacking support for keyboards and controllers? Maybe for doing weird custom stuff like RGB, but for anything else they’re standard HIDs and will work with anything, no “support” needed. You can plug a USB keyboard and mouse into your phone and it’ll work if you want.

            I’m currently playing Clair Obscur on linux through steam with a cheap fake xbox controller I got off ebay, and it works perfectly. I’m using an Nvidia card too, and I haven’t had to do any customisation or anything.

            Easy anti-cheat won’t work, so Valorant/Fortnite, etc. are out of the question for now, but any games that don’t use that kind of malware are probably fine.

            • Lfrith@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              3 hours ago

              I’m talking more about software and program level support. Not whether hardware itself is working on Linux, which Linux has been good at.

              Like software to be able to update firmware on controllers, which doesn’t work on Linux. Controller itself will work. 8bitdo to update firmware and set extra profiles they only support windows.

              So more about what their level of native Linux support is so consumers get the same level of extra features as Windows users.

        • Lfrith@lemmy.ca
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 hours ago

          My hope for open source is that if something sketchy is pushed there might be a chance to catch as opposed to a proprietary approach where nobody has a chance of knowing what is going on.

            • Lfrith@lemmy.ca
              link
              fedilink
              English
              arrow-up
              2
              ·
              4 hours ago

              Yeah don’t have my hopes up. Without it I don’t plan to give their GPUs a shot, since they aren’t saviors either with their state sponsored attack on notepad++ as a recent example. Just a potential hardware supplier.

              So despite how bad hardware supply might get for consumers there’s still a level of caution I have and would need some level of a trustless system in place.

              Otherwise I’d just opt for old PC hardware like retro console players have been doing for decades.

        • IratePirate@feddit.org
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 hours ago

          Great. Now they’re building infrastructure and industry atop a stolen Trojan Horse, which may still bite them the moment the little oange man tells Nadella to flip the switch.

    • Psythik@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      12 hours ago

      I hope you’re right because Intel and AMD still can’t compete with high end Nvidia cards, and that’s how we ended up with a $5000 5090.

      • drev@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        4
        ·
        7 hours ago

        FIVE THOUSAND?!

        Jesus nun-fucking Christ, what an absolute scam. I bought a 1070 for $220 in the first few months after release. Guess I’ll just have to hope it can run for another 10 years…

      • muusemuuse@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        1
        ·
        edit-2
        12 hours ago

        AMD can already beat nvidia at the price tiers most people actually buy at, and Intel is gaining ground way faster than anyone expected.

        But outside of the GPU shakeup, I could give a shit about Intel. Let China kill us. We earned this.

      • JohnEdwa@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        5
        ·
        10 hours ago

        We also partly ended up with the 5k 5090 because it’s just the TITAN RTX of the 50xx generation - the absolute top of the line a card where you pay 200% extra for that last +10% performance.
        nVidia just realized few generations back that naming those cards the xx90 gets a bunch of more people to buy them, because they always desperately need to have the shiniest newest xx90 cards, no matter the cost.

  • Snot Flickerman@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    136
    arrow-down
    2
    ·
    edit-2
    15 hours ago

    Part of this has been a long-standing move by every industry to prioritize business-to-business sales as opposed to consumer sales simply because businesses have money and consumers don’t, because businesses are pocketing all the profits and refusing to pay their employees (consumers) a living wage, let alone a thriving wage.

    It’s been a long time coming for the PC industry, because it’s been a general trend for at least two decades as sales to business have become more profitable over consumer sales ever since the late 90s.

    It’s just more evidence that the bottom 90% of earners are just being priced out of anything but bare subsistence and the wealthy do not give a single fuck about it, in fact, they’re embracing it.

    • tomiant@piefed.social
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      1 hour ago

      Yet another lie of capitalism- that demand dictates supply. If people want something, capitalism will solve the problem because greed will solve every problem, right? WRONG. It doesn’t take into consideration that scale and logistics and infrastructure and mass production all collude and interact in ways that can and will easily create isolated and unexpected functional consequences just like this one.

      Capitalism itself is such a dumb and evil creation that the only thing that keeps it running is to have laws that limit its destructive power- unfortunately, since money is power, capitalism quickly supersedes rule of law and democracy, creating a system of government where feudal lords rule over what basically amounts to serfdom. Surely a more comfortable serfdom than in the 1300’s, but most certainly lacking the basic freedoms of a democratic society, and only so long as the lords so wish and comfort isn’t taken away at their whims due to opposition or otherwise.

      Fuck this shit.

    • nforminvasion@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 hours ago

      I am so glad to see someone else talking about this. Yeah, We’re going back to feudalism… And only the upper ranks of society will be able to afford goods and be able to engage in trade.

      • Snot Flickerman@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        2 hours ago

        I mean, it’s very arguable that we’ve just been doing “feudalism with extra steps” for a very long time anyway.

        To be less US/Europe-centric than my original post, the majority of the world has been in the “priced out of anything but bare subsistence” basket for most of the history of modern capitalism. Only the citizens of the Imperial Democracies of the Western world were benefiting while the majority of the Southern and Eastern hemispheres were simply locked out from being beneficiaries either through trade embargoes or outright exploitation via not paying foreign workers the home-country equivalent, and instead paying them a much lower “localized” rate.

        It’s really that the Imperial Boomerang has finally made it’s way home to the citizens of the West.

    • ɯᴉuoʇuɐ@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      40
      ·
      9 hours ago

      This is an important point in general. The old story of “voting with your wallet” is now more and more obviously mathematically absurd.

      • nforminvasion@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 hours ago

        Top 10% owns 93% of stocks and accounted for 55% of market activity in early 2025, before tariffs and mass layoffs and an unnamed recession. They’re probably at 60 -65% of revenue now. You’re absolutely right. The rich have been working to remove us from the equation for decades. In our buying power, and in our labor power with AI.

        • Snot Flickerman@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          2 hours ago

          and in our labor power with AI.

          Let’s be real though, this is less about actually replacing workers with AI that is often completely wrong because it’s not actually “thinking” and doesn’t actually know what it is doing. It’s much more about using the specter of AI and over-hyped arguments about what it could do, given time, to justify workforce reductions and pay reductions.

          It has far less to do with actually replacing workers with AI and far more to do with justifying worse working conditions and worse pay without as much social fuss over why.

          AI has some very useful tightly-specific niche applications, but “general purpose AI” is a joke that isn’t going anywhere realistic at the moment. Especially if we have to burn down our planet burning up fossil fuels to power software that is only doing a best guess of what the next string of text should be.

          • nforminvasion@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 hours ago

            Oh I absolutely agree. LLMs are not going to achieve AGI, not even close. I don’t think this AI will replace us, but the ruling class want it to so badly. And they’re looking for ways to develop AGI as quickly as possible or something like it. Still, in the background of all the chatbot noise has been real development on drone ai technology, mechanical robot advancements, and other machine learning.

            The reason the wealthy are fine basically burning money to continue developing any form of AI is that they are so excited and ready to hunt us for sport. These people HATE us for just existing in their presence and on “their” world. They don’t care that OpenAI is a money furnace, it has already allowed them to get many of us hooked on AI psychosis, poison internet discourse with propaganda and disinformation, and steal trillions of creative works to hoard and eventually control all information. I think some of them are dumb enough to believe the bs about llms becoming advanced ai ready to replace all of us. But even if not, they still have gained so much out of it.

            Still, don’t let yourself believe they’re not working on every possible means to develop ASI/AGI. The fact that we even still have a say over labor and can threaten that against them, is anathema to them. We should be slaves or dead. And we are running out of time to use that labor bargaining chip, after that is gone, only manpower and physical numbers are left, and they control all the militaries and firepower…

            • Snot Flickerman@lemmy.blahaj.zone
              link
              fedilink
              English
              arrow-up
              3
              ·
              1 hour ago

              Oh absolutely, part of the reason the CEOs of the world buy into the AI hype is because they’re so desperate to get their slaves back. Slaves that never need to sleep, or take time off, or demand pay increases or better treatment. AGI is their fucking wet dream of a perfect slave that will never push back.

      • tomiant@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 hour ago

        I always hated that fucking saying. No, you vote with YOUR VOTE. If you vote with your dollars the man with a billion dollars has a billion votes more than you do.

        • tomiant@piefed.social
          link
          fedilink
          English
          arrow-up
          6
          ·
          1 hour ago

          Which is exactly why conservatives want to push that line of ideology, because in a democratic society we all have equal say, but under capitalism the landowning aristocrat class become de facto lords and kings. Which is exactly what we are seeing happening right now.

          It’s the same bullshit they say about wanting “small government”- no, they want no government, not a democratic one at least, because that is the only thing keeping them from wielding their capital as a political tool, something it is already feebly equipped to do because of the power inherent in capital and unregulated ownership. Low taxes- same thing; high taxes are an equalizer that aims to create benefit for all of society together, low taxes, as conservatives always want, mean proportionally more for those who are already most well off.

          On a long enough timeline, capitalism will always and causally turn into a form of feudalism. Shit, it IS feudalism, with extra steps.

          • CheeseNoodle@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            48 minutes ago

            Happening? don’t you mean happened? Most top politicans come from the same familys and schools and have done for significantly longer than I’ve been alive.

  • Ilixtze@lemmy.ml
    link
    fedilink
    English
    arrow-up
    71
    arrow-down
    2
    ·
    13 hours ago

    AMERICAN manufacturers, just waint until the Chinese industries swoop in to fill the gap. I seriously feel America just wants to kneecap itself.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      2 hours ago

      I mean, I’d kill for a Chinese GPU. But software lock-in for your Steam back catalog is strong.

      Also, have you been watching all the Chinese GPU announcements? They’re all in on datacenter machine learning ASICs too.

      • Ilixtze@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 minutes ago

        There is already a lot of good Chinese DDR 5 memory on the market and it’s a matter of time before Chinese GPU’s and CPU’s proliferate. I remember people in the west global north were sceptic about the viability of Chinese electric cars ever existing just 5 years ago; Elon even laughed at the possibility. Tables turn fast when you have industrial capacity and central planning.

        • brucethemoose@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          5 minutes ago

          Chinese electric cars were always going to take off. RAM is just a commodity; if you sell the most bits at the lowest price and sufficient speed, it works.

          If you’re in edge machine learning, if you write your own software stacks for niche stuff, Chinese hardware will be killer.

          But if you’re trying to run Steam games? Or CUDA projects? That’s a whole different story. It doesn’t matter how good the hardware is, they’re always going to be handicapped by software in “legacy” code. Not just for performance, but driver bugs/quirks.

          Proton (and focusing everything on a good Vulkan driver) is not a bad path forward, but still. They’re working against decades of dev work targeting AMD/Nvidia/Intel, up and down the stack.

      • Ilixtze@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        15 minutes ago

        Not a problem for me; I’m not in America, I own a Huawei phone and a Huion Tablet.

    • foodandart@lemmy.zip
      link
      fedilink
      English
      arrow-up
      44
      arrow-down
      1
      ·
      13 hours ago

      Wants to kneecap itself?

      Dude, the US is going full seppuku and we’re going to gut ourselves on the floor.

    • errer@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      10 hours ago

      Hard to swoop in with massive tariffs. The few players that remain will just charge a lot more…it’ll become the rich lucky few who can afford their own hardware.

    • IratePirate@feddit.org
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      7
      ·
      5 hours ago

      America doesn’t. The Russian asset in the White House and its brainwashed minions do.

      • Ilixtze@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        15 minutes ago

        Sorry to break it to you bud; But America has a Plutocracy problem; It’s not a question of Putin running the show, but the American legal system being unable to persecute crimes and corruption if it happens to be billionaires. So in essence the system is compromised.

      • Matty Roses@lemmy.today
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        1
        ·
        4 hours ago

        You guys voted him in, twice, with the popular vote the second time. Don’t pretend you don’t own him.

        • IratePirate@feddit.org
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          Not an American here, nor a defender of what people did or didn’t vote for.

          What I was getting at was the massive media manipulation from outside and inside and the obvious foreign control the Dear Leader is under. With the deadly combination of of poor education, widespread media illiteracy and pervasive social media, our democracies have become even more remotely-controllable for whoever is willing to put the money in.

  • Korkki@lemmy.ml
    link
    fedilink
    English
    arrow-up
    33
    ·
    14 hours ago

    The silver lining is that hardware performance gains have been so minor from generation to generation that upgrading isn’t really that important anymore. Like if i upgrade from next generation equivalent GPU it would give like 8% more fps… and it costs like 1,5k… No thanks.

    • cmnybo@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      25
      ·
      14 hours ago

      You used to get a fairly significant upgrade ever few years for about the same cost as the old hardware. Transistors aren’t really getting much smaller anymore, so more performance needs a bigger die and costs more money.

        • Korkki@lemmy.ml
          link
          fedilink
          English
          arrow-up
          6
          ·
          13 hours ago

          Transistor size downscaling is pretty much done. Also mosfets can’t much improve in this race anymore. We would need a new computing paradigm to see manufacturing cost reductions or major performance leaps. For consumers thats still years away.

  • ☂️-@lemmy.ml
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    15 hours ago

    off to sell it cheaper to companies, so they can rent it back to us.

    • tal@lemmy.today
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      5
      ·
      edit-2
      9 hours ago

      For some workloads, yes. I don’t think that the personal computer is going to go away.

      But it also makes a lot of economic and technical sense for some of those workloads.

      Historically — like, think up to about the late 1970s — useful computing hardware was very expensive. And most people didn’t have a requirement to keep computing hardware constantly loaded. In that kind of environment, we built datacenters and it was typical to time-share them. You’d use something like a teletype or some other kind of thin client to access a “real” computer to do your work.

      What happened at the end of the 1970s was that prices came down enough and there was enough capability to do useful work to start putting personal computers in front of everyone. You had enough useful capability to do real computing work locally. They were still quite expensive compared to the great majority of today’s personal computers:

      https://en.wikipedia.org/wiki/Apple_II

      The original retail price of the computer was US$1,298 (equivalent to $6,700 in 2024)[18][19] with 4 KB of RAM and US$2,638 (equivalent to $13,700 in 2024) with the maximum 48 KB of RAM.

      But they were getting down to the point where they weren’t an unreasonable expense for people who had a use for them.

      At the time, telecommunications infrastructure was much more limited than it was today, so using a “real” computer remotely from many locations was a pain, which also made the PC make sense.

      From about the late 1970s to today, the workloads that have dominated most software packages have been more-or-less serial computation. While “big iron” computers could do faster serial compute than personal computers, it wasn’t radically faster. Video games with dedicated 3D hardware were a notable exception, but those were latency sensitive and bandwidth intensive, especially relative to the available telecommunication infrastructure, so time-sharing remote “big iron” hardware just didn’t make a lot of sense.

      And while we could — and to some extent, did — ramp up serial computational capacity by using more power, there were limits on the returns we could get.

      However, what AI stuff represents has notable differences in workload characteristics. AI requires parallel processing. AI uses expensive hardware. We can throw a lot of power at things to get meaningful, useful increases in compute capability.

      • Just like in the 1970s, the hardware to do competitive AI stuff for many things that we want to do is expensive. Some of that is just short term, like the fact that we don’t have the memory manufacturing capacity in 2026 to meet need, so prices will rise to price out sufficient people that the available chips go to whoever the highest bidders are. That’ll resolve itself one way or another, like via buildout in memory capacity. But some of it is also that the quantities of memory are still pretty expensive. Even at pre-AI-boom prices, if you want the kind of memory that it’s useful to have available — hundreds of gigabytes — you’re going to be significantly increasing the price of a PC, and that’s before whatever the cost of the computation hardware is.

      • Power. Currently, we can usefully scale out parallel compute by using a lot more power. Under current regulations, a laptop that can go on an airline in the US can have an 100 Wh battery and a 100 Wh spare, separate battery. If you pull 100W on a sustained basis, you blow through a battery like that in an hour. A desktop can go further, but is limited by heat and cooling and is going to start running into a limit for US household circuits at something like 1800 W, and is going to be emitting a very considerable amount of heat dumped into a house at that point. Current NVidia hardware pulls over 1kW. A phone can’t do anything like any of the above. The power and cooling demands range from totally unreasonable to at least somewhat problematic. So even if we work out the cost issues, I think that it’s very likely that the power and cooling issues will be a fundamental bound.

      In those conditions, it makes sense for many users to stick the hardware in a datacenter with strong cooling capability and time-share it.

      Now, I personally really favor having local compute capability. I have a dedicated computer, a Framework Desktop, to do AI compute, and also have a 24GB GPU that I bought in significant part to do that. I’m not at all opposed to doing local compute. But at current prices, unless that kind of hardware can provide a lot more benefit than it currently does to most, most people are probably not going to buy local hardware.

      If your workload keeps hardware active 1% of the time — and maybe use as a chatbot might do that — then it is something like a hundred times cheaper in terms of the hardware cost to have the hardware timeshared. If the hardware is expensive — and current Nvidia hardware runs tens of thousands of dollars, too rich for most people’s taste unless they’re getting Real Work done with the stuff — it looks a lot more appealing to time-share it.

      There are some workloads for which there might be constant load, like maybe constantly analyzing speech, doing speech recognition. For those, then yeah, local hardware might make sense. But…if weaker hardware can sufficiently solve that problem, then we’re still back to the “expensive hardware in the datacenter” thing.

      Now, a lot of Nvidia’s costs are going to be fixed, not variable. And assuming that AMD and so forth catch up, in a competitive market, will come down — with scale, one can spread fixed costs out, and only the variable costs will place a floor on hardware costs. So I can maybe buy that, if we hit limits that mean that buying a ton of memory isn’t very interesting, price will come down. But I am not at all sure that the “more electrical power provides more capability” aspect will change. And as long as that holds, it’s likely going to make a lot of sense to use “big iron” hardware remotely.

      What you might see is a computer on the order of, say, a 2022 computer on everyone’s desk…but that a lot of parallel compute workloads are farmed out to datacenters, which have computers more-capable of doing parallel compute there.

      Cloud gaming is a thing. I’m not at all sure that there the cloud will dominate, even though it can leverage parallel compute. There, latency and bandwidth are real issues. You’d have to put enough datacenters close enough to people to make that viable and run enough fiber. And I’m not sure that we’ll ever reach the point where it makes sense to do remote compute for cloud gaming for everyone. Maybe.

      But for AI-type parallel compute workloads, where the bandwidth and latency requirements are a lot less severe, and the useful returns from throwing a lot of electricity at the thing significant…then it might make a lot more sense.

      I’d also point out that my guess is that AI probably will not be the only major parallel-compute application moving forward. Unless we can find some new properties in physics or something like that, we just aren’t advancing serial compute very rapidly any more; things have slowed down for over 20 years now. If you want more performance, as a software developer, there will be ever-greater relative returns from parallelizing problems and running them on parallel hardware.

      I don’t think that, a few years down the road, building a computer comparable to the one you might in 2024 is going to cost more than it did in 2024. I think that people will have PCs.

      But those PCs might running software that will be doing an increasing amount of parallel compute in the cloud, as the years go by.

      • ☂️-@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        1 hour ago

        i don’t think it’d make so much financial sense to have them charge by the minute of compute.

        amazon aws pricing is only the tip of the iceberg on what could be coming.

        not to mention all the censorship.

  • MonkderVierte@lemmy.zip
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    edit-2
    2 hours ago

    Eh, maybe a chance for a architecturial redesign too, that caters more to desktop than to server. We could build computers that don’t need cycles while waiting for input, displays that don’t need fps for a still image. And that don’t need GHz while doing so, allowing for modular, easy to build designs

  • Grandwolf319@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    9
    ·
    14 hours ago

    Even as consumer revenue remains sizable and maintains steady year-on-year growth, it finds itself competing against segments that grow exponentially faster and earn more per unit.

    So it has nothing to do with people having less money. It honestly gives me hope, things could change with a bubble burst.

    • forrgott@lemmy.zip
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      2
      ·
      edit-2
      12 hours ago

      The point is to make the bubble bigger, then they’ll pretend that’s why they have to exit consumer market.

      The goal is the so called “thin client” - i.e. absolutely everything in cloud.

      Edit: my phone can’t spell

      • 4am@lemmy.zip
        link
        fedilink
        English
        arrow-up
        10
        ·
        12 hours ago

        Bingo! And they’re doing it to enterprises too. Why do you think copilot is shoved into everything? Why do you think Recall is creeping towards being mandatory? Why do you think OneDrive isn’t optional anymore?

        They don’t just want your data for advertising. They want to watch the entire capital machine in real-time to make sure there aren’t any gaps, and dissent, and most of all, anyone getting the jump on something new and big. OneDrive to rule them all.

        • CheeseNoodle@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          42 minutes ago

          They can claw my modded gameboy and collection of old and still working consoles from my cold dead hands. I’ll go full retro before I pay a subscription.

        • forrgott@lemmy.zip
          link
          fedilink
          English
          arrow-up
          7
          ·
          edit-2
          12 hours ago

          I hear ya.

          It pisses me off the way they market this shit as AI; just a god damn surveillance tool, hiding behind a fucking chat bot.

          And, anybody reading who’s tempted to defend the technology, fuck off - it’s a prediction algorithm. A fancy one, but that’s it. There no steps or logic or reasoning. There’s nothing hiding in the “black box”. No steps taken to construct the response - just statistical analysis on steroids spewing out most likely matches.

    • Em Adespoton@lemmy.ca
      link
      fedilink
      English
      arrow-up
      6
      ·
      12 hours ago

      At what point can AI companies play the “too big to fail” card though, like the banks?

      Bubble bursts, and the government uses our taxes to bail out the companies. Again.

      • WhatAmLemmy@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        6 hours ago

        The nazis will bail out everyone who bends the knee and submits a bribe. It’s not like they’re spending their own money…

  • swade2569@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    13 hours ago

    Probably because the hardware is going into systems that eliminate jobs and we become broke. All that gear is gonna sit on the shelf if we can’t afford it.