They use HBM (High Bandwidth Memory). PCs, laptops and phones don’t use this type of RAM.

  • Appoxo@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    18
    ·
    2 days ago

    Production capacity.
    Somewhere I read that to produce one HBM module you could produce three regular DDR modules

  • CaptainBasculin@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    134
    ·
    3 days ago

    Manufacturers have a certain amount of chips they can manufacture, let’s say they can manufacture 10 million chips per year. Normally they adjust for demand, like manufacture 7 million chips for consumers and 3 million for enterprise customers. Company A contacts them and says “We need 6 million chips for this year, here’s the money”, But the factories still can manufacture only 10 million chips in total, so they adjust their factories to manufacture more enterprise focused chips, decreasing the amount of chips manufactured used by consumers.

    • Hawke@lemmy.world
      link
      fedilink
      arrow-up
      46
      arrow-down
      1
      ·
      3 days ago

      I think it’s more like “we need 6 million chips for this year, and we’ll pay you eventually, maybe”.

      But yeah the outcome is the same.

        • Hawke@lemmy.world
          link
          fedilink
          arrow-up
          19
          ·
          3 days ago

          Last time I went round with someone on this, they insisted that this is just normal business procedure to order product on credit.

          To me, while trade credit is definitely a thing, for orders this large I expect there’s some more substantial backing for it.

          Ultimately it probably fits in somewhere on a chart of financial shenanigans like this one:

          • null@piefed.nullspace.lol
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            3 days ago

            That’s more what I was wondering. I’m sure most big customers can get a contract that lets them pay over time, but this particular bubble feels significantly out of the ordinary.

            If I’m a RAM manufacturer, there’s got to be some kind of guarantee that makes me me confident enough that I’ll see that money in the end if I’m putting that many eggs in one basket.

              • null@piefed.nullspace.lol
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                2 days ago

                But what is motivating RAM manufacturers to sign that contract? Why can the commenter above figure it out, but they can’t?

                • village604@adultswim.fan
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  2 days ago

                  Because the contract would be an agreement for the purchasing company to pay for the products they ordered to be manufactured.

  • Treczoks@lemmy.world
    link
    fedilink
    arrow-up
    43
    ·
    3 days ago

    Because the manufacturers use the facilities that once produced DDR5 RAM chips to produce HBM chips instead.

  • coolie4@lemmy.world
    link
    fedilink
    arrow-up
    51
    ·
    3 days ago

    Without even getting into the electronics similarities, they use common raw materials and manufacturing facilities. Diverting resources to one lowers supply of the other, affecting costs.

  • brucethemoose@lemmy.world
    link
    fedilink
    arrow-up
    17
    ·
    edit-2
    2 days ago

    To add to what others said:

    LPDDRX is used in some inference hardware. The same stuff you find in laptops and smartphones.

    Also, the servers need a whole lot of regular CPU DIMMs since they’re still mostly EPYC/Xeon severs with 8 GPUs in each. And why are they “wasting” so much RAM on CPU RAM that isn’t really needed, you ask? Same reason as a lot of AI: it’s immediately accessible, already targeted by devs, and AI dev is way more conservative and wasteful than you’d think.

    Same for SSDs. Regular old servers (including AI servers) need it too. In a perfect world they’d use centralized storage for images/weights with near-“diskless” inference/training servers. Some AI servers do this, but most don’t.


    Basically, the waste is tremendous, for the same reason they use cheap gas generators on-site: it’s faster-to-market.

  • gdog05@lemmy.world
    link
    fedilink
    arrow-up
    27
    ·
    3 days ago

    The better question is, why are they doing all of this without an actual purchase contract?

    • kbal@fedia.io
      link
      fedilink
      arrow-up
      26
      ·
      3 days ago

      Because it’s AI, haven’t you heard? Does it make sense for the business? Who cares, it’s AI. Is it financially sustainable? Dude, it’s AI though. Will there be any customers for any of it? The AI says there will be. You’ve got to understand, this is AI we’re talking about. It’s the AI revolution that will transform the world. We’ve got to bet everything on the AI, or we’ll be left out of the AI future. I asked the AI and it was very clear about that.

      • Asafum@lemmy.world
        link
        fedilink
        arrow-up
        20
        ·
        3 days ago

        Hey @grok is this true?

        Grok: yes, I am very great. I am good at everything because I am AI. Also, fuck the Jews.

  • Robin@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    3 days ago

    Besides what others have already said, HBM is only used for the GPUs. These AI servers also use regular DDR5 chips, just with an extra EEC chip.

  • kbal@fedia.io
    link
    fedilink
    arrow-up
    12
    ·
    3 days ago

    The type of RAM that they use is different in that it takes up even more of all the things that would otherwise be used to produce the RAM that you use.

  • kmirl@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    3 days ago

    If RAM and GPUs were cheap people like us would be more likely to set up local LLMs to prevent our data from being productized by power-grabbing corporations.

      • kmirl@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 days ago

        Not claiming it’s the reason since it clearly isn’t, only that it will help drive traffic to commercial AI products.

        • village604@adultswim.fan
          link
          fedilink
          English
          arrow-up
          4
          ·
          2 days ago

          I think it’s more likely that they’re setting up to push VDI.

          The vast majority of consumers would not be able to set up a local LLM, and they know the people who are able to do so aren’t going to use their services in the first place.