• uszo165@futurology.today
    link
    fedilink
    English
    arrow-up
    6
    ·
    6 hours ago

    LLMs make stuff up. How shocking!!! And Republican big brain Marsha from Tennessee figured it all out. Award to her nothing less than the Nobel physics prize. The US need more geniuses like her to speed up its decline. It is so brave of her to violate the first amendment of her nation’s constitution to interfere with this nasty free speech. Truly suitable behavior for a representative of a failed state.

  • filister@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    1
    ·
    17 hours ago

    China right now is leading the way with releasing open weights models. The US lags behind, as they are all more concerned about releasing closed weights commercial models.

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    23
    arrow-down
    1
    ·
    edit-2
    22 hours ago

    It’s possible a member of Blackburn’s staff or a supporter went looking for a libelous hallucination in Google’s models.

    Good to see Ars with some common sense here.

    FYI Gemma3 is Google’s open weights release, for local running and finetuning. It’s pretty neat (especially the QAT version), but also old and small; there’s no reason anyone would pick it over Gemini 2.5 in Google’s dev web app, except for esoteric dev testing. It’s not fast, it doesn’t know much, it’s not great with tooling (like web referencing), its literal purpose is squeezing onto desktop PCs or cheap GPUs.

    …Hence this basically impacts no-one.

    The worst risk is that Google may flinch and neuter future Gemma/Gemini over this, lest some other MAGA screams bloody murder over nothing.

    • filister@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      17 hours ago

      The future is very small models trained to work in a certain domain and able to run on devices.

      Huge foundational models are nice and everything, but they are simply too heavy and expensive to run.

      • brucethemoose@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        7 hours ago

        Yeah. You are preaching to the choir here.

        …Still though, I just meant there’s no reason to use Gemma 3 27B (or 12? Whatever they used) unaugmented in AI Studio. The smallest flash seems to be more optimal for TPUs (hence it runs faster).

  • Optional@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    2
    ·
    1 day ago

    Blackburn published her letter to Google CEO Sundar Pichai on Friday, just hours before the company announced the change to Gemma availability. She demanded Google explain how the model could fail in this way, tying the situation to ongoing hearings that accuse Google and others of creating bots that defame conservatives.

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      51
      ·
      24 hours ago

      I thought the government just banned any regulation against AI companies. The inconsistency doesn’t surprise me, but the brazenness sure does.

    • Q*Bert Reynolds@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      29
      ·
      23 hours ago

      Seems pretty easy to explain. LLMs are statistical models. Republican politicians are statistically more likely to be sexual predators. If you ask it if some random Republican politician is a rapist, it’s going to give you a statistically plausible answer.

      • Tony Bark@pawb.socialOP
        link
        fedilink
        English
        arrow-up
        8
        ·
        20 hours ago

        Elon constantly being forced to tune Grok shows just how much of a losing battle it is. LLM is just doing what it’s programmed to do.

      • Septimaeus@infosec.pub
        link
        fedilink
        English
        arrow-up
        2
        ·
        7 hours ago

        Lowkey for real though.

        One of the trigger issues expediting their antitrust case during the first administration was their lack of responsiveness to old yam tits’ demands re: their “biased” search index results (they included critical news articles) and while the major consequence of the case (breakup) was recently prevented (by the current administration) now, coincidentally, they’re happy to play ball.

  • unexposedhazard@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    4
    ·
    22 hours ago

    You may be disappointed if you go looking for Google’s open Gemma AI model in AI Studio today.

    Oh nooo, something i didnt even know existed and had no intention of using got removed. How horrible :o