• corroded@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    ·
    2 days ago

    They say they did this by “finetuning GPT 4o.” How is that even possible? Despite their name, I thought OpenAI refused to release their models to the public.

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      edit-2
      2 days ago

      They kind of have to now though. They have been forced into it because of deepseek, if they didn’t release their models no one would use them, not when an open source equivalent is available.

      • corroded@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        2 days ago

        I feel like the vast majority of people just want to log onto Chat GPT and ask their questions, not host an open source LLM themselves. I suppose other organizations could host Deepseek, though.

        Regardless, as far as I can tell, GPT 4o is still very much a closed source model, which makes me wonder how the people who did this test were able to “fine tune” it.

        • Echo Dot@feddit.uk
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 day ago

          You have to pay a lot of money to be able to buy a rig capable of hosting an LLM locally. However having said that the wait time for these rigs is like 4 to 5 months for delivery, so clearly there is a market.

          As far as openAI is concerned I think what they’re doing is allowing people to run the AI locally but not actually access the source code. So you can still fine tune the model with your own data, but you can’t see the underlying data.

          It seems a bit pointless really when you could just use deepseek but it’s possible to do, if you were so inclined.