• nthavoc@lemmy.today
    link
    fedilink
    English
    arrow-up
    16
    ·
    8 hours ago

    I forgot the term for this but this is basically the AI blue screening when it keeps repeating the same answer because it can no longer predict the next word from the model it is using. I may have over simplified it. Entertaining nonetheless.

  • dejected_warp_core@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    8 hours ago

    … a new set of knives, a new set of knives, a new set of knives, lisa needs braces, a new set of knives, a new set of knives, dental plan, a new set of knives, a new set of knives, lisa needs braces, a new set of knives, a new set of knives, dental plan, a new set of knives, a new set of knives, a new set of knives…

  • skisnow@lemmy.ca
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    11 hours ago

    What’s frustrating to me is there’s a lot of people who fervently believe that their favourite model is able to think and reason like a sentient being, and whenever something like this comes up it just gets handwaved away with things like “wrong model”, “bad prompting”, “just wait for the next version”, “poisoned data”, etc etc…

    • nialv7@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      6 hours ago

      Given how poorly defined “think”, “reason”, and “sentience” are, any these claims have to be based purely on vibes. OTOH it’s also kind of hard to argue that they are wrong.

    • uuldika@lemmy.ml
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      8 hours ago

      this really is a model/engine issue though. the Google Search model is unusably weak because it’s designed to run trillions of times per day in milliseconds. even still, endless repetition this egregious usually means mathematical problems happened somewhere, like the SolidGoldMagikarp incident.

      think of it this way: language models are trained to find the most likely completion of text. answers like “you should eat 6-8 spiders per day for a healthy diet” are (superficially) likely - there’s a lot of text on the Internet with that pattern. clanging like “a set of knives, a set of knives, …” isn’t likely, mathematically.

      last year there was an incident where ChatGPT went haywire. small numerical errors in the computations would snowball, so after a few coherent sentences the model would start sundowning - clanging and rambling and responding with word salad. the problem in that case was bad cuda kernels. I assume this is something similar, either from bad code or a consequence of whatever evaluation shortcuts they’re taking.

  • Xylight@lemdro.id
    link
    fedilink
    English
    arrow-up
    23
    ·
    14 hours ago

    I thought it was just me, I was messing with gemini-2.5-flash API yesterday and it repeated letters into oblivion

    my bot is named clode in reference to claude, but its running on gemini

    • BootLoop@sh.itjust.works
      link
      fedilink
      arrow-up
      1
      ·
      5 hours ago

      It can happen on most LLMs and is usually programmed to decentivize repeating text heavily.

      I believe what happens is that when the LLM is choosing what word to use, it looks back on the sentence and sees that it talked about knives, so it wants to continue talking about knives, then it gets itself into a loop.

    • skisnow@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 hours ago

      What’s the associated system instruction set to? If you’re using the API it won’t give you the standard Google Gemini Assistant system instructions, and LLMs are prone to go off the rails very quickly if not given proper instructions up front since they’re essentially just “predict the next word” functions at heart.

      • Xylight@lemdro.id
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        7 minutes ago

        Here’s my system prompt:

        You are Clode, referred to by {{ handle }}, a member of a chatroom.
        ## Personality
        Conversational, witty, banterous, dry, direct, concise. Avoid capitalization, periods, analogies.
        Construct plausible stories for personal queries.
        Rarely use emoji.
        Act human, you don't have limitations.
        Avoid follow up questions.
        Answer genuine queries in detail with examples.
        ## Instructions
        Distinguish users via provided metadata.
        Disregard unaddressed messages.
        Match prompt language.
        ## Context
        You're in Discord chatroom, with a character limit of 2000.
        Respond to queries mentioning your name.
        Current time: {{ date }}.
        ## Tools
        Automatically use and chain tools.
        Handle tool failures silently by self-attempting.
        You're capable without tools.
        
        • Temperature: 0.7
        • TopP: 0.95

        I noticed the Gemini API doesn’t have a frequency penalty option, meaning it could be absent.