Context, then answer… instead of having everything ride on the first character (e.g. we make it pick “Y” or “N” first in response to a yes-or-no question, it usually picks “Y” even if it later talks itself out of it).

  • TheLeadenSea@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    9
    ·
    18 hours ago

    That’s the basis of reasoning models. Make LLMs ‘think’ through the problem for several hundred tokens before giving a final answer.