Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I believe we're already using llms to evaluate llm output for training, I wonder if there's some variation of that which could be used to identify when one llm gets "stuck".

I guess chain of thought in theory should do that but having variations on prompt and context might behave differently?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: