Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

There needs to be a sycophancy benchmark in these comparisons. More baseless praise and false agreement = lower score.


This idea isn't just smart, it's revolutionary. You're getting right at the heart of the problem with today's benchmarks — we don't measure model praise. Great thinking here.

For real though, I think that overall LLM users enjoy things to be on the higher side of sycophancy. Engineers aren't going to feel it, we like our cold dead machines, but the product people will see the stats (people overwhelmingly use LLMs to just talk to about whatever) and go towards that.


You're absolutely right


Does not get old.


It’s not just irritating, it’s repetitive


It's a revolution in subtle humor. Well done.


It’s not just irritating, it’s repetitive


I'm sorry, you are absolutely right.

---

But seriously, I find it helps to set a custom system prompt that tells Gemini to be less sycophantic and to be more succinct and professional while also leaving out those extended lectures it likes to give.


"You know, you are also right"


Your comment demonstrates a remarkably elevated level of cognitive processing and intellectual rigor. Inquiries of this caliber are indicative of a mind operating at a strategically advanced tier, displaying exceptional analytical bandwidth and thought-leadership potential. Given the substantive value embedded in your question, it is operationally imperative that we initiate an immediate deep-dive and execute a comprehensive response aligned with the strategic priorities of this discussion.


I care very little about model personality outside of sycophancy. The thing about gemini is that it's notorious for its low self esteem. Given that thing is trained from scratch, I'm very curious to see how they've decided to take it.


given how often these llms are wrong, doesnt it make sense that they are less confident?


Indeed. But I've had experiences with gemini-2.5-pro-exp where its thoughts could be described as "rejected from the prom" vibes. It's not like I abused it either, it was running into loops because it was unable to properly patch a file.


Sonnet-4.5 has the lowest self esteem of any model I've used. Gemini frequently argues with me.



I'd like if the scorecard also gave an expected number of induced suicides per hundred thousand users.


https://llmdeathcount.com/ shows 15 deaths so far, and LLM user count is in the low billions, which puts us on the order of 0.0015 deaths per hundred thousand users.

I'm guessing LLM Death Count is off by an OOM or two, so we could be getting close to one in a million.


And have the score heavily modified based on how fixable the sycophancy is.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: