Hacker News new | past | comments | ask | show | jobs | submit login

LLM cheating detection is an interesting case of the toupee fallacy.

The most obvious ChatGPT cheating, like that mentioned in this article, is pretty easy to detect.

However, a decent cheater will quickly discover ways to conduce their LLM into producing text that is very difficult to detect.

I think if I was in the teaching profession I'd just leave, to be honest. The joy of reviewing student work will inevitably be ruined by this: there is 0 way of telling if the work is real or not, at which point why bother?






You assume that the teachers job is to catch when someone is cheating; its not. The teachers job is to teach, and if the kids don't learn because their parents allow them to cheat, don't check them at all, and let them behave like shitheads, then the kids will fail in life.

> then the kids will fail in life.

Quite the assertion. If anything the evidence is in favor of the other direction.

It was eye opening to see that most students cheat. By the same token, most students end up successful. It’s why everyone wants their kids to go to college.


In many current-day school systems, the teachers job is to get the required percentage of students to pass the state assessment for their grade level.

They don’t get an exemption if the parents don’t care.


This isn't the way reality works.

Or, bad money chases out good. Idiots that cheat will get the recommendations for jobs where by maxing the grade. The person that actually works gets set back. Even worse society at large loses and actually educated person. And lastly a school is going to attempt to protect their name by preventing cheating.


On reviewing students' work: people exchange copies, get their hands on past similar assignments, get friends to do their homework , potentially each of them shadow the other in fields they're good at etc.

There always was a bunch of realistic options to not actually do your submitted work, and AI is merely makes it easier, more detectable and more scalable.

I think it moves the needle from 40 to 75, which is not great, but you'd already be holding your nose at student work half of the time before AI, so teaching had to be about more than that (and TBH it was, when I was in school teachers gave no fuck about submitted work if they didn't validate it by some additional face to face or test time)


> a decent cheater will quickly discover ways to conduce their LLM into producing text that is very difficult to detect

Do you have any examples of this? I've never been able to get direct LLM output that didn't feel distinctly LLM-ish.


this immediately comes to mind https://regmedia.co.uk/2025/04/29/supplied_can_ai_change_you...

A study on whether LLMs can influence people on r/changemymind


This only came to light after the study had already been running for a few months. That proves that we can no longer tell for certain unless it's literal GPT-speak the author was too lazy to edit themselves.

Teachers will lament the rise of AI-generated answers, but they will only ever complain about the blatantly obvious responses that are 100% copy-pasted. This is only an emerging phenomenon, and the next wave of prompters will learn from the mistakes of the past. From now on, unless you can proctor a room full of students writing their answers with nothing but pencil and paper, there will be no way to know for certain how much was AI and how much was original/rewritten.


> This only came to light after the study had already been running for a few months. That proves that we can no longer tell for certain unless it's literal GPT-speak the author was too lazy to edit themselves.

Rule 3 of the subreddit quite literally bars people from accusing posts of being AI-generated. I have only visited it a few times in recent times, but I noticed quite a few GPT-speak posts with comments calling it out getting removed and punished.


Maybe it will get us to rethink the grading system. Do we need to grade them, or do we need students to learn things? After all, if they grow up to be incompetent, they will be the ones suffering from it.

But I know it's easier said than done: if you get a student to realise that the time they spend at school is a unique opportunity for them to learn and grow, then you're job is almost done already.


> there is 0 way of telling if the work is real or not, at which point why bother?

I might argue you couldn't really tell if it was "real" before LLMs, either. But also, reviewing work without some accompanying dialogue is probably rarely considered a joy anyway.


> there is 0 way of telling if the work is real or not

Talk to the student, maybe?

I have been an interviewer in some startups. I was not asking leetcode questions or anything like that. My method was this: I would pretend that the interviewee is a new colleague and that I am having coffee with them for the first time. I am generally interested in my colleagues: who are they, what do they like, where do they come from? And then more specifically, what do they know that relates to my work? I want to know if that colleague is interested in a topic that I know better, so that I could help them. And I want to know if that colleague is an expert in a topic where they could help me.

I just have a natural discussion. If the candidate says "I love compilers", I find this interesting and ask questions about compilers. If the person is bullshitting me, they won't manage to maintain an interesting discussion about compilers for 15 minutes, will they?

It was a startup, and the "standard" process became some kind of cargo culting of whatever they thought the interviews at TooBigTech were like: leetcode, system design and whatnot. Multiple times, I could obviously tell in advance that even if this person was really good at passing the test, I didn't think it would be a good fit for the position (both for the company and for them). But our stupid interviews got them hired anyway and guess what? It wasn't a good match.

We underestimate how much we can learn by just having a discussion with a person and actually being interested in whatever they have to say. As opposed to asking them to answer standard questions.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: