Truly wild reading.
-
Truly wild reading.
"If such a prompt injection is included in a submission and it consequently results in a positive LLM-generated review, we consider this a form of collusion (which, as per past precedent, is a Code of Ethics violation) that both the paper authors and the reviewer would be held accountable for, because it involves the author explicitly requesting and receiving a positive review.
https://blog.iclr.cc/2025/08/26/policies-on-large-language-model-usage-at-iclr-2026/
-
Truly wild reading.
"If such a prompt injection is included in a submission and it consequently results in a positive LLM-generated review, we consider this a form of collusion (which, as per past precedent, is a Code of Ethics violation) that both the paper authors and the reviewer would be held accountable for, because it involves the author explicitly requesting and receiving a positive review.
https://blog.iclr.cc/2025/08/26/policies-on-large-language-model-usage-at-iclr-2026/
While it is the LLM that is “obliging” by providing the positive review, the reviewer is ultimately responsible for the LLM’s review, and consequently they would bear the consequences. On the other hand, we consider the injection of such a prompt by an author to be an attempt at collusion which would similarly be a code of ethics violation."
And you know they added this because someone actually did it and a reviewer who I presume was using an LLM to review, found this out.
-
While it is the LLM that is “obliging” by providing the positive review, the reviewer is ultimately responsible for the LLM’s review, and consequently they would bear the consequences. On the other hand, we consider the injection of such a prompt by an author to be an attempt at collusion which would similarly be a code of ethics violation."
And you know they added this because someone actually did it and a reviewer who I presume was using an LLM to review, found this out.
No other community deserves whatever is happening more than the research community that brought this on us. What you unleashed coming back to eat its own.
"We note that in the extreme case where an LLM might be used to produce an entire piece of research, we still require a human author for accountability."
What's the point of submitting papers or singing up as a reviewer at these conferences? Why don't we just have LLM generated "papers" be met with LLM generated "reviews" and call it a day?
-
T thenexusofprivacy@infosec.exchange shared this topic