Menu
Log in


Using LLM for academic reviews

  • 7 May 2025 1:57 PM
    Reply # 13496203 on 13488420

    hmmm, good point Ben. I guess it is a proofing service!

    I feel like a pre-reviewed copy both reviewers get makes the most sense? But then I guess someone has to review the LLM's review ..... so now they have to comment on that too.  So almost like a 3rd reviewer which does a 1st pass to proof it?

    If it's available to researchers to use prior to submission one would hope they use it so it doesn't find much - so maybe not much of an imposition particulalry if it removes a lot of the mundine.


  • 7 May 2025 12:48 PM
    Reply # 13496198 on 13496166
    Chris Howden wrote:

    I think LLM’s could be very useful in picking up common methodological and statistical mistakes as well as deviations from the journals style. Made even more useful if each journal trained up their own and made it available to people submitting papers.


    That's a really interesting idea! I'd be fascinated to see how journals would implement that - would they run a pre-review through the LLM and provide the pre-reviewed copy to the reviewers? Or get the reviewer to do it themselves? It sounds almost like a proofing service.

  • 7 May 2025 10:48 AM
    Reply # 13496166 on 13488420

    I think LLM’s could be very useful in picking up common methodological and statistical mistakes as well as deviations from the journals style. Made even more useful if each journal trained up their own and made it available to people submitting papers. It could really boost the quality of statistics if done right! It’s exactly what LLM’s are good at, picking up common mistakes which are highly represented in the training set.

    What it can’t do very well though (if at all) is provide useful commentary on new ideas or discoveries – particularly in the context of a fast moving field. There simply isn’t enough information in the training set. It would be spectacularly bad at anything that goes against conventional wisdom too, which would be disastrous for pretty much every domain I can think of. Bad in science, but imagine its impact on philosophy, history or the different interpretations of art?? What are all the art critics going to argue about!!?? ;)

    So seems to me that LLM’s have the potential to be very useful in taking out the mundane from reviewing and also improving statistical methodology. Which would leave reviewers the task of focusing on the intellectual contribution of the paper. Which sounds like the fun bit really!!!


  • 6 May 2025 12:30 PM
    Reply # 13495695 on 13488420

    I totally agree that the current peer review system is poor, but I'm not convinced that LLMs are the solution. Maybe for things like formatting your review, fixing typos, changing the tone - but I'd draw the line at completing a full peer review.

    If LLMs are the only ones performing peer review, it feels like a slippery slope toward LLMs also being the only ones generating evidence syntheses - at which point I'd wonder what's the point of even writing scientific papers any more! 

  • 6 May 2025 2:35 AM
    Reply # 13495446 on 13488420
    While I understand your points, Ben, we need to consider how poor the current peer review system really is. I suspect only a small proportion of papers are getting a competent review. I do think the future is AI-assisted reviews in some form. Obviously, there are things to be worked out, such as privacy, etc., but after a recent review experience, I was left wondering how many reviewers actually even read what they were reviewing. I strongly suspect Chat GPT would have done a much better job. This would be an interesting research project!
  • 5 May 2025 11:14 AM
    Reply # 13495200 on 13488420

    Using LLMs to provide peer review defeats the purpose of peer review IMO, the point is to get expert human review of your work in the context of broader scientific knowledge. Also, LLMs aren't our peers. I'd be disappointed if a reviewer used an LLM to review my work (and alarmed if they uploaded my unpublished work).

    That said, without it being blindingly obvious that an LLM was used (such as the review including text like "Yes I can do that, here's an improved version of your review"), I would be very hesitant to claim the reviewers used LLMs in their reviews. It feels difficult to prove, especially when the reviews are not independent - they are both reviews of the same piece of work. The intro text where they describe what your paper is about feels robotic but seems to be an academic norm. If the work does indeed have the issues identified by one reviewer, it shouldn't be surprising that the other reviewer also picks up on the same issues.

    Whether the things like critiquing your work for issues that are either non-existent in it or already addressed by the text are the result of LLM usage or human error/laziness is also difficult to say. Humans make mistakes too and I find it more useful to be generous in your assumptions. Replying under the assumption that a mistake was made, it gives the reviewer/editor an 'out' and hopefully leads to a productive conversation where maybe you can get to your desired outcome.

    Last modified: 7 May 2025 12:49 PM | Ben Harrap
  • 1 May 2025 9:34 AM
    Reply # 13493761 on 13488420

    I don't really want to comment but hate being part of an anodyne forum - I could comment on that at length too - but if Chris can point me at anywhere that is not anodyne these days etc

    I have strong views about the level of most people's knowledge of LLMs to use them for such important professional tasks - I could comment at lenght on that too but etc.

    According to Chat GPT those two reviews had different style and appeared not to be from the same source but could possibly have been through some later review process leading to similatirites

    PS Thanks to Beth for posting the clause that says we can engage in controversy

    “I do not seek controversy for the sake of noise, but when truth is buried beneath the weight of silence, I have no choice but to speak—bearing only the evidence, and nothing more.”

    Last modified: 1 May 2025 11:02 AM | Duncan Lowes
  • 30 Apr 2025 1:06 PM
    Reply # 13493329 on 13488420

    Hi Chris,


    While I understand your frustration with the slow response/limited engagement, I'd like to kindly suggest reconsidering your approach.

    Publicly threatening to contact publishers to have an editor removed for "incompetence" could potentially be viewed as casting doubt on another professional's competence without good cause, which may conflict with the SSA Code of Conduct.

    3.4.13

    "Whilst members of the Society are free to engage in controversy, no member shall cast doubt on the professional competence of another without good cause."


  • 17 Apr 2025 11:36 AM
    Message # 13488420

    I have heard that referees are using LLM’s to “speed up” the reviewing process. I recently received two reviews from SPL. The paper is just a silly little result which works out very well, so I wanted to send it somewhere. The reviews are attached.

    These two reviews do not seem independent to me. I have annotated (and sent back to the editor). Check the attached if you are interested. Some comments indicate that the reviewer really did not read even the abstract of the paper. (Yet it looks like some human input has been added).

    Would LLM’s produce identical reviews or just similar? I guess it depends a lot on the prompt and the exact model being used.

    SPL (in rejecting the paper) also sent me a list of suggested journal where I might sent it and they would facilitate transfer.

    • Journal of Multivariate Analysis (there is nothing multivariate in my paper).
    • International Journal of Approximate Reasoning
    • Informatics in Medicine Unlocked (open access, costs $3000, seems predatory)
    • Results in Applied Mathematics (open access, costs $2500, seems predatory)
    • Franklin Open (what the hell?) (open access, costs $3000, seems predatory)

    This list was also surely generated by a LLM. No academic with a passing knowledge of statistics or data science who has seen the abstract (or even title) of my paper would suggest this.

    Can LLM’s be used for good instead of evil?

    An experienced colleague, who is pretty bullish on AI, recently had two LLM’s produce hostile reports of his own unpublished research. Some of it was nonsense, but it did alert him to some clear weaknesses which allowed him to revise. It also revealed some weaknesses that he could not address. (He told me he just hopes that the referees do not use the same LLM!).



    2 files
Powered by Wild Apricot Membership Software