Skip to main content

Writing That is Language Now?

November 15, 2025

A follow-up to "Writing That's Not Language (Yet)."

Recently, a nonsense LLM-generated paper landed in the top 17% of ICLR submissions, with two reviewers giving it a score of 8/10. Although it was flagged by a third reviewer and later desk-rejected, this kicked up a flurry of outrage among AI researchers. Personally, I was struck by the review that initially flagged the paper as AI-generated:

I found this paper very difficult to read and comprehend. Beginning with the abstract—which conveys almost no meaningful insight to non-expert readers—the paper remains largely opaque throughout. It introduces numerous topics without proper context or explanation, ultimately leading to the unsubstantiated claim that a “verification framework” has been established.
There are two possible explanations for this lack of clarity: (1) The paper may be written in an extremely dense and narrow style, understandable only to experts working directly in this specific subarea (which I am not), or (2) The extensive use of LLM-assisted writing tools may have resulted in text that appears technically sophisticated but lacks genuine substance or coherence.

As soon as I saw the phrase “appears technically sophisticated but lacks genuine substance,” I felt that there was something interesting here. The frustration here was palpable—I had experienced the exact same thing years before.

Writing That’s Not Language

In 2020, I close-read approximately 118 AI-generated documents on cannabis legalization (plus an equal number of human-written ones). It was a painstaking task: I needed to classify the lexical aspectual class of every clause, mark coherence relations between clauses, and rate the argumentation quality of each document. But there were two things that made this difficult. First, I didn’t know which articles were human-written and which were AI-generated. Second, the AI-generated documents were extremely uncanny. Take this sentence, for example:

If weed’s not really a public health issue and you’re really happy about it, get an understanding about the ways in which it will be able to influence your behaviour.

Is this someone’s Reddit comment, posted without a second thought? Or is it a semi-competent language model’s attempt to imitate the surface form of an argument? I was never really sure. But the quality of my annotations depended on actually understanding what was being said here, so I spent a lot of time re-reading these kinds of sentences over and over, trying to grasp some kind of meaning from them. It felt like I was having a stroke, or like I was being gaslit by the text. But then, I’d read something like

BART police already have a “marijuana alley” where potential customers could find sprayers and pagers ready to use and find it where they’re supposed to.

and breathe a sigh of relief. This document is nonsense—it was generated by GPT-2, a disembodied probabilistic model that does not vote and cannot smoke weed. So, it was safe to say that there was never any meaning here for me to extract in the first place.

The following summer (2021), I wrote an essay about this experience for a class. It includes some more cool examples of AI-generated nonsense that has the shape of a sensible argument. In that essay, I argued that the text generated by LLMs is odd in that it is not language yet—not until a human is able to read and extract meaning from that text. In this way, LLM-generated text is strange and beautiful.

Ants Fornicating Meaninglessly In the Sand

  • ants fornicating meaninglessly in the sand: https://arxiv.org/pdf/2308.05576

  • doesn’t matter if the meaning is there in the first place, meaning comes from us getting it? [I WOULD ARGUE THAT THIS IS WHAT READING IS!!!! “MY READING” OF A TEXT. art, philosophy, derrida, etc. just degree of difficulty?] metaphors we live by.

  • badly written human text: derrida, bad paper, bad review replies. show the actual paper!! compare to actual legit paper!!!!!

ground back into the iclr paper

if before it was writing that is not language… now it is language that is without meaning? show example of the abstract. this is the same thing as before, just kicked up a level. now the model can imitate the form of a whole paper pretty well. but there is still a vacancy in what it is actually trying to say. . just on the entire level of the paper.

  • fake paper examples that david mentioned. yeah like who cares if the author thought it was fake, if they did actually say something interesting
  • can we only do this if we have faith that the writer is a human?