Writing is an essential element of academic research — as researchers, we have to write papers, proposals, (long) reports, code, and, of course, lots of emails. As writing takes up a lot of our time, being able to write well, and quickly, is a very important skill for any researcher.
The ability of large language models (e.g. in the form of ChatGPT) to produce well-written language is rather fascinating, so it is no surprise that many have considered using them (or have actually used them) for academic writing as well; see for example [1-3] for an overview and a discussion of some benefits and challenges. In general, large language models have many possible applications; in the context of academic writing, the goal of such models is to assist or replace the writing process, so, in this post, I will refer to them as AI-based writing assistants (AIWAs).
As I am a researcher in the field of learning-based robots, one might expect that I would be very positive about AIWAs and would fully embrace them in my work. Well, that is not really the case. In this post, I would like to actually discuss why I am rather skeptical of AIWAs, or why I value the process of writing rather than simply the end result (the written text). I will not discuss issues related to data privacy; those fall beyond the scope of what I want to cover here. I will also not touch on the issue of hallucinations by LLMs, as there are already many good discussions on this (e.g. this one).
Before I discuss my generally skeptical view of AIWAs, I would like to point out what I think are very positive aspects of using them for academic writing:
My skepticism of using AIWAs for academic writing is based on the following arguments:
I expand on each of these points below.
At least for me, writing is an essential element that supports my thinking process, particularly in two stages of my work:
An astute reader may read the discussion of the second point and ask “Well, when we write collaboratively, it is anyway not the case that we write all the text ourselves, but we have to work with text written by our co-authors. How is working with text generated by an AIWA different from text written by a co-author?” That is indeed a valid question; there are, however, a couple of differences I see, namely when I write together with a co-author:
The unclear text ownership is a related reason why I am skeptical about AIWAs in academic writing. Here, I am not actually talking about plagiarism (which has been demonstrated, for example as evidenced by the New York Times lawsuit against OpenAI, and which I think is the biggest potential ethical issue with AI-generated text), but about a more subtle issue, namely that of creating a connection with the written text as a writer.
As I have already mentioned before, at least for me, the process of writing is where I can fully demonstrate my engagement with the work, but is also what allows me to really claim ownership of the written text. Writing can be a relatively slow process, but it is this slowness that makes it possible to really internalise the contents and the logical flow of the text. In fact, I would say that, even years after writing some of my papers, I am still able to recall the exact logical arguments that I was trying to make there, precisely because I took the time to prepare everything carefully. In this respect, writing can be compared to proving a mathematical theorem: it can be helpful to work with a ready-made proof, but true understanding and ownership of a proof only happens once we have taken the time to construct the argument ourselves. Following this analogy, using an AIWA to produce text for us is a bit like having someone else write parts of the proof; the end result may be complete and correct, but the fact remains that we haven’t fully created the proof.
The above co-author argument can be made here as well: “When we write papers with co-authors, we anyway don’t have full ownership of the text, so how is that different from working with AI-generated text?” Another valid question, but the difference is that, in this case, we don’t claim (sole) ownership of the text written by the co-authors; as they are co-authors, we now have shared ownership of the text. But does that imply that an AIWA should become a co-author? Would that solve the problem? Not quite. Ignoring cases in which pets have been added as co-authors, when we have a co-author, it means that they can take accountability for the written text; an AIWA cannot do that.
In this context, I suppose the quantity of generated text matters. If only a short text is generated, it can be relatively simple to modify it so that one “makes it” their own; the problem becomes much more serious when large quantities of text are generated by an AIWA.
In modern society, productivity is perhaps one of the most coveted characteristic of an individual: the more productive we are, the more work we can do. In principle, there is nothing wrong about that; this is also one of the reasons why there are countless productivity techniques and guides out there. As AIWAs can take away some burden of the writing process, they are obviously a tool that can increase productivity. For instance, this might mean that we are able to write more papers in a shorter time, or that we can prepare project proposals quicker than we would otherwise. Sounds perfect, right?
In my opinion, not necessarily. I think this can go somewhat against the spirit of scientific research, or at least violates my understanding thereof. Research is not necessarily about finding answers fast (although there are, of course, fields such as medicine, where this can be very desirable), but is about finding the right answers after a thorough examination of different possibilities and outcomes. In addition, research should be about stepping back and considering the consequences of the work we are doing. All of this can actually benefit from less productivity at certain times; quite often, it can be very helpful to take some time to simply “sleep on things”, namely to take some distance from the work and then get back to it with a fresh mind. AIWAs can potentially disrupt this process, as they can encourage settling in on a solution quickly; taking a step back may be seen as unnecessary when using an AIWA as a productivity tool, as one can quickly finish a writing task and immediately move on to another one. But, due to the reduced time investment, this can mean that important considerations were actually not made in the process; the temptation to move on to the next thing can be overwhelming.
I do not consider myself an exceptional researcher by any means. I simply believe I am good at what I do and have been known to be quite productive when it comes to writing; however, as alluded to in the previous sections, my writing productivity has, perhaps counterintuitively, typically come through periods of lower productivity, where I have taken some time to think about how to achieve the goals of my work or have gone through an iterative process of writing (text or code) and actively examining the consequences of what I have written. While there are undoubtedly cases where I would have benefited from a productivity boost, in most cases, slower pace has been exactly what I had needed.
In this post, I have shared my critical view of the adoption of AI-based writing assistants (AIWAs) for academic writing, namely why I currently do not consider them to be a viable option for my writing activities. Even though there are undoubtedly some positive aspects of using AIWAs, particularly when observing them as language assistants, I see their use in academic writing as generally undesirable. To support this view, I focused my discussion on three main aspects, namely I discussed writing as a tool that supports the thinking process, the challenges of using AIWAs with respect to text ownership, as well as the necessity to sometimes be less productive (in the short term) for increased long-term productivity. Despite all this, it has to be mentioned that AIWAs have undoubtedly already had an effect on academic writing; as a matter of fact, a lot of the effects will likely be visible only in the long term. I might be in the minority with my views, but I suppose that remains to be seen.