The researchers have mapped this rapidly evolving landscape in a paper and a live-updated. This shows there are many working alternative ’open source’ text generators, but also that openness comes in degrees and that many models inherit legal restrictions. They sound a note of cautious optimism. Lead researcher Andreas Liesenfeld: "It’s good to see so many open alternatives emerging. ChatGPT is so popular that it is easy to forget that we don’t know anything about the training data or other tricks being played behind the scenes. This is a liability for anyone who wants to better understand such models or build applications on them. Open alternatives enable critical and fundamental research."
More and more open
Corporations like OpenAI sometimes claim that AI must be kept under wraps because openness may bring ’existential risks’, but the researchers are not impressed. Senior researcher Mark Dingemanse: "Keeping everything closed has allowed OpenAI to hide exploitative labour practices. And talk of so-called existential risk distracts from real and current harms like confabulation, biased output and tidal waves of spam content." Openness, the researchers argue, makes it easier to hold companies responsible and accountable for the models they make, the data that goes into them (often copyrighted), and the texts that come out of them.The research shows that models vary in how open they are: many only share the language model, others also provide insight into the training data, and quite a few are extensively documented. Mark Dingemanse: "In its present form, ChatGPT is unfit for responsible use in research and teaching. It can regurgitate words but has no notion of meaning, authorship, or proper attribution. And that it’s free just means we’re providing OpenAI with free labour and access to our collective intelligence. With open models, at least we can take a look under the hood and make mindful decisions about technology."
Some additional points
New models appear every month, so the paper is mainly a call for action to track their openness and transparency in a systematic way. https://opening-up-chatgpt.ioThe researchers will present their findings at the international conference on Conversational User Interfaces in Eindhoven, NL, July 19-21.
Literature reference Liesenfeld, Andreas, Alianda Lopez, and Mark Dingemanse. 2023. "Opening up ChatGPT: Tracking Openness, Transparency, and Accountability in Instruction-Tuned Text Generators." In ACM Conference on Conversational User Interfaces (CUI ’23). July 19-21, Eindhoven. doi: 10.1145/3571884.3604316