Over the past years, foundational models, including large-language models (LLMs) and multi-modal systems, have significantly advanced the possibilities regarding the understanding, analysis, and generation of human language [1]. These models, based on artificial neural network computations trained on a large scale of documents, open the possibility of performing advanced tasks related to language, audio, video, and image processing.
However, these models are relatively novel and there are several crucial open issues that need to be addressed. These issues are related both to the development and the impact of these models on society, especially on media and democracy.
Reliability and transparency
Technical issues regard the reliability and the transparency of LLMs. While LLMs are very good at providing believable and convincing responses to users, assurances on the quality of the information provided should be given. For example, LLMs often suffer from the so-called hallucination issue, that is, they generate believable but false statements. Also, they are often trained on large but biased datasets, and this might exacerbate the presence of stereotypes and undesired associations in the responses given. While preliminary solutions to these problems have been proposed (e.g. [4,5]), mapping these open issues and discussing potential solutions is important to enhance a conscious use of these models and an understanding of the pitfalls and limitations of the solutions proposed so far to known issues. For example, debiasing may result in bleached and uninteresting responses.
Use and misuse
Societal issues concern the possible use and misuse of LLMs. Media, as communication tools for storing and delivering content, can find in LLMs a powerful tool to automate and accelerate the information generation process. However, such automation should be performed by understanding the limits, risks, and capabilities of LLMs first, in order to guarantee the quality of the output generated. For example, while it could be hazardous to create content exclusively based on LLMs, other tasks like content curation or engagement enhancement could significantly benefit from LLMs.
Disinformation
Besides their effect through the media, democracy can be affected directly by LLMs, because of the influence that these models can have on the generation and propagation of disinformation [2], e.g. by deliberately prompting to generate false or unethical information. However, LLMs can also be part of the strategy to contrast misinformation spreading, by allowing a deeper and more advanced assessment of the information spread online.
In light of these considerations, the legal domain should provide tools to govern the adoption of these models and their effect on society. The AI ACT represents a beginning in this direction, although further developments are needed [3].
In this workshop, we intend to showcase, discuss, and advance recent developments regarding LLMs, both from the technical and the societal point of view.