Prompting fairness: Learning prompts for debiasing large language models

Center for AI Measurement

Authors: Lemnaru Camelia, Rad Cristian Andrei

Large language models are prone to internalize social biases due to the characteristics of the data used for their self-supervised training scheme. Considering their recent emergence and wide availability to the general public, it is mandatory to identify and alleviate these biases to avoid perpetuating stereotypes towards underrepresented groups. We present a novel prompt-tuning method for reducing biases in encoder models such as BERT or RoBERTa. Unlike other methods, we only train a small set of additional reusable token embeddings that can be concatenated to any input sequence to reduce bias in the outputs. We particularize this method to gender bias by providing a set of templates used for training the prompts. Evaluations on two benchmarks show that our method is on par with the state of the art while having a limited impact on language modeling ability

A Comparative Survey of Social Bias in Text and Image Generation: Gaps, Directions and Compliance with the EU AI Act

2025Center for AI Measurement

Authors: Rad Cristian Andrei, Lemnaru Camelia

Generative artificial intelligence models, including large language models and image generation models, are increasingly deployed in socially impactful domains. However, these models often exhibit social biases that can amplify stereotypes and produce harmful, discriminatory outputs. In this paper, we present a modality-comparative survey of social bias in text and image generation, structured around four components: benchmarks, bias identification, measurement, and mitigation. We systematically analyze methodological parallels and divergences across the two modalities, highlighting emerging research trends and identifying gaps. Finally, we map current image generation research efforts to the EU AI Act’s technical requirements, offering insights into how the community can advance towards more fair, safe, and trustworthy systems.

Fine-Grained Complexity of Ontology Mediated Queries

2025Software and Hardware for AI

Authors: Feier Cristina

Prompts and Prayers: the Rise of GPTheology

2025Center for AI Measurement

Authors: Groza Adrian Petru

Increasingly artificial intelligence (AI) has been cast in “god-like” roles (to name a few: film industry – Matrix, The Creator, Mission Impossible, Foundation, Dune etc.; literature – Children of Time, Permutation City, Neuromancer, I Have no Mouth and I Must Scream, Alphaville etc.). This trend has accelerated with the advent of sophisticated Large Language Models such as ChatGPT. For this phenomenon, where AI is perceived as divine, we use the term GPTheology, where ChatGPT and other AI models are treated as potential oracles of a semi-divine nature. This paper explores the emergence of GPTheology as a form of techno-religion, examining how narratives around AI echo traditional religious constructs. We draw on community narratives from online forums – Reddit – and recent projects – AI-powered Mazu Statue in Malaysia (Lu, 2025); “ShamAIn” Project in Korea (He-rim, 2025); AI Jesus in a Swiss Church (Kennedy, 2024). These examples show striking similarities to technological notions of the Singularity and the development of Artificial General Intelligence (AGI). Additionally, we analyse how daily interactions with AI are acquiring ritualistic associations and how AI-centric ideologies clash with or are integrated into established religions. This study uses a dataset of Reddit posts discussing AI to identify recurring themes of salvation, prophecy, and demonization surrounding AI. Our findings suggest that new belief systems are developing around AI, and this carries both philosophical and sociotechnical implications. Our paper critically analyses the benefits and dangers, as well as the social, political and ethical challenges of this development. This transdisciplinary inquiry highlights how AI and religion are increasingly intertwined, prompting necessary questions about humanity’s relationship with its creations and the future of belief.

Deep Clustering for Blood Cell Classification and Quantification

2024ArticleRobotics for healthcare

Authors: Groza Adrian Petru

Accurate classification of blood cells plays a key role in improving automated blood analysis for both medical and veterinary applications. This work presents a two-stage deep clustering method for classifying blood cells from high-dimensional signal data. In the first stage, red blood cells (RBCs) and platelets (PLTs) are separated using a combination of an improved autoencoder and the IDEC algorithm. The second stage further classifies RBC subtypes, pure RBCs, reticulocytes, and clumped RBCs, through a variational deep embedding (VaDE) approach. Due to the lack of detailed cell-level labels, soft classification probabilities are generated from sample-level data to approximate the true distributions. The aim is to contribute to the development of low-cost, automated blood analysis systems suitable for veterinary and biomedical use. Initial results indicate this method shows promise in effectively distinguishing different blood cell populations, even with limited supervision.

Emerging policies for the regulation of AI

2021Center for AI Measurement

Authors: Groza Adrian Petru, Lemnaru Camelia

The report analyzes emerging policies for the regulation of artificial intelligence (AI) and their interaction with existing regulations in related fields. The report has two parts. The first part examines the European Commission’s regulatory framework for artificial intelligence, its integration into emerging policies for the development of ethical AI, and its interaction with regulations in other sectors. The second part presents a possible institutional architecture for the regulation of AI in Romania. This vision involves a competent national authority for monitoring and regulating ethical AI, as well as a decentralized model in which conformity assessment bodies (e.g. audit centers, whether private or public) have responsibilities for monitoring, verification, and certification in different subdomains and technologies of artificial intelligence.