alignAI

Aligning LLM Technologies with Societal Values

About alignAI

About the project 

The alignAI Doctoral Network will train 17 doctoral candidates (DCs) to work in the international and highly interdisciplinary field of LLM research and development. The core of the project focuses on the alignment of LLMs with human values, identifying relevant values and methods for alignment implementation. Two principles provide a foundation for the approach. First, explainability is a key enabler for all aspects of trustworthiness, accelerating development, promoting usability, and facilitating human oversight and auditing of LLMs. Second, fairness is a key aspect of trustworthiness, facilitating access to AI applications and ensuring equal impact of AI-driven decision-making. The practical relevance of the project is ensured by three use cases in education, positive mental health, and news consumption. This approach allows us to develop specific guidelines and test prototypes and tools to promote value alignment. We follow a unique methodological approach, with DCs from social sciences and humanities “twinned” with DCs from technical disciplines for each use case (9 DCs in total), while the other 8 DCs carry out horizontal research across the use cases.

About Large Language Models

Large Language Models (LLMs) are trained on broad data, using self-supervision at scale, to complete a wide range of tasks. Wider use of LLMs has risen in recent months due to applications such as ChatGPT. Although LLMs bring many opportunities to improve our everyday lives, the impacts on humans and society have not yet been prioritized or fully understood. Given the rapid development of these tools, the risk of negative implications is significant if LLMs are not developed and deployed in a way that is aligned with human values and responds to individual needs and preferences. To mitigate any negative consequences, academia, in close collaboration with industry, needs to train the next generation of researchers to understand the complexities of the socio-technical implications surrounding the use of LLMs.

Chat AI screen (source: Canva)

Participating Organisations

Project Map

The alignAI project is built around a highly interdisciplinary training program
and research methodology designed to achieve the DN’s five research objectives:

  • O1. Establish a unique doctoral training programme (i) equipping DCs with the capacity to work in interdisciplinary environments, (ii) providing high quality scientific training, (iii) equipping DCs with communication capacities and (iv) Disseminating knowledge beyond the beneficiary institutions
  • O2. Identify the human values and user requirements/preferences that LLMs should align with
  • O3. Explore implementable ways for applying the principles of explainability (XAI) and fairness in
    the specific context of LLM use to enable alignment with values identified in RIO1
  • O4. Design and build value aligned LLM prototype tools based on outcomes from RIO1 and RIO2
  • O5. Test & validate the technical prototype tools from RIO3 and the non-technical
    tools/methods/models from RIO1 and RIO2
  • O6. Translate learnings from RIO1-RIO4 into research outputs, contextualising an “enabling
    environment” for value-aligned LLMs
    The doctoral training objective O1 is described in detail in Section 1.3.
    The five research objectives will be addressed in a context-specific way throughout the project by
    investigating them as part of three use-cases in: (i) Education, (ii) Positive Mental Health and (iii) Online
    News Consumption. Fig. 3. presents the proposed research methodology. This is followed by a detailed
    description of the research activities and their relevance for the project objectives.
alignAI Project Map
alignAI Project Map

Newsroom

Beyond the Hype: What Actually Makes AI Design Different

Just a few years ago, conversation flow design was at the heart of chatbot research (Cho et al., 2025). Designers developed detailed guidelines to structure dialogues, crafted messaging frameworks for seamless interactions, and carefully designed output messages to align with chatbot personas. Then as transformer-based large language models (LLMs) arrived, the rigid and predefined conversation structures that worked for rule-based systems couldn’t accommodate LLMs’ dynamic, context-aware response. Research priorities shifted from designing fixed dialogue trees to exploring prompt engineering and interaction patterns (Cho et al., 2025). The expertise built over years around conversation flow design was fundamental but it needed rapid reframing; designers had to rethink how to guide conversations without prescribing every turn, how to maintain coherence without rigid structures, and how to evaluate interactions that varied with each user (Subramonyam et al., 2024). This narrative about change isn’t just applicable to chatbots. It’s applicable to Generative AI’s (GenAI) unique temporal challenge and it raises a critical question for anyone designing with or for AI – are our design methods keeping up, or do we need new ones?

Read More »
PI Daniel Gatica-Perez

Q&A with PI Daniel Gatica-Perez

In this video interview, we speak with Professor Daniel Gatica-Perez, Head of Social Computing Group at Idiap and Professor at École Polytechnique Fédérale de Lausanne. He emphasises the importance of a human-centred approach and shares how he sees the alignAI project can help society.

Read More »

The Moral Panic Around AI Mental Health

Trigger Warning/Disclaimer: This blog post mentions suicide.

Governments, startup founders, academics, mental health professionals and others wrestle over who gets to define the future of AI mental health care.
Amidst a lack of regulatory oversight regarding AI-based mental health chatbots, some states in the US have taken steps to ban these systems in order to protect the public. Full bans are in place in Illinois and Nevada, and although Utah has not banned it outright, it still imposes strong restrictions and requirements around transparency, advertising, data use and human professional involvement. Bans as a political strategy and policy risk unintended consequences on a population-wide scale (Oliver et al., 2019).

Read More »

Can You Trust the Machine? alignAI Doctoral Candidates Hold Workshop at Samuel-Heinicke-Fachoberschule

On November 21st, alignAI doctoral candidates Julia Li and Simay Toplu held an interactive workshop with 32 students at Samuel-Heinicke-Fachoberschule, organized together with the Europe Direct Network. The session introduced students to the everyday presence of AI systems and encouraged them to reflect on the risks, benefits and responsible use of AI in real-life situations in the EU and beyond.

Read More »

Q&A with PI Avigdor Gal

In this video interview, we speak with Professor Avigdor Gal, Benjamin and Florence Free Chaired Professor of Data Science at Technion – Israel Institute of Technology and one of the Principal Investigators in the alignAI project. He explores his role with alignAI, how his research on data integration, uncertain data and machine learning strengthens our network and his vision for how the AI ecosystem might evolve in the future.

Read More »

Contact Us

FIll out the form below and we will contact you as soon as possible