alignAI

Aligning LLM Technologies with Societal Values

About alignAI

About the project 

The alignAI Doctoral Network will train 17 doctoral candidates (DCs) to work in the international and highly interdisciplinary field of LLM research and development. The core of the project focuses on the alignment of LLMs with human values, identifying relevant values and methods for alignment implementation. Two principles provide a foundation for the approach. First, explainability is a key enabler for all aspects of trustworthiness, accelerating development, promoting usability, and facilitating human oversight and auditing of LLMs. Second, fairness is a key aspect of trustworthiness, facilitating access to AI applications and ensuring equal impact of AI-driven decision-making. The practical relevance of the project is ensured by three use cases in education, positive mental health, and news consumption. This approach allows us to develop specific guidelines and test prototypes and tools to promote value alignment. We follow a unique methodological approach, with DCs from social sciences and humanities “twinned” with DCs from technical disciplines for each use case (9 DCs in total), while the other 8 DCs carry out horizontal research across the use cases.

About Large Language Models

Large Language Models (LLMs) are trained on broad data, using self-supervision at scale, to complete a wide range of tasks. Wider use of LLMs has risen in recent months due to applications such as ChatGPT. Although LLMs bring many opportunities to improve our everyday lives, the impacts on humans and society have not yet been prioritized or fully understood. Given the rapid development of these tools, the risk of negative implications is significant if LLMs are not developed and deployed in a way that is aligned with human values and responds to individual needs and preferences. To mitigate any negative consequences, academia, in close collaboration with industry, needs to train the next generation of researchers to understand the complexities of the socio-technical implications surrounding the use of LLMs.

Chat AI screen (source: Canva)

Participating Organisations

Project Map

The alignAI project is built around a highly interdisciplinary training program
and research methodology designed to achieve the DN’s five research objectives:

  • O1. Establish a unique doctoral training programme (i) equipping DCs with the capacity to work in interdisciplinary environments, (ii) providing high quality scientific training, (iii) equipping DCs with communication capacities and (iv) Disseminating knowledge beyond the beneficiary institutions
  • O2. Identify the human values and user requirements/preferences that LLMs should align with
  • O3. Explore implementable ways for applying the principles of explainability (XAI) and fairness in
    the specific context of LLM use to enable alignment with values identified in RIO1
  • O4. Design and build value aligned LLM prototype tools based on outcomes from RIO1 and RIO2
  • O5. Test & validate the technical prototype tools from RIO3 and the non-technical
    tools/methods/models from RIO1 and RIO2
  • O6. Translate learnings from RIO1-RIO4 into research outputs, contextualising an “enabling
    environment” for value-aligned LLMs
    The doctoral training objective O1 is described in detail in Section 1.3.
    The five research objectives will be addressed in a context-specific way throughout the project by
    investigating them as part of three use-cases in: (i) Education, (ii) Positive Mental Health and (iii) Online
    News Consumption. Fig. 3. presents the proposed research methodology. This is followed by a detailed
    description of the research activities and their relevance for the project objectives.
alignAI Project Map
alignAI Project Map

Newsroom

Navigating Truth and Accountability in the Age of AI Information

Navigating Truth and Accountability in the Age of AI Information

Journalism, as one of the main driving forces behind information flows in modern societies, has traditionally promoted itself as the medium of truth. The credibility of news institutions and the legitimacy of journalism as a profession have long rested on their ability to produce, verify and disseminate information grounded in factual accuracy and editorial integrity. Yet, in the era of artificial intelligence, these epistemic foundations are being profoundly challenged: generative AI does not only replicate or automate journalistic processes, but also potentially transforms them. The generative potential of AI introduces a new layer of uncertainty to news production, as tools that are neither human nor conscious are now producing texts with the marks of human authorship, originality and even moral voice.

Read More »
Tying the Knots of Trust: Understanding the Evolving Sociotechnical Ecosystem of Trust in LLMs

Tying the Knots of Trust: Understanding the Evolving Sociotechnical Ecosystem of Trust in LLMs

When we interact with a chatbot, ask a digital assistant for advice or rely on LLMs to summarise a long document, we are doing something profoundly human: we are trusting. Trust is part of what makes cooperation possible between people, but increasingly, also between people and machines. In the age of artificial intelligence (AI), and particularly with the rapid rise of large language models (LLMs), trust has become a central issue.

Read More »
PI Line Clemmensen

Q&A with PI Line Clemmensen

In this video interview, we speak with Professor Line Clemmensen, Professor of Machine Learning at the Technical University of Denmark. She shares what her team contributes to the project, her approach to supervising PhD candidates, and how she helps them grow in both technical expertise and ethical responsibility.

Read More »
PI Sneha Das

Q&A with PI Sneha Das

In this video interview, we speak with Assistant Professor Sneha Das, a researcher at DTU whose work focuses on trustworthy AI and responsible data driven systems. She discusses the perspective her institute contributes to the project, and how the program can guide policy and public understanding related to AI and alignment.

Read More »
Cen Lu

Exploring AI Advancement at NeurIPS 2025

The Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS 2025), was held from December 2-7 in San Diego. Our alignAI doctoral candidate Candidate Cen Lu attended and presented his poster “Chain-of-Model Learning for Language Model” at the poster session.

Read More »

AI-nxiety and AI-gency: Young Adults Navigating Generative AI

As part of the TUM Institute for Ethics in Artificial Intelligence (IEAI) Speaker Series, a December 2025 session focused on young adults and generative AI. The talk, titled “AI-nxiety & AI-gency: Young Adults Navigating Generative AI” was delivered by Dr. Jaimee Stuart, Senior Researcher and Team Lead at United Nations University Macau.

Read More »

Contact Us

FIll out the form below and we will contact you as soon as possible