Newsroom

Latest news from the alignAI doctoral network:

Doctoral Candidates - Simay Toplu

Q&A with DC Simay Toplu

What inspired you to join the alignAI project? I’ve always been drawn to the intersection of technology and mental health. I have a background in

Read More »

More Than Just Math – How Fairness is Being Approached in AI

In our last blog post, we unpacked what fairness in AI means, why it matters, and why technical fixes alone can not be enough to solve the deeper social and structural challenges at play. Still, just because fairness isn’t the whole answer doesn’t mean it isn’t a necessary part of the solution. Improving how AI systems treat individuals and groups, especially in high-stakes areas, is a critical step toward reducing harm and moving toward more equitable outcomes. That’s why in this second post, we shift our focus to the practical: How is fairness currently being pursued in AI systems today?

Read More »

“Fair Enough?” – Who Wins, Who Loses and Why AI Needs to do Better Than Just Working for Most

When people think about AI, they often imagine objectivity. They imagine algorithms that soberly follow data and numbers, unaffected by personal opinions, emotions, or prejudice. But here’s the problem: AI systems don’t fall out of the sky. Humans develop them, they’re trained on human-generated data, shaped by human choices, and deployed in human contexts – all of which are far from neutral.

Read More »

Opening the Black Box – How AI Explainability Is Being Approached

At the technical level, researchers have developed a range of methods aimed at interpreting how models function and make decisions. “Feature importance” techniques like LIME and SHAP help to identify the importance of specific input variables that particularly influence the output (Holzinger et al., 2022). This could mean for example that a person who was denied a loan could specifically inquire about what input features like credit score, age, employment, etc., influenced the decision the greatest. “Example-based methods” explain decisions by comparing them to similar training examples (McDermid et al., 2021). Yet other approaches aim to simplify complexity by, for example, approximating a black-box system with a more interpretable one, such as a decision tree or a linear model (Guidotti et al., 2018). Other researchers advocate for prioritising the use of inherently interpretable models like rule-based systems or decision trees over more complex models (Ali et al., 2023; McGrath & Jonker, 2024). Furthermore, there are a range of visualization techniques that can be employed to highlight which parts of an image or sentence a model was “focused on” during prediction (Holzinger et al., 2022; Miller, 2018).

Read More »

Can You Trust What You Don’t Understand? Why AI Needs to Explain Itself!

Most of the time, we don’t question the systems around us until they fail. When planes crash, treatments go wrong, or loans are denied, we ask, “What happened, why, and who’s responsible?”

As the development of AI is rapidly progressing and systems take on more and more power in deciding what we see, what we get, and what we do, a true understanding of how they get to those decisions is crucial in order for us not to lose control. Today’s systems provide outputs with a suspicious amount of confidence. However, when asked why, they often can’t or won’t tell us, revealing a complex, data-driven, and incomprehensible “black box” even for its creators (Maclure, 2021; Guidotti et al., 2018; Kosinski, 2024).

Read More »

Making LLM Alignment Work – The Need for Collaborative Research

Ensuring that LLMs align with human values is not an easy task. Alignment is particularly challenging because human values are not static, universal, or easily quantifiable and codifiable. What is considered ethical, fair, or appropriate varies significantly across cultures, political ideologies, and social contexts, making it difficult to establish a one-size-fits-all alignment approach (Liu et al., 2023; Shen et al., 2023). An output considered neutral or factual in one country might be seen as biased or controversial in another, specifically when thinking about political values like democracy. Similarly, ethical priorities and dilemmas like whether AI should prioritise free speech or harm prevention are often in direct conflict with one another.

Read More »

Contact Us

FIll out the form below and we will contact you as soon as possible