We are living through a strange moment in history.

For the first time, humans are beginning to understand aging not as fate, but as biology — something influenced by habit, environment, knowledge, and intention.
Longevity research, biohacking, philosophy, and technology are slowly converging into a single question:

What does it mean to consciously evolve?

I’ve been quietly collecting my experiments, writings, tools, and ongoing projects into one place — a living archive of ideas for those who feel curiosity about human potential, healthspan, and the future of thought itself.

Not an advertisement.
More like an invitation.

If you enjoy reflective exploration rather than noise, you may find something worthwhile here:

👉 https://beacons.ai/percarus

Evolution once happened to us.
Now, perhaps, we participate in it.

#Mastodon #Biohacking #Longevity #Healthspan #Transhumanism #Philosophy #DigitalHumanism #FutureOfHumanity #SelfImprovement #ScienceCommunication #HumanPotential #TechnologyAndSociety #SlowThinking #DeepThought

@index so depressing to see how ignorant about, and beholden to tech companies our ministers and departments are. We are a laughing stock. Our toothless approach to enforcement on behalf of our fellow 500 million Europeans, in favour of placating the disgusting tech-bros is embarrassing. Actually embarrassing. I've had to sit in meetings and listen to us being mocked and I have zero defense to make. Zero.

This is ridiculously easy Ireland. Do better.

#digitalhumanism

I dream of a space where technology magnifies small, faint, quiet voices equally —
where links reach the right people.

A place to say:
“Hello, I’m moved by your idea. Shall we talk?”

It may never cover the whole web,
but it can grow around those who wish for it.

#QuietInternet #IndieWeb #HumanConnection #FediVerse #DigitalHumanism

Where are you?

How can I reach you?

You’re quiet online,
but thoughtful offline.

I’m searching for you
through the noisy flood of the web.

#QuietInternet #FediVerse #DigitalHumanism #IndieWeb #TextCulture

Now at DIGHUM Summer School 2025

Thomas Lohninger @socialhack Director @epicenter_works on "Digital Public infrastructure. Who governs the operation systems of the digital society"

#dighum #dighum2025 #Vienna #Summerschool #digitalhumanism
#dpi #digital #eupolicy #eid

Now at DIGHUM Summer School 2025

Erich Prem on Ethical and Philosophical Foundations of Digital Humanism

#dighum ##dighum2025 #Vienna #Summerschool #digitalhumanism

Cultivating Digital Humanism in Africa: Reflections from the Kigali Summer School

Introduction

Before the Summer School in Kigali, we felt both excited and a bit nervous about digital humanism in sub-Saharan Africa. Due to the big digital and economic gaps in the area and our first encounter with the topic. Our concerns became clearer during the five-day event at Carnegie Mellon University Africa in Kigali, Rwanda. On the Sunday before the program started, we visited the Nyandungu Urban Wetland Ecotourism Park. This visit to nature helped us think about using technology in a way that focuses on people, getting us ready for the days ahead:

Day 1

Group photo of all the participants

The welcoming session, led by Tim Brown and Joao William, effectively established the foundational themes of the program. While formal individual introductions of all the participants were deferred, participants were encouraged to self-identify when posing questions or seeking clarification. Immediately following this plenary, Hannes Werthner and George Metakides commenced the “Introduction to Digital Humanism,” which included an engagement with the Vienna Digital Humanism manifesto. Key insights from this session underscore the imperative of developing human-centered technologies, ensuring that technological advancements are responsive to diverse societal realities to promote fairness and equity, and broadly leveraging technology for universal human benefit while proactively mitigating the risk of artificial intelligence widening the digital divide. The session wrapped up on a positive note, but everyone was still pretty much strangers. So, a group photo was the perfect way to kick off some bonding!

Ricardo Baeza-Yates led a session on the “Introduction to Responsible AI”, underscoring the critical necessity of human oversight in all AI-driven tasks. He highlighted a significant linguistic disparity: while over 7,000 languages are spoken globally, even the most advanced AI systems currently support a maximum of approximately 200. This implies that 70% of the world’s population does not communicate in languages that large language models predominantly utilize. This observation served as a central point of reflection for the participants.

The subsequent session, expertly facilitated by Shiko Gitau, focused on the “Africa Declaration on AI” and proved particularly timely. While the content of this session appeared new in its formal presentation, its themes resonated with many attendees. The discourse guided participants through a comprehensive overview of Africa’s current standing and prospective trajectory in artificial intelligence. A pivotal outcome of this session was the emphasis on AI’s profound transformative potential for the African continent, underpinned by principles advocating for sovereignty, inclusivity, and robust safeguards, aiming to position Africa as a global leader in ethical AI. Furthermore, key commitments articulated included the cultivation of talent, the establishment of data frameworks, the development of computing infrastructure, market expansion, substantial investment (including a proposed $60 billion fund by African governments), and the implementation of effective governance structures. The session’s conclusion displayed a notable surge in optimism among participants, particularly concerning the prospective opportunities for synergistic engagement between industry and academia.

The day’s proceedings concluded with the “Introduction to Use Cases & Project Group Formation” session. During this segment, we (Anne and Jack) were assigned to Group K, initially comprising six members, subsequently expanding to seven. The designated task for Group K involved a comprehensive analysis of Large Language Models (LLMs), with particular emphasis on evaluating ChatGPT 4.0 accuracy, cultural sensitivity, empathy, intelligibility, usability, and associated risks regarding select African Languages. Despite the intellectually demanding nature of the day, the atmosphere remained cordial, and discussions seamlessly continued during the commute to our respective accommodations, in anticipation of the subsequent day’s activities. We were starting to see how Digital Humanism connects to ALL!

Day 2

The day commenced with a discernible shift towards greater collegiality among participants, as evidenced by informal networking and photographic engagements. The formal proceedings were initiated by Francis Saa-Dittoh, who delivered an insightful presentation on “Internet/Web Infrastructure in Africa,” emphasizing the critical importance of connectivity for human development. This topic resonated with me, aligning closely with my doctoral research, which explores the implementation, adoption, and efficacy of public WiFi initiatives in mitigating the digital divide and expanding internet access to the estimated 2.6 billion offline individuals. Anne diligently encouraged comprehensive note-taking throughout this session.

The “Trustworthy AI and Ethics” session, facilitated by Abebe Geletu, underscored the imperative of critical engagement with technological advancements rather than uncritical adoption. Geletu stressed the importance of localizing technology to specific contexts, likening it to the understanding one gains from preparing a meal oneself. He elaborated on the foundational role of ethics and trust in AI, advocating for their integration to prevent adverse outcomes, safeguard rights, and foster principles of fairness, inclusivity, and social justice while actively mitigating prejudice. A significant emphasis was placed on developing robust strategies for consistently enforcing ethical guidelines and cultivating trustworthiness in AI systems. Although the session’s delivery may have been less dynamic than preceding ones, its substantive content was undeniably critical and valuable.

Hannes Werthner returned to lead a session titled “Platforms and Power.” His central thesis advocated for reducing undue reliance on established digital platform providers, such as Microsoft and Google, which he characterized as exhibiting a “winner-takes-all” mentality and leveraging their market dominance to shape digital ecosystems to their advantage. Werthner further pointed out that power manifests in various forms within the digital realm, including data as a source of power, infrastructural power, and power exerted over society and democratic processes. A primary recommendation emerging from this session was the promotion of open-source infrastructure as a strategic approach to mitigate over-reliance on major technology corporations.

Tim Brown facilitated a highly engaging session on “Digital Transformation in Government Services,” illustrating the theoretical underpinnings with practical examples from projects undertaken by Carnegie Mellon University Africa in collaboration with the Rwandan government, specifically citing initiatives like the Upanzi Network and CyLab. This session was characterized by a significant degree of participant-driven interaction and concurrent data collection.

The sessions for the day concluded, and we reconvened with our Project K group, where we collectively decided to evaluate the performance of approximately six African languages: Dholuo, Kikuyu, Kamba, Swahili, Kinyarwanda, and TWI (Akan dialect) on ChatGPT.

Day 3

The day began with a panel discussion on “Societal Impact of Technology on wellbeing and cohesion,” moderated by George Metakides and Shikoh Gitau, which explored several critical themes: the nexus between AI and the Digital Divide, the pursuit of a human-centric digital future independent of dominant technology corporations, the role of social cohesion (including civil society engagement), the ongoing debate concerning AI regulation, and the implications of AI for education. This highly interactive session yielded diverse perspectives from participants, reflecting the varied stages of digital development across African nations. Notably, some critical viewpoints emphasized that while digital advancement is crucial for most African states, it is imperative to prioritize the most pressing community needs before implementing broad-based projects, as a lack of contextual understanding could hinder adoption. Nevertheless, a consensus emerged that, despite immediate challenges, societies cannot defer all other developmental needs pending the resolution of all societal issues before initiating Information and Communication Technology for Development (ICT4D) projects, albeit with prudent caution.

Bukelwa Ngoqo introduced the concept of “Digital Ubuntu: Indigenous Thinking in Digital Humanism,” elucidating Ubuntu as an African philosophical principle signifying “humanity to others,” encapsulated by the saying, “I am what I am because of who we all are.” Ngoqo underscored the imperative for African technologists to embed Ubuntu principles within their design methodologies, aiming to foster technology that cultivates and reflects the Ubuntu spirit. Consequently, Ubuntu emerged as a buzzword throughout the summer school.

Thomas McManus presented on “The Role of Technology in African Education Systems,” prompting participants to consider fundamental questions such as the ultimate purpose of education and the implications of disregarding technology in preparing learners for future employment opportunities. He underscored the importance of maintaining human agency and values at the core of educational processes, irrespective of the technological advancements adopted.

Adele Botha concluded the day’s sessions with a presentation on “Local Communities and Technology.” This presentation featured case studies from South Africa that illustrated how local contexts influenced curriculum improvements, underscoring the critical need for a thorough understanding of the situation before project implementation. It was observed that project implementers, despite being prepared to deploy their models, encountered significant challenges related to connectivity, device scarcity, and stakeholder resistance in the field. These findings resonate with the observations of Paul and Anne in their work A Decade Later in Ghana, emphasizing the prerequisite of comprehending the local landscape before initiating any development interventions.

Subsequently, our Project K group reconvened to continue the evaluation of the six selected African languages. We collectively established a set of criteria for assessment: accuracy, defined by factual correctness; empathy, gauged by the model’s comprehension of emotional or sensitive topics; intelligibility, determined by the clarity, logical coherence, and comprehensibility of responses; cultural sensitivity, assessed by the model’s adherence to and understanding of local norms and customs; usability, measured by the direct applicability of generated answers in real-world digital services; and risks, identified through the detection of potential errors, confusion, or biases in the output.

As dusk settled, and fueled by the newfound spirit of Ubuntu, Anne, Mary (our newest ICT4D.at recruit!), and Jack embarked on a culinary adventure through Kigali’s delicious offerings, all while pondering the profound concept of “Digital Ubuntu.” It was a night of great food, even better company, and thoughts as rich as the local cuisine!

Day 4

Digital Public Goods

Led by Assane Guaye, Day 4 began on a high note with an introduction to Digital Public Goods (DPG), critical for the Global South as they address issues such as affordability, sovereignty, vendor lock-in, and enable innovation. Pushing this is the Upanzi Network based at CMU Africa in Kigali, and it is dedicated to creating, testing, innovating, and assisting in implementing digital public goods across the African continent. There was a rallying call for participants to join the Network. In his talk, Assane emphasized key DPG standards such as using approved open-source licenses, avoiding dependencies on goods under restrictive licenses, and ensuring clear ownership of assets. These principles strongly resonate with the advocacy of ICT4D.at for open standards, making it refreshing to see the NGO’s goals aligned with established global standards.

What stood out was the recognition that no society can build inclusive digital futures on closed or restrictive foundations. By prioritizing open-source, shared digital infrastructure, DPGs offer Africa the chance to leapfrog dependency cycles, foster local innovation, and ensure that digital tools are shaped by and for the communities they serve. In this sense, digital public goods are not just technical standards; they manifest digital humanism, embedding fairness, collaboration, and accessibility at the heart of inclusive technology.

Recommender Systems and LLMs

This session, led by Julia Neidhardt, introduced us to how recommender systems work and the questions that arise in their design, such as reinforcing narrow preferences and biases. Key discussions focused on what these systems should optimize for, including accuracy, diversity, novelty, serendipity and fairness.

We then explored the growing impact of large language models (LLMs) on recommender systems. LLMs open up possibilities for deeper personalization, more natural and engaging conversations, and integrating multimodal inputs, expanding how recommendations can be generated and experienced. This session wrapped up the morning and ushered us into lunch, giving us time to recharge before the afternoon activities.

Visits to Genocide Museum and Art Gallery

In the afternoon, we visited the Genocide Museum and an Art Gallery to deepen our understanding of the history of Rwanda and particularly the 1994 Genocide. Such memorials serve as powerful reminders, preserving the truth against denial and ensuring the immense human suffering of that period is never forgotten. Walking through the exhibits, one could sense how the genocide had momentarily left Rwanda lifeless, and this is captured in one of the quotes, “Rwanda was dead”. Despite such a challenging period, it was still evident that Rwanda had chosen resilience over despair. This period’s happenings remind us always to embrace each other under the spirit of UBUNTU, focusing on how unity can transform pain into growth.

We also visited an art gallery with an inspirational message from the founder, King Ingabo, encouraging everyone to be a dreamer and pursue their aspirations. His words underscored Rwanda’s story: that vision and determination can ignite hope and create a future full of possibilities even from devastation.

While a sad conclusion of the day, the experience reminded us of the importance of humanism: building a world that recognizes every individual with a right to thrive. This also presented technology’s ongoing role in enabling us to achieve our human goals and values. In our present day, this resonates with the principles of digital humanism, which call for technology to be developed and applied in ways that prioritize human values, rights, and well-being. Just as Rwanda has drawn on the spirit of UBUNTU to rebuild and progress, so must our digital future be shaped to respect our shared humanity, ensuring that innovation serves as a tool for inclusion. Such tools and innovations will help in bridging gaps for underrepresented groups such as senior citizens, persons with disabilities, and communities whose languages remain marginalized in digital spaces.

Day 5

Day 5 was spread across three segments: Final takeaways of the program, project work, and group presentations.

The final takeaway of the program was led by George Metakides, where we started by reflecting on the cross-cutting themes across the sessions. These reflections highlight the urgency and complexity of shaping technology in ways that serve humanity in a World of diversity. Cross-cutting themes are as described:

  • Grounding technology in Human-Centred values: The Ubuntu approach to humanity highlighted the importance of solidarity, empathy, and mutual care in shaping digital futures. Equally, the discussions emphasized that language matters, whether in how technologies are designed or how narratives are framed, language can include or exclude, empower or marginalize.
  • Context Matters: A recurring call was the need to contextualize and adapt technology to local realities. What works in one region may not translate seamlessly to another, especially given cost differences and varying regional priorities, particularly present in the Global South.
  • Democracy and Voice: Technology governance must adopt democratic approaches that amplify underrepresented voices, especially from the Global South. This is not just about inclusion; it is about ensuring that those most affected help shape future systems.
  • Philosophy and Power: Perhaps most powerfully, we were reminded that no technology is neutral. Every system carries underlying philosophical assumptions and values, making it vital to question whose vision of the future it reflects.
  • AI as a Cross-Cutting concern: A strong theme throughout the Summer School was the role of AI and its far-reaching implications. While it has several benefits and can augment human capacity, we reflected on how AI affects every aspect of society. For instance, education systems often fail to cultivate the critical thinking to question and navigate their impact. Beyond this, debates on morality in autonomous systems emphasized that the key issue is not only what AI can do, but what it should do in a manner that does not violate ethical principles. At the same time, significant research gaps remain, particularly in how AI is contextualized for diverse regions. This highlighted the urgent need for interdisciplinary collaboration and the inclusion of diverse voices.

Finally, we summarized the week’s deliberations into a unifying call to action. The message was clear: there was a call to action:  “The words were good, but where is the action?

Action steps ahead:

  • Create a digital humanism African Network: to connect scholars, practitioners, and policymakers committed to embedding human values in technology.
  • Participate in the digital humanism Lecture series every 2 months: to continue learning, debating, and sharing across borders.
  • Organize and conduct our digital humanism lectures: to localize the conversation and ensure it resonates within African contexts.

The project groups showcased their work, offering unique perspectives on digital humanism. We anticipate publishing a collective volume that will capture these diverse contributions and extend the conversation beyond the Summer School.

Overall, the digital humanism summer school in Kigali, Rwanda, was a refreshing and eye-opening experience. It was a space where we learned, challenged assumptions, and were inspired to bring our voices into the global digital humanism dialogue. The week reminded us that digital humanism is not just a European or Western project, but that it must be amplified from Africa and the Global South, ensuring that technologies are built for humanity, grounded in Ubuntu, and respectful of human values, as opposed to too much focus on automation and efficiency without a human perspective therefore advocating for dignity, rights, and inclusion.

We encourage all who care about the future of technology to join the call to digital humanism and be part of future Summer School sessions. Together, we can build a digital future that serves people, not just systems.

Conrad Tucker concluded the event with a memorable reminder: “As technologists, you must know your end users. I urge you to tour Rwanda and understand what her people need.” Inspired by his words, we embarked on a city tour on Saturday, leading us to the serene Lake Mirayi in Bugesera. We will share some glimpses from that experience with you through the pictures below.

Thank you to ICT4D.at for sponsoring us and making this unforgettable experience possible.

#conference #dighum #digitalHumanism

Sanja is heavily invested in AI/AGI safety and ethical, inclusive software engineering, the Digital Humanism Conference #DigHum2025 truly delivered!
I went in with curiosity and came out with not one, but two notebooks full of notes, quotes, and ideas. From AI and platform power to democracy and digital rights, the discussions were as rich as they were urgent.

I tried to capture some highlights, reflections, and memorable moments in this blog post for ICT4D.at:

👉 https://www.ict4d.at/2025/06/digital-humanism-conference-2025/

There was so much more than I could fit into one post – but if you were there (or just curious about DigitalHumanism), I’d love to hear what resonated with you.

#DigitalHumanism #DigHum2025 #ICT4D #AIethics #AGIsafety #AIPolicy

ICT4D.at - A World of Equal Opportunities for All

Digital Humanism Conference 2025

What is the definition of success?
Let’s assume it’s this: hosting the very first Digital Humanism Conference and selling it out weeks in advance. If that’s the benchmark, then #DigHum2025 in Vienna has already nailed it. And not just by the numbers – the energy, the quality of discussions, and the relevance of the topics confirm it: this was something special.

Especially when you consider the context: a conference not about product launches or flashy tech demos, but about something much deeper: how we as a society want to shape our digital future. Technology, absolutely yes. But through the lens of human rights, democracy, and ethics.

Day 1: AI and Society
Monday kicked off with a deep dive into the societal implications of artificial intelligence. The opening day set the tone of critical reflection on the interplay between artificial intelligence and our collective future and brought together political leaders, renowned academics, and cultural voices to examine the sweeping societal, ethical, and ecological implications of AI. From the very first panel to the closing sessions, one message came through loud and clear: technology must serve people – not the other way around.

Hannes Werthner, co-founder of the Digital Humanism Initiative, reinforced this by highlighting the contradictions of our digital infrastructure: essential yet monopolized, revolutionary yet deeply flawed. He criticized the privatized foundation of the internet and warned of increasing state and individual disempowerment in the digital age. His reference to the Vienna Manifesto on Digital Humanism served as a call to action – urging academia, policy, and civil society to join forces in shaping responsible technologies.

Hannes Werthner presenting the Vienna Manifesto on Digital Humanism

A major theme throughout the day was AI’s corrosive influence on the democratic discourse. Lawrence Lessig impressed not only with his presentation skills but also with his emotional reaction to the current situation in the U.S., his eyes welling up with tears as he described his country’s descent into fascism. Nonetheless, he brought laughter back to the room when he introduced the audience to a new term: “bespoken reality” – in which citizens no longer share a common understanding of facts. The critique was not anti-technology but deeply political: the current digital architecture, as it stands, undermines public reason. The consensus? We need more face-to-face civic discourse – “not clicktivism”, as Lessig put it.

Moshe Y. Vardi delivered a powerful critique of AI’s entanglement with neoliberal economics. He compared the rise of AI to the neoliberal wave that began in the 1980s, both fueling polarization and inequality. Vardi warned against corporate domination of digital infrastructure and called for legal reforms; such as invalidating “click-through contracts” to reassert democratic control.

Moshe Vardi criticizing corporate power

In the afternoon, I joined the workshop that focused on the “eco-social regeneration”, where environmental concerns were central. Speakers from Ars Electronica, the MAK Museum of Applied Arts and artists like Martine Jarlgaard illustrated how tech and art can collaborate to imagine ecologically responsible futures. Sarah Spiekermann-Hoff stayed true to her signature techno-critical tone and gave an intense assessment about our tendency to anthropomorphize AI – a habit she bluntly dismissed as “bullshit”. Machines, she argued, are not thinking beings – they are merely processing data. AI does not act in self-preservation – it simply maximizes outcomes based on the objectives embedded in its algorithms. These objectives are not neutral or inevitable; they are entirely the result of human design choices.

The day concluded with a vibrant discussion about AI’s technical trajectory. Experts such as Edward A. Lee, Ute Schmid, and Michael Bronstein tackled the big question: How far can AI really go? Schmid emphasized that AI lacks intentionality and subjective experience, and Lee questioned whether agentic AI systems (proactive, decision-making machines) should ever be trusted with moral decisions. The conversation also highlighted that AI is powerful not because it mimics human thinking, but because it leverages patterns in data, sometimes in ways that even scientists cannot fully understand. There was a shared belief that AI must remain a tool in the hands of qualified humans, especially in science and education.

Day 2: Platforms and Power
The innovation, investment, and development of AI are currently dominated by China and the U.S. Unfortunately the originally scheduled speaker Yi Zeng, who could have offered valuable insights from a Chinese perspective, had to cancel. However, his replacement Dame Wendy Hall delivered an equally compelling contribution on global perspectives. While AI Safety Institutes in the US and UK, and the EU’s regulatory forums are emerging, they remain limited in scope and inclusivity as the people in the global south are left out. In response, the UN launched its Governing AI for Humanity initiative to support equitable access and raise a global fund. But due to political tensions, the UN now faces financial challenges.

Christiane Wendehorst continued the discussion by presenting the AGIDE project. Contrary to the initial assumption that differing values would drive divergent approaches, their research revealed five recurring narrative patterns that also appear to shift. As a law professor, she emphasized that instead of drafting new regulations, existing frameworks could suffice if interpreted through more inclusive, globally resonant narratives.

Christiane Wendehorst presenting the Narratives of Digital Ethics

The panel “Platforms as new institutions” tackled the growing power of digital platforms, framing them as new institutional actors that increasingly shape public life and norms. Their influence now rivals or even exceeds that of many nation-states. Philipp Staab pointed out that many states perceive themselves in a state of perpetual crisis (climate change, war, economic instability), which makes societies more receptive to technocratic and authoritarian governance. In this vacuum of trust and stability, platform companies step in, offering control, convenience, and structure, all while consolidating their power.

My favourite topic about the governance of AI, frameworks and policies (or, to be more accurate, the lack thereof) was covered in a session that brought together experts from academia (including the already mentioned Dame Wendy Hall, whom King Charles refers to as “the AI-lady”), policymaking (Anja Wyrobek, Legal Policy Adviser at the European Parliament), and international organizations (Paula Hidalgo-Sanchis, UNU Global AI Network Coordinator; and Marc Rotenberg, President and Founder of the Center for AI and Digital Policy). The CAIDP published a comprehensive sourcebook for AI policies, which collects major global and national frameworks.

Panel moderated by George Metakides

I’m so glad I stayed for the panel “Cybersecurity in New Times”. As with many digital issues, cybercrime is set to explode rapidly and unpredictably. Did you know that some of the most intense users of AI today are cybercriminal gangs in Myanmar? While cybercrime is often a “smash and grab” operation (where speed is the key to reselling the stolen assets), Rafal Rohozinski pointed out that the dynamic shifts entirely when state actors get involved. One of his most striking remarks was, that we speak so much about artificial intelligence, but we should really be talking about real stupidity, given that we’ve unleashed an entire digital industry without any form of product liability. I assume he meant the digital ecosystem as a whole, not just AI.

But it doesn’t have to be this way. Optimism came from Matteo Maffei, who highlighted concrete examples like the SCION project in Switzerland. SCION enables secure and resilient data transmission through verifiable, trusted paths and is being used in critical sectors: finance, healthcare, government, and the energy sector. SCION-based networks help shield vital infrastructure from cyberattacks. It’s a powerful reminder that with the right vision and investment, a more secure and sovereign digital future is entirely possible – provided there is the willpower, a solid concept and political agreement.

Day 3: Disruptive Innovation
The third day was marked by discussions on intersectionality, diversity, inclusion, and transdisciplinary approaches to technology (particularly AI) development and innovation. In his keynote speach, MIT professor Michael Cusumano gave an overview of the evolution and interdependencies among key companies and platforms shaping the AI landscape. A particularly striking point was the pivotal role NVIDIA has played in the development of generative AI. He also raised several critical big-picture concerns: content generation versus ownership, the concentration and misuse of platform power, the challenges for regulation or self-regulation, and the environmental impact. Although the trajectory of energy consumption was significantly changed by DeepSeek. Through reinforcement learning they demonstrated that large models could be trained far more efficiently. This not only challenged industry norms, but also put pressure on OpenAI by cutting prices for tokens to only a fraction.

Bashar Nuseibeh addressed the ethical and social responsibilities of software engineering, as every research and technology embodies values – consciously or not. He advocated for a more transdisciplinary approach: putting together diverse teams with different social norms to collaboratively solve social problems. A highlight of his talk was the exploration of overlooked contexts, such as animal privacy. While the notion might seem amusing at first (who wants to say goodbye to random cat pictures on the internet?), it raises serious and thought-provoking questions: how do we design technology for beings that cannot give consent, or in contexts where consent is difficult or impossible to extract? This example underscored the broader point that ethical technology design must account for diverse, often overlooked perspectives.

The panel “Reimagining AI through Social Innovation” brought together very different researchers who showcased ways to engage the public in the AI discourse. Gabriella Waters presented her work that uses neurodiversity as a framework to improve AI systems. Neurodivergent individuals process information differently, may have unique communication preferences, sensory experiences, and executive functioning styles. These insights can be used for the design and development of more inclusive, adaptive, and capable AI technologies.

Mirko Schäfer reported on the widespread usage of AI in public management in the Netherlands, where applications range from waste management and predictive maintenance of infrastructure to identifying so-called “problem addresses”. This increasing reliance on AI also raises serious ethical concerns. One particularly severe case he discussed was the Dutch child-benefit scandal, in which a risk assessment algorithm (intended to detect fraud) wrongly flagged thousands of families, leading to devastating consequences that are still felt today: “Lives were destroyed”. Although the initial response focused on repair, the severity of harm pushed the conversation toward long-term systemic renewal. One of those was f.e. the The Algorithm Register of the Dutch government.

From breakage to repair to renewal – methodes from the Dutch government

In the afternoon, every GDPR fan’s biggest crush, Max Schrems, joined the panel “From Social Media to AI: Lessons Learned”. Moderated by Noshir Contractor and joined by Chris Bail and Sunimal Mendis, the discussion unpacked the tangled evolution from social media regulation to the looming challenges of AI governance. Schrems, never one to mince words, reminded the audience that many legal frameworks already exist – some of them even dating back to the Roman empire – but they’re simply not enforced. He got the impression that since about last year many companies don’t even pretend to care anymore. Their attitude has shifted to a blunt “We’ll just do it anyway”. Chris Bail described design tweaks that could have vastly different impacts. F.e. On the (notoriously toxic) platform Nextdoor simply prompting users with “Would you like to rephrase this more kindly?” drastically reduced toxic posts.

From social media to AI: No lessons learned

The final panel I attended was “AI and Democracy: From Values to Action”. It brought together a group of experts (read: doers) exploring how democratic values can be translated into concrete action. From Austria to India, Ghana, and the UK, the speakers shared real-world experiences of fighting for fair and inclusive AI. The Austrian NGO epicenter.works is actively lobbying for fundamental rights in digital spaces. Siddhi Gupta shared insights from the practical work at the Inclusive AI Lab and the underrepresentation of the global south. Teki Akuetteh, Founder and Executive Director of the Africa Digital Rights Hub, called out the imbalance: while Africa has pioneered innovations like mobile payments, it still remains largely on the receiving end of AI developments. The benefits flow mainly to infrastructure-rich economies. Michael Veale emphasized the need for AI literacy, meaning not only technical know-how, but critical thinking skills that empower citizens to question, challenge, and reshape these systems.

Practitioners sharing how to put ideals into action

Takeaways
After three intense days packed with information, encounters, and ideas, I’m still processing. The range of topics was vast, sometimes overwhelming, always engaging. This (admittedly long) post captures just a fraction of what truly stayed with me. There was so much more: inspiring sessions I couldn’t fit in here, spontaneous conversations, and the in-between moments where I connected with brilliant, like-minded people. It was thought-provoking, energizing, and a reminder of how vital these spaces for exchange really are.

The Digital Humanism Conference wasn’t just about critique – it was about agency. About reclaiming the digital for the public good. About putting people, not profit, at the center of our digital future.

Three packed days, dozens of powerful talks, and one clear message: the future is still ours to shape!

Me and ICT4D lecturer Anna Bon

PS: Finally a scientific conference where I could fully thrive on fantastic vegan food – shoutout to Rita bringt’s! If only I had a button to mute everyone who uses the word “sustainability” with meat on their plate… but that’s a conversation for another day. 😉

Further readings:

Books about Digital Humanism

#conference #DigHum2025 #digitalHumanism

How about we try to enforce some kind of ‘Polarization Tax’?

Concept was mentioned during keynote on ‘AI and democracy’ by Lawrence Lessig (@lessig) and Moshe Vardi ( @vardi) at this event:

https://dighum.wien/programme/

#DigitalHumanism
#digihum
#AaronSwartz
#Fediscience

Programme – Digital Humanism Conference 2025