Think with us, not for us

Hi all!
First of all. Let me introduce myself to fresh readers. This newsletter is my personal weekly digest of the news of last week, of course through the lens of topics that I think are worth capturing, and reflecting upon: Human-AI collabs, physical AI (things and beyond), and tech and society. I always take one topic to reflect on a bit more, allowing the triggered thoughts to emerge. And I share what I noticed as potentially worthwhile things to do in the coming week.
If you'd like to know more about me, I added a bit more of my background at the end of the newsletter.
Enjoy! Iskander
What did happen last week?
Last week was a mix of curating, writing, and engaging in conversations with interesting people. Sounds generic, but I cannot share everything in detail. It was about the internet of entities and what that might entail. About makerlabs and responsible AI in human-AI teams, and of course, on civic protocol economies. Speaking about the last, we had a good turnout with the first round of the call for participation. I look forward to reviewing them in more detail later this week.
What did I notice last week?
Scroll down for all the notions from last week’s news, divided into human-AI partnerships, robotic performances, immersive connectedness, and tech societies. Let me choose one per category here:
Sens-AI framework teaches developers how to think with AI.
We need to rethink existing fundamentals… apparently.
What triggered my thoughts?
Initially, my thoughts were triggered listening to an interview with Liv Boeree, specifically the Moloch trap angle she connects to the outcome when we mix competitiveness and collective intelligence. The thoughts get more layers, combining it with a post of Ethan Mollick yesterday, “Against "Brain Damage; AI can help, or hurt, our thinking”. Even more than usual, writing these thoughts, I had a good conversation with Claude to sharpen my own thinking.
AI should think with us, not for us
There is still a lot to do about the research that AI makes us less intelligent. Ethan Mollick's recent analysis made me think of an interview I'd just heard with Liv Boeree about the Moloch trap: "This concept describes a situation where individual or group competition for a specific goal leads to a worse overall outcome for everyone involved"
Boeree's example was personal: she wants to learn and grow, but social media algorithms want engagement and entertainment. It's a perfect illustration of misaligned incentives. TikTok even artificially boosts your first videos to hook you as a creator, not just a consumer.
Here's where Mollick's analysis becomes crucial. He argues that the problem is not the use of AI as tools to extend our capabilities, but the design of AI that encourages us to be lazy. "The problem is that even honest attempts to use AI for help can backfire because the default mode of AI is to do the work for you, not with you."
But here's my thought: what if you really want to be whole? Current social media uses algorithms, not true AI. What if we replaced them with AI that we control? Imagine having a conversation with AI on a deeper level, that really understands what you strive for beyond instant gratification.
The algorithm itself isn't the threat - it's who controls its intentions. What if that person could be you? The intelligence tools we need should focus on reasoning and the exchange of insights, not just producing outcomes. AI should make us smarter through dialogue, not lazier through automation.
Mollick proposes sequencing: "Always generate your own ideas before turning to AI." I agree. That's exactly how I write these columns - my thoughts first, AI as editor second.
Mollick concludes: "Our fear of AI 'damaging our brains' is actually a fear of our own laziness. The technology offers an easy out from the hard work of thinking, and we worry we'll take it. We should worry. But we should also remember that we have a choice."
The Moloch trap isn't inevitable. We can choose AI that aligns with our deeper goals—growth, understanding, and wholeness—rather than just engagement. However, first, we need to recognize that we're the ones who should set those goals. And having an AI that thinks with us, not for us.
What inspiring paper to share?
I have been following this research for some time. We had a presentation by Seowoo Nam at ThingsCon some years ago, and it's great to see how it has evolved. Diffractive Interfaces: Facilitating Agential Cuts in Forest Data Across More-than-human Scales
This pictorial challenges these limitations by exploring how interface design can transcend reductive, agent-centric representations to foster relational understandings of forest ecosystems as more-than-human bodies. Drawing on feminist theorist Karen Barad's concepts of “diffraction” and “agential cuts,” we craft a repertoire of diffractive interfaces that engage with forest simulation data, revealing how more-than-human bodies can be encountered across diverse temporal, spatial, and agential scales.
*Elisa Giaccardi, Seowoo Nam, and Iohanna Nicenboim. 2025. Diffractive Interfaces: Facilitating Agential Cuts in Forest Data Across More-than-human Scales. In Proceedings of the 2025 ACM Designing Interactive Systems Conference (DIS '25). Association for Computing Machinery, New York, NY, USA, 135–147. https://doi.org/10.1145/3715336.3735404*
What are the plans for the coming week?
This is the last week before many people here will start their holiday break. Not me, but it will change the pace next week. And this week, too, having more meetings before people are gone. And summer drinks.
Also, checking the entries for the call for participation, preparing an inspiration session on agentic AI, I will do next week, and finalizing outlines for responsible AI programs.
We are planning the next ThingsCon Salon, organized in collaboration with the Human Values for Smarter Cities research program, just as we did last year and in 2023. This time, it will be on 4th September, and the preliminary topic is: “Hold on to good intentions.” More to follow later this week.
I will attend an afternoon by Future Society Lab in Rotterdam this afternoon. A Service Designer in Amsterdam might check the summer drinks this Thursday. This looks interesting: Divination, Prediction, and AI. Monday in London.
References to the notions
Human-AI partnerships
Does AI require us to be generalists or specialists? Or can you only be a good generalist leveraging AI when you have an expert foundation? (That is a rhetorical question.)
And more from Dan Shipper: finding a definition of AGI.

And more on AI vs critical thinking.
An arms race…

If you believe that everything that is computational in our workflows becomes AI, than this strategy of Grammarly becoming a productivity platform, makes sense.

Testing consciousness of AI via ACT. Check the two posts about is by Matt.

Is there an AI model now that can simulate the human mind. Or maybe not.
Sens-AI framework teaches developers how to think with AI.

How AI is shaping human-centered design

From humanware to machineware, engineers stop coding and start conducting.

Robotic performances
Creating robotic performances on a large scale.

If you're in LA, you might want to check out this art installation.

That feels like a fun use that makes sense: let this robot draw your building plan of your new house on the ground.

Self-driving buzz on German roads?
A different form of humanoid. Cyborg.

Immersive connectedness
A new type of tattoo vending machine?
Find hidden objects

AI immerses in your home.

Tech societies
The power of the lock in

Not sure if you should applaud this collaboration…

The exponential growth of LLMs.
The value of good policy in growing technology like AI.

We need to rethink existing fundamentals… apparently.

It will be easy to become a model yourself, I reckon.
Data is driving AI development more than ideas. Not sure, define ideas and data. Or is this opening up for a new round of ideas?

Who is surprised?

How to preserve attention.

AI and the rebirth of media

In the early days of the web, you had hidden pixels to count your visitors; now we have hidden prompts.
Silicon Valley wins big in the accelerated authoritarianism of Trump's beautiful bill

Have a great week!
About me
I'm an independent researcher, designer, curator, and “critical creative”, working on human-AI-things relationships. I am available for short or longer projects, leveraging my expertise as a critical creative director in human-AI services, as a researcher, or a curator of co-design and co-prototyping activities.
Contact me if you are looking for exploratory research into human-AI co-performances, inspirational presentations on cities of things, speculative design masterclasses, research through (co-)design into responsible AI, digital innovation strategies, and advice, or civic prototyping workshops on Hoodbot and other impactful intelligent technologies.
My guiding lens is cities of things as citizens, a research program that started in 2018, when I was a visiting professor at TU Delft's Industrial Design faculty. Since 2022, Cities of Things has become a foundation dedicated to collaborative research and sharing knowledge. In 2014, I co-initiated the Dutch chapter of ThingsCon—a platform that connects designers and makers of responsible technology in IoT, smart cities, and physical AI.
Signature projects are our 2-year program (2022-2023) with Rotterdam University and Afrikaander Wijkcooperatie has created a civic prototyping platform that helps citizens, policymakers, and urban designers shape living with urban robots: Wijkbot.
Recently, I've been developing programs on intelligent services for vulnerable communities, and contributing to the "power of design" agenda of CLICKNL, and since October 2024, co-developing a new research program on Civic Protocol Economies with Martijn de Waal at Amsterdam University of Applied Sciences.