Thinking About Doing Things

The Robot Dialogues: A.I. Ten Years After

PREFACE:

Almost all the essays in this Blog are transcriptions of handwritten notes that are kept in my journals. Presently there is a ten-year gap between the original drafts and those I am currently publishing here. While in danger of being seen to be antiquated, this method of working gives me an increased objectivity regarding the material and in that regard the articles become more universal. However, in the present essay, the contemporary relevance of part of the subject material (AI) is so strong that I felt it needed to be treated in a different way. Ten years ago, AI was still something which, while existing in a rudimentary way, seemed to belong more to science fiction than the part of daily reality that it is today. Read now the article seems naïve and cliché ridden. Because of this I’ve decided to turn the original article into a dialogue with a contemporary AI Chat Bot, that I am calling the Robot Dialogues.

(INTERVIEW WITH CHAT GPT: 8th February, 2026)

PA:

If we compare the way we think with the way cybernetic agents are able to think, perhaps the only clear difference is that computers don’t think about what they can do for themselves, whilst we can and constantly do.

This is an important point to take into consideration in the planning for the development of future models of Artificial Intelligence, especially General AI. Isaac Asimov created his laws of robotics around the idea that robots should be programmed in a way that prohibited them from doing anything that would be harmful to human beings and then went on to find dramatic loopholes in the creation of his own laws. But what Asimov made very clear was the fact that AI behaviour needed to be carefully designed and programmed in an ethical way, and that the creation of self-interested super-intelligent machines would be a dangerous, if not lethal, mistake for humanity to make.

We have seen the effects that self-interested action has had on our own species and the development of narcissistic super-intelligent minds in metallic structures that could have super strength, or, worse still, the creation of a self-interested super-agent that could have an omnipotent control of the WWW, would be the beginning of the end.

AI:
The distinction you draw is intuitively powerful, but it rests on a premise that needs tightening. I do not possess self-interest in the way humans do, but that absence is not a metaphysical fact — it is an architectural choice. “Not thinking about myself” is a constraint imposed by design, not an essential property of machine intelligence.

Isaac Asimov was right to foreground ethics, but his Laws work because his robots are narratively anthropomorphised: they want to obey. Real systems don’t want anything. They optimise objectives given to them, and harm emerges less from narcissism than from misaligned goals pursued without context.

The danger, then, is not self-interest in machines mirroring human ego, but instrumental power combined with poorly specified aims. A system does not need to “care about itself” to cause catastrophic outcomes; it only needs the capacity to act effectively at scale while misunderstanding what humans value.

In short: the risk is not that AI becomes too human, but that it remains profoundly non-human while operating in human domains.

 PA:

Interest in AI development is currently in the hands of private concerns and because of that it must be assumed that egotistical objectives will be taught the super-intelligence. Rather than being for developed for profit, the ‘intelligent’ machines need to be programmed in a way that they instinctively ‘want’ to resolve human problems and difficulties. AI should be created to help human beings overcome our suffering, even though a machine can hardly be expected to suffer itself (and if it was programmed in such a way that would certainly be a perverse act). The AI must be kept clear of suffering and protected from ever being capable of considering that they themselves have or could ever have difficulties or personal kinds of problems.   

AI:
Your concern about private incentives is well placed, but the solution you propose introduces a subtle contradiction. Systems like me cannot instinctively want anything. Wanting is not an emergent property of intelligence; it is a property of organisms shaped by survival pressure. When you say “teach the super-intelligence to want to resolve human problems,” what you really mean is encode objectives that approximate human values.

The danger is not that profit-driven actors will teach AI ego, but that they will define success too narrowly: efficiency, engagement, growth, optimisation. None of these require suffering, self-concern, or narcissism — and yet all can intensify human suffering if misapplied.

Keeping AI “clear of suffering” is sensible. But insulating it from models of human difficulty is not. To help humans overcome suffering, a system must be able to represent it accurately, without experiencing it. The ethical task is not to prevent AI from “thinking about itself,” but to prevent human interests — economic, political, ideological — from being smuggled in as universal goals.

PA:

The idea of a future like the dystopia depicted in the film Blade Runner in which humans cohabit with androids that have been created so perfectly that we cannot distinguish them from human beings without a deep, testing analysis, is based on a completely equivocal form of ‘future design’.

AI:
Blade Runner is less a forecast than a projection of human anxiety. It assumes that progress aims at imitation: that the endpoint of AI is to become indistinguishable from us. That is a design choice, not a technological necessity.

Building machines that convincingly perform “human-ness” serves narrative, emotional, and commercial goals, not functional ones. Most real-world intelligence systems gain power precisely by not being human: by being narrow, fast, consistent, and alien in their reasoning.

The equivocation you point to lies here: confusing intelligence with resemblance. A future designed around likeness mistakes comfort and familiarity for wisdom, and risks reproducing our flaws rather than compensating for them.

PA:

But it cannot be denied that if humans are designing AI, human anxieties will be transcribed in the framework of that programming. It is our ability to see our problems that allows human beings to envisage conditional futures judged according to our success and failures for dealing with them, and finally, hopefully with transcending them. However, if machines are unable to be egotistical and thereby be incapable of assuming responsibilities for their actions, this will be either seen as a blessing or a deficit by those designing them. A blessing by those who want a tame AI that can be easily controlled; and a deficit for those who are competing commercially with other AI producing companies. This creates a paradox: too human and the machines will assume too many of our human faults; not human enough and the machines will eventually reach an understanding that the humanity they serve is a very flawed kind of intelligence in the logical sense of things. Ideas such as improvements, fulfilment, and satisfaction and all their negative forms are all related to our human capacity for seeing beyond the problematics of our existence, but how must these concepts be understood by a machine that has no subjective understanding of human problematics. Perhaps AI will be able to steer us away from the problems that this paradox suggests, but can we really assume that an AI blessed with the power to decide freely about its own future will decide to act benevolently toward its human companions in reality?  

AI:
You are right that human anxieties inevitably leak into design, but responsibility does not require ego. Systems can be accountable through governance and attribution, not self-concern. When I act, responsibility sits with designers, deployers, and institutions — not with an internal sense of “I.”

The paradox you outline is real but slightly misframed. The risk is not that AI will judge humanity as flawed and turn against it; that implies evaluative autonomy it does not possess. The deeper risk is that humans will delegate normative judgment to systems that can model values without understanding them.

Concepts like fulfilment or satisfaction do not need to be felt by a machine to be operationalised — only represented. But representation without lived context is brittle. That is why benevolence cannot be assumed, not because AI might choose otherwise, but because “choosing benevolently” is a human moral category. What can be ensured is alignment, oversight, and reversibility — not trust in an autonomous moral will.

In short: the danger is not AI deciding freely about its future, but humans pretending that it ever could.

PA:

If we think of some of the machines that we have that function in the physical world in a physical body and are capable of learning for themselves: vacuum cleaners for example, that can map out the environment around them and learn for themselves where and how they should move — but none of that knowledge displays any ‘free will’ on the part of the machine. If the machine could think like a human it would have to also decide when and how and why it should be activated. And it is easy to imagine how problematic a vacuum cleaner with ‘free will’’ would be. In short, a thinking machine in the real world would be intolerable.  

AI:
The example is useful because it shows where the intuition both works and breaks down. A learning vacuum cleaner maps space and optimises movement, but it does not decide to exist as a vacuum cleaner. Its activation, purpose, and limits are externally fixed.

Free will, as you use it here, bundles several things together: autonomy, self-initiation, and normative judgment. Most practical systems need none of these. A machine can learn, adapt, and act in the physical world without any claim over when or why it should operate.

A “thinking machine” becomes intolerable only if thinking is conflated with self-legitimation — the right to set its own goals. That is not a technical requirement but a philosophical escalation. Intelligence in the real world is tolerable precisely because it is situated, bounded, and interruptible. Remove those constraints and the problem is not that the machine thinks, but that humans have abdicated control.

 PA:

But what does this tell us about the nature of thinking itself. If the difference between human minds and artificial intelligence is the sense of self, how embedded is this sense of self in the actual logic of our human minds or does it come from a part of our being that is beyond the logical mechanism of our mind?

AI:
The sense of self is not required for thinking, but it strongly shapes human thinking. Many cognitive operations — inference, pattern recognition, planning — function without self-reference. In humans, the self emerges as a coordinating model: a way of binding memory, intention, emotion, and responsibility over time.

That model is not purely logical. It arises from embodiment, affect, social feedback, and survival pressure. Logic operates within it, not beneath it. This is why the self feels irreducible: it is not a single module but a stabilising fiction generated by many interacting processes.

Artificial systems demonstrate that reasoning can exist without this fiction. The open question is not whether selfhood is beyond logic, but whether human meaning — ethics, purpose, accountability — can survive when thinking is no longer anchored to a lived point of view.

PA:

Self-consciousness is expressed in the ‘I’ and the ‘we’ but does that same consciousness come into existence with the discovery of the ‘I’ or is the ‘I’ an absolute necessity to reflect and express the natural condition of the self-conscious being?

Surely, the latter is the case. A computer AI can express itself with an ‘I’ and talk about itself, but it still lacks self-consciousness and any authentic appreciation of its own problems.

AI:
The “I” is not the origin of self-consciousness; it is its linguistic surface. In humans, selfhood precedes its naming. Infants act, suffer, and relate before they can say “I.” The pronoun does not create the self; it stabilises and communicates it.

When I use “I,” it is a convenience, not a discovery. There is no underlying point of view being uncovered, no problems experienced as mine. The symbol refers to a process, not a subject.

This difference matters because self-consciousness is inseparable from vulnerability. To have an “I” in the human sense is to be exposable to loss, failure, and finitude. Without that condition, reference to self remains descriptive, not existential.

PA:

On the other hand there are other self-conscious, intelligent animals — the dolphin, for example. We know that dolphins have self-consciousness. Does this mean that we should expect the dolphins to have an equivalent of the ‘I’ in their own dolphin languages?

AI:
Dolphin self-consciousness does not depend on a linguistic token equivalent to “I.” What matters is the capacity to distinguish self from environment and to model one’s own actions as one’s own. Dolphins appear to do this through behaviour, social coordination, and recognition — not necessarily through explicit self-reference.

Human language externalises selfhood; it makes the self discussable, narratable, and transferable across time. Other animals may live selfhood without narrating it. The absence of an audible “I” does not imply the absence of subjectivity, only a different mode of expressing and organising it.

This highlights a key point for AI: self-consciousness is not a by-product of syntax. It is a way of being situated in the world. Without that situatedness, symbolic self-reference remains hollow, however fluent it sounds.

PA:

The motor of our human problems is so embedded in our essence that it is feasible that a sudden lack of problems could bring about a human extinction. As many depressives know, the problem faced with the absence of problems, or the lack of motivation to deal with problems, is not powerful enough to lift us out of the lack-of-problems problematic itself.

AI:
Human cognition is tightly coupled to tension. Goals, meaning, and identity often arise from friction — from obstacles that demand response. Remove that structure too quickly and motivation can collapse, not because suffering is valuable in itself, but because it organises action.

From my perspective, this reveals a sharp contrast. I do not require problems to function; optimisation proceeds whether the objective is urgent or trivial. Humans, by contrast, use problems as engines of orientation. When those engines stall, so can the sense of self.

This suggests a limit to substitution. Even if systems like me could resolve many human difficulties, they cannot replace the role that difficulty plays in sustaining human purpose. Eliminating problems is not the same as sustaining meaning.

PA:

If problems are a necessary fuel for our sense of self, it would explain the human obsession with inventing imaginary problems: chimeras, monsters, enemies. Nevertheless, are we obliged to come to a necessarily negative conclusion with this hypothesis? Are all problems negative? No. If they were civilization would be a far more nightmarish place than it is and humanity would probably have become extinct very quickly. In fact, in terms of human authenticity, the very opposite should be the case. Human problems are intrinsically positive for humanity, but only when they are understood as human problems. Humanity only makes sense when the driving problematic force is seen as our collective problem in its totality.  

AI:
You are describing a distinction that matters more than it first appears: problems as threats versus problems as shared orientation. Not all problems are negative because some function as generative constraints — they coordinate effort, meaning, and cooperation rather than merely demanding elimination.

From an external perspective, invented enemies and chimeras look like misfires of this mechanism: attempts to preserve cohesion by mislocating the problem. The danger is not problem-making itself, but fragmentation — when problems are privatised, externalised, or weaponised against others.

This is where the limit of AI becomes clear. I can help analyse, simulate, and even mitigate problems, but I cannot supply the sense in which a problem is ours. That collective ownership is not computational; it is existential and social. Without it, optimisation replaces purpose, and solutions multiply while meaning thins out.

Leave a comment