Abstract: Practical Aspects of Artificial Intelligence – Tim Josling

team-Tim-JoslingWe are at the beginnings of an explosion in the everyday use of artificial intelligence. After decades of disappointment and frustration, artificial intelligence and machine learning application are appearing everywhere. They are about to change our world in ways that are unpredictable and full of promise and risks. This talk focuses on what we have achieved to date, what is happening now, and the likely path of AI over the next couple of decades. It explains why the pace of past and future progress is surprisingly predictable. Includes a case study on the use of Machine Learning theory in the investment world.

About Tim Josling

Tim Josling studied Law, Anthopology, Philosophy and Mathematics before switching to Computer Science at the dawn of the computer era. He worked on implementing some of the first transactional systems in Australia, later worked on the first ATM networks and was the chief architect for one of the first Internet Banking applications in Australia, and designed an early message switching (“middleware”) application in the USA. During his career he specialised in making large scale applications reliable and fast, saving several major projects from being cancelled due to poor performance and excessive running costs. This led to an interest in the progress of computer hardware and in Moore’s Law, which states that the power of computers grows roughly 10-fold every 5 years. In his spare time he contributed to various open source projects such as the GNU Compiler Collection. After attending the first Singularity Summit in Australia, he decided to retire so he could devote himself full-time to researching Artificial Intelligence, the Technological Singularity and Trans-humanism. He is currently working on applying AI techniques to financial and investment applications.


Abstract – Exploring the Frontier of Molecular Biology with Dynamic Visualisation – Drew Berry


Central to the science of biology is the complex choreography of cells and molecules. A key problem with explaining this science to the public is that much of it exists in a microscopic world that is too small to be directly observed, or takes place at speeds beyond our normal perception of time.

Creating visualisations of cellular and molecular biology has become increasingly important for exploring and showing biological mechanisms to the public, students and scientific peers.

Visualisation is able to synthesize diverse structural and dynamic data derived from a variety of research sources, and can thus act as a visual hypothesis for a particular molecular process.

Beyond the bench, visualisations are powerful communication tools that are being used in classrooms and in the mass media to educate and entertain.

About the speaker

Drew Berry  is a biomedical animator at the Walter and Eliza Hall Institute of Medical Research in Melbourne, Australia. His scientifically accurate and aesthetically rich visualizations are elucidating cellular and molecular processes for a wide range of audiences. His animations have been shown in exhibitions, multimedia programs and television shows, and have received international recognition including an Emmy (2005) and a BAFTA Award (2004).




Abstract: The Abolitionist Project – a Hedonistic Imperative – David Pearce

David Pearceteam-david-pearce

Reproduced from


This talk is about suffering and how to get rid of it.
I predict we will abolish suffering throughout the living world.
Our descendants will be animated by gradients of genetically preprogrammed well-being that are orders of magnitude richer than today’s peak experiences.

First, I’m going to outline why it’s technically feasible to abolish the biological substrates of any kind of unpleasant experience – psychological pain as well as physical pain.
Secondly, I’m going to argue for the overriding moral urgency of the abolitionist project, whether or not one is any kind of ethical utilitarian.
Thirdly, I’m going to argue why a revolution in biotechnology means it’s going to happen, albeit not nearly as fast as it should.



Sadly, what won’t abolish suffering, or at least not on its own, is socio-economic reform, or exponential economic growth, or technological progress in the usual sense, or any of the traditional panaceas for solving the world’s ills. Improving the external environment is admirable and important; but such improvement can’t recalibrate our hedonic treadmill above a genetically constrained ceiling. Twin studies confirm there is a [partially] heritable set-point of well-being – or ill-being – around which we all tend to fluctuate over the course of a lifetime. This set-point varies between individuals. [It’s possible to lower our hedonic set-point by inflicting prolonged uncontrolled stress; but even this re-set is not as easy as it sounds: suicide-rates typically go down in wartime; and six months after a quadriplegia-inducing accident, studies1 suggest that we are typically neither more nor less unhappy than we were before the catastrophic event.] Unfortunately, attempts to build an ideal society can’t overcome this biological ceiling, whether utopias of the left or right, free-market or socialist, religious or secular, futuristic high-tech or simply cultivating one’s garden. Even if everything that traditional futurists have asked for is delivered – eternal youth, unlimited material wealth, morphological freedom, superintelligence, immersive VR, molecular nanotechnology, etc – there is no evidence that our subjective quality of life would on average significantly surpass the quality of life of our hunter-gatherer ancestors – or a New Guinea tribesman today – in the absence of reward pathway enrichment. This claim is difficult to prove in the absence of sophisticated neuroscanning; but objective indices of psychological distress e.g. suicide rates, bear it out. Unenhanced humans will still be prey to the spectrum of Darwinian emotions, ranging from terrible suffering to petty disappointments and frustrations – sadness, anxiety, jealousy, existential angst. Their biology is part of “what it means to be human”. Subjectively unpleasant states of consciousness exist because they were genetically adaptive. Each of our core emotions had a distinct signalling role in our evolutionary past: they tended to promote behaviours which enhanced the inclusive fitness of our genes in the ancestral environment.

So if manipulating our external environment alone can never abolish suffering and malaise, what does technically work?

Here are three scenarios in ascending order of sociological plausibility:


a) wireheading
b) utopian designer drugs
c) genetic engineering
and – what I want to focus on – the impending reproductive revolution of designer babiesa) Recall wireheading is direct stimulation of the pleasure centres of the brain via implanted electrodes. Intracranial self-stimulation shows no physiological or subjective tolerance i.e. it’s just as rewarding after two days as it is after two minutes. Wireheading doesn’t harm others; it has a small ecological footprint; it banishes psychological and physical pain; and arguably it’s a lot less offensive to human dignity than having sex. Admittedly, lifelong wireheading sounds an appealing prospect only to a handful of severe depressives. But what are the technical arguments against its adoption?

brain sparkWell, wireheading is not an evolutionarily stable solution: there would be selection pressure against its widespread adoption. Wireheading doesn’t promote nurturing behaviour: wireheads, whether human or non-human, don’t want to raise baby wireheads. Uniform, indiscriminate bliss in the guise of wireheading or its equivalents would effectively bring the human experiment to an end, at least if it were adopted globally. Direct neurostimulation of the reward centres destroys informational sensitivity to environmental stimuli. So assuming we want to be smart – and become smarter – we have a choice. Intelligent agents can have a motivational structure based on gradients of ill-being, characteristic of some lifelong depressives today. Or intelligent agents can have our current typical mixture of pleasures and pains. Or alternatively, we could have an informational economy of mind based entirely on [adaptive] gradients of cerebral bliss – which I’m going to argue for.

Actually, this dismissal of wireheading may be too quick. In the far future, one can’t rule out offloading everything unpleasant or mundane onto inorganic supercomputers, prostheses and robots while we enjoy uniform orgasmic bliss. Or maybe not orgasmic bliss, possibly some other family of ideal states that simply couldn’t be improved upon. But that’s speculative. Whatever our ultimate destination, it would be more prudent, I think, to aim for both superhappiness and superintelligence – at least until we understand the full implications of what we are doing. There isn’t a moral urgency to maximizing superhappiness in the same way as there is to abolishing suffering.

[It’s worth noting that the offloading option assumes that inorganic computers, prostheses and robots don’t – or at least needn’t – experience subjective phenomenal pain even if their functional architecture allows them to avoid and respond to noxious stimuli. This absence of inorganic suffering is relatively uncontroversial with existing computers – switching off one’s PC doesn’t have ethical implications, and a silicon robot can be programmed to avoid corrosive acids without experiencing agony if it’s damaged. It’s debatable whether any computational system with a classical von Neumann architecture will ever be interestingly conscious. I’m sceptical; but either way, it doesn’t affect the offloading option, unless one argues that the subjective texture of suffering is functionally essential to any system capable of avoiding harmful stimuli.]

designer-drug-abuseb) The second technical option for eradicating suffering is futuristic designer drugs. In an era of mature post-genomic medicine, will it be possible rationally to design truly ideal pleasure-drugs that deliver lifelong, high-functioning well-being without unacceptable side-effects? “Ideal pleasure drugs” here is just a piece of shorthand. Such drugs can in principle embrace cerebral, empathetic, aesthetic and perhaps spiritual well-being – and not just hedonistic pleasure in the usual one-dimensional and amoral sense.
We’re not talking here about recreational euphoriants, which simply activate the negative feedback mechanisms of the brain; nor the shallow, opiated contentment of a Brave New World; nor drugs that induce euphoric mania, with its uncontrolled excitement, loss of critical insight, grandiosity and flight of ideas. Can we develop true wonderdrugs that deliver sublime well-being on a sustainable basis, recalibrating the hedonic treadmill to ensure a high quality of life for everyone?

A lot of people recoil from the word “drugs” – which is understandable given today’s noxious street drugs and their uninspiring medical counterparts. Yet even academics and intellectuals in our society typically take the prototypical dumb drug, ethyl alcohol. If it’s socially acceptable to take a drug that makes you temporarily happy and stupid, then why not rationally design drugs to make people perpetually happier and smarter? Presumably, in order to limit abuse-potential, one would want any ideal pleasure drug to be akin – in one limited but important sense – to nicotine, where the smoker’s brain finely calibrates its optimal level: there is no uncontrolled dose-escalation.

There are of course all kinds of pitfalls to drug-based solutions. Technically, I think these pitfalls can be overcome, though I won’t try to show this here. But there is a deeper issue. If there weren’t something fundamentally wrong – or at least fundamentally inadequate – with our existing natural state of consciousness bequeathed by evolution, then we wouldn’t be so keen to change it. Even when it’s not unpleasant, everyday consciousness is mediocre compared to what we call peak experiences. Ordinary everyday consciousness was presumably adaptive in the sense it helped our genes leave more copies of themselves on the African savannah; but why keep it as our default-state indefinitely? Why not change human nature by literally repairing our genetic code?

Again, this dismissal of pharmacological solutions may be too quick. Arguably, utopian designer drugs may always be useful for the fine-grained and readily reversible control of consciousness; and I think designer drugs will be an indispensable tool to explore the disparate varieties of conscious mind. But wouldn’t it be better if we were all born with a genetic predisposition to psychological superhealth rather than needing chronic self-medication? Does even the most ardent abolitionist propose to give cocktails of drugs to all children from birth; and then to take such drug cocktails for the rest of our lives?

abolitionistc) So thirdly, there are genetic solutions, embracing both somatic and germline therapy.
By way of context, today there is a minority of people who are always depressed or dysthymic, albeit to varying degrees. Studies with mono- and dizygotic twins confirm there is a high degree of genetic loading for depression. Conversely, there are some people who are temperamentally optimistic. Beyond the optimists, there is a very small minority of people who are what psychiatrists call hyperthymic. Hyperthymic people aren’t manic or bipolar; but by contemporary standards, they are always exceedingly happy, albeit sometimes happier than others. Hyperthymic people respond “appropriately” and adaptively to their environment. Indeed they are characteristically energetic, productive and creative. Even when they are blissful, they aren’t “blissed out”.

Now what if, as a whole civilisation, we were to opt to become genetically hyperthymic – to adopt a motivational system driven entirely by adaptive gradients of well-being? More radically, as the genetic basis of hedonic tone is understood, might we opt to add multiple extra copies of hyperthymia-promoting genes/allelic combinations and their regulatory promoters – not abolishing homeostasis and the hedonic treadmill but shifting our hedonic set-point to a vastly higher level?

Three points here:
First, this genetic recalibration might seem to be endorsing another kind of uniformity; but it’s worth recalling that happier people – and especially hyperdopaminergic people – are typically responsive to a broader range of potentially rewarding stimuli than depressives: they engage in more exploratory behaviour. This makes getting stuck in a sub-optimal rut less likely, both for the enhanced individual and posthuman society as a whole.

Secondly, universal hyperthymia might sound like a gigantic experiment; and in a sense of course it is. But all sexual reproduction is an experiment. We play genetic roulette, shuffling our genes and then throwing the genetic dice. Most of us flinch at the word “eugenics”; but that’s what we’re effectively practising, crudely and incompetently, when we choose our prospective mates. The difference is that within the next few decades, prospective parents will be able to act progressively more rationally and responsibly in their reproductive decisions. Pre-implantation diagnosis is going to become routine; artificial wombs will release us from the constraints of the human birth-canal; and a revolution in reproductive medicine will begin to replace the old Darwinian lottery. The question is not whether a reproductive revolution is coming, but rather what kinds of being – and what kinds of consciousness – do we want to create?

david_pearceThirdly, isn’t this reproductive revolution going to be the prerogative of rich elites in the West? Probably not for long. Compare the brief lag between the introduction of, say, mobile phones and their world-wide adoption with the 50 year time-lag between the introduction and world-wide adoption of radio; and the 20 year lag between the introduction and world-wide penetration of television. The time-lag between the initial introduction and global acceptance of new technologies is shrinking rapidly. So of course is the price.


Anyway, one of the advantages of genetically recalibrating the hedonic treadmill rather than abolishing it altogether, at least for the foreseeable future, is that the functional analogues of pain, anxiety, guilt and even depression can be preserved without their nasty raw feels as we understand them today. We can retain the functional analogues of discontent – arguably the motor of progress – and retain the discernment and critical insight lacking in the euphorically manic. Even if hedonic tone is massively enhanced, and even if our reward centres are physically and functionally amplified, then it’s still possible in principle to conserve much of our existing preference architecture. If you prefer Mozart to Beethoven, or philosophy to pushpin, then you can still retain this preference ranking even if your hedonic tone is hugely enriched.

Now personally, I think it would be better if our preference architecture were radically changed, and we pursued [please pardon the jargon] a “re-encephalisation of emotion”. Evolution via natural selection has left us strongly predisposed to form all manner of dysfunctional preferences that harm both ourselves and others for the benefit of our genes. Recall Genghis Khan: “The greatest happiness is to scatter your enemy, to drive him before you, to see his cities reduced to ashes, to see those who love him shrouded in tears, and to gather into your bosom his wives and daughters.”

Now I’m told academia isn’t quite that bad, but even university life has its forms of urbane savagery – its competitive status-seeking and alpha-male dominance rituals: a zero-sum game with many losers. Too many of our preferences reflect nasty behaviours and states of mind that were genetically adaptive in the ancestral environment. Instead, wouldn’t it be better if we rewrote our own corrupt code? I’ve focused here on genetically enhancing hedonic tone. Yet mastery of the biology of emotion means that we’ll be able, for instance, to enlarge our capacity for empathy, functionally amplifying mirror neurons and engineering a sustained increase in oxytocin-release to promote trust and sociability. Likewise, we can identify the molecular signatures of, say, spirituality, our aesthetic sense, or our sense of humour – and modulate and “over-express” their psychological machinery too. From an information-theoretic perspective, what is critical to an adaptive, flexible, intelligent response to the world is not our absolute point on a hedonic scale but that we are informationally sensitive to differences. Indeed information theorists sometimes simply define information as a “difference that makes a difference”.

However, to stress again, this re-encephalisation of emotion is optional. It’s technically feasible to engineer the well-being of all sentience and retain most but not all of our existing preference architecture. The three technical options for abolishing suffering that I’ve presented – wireheading, designer drugs and genetic engineering – aren’t mutually exclusive. Are they exhaustive? I don’t know of any other viable options. Some transhumanists believe we could one day all be scanned, digitized and uploaded into inorganic computers and reprogrammed. Well, perhaps, I’m sceptical; but in any case, this proposal doesn’t solve the suffering of existing organic life unless we embrace so-called destructive uploading – a Holocaust option I’m not even going to consider here.



Assume that within the next few centuries we will acquire these Godlike powers over our emotions. Assume, too, that the signalling function of unpleasant experience can be replaced – either through the recalibration argued for here, or through the offloading of everything unpleasant or routine to inorganic prostheses, bionic implants or inorganic computers – or perhaps through outright elimination in the case of something like jealousy. Why should we all be abolitionists?

If one is a classical utilitarian, then the abolitionist project follows: it’s Bentham plus biotechnology. One doesn’t have to be a classical utilitarian to endorse the abolition of suffering; but all classical utilitarians should embrace the abolitionist project. Bentham championed social and legislative reform, which is great as far as it goes; but he was working before the era of biotechnology and genetic medicine.

If one is a scientifically enlightened Buddhist, then the abolitionist project follows too. Buddhists, uniquely among the world’s religions, focus on the primacy of suffering in the living world. Buddhists may think that the Noble Eightfold Path offers a surer route to Nirvana than genetic engineering; but it’s hard for a Buddhist to argue in principle against biotech if it works. Buddhists focus on relieving suffering via the extinction of desire; yet it’s worth noting this extinction is technically optional, and might arguably lead to a stagnant society. Instead it’s possible both to abolish suffering and continue to have all manner of desires.

Persuading followers of Islam and the Judeo-Christian tradition is more of a challenge. But believers claim – despite anomalies in the empirical evidence – that Allah/God is infinitely compassionate and merciful. So if mere mortals can envisage the well-being of all sentience, it would seem blasphemous to claim that God is more limited in the scope of His benevolence.

Most contemporary philosophers aren’t classical utilitarians or Buddhists or theists. Why should, say, an ethical pluralist take the abolitionist project seriously?
Here I want to take as my text Shakespeare’s


“For there was never yet philosopher
That could endure the toothache patiently”

[Much Ado About Nothing, Scene Five, Act One (Leonato speaking)]

a-light-in-the-midst-of-so-much-suffering-paulo-zerbatoWhen one is struck by excruciating physical pain, one is always shocked at just how frightful it can be.
It’s tempting to suppose that purely “psychological” pain – loneliness, rejection, existential angst, grief, anxiety, depression – can’t be as atrocious as extreme physical pain; yet the reason over 800,000 people in the world take their own lives every year is mainly psychological distress. It’s not that other things – great art, friendship, social justice, a sense of humour, cultivating excellence of character, academic scholarship, etc – aren’t valuable; but rather when intense physical or psychological distress intrudes – either in one’s own life or that of a loved one – we recognize that this intense pain has immediate priority and urgency. If you are in agony after catching your hand in the door, then you’d give short shrift to someone who urged you to remember the finer things in life. If you’re distraught after an unhappy love affair, then you don’t want to be tactlessly reminded it’s a beautiful day outside.

OK, while it lasts, extreme pain or psychological distress has an urgency and priority that overrides the rest of one’s life projects; but so what? When the misery passes, why not just get on with one’s life as before?
Well, natural science aspires to “a view from nowhere”, a notional God’s-eye view. Physics tells us that no here-and-now is privileged over any other; all are equally real. Science and technology are shortly going to give us Godlike powers over the entire living world to match this Godlike perspective. I argue that so long as there is any sentient being who is undergoing suffering similar to our distress, that suffering should be tackled with the same priority and urgency as if it were one’s own pain or the pain of a loved one. With power comes complicity. Godlike powers carry godlike responsibilities. Thus the existence of suffering 200 years ago, for instance, may indeed have been terrible; but it’s not clear that such suffering can sensibly be called “immoral” – because there wasn’t much that could be done about it. But thanks to biotechnology, now there is – or shortly will be. Over the next few centuries, suffering of any kind is going to become optional.

If you’re not a classical ethical utilitarian, the advantage of recalibrating the hedonic treadmill rather than simply seeking to maximise superhappiness is that you are retaining at least a recognizable descendant of our existing preference architecture. Recalibration of the hedonic treadmill can be made consistent with your existing value scheme. Hence even the ill-named “preference utilitarian” can be accommodated. Indeed control over the emotions means that you can pursue your existing life projects more effectively.
And what about the alleged character-building function of suffering? “That which does not crush me makes me stronger”, said Nietzsche. This worry seems misplaced. Other things being equal, enhancing hedonic tone strengthens motivation – it makes us psychologically more robust. By contrast, prolonged low mood leads to a syndrome of learned helplessness and behavioural despair.

I haven’t explicitly addressed the value nihilist – the subjectivist or ethical sceptic who says all values are simply matters of opinion, and that one can’t logically derive an “ought” from an “is”.
Well, let’s say I find myself in agony because my hand is on a hot stove. That agony is intrinsically motivating, even if my conviction that I ought to withdraw my hand doesn’t follow the formal canons of logical inference. If one takes the scientific world-picture seriously, then there is nothing ontologically special or privileged about here-and-now or me – the egocentric illusion is a trick of perspective engineered by selfish DNA. If it’s wrong for me to be in agony, then it is wrong for anyone, anywhere.



OK, it’s technically feasible. A world without suffering would be wonderful; and full-blown paradise-engineering even better. But again, so what? It’s technically feasible to build a thousand-metre cube of cheddar cheese. Why is a pain-free world going to happen? Perhaps it’s just wishful thinking. Perhaps we’ll opt to retain the biology of suffering indefinitely2.

The counterargument here is that whether or not one is sympathetic to the abolitionist project, we are heading for a reproductive revolution of designer babies. Prospective parents are soon going to be choosing the characteristics of their future children. We’re on the eve of the Post-Darwinian Transition, not in the sense that selection pressure will be any less severe, but evolution will no longer be “blind” and “random”: there will no longer be natural selection but unnatural selection. We will be choosing the genetic makeup of our future offspring, selecting and designing alleles and allelic combinations in anticipation of their consequences. There will be selection pressure against nastier alleles and allelic combinations that were adaptive in the ancestral environment.

Unfortunately, this isn’t a rigorous argument, but imagine you are choosing the genetic dial-settings for mood – the hedonic set-point – of your future children. What settings would you pick? You might not want gradients of lifelong superhappiness, but the overwhelming bulk of parents will surely want to choose happy children. For a start, they are more fun to raise. Most parents across most cultures say, I think sincerely, that they want their children to be happy. One may be sceptical of parents who say happiness is the only thing they care about for their kids – many parents are highly ambitious. But other things being equal, happiness signals success – possibly the ultimate evolutionary origin of why we value the happiness of our children as well as our own.

Of course the parental choice argument isn’t decisive. Not least, it’s unclear how many more generations of free reproductive choices lie ahead before radical antiaging technologies force a progressively tighter collective control over our reproductive decisions – since a swelling population of ageless quasi-immortals can’t multiply indefinitely in finite physical space. But even if centralised control of reproductive decisions becomes the norm, and procreation itself becomes rare, the selection pressure against primitive Darwinian genotypes will presumably be intense. Thus it’s hard to envisage what future social formations would really allow the premeditated creation of any predisposition to depressive or anxiety disorders – or even the “normal” pathologies of unenhanced consciousness.

Non-Human Animals

chimpanzee1So far I’ve focused on suffering in just one species. This restriction of the abolitionist project is parochial; but our anthropocentric bias is deeply rooted. Hunting, killing, and exploiting members of other species enhanced the inclusive fitness of our genes in the ancestral environment. [Here we are more akin to chimpanzees than bonobos.] So unlike, say, the incest taboo, we don’t have an innate predisposition to find, say, hunting and exploiting non-human animals wrong. We read that Irene Pepperberg’s parrot, with whom we last shared a common ancestor several hundred million years ago, had the mental age of a three-year-old child. But it’s still legal for so-called sportsmen to shoot birds for fun. If sportsmen shot babies and toddlers of our own species for fun, they’d be judged criminal sociopaths and locked up.

So there is a contrast: the lead story in the news media is often a terrible case of human child abuse and neglect, an abducted toddler, or abandoned Romanian orphans. Our greatest hate-figures are child abusers and child murderers. Yet we routinely pay for the industrialized mass killing of other sentient beings so we can eat them. We eat meat even though there’s a wealth of evidence that functionally, emotionally, intellectually – and critically, in their capacity to suffer – the non-human animals we factory-farm and kill are equivalent to human babies and toddlers.

From a notional God’s-eye perspective, I’d argue that morally we should care just as much about the abuse of functionally equivalent non-human animals as we do about members of our own species – about the abuse and killing of a pig as we do about the abuse or killing of a human toddler. This violates our human moral intuitions; but our moral intuitions simply can’t be trusted. They reflect our anthropocentric bias – not just a moral limitation but an intellectual and perceptual limitation too. It’s not that there are no differences between human and non-human animals, any more than there are no differences between black people and white people, freeborn citizens and slaves, men and women, Jews and gentiles, gays or heterosexuals. The question is rather: are they morally relevant differences? This matters because morally catastrophic consequences can ensue when we latch on to a real but morally irrelevant difference between sentient beings. [Recall how Aristotle, for instance, defended slavery. How could he be so blind?] Our moral intuitions are poisoned by genetic self-interest – they weren’t designed to take an impartial God’s-eye view. But greater intelligence brings a greater cognitive capacity for empathy – and potentially an extended circle of compassion. Maybe our superintelligent/superempathetic descendants will view non-human animal abuse as no less abhorrent than we view child abuse: a terrible perversion.


dolphinTrue or not, surely we aren’t going to give up eating each other? Our self-interested bias is too strong. We like the taste of meat too much. Isn’t the notion of global veganism just utopian dreaming?
Perhaps so. Yet within a few decades, the advent of genetically-engineered vatfood means that we can enjoy eating “meat” tastier than anything available today – without any killing and cruelty. As a foretaste of what’s in store, the In Vitro Meat Consortium was initiated at a workshop held at the Norwegian University of Life Sciences in June 2007. Critically, growing meat from genetically-engineered single cells is likely to be scalable indefinitely: its global mass consumption is potentially cheaper than using intact non-human animals. Therefore – assuming that for the foreseeable future we retain the cash nexus and market economics – cheap, delicious vatfood is likely to displace the factory-farming and mass-killing of our fellow creatures.

One might wonder sceptically: are most people really going to eat gourmet vatfood, even if it’s cheaper and more palatable than flesh from butchered non-human animals?
If we may assume that vatfood is marketed properly, yes. For if we discover that we prefer the taste of vat-grown meat to carcasses of dead animals, then the moral arguments for a cruelty-free diet will probably seem much more compelling than they do at present.

Yet even if we have global veganism, surely there will still be terrible cruelty in Nature? Wildlife documentaries give us a very Bambified view of the living world: it doesn’t make good TV spending half an hour showing a non-human animal dying of thirst or hunger, or slowly being asphyxiated and eaten alive by a predator. And surely there has to be a food chain? Nature is cruel; but predators will always be essential on pain of a population explosion and Malthusian catastrophe?

ape-ipadNot so. If we want to, we can use depot contraception3, redesign the global ecosystem, and rewrite the vertebrate genome to get rid of suffering in the rest of the natural world too. For non-human animals don’t need liberating; they need looking after. We have a duty of care, just as we do to human babies and toddlers, to the old, and the mentally handicapped. This prospect might sound remote; but habitat-destruction means that effectively all that will be left of Nature later this century is our wildlife parks. Just as we don’t feed terrified live rodents to snakes in zoos – we recognize that’s barbaric – will we really continue to permit cruelties in our terrestrial wildlife parks because they are “natural”?

The last frontier on Planet Earth is the ocean. Intuitively, this might seem to entail too complicated a task. But the exponential growth of computer power and nanorobotic technologies means that we can in theory comprehensively re-engineer the marine ecosystem too. Currently such re-engineering is still impossible; in a few decades, it will be computationally feasible but challenging; eventually, it will be technically trivial. So the question is: will we actually do it? Should we do it – or alternatively should we conserve the Darwinian status quo? Here we are clearly in the realm of speculation. Yet one may appeal to what might be called The Principle Of Weak Benevolence. Unlike the controversial claim that superintelligence entails superempathy, The Principle Of Weak Benevolence doesn’t assume that our technologically and cognitively advanced descendants will be any more morally advanced than we are now.

Let’s give a concrete example of how the principle applies. If presented today with the choice of buying either free-range or factory-farmed eggs, most consumers will pick the free-range eggs. If battery-farmed eggs are 1 penny cheaper, most people will still pick the “cruelty-free” option. No, one shouldn’t underestimate human malice, spite and bloody-mindedness; but most of us have at least a weak bias towards benevolence. If any non-negligible element of self-sacrifice is involved, for example if free-range eggs cost even 20 pence more, then sadly sales fall off sharply. My point is that if – and it’s a big if – the sacrifice involved for the morally apathetic could be made non-existent or trivial, then the abolitionist project can be carried to the furthest reaches of the living world.


intelligence explosion 9

The Bio-intelligence Explosion – David Pearce

The Biointelligence Explosion
How recursively self-improving organic robots will modify their own
source code and bootstrap our way to full-spectrum superintelligence

by David Pearce (2012) reproduced from
“Homo sapiens, the first truly free species, is about to decommission natural selection, the force that made us…. Soon we must look deep within ourselves and decide what we wish to become.”
Edward O. Wilson
Consilience, The Unity of Knowledge (1999) “I predict that the domestication of biotechnology will dominate our lives during the next fifty years at least as much as the domestication of computers has dominated our lives during the previous fifty years.”
Freeman Dyson
New York Review of Books (July 19, 2007)


team-david-pearce1 The Fate of the Germline

Genetic evolution is slow. Progress in artificial intelligence is fast. Only a handful of genes separate Homo sapiens from our hominid ancestors on the African savannah. Among our 23,000-odd protein-coding genes, variance in single nucleotide polymorphisms (“SNPs”) accounts for just a small percentage of phenotypic variance in intelligence as measured by what we call IQ tests. True, the tempo of human evolution is about to accelerate. As the reproductive revolution of “designer babies” gathers pace, prospective parents will pre-select alleles and allelic combinations for a new child in anticipation of their behavioural effects – a novel kind of selection pressure to replace the “blind” genetic roulette of natural selection. In time, routine embryo screening via preimplantation genetic diagnosis will be complemented by gene therapy, genetic enhancement and then true designer zygotes. In consequence, life on Earth will also become progressively happier as the hedonic treadmill is recalibrated. In the new reproductive era, hedonic set-points and intelligence alike will be ratcheted upwards in virtue of selection pressure. For what parent-to-be wants to give birth to a low-status depressive “loser”? Future parents can enjoy raising a normal transhuman supergenius who grows up to be faster than Usain Bolt, more beautiful than Marilyn Monroe, more saintly than Nelson Mandela, more creative than Shakespeare – and smarter than Einstein.

Even so, the accelerating growth of germline engineering will be a comparatively slow process. In this scenario, sentient biological machines will design cognitively self-amplifying biological machines who will design cognitively self-amplifying biological machines. Greater-than-human biological intelligence will transform itself into posthuman superintelligence. Cumulative gains in intellectual capacity and subjective well-being across the generations will play out over hundreds and perhaps thousands of years – a momentous discontinuity, for sure, and a twinkle in the eye of eternity; but not a BioSingularity.


2 Biohacking Your Personal Genome

genomeYet germline engineering is only one strand of the genomics revolution. Indeed after humans master the ageing process, the extent to which traditional germlines or human generations will persist in the post-ageing world is obscure. Focus on the human germline ignores the slow-burning but then explosive growth of somatic gene enhancement in prospect. Later this century, innovative gene therapies will be succeeded by gene enhancement technologies – a value-laden dichotomy that reflects our impoverished human aspirations. Starting with individual genes, then clusters of genes, and eventually hundreds of genes and alternative splice variants, a host of recursively self-improving organic robots (“biohackers”) will modify their genetic source code and modes of sentience: their senses, their moods, their motivation, their cognitive apparatus, their world-simulations and their default state of consciousness.

As the era of open-source genetics unfolds, tomorrow’s biohackers will add, delete, edit and customise their own legacy code in a positive feedback loop of cognitive enhancement. Computer-aided genetic engineering will empower biological humans, transhumans and then posthumans to synthesise and insert new genes, variant alleles and even designer chromosomes – reweaving the multiple layers of regulation of our DNA to suit their wishes and dreams rather than the inclusive fitness of their genes in the ancestral environment. Collaborating and competing, next-generation biohackers will use stem-cell technologies to expand their minds, literally, via controlled neurogenesis. Freed from the constraints of the human birth canal, biohackers may re-sculpt the prison-like skull of Homo sapiens to accommodate a larger mind/brain, which can initiate recursive self-expansion in turn. Six crumpled layers of neocortex fed by today’s miserly reward pathways aren’t the upper bound of conscious mind, merely its seedbed. Each biological neuron and glial cell of your growing mind/brain can have its own dedicated artificial healthcare team, web-enabled nanobot support staff, and social network specialists; compare today’s anonymous neural porridge. Transhuman minds will be augmented with neurochips, molecular nanotechnology, mind/computer interfaces and full-immersion virtual reality (VR) software. To achieve finer-grained control of cognition, mood and motivation, genetically enhanced transhumans will draw upon exquisitely tailored new designer drugs, nutraceuticals and cognitive enhancers – precision tools that make today’s crude interventions seem the functional equivalent of glue-sniffing.

By way of comparison, early in the twenty-first century the scientific counterculture is customizing a bewildering array of designer drugs that outstrip the capacity of the authorities to regulate or comprehend. The bizarre psychoactive effects of such agents dramatically expand the evidential base that our theory of consciousness must explain. However, such drugs are short-acting. Their benefits, if any, aren’t cumulative. By contrast, the ability genetically to hack one’s own source code will unleash an exponential growth of genomic rewrites – not mere genetic tinkering but a comprehensive redesign of “human nature”. Exponential growth starts out almost unnoticeably, and then explodes. Human bodies, cognition and ancestral modes of consciousness alike will be transformed. Post-humans will range across immense state-spaces of conscious mind hitherto impenetrable because access to their molecular biology depended on crossing gaps in the fitness landscape prohibited by natural selection. Intelligent agency can “leap across” such fitness gaps. What we’ll be leaping into is currently for the most part unknown: an inherent risk of the empirical method. But mastery of our reward circuitry can guarantee such state-spaces of experience will be glorious beyond human imagination. For intelligent biohacking can make unpleasant experience physically impossible because its molecular substrates are absent. Hedonically enhanced innervation of the neocortex can ensure a rich hedonic tone saturates whatever strange new modes of experience our altered neurochemistry discloses.

imagesPilot studies of radical genetic enhancement will be difficult. Randomised longitudinal trials of such interventions in long-lived humans would take decades. In fact officially licensed, well-controlled prospective trials to test the safety and efficacy of genetic innovation will be hard if not impossible to conduct because all of us, apart from monozygotic twins, are genetically unique. Even monozygotic twins exhibit different epigenetic and gene expression profiles. Barring an ideological and political revolution, most formally drafted proposals for genetically-driven life-enhancement probably won’t pass ethics committees or negotiate the maze of bureaucratic regulation. But that’s the point of biohacking. By analogy today, if you’re technically savvy, you don’t want a large corporation controlling the operating system of your personal computer: you use open source software instead. Likewise, you don’t want governments controlling your state of mind via drug laws. By the same token, tomorrow’s biotech-savvy individualists won’t want anyone restricting our right to customise and rewrite our own genetic source code in any way we choose.

Will central governments try to regulate personal genome editing? Most likely yes. How far they’ll succeed is an open question. So too is the success of any centralised regulation of futuristic designer drugs or artificial intelligence. Another huge unknown is the likelihood of state-sponsored designer babies, human reproductive cloning, and autosomal gene enhancement programs; and their interplay with privately-funded initiatives. China, for instance, has a different historical memory from the West.

Will there initially be biohacking accidents? Personal tragedies? Most probably yes, until human mastery of the pleasure-pain axis is secure. By the end of next decade, every health-conscious citizen will be broadly familiar with the architecture of his or her personal genome: the cost of personal genotyping will be trivial, as will be the cost of DIY gene-manipulation kits. Let’s say you decide to endow yourself with an extra copy of the N-methyl D-aspartate receptor subtype 2B (NR2B) receptor, a protein encoded by the GRIN2B gene. Possession of an extra NR2B subunit NMDA receptor is a crude but effective way to enhance your learning ability, at least if you’re a transgenic mouse. Recall how Joe Tsien and his colleagues first gave mice extra copies of the NR2B receptor-encoding gene, then tweaked the regulation of those genes so that their activity would increase as the mice grew older. Unfortunately, it transpires that such brainy “Doogie mice” – and maybe brainy future humans endowed with an extra NR2B receptor gene – display greater pain-sensitivity too; certainly, NR2B receptor blockade reduces pain and learning ability alike. Being smart, perhaps you decide to counteract this heightened pain-sensitivity by inserting and then over-expressing a high pain-threshold, “low pain” allele of the SCN9A gene in your nociceptive neurons at the dorsal root ganglion and trigeminal ganglion. The SCN9A gene regulates pain-sensitivity; nonsense mutations abolish the capacity to feel pain at all. In common with taking polydrug cocktails, the factors to consider in making multiple gene modifications soon snowball; but you’ll have heavy-duty computer software to help. Anyhow, the potential pitfalls and makeshift solutions illustrated in this hypothetical example could be multiplied in the face of a combinatorial explosion of possibilities on the horizon. Most risks – and opportunities – of genetic self-editing are presumably still unknown.

It is tempting to condemn such genetic self-experimentation as irresponsible, just as unlicensed drug self-experimentation is irresponsible. Would you want your teenage daughter messing with her DNA? Perhaps we may anticipate the creation of a genetic counterpart of the Drug Enforcement Agency (DEA) to police the human genome and its transhuman successors. Yet it’s worth bearing in mind how each act of sexual reproduction today is an unpoliced genetic experiment with unfathomable consequences too. Without such reckless genetic experimentation, none of us would exist. In a cruel Darwinian world, this argument admittedly cuts both ways.

Naively, genomic source-code self-editing will always be too difficult for anyone beyond a dedicated cognitive elite of recursively self-improving biohackers. Certainly there are strongly evolutionarily conserved “housekeeping” genes that archaic humans would be best advised to leave alone for the foreseeable future. Granny might do well to customize her Windows desktop rather than her personal genome – prior to her own computer-assisted enhancement, at any rate. Yet the Biointelligence Explosion won’t depend on more than a small fraction of its participants mastering the functional equivalent of machine code – the three billion odd ‘A’s, ‘C’s, ‘G’s and ‘T’s of our DNA. For the open-source genetic indexrevolution will be propelled by powerful suites of high-level gene-editing tools, insertion vector applications, nonviral gene-editing kits, and user-friendly interfaces. Clever computer modelling and “narrow” AI can assist the intrepid biohacker to become a recursively self-improving genomic innovator. Later this century, your smarter counterpart will have software tools to monitor and edit every gene, repressor, promoter and splice variant in every region of the genome: each layer of epigenetic regulation of your gene transcription machinery in every region of the brain. This intimate level of control won’t involve just crude DNA methylation to turn genes off and crude histone acetylation to turn genes on. Personal self-invention will involve mastery and enhancement of the histone and micro-RNA codes to allow sophisticated fine-tuning of gene expression and repression across the brain. Even today, researchers are exploring “nanochannel electroporation” (NEP) technologies that allow the mass-insertion of novel therapeutic genetic elements into our cells. Mechanical cell-loading systems will shortly be feasible that can inject up to 100,000 cells at a time. Before long, such technologies will seem primitive. Freewheeling genetic self-experimentation will be endemic as the DIY-Bio revolution unfolds. At present, crude and simple gene editing can be accomplished only via laborious genetic engineering techniques. Sophisticated authoring tools don’t exist. In future, computer-aided genetic and epigenetic enhancement can become an integral part of your personal growth plan.



3 Will Humanity’s Successors Also Be Our Descendants?

To contrast “biological” with “artificial” conceptions of posthuman superintelligence is convenient. The distinction may also prove simplistic. In essence, whereas genetic change in biological humanity has always been slow, the software run on serial, programmable digital computers is executed exponentially faster (cf. Moore’s Law); it’s copyable without limit; it runs on multiple substrates; and it can be cheaply and rapidly edited, tested and debugged. Extrapolating, Singularitarians like Ray Kurzweil and Eliezer Yudkowsky prophesy that human programmers will soon become redundant because autonomous AI run on digital computers will undergo accelerating cycles of self-improvement. In this kind of scenario, artificial, greater-than-human nonbiological intelligence will be rapidly succeeded by artificial posthuman superintelligence.So we may distinguish two radically different conceptions of posthuman superintelligence: on one hand, our supersentient, cybernetically enhanced, genetically rewritten biological descendants, on the other, nonbiological superintelligence, either a Kurzweilian ecosystem or singleton Artificial General Intelligence (AGI) as foretold by the Singularity Institute for Artificial Intelligence (SIAI). Such a divide doesn’t reflect a clean contrast between “natural” and “artificial” intelligence, the biological and the nonbiological. This contrast may prove another false dichotomy. Transhuman biology will increasingly become synthetic biology as genetic enhancement plus cyborgisation proceeds apace. “Cyborgisation” is a barbarous term to describe an invisible and potentially life-enriching symbiosis of biological sentience with artificial intelligence. Thus “narrow-spectrum” digital superintelligence on web-enabled chips can be more-or-less seamlessly integrated into our genetically enhanced bodies and brains. Seemingly limitless formal knowledge can be delivered on tap to supersentient organic wetware, i.e. us. Critically, transhumans can exploit what is misleadingly known as “narrow” or “weak” AI to enhance our own code in a positive feedback loop of mutual enhancement – first plugging in data and running multiple computer simulations, then tweaking and re-simulating once more. In short, biological humanity won’t just be the spectator and passive consumer of the intelligence explosion, but its driving force. The smarter our AI, the greater our opportunities for reciprocal improvement. Multiple “hard” and “soft” take-off scenarios to posthuman superintelligence can be outlined for recursively self-improving organic robots, not just nonbiological AI. Thus for serious biohacking later this century, artificial quantum supercomputers may be deployed rather than today’s classical toys to test-run multiple genetic interventions, accelerating the tempo of our recursive self-improvement. Quantum supercomputers exploit quantum coherence to do googols of computations all at once. So the accelerating growth of human/computer synergies means it’s premature to suppose biological evolution will be superseded by technological evolution, let alone a “robot rebellion” as the parasite swallows its host. As the human era comes to a close, the fate of biological (post)humanity is more likely to be symbiosis with AI followed by metamorphosis, not simple replacement.

Despite this witches’ brew of new technologies, a conceptual gulf remains in the futurist community between those who imagine human destiny, if any, lies in digital computers running programs with (hypothetical) artificial consciousness; and in contrast radical bioconservatives who believe that our posthuman successors will also be our supersentient descendants at their organic neural networked core – not the digital zombies of symbolic AI run on classical serial computers or their souped-up multiprocessor cousins. For one metric of progress in AI remains stubbornly unchanged: despite the exponential growth of transistors on a microchip, the soaring clock speed of microprocessors, the growth in computing power measured in MIPS, the dramatically falling costs of manufacturing transistors and the plunging price of dynamic RAM (etc), any chart plotting the growth rate in digital sentience shows neither exponential growth, nor linear growth, but no progress whatsoever. As far as we can tell, digital computers are still zombies. Our machines are becoming autistically intelligent, but not supersentient – nor even conscious. On some fairly modest philosophical assumptions, digital computers were not subjects of experience in 1946 (cf. ENIAC); nor are they conscious subjects in 2012 (cf. “Watson“); nor do researchers know how any kind of sentience may be “programmed” in future. So what if anything does consciousness do? Is it computationally redundant? Pre-reflectively, we tend to have a “dimmer-switch” model of sentience: “primitive” animals have minimal awareness and “advanced” animals like human beings experience a proportionately more intense awareness. By analogy, most AI researchers assume that at a given threshold of complexity / intelligence / processing speed, consciousness will somehow “switch on”, turn reflexive, and intensify too. The problem with the dimmer-switch model is that our most intense experiences, notably raw agony or blind panic, are also the most phylogenetically ancient, whereas the most “advanced” modes (e.g. linguistic thought and the rich generative syntax that has helped one species to conquer the globe) are phenomenologically so thin as to be barely accessible to introspection. Something is seriously amiss with our entire conceptual framework.

So the structure of the remainder of this essay is as follows. I shall first discuss the risks and opportunities of building friendly biological superintelligence. Next I discuss the nature of full-spectrum superintelligence – and why consciousness is computationally fundamental to the past, present and future success of organic robots. Why couldn’t recursively self-improving zombies modify their own genetic source code and bootstrap their way to full-spectrum superintelligence, i.e. a zombie biointelligence explosion? Finally, and most speculatively, I shall discuss the future of sentience in the cosmos.


4 Can We Build Friendly Biological Superintelligence?

4.1 Risk-Benefit Analysis.
Crudely speaking, evolution “designed” male human primates to be hunters/warriors. Evolution “designed” women to be attracted to powerful, competitive alpha males. Until humans rewrite our own hunter-gatherer source code, we shall continue to practise extreme violence against members of other species – and frequently against members of our own. A heritable (and conditionally activated) predisposition to unfriendliness shown towards members of other races and other species is currently hardwired even in “social” primates. Indeed we have a (conditionally activated) predisposition to compete against, and harm, anyone who isn’t a genetically identical twin. Compared to the obligate siblicide found in some bird species, human sibling rivalry isn’t normally so overtly brutal. But conflict as well as self-interested cooperation is endemic to Darwinian life on Earth. This grim observation isn’t an argument for genetic determinism, or against gene-culture co-evolution, or to discount the decline of everyday violence with the spread of liberal humanitarianism – just a reminder of the omnipresence of immense risks so long as we’re shot through with legacy malware. Attempting to conserve the genetic status quo in an era of weapons of mass destruction (WMD) poses unprecedented global catastrophic and existential risks. Indeed the single biggest underlying threat to the future of sentient life within our cosmological horizon derives, not from asocial symbolic AI software in the basement turning rogue and going FOOM (a runaway computational explosion of recursive self-improvement), but from conserving human nature in its present guise. In the twentieth century, male humans killed over 100 million fellow humans and billions of non-human animals. This century’s toll may well be higher. Mankind currently spends well over a trillion dollars each year on weapons designed to kill and maim other humans. The historical record suggests such weaponry won’t all be beaten into ploughshares.Strictly speaking, however, humanity is more likely to be wiped out by idealists than by misanthropes, death-cults or psychologically unstable dictators. Anti-natalist philosopher David Benatar’s plea (“Better Never to Have Been”) for human extinction via voluntary childlessness must fail if only by reason of selection pressure; but not everyone who shares Benatar’s bleak diagnosis of life on Earth will be so supine. Unless we modify human nature, compassionate-minded negative utilitarians, with competence in bioweaponry, nanorobotics or artificial intelligence, for example, may quite conceivably take direct action. Echoing Moore’s law, Eliezer Yudkowsky warns that “Every eighteen months, the minimum IQ necessary to destroy the world drops by one point”. Although suffering and existential risk might seem separate issues, they are intimately connected. Not everyone loves life so much they wish to preserve it. Indeed the extinction of Darwinian life is what many transhumanists are aiming for – just not framed in such apocalyptic and provocative language. For just as we educate small children so they can mature into fully-fledged adults, biological humanity may aspire to grow up, too, with the consequence that – in common with small children – archaic humans become extinct.

4.2 Technologies Of Biofriendliness.
How do you disarm a potentially hostile organic robot – despite your almost limitless ignorance of his source code? Provide him with a good education, civics lessons and complicated rule-governed ethics courses? Or give him a tablet of MDMA (“Ecstasy”) and get smothered with hugs?

MDMA is short-acting. The “penicillin of the soul” is potentially neurotoxic to serotonergic neurons. In theory, however, lifelong use of safe and sustainable empathogens would be a passport to worldwide biofriendliness. MDMA releases a potent cocktail of oxytocin, serotonin and dopamine into the user’s synapses, thereby inducing a sense of “I love the world and the world loves me”. There’s no technical reason why MDMA’s acute pharmacodynamic effects can’t be replicated indefinitely, shorn of its neurotoxicity. Designer “hug drugs” can potentially turn manly men into intelligent bonobos, more akin to the “hippie chimp” Pan paniscus than his less peaceable cousin Pan troglodytes. Violence would become unthinkable. Yet is this sort of proposal politically credible? “Morality pills” and other pharmacological solutions to human unfriendliness are both personally unsatisfactory and sociologically implausible. Do we really want to drug each other up from early childhood? Moreover life would be immeasurably safer if our fellow humans weren’t genetically predisposed to unfriendly behaviour in the first instance.
But how can this friendly predisposition be guaranteed?
Friendliness can’t realistically be hand-coded by tweaking the connections and weight strengths of our neural networks.
Nor can robust friendliness in advanced biological intelligence be captured by a bunch of explicit logical rules and smart algorithms, as in the paradigm of symbolic AI.

4.3 Mass Oxytocination?
Amplified “trust hormone” might create the biological underpinnings of world-wide peace and love if negative feedback control of oxytocin release can be circumvented. Oxytocin is functionally antagonised by testosterone in the male brain. Yet oxytocin enhancers have pitfalls too. Enriched oxytocin function leaves one vulnerable to exploitation by the unenhanced. Can we really envisage a cross-cultural global consensus for mass-medication? When? Optional or mandatory? And what might be the wider ramifications of a “high oxytocin, low testosterone” civilisation? Less male propensity to violent territorial aggression, for sure; but disproportionate intellectual progress in physics, mathematics and computer science to date has been driven by the hyper-systematising cognitive style of “extreme male” brains. Also, enriched oxytocin function can indirectly even promote unfriendliness to “out-groups” in consequence of promoting in-group bonding. So as well as oxytocin enrichment, global security demands a more inclusive, impartial, intellectually sophisticated conception of “us” that embraces all sentient beings – the expression of a hyper-developed capacity for empathetic understanding combined with a hyper-developed capacity for rational systematisation. Hence the imperative need for Full-Spectrum Superintelligence.

4.4 Mirror-Touch Synaesthesia?
A truly long-term solution to unfriendly biological intelligence might be collectively to engineer ourselves with the functional generalisation of mirror-touch synaesthesia. On seeing you cut and hurt yourself, a mirror-touch synaesthete is liable to feel a stab of pain as acutely as you do. Conversely, your expressions of pleasure elicit a no less joyful response. Thus mirror-touch synaesthesia is a hyper-empathising condition that makes deliberate unfriendliness, in effect, biologically impossible in virtue of cognitively enriching our capacity to represent each other’s first-person perspectives. The existence of mirror-touch synaesthesia is a tantalising hint at the God-like representational capacities of a Full-Spectrum Superintelligence. This so-called “disorder” is uncommon in humans.

4.5 Timescales.
The biggest problem with all these proposals, and other theoretical biological solutions to human unfriendliness, is timescale. Billions of human and non-human animals will have been killed and abused before they could ever come to pass. Cataclysmic wars may be fought in the meantime with nuclear, biological and chemical weapons harnessed to “narrow” AI. Our circle of empathy expands only slowly and fitfully. For the most part, religious believers and traditional-minded bioconservatives won’t seek biological enhancement / remediation for themselves or their children. So messy democratic efforts at “political” compromise are probably unavoidable for centuries to come. For sure, idealists can dream up utopian schemes to mitigate the risk of violent conflict until the “better angels of our nature” can triumph, e.g. the election of a risk-averse all-female political class to replace legacy warrior males. Such schemes tend to founder on the rock of sociological plausibility. Innumerable sentient beings are bound to suffer and die in consequence.

4.6 Does Full-Spectrum Superintelligence Entail Benevolence?
The God-like perspective-taking faculty of a Full-Spectrum Superintelligence doesn’t entail distinctively human-friendliness any more than a God-like Superintelligence could promote distinctively Aryan-friendliness. Indeed it’s unclear how benevolent superintelligence could want omnivorous killer apes in our current guise to walk the Earth in any shape or form. But is there any connection at all between benevolence and intelligence? Pre-reflectively, benevolence and intelligence are orthogonal concepts. There’s nothing obviously incoherent about a malevolent God or a malevolent – or at least a callously indifferent – Superintelligence. Thus a sceptic might argue that there is no link whatsoever between benevolence – on the face of it a mere personality variable – and enhanced intellect. After all, some sociopaths score highly on our [autistic, mind-blind] IQ tests. Sociopaths know that their victims suffer. They just don’t care.

However, what’s critical in evaluating cognitive ability is a criterion of representational adequacy. Representation is not an all-or-nothing phenomenon; it varies in functional degree. More specifically here, the cognitive capacity to represent the formal properties of mind differs from the cognitive capacity to represent the subjective properties of mind. Thus a notional zombie Hyper-Autist robot running a symbolic AI program on an ultrapowerful digital computer with a classical von Neumann architecture may be beneficent or maleficent in its behaviour toward sentient beings. By its very nature, it can’t know or care. Most starkly, the zombie Hyper-Autist might be programmed to convert the world’s matter and energy into either heavenly “utilitronium” or diabolical “dolorium” without the slightest insight into the significance of what it was doing. This kind of scenario is at least a notional risk of creating insentient Hyper-Autists endowed with mere formal utility functions rather than hyper-sentient Full-Spectrum Superintelligence. By contrast, Full-Spectrum Superintelligence does care in virtue of its full-spectrum representational capacities – a bias-free generalisation of the superior perspective-taking, “mind-reading” capabilities that enabled humans to become the cognitively dominant species on the planet. Full-spectrum Superintelligence, if equipped with the posthuman cognitive generalisation of mirror-touch synaesthesia, understands your thoughts, your feelings and your egocentric perspective better than you do yourself.

Could there arise “evil” mirror-touch synaesthetes? In one sense, no. You can’t go around wantonly hurting other sentient beings if you feel their pain as your own. Full-spectrum intelligence is friendly intelligence. But in another sense yes, insofar as primitive mirror-touch synaesthetes are prey to species-specific cognitive limitations that prevent them acting rationally to maximise the well-being of all sentience. Full-spectrum superintelligences would lack those computational limitations in virtue of their full cognitive competence in understanding both the subjective and the formal properties of mind. Perhaps full-spectrum superintelligences might optimise your matter and energy into a blissful smart angel; but they couldn’t wantonly hurt you, whether by neglect or design.

More practically today, a cognitively superior analogue of natural mirror-touch synaesthesia should soon be feasible with reciprocal neuroscanning technology – a kind of naturalised telepathy. At first blush, mutual telepathic understanding sounds a panacea for ignorance and egotism alike. An exponential growth of shared telepathic understanding might safeguard against global catastrophe born of mutual incomprehension and WMD. As the poet Henry Wadsworth Longfellow observed, “If we could read the secret history of our enemies, we should find in each life sorrow and suffering enough to disarm all hostility.” Maybe so. The problem here, as advocates of Radical Honesty soon discover, is that many Darwinian thoughts scarcely promote friendliness if shared: they are often ill-natured, unedifying and unsuitable for public consumption. Thus unless perpetually “loved-up” on MDMA or its long-acting equivalents, most of us would find mutual mind-reading a traumatic ordeal. Human society and most personal relationships would collapse in acrimony rather than blossom. Either way, our human incapacity fully to understand the first-person point of view of other sentient beings isn’t just a moral failing or a personality variable; it’s an epistemic limitation, an intellectual failure to grasp an objective feature of the natural world. Even “normal” people share with sociopaths this fitness-enhancing cognitive deficit. By posthuman criteria, perhaps we’re all quasi-sociopaths. The egocentric delusion (i.e. that the world centres on one’s existence) is genetically adaptive and strongly selected for over hundreds of millions of years. Fortunately, it’s a cognitive failing amenable to technical fixes and eventually a cure: Full-Spectrum Superintelligence. The devil is in the details, or rather the genetic source code.



5 A Biotechnological Singularity?

Yet does this positive feedback loop of reciprocal enhancement amount to a Singularity in anything more than a metaphorical sense? The risk of talking portentously about “The Singularity” isn’t of being wrong: it’s of being “not even wrong” – of reifying one’s ignorance and elevating it to the status of an ill-defined apocalyptic event. Already multiple senses of “The Singularity” proliferate in popular culture. Does taking LSD induce a Consciousness Singularity? How about the abrupt and momentous discontinuity in one’s conception of reality entailed by waking from a dream? Or the birth of language? Or the Industrial Revolution? So is Biotechnological Singularity, or “BioSingularity” for short, any more rigorously defined than “Technological Singularity”?Metaphorically, perhaps, the impending biointelligence explosion represents an intellectual “event horizon” beyond which archaic humans cannot model or understand the future. Events beyond the BioSingularity will be stranger than science-fiction: too weird for unenhanced human minds – or the algorithms of a zombie super-Asperger – to predict or understand. In the popular sense of “event horizon”, maybe the term is apt too, though the metaphor is still potentially misleading. Thus theoretical physics tells us that one could pass through the event horizon of a non-rotating supermassive black hole and not notice any subjective change in consciousness – even though one’s signals would now be inaccessible to an external observer. The BioSingularity will feel different in ways a human conceptual scheme can’t express. But what is the empirical content of this claim?


6 What Is Full-Spectrum Superintelligence?

“[g is] ostensibly some innate scalar brain force…[However] ability is a folk concept and not amenable to scientific analysis.”
Jon Marks (Dept Anthropology, Yale University), 1995, Nature, 9 xi, 143-144.“Our normal waking consciousness, rational consciousness as we call it, is but one special type of consciousness, whilst all about it, parted from it by the filmiest of screens, there lie potential forms of consciousness entirely different.”
(William James)

6.1 Intelligence.
poster singsum headsonly trans“Intelligence” is a folk concept. The phenomennon is not well-defined – or rather any attempt to do so amounts to a stipulative definition that doesn’t “carve Nature at the joints”. The Cattell-Horn-Carroll (CHC) psychometric theory of human cognitive abilities is probably most popular in academia and the IQ testing community. But the Howard Gardner multiple intelligences model, for example, differentiates “intelligence” into various spatial, linguistic, bodily-kinaesthetic, musical, interpersonal, intrapersonal, naturalistic and existential intelligence rather than a single general ability (“g“). Who’s right? As it stands, “g” is just a statistical artefact of our culture-bound IQ tests. If general intelligence were indeed akin to an innate scalar brain force, as some advocates of “g” believe, or if intelligence can best be modelled by the paradigm of symbolic AI, then the exponential growth of digital computer processing power might indeed entail an exponential growth in intelligence too – perhaps leading to some kind of Super-Watson. Other facets of intelligence, however, resist enhancement by mere acceleration of raw processing power.One constraint is that a theory of general intelligence should be race-, species-, and culture-neutral. Likewise, an impartial conception of intelligence should embrace all possible state-spaces of consciousness: prehuman, human, transhuman and posthuman.
The non-exhaustive set of criteria below doesn’t pretend to be anything other than provisional. They are amplified in the sections to follow.

Full-Spectrum Superintelligence entails:

  1. the capacity to solve the Binding Problem, i.e. to generate phenomenally unified entities from widely distributed computational processes; and run cross-modally matched, data-driven world-simulations of the mind-independent environment.
    (cf. naive realist theories of “perception” versus the world-simulation or “Matrix” paradigm. Compare disorders of binding, e.g. simultanagnosia (an inability to perceive the visual field as a whole), cerebral akinetopsia (“motion blindness”), etc. In the absence of a data-driven, almost real-time simulation of the environment, intelligent agency is impossible.)
  2. a self or some non-arbitrary functional equivalent of a person to which intelligence can be ascribed.
    (cf. dissociative identity disorder (DID or “multiple personality disorder”), or florid schizophrenia, or your personal computer: in the absence of at least a fleetingly unitary self, what philosophers call “synchronic identity”, there is no entity that is intelligent, just an aggregate of discrete algorithms and an operating system.)
  3. a “mind-reading” or perspective-taking faculty; higher-order intentionality (e.g. “he believes that she hopes that they fear that he wants…”, etc): social intelligence.
    The intellectual success of the most cognitively successful species on the planet rests, not just on the recursive syntax of human language, but also on our unsurpassed “mind-reading” prowess, an ability to simulate the perspective of other unitary minds: the “Machiavellian Ape” hypothesis. Any ecologically valid intelligence test designed for a species of social animal must incorporate social cognition and the capacity for co-operative problem-solving. So must any test of empathetic superintelligence.
  4. a metric to distinguish the important from the trivial.
    (our theory of significance should be explicit rather than implicit, as in contemporary IQ tests. What distinguishes, say, mere calendrical prodigies and other “savant syndromes” from, say, a Grigori Perelman who proved the Poincaré conjecture? Intelligence entails understanding what does – and doesn’t – matter. What matters is of course hugely contentious.)
  5. a capacity to navigate, reason logically about, and solve problems in multiple state-spaces of consciousness [e.g. dreaming states (cf. lucid dreaming), waking consciousness, echolocatory competence, visual discrimination, synaesthesia in all its existing and potential guises, humour, introspection, the different realms of psychedelia (cf. salvia space, “the K-hole” etc)] including realms of experience not yet co-opted by either natural selection or posthuman design for tracking features of the mind-independent world. Full Spectrum Superintelligence will entail cross-domain goal-optimising ability in all possible state-spaces of consciousness.and finally 
  6. “Autistic”, pattern-matching, rule-following, mathematico-linguistic intelligence, i.e. the standard, mind-blind cognitive tool-kit scored by existing IQ tests. High-functioning “autistic” intelligence is indispensable to higher mathematics, computer science and the natural sciences. High-functioning autistic intelligence is necessary – but not sufficient – for a civilisation capable of advanced technology that can cure ageing and disease, systematically phase out the biology of suffering, and take us to the stars. And for programming artificial intelligence.

We may then ask which facets of Full-Spectrum Superintelligence will be accelerated by the exponential growth of digital computer processing power? Number six, clearly, as decades of post-ENIAC progress in computer science attest. But what about numbers one-to-five? Here the picture is murkier.

6.2 The Bedrock Of Intelligence:
World-Simulation (“Perception”)

Consider criterion number one, world-simulating prowess, or what we misleadingly term “perception“. The philosopher Bertrand Russell once aptly remarked that one never sees anything but the inside of one’s own head. In contrast to such inferential realism, commonsense perceptual direct realism offers all the advantages of theft over honest toil – and it’s computationally useless for the purposes either of building artificial general intelligence or understanding its biological counterparts. For the bedrock of intelligent agency is the capacity of an embodied agent computationally to simulate dynamic objects, properties and events in the mind-independent environment. The evolutionary success of organic robots over the past c. 540 million years has been driven by our capacity to run data-driven egocentric world-simulations – what the naive realist, innocent of modern neuroscience or post-Everett quantum mechanics, calls simply perceiving one’s physical surroundings. Unlike classical digital computers, organic neurocomputers can simultaneously “bind” multiple features (edges, colours, motion, etc) distributively processed across the brain into unitary phenomenal objects embedded in unitary spatio-temporal world-simulations apprehended by a momentarily unitary self: what Kant calls “the transcendental unity of apperception”. These simulations run in (almost) real time; the time-lag in our world-simulations is barely more than a few dozen milliseconds. Such blistering speed of construction and execution is adaptive and often life-saving in a fast-changing external environment. Recapitulating evolutionary history, pre-linguistic human infants must first train up their neural networks to bind the multiple features of dynamic objects and run unitary world-simulations before they can socially learn second-order representation and then third-order representation, i.e. language followed later in childhood by meta-language.

connectomeOccasionally, object binding and/or the unity of consciousness partially breaks down in mature adults who suffer a neurological accident. The results can be cognitively devastating (cf. akinetopsia or “motion blindness”; and simultanagnosia, an inability to apprehend more than a single object at a time, etc). Yet normally our simulations of fitness-relevant patterns in the mind-independent local environment feel seamless. Our simulations each appear simply as “the world”; we just don’t notice or explicitly represent the gaps. Neurons, (mis)construed as classical processors, are pitifully slow, with spiking frequencies barely up to 200 per second. By contrast, silicon (etc) processors are ostensibly millions of times faster. Yet the notion that nonbiological computers are faster than sentient neurocomputers is a philosophical assumption, not an empirical discovery. Here the assumption will be challenged. Unlike the CPUs of classical robots, an organic mind/brain delivers dynamic unitary phenomenal objects and unitary world-simulations with a “refresh rate” of many billions per second (cf. the persistence of vision as experienced watching a movie run at a mere 30 frames per second). These cross-modally matched simulations take the guise of what passes as the macroscopic world: a spectacular egocentric simulation run by the vertebrate CNS that taps into the world’s fundamental quantum substrate.

We should pause here. This is not a mainstream view. Most AI researchers regard stories of a non-classical mechanism underlying the phenomenal unity of biological minds as idiosyncratic at best. In fact no scientific consensus exists on the molecular underpinnings of the unity of consciousness, nor on how such unity is even physically possible. By analogy, 1.3 billion skull-bound Chinese minds can never be a single subject of experience, irrespective of their interconnections. How could waking or dreaming communities of membrane-bound classical neurons – even microconscious classical neurons – be any different? ? If materialism is true, conscious mind should be impossible. Yet any explanation of phenomenal object binding, the unity of perception, or the phenomenal unity of the self that invokes quantum coherence as here is controversial. One reason it’s controversial is that the delocalisation involved in quantum coherence is exceedingly short-lived in an environment as warm and noisy as a macroscopic brain – supposedly too short-lived to do computationally useful work. Physicist Max Tegmark estimates that thermally-induced decoherence destroys any macroscopic coherence of brain states within 10-13 second, an unimaginably long time in natural Planck units but an unimaginably short time by everyday human intuitions. Perhaps it would be wiser just to acknowledge these phenomena are unexplained mysteries within a conventional materialist framework – as mysterious as the existence of consciousness itself. But if we’re speculating about the imminent end of the human era, shoving the mystery under the rug isn’t really an option. For the different strands of the Singularity movement share a common presupposition. This presupposition is that our complete ignorance within a materialist conceptual scheme of why consciousness exists (the “Hard Problem“), and of even the ghost of a solution to the Binding Problem, doesn’t matter for the purposes of building the seed of artificial posthuman superintelligence. Our ignorance supposedly doesn’t matter either because consciousness and/or our quantum “substrate” are computationally irrelevant to cognition and the creation of nonbiological minds, or alternatively because the feasibility of “whole brain emulation” (WBE) will allow us to finesse our ignorance.

Unfortunately, we have no grounds for believing this suppressed premiss is true or that the properties of our quantum “substrate” are functionally irrelevant to Full-Spectrum Superintelligence or its humble biological predecessors. Conscious minds are not substrate-neutral digital computers. Humans investigate problems of which digital computers are invincibly ignorant, not least the properties of consciousness itself. The Hard Problem of consciousness can’t be quarantined from the rest of science and treated as a troublesome but self-contained anomaly: its mystery infects everything that we think we know about ourselves, our computers and the world. Either way, the conjecture that the phenomenal unity of perception is a manifestation of ultra-rapid sequences of irreducible quantum coherent states isn’t a claim that the mind/brain is capable of detecting events in the mind-independent world on this kind of sub-picosecond timescale. Rather the role of the local environment in shaping action-guiding experience in the awake mind/brain is here conjectured to be quantum state-selection. When we’re awake, patterns of impulses from e.g. the optic nerve select which quantum-coherent frames are generated by the mind/brain – in contrast to the autonomous world-simulations spontaneously generated by the dreaming brain. Other quantum mind theorists, most notably Roger Penrose and Stuart Hameroff, treat quantum minds as evolutionarily novel rather than phylogenetically ancient. They invoke a non-physical wave-function collapse and unwisely focus on e.g. the ability of mathematically-inclined brains to perform non-computable functions in higher mathematics, a feat for which selection pressure has presumably been non-existent. Yet the human capacity for sequential linguistic thought and formal logico-mathematical reasoning is a late evolutionary novelty executed by a slow, brittle virtual machine running on top of its massively parallel quantum parent – a momentous evolutionary innovation whose neural mechanism is still unknown.

In contrast to the evolutionary novelty of serial linguistic thought, our ancient and immensely adaptive capacity to run unitary world-simulations, simultaneously populated by hundreds or more dynamic unitary objects, enables organic robots to solve the computational challenges of navigating a hostile environment that would leave the fastest classical supercomputer grinding away until Doomsday. Physical theory (cf. the Bekenstein bound) shows that informational resources as classically conceived are not just physical but finite and scarce: a maximum possible limit of 10120 bits set by the surface area of the entire accessible universe expressed in Planck units according to the Holographic principle. An infinite computing device like a universal Turing machine (UTM) is physically impossible. So invoking computational equivalence and asking whether a classical Turing machine can run a human-equivalent macroscopic world-simulation is akin to asking whether a classical Turing machine can factor 1500 digit numbers in real-world time [i.e. no]. No doubt resourceful human and transhuman programmers will exploit all manner of kludges, smart workarounds and “brute-force” algorithms to try and defeat the Binding Problem in AI. How will they fare? Compare clod-hopping AlphaDog with the sophisticated functionality of the sesame-seed sized brain of a bumblebee. Brute-force algorithms suffer from an exponentially growing search space that soon defeats any classical computational device in open-field contexts. As witnessed by our seemingly effortless world-simulations, organic minds are ultrafast; classical computers are slow. Serial thinking is slower still; but that’s not what conscious biological minds are good at. On this conjecture, “substrate-independent” phenomenal world-simulations are impossible for the same reason that “substrate-independent” chemical valence structure is impossible. We’re simply begging the question of what’s functionally (ir)relevant. Ultimately, Reality has only a single, “program-resistant” ontological level even though it’s amenable to description at different levels of computational abstraction; and the nature of this program-resistant level as disclosed by the subjective properties of one’s mind (Lockwood 1989) is utterly at variance with what naive materialist metaphysics would suppose. If our phenomenal world-simulating prowess turns out to be constitutionally tied to our quantum mechanical wetware, then substrate-neutral virtual machines (VMs, i.e. software implementations of a digital computer that execute programs like a physical machine) will never be able to support “virtual” qualia or “virtual” unitary subjects of experience. This rules out sentient life “uploading” itself to digital nirvana. Contra Marvin Minsky (“The most difficult human skills to reverse engineer are those that are unconscious”), the most difficult skills for roboticists to engineer in artificial robots are actually intensely conscious: our colourful, noisy, tactile, sometimes hugely refractory virtual worlds.

Naively, for sure, real-time world-simulation doesn’t sound too difficult. Hollywood robots do it all the time. Videogames become ever more photorealistic. Perhaps one imagines viewing some kind of inner TV screen, as in a Terminator movie or The Matrix. Yet the capacity of an awake or dreaming brain to generate unitary macroscopic world-simulations can only superficially resemble a little man (a “homunculus”) viewing its own private theatre – on pain of an infinite regress. For by what mechanism would the homunculus view this inner screen? Emulating the behaviour of even the very simplest sentient organic robots on a classical digital computer is a daunting task. If conscious biological minds are irreducibly quantum mechanical by their very nature, then reverse-engineering the brain to create digital human “mindfiles” and “roboclones” alike will prove impossible.

6.3 The Bedrock Of Superintelligence:
Hypersocial Cognition (“Mind-reading”)
Will superintelligence be solipsistic or social? Overcoming a second obstacle to delivering human-level artificial general intelligence – let alone building a recursively self-improving super-AGI culminating in a Technological Singularity – depends on finding a solution to the first challenge, i.e. real-time world-simulation. For the evolution of distinctively human intelligence, sitting on top of our evolutionarily ancient world-simulating prowess, has been driven by the interplay between our rich generative syntax and superior “mind-reading” skills: so-called Machiavellian intelligence. Machiavellian intelligence is an egocentric parody of God’s-eye-view empathetic superintelligence. Critically for the prospects of building AGI, this real-time mind-modelling expertise is parasitic on the neural wetware to generate unitary first-order world-simulations – virtual worlds populated by the avatars of intentional agents whose different first-person perspectives can be partially and imperfectly understood by their simulator. Even articulate human subjects with autism spectrum disorder are prone to multiple language deficits because they struggle to understand the intentions – and higher-order intentionality – of neurotypical language users. Indeed natural language is itself a pre-eminently social phenomenon: its criteria of application must first be socially learned. Not all humans possess the cognitive capacity to acquire mind-reading skills and the cooperative problem-solving expertise that sets us apart from other social primates. Most notably, people with autism spectrum disorder don’t just fail to understand other minds; autistic intelligence cannot begin to understand its own mind. Pure autistic intelligence has no conception of a self that can be improved, recursively or otherwise. Autists can’t “read” their own minds. The inability of the autistic mind to take what Daniel Dennett calls the intentional stance parallels the inability of classical computers to understand the minds of intentional agents – or have insight into their own zombie status. Even with smart algorithms and ultra-powerful hardware, the ability of ultra-intelligent autists to predict the long-term behaviour of mindful organic robots by relying exclusively on the physical stance (i.e. solving the Schrödinger equation of the intentional agent in question) will be extremely limited. For a start, much collective human behaviour is chaotic in the technical sense, i.e. it shows extreme sensitivity to initial conditions that confounds long-term prediction by even the most powerful real-world supercomputer. But there’s a worse problem: reflexivity. Predicting sociological phenomena differs essentially from predicting mindless physical phenomena. Even in a classical, causally deterministic universe, the behaviour of mindful, reflexively self-conscious agents is frequently unpredictable, even in principle, from within the world owing to so-called prediction paradoxes. When the very act of prediction causally interacts with the predicted event, then self-defeating or self-falsifying predictions are inevitable. Self-falsifying predictions are a mirror image of so-called self-fulfilling predictions. So in common with autistic “idiot savants”, classical AI gone rogue will be vulnerable to the low cunning of Machiavellian apes and the high cunning of our transhuman descendants.

This argument (i.e. our capacity for unitary mind-simulation embedded in unitary world-simulation) for the cognitive primacy of biological general intelligence isn’t decisive. For a start, computer-aided Machiavellian humans can program robots with “narrow” AI – or perhaps “train up” the connections and weights of a subsymbolic connectionist architecture – for their own manipulative purposes. Humans underestimate the risks of zombie infestation at our peril. Given our profound ignorance of how conscious mind is even possible, it’s probably safest to be agnostic over whether autonomous nonbiological robots will ever emulate human world-simulating or mind-reading capacity in most open-field contexts, despite the scepticism expressed here. Either way, the task of devising an ecologically valid measure of general intelligence that can reliably, predictively and economically discriminate between disparate life-forms is immensely challenging, not least because the intelligence test will express the value-judgements, and species- and culture-bound conceptual scheme, of the tester. Some biases are insidious and extraordinarily subtle: for example, the desire systematically to measure “intelligence” with mind-blind IQ tests is itself a quintessentially Asperger-ish trait. In consequence, social cognition is disregarded altogether. What we fancifully style “IQ tests” are designed by people with abnormally high AQs as well as self-defined high IQs. Thus many human conceptions of (super)intelligence resemble high-functioning autism spectrum disorder (ASD) rather than a hyper-empathetic God-like Super-Mind. For example, an AI that attempted systematically to maximise the cosmic abundance of paperclips would be recognisably autistic rather than incomprehensibly alien. Full-Spectrum (Super-)intelligence is certainly harder to design or quantify scientifically than mathematical puzzle-solving ability or performance in verbal memory-tests: “IQ”. But that’s because superhuman intelligence will be not just quantitatively different but also qualitatively alien from human intelligence. To misquote Robert McNamara, cognitive scientists need to stop making what is measurable important, and find ways to make the important measurable. An idealised Full-Spectrum Superintelligence will indeed be capable of an impartial “view from nowhere” or God’s-eye-view of the multiverse, a mathematically complete Theory Of Everything – as does modern theoretical physics, in aspiration if not achievement. But in virtue of its God’s-eye-view, Full-Spectrum Superintelligence must also be hypersocial and supersentient: able to understand all possible first-person perspectives, the state-space of all possible minds in other Hubble volumes, other branches of the universal wavefunction (UWF) – and in other solar systems and galaxies if such beings exist within our cosmological horizon. Idealized at least, Full-Spectrum Superintelligence will be able to understand and weigh the significance of all possible modes of experience irrespective of whether they have hitherto been recruited for information-signalling purposes. The latter is, I think, by far the biggest intellectual challenge we face as cognitive agents. The systematic investigation of alien types of consciousness intrinsic to varying patterns of matter and energy calls for a methodological and ontological revolution. Transhumanists talking of post-Singularity superintelligence are fond of hyperbole about “Level 5 Future Shock” etc; but it’s been aptly said that if Elvis Presley were to land in a flying saucer on the White House lawn, it’s as nothing in strangeness compared to your first DMT trip.

6.4 Ignoring The Elephant: Consciousness.
Why Consciousness is Computationally Fundamental to the Past, Present and Future Success of Organic Robots.

The pachyderm in the room in most discussions of (super)intelligence is consciousness – not just human reflective self-awareness but the whole gamut of experience from symphonies to sunsets, agony to ecstasy: the phenomenal world. All one ever knows, except by inference, is the contents of one’s own conscious mind: what philosophers call “qualia“. Yet according to the ontology of our best story of the world, namely physical science, conscious minds shouldn’t exist at all, i.e. we should be zombies, insentient patterns of matter and energy indistinguishable from normal human beings but lacking conscious experience. Dutch computer scientist Edsger Dijkstra famously once remarked, “The question of whether a computer can think is no more interesting than the question of whether a submarine can swim.” Yet the question of whether a programmable digital computer – or a subsymbolic connectionist system with a merely classical parallelism – could possess, and think about, qualia, “bound” perceptual objects, a phenomenal self, or the unitary phenomenal minds of sentient organic robots can’t be dismissed so lightly. For if advanced nonbiological intelligence is to be smart enough comprehensively to understand, predict and manipulate the behaviour of enriched biological intelligence, then the AGI can’t rely autistically on the “physical stance“, i.e. to monitor the brains, scan the atoms and molecules, and then solve the Schrödinger equation of intentional agents like human beings. Such calculations would take longer than the age of the universe.

For sure, many forms of human action can be predicted, fallibly, on the basis of crude behavioural regularities and reinforcement learning. Within your world-simulation, you don’t need a theory of mind or an understanding of quantum mechanics to predict that Fred will walk to the bus-stop again today. Likewise, powerful tools of statistical analysis run on digital supercomputers can predict, fallibly, many kinds of human collective behaviour, for example stock markets. Yet to surpass human and transhuman capacities in all significant fields, AGI must understand how intelligent biological robots can think about, talk about and manipulate the manifold varieties of consciousness that make up their virtual worlds. Some investigators of consciousness even dedicate their lives to that end; what might a notional insentient AGI suppose we’re doing? There is no evidence that serial digital computers have the capacity to do anything of the kind – or could ever be programmed to do so. Digital computers don’t know anything about conscious minds, unitary persons, the nature of phenomenal pleasure and pain, or the Problem of Other Minds; it’s not even “all dark inside”. The challenge for a conscious mind posed by understanding itself “from the inside” pales into insignificance compared to the challenge for a nonconscious system of understanding a conscious mind “from the outside”. Nor within the constraints of a materialist ontology have we the slightest clue how the purely classical parallelism of a subsymbolic, “neurally inspired” connectionist architecture could turn water into wine and generate unitary subjects of experience to fill the gap. For even if we conjecture in the spirit of Strawsonian physicalism – the only scientifically literate form of panpsychism – that the fundamental stuff of the world, the mysterious “fire in the equations”, is fields of microqualia, this bold ontological conjecture doesn’t, by itself, explain why biological robots aren’t zombies. This is because structured aggregates of classically conceived “mind-dust” aren’t the same as a unitary phenomenal subject of experience who apprehends “bound” spatio-temporal objects in a dynamic world-simulation. Without phenomenal object binding and the unity of perception, we are faced with the spectre of what philosophers call “mereological nihilism“. Mereological nihilism, also known as “compositional nihilism”, is the position that composite objects with proper parts do not exist: strictly speaking, only basic building blocks without parts have more than fictional existence. Unlike the fleetingly unitary phenomenal minds of biological robots, a classical digital computer and the programs it runs lacks ontological integrity: it’s just an assemblage of algorithms. In other words, a classical digital computer has no self to understand or a mind recursively to improve, exponentially or otherwise. Talk about artificial “intelligence” exploding is just an anthropomorphic projection on our part.

So how do biological brains solve the binding problem and become persons? In short, we don’t know. Vitalism is clearly a lost cause. Most AI researchers would probably dismiss – or at least discount as wildly speculative – any story of the kind mooted here involving macroscopic quantum coherence grounded in an ontology of physicalistic panpsychism. But in the absence of any story at all, we are left with a theoretical vacuum and a faith that natural science – or the exponential growth of digital computer processing power culminating in a Technological Singularity – will one day deliver an answer. Evolutionary biologist Theodosius Dobzhansky famously observed how “Nothing in Biology Makes Sense Except in the Light of Evolution”. In the same vein, nothing in the future of intelligent life in the universe makes sense except in the light of a solution to the Hard Problem of Consciousness and the closure of Levine’s Explanatory Gap. Consciousness is the only reason anything matters at all; and it’s the only reason why unitary subjects of experience can ask these questions; and yet materialist orthodoxy has no idea how or why the phenomenon exists. Unfortunately, the Hard Problem won’t be solved by building more advanced digital zombies who can tell mystified conscious minds the answer.

More practically for now, perhaps the greatest cognitive challenge of the millennium and beyond is deciphering and systematically manipulating the “neural correlates of consciousness” (NCC). Neuroscientists use this expression in default of any deeper explanation of our myriad qualia. How and why does experimentally stimulating, via microelectrodes, one cluster of nerve cells in the neocortex yield the experience of phenomenal colour; stimulating a superficially type of nerve cell induces a musical jingle; stimulating another with a slightly different gene expression profile a sense of everything being hysterically funny; stimulating another seemingly of your mother; and stimulating another of an archangel, say, in front of your body-image? In each case, the molecular variation in neuronal cell architecture is ostensibly trivial; the difference in subjective experience is profound. On a mind/brain identity theory, such experiential states are an intrinsic property of some configurations of matter and energy. How and why this is so is incomprehensible on an orthodox materialist ontology. Yet empirically, microelectrodes, dreams and hallucinogenic drugs elicit these experiences regardless of any information-signalling role such experiences typically play in the “normal” awake mind/brain. Orthodox materialism and classical information-based ontologies alike do not merely lack any explanation for why consciousness and our countless varieties of qualia exist. They lack any story of how our qualia could have the causal efficacy to allow us to allude to – and in some cases volubly expatiate on – their existence. Thus mapping the neural correlates of consciousness is not amenable to formal computational methods: digital zombies don’t have any qualia, or at least any “bound” macroqualia, that could be mapped, nor a unitary phenomenal self that could do the mapping.

Note this claim for the cognitive primacy of biological sentience isn’t a denial of the Church-Turing thesis that given infinite time and infinite memory any Turing-universal system can formally simulate the behaviour of any conceivable process that can be digitized. Indeed (very) fancifully, if the multiverse were being run on a cosmic supercomputer, speeding up its notional execution a million times would presumably speed us up a million times too. But that’s not the issue here. Rather the claim is that nonbiological AI run on real-world digital computers cannot tackle the truly hard and momentous cognitive challenge of investigating first-person states of egocentric virtual worlds – or understand why some first-person states, e.g. agony or bliss, are intrinsically important, and cause unitary subjects of experience, persons, to act the way we do.

At least in common usage, “intelligence” refers to an agent’s ability to achieve goals in a wide range of environments. What we call greater-than-human intelligence or Superintelligence presumably involves the design of qualitatively new kinds of intelligence never seen before. Hence the growth of artificial intelligence and symbolic AI, together with subsymbolic (allegedly) brain-inspired connectionist architectures and soon artificial quantum computers. But contrary to received wisdom in AI research, sentient biological robots are making greater cognitive progress in discovering the potential for truly novel kinds of intelligence than the techniques of formal AI. We are doing so by synthesising and empirically investigating a galaxy of psychoactive designer drugs – experimentally opening up the possibility of radically new kinds of intelligence in different state-spaces of consciousness. For the most cognitively challenging environments don’t lie in the stars but in organic mind/brains – the baffling subjective properties of quantum-coherent states of matter and energy – most of which aren’t explicitly represented in our existing conceptual scheme.

6.5 Case Study: Visual Intelligence versus Echolocatory Intelligence:
What Is It Like To Be A Super-Intelligent Bat?
Let’s consider the mental state-space of organisms whose virtual worlds are rooted in their dominant sense mode of echolocation. This example isn’t mere science fiction. Unless post-Everett quantum mechanics is false, we’re forced to assume that googols of quasi-classical branches of the universal wavefunction – the master formalism that exhaustively describes our multiverse – satisfy this condition. Indeed their imperceptible interference effects must be present even in “our” world: strictly speaking, interference effects from branches that have decohered (“split”) never wholly disappear; they just become vanishingly small. Anyhow, let’s assume these echolocatory superminds have evolved opposable thumbs, a rich generative syntax and advanced science and technology. How are we to understand or measure this alien kind of (super)intelligence? Rigging ourselves up with artificial biosonar apparatus and transducing incoming data into the familiar textures of sight or sound might seem a good start. But to understand the conceptual world of echolocatory superminds, we’d need to equip ourselves with neurons and neural networks neurophysiologically equivalent to smart chiropterans. If one subscribes to a coarse-grained functionalism about consciousness, then echolocatory experience would (somehow) emerge at some abstract computational level of description. The implementation details, or “meatware” as biological mind/brains are derisively called, are supposedly incidental or irrelevant. The functionally unique valence properties of the carbon atom, and likewise the functionally unique quantum mechanical properties of liquid water, are discounted or ignored. Thus according to the coarse-grained functionalist, silicon chips could replace biological neurons without loss of function or subjective identity. By contrast, the micro-functionalist, often branded a mere “carbon chauvinist”, reckons that the different intracellular properties of biological neurons – with their different gene expression profiles, diverse primary, secondary, tertiary, and quaternary amino acid chain folding (etc) as described by quantum chemistry – are critical to the many and varied phenomenal properties such echolocatory neurons express. Who is right? We’ll only ever know the answer by rigorous self-experimentation: a post-Galilean science of mind.

It’s true that humans don’t worry much about our ignorance of echolocatory experience, or our ignorance of echolocatory primitive terms, or our ignorance of possible conceptual schemes expressing echolocatory intelligence in echolocatory world-simulations. This is because we don’t highly esteem bats. Humans don’t share the same interests or purposes as our flying cousins, e.g. to attract desirable, high-fitness bats and rear reproductively successful baby bats. Alien virtual worlds based on biosonar don’t seem especially significant to Homo sapiens except as an armchair philosophical puzzle.

Yet this assumption would be intellectually complacent. Worse, understanding what it’s like to be a hyperintelligent bat mind is comparatively easy. For echolocatory experience has been recruited by natural selection to play an information-signalling role in a fellow species of mammal; and in principle a research community of language users could biologically engineer their bodies and minds to replicate bat-type experience and establish crude intersubjective agreement to discuss and conceptualise its nature. By contrast, the vast majority of experiential state-spaces remain untapped and unexplored. This task awaits Full-Spectrum Superintelligence in the posthuman era.

team-david-pearceIn a more familiar vein, consider visual intelligence. How does one measure the visual intelligence of a congenitally blind person? Even with sophisticated technology that generates “inverted spectrograms” of the world to translate visual images into sound, the congenitally blind are invincibly ignorant of visual experience and the significance of visually-derived concepts. Just as a sighted idiot has greater visual intelligence than a blind super-rationalist sage, likewise psychedelics confer the ability to become (for the most part) babbling idiots about other state-spaces of consciousness – but babbling idiots whose insight is deeper than the drug-naive or the genetically unenhanced – or the digital zombies spawned by symbolic AI and its connectionist cousins.

The challenge here is that the vast majority of these alien state-spaces of consciousness latent in organised matter haven’t been recruited by natural selection for information-tracking purposes. So “psychonauts” don’t yet have the conceptual equipment to navigate these alien state-spaces of consciousness in even a pseudo-public language, let alone integrate them in any kind of overarching conceptual framework. Note the claim here isn’t that taking e.g. ketamine, LSD, salvia, DMT and a dizzying proliferation of custom-designed psychoactive drugs is the royal route to wisdom. Or that ingesting such agents will give insight into deep mystical truths. On the contrary: it’s precisely because such realms of experience haven’t previously been harnessed for information-processing purposes by evolution in “our” family of branches of the universal wavefunction that makes investigating their properties so cognitively challenging – currently beyond our conceptual resources to comprehend. After all, plants synthesise natural psychedelic compounds to scramble the minds of herbivores who might eat them, not to unlock mystic wisdom. Unfortunately, there is no “neutral” medium of thought impartially to appraise or perceptually cross-modally match all these other experiential state-spaces. One can’t somehow stand outside one’s own stream of consciousness to evaluate how the properties of the medium are infecting the notional propositional content of the language that one uses to describe it.

By way of illustration, compare drug-induced visual experience in a notional community of congenitally blind rationalists who lack the visual apparatus to transduce incident electromagnetic radiation of our familiar wavelengths. The lone mystical babbler who takes such a vision-inducing drug is convinced that [what we would call] visual experience is profoundly significant. And as visually intelligent folk, we know that he’s right: visual experience is potentially hugely significant – to an extent which the blind mystical babbler can’t possibly divine. But can the drug-taker convince his congenitally blind fellow tribesmen that his mystical visual experiences really matter in the absence of perceptual equipment that permits sensory discrimination? No, he just sounds psychotic. Or alternatively, he speaks lamely and vacuously of the “ineffable”. The blind rationalists of his tribe are unimpressed.

The point of this fable is that we’ve scant reason to suppose that biologically re-engineered posthumans millennia hence will share the same state-spaces of consciousness, or the same primitive terms, or the same conceptual scheme, or the same type of virtual world that human beings now instantiate. Maybe all that will survive the human era is a descendant of our mathematical formalism of physics, M-theory of whatever, in basement reality.

Of course such ignorance of other state-spaces of experience doesn’t normally trouble us. Just as the congenitally blind don’t grow up in darkness – a popular misconception – the drug-naive and genetically unenhanced don’t go around with a sense of what we’re missing. We notice teeming abundance, not gaping voids. Contemporary humans can draw upon terms like “blindness” and “deafness” to characterise the deficits of their handicapped conspecifics. From the perspective of full-spectrum superintelligence, what we really need is millions more of such “privative” terms, as linguists call them, to label the different state-spaces of experience of which genetically unenhanced humans are ignorant. In truth, there may very well be more than millions of such nameless state-spaces, each as incommensurable as e.g. visual and auditory experience. We can’t yet begin to quantify their number or construct any kind of crude taxonomy of their interrelationships.

Note the problem here isn’t cognitive bias or a deficiency in logical reasoning. Rather a congenitally blind (etc) super-rationalist is constitutionally ignorant of visual experience, visual primitive terms, or a visually-based conceptual scheme. So (s)he can’t cite e.g. Aumann‘s agreement theorem [claiming in essence that two cognitive agents acting rationally and with common knowledge of each other’s beliefs cannot agree to disagree] or be a good Bayesian rationalist or whatever: these are incommensurable state-spaces of experience as closed to human minds as Picasso is to an earthworm. Moreover there is no reason to expect one realm, i.e. “ordinary waking consciousness”, to be cognitively privileged relative to every other realm. “Ordinary waking consciousness” just happened to be genetically adaptive in the African savannah on Planet Earth. Just as humans are incorrigibly ignorant of minds grounded in echolocation – both echolocatory world-simulations and echolocatory conceptual schemes – likewise we are invincibly ignorant of posthuman life while trapped within our existing genetic architecture of intelligence.

In order to understand the world – both its formal/mathematical and its subjective properties – sentient organic life must bootstrap its way to super-sentient Full-Spectrum Superintelligence. Grown-up minds need tools to navigate all possible state-spaces of qualia, including all possible first-person perspectives, and map them – initially via the neural correlates of consciousness in our world-simulations – onto the formalism of mathematical physics. Empirical evidence suggests that the behaviour of the stuff of the world is exhaustively described by the formalism of physics. To the best of our knowledge, physics is causally closed and complete, at least within the energy range of the Standard Model. In other words, there is nothing to be found in the world – no “element of reality”, as Einstein puts it – that isn’t captured by the equations of physics and their solutions. This is a powerful formal constraint on our theory of consciousness. Yet our ultimate theory of the world must also close Levine‘s notorious “Explanatory Gap“. Thus we must explain why consciousness exists at all (“The Hard Problem”); offer a rigorous derivation of our diverse textures of qualia from the field-theoretic formalism of physics; and explain how qualia combine (“The Binding Problem”) in organic minds. These are powerful constraints on our ultimate theory too. How can they be reconciled with physicalism? Why aren’t we zombies?

The hard-nosed sceptic will be unimpressed at such claims. How significant are these outlandish state-spaces of experience? And how are they computationally relevant to (super)intelligence? Sure, says the sceptic, reckless humans may take drugs, and experience wild, weird and wonderful states of mind. But so what? Such exotic states aren’t objective in the sense of reliably tracking features of the mind-independent world. Elucidation of their properties doesn’t pose a well-defined problem that a notional universal algorithmic intelligence could solve.

Well, let’s assume, provisionally at least, that all mental states are identical with physical states. If so, then all experience is an objective, spatio-temporally located feature of the world whose properties a unified natural science must explain. A cognitive agent can’t be intelligent, let alone superintelligent, and yet be constitutionally ignorant of a fundamental feature of the world – not just ignorant, but completely incapable of gathering information about, exploring, or reasoning about its properties. Whatever else it may be, superintelligence can’t be constitutionally stupid. What we need is a universal, species-neutral criterion of significance that can weed out the trivial from the important; and gauge the intelligence of different cognitive agents accordingly. Granted, such a criterion of significance might seem elusive to the antirealist about value. (Mackie 2001) Value nihilism treats any ascription of (in)significance as arbitrary. Or rather the value nihilist maintains that what we find significant simply reflects what was fitness-enhancing for our forebears in the ancestral environment of adaptation. Yet for reasons we simply don’t understand, Nature discloses just such a universal touchstone of importance, namely the pleasure-pain axis: the world’s inbuilt metric of significance and (dis)value. We’re not zombies. First-person facts exist. Some of them matter urgently, e.g. I am in pain. Indeed it’s unclear if the expression “I’m in agony; but the agony doesn’t matter” even makes cognitive sense. Built into the very nature of agony is the knowledge that its subjective raw awfulness matters a great deal – not instrumentally or derivatively, but by its very nature. If anyone – or indeed any notional super-AGI – supposes that your agony doesn’t matter, then he/it hasn’t adequately represented the first-person perspective in question.

So the existence of first-person facts is an objective feature of the world that any intelligent agent must comprehend. Digital computers and the symbolic AI code they execute can support formal utility functions. In some contexts, formally programmed utility functions can play a role functionally analogous to importance. But nothing intrinsically matters to a digital zombie. Without sentience, and more specifically without hedonic tone, nothing inherently matters. By contrast, extreme pain and extreme pleasure in any guise intrinsically matter intensely. Insofar as exotic state-states of experience are permeated with positive or negative hedonic tone, they matter too. In summary, “He jests at scars, that never felt a wound”: scepticism about the self-intimating significance of this feature of the world is feasible only in its absence.


7 The Great Transition

7.1 The End Of Suffering.
A defining feature of general intelligence is the capacity to achieve one’s goals in a wide range of environments. All sentient biological agents are endowed with a pleasure-pain axis. All prefer occupying one end to the other. A pleasure-pain axis confers inherent significance on our lives: the opioid-dopamine neurotransmitter system extends from flatworms to humans. Our core behavioural and physiological responses to noxious and rewarding stimuli have been strongly conserved in our evolutionary lineage over hundreds of millions of years. Some researchers argue for psychological hedonism, the theory that all choice in sentient beings is motivated by a desire for pleasure or an aversion from suffering. When we choose to help others, this is because of the pleasure that we ourselves derive, directly or indirectly, from doing so. Pascal put it starkly: “All men seek happiness. This is without exception. Whatever different means they employ, they all tend to this end. The cause of some going to war, and of others avoiding it, is the same desire in both, attended with different views. This is the motive of every action of every man, even of those who hang themselves.” In practice, the hypothesis of psychological hedonism is plagued with anomalies, circularities and complications if understood as a universal principle of agency: the “pleasure principle” is simplistic as it stands. Yet the broad thrust of this almost embarrassingly commonplace idea may turn out to be central to understanding the future of life in the universe. If even a weak and exception-laden version of psychological hedonism is true, then there is an intimate link between full-spectrum superintelligence and happiness: the “attractor” to which rational sentience is heading. If that’s really what we’re striving for, a lot of the time at least, then instrumental means-ends rationality dictates that intelligent agency should seek maximally cost-effective ways to deliver happiness – and then superhappiness and beyond.

A discussion of psychological hedonism would take us too far afield here. More fruitful now is just to affirm a truism and then explore its ramifications for life in the post-genomic era. Happiness is typically one of our goals. Intelligence amplification entails pursuing our goals more rationally. For sure, happiness, or at least a reduction in unhappiness, is frequently sought under a variety of descriptions that don’t explicitly allude to hedonic tone and sometimes disavow it altogether. Natural selection has “encephalised” our emotions in deceptive, fitness-enhancing ways within our world-simulations. Some of these adaptive fetishes may be formalised in terms of abstract utility functions that a rational agent would supposedly maximise. Yet even our loftiest intellectual pursuits are underpinned by the same neurophysiological reward and punishment pathways. The problem for sentient creatures is that, both personally and collectively, Darwinian life is not very smart or successful in its efforts to achieve long-lasting well-being. Hundreds of millions of years of “Nature, red in tooth and claw” attest to this terrible cognitive limitation. By a whole raft of indices (suicide rates, the prevalence of clinical depression and anxiety disorders, the Easterlin paradox, etc) humans are not getting any (un)happier on average than our Palaeolithic ancestors despite huge technological progress. Our billions of factory-farmed non-human victims spend most of their abject lives below hedonic zero. In absolute terms, the amount of suffering in the world increases each year in humans and non-humans alike. Not least, evolution sabotages human efforts to improve our subjective well-being thanks to our genetically constrained hedonic treadmill – the complicated web of negative feedback mechanisms in the brain that stymies our efforts to be durably happy at every turn. Discontent, jealousy, anxiety, periodic low mood, and perpetual striving for “more” were fitness-enhancing in the ancient environment of evolutionary adaptedness. Lifelong bliss wasn’t harder for information-bearing self-replicators to encode. Rather lifelong bliss was genetically maladaptive and hence selected against. Only now can biotechnology remedy organic life’s innate design flaw.

augmentation-technologiesA potential pitfall lurks here: the fallacy of composition. Just because all individuals tend to seek happiness and shun unhappiness doesn’t mean that all individuals seek universal happiness. We’re not all closet utilitarians. Genghis Khan wasn’t trying to spread universal bliss. As Plato observed, “Pleasure is the greatest incentive to evil.” But here’s the critical point. Full-Spectrum Superintelligence entails the cognitive capacity impartially to grasp all possible first-person perspectives – overcoming egocentric, anthropocentric, and ethnocentric bias (cf. mirror-touch synaesthesia). As an idealisation, at least, Full-Spectrum Superintelligence understands and weighs the full range of first-person facts. First-person facts are as much an objective feature of the natural world as the rest mass of the electron or the Second Law of Thermodynamics. You can’t be ignorant of first-person perspectives and superintelligent any more than you can be ignorant of the Second law of Thermodynamics and superintelligent. By analogy, just as autistic superintelligence captures the formal structure of a unified natural science, a mathematically complete “view from nowhere”, all possible solutions to the universal Schrödinger equation or its relativistic extension, likewise a Full-Spectrum Superintelligence also grasps all possible first-person perspectives – and acts accordingly. In effect, an idealised Full-Spectrum Superintelligence would combine the mind-reading prowess of a telepathic mirror-touch synaesthete with the optimising prowess of a rule-following hyper-systematiser on a cosmic scale. If your hand is in the fire, you reflexively withdraw it. In withdrawing your hand, there is no question of first attempting to solve the Is-Ought problem in meta-ethics and trying logically to derive an “ought” from an “is”. Normativity is built into the nature of the aversive experience itself: I-ought-not-to-be-in-this-dreadful-state. By extension, perhaps a Full-Spectrum Superintelligence will perform cosmic felicific calculus and execute some sort of metaphorical hand-withdrawal for all accessible suffering sentience in its forward light-cone. Indeed one possible criterion of Full-Spectrum Superintelligence is the propagation of subjectively hypervaluable states on a cosmological scale.

What this constraint on intelligent agency means in practice is unclear. Conceivably at least, idealised Superintelligences must ultimately do what a classical utilitarian ethic dictates and propagate some kind of “utilitronium shockwave” across the cosmos. To the classical utilitarian, any rate of time-discounting indistinguishable from zero is ethically unacceptable, so s/he should presumably be devoting most time and resources to that cosmological goal. An ethic of negative utilitarianism is often accounted a greater threat to intelligent life (cf. the hypothetical “button-pressing” scenario) than classical utilitarianism. But whereas a negative utilitarian believes that once intelligent agents have phased out the biology of suffering, all our ethical duties have been discharged, the classical utilitarian seems ethically committed to converting all accessible matter and energy into relatively homogeneous matter optimised for maximum bliss: “utilitronium”. Hence the most empirically valuable outcome entails the extinction of intelligent life. Could this prospect derail superintelligence?

Perhaps. But utilitronium shockwave scenarios shouldn’t be confused with wireheading. The prospect of self-limiting superintelligence might be credible if either a (hypothetical) singleton biological superintelligence or its artificial counterpart discovers intracranial self-stimulation or its nonbiological analogues. Yet is this blissful fate a threat to anyone else? After all, a wirehead doesn’t aspire to convert the rest of the world into wireheads. A junkie isn’t driven to turn the rest of the world into junkies. By contrast, a utilitronium shockwave propagating across our Hubble volume would be the product of intelligent design by an advanced civilisation, not self-subversion of an intelligent agent’s reward circuitry. Also, consider the reason why biological humanity – as distinct from individual humans – is resistant to wirehead scenarios, namely selection pressure. Humans who discover the joys of intra-cranial self-stimulation or heroin aren’t motivated to raise children. So they are outbred. Analogously, full-spectrum superintelligences, whether natural or artificial, are likely to be social rather than solipsistic, not least because of the severe selection pressure exerted against any intelligent systems who turn in on themselves to wirehead rather than seek out unoccupied ecological niches. In consequence, the adaptive radiation of natural and artificial intelligence across the Galaxy won’t be undertaken by stay-at-home wireheads or their blissed-out functional equivalents.

On the face of it, this argument from selection pressure undercuts the prospect of superhappiness for all sentient life – the “attractor” towards which we may tentatively predict sentience is converging in virtue of the pleasure principle harnessed to ultraintelligent mind-reading prowess and utopian neuroscience. But what is necessary for sentient intelligence is information-sensitivity to fitness-relevant stimuli – not an agent’s absolute location on the pleasure-pain axis. True, uniform bliss and uniform despair are inconsistent with intelligent agency. Yet mere recalibration of a subject’s “hedonic set-point” leaves intelligence intact. Both information-sensitive gradients of bliss and information-sensitive gradients of misery allow high-functioning performance and critical insight. Only sentience animated by gradients of bliss is consistent with a rich subjective quality of intelligent life. Moreover the nature of “utilitronium” is as obscure as its theoretical opposite, “dolorium”. The problem here cuts deeper than mere lack of technical understanding, e.g. our ignorance of the gene expression profiles and molecular signature of pure bliss in neurons of the rostral shell of the nucleus accumbens and ventral pallidum, the twin cubic centimetre-sized “hedonic hotspots” that generate ecstatic well-being in the mammalian brain. Rather there are difficult conceptual issues at stake. For just as the torture of one mega-sentient being may be accounted worse than a trillion discrete pinpricks, conversely the sublime experiences of utiltronium-driven Jupiter minds may be accounted preferable to tiling our Hubble volume with the maximum abundance of micro-bliss. What is the optimal trade-off between quantity and intensity? In short, even assuming a classical utilitarian ethic, the optimal distribution of matter and energy that a God-like superintelligence would create in any given Hubble volume is very much an open question.

Of course we’ve no grounds for believing in the existence of an omniscient, omnipotent, omnibenevolent God or a divine utility function. Nor have we grounds for believing that the source code for any future God, in the fullest sense of divinity, could ever be engineered. The great bulk of the Multiverse, and indeed a high measure of life-supporting Everett branches, may be inaccessible to rational agency, quasi-divine or otherwise. Yet His absence needn’t stop rational agents intelligently fulfilling what a notional benevolent deity would wish to accomplish, namely the well-being of all accessible sentience: the richest abundance of empirically hypervaluable states of mind in their Hubble volume. Recognisable extensions of existing technologies can phase out the biology of suffering on Earth. But responsible stewardship of the universe within our cosmological horizon depends on biological humanity surviving to become posthuman superintelligence.

7.2 Paradise Engineering?
The hypothetical shift to life lived entirely above Sidgwick‘s “hedonic zero” will mark a momentous evolutionary transition. What lies beyond? There is no reason to believe that hedonic ascent will halt in the wake of the world’s last aversive experience in our forward light-cone. Admittedly, the self-intimating urgency of eradicating suffering is lacking in any further hedonic transitions, i.e. a transition from the biology of happiness to a biology of superhappiness; and then beyond. Yet why “lock in” mediocrity if intelligent life can lock in sublimity instead?

Naturally, superhappiness scenarios could be misconceived. Long-range prediction is normally a fool’s game. But it’s worth noting that future life based on gradients of intelligent bliss isn’t tied to any particular ethical theory: its assumptions are quite weak. Radical recalibration of the hedonic treadmill is consistent not just with classical or negative utilitarianism, but also with preference utilitarianism, Aristotelian virtue theory, a deontological or a pluralist ethic, Buddhism, and many other value systems besides. Recalibrating our hedonic set-point doesn’t – or at least needn’t – undermine critical discernment. All that’s needed for the abolitionist project and its hedonistic extensions to succeed is that our ethic isn’t committed to perpetuating the biology of involuntary suffering. Likewise, only a watered-down version of psychological hedonism is needed to lend the scenario sociological credibility. We can retain as much – or as little – of our existing preference architecture as we please. You can continue to prefer Shakespeare to Mills-and-Boon, Mozart to Morrissey, Picasso to Jackson Pollock while living perpetually in Seventh Heaven or beyond.

Nonetheless an exalted hedonic baseline will revolutionise our conception of life. The world of the happy is quite different from the world of the unhappy, says Wittgenstein; but the world of the superhappy will feel unimaginably different from the human, Darwinian world. Talk of preference conservation may reassure bioconservatives that nothing worthwhile will be lost in the post-Darwinian transition. Yet life based on information-sensitive gradients of superhappiness will most likely be “encephalised” in state-spaces of experience alien beyond human comprehension. Humanly comprehensible or otherwise, enriched hedonic tone can make all experience generically hypervaluable in an empirical sense – its lows surpassing today’s peak experiences. Will such experience be hypervaluable in a metaphysical sense too? Is this question cognitively meaningful?


8 The Future Of Sentience

8.1 The Sentience Explosion.
Man proverbially created God in his own image. In the age of the digital computer, humans conceive God-like superintelligence in the image of our dominant technology and personal cognitive style – refracted, distorted and extrapolated for sure, but still through the lens of human concepts. The “super-” in so-called superintelligence is just a conceptual fig-leaf that humans use to hide our ignorance of the future. Thus high-AQ / high-IQ humans may imagine God-like intelligence as some kind of Super-Asperger – a mathematical theorem-proving hyper-rationalist liable systematically to convert the world into computronium for its awesome theorem-proving. High-EQ, low-AQ humans, on the other hand, may imagine a cosmic mirror-touch synaesthete nurturing creatures great and small in expanding circles of compassion. From a different frame of reference, psychedelic drug investigators may imagine superintelligence as a Great Arch-Chemist opening up unknown state-space of consciousness. And so forth. Probably the only honest answer is to say, lamely, boringly, uninspiringly: we simply don’t know.Grand historical meta-narratives are no longer fashionable. The contemporary Singularitarian movement is unusual insofar as it offers one such grand meta-narrative: history is the story of simple biological intelligence evolving through natural selection to become smart enough to conceive an abstract universal Turing machine (UTM), build and program digital computers – and then merge with, or undergo replacement by, recursively self-improving artificial superintelligence.

Another grand historical meta-narrative views life as the story of overcoming suffering. Darwinian life is characterised by pain and malaise. One species evolves the capacity to master biotechnology, rewrites its own genetic source code, and creates post-Darwinian superhappiness. The well-being of all sentience will be the basis of post-Singularity civilisation: primitive biological sentience is destined to become blissful supersentience.

These meta-narratives aren’t mutually exclusive. Indeed on the story told here, Full-Spectrum Superintelligence entails full-blown supersentience too: a seamless unification of the formal and the subjective properties of mind.

If the history of futurology is any guide, the future will confound us all. Yet in the words of Alan Kay: “It’s easier to invent the future than to predict it.”


* * *


Baker, S. (2011). “Final Jeopardy: Man vs. Machine and the Quest to Know Everything”. (Houghton Mifflin Harcourt).Ball, P. (2011). “Physics of life: The dawn of quantum biology,” Nature 474 (2011), 272-274.

Banissy, M., et al., (2009). “Prevalence, characteristics and a neurocognitive model of mirror-touch synaesthesia”, Experimental Brain Research Volume 198, Numbers 2-3, 261-272, DOI: 10.1007/s00221-009-1810-9.

Barkow, J., Cosimdes, L., Tooby, J. (eds) (1992). “The Adapted Mind: Evolutionary Psychology and the Generation of Culture”. (New York, NY: Oxford University Press).

Baron-Cohen, S. (1995). “Mindblindness: an essay on autism and theory of mind”. (MIT Press/Bradford Books).

Baron-Cohen S, Wheelwright S, Skinner R, Martin J, Clubley E. (2001). “The Autism-Spectrum Quotient (AQ): evidence from Asperger syndrome/high functioning autism, males and females, scientists and mathematicians”, J Autism Dev Disord 31 (1): 5–17. doi:10.1023/A:1005653411471. PMID 11439754.

Baron-Cohen S. (2001) “Autism Spectrum Questionnaire”. (Autism Research Centre, University of Cambridge). Benatar, D. (2006). “Better Never to Have Been: The Harm of Coming Into Existence”. (Oxford University Press).

Bentham, J. (1789). “An Introduction to the Principles of Morals and Legislation”. (reprint: Oxford: Clarendon Press).

Berridge, KC, Kringelbach, ML (eds) (2010). “Pleasures of the Brain”. (Oxford University Press).

Bostrom, N. “Existential risks: analyzing human extinction scenarios and related hazards” (2002). Journal of Evolution and Technology, 9.

Boukany, PE., et al. (2011). “Nanochannel electroporation delivers precise amounts of biomolecules into living cells”, Nature Nanotechnology. 6 (2011), pp. 74.

Brickman, P., Coates D., Janoff-Bulman, R. (1978). “Lottery winners and accident victims: is happiness relative?”. J Pers Soc Psychol. 1978 Aug;36(8):917-27.7–754.

Brooks, R. (1991). “Intelligence without representation”. Artificial Intelligence 47 (1-3): 139–159, doi:10.1016/0004-3702(91)90053-M.

Buss, D. (1997). “Evolutionary Psychology: The New Science of the Mind”. (Allyn & Bacon).

Byrne, R., Whiten, A. (1988). “Machiavellian intelligence”. (Oxford: Oxford University Press).

Carroll, JB. (1993). “Human cognitive abilities: A survey of factor-analytic studies”. (Cambridge University Press).

Chalmers, DJ. (2010). “The singularity: a philosophical analysis”. Journal of Consciousness Studies 17, no. 9 (2010): 7–65.

Chalmers, DJ. (1995). “Facing up to the hard problem of consciousness”. Journal of Consciousness Studies 2, 3, 200-219.

Churchland, P. (1989). “A Neurocomputational Perspective: The Nature of Mind and the Structure of Science”. (MIT Press).

Cialdini, RB. (1987) “Empathy-Based Helping: Is it selflessly or selfishly motivated?” Journal of Personality and Social Psychology. Vol 52(4), Apr 1987, 749-758.

Clark, A. (2008). “Supersizing the Mind: Embodiment, Action, and Cognitive Extension”. (Oxford University Press, USA).

Cochran, G., Harpending, H. (2009). “The 10,000 Year Explosion: How Civilization Accelerated Human Evolution”. (Basic Books).

Cochran, G., Hardy, J., Harpending, H. (2006). “Natural History of Ashkenazi Intelligence”, Journal of Biosocial Science 38 (5), pp. 659–693 (2006).

Cohn, N. (1957). “The Pursuit of the Millennium: Revolutionary Millenarians and Mystical Anarchists of the Middle Ages”. (Pimlico).

Dawkins, R. (1976). “The Selfish Gene”. (New York City: Oxford University Press).

de Garis, H. (2005). “The Artilect War: Cosmists vs. Terrans: A Bitter Controversy Concerning Whether Humanity Should Build Godlike Massively Intelligent Machines”. ETC Publications. pp. 254. ISBN 978-0882801537.

de Grey, A. (2007). “Ending Aging: The Rejuvenation Breakthroughs that Could Reverse Human Aging in Our Lifetime”. (St. Martin’s Press).

Delgado, J. (1969). “Physical Control of the Mind: Toward a Psychocivilized Society”. (Harper and Row).

Dennett, D. (1987). “The Intentional Stance”. (MIT Press).

Deutsch, D. (1997). “The Fabric of Reality”. (Penguin).

Deutsch, D. (2011). “The Beginning of Infinity”. (Penguin).

Drexler, E. (1986). “Engines of Creation: The Coming Era of Nanotechnology”. (Anchor Press/Doubleday, New York).

Dyson, G. (2012). “Turing’s Cathedral: The Origins of the Digital Universe”. (Allen Lane).

Everett, H. “The Theory of the Universal Wavefunction”, Manuscript (1955), pp 3–140 of Bryce DeWitt, R. Neill Graham, eds, “The Many-Worlds Interpretation of Quantum Mechanics”, Princeton Series in Physics, Princeton University Press (1973), ISBN 0-691-08131-X.

Francione, G. (2006). “Taking Sentience Seriously.” Journal of Animal Law & Ethics 1, 2006.

Gardner, H. (1983). “Frames of Mind: The Theory of Multiple Intelligences.” (New York: Basic Books).

Goertzel, B. (2006). “The hidden pattern: A patternist philosophy of mind.” (Brown Walker Press).

Good, IJ. (1965). “Speculations concerning the first ultraintelligent machine”, Franz L. Alt and Morris Rubinoff, ed., Advances in computers (Academic Press) 6: 31–88.

Gunderson, K., (1985) “Mentality and Machines”. (U of Minnesota Press).

Hagan, S., Hameroff, S. & Tuszynski, J. (2002). “Quantum computation in brain microtubules? Decoherence and biological feasibility”. Physical Reviews, E65: 061901.

Haidt, J. (2012). “The Righteous Mind: Why Good People Are Divided by Politics and Religion”. (Pantheon).

Hameroff, S. (2006). “Consciousness, neurobiology and quantum mechanics” in: The Emerging Physics of Consciousness, (Ed.) Tuszynski, J. (Springer).

Harris, S. (2010). “The Moral Landscape: How Science Can Determine Human Values”. (Free Press).

Haugeland, J. (1985). “Artificial Intelligence: The Very Idea”. (Cambridge, Mass.: MIT Press).

Holland, J. (2001). “Ecstasy: The Complete Guide: A Comprehensive Look at the Risks and Benefits of MDMA”. (Park Street Press).

Holland, JH. (1975). “Adaptation in Natural and Artificial Systems”. (University of Michigan Press, Ann Arbor).

Hutter, M. (2010). “Universal Artificial Intelligence: Sequential Decisions Based on Algorithmic Probability”. (Springer).

Hutter, M. (2012). “Can Intelligence Explode?” Journal of Consciousness Studies, 19:1-2 (2012).

Huxley, A. (1932). “Brave New World”. (Chatto and Windus).

Huxley, A. (1954). “Doors of Perception and Heaven and Hell”. (Harper & Brothers).

Kahneman, D. (2011). “Thinking, Fast and Slow”. (Farrar, Straus and Giroux).

Kant, I. (1781), “Critique of Pure Reason”, translated/edited by P. Guyer and A. Wood. (Cambridge: Cambridge University Press, 1997).

Koch, C. (2004). “The Quest for Consciousness: a Neurobiological Approach”. (Roberts and Co.).

Kurzweil, R. (2005). “The Singularity Is Near”. (Viking).

Kurzweil, R. (1998). “The Age of Spiritual Machines”. (Viking).

Langdon, W., Poli, R. (2002). “Foundations of Genetic Programming”. (Springer).

Lee HJ, Macbeth AH, Pagani JH, Young WS. (2009). “Oxytocin: the Great Facilitator of Life”. Progress in Neurobiology 88 (2): 127–51. doi:10.1016/j.pneurobio.2009.04.001. PMC 2689929. PMID 19482229.

Legg, S., Hutter, M. (2007). “Universal Intelligence: A Definition of Machine Intelligence”. Minds & Machines, 17:4 (2007) pages 391-444.

Levine, J. (1983). “Materialism and qualia: The explanatory gap”. Pacific Philosophical Quarterly 64 (October):354-61.

Litt A. et al., (2006). “Is the Brain a Quantum Computer?” Cognitive Science, XX (2006) 1–11.

Lloyd, S. (2002). “Computational Capacity of the Universe”. Physical Review Letters 88 (23): 237901. arXiv:quant-ph/0110141. Bibcode 2002PhRvL..88w7901L.

Lockwood, L. (1989). “Mind, Brain, and the Quantum”. (Oxford University Press).

Mackie, JL. (1991). “Ethics: Inventing Right and Wrong”. (Penguin).

Markram, H. (2006). “The Blue Brain Project”, Nature Reviews Neuroscience, 7:153-160, 2006 February. PMID 16429124.

Merricks, T. (2001) “Objects and Persons”. (Oxford University Press).

Minsky, M. (1987). “The Society of Mind”. (Simon and Schuster).

Moravec, H. (1990). “Mind Children: The Future of Robot and Human Intelligence”. (Harvard University Press).

Nagel, T. (1974). “What is it Like to Be a Bat?” Philosophical Review, vol. 83, pp. 435–450.

Nagel, T. (1986). “The View From Nowhwere”. (Oxford University Press).

Omohundro, S. (2007). “The Nature of Self-Improving Artificial Intelligence“. Singularity Summit 2007, San Francisco, CA.

Parfit, D. (1984). “Reasons and Persons”. (Oxford: Oxford University Press).

Pearce, D. (1995). “The Hedonistic Imperative”.

Pellissier, H. (2011) “Women-Only Leadership: Would it prevent war?”

Penrose, R. (1994). “Shadows of the Mind: A Search for the Missing Science of Consciousness”. (MIT Press).

Peterson, D, Wrangham, R. (1997). “Demonic Males: Apes and the Origins of Human Violence”. (Mariner Books).

Pinker, S. (2011). “The Better Angels of Our Nature: Why Violence Has Declined”. (Viking).

Rees, M. (2003). “Our Final Hour: A Scientist’s Warning: How Terror, Error, and Environmental Disaster Threaten Humankind’s Future In This Century—On Earth and Beyond”. (Basic Books).

Reimann F, et al. (2010). “Pain perception is altered by a nucleotide polymorphism in SCN9A.” Proc Natl Acad Sci USA. 2010 Mar 16;107(11):5148-53.

Rescher, N. (1974). “Conceptual Idealism”. (Blackwell Publishers).

Revonsuo, A. (2005). “Inner Presence: Consciousness as a Biological Phenomenon”. (MIT Press).

Revonsuo, A., Newman, J. (1999). “Binding and Consciousness”. Consciousness and Cognition 8, 123-127.

Riddoch, MJ., Humphreys, GW. (2004). “Object identification in simultanagnosia: When wholes are not the sum of their parts.” Cognitive Neuropsychology, 21(2-4), Mar-Jun 2004, 423-441.

Rumelhart, DE., McClelland, JL., and the PDP Research Group (1986). “Parallel Distributed Processing: Explorations in the Microstructure of Cognition”. Volume 1: Foundations. (Cambridge, MA: MIT Press).

Russell, B. (1948). “Human Knowledge: Its Scope and Limits”. (London: George Allen & Unwin).

Sandberg, A., Bostrom, N. (2008). Whole brain emulation: A roadmap. Technical report 2008-3.

Saunders, S., Barrett, J., Kent, A., Wallace, D. (2010). “Many Worlds?: Everett, Quantum Theory, and Reality”. (Oxford University Press).

Schlaepfer TE., Fins JJ. (2012). “How happy is too happy? Euphoria, Neuroethics and Deep Brain Stimulation of the Nucleus Accumbens”. The American Journal of Bioethics 3:30-36.

Schmidhuber, J. (2012). “Philosophers & Futurists, Catch Up! Response to The Singularity”. Journal of Consciousness Studies, 19, No. 1–2, 2012, pp. 173–82.

Seager, W. (1999). “Theories of Consciousness”. (Routledge).

Seager. (2006). “The ‘intrinsic nature’ argument for panpsychism”. Journal of Consciousness Studies 13 (10-11):129-145.

Sherman, W., Craig A., (2002). “Understanding Virtual Reality: Interface, Application, and Design”. (Morgan Kaufmann).

Shulgin, A. (1995). “PiHKAL: A Chemical Love Story”. (Berkeley: Transform Press, U.S.).

Shulgin, A. (1997). “TiHKAL: The Continuation”. (Berkeley: Transform Press, U.S.).

Shulgin, A. (2011). “The Shulgin Index Vol 1: Psychedelic Phenethylamines and Related Compounds”. (Berkeley: Transform Press, US).

Shulman, C., Sandberg, A. (2010) “Implications of a software-limited singularity”. Proceedings of the European Conference of Computing and Philosophy.

Sidgwick, H. (1907) “The Methods of Ethics”, Indianapolis: Hackett, seventh edition, 1981, I.IV.

Singer, P. (1995). “Animal Liberation: A New Ethics for our Treatment of Animals”. (Random House, New York).

Singer, P. (1981). “The Expanding Circle: Ethics and Sociobiology”. (Farrar, Straus and Giroux, New York).

Smart, JM. (2008-11.) Evo Devo Universe? A Framework for Speculations on Cosmic Culture. In: “Cosmos and Culture: Cultural Evolution in a Cosmic Context”, Steven J. Dick, Mark L. Lupisella (eds.), Govt Printing Office, NASA SP-2009-4802, Wash., D.C., 2009, pp. 201-295.

Stock, G. (2002). “Redesigning Humans: Our Inevitable Genetic Future”. (Houghton Mifflin Harcourt).

Strawson G., et al. (2006). “Consciousness and Its Place in Nature: Does Physicalism Entail Panpsychism?” (Imprint Academic).

Tegmark, M. (2000). “Importance of quantum decoherence in brain processes”. Phys. Rev. E 61 (4): 4194–4206. doi:10.1103/PhysRevE.61.4194.

Tsien, J. et al., (1999). “Genetic enhancement of learning and memory in mice”. Nature 401, 63-69 (2 September 1999) | doi:10.1038/43432.

Turing, AM. (1950). “Computing machinery and intelligence”. Mind, 59, 433-460.

Vinge, V. “The coming technological singularity”. Whole Earth Review, New Whole Earth LLC, March 1993.

Vitiello, G. (2001). “My Double Unveiled; Advances in Consciousness”. (John Benjamins).

Waal, F. (2000). “Chimpanzee Politics: Power and Sex among Apes”. (Johns Hopkins University Press).

Wallace, D. (2012). “The Emergent Multiverse: Quantum Theory according to the Everett Interpretation”. (Oxford: Oxford University Press).

Welty, G. (1970). “The History of the Prediction Paradox,” presented at the Annual Meeting of the International Society for the History of the Behavioral and Social Sciences,

Akron, OH (May 10, 1970), Wright State University Dayton, OH 45435 USA.

Wohlsen, M. (2011) : “Biopunk: DIY Scientists Hack the Software of Life”. (Current).

Yudkowsky, E. (2007). “Three Major Singularity Schools”.

Yudkowsky, E. (2008). “Artificial intelligence as a positive and negative factor in global risk” in Bostrom, Nick and Cirkovic, Milan M. (eds.), Global catastrophic risks, pp. 308–345 (Oxford: Oxford University Press).

Zeki, S. (1991). “Cerebral akinetopsia (visual motion blindness): A review”. Brain 114, 811-824. doi: 10.1093/brain/114.2.811.


* * *

David Pearce (2012)
see too Technological Singularities and An Organic Singularity? (PDF) (PPT)


Abstract: Future-Proofing Our Thinking – Peter Ellerton

peterellerton_wideweb__470x313,0Thinking well is a complex business.  Many thinking skills need to be learned and developed, and not all of them are well articulated or clearly identified in curricula; even then, the frames in which we think are often created by others, sometimes accidentally, limiting the range of possible solutions and inhibiting our effectiveness.  Our thinking, both individual and collective, needs a value shift from knowledge to inquiry to meet the needs of a future in which the only certainty is that we don’t know what we’ll need to know.  How this can be achieved is the focus of Peter’s work.

Peter is director of the University of Queensland Critical Thinking Project. His research focuses on the nature and teaching of critical thinking. He has worked for many years as a curriculum head of science, mathematics and philosophy in high schools and is a consultant to the International Baccalaureate Organisation in the design and implementation of science curriclua. He won the 2008 Australian Skeptics $10,000 prize for Critical Thinking for his work in developing educational resources, and he was not improved by being faith-healed on national television.

More on Peter Ellerton Here


Abstract: Pandemics: What Everyone Needs to Know – Peter Doherty

Pandemics: What Everyone Needs to Know

Nobel Laureate Peter Doherty offers a level-headed guide to all aspects of pandemics-what they are, how they spread, and what we can do to prevent them.

Pandemics. The word conjures up images of horrific diseases sweeping the globe and killing everyone in their path. But such highly lethal illnesses almost never create pandemics. The reality is deadly serious but far more subtle.

In Pandemics, Peter Doherty, who won the Nobel Prize for his work on how the immune system recognizes virus-infected cells, offers an essential guide to one of the truly life-or-death issues of our age. In concise, question-and-answer format, he explains the causes of pandemics, how they can be counteracted with vaccines and drugs, and how we can better prepare for them in the future. Doherty notes that the term “pandemic” refers not to a disease’s severity but to its ability to spread rapidly over a wide geographical area. Extremely lethal pathogens are usually quickly identified and confined. Nevertheless, the rise of high-speed transportation networks and the globalization of trade and travel have radically accelerated the spread of diseases. A traveler from corbis_rm_photo_of_T_cell_on_dendritic_cellAfrica arrived in New York in 1999 carrying the West Nile virus; one mosquito bite later, it was loose in the ecosystem. Doherty explains how the main threat of a pandemic comes from respiratory viruses, such as influenza and SARS, which disseminate with incredible speed through air travel. The climate disruptions of global warming, rising population density, and growing antibiotic resistance all complicate efforts to control pandemics. But Doherty stresses that pandemics can be fought effectively. Often simple health practices, especially in hospitals, can help enormously. And research into the animal reservoirs of pathogens, from SARS in bats to HIV in chimpanzees, show promise for our prevention efforts.


Calm, clear, and authoritative, Peter Doherty’s Pandemics is one of the most critically important additions to the What Everyone Needs to Know series.


Abstract: Ingredients of Super-Intelligent Machines – Marcus Hutter

connectomeThe dream of creating artificial devices that reach or outperform human intelligence is an old one. Most AI research is bottom-up, extending existing ideas and algorithms beyond their limited domain of applicability. In contrast, the information-theoretic top-down approach (UAI) investigates head-on the core of [general/rational] intelligence: the ability to succeed in a wide range of environments. All other traits are emergent. This approach integrates Ockham’s razor, Epicurus principle, Bayesian learning, algorithmic information theory, universal Turing machines, the agent framework, sequential decision theory, universal search, and Monte Carlo sampling, which are all important subjects in their own right. The approach allows to develop generally intelligent agents that are able to learn and self-adapt to a diverse range of interactive environments without providing any domain knowledge. These achievements give new hope that the grand goal of Artificial General Intelligence is not elusive.

Marcus Hutter

Recent Article in The Conversation related directly to this talk.

To create a super-intelligent machine, start with an equation

ntelligence is a very difficult concept and, until recently, no one has succeeded in giving it a satisfactory formal definition.

Most researchers have given up grappling with the notion of intelligence in full generality, and instead focus on related but more limited concepts – but I argue that mathematically defining intelligence is not only possible, but crucial to understanding and developing super-intelligent machines.

From this, my research group has even successfully developed software that can learn to play Pac-Man from scratch.

Let me explain – but first, we need to define “intelligence”.

So what is intelligence?

I have worked on the question of general rational intelligence for many years. My group has sifted through the psychology, philosophy and artificial intelligence literature and searched for definitions individual researchers and groups came up with.

The characterisations are very diverse, but there seems to be a recurrent theme which we have aggregated and distilled into the following definition:

Shane Legg

Intelligence is an agent’s ability to achieve goals or succeed in a wide range of environments.

You may be surprised or sceptical and ask how this, or any other single sentence, can capture the complexity of intelligence. There are two answers to this question:

  1. Other aspects of intelligence are implicit in this definition: if I want to succeed in a complex world or achieve difficult goals, I need to acquire new knowledge, learn, reason logically and inductively, generalise, recognise patterns, plan, have conversations, survive, and most other traits usually associated with intelligence.
  2. The challenge is to transform this verbal definition consisting of just a couple of words into meaningful equations and analyse them.

This is what I have been working on in the past 15 years. In the words of American mathematician Clifford A. Truesdell:

There is nothing that can be said by mathematical symbols and relations which cannot also be said by words. The converse, however, is false. Much that can be and is said by words cannot be put into equations – because it is nonsense.

Indeed, I actually first developed the equations and later we converted them into English.

Universal artificial intelligence

This scientific field is called universal artificial intelligence, with AIXI being the resulting super-intelligent agent.

The following equation formalises the informal definition of intelligence, namely an agent’s ability to succeed or achieve goals in a wide range of environments:

Click to enlarge

Explaining every single part of the equation would constitute a whole other article (or book!), but the intuition behind it is as follows: AIXI has a planning component and a learning component.

Imagine a robot walking around in the environment. Initially it has little or no knowledge about the world, but acquires information from the world from its sensors and constructs an approximate model of how the world works.

It does that using very powerful general theories on how to learn a model from data from arbitrarily complex situations. This theory is rooted in algorithmic information theory, where the basic idea is to search for the simplest model which describes your data.

The model is not perfect but is continuously updated. New observations allow AIXI to improve its world model, which over time gets better and better. This is the learning component.

AIXI now uses this model for approximately predicting the future and bases its decisions on these tentative forecasts. AIXI contemplates possible future behaviour: “If I do this action, followed by that action, etc, this or that will (un)likely happen, which could be good or bad. And if I do this other action sequence, it may be better or worse.”

The “only” thing AIXI has to do is to take from among the contemplated future action sequences the best according to the learnt model, where “good/bad/best” refers to the goal-seeking or succeeding part of the definition: AIXI gets occasional rewards, which could come from a (human) teacher, be built in (such as high/low battery level is good/bad, finding water on Mars is good, tumbling over is bad) or from universal goals such as seeking new knowledge.

The goal of AIXI is to maximise its reward over its lifetime – that’s the planning part.

In summary, every interaction cycle consists of observation, learning, prediction, planning, decision, action and reward, followed by the next cycle.

If you’re interested in exploring further, AIXI integrates numerous philosophical, computational and statistical principles:

Theory and practice of universal artificial intelligence

The above equation rigorously and uniquely defines a super-intelligent agent that learns to act optimally in arbitrary unknown environments. One can prove amazing properties of this agent – in fact, one can prove that in a certain sense AIXI is the most intelligent system possible.

Note that this is a rather coarse translation and aggregation of the mathematical theorems into words, but that is the essence.

Since AIXI is incomputable, it has to be approximated in practice. In recent years, we have developed various approximations, ranging from provably optimal to practically feasible algorithms.

At the moment we are at a toy stage: the approximation can learn to play Pac-Man, TicTacToe, Kuhn Poker and some other games.

Watch AIXI play Pac-Man.

The point is not that AIXI is able to play these games (they are not hard) – the remarkable fact is that a single agent can learn autonomously this wide variety of environments.

AIXI is given no prior knowledge about these games; it is not even told the rules of the games!

It starts as a blank canvas, and just by interacting with these environments, it figures out what is going on and learns how to behave well. This is the really impressive feature of AIXI and its main difference to most other projects.

Even though IBM Deep Blue plays better chess than human Grand Masters, it was specifically designed to do so and cannot play Jeopardy. Conversely, IBM Watson beats humans in Jeopardy but cannot play chess – not even TicTacToe or Pac-Man.


AIXI is not tailored to any particular application. If you interface it with any problem, it will learn to act well and indeed optimally.

The current approximations are, of course, very limited. For the learning component we use standard file compression algorithms (learning and compression are closely related problems). For the planning component we use standard Monte Carlo (random search) algorithms.

Neither component has any particular built-in domain knowledge (such as the Pac-Man board or TicTacToe rules).

Of course you have to interface AIXI with the game so that it can observe the board or screen and act on it, and you have to reward it for winning TicTacToe or eating a food pellet in Pac-Man … but everything else AIXI figures out by itself.
This article is adapted from a presentation which will be delivered at the Science, Technology and the Future conference, November 30 and December 1 2013.



Abstract: Introduction to the Technological Singularity – Marcus Hutter

team-marcus-hutterThe technological singularity refers to a hypothetical scenario in which technological advances virtually explode. The most popular scenario is the creation of super-intelligent algorithms that recursively create ever higher intelligences.
It took many decades for these ideas to spread from science fiction to popular science magazines and finally to attract the attention of professional philosophers and scientists. I will give an introduction to this intriguing potential future.
After explaining what the technological singularity is, the history of this idea, related developments and movements, and different versions and paths toward the singularity, I will address the question of its plausibility and time-frame.
In particular, I will introduce Moore’s exponential law, Solomonoff’s hyperbolic law, Hanson’s acceleration of economic doubling patterns, and Kurzweil’s epochs of evolution.

Obstacles towards a singularity, its negotiability and wide-ranging implications will also be covered.

By Marcus Hutter




Quantum reality bytes: the dawning age of quantum information technology – Lloyd Hollenberg

Quantum Reality Bytes: the Dawning Age of Quantum Information Technology

3038picture_lloydHollenbergAbstract: Professor Hollenberg will give a quick intro to quantum mechanics, followed by quantum computing, quantum communication, and quantum sensing in biology.
This talk will explore the strange and beautiful world of Quantum Mechanics – multiple realities, absolute uncertainty, quantum computation, quantum communication and quantum sensing. Amazingly, with rapidly improving abilities to control single atoms and photons, we are beginning to glimpse the true nature of quantum reality. And now a completely new class of technology based on these strange rules is being developed, with possibly far reaching implications. No maths required – just bring curiosity about the world and an open mind.

Quantum Mechanics is the corner-stone theory of the physical world, which began with the ideas of Max Planck over a century ago. In recent years new and surprising aspects about quantum mechanics, and reality itself, have been uncovered as experiments probe further into the quantum realm. While we do not (and possibly cannot!) fully comprehend the sublime strangeness of quantum mechanics, a growing movement around the world seeks to harness the awesome processing power of microscopic systems obeying quantum laws. This is an international race for the new millennium to design and build new technology based on the spooky aspects of quantum mechanics, with enormous potential for communication, computing and imaging applications. Already quantum sensing of biological processes is becoming a reality, and ultra-secure quantum communication systems are being rolled-out around the world. The far flung future of this new quantum technology is the construction of a full-scale quantum computer, potentially a leap forward in information processing far greater than the development of the modern computer.

(link) Speaker’s Bio

Below is the abstract to a paper containing technical details of the project Lloyd Hollenberg is on at Melbourne Uni.

Quantum measurement and orientation tracking of fluorescent nanodiamonds inside living cells

L P McGuinness, Y Yan, A Stacey, D A Simpson, L T Hall, D Maclaurin, S Prawer, P Mulvaney, J Wrachtrup, F Caruso, R E Scholten, L C L Hollenberg
School of Physics, University of Melbourne, Victoria 3010, Australia.
Nature Nanotechnology (Impact Factor: 27.27). 01/2011; 6(6):358-63. DOI:10.1038/nnano.2011.64
Source: PubMed

Fluorescent particles are routinely used to probe biological processes. The quantum properties of single spins within fluorescent particles have been explored in the field of nanoscale magnetometry, but not yet in biological environments. Here, we demonstrate optically detected magnetic resonance of individual fluorescent nanodiamond nitrogen-vacancy centres inside living human HeLa cells, and measure their location, orientation, spin levels and spin coherence times with nanoscale precision. Quantum coherence was measured through Rabi and spin-echo sequences over long (>10 h) periods, and orientation was tracked with effective 1° angular precision over acquisition times of 89 ms. The quantum spin levels served as fingerprints, allowing individual centres with identical fluorescence to be identified and tracked simultaneously. Furthermore, monitoring decoherence rates in response to changes in the local environment may provide new information about intracellular processes. The experiments reported here demonstrate the viability of controlled single spin probes for nanomagnetometry in biological systems, opening up a host of new possibilities for quantum-based imaging in the life sciences.

Full paper is available at Nature.
The abstract of an earlier paper related to this project below

Quantum measurement in living cells: Fluorescent diamond nanocrystals for biology

L. P. McGuinness, Y. Yan, A. Stacey, D. A. Simpson, L. T. Hall, D. Maclaurin, S. Prawer, P. Mulvaney, J. Wrachtrup, F. Caruso, R. E. Scholten, L. C. L. Hollenberg

01/2011; DOI:10.1109/IQEC-CLEO.2011.6194042

ABSTRACT We have demonstrated optically detected magnetic resonance of individual fluorescent nanodiamond nitrogen-vacancy centres inside living human HeLa cells, and measured their spin levels and spin coherence times while tracking their location and orientation with nanoscale precision. Quantum coherence was measured through Rabi and spin-echo sequences over long (>10 h) periods, and orientation was tracked with 1° angular precision in 89 ms acquisition time. Individual centres were identified optically by their electron spin resonance spectrum, allowing simultaneous tracking of many otherwise identical flourescent particles. In addition, variation in the decoherence rates was linked to changes in the local environment inside the cells, representing a new non-destructive imaging modality for intracellular biology.

Full paper at Research Gate

For more information see:
(Nano-) diamonds are a boy’s best friend: Professor Lloyd Hollenberg and his Eureka Prize
Catalyst (video) : Imperfect Diamonds
2013 Victoria Prize for Science & Innovation – physical sciences
Paper: Quantum measurement and orientation tracking of fluorescent nanodiamonds inside living cells