Tuesday, 3 October 2017

The End of Cassini: Why We Must Keep Exploring

An illustration of Cassini’s “Grand Finale” at Saturn (NASA/JPL)

Semper Exploro. Always exploring. The spirit of this motto filled my heart as I watched the planned disintegration of the 20 year old Saturn orbiter Cassini unfold on social media. When the spacecraft finally stopped signalling the homeworld at 1155 UTC on September 15th, I realised that, although the mission was a resounding success, the spirit of Semper Exploro demands that we should send another probe in its place. After all, there is so much left undiscovered at the Saturnian system; what lies under the icy crust of the geyser-spewing moon Enceladus? What do the seas of Titan actually look like? Cassini has left us with even more questions than before it entered orbit in 2004 after a 7 year journey across the solar system. And those questions should be answered not merely because it is scientifically relevant (it is a given), but because the spirit of exploration which Cassini embodied and which Semper Exploro captures so well has the potential to further unite humanity and bring about the best of us in more ways than what any formal ideological framework can do. And we need a better alternative today, now more than ever.

Exploring can be a risky venture, but its a worthy risk. Indeed, an explorer’s death may be the only kind of death worthy of glorification. We stand today only because of a few men and women who risked it all stepping into the unknown in all kinds of fields. With exploration comes advancement. With curiosity comes the gifts of innovation. With ventures comes prosperity. With each expedition into the unknown comes priceless knowledge that uplifts us all as a species. How much more prosperous would we be today had we dedicated all our efforts to kill or dominate one another towards instead settling space, curing illnesses and so forth? The most logical answer would be: many times over, perhaps a thousand fold.

Film works like Star Trek that attempt to portray the future as an advanced utopia are sometimes criticised of being overoptimistic, naive and ignorant of ‘reality on the ground’. But reality is only what we allow it to be. If we want, we could have a Star Trek kind of future right now. If we want, we could have institutions whose only purpose is to explore, discover and advance peaceful prosperity across the stars. Yes, even with our severely flawed humanity we can still have our cake and eat it IF we believe we can. After all, this flawed humanity is the same humanity that has created the better present we see today. And experience can serve as a catalyst; people today are more motivated than ever to improve not just their own circumstances, but the circumstances of everyone else around them, simply because of inspired hope. The momentum created by this hopeful belief means that, for the most part, we have no where else to go but up.

We cannot hope to have a smooth ride to the future. But fantastic endeavours like the Cassini mission can help remind and solidify our global desire as a species: to see what’s on the other side of the distant horizon. To learn and grow wiser. And to do it with everyone around us.

If we can just keep on exploring, perhaps one day we might discover an even better version of ourselves than we could ever have imagined. But we must keep exploring. Semper exploro forever Cassini.

Saturday, 27 May 2017

On Designing Relevant AI Products for Humans

Many attempts at creating effective user experiences with artificial intelligence (AI) tend to produce news worthy results but not so much in terms of upping mainstream adoption.

Routine AI use for most consumers is still not an everyday thing (broadly speaking anyway, it may not be so if you look at certain demographic groups and interest groups such as tech enthusiasts). No one’s as dependent on AI assistants in the same way as we are so dependent on, say, Word processors or Google Search or mobile devices.

Intelligent assistants are still a novelty, and that may stem from the problem of their design. Most people don’t want to converse with machines on philosophical matters (though it ought to be cool), they just need to make a reference. People don’t want virtual girls in holographic displays asking them how their day went after coming home from work. And people seem to naturally want to troll chatbots in an attempt to explore the limits of their “intelligence”, and perhaps have a laugh while at it.

In fact, the more we attempt to define AI use in the context of human specific activities, the less it is used. It is perhaps a result of our bias towards emulating the all-knowing, all-powerful and very personal/witty AI found in films like Her or in sci-fi games like Halo. This kind of cultural hinting naturally leads many people to believe that, because the AI is being presented in a very human-like fashion, then it should act like a human. The fact of the matter is that, in almost all cases, you’re probably going to see through the facade before 10mins of continuous use have elapsed.
AI simply isn’t human, and designers shouldn’t help us pretend that they are.
Instead of trying to approach the question of AI usage when designing interfaces purely from the perspective of our sci-fi fantasies (who knows the future anyway?), we could perhaps start with the user; what do they actually need? They probably might not need an AI with a built-in personality.
They might instead need an AI that can do small but useful stuff efficiently and reliably. Successful intelligent agents that do just that usually lack any form of personality or similar bells and whistles. 

Take Google’s strategy; Google’s assistant and intelligent products haven’t a hint of personality besides the voice used for their voice search functionality. This cuts out a lot of cultural and psychological baggage from the conversation. Since the product clearly appears as a machine, people will not lead themselves to believe that their assistants can do human-like stuff (sparing the occasional witty answers and trivia). This can be beneficial because the user isn’t distracted by the product’s attempts at sounding human. The user is instead empowered without him/her even realising it. It just works.

I used to think building an assistant as cold and impersonal as Google Now would be a bad move, but I can now see the logic when comparing Google Now to the competition. Cortana looks so lively, and yet she disappoints me precisely because you keep looking for the human-likeness as it all but disappears with continued usage. Ditto Siri. I get so distracted by their apparent witty personality that I can’t seem to get at their actual functionality. Why do I need a witty robot in the house?

By far, the most interesting thing that Google has ever added to it’s Google Now app was no less impersonal, but it was extremely useful IMO. Now on Tap is able to recognise elements displayed at anytime on a phone, allowing a search to be done without leaving any given opened app. If it doesn’t detect what you want, you can simply highlight it and it would search for that. It is a perfect design; minimalistic, useful and brutally impersonal. It just works.

Intelligent apps shouldn’t be built to seem explicitly human; they should be built to get some actual work done. And they should do it without distracting. I know this might not be the kind of interesting personal robot some of us would have imagined, but it is probably the only way to make things less awkward between man and machine. The uncanny valley is not a place that you’d like your app to end up in, certainly not the rest of the consumer AI market and industry.

AI is still an emerging technology, like the Internet before it went mainstream. We are still trying to understand how to integrate the technology into the normal, everyday workflow of average human beings. Human-AI UX design choices may prove to be what makes or breaks consumer AI applications. The potential rewards for successfully breaking the design problem are tantalising; designers must keep seeking the perfect problem-solution fits that real people might care about and design around that.

Tuesday, 16 May 2017

Extending Our Intelligence: The Coming Era of Binary Sapience

Our disembodied extensions of our brains are set to become smarter and more closely linked with us

Our phones and other computing devices are more or less passive creations, bound at every step to man's programming. Now, however, they are becoming more and more responsive to us. More anticipatory towards our various needs, subtle tastes and wants; they are becoming more intelligent.
What does it mean for humanity exactly? How far can we take this?

If an intelligent algorithm can read and classify your current mood, and map your different mood swings over the course of a day through different data inputs with a higher accuracy than your spouse, could it be used to subtly use that data to make your mood better? Is such a state of affairs desirable, assuming it is possible given current technology. And what of that data anyway?

A broad personal artificial intelligence (AI) suite, perhaps more powerful and capable than anything our current best could one day be set lose on every individual on Earth to map out what makes them tick exactly (the data profile on each individual could be worth quite a lot of money to say the least), and then actively engage with them to achieve a desired individual state for each person as set by themselves (or others?). Such a machine could become the sort of ancillary intelligence we see in many works of science fiction; a true intelligence that will allow each person to understand themselves COMPLETELY, and then proceed in assisting us to make ourselves better than we can possibly ever be on our own, or even with the help of HUMAN professionals.

An individual human will cease being a single intelligence, but will instead become a binary intelligence, with the personal machine significantly expanding our cognition and identity at an almost subconscious level, while itself advancing itself via software upgrades and learning.

We can imagine a remote future where a newly born child is assigned such a personal intelligence the moment he or she is born. The intelligence will observe his development and consult the corpus of our knowledge to extrapolate the future possibilities of this child's personality and behaviour, to better understand how to manage them as he/she grows up.

Meanwhile, the child's parent's own intelligent aides can be synchronised with their child's, to enhance not just their parenting experience, but to enhance the way the literally think about their role as parents. We can keep going; we can even imagine training AI to school children (a personal tutor from the future). Given the knowledge these AIs will have access to, it is reasonable to assume that they will probably evolve to become the best tutors, or at the very least, tutor-student aids possible.

Why should any of this be desirable? Why would anyone want to create machines that not only replace humans in a variety of high level skilled tasks, but also become a very part of our very being. And we haven't gone into the ethics of the data used produced in the meantime. But make no mistake, we're heading in that general direction very fast. And the reason is painfully obvious.

While human beings are versatile in terms of initiative and creativity, we don't scale very well in the realm of perfect recall, infinite patience, logic, reasoning, et cetera. Obviously we will need tools to give ourselves these superhuman abilities. And of course, sometimes we might need something to light our way forwards, like when a writer has writer's block and needs a dose of much needed inspiration from his personal intelligence aid. The more these tools actively interact with us, the better we'll become at handling not just the modern world, but ourselves as well.

However, we shouldn't kid ourselves. Advancement comes with risks. The Wanna Cry worm that rocked the world over the weekend demonstrates not the fragility of our systems, but the fragility of our ability to get a grip on our systems. Our minds are the weak point (which the hacker knows and exploits), and at the same time our strongest point. We evolve through experience, and in this century, the sum total of our experiential knowledge is more accessible and more voluminous in scope and depth than ever before. It would be folly to not add on an extra layer of thinking cortex, our exocortex in the form of actively engaged, thinking machines to ourselves in order to make efficient use of our species' knowledge vault.