How can we help you today?

Fill in the form below so we can explore ways to reach your goals or call us at 1800 577 346.

1 / 2
x
How can we help you?
One last step

Leave your details below and we'll be in touch.

Confirmation
2 / 2
x
Previous
Next step
Thanks! We have received your form submission, I'll get back to you shortly!
Oops! Something went wrong while submitting the form

Why We Should Fear and Hail A.I.

Why We Should Fear and Hail A.I.
What's new: K-Startup Grand Challenge 2020 for Australian/New Zealand Startups! More information here.

I recently had the pleasure of speaking with Neil Sahota for an episode of the Future Squared podcast.

Sahota is an IBM Master Inventor, United Nations A.I. subject matter expert, professor at UC Irvine.

He’s a founding member of the UN’s Artificial Intelligence for Social Good Committee, and he co-wrote Own the A.I. Revolution, which provides a future-forward look at A.I., focusing on how businesses can use it to commercialize while doing good in the world.

We explored the depths of the esoteric, covering questions like:

  • Can A.I. become conscious?
  • If humans merge with A.I., at what point do we stop being human?
  • How do we get around embedding our own cogitive biases into A.I.?

…and so much more.

Trending AI Articles:

1. Introducing Ozlo

2. Basics of Neural Network

3. Bursting the Jargon bubbles — Deep Learning

4. How Can We Improve the Quality of Our Data?

You can listen to the entire conversation below.

Episode #362: Own the AI Revolution with Neil Sahota

Neil Sahota is an IBM Master Inventor, United Nations A.I. subject matter expert, professor at UC Irvine, and…

www.futuresquared.xyz

As I like to do, to both reiterate my own learning, and share lessons learned with the world, find below key take-aways from my conversation with Neil Sahota.

The Definition of True A.I.

So many companies peddle what they purport to be A.I., including IBM, but for a product to truly earn this definition, it needs to meet the following three criteria.

1 — it learns from experience and consumptions

2 — understands natural language

3 — interacts like a human being

IBM’s Watson not A.I.

Based on these criteria, IBM’s Watson is not actually A.I., despite its success at Jeopardy all of those years ago!

As Berkeley’s John Searle noted, Watson manipulates symbols but doesn’t understand the meaning behind the symbols as a human would.

AGI: artificial general intelligence (think human cognition — we are not there yet)

ANI: artificial narrow intelligence (think A.I. that does just one thing really well — this is where we are at today. eg. Google Translate)

It’s difficult to determine how far away AGI. It could be next month or could be 50 years away.

Economic incentives don’t support AGI

Commercial incentives don’t support development of AGI.

There are much better rewards to be reaped in the short-term based on investing in narrow A.I., for which there are numerous use cases, and for which the cost of investment and unknowns are much lower. This makes the ROI higher on narrow A.I., and delays the development of AGI.

Decision-Making Challenges

AI helps us make better decisions, but it has three key challenges:

  1. It does so based on the data we feed it. So it’s case of rubbish in, rubbish out, bias in, bias out, racism in, racism out (see below)…
  2. It needs to control for factors we can’t see or reliably attribute (e.g. luck, to counter this, it can process millions of records of data to identify hidden but influential patterns in data)
  3. It needs to adapt its decision-making rules to account for changing external circumstances, much like Darwinism itself (this is called updating its ‘ground truths’)

Influencing Our World View

People might have different experiences of A.I., as we do with Google search.

This means we might develop different world views as a consequence.

As Sahota put it, “the truth may change but the facts remain the same”.

The Paperclip Maximizer and Perverse Instantiation

Who thought that this little ClipArt guy could be the death of us?!

The paperclip maximizer is a thought experiment showing how AGI, even one designed competently and without malice, could ultimately destroy humanity. An extremely powerful A.I. could seek goals that are completely alien to ours, and as a side-effect destroy us by consuming resources essential to our survival.

This tendency to optimize for a particular outcome, at the expense of ethics, morality or reason, is known as ‘perverse instantiation’.

Sahota says that the paperclip maximizer is real, insofar as it is a possibility. “It could actually happen”. In order to counter this, we need to set constraints to avoid adverse outcomes. With so many potential constraints to account for, this represents challenges.

The Future of Work

A.I. may, in a single generation, produce more technological breakthroughs than humankind has managed during the first 20,000 years of its existence.

47% of US jobs are likely to be automated by 2050.

The goal of A.I. is to free people up for higher-value tasks.

“Jobs will go away, but new jobs will be created”, says Sahota.

On the other side of the spectrum, many fear that there won’t be enough work to go around, which is why universal basic income (UBI) has become such a big talking point as of late.

Since 1980, the gap between US productivity and labor compensation has gotten larger, thanks to technology doing more of what humans once did.

This gap is set to get larger and with each disruptive innovation, it can take decades for organizations and societies to reorganize around it, during which time we might experience a downturn in both productivity and potentially, average living standards. This was true of the transition from steam to electricity, and it is known in economic circles as a ‘productivity paradox’.

Socrates, meet Watson

Given the pushback against big-tech that we’re seeing, thanks partially to the attention merchant economy and the Big Brother nature of companies like Facebook, it’s critical that we embed philosophy and the arts into the design of technology.

The challenge then becomes, which philosophy? Since Socrates, there have been debates between rival schools of philosophy, and ideas in general. But the need to reconcile the gaps between being able to commercialize technology, and being able to commercialize technology that does good for humanity, is painfully evident each time you walk onto a Subway train. I’m talking about the army of people mindlessly staring at their screens.

Workflow Podcast

The WorkFlow podcast is hosted by Steve Glaveski with a mission to help you unlock your potential to do more great work in far less time, whether you're working as part of a team or flying solo, and to set you up for a richer life.

No items found.
FREE EBOOK

100 DOS AND DON'TS FOR CORPORATE INNOVATION

To help you avoid stepping into these all too common pitfalls, we’ve reflected on our five years as an organization working on corporate innovation programs across the globe, and have prepared 100 DOs and DON’Ts.

No items found.
No items found.

STEP INTO THE METAVERSE

Unlock new opportunities and markets by taking your brand into the brave new world.

Thanks for your submission. We will be in touch shortly!
Oops! Something went wrong while submitting the form.

Steve Glaveski

Steve Glaveski is the co-founder of Collective Campus, author of Time Rich, Employee to Entrepreneur and host of the Future Squared podcast. He’s a chronic autodidact, and he’s into everything from 80s metal and high-intensity workouts to attempting to surf and do standup comedy.

Ask me a question!