Difference between revisions of "Welcome to the Robustly Beneficial Wiki"

From RB Wiki
 
Line 1: Line 1:
<strong>MediaWiki has been installed.</strong>
+
Welcome to the Robustly Beneficial wiki!!
  
Consult the [https://www.mediawiki.org/wiki/Special:MyLanguage/Help:Contents User's Guide] for information on using the wiki software.
+
== What is this wiki about? And why? ==
  
 +
This wiki aims to list references, ideas and research questions in AI ethics. We hope to help better visualize the scope and the limits of current AI ethics.
 +
 +
The wiki is managed by the Lausanne Alignment Club, based in EPFL, Switzerland. Please check also our Robustly Beneficial YouTube channel, our Robustly Beneficial Podcast and our Robustly Beneficial Talks.
 +
 +
The wiki has just been launched, so most pages are still being written. But they will never be finished — this is the whole point of a wiki!
 +
 +
== New here? How about starting with these pages? ==
 +
 +
The pages can be roughly divided into 4 categories.
 +
 +
=== Why AI ethics is becoming critical ===
 +
 +
If you are new to AI ethics, you should probably start with the [[AI risks]] page. You could then go into arguably today's most important case of AI ethics, namely [[YouTube]]. Note that algorithms also offer formidable [[AI opportunities]] that are definitely worth considering. Find out more by reading about [[online polarization]], [[misinformation]], [[addiction]], [[mental health]] or [[hate]].
 +
 +
And if you know little about the current state of algorithmic research, you might want to check the latest [[impressive advances in AI]]. Or you could check some [[funny applications of AI]]. You can also read my [[rant against semantic debates]] and my [https://www.lesswrong.com/posts/bwqDrSZvhEDKxRf6z/a-rant-against-robots rant against robots].
 +
 +
=== How today's (and probably tomorrow's) AIs work ===
 +
 +
The most important principle of today's AI is surely [[machine learning]]. Today, it mostly relies on [[stochastic gradient descent]] for (deep) [[neural networks]], which allow [[representational learning]] (see [[convolutional neural network]], [[residual network]], [[LSTM]]). See also [[Turing 1950]], [[convexity]], [[generative adversarial network]], [[transformer]] and [[linear systems]].
 +
 +
[[Bayesianism]] has been argued to be the ideal form of supervised and unsupervised learning, if we had infinite computational power (see [[Solomonoff's demon]]). It has numerous desirable properties, like [[statistical admissibility]], [[Bayesian agreement]] or the [[Bayesian brain]] hypothesis. See also [[Bayesian examination]] and [[conjugate priors]].
 +
 +
A branch of learning called [[reinforcement learning]], which relies on [[Q-learning]] or [[policy learning]], seems likely to become the core framework of today's and tomorrow's AIs. [[AIXI]] achieves the upper-bound for [[Legg-Hutter intelligence]], which aims to measure [[artificial general intelligence|general intelligence]].
 +
 +
To understand the gap between Bayesianism/AIXI and practical machine learning, we need to understand the constraints of computational [[complexity]] theory. By building upon the [[Church-Turing thesis]] and knowledge from [[human brain computations]], this allows some insights into [[human-level AI]], in addition to [[experts' AI predictions]]. See also [[entropy]] and [[sophistication]].
 +
 +
AIs are already doing [[distributed learning]], which raises numerous challenges, like [[Byzantine fault tolerance]] and [[model drift]].
 +
 +
=== Why AI safety and ethics is harder than meets the eye ===
 +
 +
We want to get algorithms to do what we would really want them to do. But this turns out to raise numerous highly nontrivial problems, like [[Goodhart's law]], [[overfitting]], [[robust statistics]], [[confounding variables]], [[adversarial attacks]], [[algorithmic bias]], [[cognitive bias]], [[backfire effect]], [[distributional shift]], [[privacy]], [[interpretability]], [[reward hacking]], [[wireheading]] and [[instrumental convergence]]. Because of all such problems, it seems crucial that AIs be able to reason about their ignorance, using [[Bayesianism|Bayesian]] principles, [[moral uncertainty]] and [[second opinion querying]].
 +
 +
AI ethics also demands that we solve thorny philosophical dilemmas, like the [[repugnant conclusion]], [[Newcomb's paradox]] and [[moral realism]]. Unfortunately, we have numerous [[cognitive bias|cognitive biases]], which seem critical to understand to solve AI ethics. Results about [[counterfactual]], [[von Neumann-Morgenstern theorem]] and [[Dutch book]] also seem useful to consider.
 +
 +
=== How to solve AI ethics (hopefully) ===
 +
 +
To solve AI ethics, [http://ceur-ws.org/Vol-2301/paper_1.pdf Hoang][https://dblp.org/rec/bibtex/conf/aaai/Hoang19 19a] proposed the [[ABCDE roadmap]], which decomposes the [[alignment]] problem into numerous (hopefully) orthogonal and complementary subproblems. Such subproblems include [[data certification]], perhaps through [[Blockchain]], [[world model inference]] through [[Bayesianism]] and/or [[representational learning]], [[volition]] learning and [[social choice]] solutions, [[corrigibility]] and safe [[reinforcement learning]].
 +
 +
If you are interested in doing AI ethics research, please check out [[AI ethics research questions]].
 +
 +
== About the authors ==
 +
 +
Lê holds a [https://publications.polymtl.ca/1668/ PhD in applied mathematics] from the École Polytechnique of Montreal. He did postdoctoral research at MIT, before joining EPFL as a science communicator. He runs EPFL YouTube channels [https://www.youtube.com/user/TheWandida Wandida] and [https://www.youtube.com/channel/UCfY6ovyFMaw30NRs-KrxrWw/ ZettaBytes], and French-speaking YouTube channels and podcasts [https://www.youtube.com/channel/UC0NCbj8CxzeCGIF6sODJ-7A Science4All], [https://www.youtube.com/channel/UCNHFiyWgsnaSOsMtSoV_Q1A/ Axiome] and [https://www.youtube.com/playlist?list=PLtzmb84AoqRQ0ikLb4yC4lKgjeDEIpE1i Probablement]. In 2020, he (will have) published <em>The Equation of Knowledge: From Bayes Rule to a Unified Philosophy of Science</em> at CRC Press (French version is out [https://www.amazon.fr/Formule-du-Savoir-Hoang-Nguyen/dp/2759822605/ Hoang18<sup>FR</sup>]). Lê is still partly involved in research project. But his current main focus is synthesizing and organizing the current effort in AI ethics.
 +
 +
In 2020, Lê and Mahdi co-wrote the book <em>The Fabulous Endeavor: Make Artificial Intelligence Robustly Beneficial</em> [https://laboutique.edpsciences.fr/produit/1107/9782759824304/Le%20fabuleux%20chantier HoangElmhamdi][https://scholar.google.ch/scholar?hl=en&as_sdt=0%2C5&q=Le+fabuleux+chantier%3A+Rendre+l%27intelligence+artificielle+robustement+b%C3%A9n%C3%A9fique&btnG= 19<sup>FR</sup>] (the English version is pending).
 +
 +
 +
<!--
 
== Getting started ==
 
== Getting started ==
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Configuration_settings Configuration settings list]
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Configuration_settings Configuration settings list]
Line 9: Line 57:
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Localisation#Translation_resources Localise MediaWiki for your language]
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Localisation#Translation_resources Localise MediaWiki for your language]
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Combating_spam Learn how to combat spam on your wiki]
 
* [https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Combating_spam Learn how to combat spam on your wiki]
 +
-->

Revision as of 21:57, 20 January 2020

Welcome to the Robustly Beneficial wiki!!

What is this wiki about? And why?

This wiki aims to list references, ideas and research questions in AI ethics. We hope to help better visualize the scope and the limits of current AI ethics.

The wiki is managed by the Lausanne Alignment Club, based in EPFL, Switzerland. Please check also our Robustly Beneficial YouTube channel, our Robustly Beneficial Podcast and our Robustly Beneficial Talks.

The wiki has just been launched, so most pages are still being written. But they will never be finished — this is the whole point of a wiki!

New here? How about starting with these pages?

The pages can be roughly divided into 4 categories.

Why AI ethics is becoming critical

If you are new to AI ethics, you should probably start with the AI risks page. You could then go into arguably today's most important case of AI ethics, namely YouTube. Note that algorithms also offer formidable AI opportunities that are definitely worth considering. Find out more by reading about online polarization, misinformation, addiction, mental health or hate.

And if you know little about the current state of algorithmic research, you might want to check the latest impressive advances in AI. Or you could check some funny applications of AI. You can also read my rant against semantic debates and my rant against robots.

How today's (and probably tomorrow's) AIs work

The most important principle of today's AI is surely machine learning. Today, it mostly relies on stochastic gradient descent for (deep) neural networks, which allow representational learning (see convolutional neural network, residual network, LSTM). See also Turing 1950, convexity, generative adversarial network, transformer and linear systems.

Bayesianism has been argued to be the ideal form of supervised and unsupervised learning, if we had infinite computational power (see Solomonoff's demon). It has numerous desirable properties, like statistical admissibility, Bayesian agreement or the Bayesian brain hypothesis. See also Bayesian examination and conjugate priors.

A branch of learning called reinforcement learning, which relies on Q-learning or policy learning, seems likely to become the core framework of today's and tomorrow's AIs. AIXI achieves the upper-bound for Legg-Hutter intelligence, which aims to measure general intelligence.

To understand the gap between Bayesianism/AIXI and practical machine learning, we need to understand the constraints of computational complexity theory. By building upon the Church-Turing thesis and knowledge from human brain computations, this allows some insights into human-level AI, in addition to experts' AI predictions. See also entropy and sophistication.

AIs are already doing distributed learning, which raises numerous challenges, like Byzantine fault tolerance and model drift.

Why AI safety and ethics is harder than meets the eye

We want to get algorithms to do what we would really want them to do. But this turns out to raise numerous highly nontrivial problems, like Goodhart's law, overfitting, robust statistics, confounding variables, adversarial attacks, algorithmic bias, cognitive bias, backfire effect, distributional shift, privacy, interpretability, reward hacking, wireheading and instrumental convergence. Because of all such problems, it seems crucial that AIs be able to reason about their ignorance, using Bayesian principles, moral uncertainty and second opinion querying.

AI ethics also demands that we solve thorny philosophical dilemmas, like the repugnant conclusion, Newcomb's paradox and moral realism. Unfortunately, we have numerous cognitive biases, which seem critical to understand to solve AI ethics. Results about counterfactual, von Neumann-Morgenstern theorem and Dutch book also seem useful to consider.

How to solve AI ethics (hopefully)

To solve AI ethics, Hoang19a proposed the ABCDE roadmap, which decomposes the alignment problem into numerous (hopefully) orthogonal and complementary subproblems. Such subproblems include data certification, perhaps through Blockchain, world model inference through Bayesianism and/or representational learning, volition learning and social choice solutions, corrigibility and safe reinforcement learning.

If you are interested in doing AI ethics research, please check out AI ethics research questions.

About the authors

Lê holds a PhD in applied mathematics from the École Polytechnique of Montreal. He did postdoctoral research at MIT, before joining EPFL as a science communicator. He runs EPFL YouTube channels Wandida and ZettaBytes, and French-speaking YouTube channels and podcasts Science4All, Axiome and Probablement. In 2020, he (will have) published The Equation of Knowledge: From Bayes Rule to a Unified Philosophy of Science at CRC Press (French version is out Hoang18FR). Lê is still partly involved in research project. But his current main focus is synthesizing and organizing the current effort in AI ethics.

In 2020, Lê and Mahdi co-wrote the book The Fabulous Endeavor: Make Artificial Intelligence Robustly Beneficial HoangElmhamdi19FR (the English version is pending).