<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://robustlybeneficial.org/wiki/index.php?action=history&amp;feed=atom&amp;title=Preference_learning_from_comparisons</id>
	<title>Preference learning from comparisons - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://robustlybeneficial.org/wiki/index.php?action=history&amp;feed=atom&amp;title=Preference_learning_from_comparisons"/>
	<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;action=history"/>
	<updated>2026-04-28T13:28:20Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.34.0</generator>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=219&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang at 22:34, 22 February 2020</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=219&amp;oldid=prev"/>
		<updated>2020-02-22T22:34:04Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 22:34, 22 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l1&quot; &gt;Line 1:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 1:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;It has been argued that we humans are much more effective at comparing alternatives than at scoring them [https://infoscience.epfl.ch/record/255399/files/EPFL_TH8637.pdf MaystrePhD][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Efficient+Learning+from+Comparisons+maystre&amp;amp;btnG= 18]. Besides, implicit observations of humans mostly provide such choice data, in the form of like/no-like, share/no-share, click/no-click, and so on. Therefore, preference learning from comparisons seems to be an important approach to preference learning.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;It has been argued that we humans are much more effective at comparing alternatives than at scoring them [https://infoscience.epfl.ch/record/255399/files/EPFL_TH8637.pdf MaystrePhD][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Efficient+Learning+from+Comparisons+maystre&amp;amp;btnG= 18&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;] [https://www.youtube.com/watch?v=bmD-myeu19Q RB5&lt;/ins&gt;]. Besides, implicit observations of humans mostly provide such choice data, in the form of like/no-like, share/no-share, click/no-click, and so on. Therefore, preference learning from comparisons seems to be an important approach to preference learning.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Classical models ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Classical models ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=198&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Gaussian process */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=198&amp;oldid=prev"/>
		<updated>2020-02-04T14:21:56Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Gaussian process&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 14:21, 4 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l29&quot; &gt;Line 29:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 29:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Perhaps one of the most promising avenues to scalable comparison-based preference learning consists of building upon the assumption that, a priori, the intrinsic preference &amp;lt;math&amp;gt;\theta_1-\theta_2&amp;lt;/math&amp;gt; between two alternatives &amp;lt;math&amp;gt;z_1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2&amp;lt;/math&amp;gt; is a Gaussian process. By then invoking the similarity between having to choose between &amp;lt;math&amp;gt;x = (z_1,z_2)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x' = (z_1',z_2')&amp;lt;/math&amp;gt;, we can then generalize from observed data by means of a [[kernel method]].&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Perhaps one of the most promising avenues to scalable comparison-based preference learning consists of building upon the assumption that, a priori, the intrinsic preference &amp;lt;math&amp;gt;\theta_1-\theta_2&amp;lt;/math&amp;gt; between two alternatives &amp;lt;math&amp;gt;z_1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2&amp;lt;/math&amp;gt; is a Gaussian process. By then invoking the similarity between having to choose between &amp;lt;math&amp;gt;x = (z_1,z_2)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x' = (z_1',z_2')&amp;lt;/math&amp;gt;, we can then generalize from observed data by means of a [[kernel method]].&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;To implement this approach, we first need a measure of the similarity between two choices &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt;. Intuitively, &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; will be similar if both &amp;lt;math&amp;gt;z_1 \approx z_1'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_2'&amp;lt;/math&amp;gt;. Note that &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; can also be &amp;quot;anti-similar&amp;quot;, if &amp;lt;math&amp;gt;z_1 \approx z_2'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_1'&amp;lt;/math&amp;gt;. Let us call &amp;lt;math&amp;gt;K(x,x')&amp;lt;/math&amp;gt; &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;this property&lt;/del&gt;.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;To implement this approach, we first need a measure of the similarity between two choices &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt;. Intuitively, &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; will be similar if both &amp;lt;math&amp;gt;z_1 \approx z_1'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_2'&amp;lt;/math&amp;gt;. Note that &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; can also be &amp;quot;anti-similar&amp;quot;, if &amp;lt;math&amp;gt;z_1 \approx z_2'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_1'&amp;lt;/math&amp;gt;. Let us call &amp;lt;math&amp;gt;K(x,x')&amp;lt;/math&amp;gt; &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;the similarity between &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;/math&amp;gt;x'&amp;lt;/math&amp;gt;. Gaussian process prior and Bayesian methods then enable to infer a posterior from revealed preferences on the preference for other dilemmas [https://dl.acm.org/doi/pdf/10.1145/1102351.1102369?download=true ChuGhahramani][https://dblp.org/rec/bibtex/conf/icml/ChuG05 05a] [http://www.jmlr.org/papers/volume6/chu05a/chu05a.pdf ChuGharamani][https://dblp.org/rec/bibtex/journals/jmlr/ChuG05 05b]&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that such kernels can be approximately implemented by [[representational learning|vector representation]], for instance by some deep neural network &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. We could then have &amp;lt;math&amp;gt;K(x,x') = f(x)^Tf(x')&amp;lt;/math&amp;gt; (anti-similarity could be naturally enforced if &amp;lt;math&amp;gt;f(z_1,z_2) = - f(z_2,z_1)&amp;lt;/math&amp;gt;, which is guaranteed if &amp;lt;math&amp;gt;f(z_1,z_2) = g(z_1)-g(z_2)&amp;lt;/math&amp;gt; for some other vector representation &amp;lt;math&amp;gt;g&amp;lt;/math&amp;gt;). The use of such vector representations could be useful to accelerate computation, as opposed to the kernel method which requires going through all the training data to make a prediction.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that such kernels can be approximately implemented by [[representational learning|vector representation]], for instance by some deep neural network &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. We could then have &amp;lt;math&amp;gt;K(x,x') = f(x)^Tf(x')&amp;lt;/math&amp;gt; (anti-similarity could be naturally enforced if &amp;lt;math&amp;gt;f(z_1,z_2) = - f(z_2,z_1)&amp;lt;/math&amp;gt;, which is guaranteed if &amp;lt;math&amp;gt;f(z_1,z_2) = g(z_1)-g(z_2)&amp;lt;/math&amp;gt; for some other vector representation &amp;lt;math&amp;gt;g&amp;lt;/math&amp;gt;). The use of such vector representations could be useful to accelerate computation, as opposed to the kernel method which requires going through all the training data to make a prediction.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;NOTE: This section needs references.&lt;/del&gt;&lt;/div&gt;&lt;/td&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Connection to sports ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Connection to sports ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Chess [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+Rating+Of+ChessPlayers+past+and+present+Players+elo+arco&amp;amp;btnG= Elo78], football [http://ceur-ws.org/Vol-1842/paper_04.pdf MKFG][https://dblp.org/rec/bibtex/conf/pkdd/MaystreKFG16 16].&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Chess [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+Rating+Of+ChessPlayers+past+and+present+Players+elo+arco&amp;amp;btnG= Elo78], football [http://ceur-ws.org/Vol-1842/paper_04.pdf MKFG][https://dblp.org/rec/bibtex/conf/pkdd/MaystreKFG16 16].&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=197&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Classical models */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=197&amp;oldid=prev"/>
		<updated>2020-02-04T13:02:51Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Classical models&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 13:02, 4 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l3&quot; &gt;Line 3:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 3:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Classical models ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Classical models ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The classical models for preference learning are due to [http://mlab.no/blog/wp-content/uploads/2009/07/thurstone94law.pdf Thurston][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=A+law+of+comparative+judgment+thurstone&amp;amp;btnG= 27], [https://idp.springer.com/authorize/casa?redirect_uri=https://link.springer.com/content/pdf/10.1007/BF01180541.pdf&amp;amp;casa_token=dMSQZQpkjt4AAAAA:4HFh3rzi3S6VqH29Pg00S64qMuFB9b6VpF8pRikHmTNWf3REi10i_dj9_Eps3Ivp6l8c-ZETTU-Cx_-L6g Zermelo][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Die+Berechnung+der+Turnier-Ergebnisse+als+ein+Maximumproblem+der+Wahrscheinlichkeitsrechnung.&amp;amp;btnG= 29], [https://&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;apps&lt;/del&gt;.&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;dtic&lt;/del&gt;.&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;mil&lt;/del&gt;/&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;dtic&lt;/del&gt;/&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;tr&lt;/del&gt;/&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;fulltext/u2&lt;/del&gt;/&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;a417190&lt;/del&gt;.pdf&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;#page=15 DavidBook&lt;/del&gt;][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;The&lt;/del&gt;+&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;Method&lt;/del&gt;+of+&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;Paired&lt;/del&gt;+&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;Comparisons+david&lt;/del&gt;&amp;amp;btnG= &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;63&lt;/del&gt;], [https://www.jstor.org/stable/pdf/2334029.pdf?casa_token=Q0ROYjxwOZEAAAAA:fuuqbCntxaNmZB0hlKk8iXfzOkLzQs9H677gh2UOQdMI4496B3YZHlUaiuMgD9zONaHYLHgtSmjMHziL29cx9d9tv2f_QwyqN0wxYUJrExI_eK8QR7sa BradleyTerry][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Rank+analysis+of+incomplete+block+designs%3A+I.+The+method+of+paired+comparisons&amp;amp;btnG= 52], [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Individual+Choice+Behavior%3A+A+Theoretical+Analysis+luce+wiley+1959&amp;amp;btnG= LuceBook59].&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The classical models for preference learning are due to [http://mlab.no/blog/wp-content/uploads/2009/07/thurstone94law.pdf Thurston][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=A+law+of+comparative+judgment+thurstone&amp;amp;btnG= 27], [https://idp.springer.com/authorize/casa?redirect_uri=https://link.springer.com/content/pdf/10.1007/BF01180541.pdf&amp;amp;casa_token=dMSQZQpkjt4AAAAA:4HFh3rzi3S6VqH29Pg00S64qMuFB9b6VpF8pRikHmTNWf3REi10i_dj9_Eps3Ivp6l8c-ZETTU-Cx_-L6g Zermelo][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Die+Berechnung+der+Turnier-Ergebnisse+als+ein+Maximumproblem+der+Wahrscheinlichkeitsrechnung.&amp;amp;btnG= 29], [https://&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;link&lt;/ins&gt;.&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;springer&lt;/ins&gt;.&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;com&lt;/ins&gt;/&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;content&lt;/ins&gt;/&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;pdf&lt;/ins&gt;/&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;10.1007&lt;/ins&gt;/&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;BF02289115&lt;/ins&gt;.pdf &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;Mosteller&lt;/ins&gt;][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;mosteller+Remarks+on+the&lt;/ins&gt;+&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;method&lt;/ins&gt;+of+&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;paired&lt;/ins&gt;+&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;comparisons&lt;/ins&gt;&amp;amp;btnG= &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;51&lt;/ins&gt;], [https://www.jstor.org/stable/pdf/2334029.pdf?casa_token=Q0ROYjxwOZEAAAAA:fuuqbCntxaNmZB0hlKk8iXfzOkLzQs9H677gh2UOQdMI4496B3YZHlUaiuMgD9zONaHYLHgtSmjMHziL29cx9d9tv2f_QwyqN0wxYUJrExI_eK8QR7sa BradleyTerry][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Rank+analysis+of+incomplete+block+designs%3A+I.+The+method+of+paired+comparisons&amp;amp;btnG= 52], [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=Individual+Choice+Behavior%3A+A+Theoretical+Analysis+luce+wiley+1959&amp;amp;btnG= LuceBook59&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;], [https://apps.dtic.mil/dtic/tr/fulltext/u2/a417190.pdf#page=15 DavidBook][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+Method+of+Paired+Comparisons+david&amp;amp;btnG= 63&lt;/ins&gt;].&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;In these models, in a choice between 1 and 2, the human implicitly the scores &amp;lt;math&amp;gt;\theta_1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\theta_2&amp;lt;/math&amp;gt; that she assigns to each alternative. But his computation of &amp;lt;math&amp;gt;\theta_1 - \theta_2&amp;lt;/math&amp;gt; is noisy, and is accompanied with some noise &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt;, to yield &amp;lt;math&amp;gt;x_{12} = \theta_1-\theta_2+\varepsilon&amp;lt;/math&amp;gt;. In the above models, the sign of &amp;lt;math&amp;gt;x_{12}&amp;lt;/math&amp;gt; then determines the choice of the human.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;In these models, in a choice between 1 and 2, the human implicitly the scores &amp;lt;math&amp;gt;\theta_1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\theta_2&amp;lt;/math&amp;gt; that she assigns to each alternative. But his computation of &amp;lt;math&amp;gt;\theta_1 - \theta_2&amp;lt;/math&amp;gt; is noisy, and is accompanied with some noise &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt;, to yield &amp;lt;math&amp;gt;x_{12} = \theta_1-\theta_2+\varepsilon&amp;lt;/math&amp;gt;. In the above models, the sign of &amp;lt;math&amp;gt;x_{12}&amp;lt;/math&amp;gt; then determines the choice of the human.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=183&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Gaussian process */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=183&amp;oldid=prev"/>
		<updated>2020-02-03T08:25:09Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Gaussian process&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 08:25, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l31&quot; &gt;Line 31:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 31:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;To implement this approach, we first need a measure of the similarity between two choices &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt;. Intuitively, &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; will be similar if both &amp;lt;math&amp;gt;z_1 \approx z_1'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_2'&amp;lt;/math&amp;gt;. Note that &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; can also be &amp;quot;anti-similar&amp;quot;, if &amp;lt;math&amp;gt;z_1 \approx z_2'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_1'&amp;lt;/math&amp;gt;. Let us call &amp;lt;math&amp;gt;K(x,x')&amp;lt;/math&amp;gt; this property.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;To implement this approach, we first need a measure of the similarity between two choices &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt;. Intuitively, &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; will be similar if both &amp;lt;math&amp;gt;z_1 \approx z_1'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_2'&amp;lt;/math&amp;gt;. Note that &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; can also be &amp;quot;anti-similar&amp;quot;, if &amp;lt;math&amp;gt;z_1 \approx z_2'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_1'&amp;lt;/math&amp;gt;. Let us call &amp;lt;math&amp;gt;K(x,x')&amp;lt;/math&amp;gt; this property.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that such kernels can be approximately implemented by [[representational learning|vector representation]], for instance by some deep neural network &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. We could then have &amp;lt;math&amp;gt;K(x,x') = f(x)^Tf(x')&amp;lt;/math&amp;gt; (anti-similarity could be naturally enforced if &amp;lt;math&amp;gt;f(z_1,z_2) = - f(z_2,z_1)&amp;lt;/math&amp;gt;). The use of such vector representations could be useful to accelerate computation, as opposed to the kernel method which requires going through all the training data to make a prediction.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that such kernels can be approximately implemented by [[representational learning|vector representation]], for instance by some deep neural network &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. We could then have &amp;lt;math&amp;gt;K(x,x') = f(x)^Tf(x')&amp;lt;/math&amp;gt; (anti-similarity could be naturally enforced if &amp;lt;math&amp;gt;f(z_1,z_2) = - f(z_2,z_1)&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;&amp;lt;/math&amp;gt;, which is guaranteed if &amp;lt;math&amp;gt;f(z_1,z_2) = g(z_1)-g(z_2)&amp;lt;/math&amp;gt; for some other vector representation &amp;lt;math&amp;gt;g&lt;/ins&gt;&amp;lt;/math&amp;gt;). The use of such vector representations could be useful to accelerate computation, as opposed to the kernel method which requires going through all the training data to make a prediction.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;NOTE: This section needs references.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;NOTE: This section needs references.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=182&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Gaussian process */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=182&amp;oldid=prev"/>
		<updated>2020-02-03T08:03:55Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Gaussian process&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 08:03, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l26&quot; &gt;Line 26:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 26:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Perhaps one of the most promising avenues to scalable comparison-based preference learning consists of building upon the assumption that, a priori, the intrinsic preference &amp;lt;math&amp;gt;\theta_1-\theta_2&amp;lt;/math&amp;gt; between two alternatives &amp;lt;math&amp;gt;z_1&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2&amp;lt;/math&amp;gt; is a Gaussian process. By then invoking the similarity between having to choose between &amp;lt;math&amp;gt;x = (z_1,z_2)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x' = (z_1',z_2')&amp;lt;/math&amp;gt;, we can then generalize from observed data by means of a [[kernel method]].&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;To implement this approach, we first need a measure of the similarity between two choices &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt;. Intuitively, &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; will be similar if both &amp;lt;math&amp;gt;z_1 \approx z_1'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_2'&amp;lt;/math&amp;gt;. Note that &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;x'&amp;lt;/math&amp;gt; can also be &amp;quot;anti-similar&amp;quot;, if &amp;lt;math&amp;gt;z_1 \approx z_2'&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_2 \approx z_1'&amp;lt;/math&amp;gt;. Let us call &amp;lt;math&amp;gt;K(x,x')&amp;lt;/math&amp;gt; this property.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Note that such kernels can be approximately implemented by [[representational learning|vector representation]], for instance by some deep neural network &amp;lt;math&amp;gt;f&amp;lt;/math&amp;gt;. We could then have &amp;lt;math&amp;gt;K(x,x') = f(x)^Tf(x')&amp;lt;/math&amp;gt; (anti-similarity could be naturally enforced if &amp;lt;math&amp;gt;f(z_1,z_2) = - f(z_2,z_1)&amp;lt;/math&amp;gt;). The use of such vector representations could be useful to accelerate computation, as opposed to the kernel method which requires going through all the training data to make a prediction.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;NOTE: This section needs references.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Connection to sports ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Connection to sports ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Chess [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+Rating+Of+ChessPlayers+past+and+present+Players+elo+arco&amp;amp;btnG= Elo78], football [http://ceur-ws.org/Vol-1842/paper_04.pdf MKFG][https://dblp.org/rec/bibtex/conf/pkdd/MaystreKFG16 16].&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Chess [https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+Rating+Of+ChessPlayers+past+and+present+Players+elo+arco&amp;amp;btnG= Elo78], football [http://ceur-ws.org/Vol-1842/paper_04.pdf MKFG][https://dblp.org/rec/bibtex/conf/pkdd/MaystreKFG16 16].&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=181&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Classical models */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=181&amp;oldid=prev"/>
		<updated>2020-02-03T07:46:11Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Classical models&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 07:46, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l12&quot; &gt;Line 12:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 12:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The Bradley-Terry model assumes that &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt; follows a [https://en.wikipedia.org/wiki/Gumbel_distribution Gumbel distribution]. Equivalently, it sets &amp;lt;math&amp;gt;\phi(z)= \frac{1}{1+\exp(-z)}&amp;lt;/math&amp;gt; (up to variance scaling). Luce's model generalizes the Bradley-Terry models, by considering choices among numerous alternatives and setting &amp;lt;math&amp;gt;\phi(z_2, z_3, ...)= \frac{1}{1+\exp(-z_2)+\exp(-z_3)+...}&amp;lt;/math&amp;gt;, where this quantity is the probability of choosing option 1, and where &amp;lt;math&amp;gt;z_i = \theta_1-\theta_i&amp;lt;/math&amp;gt;. Interestingly, Luce proved that this framework was equivalent to demanding independence of irrelevant alternatives.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The Bradley-Terry model assumes that &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt; follows a [https://en.wikipedia.org/wiki/Gumbel_distribution Gumbel distribution]. Equivalently, it sets &amp;lt;math&amp;gt;\phi(z)= \frac{1}{1+\exp(-z)}&amp;lt;/math&amp;gt; (up to variance scaling). Luce's model generalizes the Bradley-Terry models, by considering choices among numerous alternatives and setting &amp;lt;math&amp;gt;\phi(z_2, z_3, ...)= \frac{1}{1+\exp(-z_2)+\exp(-z_3)+...}&amp;lt;/math&amp;gt;, where this quantity is the probability of choosing option 1, and where &amp;lt;math&amp;gt;z_i = \theta_1-\theta_i&amp;lt;/math&amp;gt;. Interestingly, Luce proved that this framework was equivalent to demanding independence of irrelevant alternatives.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;== Inference ==&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice. Approximate Bayesian methods are needed.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice. Approximate Bayesian methods are needed.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=180&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Classical models */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=180&amp;oldid=prev"/>
		<updated>2020-02-03T07:45:51Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Classical models&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 07:45, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l13&quot; &gt;Line 13:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 13:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The Bradley-Terry model assumes that &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt; follows a [https://en.wikipedia.org/wiki/Gumbel_distribution Gumbel distribution]. Equivalently, it sets &amp;lt;math&amp;gt;\phi(z)= \frac{1}{1+\exp(-z)}&amp;lt;/math&amp;gt; (up to variance scaling). Luce's model generalizes the Bradley-Terry models, by considering choices among numerous alternatives and setting &amp;lt;math&amp;gt;\phi(z_2, z_3, ...)= \frac{1}{1+\exp(-z_2)+\exp(-z_3)+...}&amp;lt;/math&amp;gt;, where this quantity is the probability of choosing option 1, and where &amp;lt;math&amp;gt;z_i = \theta_1-\theta_i&amp;lt;/math&amp;gt;. Interestingly, Luce proved that this framework was equivalent to demanding independence of irrelevant alternatives.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The Bradley-Terry model assumes that &amp;lt;math&amp;gt;\varepsilon&amp;lt;/math&amp;gt; follows a [https://en.wikipedia.org/wiki/Gumbel_distribution Gumbel distribution]. Equivalently, it sets &amp;lt;math&amp;gt;\phi(z)= \frac{1}{1+\exp(-z)}&amp;lt;/math&amp;gt; (up to variance scaling). Luce's model generalizes the Bradley-Terry models, by considering choices among numerous alternatives and setting &amp;lt;math&amp;gt;\phi(z_2, z_3, ...)= \frac{1}{1+\exp(-z_2)+\exp(-z_3)+...}&amp;lt;/math&amp;gt;, where this quantity is the probability of choosing option 1, and where &amp;lt;math&amp;gt;z_i = \theta_1-\theta_i&amp;lt;/math&amp;gt;. Interestingly, Luce proved that this framework was equivalent to demanding independence of irrelevant alternatives.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;. Approximate Bayesian methods are needed&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that in the Luce model, the log-likelihood &amp;lt;math&amp;gt;\log \mathbb P[\mathcal D|\theta] = \sum \left( \theta_i - \log \sum_{(j \succ i) \in \mathcal D} \exp \theta_j \right)&amp;lt;/math&amp;gt; is a strictly concave function in &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if the comparison graph is strongly connected (for any &amp;lt;math&amp;gt;i,j&amp;lt;/math&amp;gt;, there exists &amp;lt;math&amp;gt;k_1, ... k_m&amp;lt;/math&amp;gt; such that there are some of the data that says &amp;lt;math&amp;gt;i \succ k_1 \succ k_2 \succ ... \succ k_m \succ j&amp;lt;/math&amp;gt;). This proves the existence and uniqueness of the maximum likelihood estimator under this assumption.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Note that in the Luce model, the log-likelihood &amp;lt;math&amp;gt;\log \mathbb P[\mathcal D|\theta] = \sum \left( \theta_i - \log \sum_{(j \succ i) \in \mathcal D} \exp \theta_j \right)&amp;lt;/math&amp;gt; is a strictly concave function in &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if the comparison graph is strongly connected (for any &amp;lt;math&amp;gt;i,j&amp;lt;/math&amp;gt;, there exists &amp;lt;math&amp;gt;k_1, ... k_m&amp;lt;/math&amp;gt; such that there are some of the data that says &amp;lt;math&amp;gt;i \succ k_1 \succ k_2 \succ ... \succ k_m \succ j&amp;lt;/math&amp;gt;). This proves the existence and uniqueness of the maximum likelihood estimator under this assumption.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=179&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Classical models */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=179&amp;oldid=prev"/>
		<updated>2020-02-03T07:45:16Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Classical models&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 07:45, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l14&quot; &gt;Line 14:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 14:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;The inference problem is then the problem of inferring the values of parameters &amp;lt;math&amp;gt;\theta_i&amp;lt;/math&amp;gt; given observational data &amp;lt;math&amp;gt;\mathcal D&amp;lt;/math&amp;gt; of choices &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; out of a set of alternatives &amp;lt;math&amp;gt;\mathcal A&amp;lt;/math&amp;gt;. Bayesian inference would suggest computing &amp;lt;math&amp;gt;\mathbb P[\theta|\mathcal D]&amp;lt;/math&amp;gt;. But as often, this approach might be too computationally costly in practice.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Note that in the Luce model, the log-likelihood &amp;lt;math&amp;gt;\log \mathbb P[\mathcal D|\theta] = \sum \left( \theta_i - \log \sum_{(j \succ i) \in \mathcal D} \exp \theta_j \right)&amp;lt;/math&amp;gt; is a strictly concave function in &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if the comparison graph is strongly connected (for any &amp;lt;math&amp;gt;i,j&amp;lt;/math&amp;gt;, there exists &amp;lt;math&amp;gt;k_1, ... k_m&amp;lt;/math&amp;gt; such that there are some of the data that says &amp;lt;math&amp;gt;i \succ k_1 \succ k_2 \succ ... \succ k_m \succ j&amp;lt;/math&amp;gt;). This proves the existence and uniqueness of the maximum likelihood estimator under this assumption.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=178&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Markov chain for preference learning */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=178&amp;oldid=prev"/>
		<updated>2020-02-03T07:45:04Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Markov chain for preference learning&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 07:45, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l17&quot; &gt;Line 17:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 17:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;Note that &lt;/del&gt;in the &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;Luce model&lt;/del&gt;, the &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;log-likelihood &lt;/del&gt;&amp;lt;math&amp;gt;&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;\log \mathbb P[\mathcal D|\theta] = \sum \left&lt;/del&gt;( &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;\theta_i - \log \sum_{(j \succ i) \in \mathcal D} &lt;/del&gt;\exp \&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;theta_j &lt;/del&gt;\&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;right&lt;/del&gt;)&amp;lt;/math&amp;gt; &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;is a strictly concave function in &lt;/del&gt;&amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;if the comparison graph &lt;/del&gt;is &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;strongly connected (for any &amp;lt;math&amp;gt;i,j&amp;lt;/math&amp;gt;, there exists &amp;lt;math&amp;gt;k_1, ... k_m&amp;lt;/math&amp;gt; such that there are some of &lt;/del&gt;the &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;data that says &amp;lt;math&amp;gt;i \succ k_1 \succ k_2 \succ ... \succ k_m \succ j&amp;lt;/math&amp;gt;)&lt;/del&gt;.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;[http://papers.nips.cc/paper/4701-iterative-ranking-from-pair-wise-comparisons.pdf NOS][https://dblp.org/rec/bibtex/conf/nips/NegahbanOS12 12] proposed a Markov chain approach to compute the maximum likelihood estimator, &lt;/ins&gt;in the &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;same vein as the PageRank algorithm [http://ilpubs.stanford.edu:8090/422/1/1999-66.pdf PBMW][https://scholar.google.ch/scholar?hl=en&amp;amp;as_sdt=0%2C5&amp;amp;q=The+PageRank+citation+ranking%3A+Bringing+order+to+the+Web+page+brin+motwani+winograd&amp;amp;btnG= 99]. Namely, by choosing adequately the transition rates of the comparison graph&lt;/ins&gt;, the &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;stationary distribution of the Markov chain thereby constructed turns out to equal the vector &lt;/ins&gt;&amp;lt;math&amp;gt;(\exp \&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;theta_i^&lt;/ins&gt;\&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;star&lt;/ins&gt;)&amp;lt;/math&amp;gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;, where &lt;/ins&gt;&amp;lt;math&amp;gt;\theta&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;^\star&lt;/ins&gt;&amp;lt;/math&amp;gt; is the &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;maximum likelihood estimator&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;−&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;del class=&quot;diffchange diffchange-inline&quot;&gt;One &lt;/del&gt;approach &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;to compute &lt;/del&gt;the &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;maximum likelihood estimator (&amp;lt;math&amp;gt;\arg\max_\theta \mathbb P[\mathcal D|\theta]&amp;lt;/math&amp;gt;) consists &lt;/del&gt;of &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;computing the stationary distribution of a Markov chain on the &amp;lt;em&amp;gt;comparison graph&amp;lt;/em&amp;gt;&lt;/del&gt;. &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;This is a graph whose nodes are alternatives&lt;/del&gt;, &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;and which contains a directed arrow from &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; &lt;/del&gt;to &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;&amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; if &lt;/del&gt;the &lt;del class=&quot;diffchange diffchange-inline&quot;&gt;user has once asserted that he prefers &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;&lt;/del&gt;.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;Note however, that this &lt;/ins&gt;approach &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;is too data and computationally expensive if &lt;/ins&gt;the &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;set &lt;/ins&gt;of &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;alternatives is combinatorial&lt;/ins&gt;. &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;In fact, for most recommender systems, it may be inapplicable&lt;/ins&gt;, &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;since most users have never even been exposed &lt;/ins&gt;to &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;most of &lt;/ins&gt;the &lt;ins class=&quot;diffchange diffchange-inline&quot;&gt;video or music contents of platforms like YouTube or Spotify&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
	<entry>
		<id>https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=177&amp;oldid=prev</id>
		<title>Lê Nguyên Hoang: /* Markov chain for preference learning */</title>
		<link rel="alternate" type="text/html" href="https://robustlybeneficial.org/wiki/index.php?title=Preference_learning_from_comparisons&amp;diff=177&amp;oldid=prev"/>
		<updated>2020-02-03T07:32:52Z</updated>

		<summary type="html">&lt;p&gt;&lt;span dir=&quot;auto&quot;&gt;&lt;span class=&quot;autocomment&quot;&gt;Markov chain for preference learning&lt;/span&gt;&lt;/span&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #222; text-align: center;&quot;&gt;Revision as of 07:32, 3 February 2020&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l17&quot; &gt;Line 17:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 17:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Markov chain for preference learning ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Note that in the Luce model, the log-likelihood &amp;lt;math&amp;gt;\log \mathbb P[\mathcal D|\theta] = \sum \left( \theta_i - \log \sum_{(j \succ i) \in \mathcal D} \exp \theta_j \right)&amp;lt;/math&amp;gt; is a strictly concave function in &amp;lt;math&amp;gt;\theta&amp;lt;/math&amp;gt; if the comparison graph is strongly connected (for any &amp;lt;math&amp;gt;i,j&amp;lt;/math&amp;gt;, there exists &amp;lt;math&amp;gt;k_1, ... k_m&amp;lt;/math&amp;gt; such that there are some of the data that says &amp;lt;math&amp;gt;i \succ k_1 \succ k_2 \succ ... \succ k_m \succ j&amp;lt;/math&amp;gt;).&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;One approach to compute the maximum likelihood estimator (&amp;lt;math&amp;gt;\arg\max_\theta \mathbb P[\mathcal D|\theta]&amp;lt;/math&amp;gt;) consists of computing the stationary distribution of a Markov chain on the &amp;lt;em&amp;gt;comparison graph&amp;lt;/em&amp;gt;. This is a graph whose nodes are alternatives, and which contains a directed arrow from &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; if the user has once asserted that he prefers &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #222; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Gaussian process ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Lê Nguyên Hoang</name></author>
		
	</entry>
</feed>