<?xml version="1.0" encoding="UTF-8" standalone="yes"?><oembed><version><![CDATA[1.0]]></version><provider_name><![CDATA[Ordinary Ideas]]></provider_name><provider_url><![CDATA[https://ordinaryideas.wordpress.com]]></provider_url><author_name><![CDATA[paulfchristiano]]></author_name><author_url><![CDATA[https://ordinaryideas.wordpress.com/author/paulfchristiano/]]></author_url><title><![CDATA[Best of]]></title><type><![CDATA[link]]></type><html><![CDATA[<p><a title="“Indirect Normativity” Write-up" href="https://ordinaryideas.wordpress.com/2012/04/21/indirect-normativity-write-up/">Indirect normative theories</a> are a general approach to formally specifying what we value, without having to actually solve any hard problems. In this post I give a very formal proposal which could be implemented with (roughly) existing technology. It&#8217;s not something I would ever recommend doing, but I think it does work as a general argument of feasibility for this class of approaches.</p>
<p>The usual universal prior is <a href="https://ordinaryideas.wordpress.com/2016/11/30/what-does-the-universal-prior-actually-look-like/">really weird</a>, and it would probably be bad if we ever actually used it to make important decisions.</p>
<p><a href="https://ordinaryideas.wordpress.com/2014/12/30/recent-thoughts/">Some interesting, optimistic ideas about AI safety</a>.</p>
<p>What happens if <a title="Incomprehensible utility functions" href="https://ordinaryideas.wordpress.com/2012/05/07/incomprehensible-utility-functions/">you can&#8217;t understand your values at all</a>? We don&#8217;t have a clear enough account of logical uncertainty to give a great answer, but I think this confusing post makes the right first observation.</p>
<p>A discussion of <a href="https://ordinaryideas.wordpress.com/2015/11/30/driving-fast-in-the-counterfactual-loop/">counterfactual oversight and reliability</a>.</p>
<p>A <a title="Specifying (non-decision-theoretic) Counterfactuals" href="https://ordinaryideas.wordpress.com/2011/12/15/specifying-non-decision-theoretic-counterfactuals/">formal specification</a> for (a narrow class of) counterfactuals, and a <a title="Abstract Randomness and Formal CDT" href="https://ordinaryideas.wordpress.com/2011/12/18/abstract-randomness-and-formal-cdt/">similar specification</a> for implementing causal decision theory.</p>
<p><a title="Motivation" href="https://ordinaryideas.wordpress.com/2012/04/27/motivation/">Why</a> think about these things at all?</p>
<p>There are some hypotheses we assign a very low probability. But if you assign a <em>simple</em> hypothesis a very low probability, you will <a title="Improbable simple hypotheses are unbelievable" href="https://ordinaryideas.wordpress.com/2012/12/11/improbable-simple-hypotheses-are-unbelievable/">never believe it</a>.</p>
<p>The speed prior <a title="Speed Prior and MWI" href="https://ordinaryideas.wordpress.com/2011/12/21/speed-prior-and-mwi/">doesn&#8217;t</a> have anything to say about MWI except by straightforward question-begging. By similar arguments, it&#8217;s not clear it has anything to say about any physical questions.</p>
<p>The amount of meaningful computation we can do in the universe may be <a title="Entropy, Reversibility, and Uncomputation" href="https://ordinaryideas.wordpress.com/2011/12/21/entropy-reversibility-and-uncomputation/">much larger than you would think</a>. I was more pleased with this post before I learned that it was an old result in complexity theory, but it&#8217;s still an interesting point.</p>
<p>Formally speaking, if it were possible a physical implementation of AIXI <a title="AIXI and Existential Despair" href="https://ordinaryideas.wordpress.com/2011/12/14/aixi-and-existential-despair/">would have problems</a>. It looks like natural realistic approximations would have problems, too.</p>
<p>Prediction markets <a title="Risk Arbitrage" href="https://ordinaryideas.wordpress.com/2011/12/16/risk-arbitrage/">don&#8217;t tell you</a> the probability of things. Probably obvious to anyone in finance.</p>
<p>&nbsp;</p>
]]></html></oembed>