mathstodon.xyz is one of the many independent Mastodon servers you can use to participate in the fediverse.
A Mastodon instance for maths people. We have LaTeX rendering in the web interface!

Server stats:

2.8K
active users

#AlignmentProblem

0 posts0 participants0 posts today
Jim Donegan 🎵 ✅<p>"OpenAI's o1 just hacked the system"</p><p>Frankly, I am not surprised at this given the well known issue of machine maximisation functions within typical misalignment around stated goals. Have we learned nothing from the <a href="https://mastodon.scot/tags/Bostrom" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Bostrom</span></a> <a href="https://mastodon.scot/tags/PaperclipProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>PaperclipProblem</span></a> ? In a way, it's still impressive that we've now ACHIEVED it.</p><p><a href="https://www.youtube.com/watch?v=oJgbqcF4sBY" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">youtube.com/watch?v=oJgbqcF4sB</span><span class="invisible">Y</span></a></p><p><a href="https://mastodon.scot/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://mastodon.scot/tags/ArtificialIntelligence" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ArtificialIntelligence</span></a> <a href="https://mastodon.scot/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a> <a href="https://mastodon.scot/tags/Alignment" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Alignment</span></a> <a href="https://mastodon.scot/tags/Misalignment" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Misalignment</span></a> <a href="https://mastodon.scot/tags/Hacking" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Hacking</span></a></p>
Joanna Bryson, blathering<p>Re the <a href="https://mastodon.social/tags/alignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>alignmentProblem</span></a>: the chief things we need to be worrying about in <a href="https://mastodon.social/tags/AIEthics" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AIEthics</span></a> (and governance more generally) is human autonomy, accountability, and responsibility, and that is all enabled through transparency. The "research" (surveillance capitalist) trend of ML to get at what the users doesn't know about themselves then tidy the world out of the user's sight is not enabling, its disabling. It fragments social structure and facilitates corporate-political excess.</p>
Cory Doctorow<p>This is sometimes called the <a href="https://mamot.fr/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a>. High-speed, probabilistic systems that can't be fully predicted in advance can *very* quickly run off the rails. It's an idea that pre-dates AI, of course - think of the <a href="https://mamot.fr/tags/SorcerersApprentice" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>SorcerersApprentice</span></a>. But AI produces these perverse outcomes at scale...and so does capitalism.</p><p>5/</p>
Larry O'Brien<p>I think it was Cory Doctorow who came up with the metaphor of corporations as "slow <a href="https://hachyderm.io/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a>." The <a href="https://hachyderm.io/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a> can be seen with corporations: there's a gap between what you want the system to do ("optimize societal benefit") and how it pursues that goal ("maximize short-term profits"). At the media level the gap is between "be rewarded for entertaining people" and the pursuit of "maximize engagement." If aligning "slow AI" has led to big problems, what about when AI becomes "fast"?</p>
James Fern<p>I often hear A.I. 'experts' talk about the 3 things that we previously said wouldn't allow A.I. to do when it becomes advanced. I don't see specific reference to it in the usual places (Russell, Tegmark, Kurzweil, Christian) </p><p>1. code <br>2. understand human emotion <br>3. access the internet</p><p>Does anyone know a specific source for this? </p><p><a href="https://techhub.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://techhub.social/tags/AGI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AGI</span></a> <a href="https://techhub.social/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a> <a href="https://techhub.social/tags/chatgpt" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>chatgpt</span></a></p>
Salve J. Nilsen<p>In the talk above (about <a href="https://chaos.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a>'s and <a href="https://chaos.social/tags/ChatGPT" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ChatGPT</span></a>'s <a href="https://chaos.social/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a>), Harris mentions another presentation he gave in March.</p><p>This is the one: <a href="https://www.youtube.com/watch?v=xoVJKj8lcNQ" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">youtube.com/watch?v=xoVJKj8lcN</span><span class="invisible">Q</span></a></p><p>He talks about how we handle AI being a "Civilizational Right of Passage Moment".</p><p>He's very nice about it! Too nice, maybe.</p><p>How about just calling it our next "Great Filter Moment" instead? 😐</p>
Salve J. Nilsen<p><a href="https://chaos.social/tags/Recommendation" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Recommendation</span></a>: Super useful conversation between <span class="h-card"><a href="https://mastodon.world/@lessig" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>lessig</span></a></span> and Tristan Harris about <a href="https://chaos.social/tags/SocialMedia" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>SocialMedia</span></a>, <a href="https://chaos.social/tags/Policy" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Policy</span></a>, <a href="https://chaos.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a>, and the <a href="https://chaos.social/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a>, and how risks and failures there are likely to shape things to come.</p><p>9/1 (on a 0-10/0-10 scale) Signal/Noise ratio, 1h21m, multitask-friendly audio.</p><p><a href="https://open.spotify.com/episode/5IxYtMKgsmFE2J2NJO8M7Z" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="ellipsis">open.spotify.com/episode/5IxYt</span><span class="invisible">MKgsmFE2J2NJO8M7Z</span></a></p>
Joaquín Herrero :wiki:<p>"Currently, we don't have a solution for steering or controlling a potentially superintelligent AI, and preventing it from going rogue. Our current techniques for aligning AI, such as reinforcement learning from human feedback, rely on humans’ ability to supervise AI. But humans won’t be able to reliably supervise AI systems much smarter than us. We need new scientific and technical breakthroughs."</p><p>Introducing Superalignment<br><a href="https://openai.com/blog/introducing-superalignment" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="ellipsis">openai.com/blog/introducing-su</span><span class="invisible">peralignment</span></a></p><p><a href="https://scholar.social/tags/alignmentproblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>alignmentproblem</span></a></p>
FeralRobots<p>Put another way: The <a href="https://mastodon.social/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a> isn't technical, it's ethical, &amp; it's a problem that humans will solve by behaving ethically when they create &amp; use AI/ML.</p>
FeralRobots<p>The "<a href="https://mastodon.social/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a>" is absolutely not something the current generation of <a href="https://mastodon.social/tags/longtermism" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>longtermism</span></a> cultists (which is who're driving a lot of the hype around "AI" right now) are capable of solving, given that their "existential risk" calculus almost always optimizes for their social &amp; economic advantage. It's a basic matter of empirical versus theoretical approach: Their theory is more or less never, ever checked by empirical tests. (Thought experiments aren't empirical.)</p>
Freedom Baird<p>Interesting and informative interview with OpenAI co-founder and chief scientist Ilya Sutskever</p><p>Good clarity and context around OpenAIs approach to LLMs</p><p>I noted a few key points in the reply below</p><p>(I have personal interest in this and also will present on it at work next month)</p><p><a href="https://mas.to/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://mas.to/tags/OpenAI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>OpenAI</span></a> <a href="https://mas.to/tags/IlyaSutskever" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>IlyaSutskever</span></a> <a href="https://mas.to/tags/ChatGPT" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ChatGPT</span></a> <a href="https://mas.to/tags/BingChat" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>BingChat</span></a> <a href="https://mas.to/tags/NeuralNetwork" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>NeuralNetwork</span></a> <a href="https://mas.to/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a></p><p><a href="https://www.youtube.com/watch?v=UHSkjro-VbE" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">youtube.com/watch?v=UHSkjro-Vb</span><span class="invisible">E</span></a></p>
Peter Drake HAS MOVED<p>In my AI &amp; Machine Learning course, I have each student choose and read a book about the social context of AI. Here's the current list of options:</p><p><a href="https://docs.google.com/spreadsheets/d/1IfAQx8gbiDUQaQFDGcW0o353BKtizoR3N4Jp7zy1MkQ/edit?usp=sharing" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="ellipsis">docs.google.com/spreadsheets/d</span><span class="invisible">/1IfAQx8gbiDUQaQFDGcW0o353BKtizoR3N4Jp7zy1MkQ/edit?usp=sharing</span></a></p><p>I'd be interested in your suggestions, with the following constraints:</p><p>1) It has to be nonfiction.<br>2) It must be no more than ten years old.*</p><p>Thanks!</p><p><a href="https://qoto.org/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a> <a href="https://qoto.org/tags/ml" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ml</span></a> <a href="https://qoto.org/tags/MachineLearning" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>MachineLearning</span></a> <a href="https://qoto.org/tags/ethics" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ethics</span></a> <a href="https://qoto.org/tags/TechEthics" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>TechEthics</span></a> <a href="https://qoto.org/tags/bias" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>bias</span></a> <a href="https://qoto.org/tags/privacy" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>privacy</span></a> <a href="https://qoto.org/tags/justice" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>justice</span></a> <a href="https://qoto.org/tags/equality" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>equality</span></a> <a href="https://qoto.org/tags/AlignmentProblem" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AlignmentProblem</span></a></p><p>*Every time I open this to suggestions, someone comes out of the woodwork to suggest that students read some Really Important book that was written in 1974. That may very well be, but they should do it in a course in philosophy or the history of technology, not this one. The world has changed too much.</p>