These "AI Ethics" guys on JRE today seem like real dinkuses
| metaphysics is fallow | 04/25/25 | | Ass Sunstein | 04/25/25 | | metaphysics is fallow | 04/25/25 | | Ass Sunstein | 04/25/25 | | internet poaster with a praise kink | 04/25/25 | | internet poaster with a praise kink | 04/25/25 | | ,.,.,.,.,.,.,..,:,,:,,.,:::,.,,.,:.,,.:.,:.,:.::,. | 04/25/25 | | Pierbattista Pizzaballa | 04/25/25 | | ,.,.,.,.,.,.,..,:,,:,,.,:::,.,,.,:.,,.:.,:.,:.::,. | 04/25/25 | | internet poaster with a praise kink | 04/25/25 | | ,.,.,.,....,.,..,.,.,. | 04/25/25 | | metaphysics is fallow | 04/25/25 | | ,.,.,.,....,.,..,.,.,. | 04/25/25 | | metaphysics is fallow | 04/25/25 | | ,.,.,.,....,.,..,.,.,. | 04/25/25 | | metaphysics is fallow | 04/25/25 | | ,.,.,.,....,.,..,.,.,. | 04/25/25 | | internet poaster with a praise kink | 04/25/25 | | Ass Sunstein | 04/25/25 | | metaphysics is fallow | 04/25/25 | | blow off some steam | 04/25/25 | | scholarship | 04/25/25 |
Poast new message in this thread
 |
Date: April 25th, 2025 6:22 PM
Author: ,.,.,.,.,.,.,..,:,,:,,.,:::,.,,.,:.,,.:.,:.,:.::,.
In 5 years you’ll be asking “WHY DIDN’T WE LISTEN??”
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880856) |
Date: April 25th, 2025 6:26 PM
Author: ,.,.,.,.,.,.,..,:,,:,,.,:::,.,,.,:.,,.:.,:.,:.::,.
Tristan Harris (near term risk) and Eliezer Yudkowsky (long term risk) are the only people you should be listening to about this stuff
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880861) |
 |
Date: April 25th, 2025 6:38 PM Author: ,.,.,.,....,.,..,.,.,.
Eliezer essentially built the community for this but I don’t know that it makes him more credible. It just makes him first.
In general, I don’t think AI safety people have updated much from LLMs. Their ethical behavior is robust already even just using reinforcement learning from human feedback on a pre-trained model. The central argument of AI risk seems to be that human values are a narrow target, so pointing an AI agent precisely at them is extremely difficult and unlikely to happen. But they aren’t hand coding this, they are pointing a pre-trained model that was already optimized hard on an objective that encourages the model to learn human preferences. Moving a model like that down a gradient to good behavior is far less difficult than creating a program from scratch that can do it. Their argument would seem to make false predictions too about the ability of stochastic gradient descent to produce generalizing solutions. LLMs could theoretically just memorize stuff they have seen and not generalize, but in practice they don’t
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880894) |
 |
Date: April 25th, 2025 6:56 PM Author: ,.,.,.,....,.,..,.,.,.
They do stupid things like that but i suspect a lot of this will go away with more training and ability to more effectively leverage inference compute. They still use transformers with minor modifications but there are other, more costly architectures that would likely be more reliable if they could be trained at scale. Things like memory augmented transformers that could read and write to a memory state over multiple passes of a neural network, so it could iteratively reason more effectively. It’s actually kind of ridiculous that transformers work as well as they do currently.
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880950) |
 |
Date: April 25th, 2025 7:05 PM Author: ,.,.,.,....,.,..,.,.,.
I have had the 2.5 pro model push back against me. I don’t think this is just a side effect of RLHF encouraging them to be sycophants. I think the models aren’t smart enough that they can always reason their way out of a false premise, so they take the easy path and bullshit.
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880965) |
 |
Date: April 25th, 2025 7:10 PM Author: ,.,.,.,....,.,..,.,.,.
That was my impression and I have had that exact same experience with coding.
(http://www.autoadmit.com/thread.php?thread_id=5716142&forum_id=2#48880978) |
|
|