Discussion about this post

User's avatar
Benjamin Riley's avatar

You're really onto something here. Early reporting from Kara Swisher suggests that Sutskever found support on the board from Helen Toner, the director of CSET. She's authored numerous reports that make clear her concerns about how generative AI models may inflict harms as they grow more powerful -- whether a closet doomer or not is almost irrelevant, given the present dangers that concern her. Here's a quote from one of her reports:

As machine learning systems become more advanced, they will

likely be deployed in increasingly complex environments to carry

out increasingly complex tasks. This is where specification

problems may begin to bite. Without significant progress in

methods to convey intentions, machine learning systems will

continue to carry out their instructions exactly as given—obeying

the letter, not the spirit, of the rules their designer gives them.

To address the challenges posed by misspecification, more

machine learning research needs to account for worst case

scenarios and develop algorithms that more explicitly incorporate

human supervision or provide theoretical guarantees for the worst

case performance under a given specification.

Almost by definition, using commercial deployment to test generative AI systems as a product runs afoul of this imperative. The (much?) safer way to account for worst-case scenarios and develop theoretical guarantees -- perhaps better described as safeguards -- is to conduct research on them in advance. This may end up being the opening salvo in an epic philosophical war.

https://cset.georgetown.edu/wp-content/uploads/Key-Concepts-in-AI-Safety-Specification-in-Machine-Learning.pdf

Expand full comment
Alexa McLain's avatar

I mean more on a human relationship/emotional level. Like maybe their philosophical alignments haven’t changed at all, but for some reason(s) there’s an inability to assume good intent that previously existed.

Expand full comment
22 more comments...

No posts