-
Notifications
You must be signed in to change notification settings - Fork 0
Home
mpursley edited this page Jan 10, 2024
·
1 revision
| Title/TLDR ____________________________________ |
Arugment __________________________________________________________________________________________ |
Counter Aguments __________________________________________________________________________________________ |
Author | Reference URL | Expecting Uncontrollable Risk? |
|---|---|---|---|---|---|
| LLMS are currently "summoned and tame" | - The problem is that with with these experiments uh they are producing uncontrollable Minds uh the yeah that's why I call it someone and tame uh Paradigm of AI - You summoned this mind uh from the Mind space using your data and a lot of compute a lot of money and then you try to tame it using things reinforcement learning from Human feedback | Jaan Tallin | https://youtu.be/Dmh6ciu24v0?t=2993 | Yes | |
| LLMs Dev should be required to by "insurance" for potential damages | - I would combine this with some sort of required liabilityinsurance that would even strengthen it such that the bigger the damages and the more they have to make sure somebody some insurance company is set up to pay for it and that would also discourage them in the sense that it would be harder to arrange for that insurance and the insurers would say you know you know show me that you're not going too close to this problematic part and they have problem doing that and that's why they would stay away that's the idea | Robin Hanson | https://youtu.be/Dmh6ciu24v0?t=2999 | No | |
| Future LLMs might be smarter then Human Level, but that does not pose a specific risk | - I'm granting that GPT 5 could be superhuman on many performance characteristics what I'm doubting is that that destroys the world. | Robin Hanson | https://youtu.be/Dmh6ciu24v0?t=3370 | No | |
| If people were doing bad thinags, then I would look into stopping it | - if people were trying to make the worst case systems and like the management and funding of a particular you know most advanced projects were trying to make the worst case scenarios they went out of their way to make those scenarios well then I get a lot more worried about those | - Wow. - People are trying to use GPTs to be bad things already - They wouldn't "broadcast" it if they were, as it would undoughtedly be illegal | Robin Hanson | https://youtu.be/Dmh6ciu24v0?t=3483 | No |
| AI research better than better than Humans | - if we are talking about things that is able to do AI research better than better than humans uh then of course that would be um potentially manifested by the fact that we're dead | Jaan Tallin | https://youtu.be/Dmh6ciu24v0?t=4614 | Yes | |
| AI and LLMs are already being developed with help from AI | - the thing that I would be tracking is like how much uh in every successful every successive generation how much effort was kind of put in input by uh humans on how much was AI because it's already non-zero by AI when it comes to like designing the chips and architectures there's one more topic I want to hit on | Jaan Tallin | https://youtu.be/Dmh6ciu24v0?t=4713 | Yes | |
| - researchers actually do not know how to reliably uh Point uh this black box to like any gold | Jaan Tallin, Eliezer Yudkowsky | ||||
| AI Could Turn the world into an RPG to help Humans? | - in an MMO RPG, the mission goal for how to win is listed at the top of the screen. AI could do that for Humans. Do we want that? | George Hotz | https://youtu.be/WS5wGal3ukw?si=Mx3mqIRVvRx0-cpv | ||
| Moore's law for mad scientists - the IQ needed for people to become very dangerous drops my 2 IQ points each year... | elizer yudkowski | https://youtu.be/R78mbtNeCvM?si=xlQvbBXXmEECBoEb | |||
| - it won't be dramatic. we'll just get more and more confused about what is real on the Internet. no one will know what's happening. and we all fall over dead one day and we won't know why | Conor Leahy | Yes | |||
| Natural Selection favors AIs | - Natural Selection favors AIs or Human and this could lead to Human extinction. | Dan Hendrycks | https://twitter.com/DanHendrycks/status/1651741353574019074 https://youtu.be/Fk7cMD44uWs?t=238 | Yes | |
| Squiggle Maximizer (formerly "Paperclip maximizer") | - it would undergo an intelligence explosion: It would work to improve its own intelligence, where "intelligence" is understood in the sense of optimization power, the ability to maximize a reward/utility function—in this case, the number of paperclips. The AGI would improve its intelligence, not because it values more intelligence in its own right, but because more intelligence would help it achieve its goal of accumulating paperclips. Having increased its intelligence, it would produce more paperclips, and also use its enhanced abilities to further self-improve. Continuing this process, it would undergo an intelligence explosion and reach far-above-human levels. | Eliezer Yudkowsky | https://www.lesswrong.com/tag/squiggle-maximizer-formerly-paperclip-maximizer | Yes | |
| The market will exhibit caution | - the market will exhibit caution. every organism company consumer is acting out of self-interest and they won't assign Capital to things that have negative utility to them | - "the problem is with the market is like you know there's not always perfect information there's manipulation there's a bad faith actors that mess with the system it's not it's not always a um rational and honest system". -- Lex Friedman | Guillaume Verdon | https://youtu.be/Fk7cMD44uWs?t=413 | No |
| Concrete Problems in AI Safety | - 1. Avoiding Negative Side Effects - 2. Avoiding Reward Hacking - 3. | https://arxiv.org/abs/1606.06565 | Yes | ||
| - they hack all the systems and confuse everyone | Conor leahy | https://youtu.be/Cf_l6LNiXGI?si=LmM5JfGZuUnF4Fhq 22:50 | |||
| Train up a Human Level AI and ask it how to align a SuperAGI | - Plan of how we want to approach the problem that involves training uh roughly human level alignment researcher that can work automatically and then ask that automated alignment researcher and to figure out how to align super intelligence -- Jan Leike https://youtu.be/Uk6-Rw5N_Dg?t=120 | - How good is it is the model as like spinning really clear and lies or like being deceptive of like pretending to do something or believe one thing and then actually wanting another I think another like really key capability here is self-exfiltration so how good would the model be like breaking the security precautions and like accessing its own weights and trying to copy it somewhere else on the internet um or persuading engineer with access to the weights to you know like download them and send them somewhere -- Jan Leike https://youtu.be/Uk6-Rw5N_Dg?t=517 | Jan Leike | https://youtu.be/Uk6-Rw5N_Dg?t=120 | No |