OpenAI’s ChatGPT Is Now Learning From Another AI!
❤️ Check out Lambda here and sign up for their GPU Cloud:
📝 The paper "Rule Based Rewards for Language Model Safety" is available here:
📝 My paper on simulations that look almost like reality is available for free here:
Or this is the orig. Nature Physics link with clickable citations:
🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
Alex Balfanz, Alex Haro, B Shang, Benji Rabhan, Gaston Ingaramo, Gordon Child, John Le, Juan Benet, Kyle Davis, Loyal Alchemist, Lukas Biewald, Martin, Michael Albrecht, Michael Tedder, Owen Skarpness, Richard Sundvall, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Tybie Fitzhugh, Ueli Gallizzi.
If you wish to appear here or pick up other perks, click here:
My research:
X/Twitter:
Thumbnail design: Felícia Zsolnai-Fehér –
it begins
lmfao i knew this would be here already.
stop watching Terminator pal
I can’t get enough of your content. 😮 What a time to be alive! ❤
lol
Anyone else remember back when two minute papers was exited when an AI could complete sentences?
That was two papers ago
What a time to be alive!
The rest of the internet: training their models using datasets generated from ChatGPT for the past 2 years
Two Minute Papers:
IMHO the true brakethrough will come when AIs will start looking for data in nature, beyond human made data. Imagine an AI that can use microscopes and telescopes without humans in the loop. That will lead to higher intelligence. Otherwise AIs will always gravitate around human intelligence… and stupidity.
We control AI and if we don’t it just hallucinates and come up with random things. Its knowledge is based of our conclusions, but there is not one right answer to everything and we change our opinions all the time. What is higher intelligence? High intelligence doesn’t mean much without real life experience. I see no fear in high intelligence. If we all have AI there is no advantage of having it.
@@hombacom Intelligence can be benchmarked in many ways. Higher intelligence, in a scientific context, would mean better scores in those benchmarks.
About real life experience, having the AIs controlling autonomously what goes into the training data – possibly going beyond human made data- would constitute just that: Experience (without human filters). Anyways think about animal intelligence, for example in octopuses: It evolved independently from us, so I assume the same cold be true for AIs. There are some examples of AIs learning some common sense physics just by observing simulations. I believe that this can be brought one step further to achieve superintelligence.
How can anyone not subscribe to this channel?
As William said:
“The future is already here – it’s just not evenly distributed.”
I see this channel in my feed all the time yet don’t subscribe. Or if you really want to watch one of his videos you can just go to the page.
What the point of subscribing?
because he sounds like he had a stroke
I’ve been surprised lately by what ChatGPT has been willing to discuss with me. I can’t remember the last time it told me no. Maybe month or more? I been discussing wide variety of things, from chemicals to hacking a firmware on a radio. I will advocate for fully open models, but have also been reasonably happy with its current safety levels.
I stopped using Claude because it wouldnt even help me decipher a blinking airbag light code.
Claude is the dude who always snitched in class
you’ll advocate for fully open models until some bad character will successfuly use such model to rob you
I really like Claude. I’m not a programmer, but I’m an amateur mathematician (great emphasis on ‘amateur’😂), and Claude writes code for me (usually Python) that allows me to rapidly test ideas that I would normally have to hand calculate. I even had it write an awesome version of Tetris, but with all polyominoes, from 1 cell, to 5 cells…and with a soundtrack! It’s truly crazy to have these tools.⚔️
“It depends” drives us humans up the wall? Are you sure? I often start my sentences with that. Always have. Because many things depend on many different factors and/or perspectives.
Saying “it depends” and stopping there
I asked about Early algorithm to ChatGPT and it refused to answer saying it’s unsafe and might be a security risk🤯
I don’t know what “early algorithm” is so I just asked ChatGPT to explain it for me to see if I would get the same result.
It started freely talking about it so I don’t know what you might be doing wrong
@@codycast i asked it to implement it in any programming language, then it said so but that was 2 months ago so…
Anyways I used my two brain 🧠 cells to figure it out myself!
“Safety” in non-agentic AI is just censorship. Safety only becomes a concern when AI agents are capable of executing long horizon tasks autonomously.
Are you sure about that? Can you not think of *any* case, where image/video based generative AI might be able to produce material that is straight up illegal, for example? Models are over-censored, but there are valid cases for safety mechanisms.
“this is anti-semitic question.”
“this question is against our community standard.”
“This question is flagged as hate speech.”
It’s self-censorship, which is fine.
1 – We dream and wake, and wake and dream, so our machines are born with more dreams than we could ever conceive. 2 – This thing that we built already builds its own things, and EVERYTHING we know is just ONE of its dreams.
I slowly find talking to some AI-brains more helpful than to some humans
Slowly? You know more helpful people than I do.
Microsoft’s Copilot is really good these days.
Often, when I ask it a question, I not only get a detailed answer but a short, intelligent conversation as it asks me a question regarding my query.
We desperately need a model that has the HR ripped out of it.
Where censorship is “needed”, it’s a massive red flag that there are deeper more serious problems at play. Those tend to be elephants in the room that people prefer to ignore. You can use duct tape to cover someone’s mouth but don’t think it will patch that crack running up the brick wall.
Chat GP 3 went through months of extensive testing and safeguard training to ensure that the outputs given are not malicious in nature. One benefit to having this separate system is that when OpenAI upgrade models, say to GPT 5 they wont need to go through the safe guarding stages again, and can just use the old watch model to monitor the new one
AIs are already overly censored… These big tech acting like we’re all first grade kids
This sparked a thought. We humans learn quite a bit from failure and the successes are enhanced by this. AI, I’m guessing, was trained on only the correct information (labeled information) and AI doesn’t have all of the missing failures that could engender a more rich AI experience.
4:37 What does it even mean for it to have 16% less bad refusals than humans do? Like, humans mistake 16% more safe tasks for unsafe ones?
Be warned, I saw how this can also create a huge mess, as ai’s copying fuzzy ai things, copying fuzzy things, might lead to more and more and more mistakes. This warning (another video) said many AI’s would slip into a mess because of AI learning from AI.
It wasn’t just found, it has been used for many months, and outside of OpenAI as well.
I feel as safe as David Bowman, when an AI tells me I’m guilty of wrong think.