OpenAI’s ChatGPT Is Now Learning From Another AI!

❤️ Check out Lambda here and sign up for their GPU Cloud:

📝 The paper "Rule Based Rewards for Language Model Safety" is available here:

📝 My paper on simulations that look almost like reality is available for free here:

Or this is the orig. Nature Physics link with clickable citations:

🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
Alex Balfanz, Alex Haro, B Shang, Benji Rabhan, Gaston Ingaramo, Gordon Child, John Le, Juan Benet, Kyle Davis, Loyal Alchemist, Lukas Biewald, Martin, Michael Albrecht, Michael Tedder, Owen Skarpness, Richard Sundvall, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Tybie Fitzhugh, Ueli Gallizzi.
If you wish to appear here or pick up other perks, click here:

My research:
X/Twitter:
Thumbnail design: Felícia Zsolnai-Fehér –

Joe Lilli
 

  • @8bit-ascii says:

    it begins

  • @marlonochoaj says:

    I can’t get enough of your content. 😮 What a time to be alive! ❤

  • @misterprofessor5038 says:

    Anyone else remember back when two minute papers was exited when an AI could complete sentences?

  • @vslaykovsky says:

    The rest of the internet: training their models using datasets generated from ChatGPT for the past 2 years
    Two Minute Papers:

  • @lucamatteobarbieri2493 says:

    IMHO the true brakethrough will come when AIs will start looking for data in nature, beyond human made data. Imagine an AI that can use microscopes and telescopes without humans in the loop. That will lead to higher intelligence. Otherwise AIs will always gravitate around human intelligence… and stupidity.

    • @hombacom says:

      We control AI and if we don’t it just hallucinates and come up with random things. Its knowledge is based of our conclusions, but there is not one right answer to everything and we change our opinions all the time. What is higher intelligence? High intelligence doesn’t mean much without real life experience. I see no fear in high intelligence. If we all have AI there is no advantage of having it.

    • @lucamatteobarbieri2493 says:

      @@hombacom Intelligence can be benchmarked in many ways. Higher intelligence, in a scientific context, would mean better scores in those benchmarks.
      About real life experience, having the AIs controlling autonomously what goes into the training data – possibly going beyond human made data- would constitute just that: Experience (without human filters). Anyways think about animal intelligence, for example in octopuses: It evolved independently from us, so I assume the same cold be true for AIs. There are some examples of AIs learning some common sense physics just by observing simulations. I believe that this can be brought one step further to achieve superintelligence.

  • @45414 says:

    How can anyone not subscribe to this channel?

  • @jamiethomas4079 says:

    I’ve been surprised lately by what ChatGPT has been willing to discuss with me. I can’t remember the last time it told me no. Maybe month or more? I been discussing wide variety of things, from chemicals to hacking a firmware on a radio. I will advocate for fully open models, but have also been reasonably happy with its current safety levels.

    I stopped using Claude because it wouldnt even help me decipher a blinking airbag light code.

    • @PuffyNavel says:

      Claude is the dude who always snitched in class

    • @miauzure3960 says:

      you’ll advocate for fully open models until some bad character will successfuly use such model to rob you

    • @maynardtrendle820 says:

      I really like Claude. I’m not a programmer, but I’m an amateur mathematician (great emphasis on ‘amateur’😂), and Claude writes code for me (usually Python) that allows me to rapidly test ideas that I would normally have to hand calculate. I even had it write an awesome version of Tetris, but with all polyominoes, from 1 cell, to 5 cells…and with a soundtrack! It’s truly crazy to have these tools.⚔️

  • @AkariTheImmortal says:

    “It depends” drives us humans up the wall? Are you sure? I often start my sentences with that. Always have. Because many things depend on many different factors and/or perspectives.

  • @ransomecode says:

    I asked about Early algorithm to ChatGPT and it refused to answer saying it’s unsafe and might be a security risk🤯

    • @codycast says:

      I don’t know what “early algorithm” is so I just asked ChatGPT to explain it for me to see if I would get the same result.
      It started freely talking about it so I don’t know what you might be doing wrong

    • @ransomecode says:

      @@codycast i asked it to implement it in any programming language, then it said so but that was 2 months ago so…

      Anyways I used my two brain 🧠 cells to figure it out myself!

  • @ThreeChe says:

    “Safety” in non-agentic AI is just censorship. Safety only becomes a concern when AI agents are capable of executing long horizon tasks autonomously.

    • @sophiophile says:

      Are you sure about that? Can you not think of *any* case, where image/video based generative AI might be able to produce material that is straight up illegal, for example? Models are over-censored, but there are valid cases for safety mechanisms.

    • @begobolehsjwjangan2359 says:

      “this is anti-semitic question.”
      “this question is against our community standard.”
      “This question is flagged as hate speech.”

    • @alansmithee419 says:

      It’s self-censorship, which is fine.

  • @TyronePost says:

    1 – We dream and wake, and wake and dream, so our machines are born with more dreams than we could ever conceive. 2 – This thing that we built already builds its own things, and EVERYTHING we know is just ONE of its dreams.

  • @user-kl6ov1cm4t says:

    I slowly find talking to some AI-brains more helpful than to some humans

    • @PuffyNavel says:

      Slowly? You know more helpful people than I do.

    • @antonystringfellow5152 says:

      Microsoft’s Copilot is really good these days.
      Often, when I ask it a question, I not only get a detailed answer but a short, intelligent conversation as it asks me a question regarding my query.

  • @tellesu says:

    We desperately need a model that has the HR ripped out of it.

    • @StickerWyck says:

      Where censorship is “needed”, it’s a massive red flag that there are deeper more serious problems at play. Those tend to be elephants in the room that people prefer to ignore. You can use duct tape to cover someone’s mouth but don’t think it will patch that crack running up the brick wall.

  • @EVILBUNNY28 says:

    Chat GP 3 went through months of extensive testing and safeguard training to ensure that the outputs given are not malicious in nature. One benefit to having this separate system is that when OpenAI upgrade models, say to GPT 5 they wont need to go through the safe guarding stages again, and can just use the old watch model to monitor the new one

  • @markmuller7962 says:

    AIs are already overly censored… These big tech acting like we’re all first grade kids

  • @PatrickHoodDaniel says:

    This sparked a thought. We humans learn quite a bit from failure and the successes are enhanced by this. AI, I’m guessing, was trained on only the correct information (labeled information) and AI doesn’t have all of the missing failures that could engender a more rich AI experience.

  • @thearchitect5405 says:

    4:37 What does it even mean for it to have 16% less bad refusals than humans do? Like, humans mistake 16% more safe tasks for unsafe ones?

  • @KootFloris says:

    Be warned, I saw how this can also create a huge mess, as ai’s copying fuzzy ai things, copying fuzzy things, might lead to more and more and more mistakes. This warning (another video) said many AI’s would slip into a mess because of AI learning from AI.

  • @amdenis says:

    It wasn’t just found, it has been used for many months, and outside of OpenAI as well.

  • @theosalmon says:

    I feel as safe as David Bowman, when an AI tells me I’m guilty of wrong think.

  • >