AI’s Dirty Little Secret

521,884
0
Published 2024-06-04
Learn more about neural networks and large language models with Brilliant! First 30 days are free and 20% off the annual premium subscription when you use our link ➜ brilliant.org/sabine.

There’s a lot of talk about artificial intelligence these days, but what I find most interesting about AI no one ever talks about. It’s that we have no idea why they work as well as they do. I find this a very interesting problem because I think if we figure it out it’ll also tell us something about how the human brain works. Let’s have a look.

🤓 Check out my new quiz app ➜ quizwithit.com/
💌 Support me on Donorbox ➜ donorbox.org/swtg
📝 Transcripts and written news on Substack ➜ sciencewtg.substack.com/
👉 Transcript with links to references on Patreon ➜ www.patreon.com/Sabine
📩 Free weekly science newsletter ➜ sabinehossenfelder.com/newsletter/
👂 Audio only podcast ➜ open.spotify.com/show/0MkNfXlKnMPEUMEeKQYmYC
🔗 Join this channel to get access to perks ➜
youtube.com/channel/UC1yNl2E66ZzKApQdRuTQ4tw/join
🖼️ On instagram ➜ www.instagram.com/sciencewtg/

#science #sciencenews #technews #tech #ai

All Comments (21)
  • @and3583
    "Alexa, I need emergency medical treatment" "I've added emergency medical treatment to your shopping list"
  • @malachimcleod
    "It's like a teenager, but without the eye-rolling." 🤣
  • @rich_tube
    As someone who works in machine learning research, I find this video a bit surprising, since 90% of what we are doing is developing approaches to fight overfitting when using big models. So we do very well know why NNs don’t overfit: stochastic/mini batch gradient descent, momentum based optimizers, norm-regularization, early stopping, batch normalization, dropout, gradient clipping, data augmentation, model pruning, and many, many more very clever ideas…
  • @Lyserg.
    Stop all trains to prevent train crashes is the same logic like cancelled trains are not delayed. I think the AI learned from Deutsche Bahn (German railway company).
  • @Pau_Pau9
    This is a story I read from a magazine long time ago: In distant future, scientists create a super complex AI computer to solve energy crisis that is plaguing mankind. So much time, resources and money was put into creating this super AI computer. Then the machine is complete and the scientists nervously turn on the machine for the first time. Then the lead scientist asks, "Almighty Super Computer, how do we resolve our current energy crisis?" Computer replies, "Turn me off."
  • One of my favorites is that in skin cancer pictures, an AI came to the conclusion that rulers cause cancer (because the malignant ones were measured in the majority of pictures)
  • @enduka
    That phenomenon is called Grokking, aka "generalizing after overfitting". There is quite some recent research in that area. Experiments on some toy datasets suggests thet the models first memorizes the data and then tries to find more efficient ways to represent the embedding space leading to better overall performance.(Source: Towards Understanding Grokking: An Effective Theory of Representation Learning)
  • @oleran4569
    And people who come to emergency medical departments by car tend toward better outcomes than those who arrive by ambulance. We should likely stop using ambulances.
  • @nickdryad
    Man, I went out with a model. I never could predict what was going to happen next
  • @user-qn8ne8lr2k
    I come here every day just to listen to how Sabine says: "No one knows"
  • @generessler6282
    Haha. The "stop all the trains" solution is a mirror of the old movie "Colossus, the Forbin Project." To prevent human race from hurting itself, enslave it.
  • @Lazdinger
    The “you can’t crash a train that never leaves the station” answer sounded kinda like a glorious StackOverflow response.
  • @beatsaway
    this is amazing u prevent the misconceptions by addressing them one by one in the intro
  • It occurs when a model is too specialized to the training data and performs poorly on new, unseen data. This can happen when a model is too complex, has too many parameters relative to the amount of training data, or when the training data itself contains a lot of noise or irrelevant information "The man with a hammer analogy perfectly captures the essence of the overfitting issue in AI. Just as the man with a hammer sees every problem as a nail, an overfitting model sees every pattern in the training data as crucial, even if it's just noise. It becomes so specialized to the training data that it loses sight of the bigger picture, much like the man who tries to hammer every problem into submission. As a result, the model performs exceptionally well on the training data but fails miserably when faced with new, unseen data. This is because it has become too good at fitting the noise and irrelevant details in the training data, rather than learning the underlying patterns that truly matter. Just as the man with a hammer needs to learn to put down his trusty tool and approach problems with a more nuanced perspective, an overfitting model needs to be reined in through regularization and other techniques to prevent it from becoming too specialized and losing its ability to generalize.
  • @AnnNunnally
    We need to use those computers that they have in 50’s movies. It is really big, but you can ask it anything and it prints out a perfect answer.
  • @washingtonx1
    This is one of the best videos I have seen on AI, and I keep up with this stuff much more than average. Well done, Sabine. This is an area to expand on. Please keep going. 🙏
  • @Li-rm2gj
    Fantastic video Sabine. Interesting, knowledgeable, highly relevant. Very impressive for people to communicate a topic this well outside of their field.