Scribe
Scribe

ชอบมันไหม? ทำให้ Scribe ดียิ่งขึ้นโดย การให้คะแนน

รับส่วนขยาย Chrome

เรียกดู

  • วิดีโอยอดนิยม
  • วิดีโอล่าสุด
  • ช่องทั้งหมด

เครื่องมือฟรี

  • ตัวดาวน์โหลดคำบรรยาย
  • ตัวสร้างเวลา
  • ตัวสรุปวิดีโอ
  • ตัวนับคำ
  • ตัววิเคราะห์ชื่อเรื่อง
  • ค้นหาบทถอดความวิดีโอ
  • การวิเคราะห์วิดีโอ
  • ตัวสร้างบท
  • ตัวสร้างแบบทดสอบ
  • แชทกับวิดีโอ

ผลิตภัณฑ์

  • ราคา
  • บล็อก

Developers

  • Transcript API
  • API Documentation

กฎหมาย

  • ข้อกำหนด
  • ความเป็นส่วนตัว
  • การสนับสนุน
  • แผนผังเว็บไซต์

ลิขสิทธิ์ © 2026 สร้างด้วยความรักโดย Scribe

— ถ้านี่ทำให้ชีวิตของคุณง่ายขึ้น (หรืออย่างน้อยก็วุ่นวายน้อยลง) กรุณาให้คะแนนเรา! เราสัญญาว่ามันจะทำให้วันของเราดีขึ้น 😊

Related Videos

The Story of Deep Learning

Video thumbnail
69.02k1,667 คำ8m readGrade 18
แชร์
Channel
Art of the Problem
since the mid 20th century people had been dreaming of making computers that mimic the human mind by letting them learn on their own from experience using a new type of computer inspired by biological brains known as neural networks neural networks have remained a more or less obscure corner of computer science as they were conceived in the 1950s but since around 2009 due to a rapid surge in computer power neural networks are suddenly performing functions we never thought possible leading to a second computing revolution we called deep learning traditionally in the 20th century the approach to
computer design was to give them a list of explicit human written instructions to define their action a program regardless of what these programs did they were essentially sequences of logical decisions however with neural networks we throw away this human centric view of programs instead of having humans input explicit instructions to define the behavior of a computer the neural network learns everything by itself using direct experience starting with a random connection pattern it gradually self wires into parallel patterns of computation in order to perform complex functions automatically often achieving human or superhuman performance on a growing
set of tasks which recently were thought far out of reach of machines let's look at a few practical problems first to get a sense of this new power take tic-tac-toe for example the board is small and the length of the game is short and so it's easy to make every decision based on checking all future moves until the game is over chess on the other hand was much harder because the board is bigger and the games are much longer and so the number of possible sequence of moves quickly explodes easily greater than the number of
atoms in the known universe the only way to play chess is to make short cuts which take the form of best guesses the chess masters described these moves as feelings not calculations in the 1990s the deep blue computer developed by IBM was programmed by asking the chess champions for the logic behind their strategies this led to sets of rules which were connected together in a huge step by step decision-making algorithm which famously beat the chess master Garry Kasparov but two people in the computer science community this felt more like a publicity stunt than a real
revolution it wasn't satisfying that a machine following many thousand human rules could do something about as well as a single human can but compare that to the recent deep learning approach applied to chess instead of using human design rules researchers throw away all hard coded knowledge about chess and gave the computer only the rules of chess and let it learn the game of chess entirely through self play starting with a random strategy and feedback on wins and losses it absorbed winning chess strategies which were far beyond human capability and it did this self learning in
a matter of hours and beat the human design computer chess programs which represented 30 plus years of expert work and it keeps getting better Kasparov called it alien chess because it would think deeper and wider than a human can and today right now chess players and masters all over the world are working with these systems to learn new strategies of chess never before seen you know it's always fun to see someone move something with a really iconoclastic style so some of the things that alpha zero is doing are just incredible another example consider translating between
languages up until recently computer translation was built on huge look-up tables of lots of existing translated text to reference along with a huge web of rules for each language at Google for example it was developed with the involvement of hundreds of human experts and linguists this resulted in a complex translation machine which worked for a word or two but quickly broke down into gibberish especially beyond a sentence however compare this to the recent deep learning approaches to translation where researchers throw away all human design knowledge instead they started from scratch and exposed a neural network
to millions of examples of human translated text and let it learn the translation process on its own by absorbing the patterns involved in translation and rapidly the quality of the translation system improved the word error rate on simple translations dropped from 20% to near 0% over the short term the power of deep learning is even more apparent when we look at intuitive problems by intuition we mean actions we perform but can't perfectly describe in words such as the problem of recognizing a face seeing a face is more of a feeling we have difficult to explain
in a sequence of instructions this is why the idea of a human writing a huge series of instructions to recognize things and images seem near impossible the best minds working together on this problem for over 50 years ended up with algorithms which failed about a quarter of the time useless for anything practical especially critical applications that was until people flipped to the deep learning approach and instead collected millions of examples of images which already had a human-made label and fed it into the neural network at the level of individual pixels to learn from using the
human provided labels as initial feedback the network then absorbed the statistical information in the images from the pixels up and separated into conceptual or word buckets where the idea of a dog versus a cat comes down to a unique pattern of activations inside the network in 2012 this was famously demonstrated with a huge breakthrough in a computer vision competition called image net where systems based on neural networks suddenly cut the error rate down by over 50 percent in one year and within just a few years the machines were superhuman at recognizing many kinds of images
so the field of image diagnosis where we are looking for an illness in an image had to adapt to this new tool which can often identify patterns that humans can miss by now the trend should be getting clear the old human engineered rural systems would hit a wall in performance while the deep learning systems just keep improving often beyond a human level but we've only been looking at this from the perspective of using neural networks to analyze information but we can also run them in Reverse to generate information for example a network trained on lots
of human text can spit out random essays on any topic which are difficult if not impossible to distinguish from human they were so good that the researchers who first saw the results decided it was best to hold the system back from the public for fear it could be used to generate fake but realistic information of all kinds perhaps in ways they haven't yet imagined the same thing happened with images for example a neural network which learned to recognize faces can endlessly generate new images of human faces that don't exist as of today you can no
longer trust a photo or a video in the same way so at first we will experience this new computer revolution both by computers which can trick us in ways like never before but also fill in our blanks in ways people have been dreaming about since the dawn of computers it's the computer science equivalent to the realization that the earth isn't the center of the universe where humans aren't the center or upper limit of intelligence and this advance really breaks down into two key ideas we will explore in this series the first was to let machines
learn knowledge instead of hard-coding it into their instructions known as the machine learning paradigm every time he does something a little closer to what we want him to do we reinforce him twenty minutes and the pigeon has learned affected this and the second idea was to move away from this idea of following a sequence of instructions based on words and move towards using parallel patterns of activity as the unit of computation known as a distributed representation having a symbol only view of the mind was described as the classic mistake by Geoffrey Hinton who is known
as the father of deep learning in fact still most people in AI they made a very naive mistake which is they thought that strings of symbols come in as words when you say something strings of symbols come out so they think what's in between must be something like a string of symbols even though you know that what's in the brain is just big vectors of neural activity there's no symbols in there instead what goes on inside the brain looks like it happens at the level of patterns not symbols these patterns take the form of a
cascade of electrical pulses which we can see are unique to different thoughts and that's not meant to be a joke that's what I believe a thought is a thought is an activity pattern in a big bunch of neurons and only later in life do humans start using symbols to express these thoughts using language working at the level of patterns of activations or feelings first not symbols first this is a biologically inspired view of thought that neural networks exhibit it gives machines what we call a sense of their inputs or a form of intuition but of
course there are many limitations concerns and open questions about learning with neural networks and parallel questions about how the brain functions of which we have a very limited understanding especially when it comes to explaining what exactly makes the human mind unique and so to begin we need to go way back in time and think about what we mean when we say intelligence [Music] you
วิดีโอที่เกี่ยวข้อง
From Bacteria to Humans (Evolution of Learning)
16:58
From Bacteria to Humans (Evolution of Lear...
Art of the Problem
48,398 views
How are memories stored in neural networks? | The Hopfield Network #SoME2
15:14
How are memories stored in neural networks...
Layerwise Lectures
700,043 views
The moment we stopped understanding AI [AlexNet]
17:38
The moment we stopped understanding AI [Al...
Welch Labs
1,150,504 views
"Godfather of AI" Geoffrey Hinton: The 60 Minutes Interview
13:12
"Godfather of AI" Geoffrey Hinton: The 60 ...
60 Minutes
2,003,376 views
The Most Important Algorithm in Machine Learning
40:08
The Most Important Algorithm in Machine Le...
Artem Kirsanov
453,677 views
Generative Model That Won 2024 Nobel Prize
33:04
Generative Model That Won 2024 Nobel Prize
Artem Kirsanov
70,693 views
Physics Informed Neural Networks (PINNs) [Physics Informed Machine Learning]
34:32
Physics Informed Neural Networks (PINNs) [...
Steve Brunton
63,063 views
How AI Learns (Backpropagation)
15:45
How AI Learns (Backpropagation)
Art of the Problem
46,681 views
Transformer Neural Networks Derived from Scratch
18:08
Transformer Neural Networks Derived from S...
Algorithmic Simplicity
142,680 views
AI can't cross this line and we don't know why.
24:07
AI can't cross this line and we don't know...
Welch Labs
971,691 views
How large language models work, a visual intro to transformers | Chapter 5, Deep Learning
27:14
How large language models work, a visual i...
3Blue1Brown
3,175,787 views
Watching Neural Networks Learn
25:28
Watching Neural Networks Learn
Emergent Garden
1,339,312 views
Why Computer Vision Is a Hard Problem for AI
8:39
Why Computer Vision Is a Hard Problem for AI
Quanta Magazine
133,135 views
But what is a neural network? | Chapter 1, Deep learning
18:40
But what is a neural network? | Chapter 1,...
3Blue1Brown
17,432,068 views
WE MUST ADD STRUCTURE TO DEEP LEARNING BECAUSE...
1:49:11
WE MUST ADD STRUCTURE TO DEEP LEARNING BEC...
Machine Learning Street Talk
88,173 views
AI Deception: How Tech Companies Are Fooling Us
18:59
AI Deception: How Tech Companies Are Fooli...
ColdFusion
1,936,074 views
Just Happened! NEURALINK Second Human Trials! Elon Musk Reveals NEW Updates
19:12
Just Happened! NEURALINK Second Human Tria...
TESLA CAR WORLD
564,283 views
How AI Learns Concepts
14:22
How AI Learns Concepts
Art of the Problem
174,658 views
Can We Build an Artificial Hippocampus?
23:51
Can We Build an Artificial Hippocampus?
Artem Kirsanov
207,894 views
Neural Network Simply Explained - Deep Learning for Beginners
6:38
Neural Network Simply Explained - Deep Lea...
Python Simplified
56,712 views