Ravid Shwartz-Ziv (Assistant Professor) 00:13.690
hi anne and welcome to the information bottleneck and i have to say this is a bit weird for me like i've known you for almost five years and we've worked closely together but this is the first time that i'm interviewing you for a podcast right usually our conversations are more
Ravid Shwartz-Ziv (Assistant Professor) 00:33.610
like yan it doesn't work what they should do OK so even though i'm sure all of our audience knows you i will say yana kun is a turing award winner one of the godfathers of deep learning the inventors of convolutional neural networks founder of metals fundamental AI research lab
Ravid Shwartz-Ziv (Assistant Professor) 00:59.790
and still their chief AI scientist and the professor at NYU so welcome
Yann LeCun (Chief AI Scientist) 01:08.520
pleasure to be here yeah
Ravid Shwartz-Ziv (Assistant Professor) 01:11.680
and it's a pleasure for me to be anywhere near you i have been you know in this industry for a lot less time than either one of you doing research for a lot less time so the fact that i'm able to publish papers somewhat regularly with ravid has been an honor and to be able to
Ravid Shwartz-Ziv (Assistant Professor) 01:33.270
start hosting this podcast has been even more of one so it's really a pleasure to sit down with you
Yann LeCun (Chief AI Scientist) 01:39.960
yeah
Ravid Shwartz-Ziv (Assistant Professor) 01:40.280
so we'll try like congratulations on the new startup you recently announced that after twelve years at meta you're starting a new startup advanced machine intelligence that you focus on world model and so first of all how does it feel to be in the in the other side going from a
Ravid Shwartz-Ziv (Assistant Professor) 02:02.630
big company to starting something from
Yann LeCun (Chief AI Scientist) 02:06.070
scratch well i co founded companies before i was you know involved more peripherally than than this new one but you know i know i know how this works what's unique about this one is a new phenomenon where there is enough hope from the part of investors that you know AI will have
Yann LeCun (Chief AI Scientist) 02:28.990
a big impact that they are ready to invest a lot of money essentially which means now you can create a startup where you know the first couple of years are essentially focused on research that just was not possible before like you know the only place to do research in industry
Yann LeCun (Chief AI Scientist) 02:45.830
before was in a large company that was you know not fighting for its survival and basically had a dominant position market and had a you know long enough view that they were willing to to fund long term projects so from you know history the the big labs that we remember like
Yann LeCun (Chief AI Scientist) 03:09.190
bell labs belong to AT and T which basically had a monopoly yield telecommunication in the US you know IBM had a monopoly on big computers essentially right and they had a good research lab xerox as a monopoly on photocopiers and that enabled them to park did not enable them to
Yann LeCun (Chief AI Scientist) 03:26.430
profit from the research going on there but that profited apple microsoft research google research and fair at meta and the industry is shifting again fair had a big influence on AI the AI research ecosystem by essentially being very open right publishing everything open
Yann LeCun (Chief AI Scientist) 03:54.190
sourcing everything with and with tools like pythorch but also like research prototypes that a lot of people have been using in industry so we caused other labs like google to become more open and other labs to also kind of publish much more systematically than before but what's
Yann LeCun (Chief AI Scientist) 04:12.030
been happening over the last couple of years is that rosso 's labs have been kind of climbing up and becoming more secretive and that's certainly the case i mean that was the case for a penny i several years ago and and now google is becoming more closed and possibly even meta
Yann LeCun (Chief AI Scientist) 04:35.200
so yeah i mean it was it was time for the type of stuff that i'm interested in to kind of do it outside so
Ravid Shwartz-Ziv (Assistant Professor) 04:46.510
to be clear then does ami advanced machine intelligence plan to do their research in the open
Yann LeCun (Chief AI Scientist) 04:56.080
yeah upstream research i mean in my opinion you cannot really call it research unless you publish what you do because otherwise you can get easily fooled by yourself you know you you come up with something you think it's the best thing since sliced bread OK if you don't actually
Yann LeCun (Chief AI Scientist) 05:14.510
submit it to the rest of the community you might just be delusional and i've seen that phenomenon many times you know in lots of industry research lab where there's sort of internal hype about you know some internal projects without kind of realizing that other people are doing
Yann LeCun (Chief AI Scientist) 05:33.360
things that actually are better right so if you if you tell the scientists like you know publish your your work first of all that is an incentive for them to do better work that is more you know whether methodology is kind of more thorough and the results are kind of more
Yann LeCun (Chief AI Scientist) 05:51.710
reliable the research is more reliable it's good for them because very often when you work on a research project the impact you may have on product could be months years or decades down the line and you cannot tell people like you know come work for us don't say what you're
Yann LeCun (Chief AI Scientist) 06:11.070
working on and maybe there is a product you will have an impact on five years from now like in the in the meantime like they can't be motivated to really do something useful so if you tell them that they they tend to work on things that have a short term impact right so if you
Yann LeCun (Chief AI Scientist) 06:26.430
really want breakthroughs you need to let people publish you can't do it any other way and this is something that a lot of the industry is forgetting at the moment
Ravid Shwartz-Ziv (Assistant Professor) 06:35.310
does AMI like what products if any does AMI plan to to produce or make is it research or more than that
Yann LeCun (Chief AI Scientist) 06:44.350
no it's more than that it's actual products OK but you know what things i have to do with with you know what models and you know planning and and basically we have the ambition of becoming kind of one of the main suppliers of intelligent systems down the line we think the the
Yann LeCun (Chief AI Scientist) 07:03.990
current architectures that are employed you know at adams or you know agent systems that are based on NLM's umm work OK for language even agenting systems really don't work very well they require a lot of data to basically clone the behavior of humans and they're not that
Yann LeCun (Chief AI Scientist) 07:21.750
reliable so we think the proper way to handle this and i've been saying this for almost ten years now is have role models that are capable of predicting what would be the consequence or the consequences of an action or a sequence of actions that an AI system might take and then
Yann LeCun (Chief AI Scientist) 07:42.510
the system arrives at a sequence of actions or an output by optimization by figuring out what sequence of actions will optimally accomplish a test you know setting for myself that's planning OK so i think in the central part of intelligence is being able to predict the
Yann LeCun (Chief AI Scientist) 08:02.040
consequences of your actions and then use them for planning and that's what we're that's what i've been working on for many years we've been making fast progress with you know a combination of projects here at NYU and also at meta and now it's time to basically make it make it
Yann LeCun (Chief AI Scientist) 08:21.990
real
Ravid Shwartz-Ziv (Assistant Professor) 08:24.160
and what do you think are the missing parts like and why you think it's taking so long because you're talking about it as you said like for many years already but it's still not better than LLM right it's
Yann LeCun (Chief AI Scientist) 08:38.150
not the same thing as LLM right it's designed to handle modalities that are high dimensional continuous and noisy and LNM's completely suck at this like they really do not work right if you try to train an LLM to kind of learn good representations of images or video they're
Yann LeCun (Chief AI Scientist) 08:57.040
really not that great you know generally vision capabilities for for AI AI systems right are trained separately they're not part of the whole LLM thing so yeah if you want to handle data that is high dimensional continuous and noisy you cannot use generative models you can
Yann LeCun (Chief AI Scientist) 09:21.360
certainly not use generative models that tokenize your data into kind of discrete symbols OK it's just no way and we have a lot of empirical evidence that this simply doesn't work very well what does work is learning an abstract representation space that illuminates a lot of
Yann LeCun (Chief AI Scientist) 09:37.670
details about the input essentially all these details that are not predictable which includes noise and make predictions in that representation space and this is the idea of jetpack johnson then in predictive architectures which you know you are as familiar to yeah
Ravid Shwartz-Ziv (Assistant Professor) 09:54.430
with sorry because you worked on this yeah so also randall was a hosted in the past in in the podcast i probably talked about this at length
Yann LeCun (Chief AI Scientist) 10:06.310
so so there's a lot of ideas around this and let me tell you my history around this OK i have been convinced for a long time probably the better part of twenty years that the the proper way to building intelligent systems was through some form of unsupervised learning i started
Yann LeCun (Chief AI Scientist) 10:26.710
working on unsupervised learning as the basis for you know making progress in the early two thousands mid two thousands before that i wasn't so convinced this was the way to go and and basically this was the idea of you know training auto encoders to learn representations right
Yann LeCun (Chief AI Scientist) 10:47.430
so you have an input you run into an encoder it finds a representation of it and then you decode so you guarantee that the representation contains all the information about the input does that that iteration is long like insisting that the representation contains all the
Yann LeCun (Chief AI Scientist) 11:01.910
information about the input is a bad idea OK i didn't know this at the time so what we worked on was you have several ways of doing this you know jeff hinton at the time was working on restricted boss machines joshua benji was working on the noisy autoeurs which actually became
Yann LeCun (Chief AI Scientist) 11:18.480
quite successful in different contexts right for NLP among others and i was working on sparse auto encoders so basically you know if you train on auto encoder you need to recognize the the representation so that the autoencoder does not trivially learn an identity function and
Yann LeCun (Chief AI Scientist) 11:33.550
this is the information bottleneck podcast listens about information bottleneck right you need to create an information bottleneck to limit the information content of the representation and i thought high dimensional sparse representations was actually a good way to go so so a
Yann LeCun (Chief AI Scientist) 11:50.190
bunch of my students did their PH D on this corelu who is not a chief AI architect at alphabet and also the CTO actually did this finished here on this with me and you know a few a few other a few other false macro on zetto and eden bro and a few others so this was kind of the
Yann LeCun (Chief AI Scientist) 12:12.080
idea and then as it turned out and the idea the reason why we worked on this was because we wanted to pre train very deep neural nets by pre training in those things that auto encoders we thought that was the way to go what happened though was that we started like you know
Yann LeCun (Chief AI Scientist) 12:30.070
experimenting with things like normalization rectification instead of hyperbole tangential to sigmoid like radius that ended up you know basically allowing us to train fairly deep network completely supervised so self supervised learning and this was at the same time that data
Yann LeCun (Chief AI Scientist) 12:51.790
set started to get bigger and so it turned out like you know supervisor it worked fine so the whole idea of self supervised or unsupervised learning was put put aside and then came resnet and you know that's a completely solved the problem of training very deep architecture
Yann LeCun (Chief AI Scientist) 13:07.400
that's right in twenty fifteen but then in twenty fifteen i started you know thinking again about like how do we push towards like human level AI which really was the original objective of fair really and my objective my life 's you know mission and realized that you know all
Yann LeCun (Chief AI Scientist) 13:26.110
the approaches of reinforcement learning and things of that type were basically not scaling you know reinforcement learning is incredibly inefficient in terms of samples and so this is not the way to go and and so the idea of world models right the system that can predict the
Yann LeCun (Chief AI Scientist) 13:44.400
consequence consequences of its action they can plan i started researching playing with this around twenty fifteen sixteen my keynote at what was still called nips at the time in twenty sixteen was on role model i was arguing for it those basically the centerpiece of my talk was
Yann LeCun (Chief AI Scientist) 14:03.590
like this is what we should be working on like you know grandma was action condition and if your residents are working on this on video prediction and things like that we had some papers on video prediction in twenty sixteen and i made a the same mistake as before and the same
Yann LeCun (Chief AI Scientist) 14:22.230
mistake that everybody is doing at the moment which is training a video prediction system to predict at the pixel level which is really impossible and you can't really represent useful probability distributions on the space of video frames those things don't work i knew for a
Yann LeCun (Chief AI Scientist) 14:42.990
fact that because the prediction was nondeterministic we had to have a model with latent variables to represent all the stuff you don't know about the variable you're supposed to predict and so we experimented with this for years i had a student here who is now a scientist at
Yann LeCun (Chief AI Scientist) 15:00.190
fair michael enough who developed a video prediction system with latent variables and he kind of solved this problems we're facing slightly i mean today the solution that a lot of people are employing is diffusion models which is a way to train in non deterministic function
Yann LeCun (Chief AI Scientist) 15:19.710
essentially or energy based models which have been advocating for decades now which also is another way of training non deterministic functions but in the end i discovered that this was all about the idea that the really the way to get around the fact that you can't predict that
Yann LeCun (Chief AI Scientist) 15:38.240
the pixel level is to just not predict the pixel level is to another representation and predict that the representation level eliminating all the details you cannot predict and i wasn't really thinking about those
Yann LeCun (Chief AI Scientist) 15:54.000
methods early on because i thought there was a huge problem of preventing collapse so i'm sure randall talked about this but you know when you train let's say you have an observed variable X and you're trying to predict a variable Y but you don't want to predict all the details
Yann LeCun (Chief AI Scientist) 16:11.800
right to run X and Y to encoders so now you have both a representation for's for X SX and representation for Y S Y you can train a predictor to produce you know predict the representation of Y from the representation of X but if you want to train this whole thing end to end
Yann LeCun (Chief AI Scientist) 16:28.870
simultaneously this is a trivial solution where the system ignores the input and produces constant representations and the predictors probably know is trivial right so if you're on the criterion to train the system is minimized dictionary it's not going to work it's going to
Yann LeCun (Chief AI Scientist) 16:45.670
collapse i knew about this problem for a very long time because i worked on joint embedding architectures we used to call them siamese networks back in the nineties
Ravid Shwartz-Ziv (Assistant Professor) 16:54.110
those are the same because people have been using that term sign means networks even recently that's right
Yann LeCun (Chief AI Scientist) 17:00.270
i mean the concept is still you know up to date right so you have you have an X and Y then think of the X as some sort of degraded transformed or corrupted version of Y OK you run both X and Y two encoders and you tell the system look X and Y really are two views of the same
Yann LeCun (Chief AI Scientist) 17:18.470
thing presentation you compute should be the same right so if you just train a neural net you know two neural nets with shared weights right to produce the same representation for slightly different versions of the same object view whatever it is it collapses it doesn't produce
Yann LeCun (Chief AI Scientist) 17:39.030
anything useful so you have to find a way to make sure that the system you know extract as much information from the input as possible and the original idea that we had you know it was a newspaper from nineteen ninety three with simmons net was to have a contrastive term right
Yann LeCun (Chief AI Scientist) 17:55.190
so you have other pairs of samples that you know are different and you train the system to produce different representations so you have a cost function that attracts the two representations when you show two examples that are identified or similar and you repel them when you
Yann LeCun (Chief AI Scientist) 18:10.390
show it two examples that are dissimilar and we came up with this idea because someone came to us and said like can you encode signatures of someone you know drawing a signature on the tablet can you encode this on less than eighty bytes because if you can encode it in less than
Yann LeCun (Chief AI Scientist) 18:27.790
eighty bytes we can write it on the magnetic tape of a credit card so we can do signature application for credit cards right and so we give up this idea i came up with this idea of training a neural net to produce ad variables that will quantize one by each and then training
Yann LeCun (Chief AI Scientist) 18:48.600
training it to kind of do this thing
Ravid Shwartz-Ziv (Assistant Professor) 18:51.110
and did they use it
Yann LeCun (Chief AI Scientist) 18:52.630
so it worked really well and they showed it to their you know business people who said oh we're just going to ask people to type pink codes we have every
Ravid Shwartz-Ziv (Assistant Professor) 19:03.080
lesson of like that like how you can integrate the technology right
Yann LeCun (Chief AI Scientist) 19:08.280
and you know i knew this thing was kind of fishy in the first place because like you know there were countries in europe that were using smart cards right and it was much better but they just didn't want to use smart crops for some reason anyway so so we had this technology in
Yann LeCun (Chief AI Scientist) 19:24.550
the mid two thousand i worked with two of my students on to revise this idea we came up with a new objective functions to train those so these are where people now call contractive methods it's a special case of contrastive methods we have like positive examples negative
Yann LeCun (Chief AI Scientist) 19:37.790
examples and you train you know on positive examples you train the system to have low energy and for negative samples you train them to have higher energy where energy is the distance between the representations so we had two papers at CDPR in two thousand five two thousand six
Yann LeCun (Chief AI Scientist) 19:52.790
by raya hadsel who is now the head of deep mind foundation the sort of fair like division of deep mind if you want and summit chopra who is actually a faculty here at NYU now working on medical imaging and so this gathered a bit of interest in the community and sort of revived a
Yann LeCun (Chief AI Scientist) 20:16.150
little bit of work on those ideas but it still wasn't working very well those contrasting methods really were producing representations of images for example that were kind of relatively low dimensional if we measured like those the eigenvalue spectrum of the coherence matrix of
Yann LeCun (Chief AI Scientist) 20:32.590
the representations that came out of those things it would fill up maybe two hundred dimensions never more like even training on imagenet and things like that even with data augmentation and so that was kind of disappointing and it did work OK there was a bunch of papers on this
Yann LeCun (Chief AI Scientist) 20:47.120
and it worked OK there was there was white paper from deeper it seemed clear that demonstrated you could get decent performance with contrastive training applied to siamese nets but then about five years ago one of my postdocs stephane denis at meta tried an idea that at first i
Yann LeCun (Chief AI Scientist) 21:13.390
didn't think it would work which was to essentially have some measure of information quantity that comes out of the encoder and then trying to maximize that OK and the reason i didn't think it would work is because i'd seen a lot of experiments along those lines that jeff hinton
Yann LeCun (Chief AI Scientist) 21:32.350
was doing in the nineteen eighties are trying to maximize information and you can never maximize information because you never have appropriate measures of information content that is at orbound if you want to maximize something you want to either be able to compute it or you
Yann LeCun (Chief AI Scientist) 21:50.890
want to a lower bound on it so you can push it up right and for information content we only have upper balance so i always thought this was completely hopeless and then you know stefan came up with a technique which was was called battle twins baru is a famous theoretical
Yann LeCun (Chief AI Scientist) 22:10.800
neuroscientist who came up with the idea information maximization and and it kind of worked it was wow so there i said like we have to push this right so we came up with another method with a student of mine adrian bard co advised with jean-ponse who's affiliated with NYU two
Yann LeCun (Chief AI Scientist) 22:32.720
technique called vikrag variance invariance covariance regularization and that's a network to be simpler and work even better and since then we made progress and randall recently you know discussed an idea with him that can be pushed and made practical it's called sigreg the
Yann LeCun (Chief AI Scientist) 22:51.350
whole system is called he's responsible for the name at all latent euclidean japa right yeah and sigreg has to do with sort of making sure that there are distribution of vectors that come out of the encoder is an isotropic gaussian that's the i in the G so i mean there's a lot
Yann LeCun (Chief AI Scientist) 23:21.110
of things happening in this domain which are really cool i think there's going to be some more progress over the next year or two and we'll get a lot of experience with this and and i think that's kind of a really good promising set of techniques to train models that learn
Yann LeCun (Chief AI Scientist) 23:37.710
abstract representations which i think is key and what do you think are the missing party like do you think like more compute will help or like we need better algorithms like it's kind of like do you believe in the bitter lessons right like do you think well and and furthermore
Yann LeCun (Chief AI Scientist) 23:55.550
what do you think about you know the data quality problems with the internet post twenty twenty two right i've heard people compare it to low background steal now to refer to all that data before LLM's came out like low background tokens i mean OK yeah i think i'm totally
Yann LeCun (Chief AI Scientist) 24:10.510
escaping that problem OK here is the thing and i've i've been you know using this argument publicly over the last couple of years training an LLM if you wanted to have any kind of you know decent performance requires training on basically all the available freely available text
Yann LeCun (Chief AI Scientist) 24:26.670
on the internet plus some you know synthetic data plus licensed data et cetera so a typical LLM like you know number three you know going back a year or two is trained on thirty trillion tokens the token is typically three bytes so that's ten to the fourteen bytes for pre
Yann LeCun (Chief AI Scientist) 24:43.790
training OK we're not talking about fine tuning ten to the fourteen bytes and for the LLM's to be able to really kind of exploit this they need to have a lot of memory storage because basically those are isolated facts there is a little bit of redundancy in text but a lot of it
Yann LeCun (Chief AI Scientist) 25:05.620
is just isolated facts right and so you need a lot of you need very big networks because you need a lot of memory to store all those facts if we go to them OK now compare this with video ten to the fourteen bytes if you count two megabytes per second for video for you know
Yann LeCun (Chief AI Scientist) 25:28.960
relatively compressed video not how you compress but a bit that would represent fifteen thousand hours of video ten to the fourteen bytes if fifteen thousand hours of video you have the same amount of data as the charity of all the texts available on the internet now fifteen
Yann LeCun (Chief AI Scientist) 25:46.870
thousand hours of video is absolutely nothing it's thirty minutes of youtube uploads OK it's the amount of visual information that a four year old has seen in his or her life the entire life waking time is about sixteen thousand hours in four years so a lot of information we
Yann LeCun (Chief AI Scientist) 26:06.990
have video models now vijay vijaypat two actually that just came out last summer that was trained on the equivalent of a century of video data and it's all public data OK much more data but much less than the biggest LNM actually because even though it's it's more bytes it's
Yann LeCun (Chief AI Scientist) 26:29.790
more redundant to say OK it's more redundant so it's less useful actually when you use self supervised learning you do need redundancy you cannot learn anything in self supervise or anything by the way if it's completely random redundancy is what you can learn and so so this
Yann LeCun (Chief AI Scientist) 26:46.790
just much richer structure in you know real world data like video than there is in text which kind of led me to claim that we absolutely never ever going to get to human level AI by just training on text it's just never going to happen right it's a big debate in philosophy of
Yann LeCun (Chief AI Scientist) 27:05.750
whether AI should be grounded in reality or whether it could be just you know in the realm of symbolic manipulation and things like this we talk about for world models and grounding i think you know there's still a lot of people who don't even understand what the idealized world
Yann LeCun (Chief AI Scientist) 27:22.350
model is in a sense right so for example i'm influenced by having watched star trek which i would hope you've seen a little bit of and thinking of the holodecks right i always thought that the holodeck was like an idealized perfect world model right even so many episodes of it
Yann LeCun (Chief AI Scientist) 27:37.950
going too far right and people walking out of it right but you know it even simulates things like smell and physical touch so do you you think that something like that is like the idealized world model or do you think like a different model or like way of defining it would be OK
Yann LeCun (Chief AI Scientist) 27:53.390
this is an excellent question and there is an excellent is because it goes to the core of really what you know what i think we should be doing which i'm doing and how wrong a secret video is OK so so people think you know think that a world model is something that reproduces all
Yann LeCun (Chief AI Scientist) 28:14.910
details of what the world does they think of it as a simulator yeah right and of course because you know deep learning is the thing you're going to use some deep learning system as a simulator a lot of people also are focused on video generation which is kind of a cool thing
Yann LeCun (Chief AI Scientist) 28:29.070
right you you produce those cool videos and they're wow you know people are sort of really impressed by them now there's no guarantee whatsoever then when you train a video generation system it actually has an accurate model of the underlying dynamics of the world and it's
Yann LeCun (Chief AI Scientist) 28:44.070
learned anything you know particularly abstract about it and so so the idea that somehow a model needs to reproduce every detail of the reality is wrong and hurtful and i'm going to tell you why OK a good example of simulation is CFD computational fluid dynamics it's used all
Yann LeCun (Chief AI Scientist) 29:10.550
the time people use supercomputers for that right so you want to simulate the flow of air around an airplane you cut up the space into little cubes and within each cube you have a small vector that represents the state of that cube which is you know velocity density or mass and
Yann LeCun (Chief AI Scientist) 29:32.670
temperature and maybe a couple of other things right so and then you solve navier stokes equations which are which is a differential partial differential equation and you can see related flow of air now the thing is this does not actually necessarily solve the equations very
Yann LeCun (Chief AI Scientist) 29:51.590
accurately if you have chaotic behavior like turbulences and stuff like that simulation is only you know approximately correct but in fact that's already an abstract representation of the underlying phenomenon the underlying phenomenon is molecules of air that bump into each
Yann LeCun (Chief AI Scientist) 30:07.590
other and bump on the wing and on the airplane right ever goes to that level to do the simulation that would be crazy right it would require an amount of computation that's just insane and it would depend on the initial condition i mean there's all kinds of reasons we don't do
Yann LeCun (Chief AI Scientist) 30:24.990
this and maybe it's not molecules maybe it's you know at a lower level we should simulate particles and like you know do the feynman diagrams and simulate you know all the different paths that those particles are employing because they don't take one path right it's not
Yann LeCun (Chief AI Scientist) 30:38.190
classical it's quantum so at the bottom it's like quantum field theory and probably already that that is an abstract representation so so you know everything that takes place between us at the moment in principle can be described through quantum field theory OK we just have to
Yann LeCun (Chief AI Scientist) 30:58.200
measure the wave function of the universe in a cube that contains all of us and even that would not be sufficient because they're entering all particles on the other side of the universe that you know we have so it wouldn't be sufficient but let's imagine OK for the sake of of
Yann LeCun (Chief AI Scientist) 31:14.790
the argument first of all we would not be able to measure this wave function and second of all the amount of competition we would need to devote to this is absolutely gigantic it was released on gigantic quantum computer that you know is the size of euros or something so no way
Yann LeCun (Chief AI Scientist) 31:35.150
we can describe anything at that level and it's very likely that our simulation would be accurate for maybe a few nanoseconds you know beyond that we'll diverge from reality so what do we do we invent abstractions we invent abstractions like particles atoms molecules in the
Yann LeCun (Chief AI Scientist) 31:53.550
living world its proteins organelles sales organs organisms societies ecosystems etc right and basically every level in this hierarchy ignores a lot of details about the level below and what that allows us to do is make longer term more reliable longer term predictions OK so we
Yann LeCun (Chief AI Scientist) 32:19.830
can describe the dynamics between us now in terms of the underlying science and in terms of psychology OK that's a much much higher level abstraction than particle physics right and in fact you know every level in the hierarchy i just i just mentioned is a different field of
Yann LeCun (Chief AI Scientist) 32:36.990
science a field of science is essentially defined by the level of abstraction at which you start making predictions right that you allow yourself to use to make predictions in fact physicists have this down to an art in the sense that you know if i give you a box full of gas you
Yann LeCun (Chief AI Scientist) 32:59.190
could in principle simulate all the molecules of the gas but nobody ever does this but at a very abstract level we can say you know PV equals NRT right you know pressure times value equals the number of particle times you know temperature blah blah blah and so you know that you
Yann LeCun (Chief AI Scientist) 33:20.870
know global emergent phenomenological level if you increase the pressure the temperature will go up or if you increase the temperature the pressure will go up right or if you let some particles out then the pressure will go down and blah blah blah right so so we all the time we
Yann LeCun (Chief AI Scientist) 33:38.710
build phenomenological models
Ravid Shwartz-Ziv (Assistant Professor) 33:41.150
of something complicated by ignoring all kinds of details that physicists call entropy but but it's really systematic that's the way we understand the world we
Ravid Shwartz-Ziv (Assistant Professor) 33:54.200
do not memorize every detail of we certainly not reconstruct it of what we perceive so world models don't have to be simulators at all well there are simulators but in abstract representation space and
Yann LeCun (Chief AI Scientist) 34:08.080
what they simulate is only the relevant part of reality OK if i ask you where is jupiter going to be one hundred years from now i mean we have an enormous amount of information about jupiter right but within this whole information that we have about jupiter to be able to make
Yann LeCun (Chief AI Scientist) 34:24.390
that prediction where jupyter is going to be one hundred years from now you need exactly six numbers three positions and three velocities and the rest doesn't matter so you don't believe in a synthetic datasets i do no it's useful you know data from games i mean there's
Yann LeCun (Chief AI Scientist) 34:40.160
certainly a lot of things that you learn from synthetic data from you know from games and things like that i mean you know children learn a huge amount from from play which basically are kind of simulations you know the world a little bit right but but in conditions where they
Yann LeCun (Chief AI Scientist) 34:58.630
can't kill themselves but i worry at least for video games that for example the green screen like actors doing the animations they're doing extremely it's designed to look good you know for like an often badass i guess for an action game but these often don't correspond very
Yann LeCun (Chief AI Scientist) 35:16.990
well to reality and so i i worry that like a physical system that's you know been trained or through or with the assistance of world models might get similar quirks at least in the very short term is this something that worries you no it depends on what level you train them so
Yann LeCun (Chief AI Scientist) 35:31.350
for example i mean sure if you use a very accurate robotic simulator for example right it's going to accurately simulate the dynamics of an arm you know when you apply torques to it it's going to move in a particular way dynamics no problem now simulating the friction that
Yann LeCun (Chief AI Scientist) 35:47.510
happens you know when you grab an object and manipulate it that's super hard to do it accurately friction is very hard to simulate OK and so those simulators are not particularly accurate for manipulation they're good enough that you know you can train a system to do it and then
Yann LeCun (Chief AI Scientist) 36:02.470
you can do you know seem to real with a little bit of adaptation so that can work but it does not i mean the point is much more important like for example there's a lot of completely basic things about the world that we completely take for granted which we can learn at a very
Yann LeCun (Chief AI Scientist) 36:18.150
abstract level but it's not language related OK so the fact for example and i've used this example before and people made fun of me for it but it's really true OK i have those objects on the table and the fact that when i push the table the object moves with it like this is
Yann LeCun (Chief AI Scientist) 36:33.670
something we learned it's not something that you're born with OK the fact that most objects will fall when you let them go right with gravity maybe it's run this around the edge of and the reason people make fun of me with this is because i said you know LLM's don't understand
Yann LeCun (Chief AI Scientist) 36:50.310
this kind of stuff right and and they absolutely do not even today but but you can train them to give the right answer when you ask them a question you know if i put an object on the table then i push the table what will happen to the object it will answer the object moves with
Yann LeCun (Chief AI Scientist) 37:07.670
it but because it's been fine tuned to do that OK so it's more like regurgitation that sort of real understanding of the underlying dynamics but if you look on i don't know sura like nano nano banana they they have a good physics of the world right they are not perfect they have
Yann LeCun (Chief AI Scientist) 37:23.150
some physics yeah they have some physics right so do you think like we can't push it or do you think like it's a one way to learn physics actually make predictions in our presentation space they use diffusion transformers and that prediction that the computation of the video
Yann LeCun (Chief AI Scientist) 37:45.470
snippet at an abstract level is done in representation space OK not always auto regressively by the way sometimes it's just in parallel and then there's a second diffusion model that turns this abstract representations into a nice looking video and that might be more collapse we
Yann LeCun (Chief AI Scientist) 38:03.190
don't know right because we can't really measure like the coverage of such systems with reality but but like you know the to the the previous point i can train like here is another completely obvious concept to us that we don't even imagine that we learn but we do learn it a
Yann LeCun (Chief AI Scientist) 38:27.390
person cannot be in two places at the same time OK we run this because very early on we learn object permanence the fact that when an object disappears still exists OK and reappears it's the same object that you saw before how can we train an AI system to learn this concept so
Yann LeCun (Chief AI Scientist) 38:47.080
object permanence you know you just show it a lot of videos where objects you know go behind the screen and then reappear on the other side or what they go behind the screen and the screen goes away and the object is still there and when you show four months old babies scenarios
Yann LeCun (Chief AI Scientist) 39:01.320
where things like this are violated their eyes open like super big and they're like super surprised because reality just you know violated their internal model the same thing when you show a scenario of like a little car on the platform you push it off the platform and it
Yann LeCun (Chief AI Scientist) 39:18.270
appears to float in the air they also look at it you know nine months ten months old babies look at it like really surprised six months old baby barely pay attention because they haven't run over gravity yet so they haven't been able to like you know incorporate the notion every
Yann LeCun (Chief AI Scientist) 39:33.720
object is supposed to fall so this kind of learning is really what's what's important and you do this you can learn this from very abstract things you know the same way babies learn about like you know social interactions by you know being told stories with like simple pictures
Yann LeCun (Chief AI Scientist) 39:51.590
it's a simulation an abstract simulation of the world but it sort of launched them you know particular behavior so you could imagine like training a system from let's say an adventure game like a top down to the adventure game where you know you you tell your character like you
Yann LeCun (Chief AI Scientist) 40:07.430
know move north and he goes to the other room and it's not in the first room anymore because it moved to the other room right now of course in adventure games you have gandalf that you can call and it just appears right so that's not physical but but like when you pick up a key
Yann LeCun (Chief AI Scientist) 40:21.470
from a you know from a treasure chest you have the key no one else can have it and you can use it to open a door like there's a lot of things that you learn that are very basic you know even in sort of abstract environments yeah and i just want to observe that some of those
Yann LeCun (Chief AI Scientist) 40:38.110
adventure games that they try to train models and one of them you might know about is nethack right sure and nethack is fascinating because it is an extraordinarily hard game like ever ascending in that game without cheats is like twenty years without you know going to the wiki
Yann LeCun (Chief AI Scientist) 40:55.510
people still don't do it from playing and my understanding is that AI agents the very best agent models we have or even world models are pathetic absolutely yeah yeah so to the point people have come up with sort of you know dumbed down version of net house mini hack mini hack
Yann LeCun (Chief AI Scientist) 41:13.790
exactly
Ravid Shwartz-Ziv (Assistant Professor) 41:14.430
mini hack they had to dumb it down just for for AI so some of my you know colleagues have been working with is actually one of my master students so and and you know michael enough who i mentioned earlier has been also doing some work there now what's interesting there is that
Ravid Shwartz-Ziv (Assistant Professor) 41:33.400
there is a type of situations like this where you need to plan OK but you need to plan in the process of uncertainty the problem with you know all games and adventure games in particular is that you don't have complete visibility of the state of the system you don't have the map
Ravid Shwartz-Ziv (Assistant Professor) 41:45.910
in advance you need to explore and blah blah blah you can get killed every time you do this and you know but the actions are essentially discreet yes correct possible actions is turn based and so in that sense it's like chess acceptance is not you know chess is fully observable
Ravid Shwartz-Ziv (Assistant Professor) 42:04.070
go also it's really observable stratego isn't that stratego isn't you know poker is not and so it makes it more difficult if you have uncertainty of course but those are games where the number of actions you can take is discrete and basically you know what you need to do is do
Ravid Shwartz-Ziv (Assistant Professor) 42:29.070
tree exploration OK and to do of course the tree of possible states you know goes exponentially with the number of moves and so you have to have some way of generating only the moves that are likely to be good and basically never generate the other ones or select them down and
Ravid Shwartz-Ziv (Assistant Professor) 42:47.550
you need to have a value function which is something that tells you OK i can't plan to the end of the game but even though i'm planning sort of nine moves ahead i have some way of estimating whether evaluating whether a position is good or bad it's going to lead me to you know
Ravid Shwartz-Ziv (Assistant Professor) 43:02.190
victory or solution right so you need those two components basically something that guesses what the good moves are and then something that you know essentially evaluates ends and if you have those both of those things you can train those functions using something like
Ravid Shwartz-Ziv (Assistant Professor) 43:18.350
reinforcement learning or behavioral cloning if you have data i mean the basic idea for this goes back to samuel 's checker players from nineteen sixty four it's not recent but but of course was you know the power of it was demonstrated with you know alphago and alpha zero and
Ravid Shwartz-Ziv (Assistant Professor) 43:37.110
things like that so that's good but that's a domain where humans suck humans are terrible at playing chess right they're playing go like machines are much better than we are because of the speed of tree exploration and because of the memory that's required for for tree
Ravid Shwartz-Ziv (Assistant Professor) 43:53.870
exploration we just don't have enough memory capacity to do breadth first tree exploration so we suck at it like you know when alphago came out you know people before that thought that the best human players were maybe two or three stones handicapped like below an ideal player
Ravid Shwartz-Ziv (Assistant Professor) 44:10.950
that they call god no like you know humans are terrible like we you know the best players in the world need like eight or nine stories well i i can't believe i i get the pleasure to talk about game AI with with yan i just have a few follow up questions on this the first one is
Ravid Shwartz-Ziv (Assistant Professor) 44:29.030
this example that you talk about around humans being terrible at chess and i'm familiar a bit with the development of chess AI over the years you know i've heard this referred to as more of X paradox and explained as you know humans have evolved over billions or millions sorry
Ravid Shwartz-Ziv (Assistant Professor) 44:46.760
it's large N number of years to physical locomotion and that's why babies and humans are very good at this but we have not evolved at all to play chess so that's one question and then a second question that's related is a lot of people today who play video games and i'm one of
Ravid Shwartz-Ziv (Assistant Professor) 45:02.870
them have observed that it feels like AI at least in terms of like enemy AI has not improved really in twenty years right that some of the best examples are still like halo one and fear from the early two thousands so when do you think that you know advancements that we've been
Ravid Shwartz-Ziv (Assistant Professor) 45:19.230
doing in the lab are going to actually have real impact on like gamers you know and and in a non like generative AI sense yeah i used to be a giver never addictive one but but my family is in it because i have three sons in their thirties and they have a video game design studio
Ravid Shwartz-Ziv (Assistant Professor) 45:37.070
between them so so i was sort of you know embedded in that culture but yeah no you're right and you know it's it's also it's also true that the you know despite the accuracy of physical simulators a lot of the a lot of those simulations are not used by studios who make animated
Ravid Shwartz-Ziv (Assistant Professor) 45:58.720
movies because they want control they don't necessarily want accuracy they want control and in games it's only the same thing it's a creative act and what you want is some control about the course of the story or the way the you know NPC kind of behave and all that stuff right
Ravid Shwartz-Ziv (Assistant Professor) 46:16.710
and the key AI kind of you know is difficult to maintain control at the moment so i mean it will come but you know there's there's some resistance from the creators but i think OK marvik paradox is very much still in force so moravec i think formulated it in nineteen eighty
Ravid Shwartz-Ziv (Assistant Professor) 46:38.030
eight if i don't know correctly and he said like yeah you know things that we think of as uniquely human intellectual tasks like PHS we can do with computers or you know computing integrals or whatever but the thing that we take for granted we don't even think is an intelligent
Ravid Shwartz-Ziv (Assistant Professor) 47:00.770
task like what a cat can do we still can do with robots and even now forty seven years later we still can't do them well i mean of course we can you know train robots you know by imitation and a bit of reinforcement learning and you know by training through simulation to kind of
Ravid Shwartz-Ziv (Assistant Professor) 47:21.870
locomote and you know what obstacles and do various things but then not nearly as inventive and creative and you know agile as a cat it's not because we can't build a robot we certainly can it's just we can't we can't make them smart enough to do all the stuff that a cat or unit
Ravid Shwartz-Ziv (Assistant Professor) 47:42.190
mouse can do let alone the dog or a monkey right so so you have all those people bloviating about like you know AGI in in a year or two
Yann LeCun (Chief AI Scientist) 47:53.800
is completely deluded just completely delusion because the real world is way more complicated are you not going to get it you're not going to get anywhere by tokenizing the world and using NMS it's just not going to happen
? (?) 48:08.000
so so what is your timelines when we will see like intel AGI whatever it means
? (?) 48:15.200
or like and also where are you and where are you on the optimist pessimist sorry because you know there's some doomers among or doomerism amongst like gary marcus and and i think well
Yann LeCun (Chief AI Scientist) 48:26.160
i know gary marcus no
? (?) 48:27.280
he's he's a critique he's critiques it sorry the dumer would be
? (?) 48:30.670
yahshua yeah yeah there you go like where do you fall on all
Yann LeCun (Chief AI Scientist) 48:34.030
these things OK i'll answer the first question first OK so first of all there is no such thing as general intelligence this concept makes absolutely no sense because it's it's really designed to designate human level intelligence the human intelligence is super specialized OK we
Yann LeCun (Chief AI Scientist) 48:57.910
can handle the real world really well like navigate and blah blah blah we can handle other humans really well because we evolved to do this and chest we suck OK so and there's a lot of tests that we suck at that where a lot of other animals are much better than we are OK so what
Yann LeCun (Chief AI Scientist) 49:13.190
that means is that we are specialized we think of ourselves as being general but it's simply an illusion because all of the problems that we can apprehend are the ones that we can think of right and vice versa and so we're in general in all the problems that we can imagine but
Yann LeCun (Chief AI Scientist) 49:32.470
these are a lot of problems that we cannot imagine and there's some mathematical arguments for this which i may not go into unless you ask me but so there is so this this this concept of general intelligence is complete BS we can talk about human level intelligence right so are
Yann LeCun (Chief AI Scientist) 49:53.360
we going to have machines that are as good as humans in all the domains where humans are good or better than humans and the answer is you know we already have machines that are better than humans in some domains like you know we have machines that can translate you know fifteen
Yann LeCun (Chief AI Scientist) 50:06.360
hundred languages into fifteen hundred other languages in any direction no humans can do this right and and you know there's a lot of examples and you know which hasn't go in various other things umm but while we have machines that are as good as humans in all domains the answer
Yann LeCun (Chief AI Scientist) 50:26.340
is absolutely yes there's no question at some point we'll have machines that are as good as humans in all domains
? (?) 50:32.500
OK and that leads but it's not going to be an event it's going to be very progressive we're going to make some conceptual
Yann LeCun (Chief AI Scientist) 50:40.630
advances maybe based on you know jetpack one models planning things like that uh over the next few years and if we're lucky if you don't hit an obstacle that we see uh perhaps this will lead to kind of good paths to human level AI but but perhaps we're still missing a lot of
Yann LeCun (Chief AI Scientist) 50:59.670
basic concepts and so the most optimistic view is that perhaps you know the you know learning goodwill models and and you know being able to do panning and you know understanding complex signals that are continuous high dimensional noise significant progress in that direction
Yann LeCun (Chief AI Scientist) 51:20.110
over the next two years the most optimistic view is that we'll have something that is close to human intelligence or maybe dog intelligence within you know five to ten years OK but that's the most optimistic it's very likely that as what happened you know
? (?) 51:38.310
multiple times in the history of AI in the past there's some obstacle we're not seeing yet which will you know actually kind of require us to invent some new conceptual new things to go beyond in which case that may take twenty years maybe maybe more OK but no question it will
? (?) 51:57.790
happen no question do you think it will be easier to get from the current level to a dog level intelligence compared to a dog to humans levels no i think i think the hardest part is to get to dog level once you get to dog level you basically have most of the ingredients right
? (?) 52:14.190
and then you know what's missing from OK what's missing from like primates to humans beyond just size of the brain is language maybe OK but language is basically handled
Yann LeCun (Chief AI Scientist) 52:25.270
by
? (?) 52:26.510
the vernicki area which is a tiny little piece of brain dust
Yann LeCun (Chief AI Scientist) 52:30.110
right here and the brochure area which is a tiny piece of brain right here both of those evolved in the last you know less than a million years maybe two and it can be that complicated and we already have it at ends that do a pretty good job at you know you know encoding
Yann LeCun (Chief AI Scientist) 52:46.310
language into abstract representations and then decoding thoughts into into text so maybe we'll use LLM for that so LLM will be like the wernicke and brochure areas in our brain what we're working on right now is the prefrontal cortex which is where one model resides well well
Yann LeCun (Chief AI Scientist) 53:03.830
this this gets me into you know a few questions about safety and the destabilizing potential impact so i'll start this with something a little bit funny which is to say if we really get dog level intelligence then the AI of tomorrow has gotten profoundly better than any human at
Yann LeCun (Chief AI Scientist) 53:18.870
smell and and something like that is just you know tip of the iceberg for the the destabilizing impacts of AI tomorrow let alone today i mean we have sam altman talking about super persuasion because AI docs is you so it it figures out who you are through the multi turn so it
Yann LeCun (Chief AI Scientist) 53:35.270
gets really good at kind of customizing its arguments towards you we've had AI psychosis right like people who have done horrible things as a result of kind of believing in a syncophantic AI that that is telling them to do things they shouldn't do happened to be by the way whoa
Yann LeCun (Chief AI Scientist) 53:53.190
you've got to tell us about that too what wendy a few months ago it was it was i didn't want you and i walked down to get lunch and there's a dude who's surrounded by a whole bunch of police officers and and security guys and i work past and the guy recognizes me and says oh
Yann LeCun (Chief AI Scientist) 54:11.070
mister lekun and the police officer kind of whisked me away outside and tells me like you don't want to talk to him turns out the guy had come from you know the midwest by bus here and he he's a kind of emotionally disturbed he you know he had gone through prison blah blah blah
Yann LeCun (Chief AI Scientist) 54:36.950
for kind of various things and he was carrying a bag with you know like a huge wrench and and pepper spray and a knife and so the the security guards got alarmed and basically called the police and then the police realized OK you know this guy is kind of weird so they you know
Yann LeCun (Chief AI Scientist) 54:54.790
took him away and had him examined and eventually he went back to the midwest but i mean he didn't feel threatening to me but the police wasn't so sure so so yeah it happens i had you know high school students writing emails to me saying i read all of those you know piece by
Yann LeCun (Chief AI Scientist) 55:16.880
doomers who said like you know AI is going to take over the world and either kill us all or take our jobs so what i'm totally depressed i don't go to school anymore and so i you know i answer to them said like you know don't i don't believe all that stuff you know but humanity
Yann LeCun (Chief AI Scientist) 55:33.150
is still going to be in control of of all of this now there's there's no question that you know every powerful technology has you know good consequences and bad side effects that sometimes are predicted and corrected sufficiently in advance and sometimes not so much right and
Yann LeCun (Chief AI Scientist) 55:54.040
it's always a trade off that's a history of technological progress right so they stay cars as an example OK cars crash sometimes and initially you know brakes were not that reliable and cars would like flip over and there was no you know seatbelts and blah blah blah right and
Yann LeCun (Chief AI Scientist) 56:12.670
eventually kind of the industry made
? (?) 56:14.110
progress and you know started putting sealbelts and and crumple zone and and and you know automatic kind of controlling systems so that you know the car doesn't go sway and doesn't flip or whatever so cars now are much safer than they used to be there's one thing that is now
? (?) 56:37.180
mandatory in every car sold in the EU and it's actually an AI system that looks out the window it's called it's called AEDS automatic emergency braking system it's basically a common shot
Yann LeCun (Chief AI Scientist) 56:50.380
there right and it looks at the windshield and it detects you know all objects and if it detects on an object is too close it just automatically breaks and if you detect that it's going to be a collision that the driver is not going to be able to avoid it just stops the car or
Yann LeCun (Chief AI Scientist) 57:12.550
sways right and that when statistics i read is that this reduces frontal collisions by forty percent and so it became mandatory equipment in every car sold in the EU even low end because it saves rise so this is AI not killing people saving lives right i mean also same thing for
Yann LeCun (Chief AI Scientist) 57:34.070
like medical imaging and everything there's a lot of live being saved by AI at the moment and like so but do you think so you jeff and joseph right like both of you won the the touring award together and like and you have different opinions about it right and jeff says like he's
Yann LeCun (Chief AI Scientist) 57:57.590
regrets and joshua works on safety and you trying to push it forward and do you think you will get to some some level of intelligent you will say oh this become too dangerous we need to work more on the safety side i mean you have to do it right i'm going to use another example
Yann LeCun (Chief AI Scientist) 58:22.510
jet engines OK i find this astonishing that you can fly halfway around the world on a two engine airplane in complete safety and i really really say halfway around the world like it's a seventeen hour flight OK from new york to singapore right on the airbus three fifty it's
Yann LeCun (Chief AI Scientist) 58:46.790
astonishing and when you look at a jet engine the turbofan it should not work right i mean there is no metal that can stand the type of temperature that takes place there in the kind of like efforts when you have like a huge turbine like you know rotating a two thousand or i
Yann LeCun (Chief AI Scientist) 59:06.230
don't know what speed like the the forest that was on it is just insane you know it's hundreds of tons so it should not be possible yet those things are incredibly reliable so what i'm saying is you can't you know build
? (?) 59:25.910
something like a turbojet the first time you build it it's not going to be safe it's going to run for ten minutes and then blow up OK and it's not going to be fuel efficient and it's you know etc it's not going to be reliable but you know as you make progress in engineering and
? (?) 59:44.030
materials et cetera there's so much you know economic motivation
Yann LeCun (Chief AI Scientist) 59:48.270
to make this good that you know eventually it's going to be the type of reliability we see today the same going to be true for AI we're going to start making systems that you know have agency can plan can reason have role models blah blah blah but we you know they're going to
Yann LeCun (Chief AI Scientist) 60:07.070
have the power of maybe a cat brain right which is about one hundred times smaller than a human brain put guardrails in them to prevent them from doing you know taking actions that are obviously uh dangerous or something you can do this at a very low level like if you have
? (?) 60:27.030
i don't know a domestic robot right that oh so so one example that stuart russell for example have used is is to say well you know if you have a robot the domestic robot and you ask you to fetch you coffee and someone is standing in front of the coffee machine if the system
? (?) 60:44.670
wants to fulfill its goal it's going to have to you know either assassinate or smash the person in front of the coffee machine to get access
Yann LeCun (Chief AI Scientist) 60:53.110
to the coffee machine and obviously you don't want that to happen now it's like the pay per click maximization it's kind of a ridiculous example because it's super easy to fix right you put some guardrail that say well you know you're a domestic robot you should stay away from
Yann LeCun (Chief AI Scientist) 61:08.030
people and maybe ask them to move if they are in a way but not actually kind of you know hurt them in any way or whatever you can do like you know you can put a whole bunch of low level conditions like this like if you have domestic robot and it's you know it's a cooking robot
Yann LeCun (Chief AI Scientist) 61:22.190
right so it has a big knife in its hand and it's you know cutting the cucumber you know don't flare your arms if there are if you're a big knife in your hand and people around OK it can be kind of a low level constraint that the system has to satisfy now some people say oh but
Yann LeCun (Chief AI Scientist) 61:39.590
you know with NLM 's we can fine tune them to not do things that are dangerous but there is always you can you can generate them you can always find prompts where they're going to kind of escape their condition you know the all the things that we stop them from from doing i
Yann LeCun (Chief AI Scientist) 61:56.870
agree that's why i'm saying we shouldn't use LMS we should use those objective driven AI architectures i was talking about earlier where you have a system that has a word model can predict the consequences of its action and can figure out a sequence of actions to accomplish a
Yann LeCun (Chief AI Scientist) 62:12.070
task but also is subject to a bunch of constraints that guarantee that whatever action is being pursued and whatever state of the world is being predicted does not endanger anybody or does not have you know when they get negative side effects right so there is it's by
Yann LeCun (Chief AI Scientist) 62:30.230
construction the system is intrinsically safe because it has all those guardrails and because it obtains its output by optimization by minimizing the objective of the task and satisfying the constraints of the guardrails it cannot escape that it's not a fine tuning right it's by
Yann LeCun (Chief AI Scientist) 62:50.310
construction yeah and and i'll there's a technique you know that that for LLM's for constraining the output space where you say that you ban all outputs except whatever you want like maybe zero to ten and everything else and they have that even for diffusion models sure do you
Yann LeCun (Chief AI Scientist) 63:08.790
think that tactics like that as they exist today significantly improve the utility of those kinds of models well they do but they're ridiculously expensive because the the way they work is that you have to have the system generate lots of proposals for an output and then have a
Yann LeCun (Chief AI Scientist) 63:23.560
shelter that says well this one is good this one is terrible etc i'll rank them and then just put out the one that has the the less toxic rating essentially so it's it's insanely expensive right so unless you have you know some sort of objective driven value function
? (?) 63:41.390
that kind of drives the system towards producing those high you know high score outputs low toxicity outputs it's going to be it's going
Yann LeCun (Chief AI Scientist) 63:51.150
to be expensive yeah and
? (?) 63:53.270
i want to change the topic just a tiny bit off we've been very technical for a moment but we you know i think our audience in the world
? (?) 64:01.310
has a few questions that are maybe a little bit more more social related you know the person who appears to be trying to fill your shoes in
Yann LeCun (Chief AI Scientist) 64:10.030
at meta alex wang where i'm curious as to do you have any thoughts or
Yann LeCun (Chief AI Scientist) 64:15.070
anything about you know kind of how how that will play out for for meta is not is not in my shoes at all he's he's he's in charge of all the R and D and product that are AI related at beta so it's not a researcher or scientist or or anything like that it's more kind of you know
Yann LeCun (Chief AI Scientist) 64:37.110
overseeing the entire operation so within meta superintelligence lab which is his organization kind of divisions if you want so one of them is fair which is long term research one of them is TBD lab which is basically building frontier models which is mostly entirely LLM focused
Yann LeCun (Chief AI Scientist) 65:04.820
a fourth organization is AI infrastructure software infrastructure hardware is some other organization one is products OK so people who take the frontier models and then turn them into actual chat bots that people can use and you know disseminate them and you know plug them into
Yann LeCun (Chief AI Scientist) 65:22.710
whatsapp and everything else right so so those are four divisions he overseas all of that so and there are several AI scientists there is AI scientists are fair that's me and i really have a long term view and basically you know i'm going to be at meta for another you know three
Yann LeCun (Chief AI Scientist) 65:43.030
weeks OK so and and fair is led by or NYU colleague rob fergus right now after joel pino left several months ago fair is being pushed towards kind of working on slightly you know shorter term projects that it has done in the in the traditionally with less emphasis on publication
Yann LeCun (Chief AI Scientist) 66:14.230
more focused on sort of helping TBD lab with the LLM's and frontier models and and you know last publication which means you know meta is becoming a little more close closed and TBD lab has achieved scientists also but which is really focused on LLM and other organizations are
Yann LeCun (Chief AI Scientist) 66:40.590
more like infrastructure and products so you know there's some appropriate research there so for example the group that works on
? (?) 66:46.310
sam segment eighteen yeah yeah that's actually part of the product division of missile there used to be at fair but because they worked on kind of relatively you know kind of outside facing kind of practical things that were kind of moved to department and and do you have any
? (?) 67:04.160
opinions on like some of the other companies that are trying to move into world models like thinking machines or even i've heard jeff bezos and and some of his it's not clear at all what seeking machine is doing maybe you have more information than me but maybe not sorry maybe
? (?) 67:21.080
i'm mixing it up here physical
Yann LeCun (Chief AI Scientist) 67:23.440
intelligence physical sorry yeah sorry and then i mix them up with like SSI as well they're all kind of like so nobody knows what they're doing including their own investors OK and he said the rumors that's a rumor i heard i don't know if it's true it's become a bit of a joke
Yann LeCun (Chief AI Scientist) 67:44.230
but but yeah physical intelligence company is is focused on you know basically producing geometrically correct videos OK where you know there is persistent geometry and you know when you look at something and you turn around and you come back it's the same object you had before
Yann LeCun (Chief AI Scientist) 68:11.030
like it doesn't change behind your back right so it's it's generative right i mean the whole idea is to generate pixels which i just spent you know a long time arguing against that was a bad idea there are other companies that are have role model a good one is wave wave W A Y V
Yann LeCun (Chief AI Scientist) 68:34.430
E so it's a company based in oxford and they i'm an advisor for full disclosure and they have they have a role model for autonomous driving and the way they're training it is that they're training a representation space by basically training a VAE or VQVAE and then training a
Yann LeCun (Chief AI Scientist) 68:52.670
predictor to do temporal prediction in that abstract representation space so they have half of it right and half of it wrong the piece they have right is that you make predictions in representation space the pieces are wrong is that they haven't figured out how to train their
Yann LeCun (Chief AI Scientist) 69:07.390
representation space in any other way than by reconstruction and i think that's bad OK but the MRL is great like it works really well i mean among all the people who kind of work in this kind of stuff they're pretty far advanced there are people who talk about similar things and
Yann LeCun (Chief AI Scientist) 69:24.790
nvidia a company called sandbox AQ the CEO of it jack hilary talks about qualitative models you know large quantitative models as opposed to large language models predictive models that can deal with continuous high dimensional noisy data right which is what also i've been kind
Yann LeCun (Chief AI Scientist) 69:44.150
of talking about and google of course has been working on you know on word models mostly using generative approaches there was an interesting effort at google by dani jar after so he built models called dreamer dreamer V one two three four yeah that was on a good path except he
Yann LeCun (Chief AI Scientist) 70:05.600
just left google to create his own startup and do you have so i'm interested so you were really criticized about the silicon valley culture that they are focusing on NLM and this is like one of the reasons that now you started with the new company is starting in paris right so
Yann LeCun (Chief AI Scientist) 70:33.360
this is something do you think that we will see more and more or do we see this is something will be very unique
? (?) 70:41.430
that only a few companies will be in europe running is global OK it has an office in paris but it's a global company has office in new york too a couple other places so OK there is an interesting phenomenon in industry which is that everybody has to do the same thing as
? (?) 71:03.510
everybody else because it's so competitive that if you start
Yann LeCun (Chief AI Scientist) 71:07.110
taking attention you're taking a good risk of falling behind because you're using a different technology than everybody else right so basically everyone is trying to catch up with the others and so that creates this herd effect and a kind of monoculture which is really specific
Yann LeCun (Chief AI Scientist) 71:23.470
to silicon valley where you know open AI beta google into our pick everybody is basically working on the same thing and you know sometimes like what happened a while back another group you know like deepseek in china comes up with kind of a new way of doing things and everybody
Yann LeCun (Chief AI Scientist) 71:43.710
is like what right you mean like other people in silicon valley are not stupid and can come up with original ideas i mean there's a bit of a you know superiority complex right but you're basically in your trench and you are you have to move as fast as possible because you can't
Yann LeCun (Chief AI Scientist) 72:00.080
afford board to kind of you know fall behind the other guys who you think are your competitors but you run the risk of being surprised by something that's completely out of the left field that uses a different set of technologies and or maybe addresses a different problem so you
Yann LeCun (Chief AI Scientist) 72:17.560
know what i've been interested in is completely orthogonal because the the whole japan idea word model is really to handle data that is not easily handled by LLM so the type of applications we're envisioning that have tons of applications in industry
? (?) 72:33.550
where the data comes to you in the form of continuous high dimensional noisy data
Yann LeCun (Chief AI Scientist) 72:38.350
including video are domains where LMS basically are not present where people are trying to use them and totally failed essentially right so if you don't want to be OK so the expression in silicon
? (?) 72:53.310
valley is that you are LLM pilled you you think that the path to superintelligence you just get up and adams you train on more synthetic data you license on more data you hire thousands of people to kind of find you know to basically school your
Yann LeCun (Chief AI Scientist) 73:09.040
system in post training you invent a new tweaks on RL
? (?) 73:13.680
and you're going to get to super intelligence and this i think is complete bullshit like it's
Yann LeCun (Chief AI Scientist) 73:18.160
just never going to work and then you add a few you know kind of reasoning techniques which basically consist in you know doing like super long chain of thoughts and then having the system generate lots and lots of different token outputs you know from which you can select good
Yann LeCun (Chief AI Scientist) 73:32.830
ones using some sort of valuation function the second LLM basically you hit i mean that's the word all those things work this is not going to take us it's just not so so yeah i mean you need to escape that culture and there are people within all the companies in silicon valley
Yann LeCun (Chief AI Scientist) 73:49.400
who think like this is never going to work i want to i want to do an event in japan blah blah blah i'm hiring them so yeah so escaping the moodle culture of silicon valley i think is important yeah this is a this part of the the story and what do you think about the competition
Yann LeCun (Chief AI Scientist) 74:17.030
between like the US china and the and europe like now that you are starting a company like do you see more i know that some there are some places are more attractive than others we're in this very paradoxical situation where all the american companies until now not meta but all
Yann LeCun (Chief AI Scientist) 74:39.840
american companies have been kind of becoming really secretive and to preserve their competitive what they think is a competitive advantage and by contrast the chinese players companies and others have been completely open so the best open source systems at the moment are
Yann LeCun (Chief AI Scientist) 75:00.630
chinese and that causes a lot of the industry to use them because they want to use a pencil system and they hold their nose a little bit because they know those models are kind of fine tuned to not answer questions about politics and stuff like that right but they don't really
Yann LeCun (Chief AI Scientist) 75:18.560
have a choice and certainly a lot of academic research now you know uses the best chinese models certainly everything that has to do with like reasoning and things like that right so so it's really paradoxical and a lot of people in the US in industry are really unhappy about
Yann LeCun (Chief AI Scientist) 75:37.430
this they really want a serious non chinese open source model there could have been was a disappointment for various reasons maybe that will get fixed with you know the the new efforts at meta or maybe meta will decide to go close as well it's still clear mistral just had a
Yann LeCun (Chief AI Scientist) 75:57.160
model early yes just be cool for cogen yeah yeah yeah that's right no it's it's it's cool so yeah they they maintain openness yeah no it's really really interesting what they're what they're doing yeah wow OK let's go to more personal questions yeah yeah so like you are sixty
Yann LeCun (Chief AI Scientist) 76:23.510
five right years old you want a turing award you just got a queen elizabeth prize basically you could retire right yeah i could that's what my wife wants me to do
? (?) 76:36.320
so why why to start a new company now like what keep you happen because i have a mission you know i mean i always thought that
Yann LeCun (Chief AI Scientist) 76:49.390
either making people smarter or more knowledgeable or making this model with the help of machines so basically increasing the amount of intelligence in the world was an intrinsically good thing OK intelligence is really kind of the commodity that is the most in demand certainly
Yann LeCun (Chief AI Scientist) 77:09.950
in like government OK so but but like in you
? (?) 77:17.080
know every aspect of of life we are limited as you know as as a species as a planet by the limited supply of intelligence
Yann LeCun (Chief AI Scientist) 77:29.480
right which is why we we we spend enormous resources educating people and and and things like that so you know increasing the amount of intelligence at the service of humanity or the planet more globally not just humans is intrinsically a good thing despite all the what the
Yann LeCun (Chief AI Scientist) 77:51.570
doomers are saying OK of course you are dangerous and you have to protect against that in the same way you have to make sure your jet engine is safe and reliable and your car doesn't kill you with a you know small crash right but that's OK that's an engineering problem there's
Yann LeCun (Chief AI Scientist) 78:06.920
no there's no like fundamental issue with that also with political department but not it's not like insurmountable so that's an interestingly good thing and if i can contribute to this i will and basically all research projects i've done in my entire career even those that were
Yann LeCun (Chief AI Scientist) 78:25.350
not related to machine learning in my professional activities were all focused on either making people smarter that's what that's why i'm
? (?) 78:34.190
a professor and
Yann LeCun (Chief AI Scientist) 78:39.640
that's why also i'm communicating publicly a lot about AI and science and things like that and a big presence on social networks and stuff like that right because i think people should know stuff right but also on machine intelligence because i think machines will assist humans
Yann LeCun (Chief AI Scientist) 78:59.470
and make them smarter OK people think there is a fundamental difference between trying to make you know machines that are intelligent and autonomous and blah blah blah and and it's a different set of technologies so i'm trying to make machines that are assistive to humans it's
Yann LeCun (Chief AI Scientist) 79:17.360
not it's the same technology it's exactly the same and it's not because the system is intelligent or even a human is intelligent that it wants to dominate or take over it's not even true of humans like it's not the humans who are the smartest that want to dominate
Yann LeCun (Chief AI Scientist) 79:32.030
others we see this on the international political scene every day it's not the smartest among among us who want to be the chief and probably many of the smartest people that we've ever met are people who basically want nothing to do with the rest of humanity right they just want
Yann LeCun (Chief AI Scientist) 79:50.480
to work on their problems you know kind of stereotyping that's what
? (?) 79:58.960
hannah wren talks about the vita contemplativa right versus like the active life or the contemplative life right and her like philosophical analysis and like making a choice kind of early on on what you work on
Yann LeCun (Chief AI Scientist) 80:11.230
right but you can be you know simultaneously kind of you know a dreamer or contemplative but have a big impact on the world right by you know your scientific production like think of einstein or something
Yann LeCun (Chief AI Scientist) 80:25.110
yeah or even newton like newton basically didn't want to meet anybody
? (?) 80:31.070
famously
Yann LeCun (Chief AI Scientist) 80:33.520
or pull the rack pull the rack was kind of you know practically autistic well well
? (?) 80:42.680
is there like a paper or idea you haven't written or or something else that you you're you know nagging that you want to get to or maybe that you don't have time or any regret
Yann LeCun (Chief AI Scientist) 80:52.590
oh yeah a lot oh my entire career has been a succession of me not devoting enough time to express my ideas and writing them down and mostly getting scooped
? (?) 81:08.920
what is the most significant one i
Yann LeCun (Chief AI Scientist) 81:13.200
don't want to go through that
Yann LeCun (Chief AI Scientist) 81:16.160
the backprop is a good one
? (?) 81:18.510
OK
Yann LeCun (Chief AI Scientist) 81:19.350
i published some sort of early version of some algorithm to trade multilayer nets which today we would call target prop and i had the back prop team figured out except i didn't write it before you know there were more heart engine they were nice enough to cite my earlier paper
Yann LeCun (Chief AI Scientist) 81:39.840
in their in theirs but so there's been a few of those yeah recurrent nets you know various other things but and things are more perhaps more recent but you know i have no regrets about this like you know this is life like you know i'm not going to say oh you know i invented this
Yann LeCun (Chief AI Scientist) 81:58.830
in nineteen ninety one and i should have
? (?) 82:01.230
like somewhere
? (?) 82:04.800
i i don't know i should say the name we
Yann LeCun (Chief AI Scientist) 82:08.120
all know well i mean if you know you know the way ideas pop up you know is is relatively complex it's rare that someone comes up with an idea in complete information and that you know nobody else comes up with similar ideas at the same time most of the time appear simultaneously
Yann LeCun (Chief AI Scientist) 82:25.110
but then there is various ways to is having the idea and then there is kind of writing it down but there is also writing it down in a sort of convincing way in a clear way and then there is kind of making it work on toy problems maybe OK and then there is making the theory that
Yann LeCun (Chief AI Scientist) 82:40.670
shows that it can work and then there is making it work on a real application right and then there is making a product out of it OK so this whole chain and you know some people are really extreme think that the only person who should get all the credit is the very first person
Yann LeCun (Chief AI Scientist) 82:55.110
who got the
? (?) 82:56.030
idea
Yann LeCun (Chief AI Scientist) 82:57.030
i think that's wrong there's there's a lot of really difficult steps to get this idea to the state what actually works so this idea of role model i mean goes back to the nineteen sixties you know people in optimal control had water models to do planning that's the way NASA you
Yann LeCun (Chief AI Scientist) 83:11.870
know planned the trajectory of the rockets to go to to orbit basically simulating the rocket and sort of bio optimization figuring out the the controller to get the rocket to where it needs to be so that's another idea very old idea the the fact that you could do some level of
Yann LeCun (Chief AI Scientist) 83:32.990
training or adaptation in this is called system identification in article control very old idea too goes back to the seventies something called yeah system identification or even MPC where you adapt the model as it goes like while you're running the system then go back to the
Yann LeCun (Chief AI Scientist) 83:48.390
seventies to some obscure paper in france and then the fact that you can just learn a model from data people have been working on this with neural net since the nineteen eighties right and and not just
Yann LeCun (Chief AI Scientist) 84:04.110
yoga it's like a whole bunch of people who have been working people who came from optimal control and realized they could use neural nets as kind of a universal function approximator and use it for direct control or feedback control or role models for planning blah blah blah and
Yann LeCun (Chief AI Scientist) 84:21.230
like a lot of things in neural nets in the nineteen eighties and nineties it kind of worked but not like to the point where it took over the industry so it's the same for you know computer vision speech recognition there were attempts at using neural nets for that back in those
Yann LeCun (Chief AI Scientist) 84:38.910
days but it started really really working well in the late two thousands where it totally took over right and then early two thousand ten S for vision mid two thousand ten's for NLP and for robotics it's starting but it's not
? (?) 84:55.030
why why i think it's only like in the this time started to get over the well
Yann LeCun (Chief AI Scientist) 85:02.150
it's combination of like having the right state of mind about it and the right mindset having the right architectures the right machine learning techniques like you know residual connections real use whatever then having powerful enough computers and having access to data and
Yann LeCun (Chief AI Scientist) 85:21.000
it's only when those planets are aligned that you get a breakthrough right which appears like a conceptual breakthrough but it's actually just a practical one like OK let's talk about confidential nets OK lots of people during the seventies had the idea or even during the
Yann LeCun (Chief AI Scientist) 85:40.190
sixties actually had the idea of using local connections like building a neural net with local connections for extracting local features and the idea that local features is like convolution like in image processing is like you know goes back to the sixties so these are not new
Yann LeCun (Chief AI Scientist) 85:56.270
concepts the fact that you can learn adaptive filters of this type using data goes back to the perceptron and adaline which is early sixties OK but that's only for one layer now the concept that you can train a system with multiple layers everybody was looking for this in the
Yann LeCun (Chief AI Scientist) 86:11.990
sixties nobody found a lot of people made proposals which kind of have worked but like none of them was convincing enough for people to say OK this is a good technique technique that was adopted is what's called polynomial classifiers we turn this into kernel methods but it's
Yann LeCun (Chief AI Scientist) 86:29.590
you know basically you sort of have a handcrafted feature extractor and then you train basically what amounts to a linear classifier on top of it that was going to come on practice in the seventies and certainly eighties but the idea that you could train a nonlinear system
Yann LeCun (Chief AI Scientist) 86:45.590
composed of multiple nonlinear steps using gradient descent the basic concept for this goes back to the kelly bison algorithm which is optimal control was mostly linear from nineteen sixty two and people in optimal control kind of word thinks about this you know in the sixties
Yann LeCun (Chief AI Scientist) 87:03.470
but nobody realized you could use this for machine learning to do pattern recognition or to do you know natural language processing that really only happened after you know the ramelight hinton williams paper in nineteen eighty five even though people had proposed the very same
Yann LeCun (Chief AI Scientist) 87:19.390
algorithm a few years before like pope werbus you don't propose you know what he called order derivatives which turns out to be backdrop but it's the same thing as the adjoint state method in optimal control so like those are ideas i mean the fact that an idea or a technique is
Yann LeCun (Chief AI Scientist) 87:34.190
reinvented multiple times in different fields and then only after the fact people say oh right it's actually the same thing and we knew about this before we didn't realize we could use this for it for this particular stuff right so all those like claims of plagiarism it's just
Yann LeCun (Chief AI Scientist) 87:49.960
it's just a complete misunderstanding of ideas
? (?) 87:56.440
OK what do you do when you're not thinking about AI
Yann LeCun (Chief AI Scientist) 88:03.840
have a whole bunch of hobbies that have very little time to actually partake in alexei aline so i go selling in the summer i like selling multi hall boats like tremor and catamarans i have a bunch of boats i like building flying contraptions
? (?) 88:31.440
so a modern da vinci i wouldn't i
Yann LeCun (Chief AI Scientist) 88:35.160
wouldn't call it airplanes because a lot of them don't look like airplanes but they don't fry OK i like the the sort of you know concrete creative act of that my dad was aerospace engineer and he mechanical engineer working in the aerospace industry and he was you know building
Yann LeCun (Chief AI Scientist) 88:54.670
airplanes as a hobby and like you know building his own radio control system and stuff like that and he got me and my brother into it my brother who works at google at google research in france in paris and and and that became kind of a family activity if you want so my brother
Yann LeCun (Chief AI Scientist) 89:12.990
and i still still do this but and then in the covid years i picked up astrophotography so i have a bunch of telescopes and pictures of the sky and i built electronics so since i was a teenager i was interested in music i was playing renaissance and baroque music and also some
Yann LeCun (Chief AI Scientist) 89:33.510
type of folk music playing wind instruments wind winds and but i was also into electronic music and my cousin who is such older than me was inspiring electronic musician so we had like analog synthesizers because i knew electronics i would like you know modify them for him and i
Yann LeCun (Chief AI Scientist) 89:54.430
was still in high school at the time and and now in my home i have a whole bunch of synthesizers and i build electronic musical instruments so so these are wind instruments you blow into them you know there's fingering and stuff but what they produce is control signals for a
Yann LeCun (Chief AI Scientist) 90:16.190
synthesizer
? (?) 90:17.950
it's cool very cool i've heard
? (?) 90:21.040
a lot of people in tech are into sailing like in yeah gotten that answer surprising amount i'm going to start trying to sail now OK
Yann LeCun (Chief AI Scientist) 90:29.040
so i tell you something about sailing it's very much like the war model story to be able to you know kind of control this elbow properly to make it go as fast as possible and everything you have to anticipate a lot of things you have to anticipate the motion of the waves like
Yann LeCun (Chief AI Scientist) 90:45.070
how the waves are going to affect your boat you know whether a gust of wind is going to come and have to you know start you know the woods going to start healing and things like that and you basically have to run CFD in your head because you have to figure out like the you know
Yann LeCun (Chief AI Scientist) 91:01.550
through dynamics you have to figure out like what is the flow of air around the around the around the sails and you know that if the angle of attack is too high it's going to be turbulent on the back and the the lift is going to be much lower so blah blah blah so like you know
Yann LeCun (Chief AI Scientist) 91:19.350
tuning sales is basically requires running cod in your head but at an abstract level you're not solving the you know navier stokes right we have really good intuitive so that's what i like about it like the whole thing that you have to build this mental you know predictive model
Yann LeCun (Chief AI Scientist) 91:36.550
of the world to be able to do a good
? (?) 91:39.390
job how many samples you need
Yann LeCun (Chief AI Scientist) 91:43.560
yeah probably a lot but but you know you get to run it on ian you know if you a few years of practice and yeah yeah
? (?) 91:55.190
OK your friends and you lived in the US for many decades already do you still feel french derek does that perspective shape your view of the of the wall of the american tech culture
Yann LeCun (Chief AI Scientist) 92:12.160
well inevitably yeah i mean you you can't completely escape your your upbringing and your culture so i mean i feel both french and american in the sense that you know i mean in the US for thirty seven years and in north america for thirty eight because i was in canada before or
Yann LeCun (Chief AI Scientist) 92:33.550
children grew up in the US and so from that point of view i'm american but i have a view certainly on you know on various aspects of science and society that probably are you know a consequence of growing up in france yeah absolutely and if you're french i mean france
? (?) 92:54.680
i'm curious i did not actually realize that you had a brother that also worked in tech i'm fast fascinated by this because yahshua bengio 's brother also works in tech and i always thought that he was the only serena venus williams situation in AI but you you too also have a
? (?) 93:12.230
brother so how many more AI research like like is it that common that it just runs in families
Yann LeCun (Chief AI Scientist) 93:17.430
i have no idea that's what i have a sister who you know is not in tech but she's also a professor my brother was a professor before he moved to google he he doesn't work on AI machine learning he's very careful not to he's a younger brother six years younger than me and he works
Yann LeCun (Chief AI Scientist) 93:39.880
on operations research and optimization essentially which now is actually also being invaded by machine learning
? (?) 93:52.400
yeah OK one more question so like if the world models work in twenty years from now what is the what is the dream like how what how does it look like how does i know like our lives will be
Yann LeCun (Chief AI Scientist) 94:10.470
total world domination OK it's a joke the the i i said this status because this is what linda starvalds used to say you say like what's your goal with linux and you say total i thought it was super funny and it actually succeeded i mean basically you know to first approximation
Yann LeCun (Chief AI Scientist) 94:37.630
every computer in the world runs linux there's only a few desktops that don't and a few iphones but you know everything else with linux so really like you know having you know pushing towards like a recipe for training and building intelligence systems perhaps all the way to
Yann LeCun (Chief AI Scientist) 94:56.910
human intelligence or more and and basically building AI systems that would you know help people and humanity more generally in their daily lives at all times amplifying human intelligence will be their boss right it's not like those things are going to dominate us because again
Yann LeCun (Chief AI Scientist) 95:16.350
it's not because something is intelligent that it wants to dominate those are two different things in humanity you know we are hardwired having to influence other people and sometimes it's through domination sometimes it's through prestige but the hardwired evolution to do this
Yann LeCun (Chief AI Scientist) 95:37.070
because we are a social species there's no reason we would build those kind of drive into into our intelligent systems and it's not like they're going to develop those kinds of drives by themselves so so yeah i'm quite optimistic
? (?) 95:54.280
me too so am i all right OK so we have final questions from the audience and so yeah let's start and if you were starting your AI career today what skills and research directions would you focus on
Yann LeCun (Chief AI Scientist) 96:12.400
i get a lot this question a lot from young students or parents of future students i mean i think you should learn things that have a long shelf life and you should learn things that help you learn to learn because technology is you know evolving so quickly that you all kind of
Yann LeCun (Chief AI Scientist) 96:33.270
you know the ability to learn really quickly and basically that can that is done by you know learning very busy so in the context of stem right science and technology engineering mathematics and i'm talking about humanities here this is although you should learn philosophy this
Yann LeCun (Chief AI Scientist) 96:58.310
is done by learning things i have a long shared life so the joke i say is that if you first of all you the things i have a long shelf life tend to not be computer science OK so here's a computer science professor you know are you against studying computer science don't count
? (?) 97:15.440
don't count to study
Yann LeCun (Chief AI Scientist) 97:19.520
and have a terrible coefficient to make which is i studied electrical engineering as an undergrad so i'm not a real computer scientist OK but what what we should do is learn kind of basic things in mathematics in modeling mathematics that can be connected with reality you tend
Yann LeCun (Chief AI Scientist) 97:37.190
to learn this kind of stuff in engineering in some schools that's linked with computer science but sort of you know electrical engineering mechanical engineering et cetera engineering disciplines you're all you're running in the US calculus one two three that gives you a good
Yann LeCun (Chief AI Scientist) 97:52.110
basis right computer science you don't you know you can get away with just calculus one that's not enough right you know we're learning you know probability theory and linear algebra you know all the stuff that are really kind of basic and then if you do network engineering
Yann LeCun (Chief AI Scientist) 98:09.680
things like i don't know control tiers or signal processing like all of those methods optimization you know all of those methods are really useful for things like AI and then you can you can basically learn similar things in physics because physics is all about like what should
Yann LeCun (Chief AI Scientist) 98:27.240
i represent about reality to be able to make predictive models right and that's really what intelligence is about so so i think you can learn most of what you need to learn also if you go through a physics a curriculum but obviously you need to learn enough computer science to
Yann LeCun (Chief AI Scientist) 98:45.190
kind of program and use computers and even though AI is going to help you be more efficient at programming you still need to know like how how to do this what
? (?) 98:56.150
do you think about vivec coding
Yann LeCun (Chief AI Scientist) 98:59.310
i mean it's cool it makes is going to cause a funny kind of thing where a lot of the code that will be written will be used only once because it's going to become so cheap to write code right you're going to ask your kind of AI assistant like you know produced this graph or like
Yann LeCun (Chief AI Scientist) 99:20.190
you know this research blah blah blah he's going to write a little piece of code to do this or maybe it's an applet that you need to play with for you know a little simulator and you can use it once and throw it away because so cheap to produce right so the idea that we're not
Yann LeCun (Chief AI Scientist) 99:34.710
going to need programmers anymore is false we're going to the cost of generating software you know has been going down continuously for decades and it's that's just the next step of the cars going down but it doesn't mean computers will be less useful they're going to be more
Yann LeCun (Chief AI Scientist) 99:53.030
useful
? (?) 99:56.930
OK one more question so what do you think about the connection between neuroscience and machine learning there are some ideas a lot of time that AI borrows from neuroscience and the other way right predictive coding for example do you think that it's useful to use ideas from but
Yann LeCun (Chief AI Scientist) 100:18.030
there's a lot of inspiration you can get from neuroscience from biology in general but neuroscience in particular i certainly was very influenced by classic work in neuroscience like you know hubble and riesel 's work on the architecture of the visual cortex is basically what
Yann LeCun (Chief AI Scientist) 100:34.190
led to convolutional nets right and you know i wasn't the first one to use those ideas in in artificial neural nets right there were people in the sixties trying to do this there were you know people in the eighties building locally connected networks with multiple layers they
Yann LeCun (Chief AI Scientist) 100:48.470
didn't have ways to train them with backdrop you know there was the the cognitron the neo cognitron from fukushima which had a lot of the ingredients just not proper learning algorithm and there was another kind of aspect of the cognitron which is that it was really meant to be
Yann LeCun (Chief AI Scientist) 101:06.350
a model of the visual cortex so it tried to reproduce every quirks of biology for example the fact that in the brain you don't have positive and negative weights you have positive and negative neurons so all the neurons all the synapses coming out of an inhibitor neurons neurons
Yann LeCun (Chief AI Scientist) 101:28.270
have negative weights OK and all the synapses coming out of non inhibitory neurons have positive weights so fukushima implemented this in his model right he implemented the fact that you know what neurons spike so we didn't have a spiking neuron model but but you cannot have a
Yann LeCun (Chief AI Scientist) 101:49.830
negative number of spikes and so so his function was basically rectification like a rail you except it had a saturation and then he knew from you know various works that there was some sort of normalization and he had to use this because otherwise there was no back props of the
Yann LeCun (Chief AI Scientist) 102:08.080
activation in this network would go haywire so we had to do like division normalization that turns out to actually be correspond to some theoretical models of the visual cortex that some of our colleagues at the center for neuroscience said NYU have been pushing like david heger
Yann LeCun (Chief AI Scientist) 102:25.230
and people like that so yeah i mean i think neuroscience is very a source of inspiration you know more more recently the sort of macro architecture of the brain in terms of you know perhaps the world model and planning and things like this like how how is that reproduced like
Yann LeCun (Chief AI Scientist) 102:45.550
why do we have a separate module in the brain for for for factual memory the hippocampus right and we see this now in certain neural net architectures that there is like a separate memory module right maybe that's a good idea i think we're going to have i think what's going to
Yann LeCun (Chief AI Scientist) 103:05.310
happen is that we're going to come up with sort of new AI neural net architectures deep learning architectures and a posterior you will discover that the characteristic that we implemented in them actually exist in the brain and in fact that's a lot of what's happening now in
Yann LeCun (Chief AI Scientist) 103:22.870
neuroscience which is that there is a lot of feedback now from AI to neuroscience where the best models of human perception are basically conventional nets today
? (?) 103:34.600
yeah OK
? (?) 103:37.440
do you have anything else that you want to add to say to the audience whatever
Yann LeCun (Chief AI Scientist) 103:43.200
you want i think we covered a lot of grounds i think you you know you want to be careful who you listen to so don't listen to AI scientists talking about economics OK so when some AI person or even a business person tells you AI is going to put everybody out of work talk to an
Yann LeCun (Chief AI Scientist) 104:07.030
economist basically none of them is saying anything anywhere close to this OK and you know the effect of technological revolutions on the labor market is something that a few people have devoted their career on doing none of them is predicting massive unemployment none of them
Yann LeCun (Chief AI Scientist) 104:30.650
is predicting that radiologists are going to be all unemployed you know et cetera right also realize that actually fielding practical applications of AI so that there are sufficiently reliable and everything is super difficult and it's very expensive and in previous waves of
Yann LeCun (Chief AI Scientist) 104:49.400
interest AI the techniques that people have put a big hope in turned out to be overly unwieldy and expensive except for a few applications so there was a big wave of interest in expert systems back in the nineteen eighties japan started a huge project called the fifth generation
Yann LeCun (Chief AI Scientist) 105:14.760
computer project which was like computers with CPU 's that were going to run lists and you know inference engines and stuff right and the hottest job in the late eighties was going to be knowledge engineer you were going to sit next to an expert and then the knowledge of the
Yann LeCun (Chief AI Scientist) 105:31.790
expert into rules and facts right and then the computer would be able to basically do what the expert was this was manual behavior clothing OK uh and it kind of worked but only for a few domains where where economically it made sense and it was doable at the level of reliability
Yann LeCun (Chief AI Scientist) 105:49.110
that was good enough but it was not a pass towards kind of human level intelligence the idea somehow like the delusion that people today have that the current AI mainstream you know fashion is going to take us to human intelligence has happened already three times during my
Yann LeCun (Chief AI Scientist) 106:12.990
career and probably five or six times before right you should you should see what people are saying about the perceptron right the new york times article people were saying oh we're going to have like super intelligent machines within ten years marvin minsky in the sixties says
Yann LeCun (Chief AI Scientist) 106:25.470
oh within ten years the best chess player in the world would be a computer it took a bit longer than that and you know and you know this you know this happened over and over again in nineteen fifty six or something when newell and simon produced the the general problem solver
Yann LeCun (Chief AI Scientist) 106:45.360
very modestly called the general problem solver OK what they what they thought was really cool they say OK the way we think is very simple we pose a problem there is a number of different solutions to that problem different proposals for solution a space of potential solutions
Yann LeCun (Chief AI Scientist) 107:03.880
like you know you do like traveling salesman right there was a number of you know factorial you know N factorial paths possible paths you just have to look for the one that is the best right and they say like every problem can be formulated this way essentially for a search for
Yann LeCun (Chief AI Scientist) 107:20.360
the best solution if you can formulate the problem as an objective by writing a program that checks whether it's a good solution or not or gives a rating to it and then you have a search algorithm that search through the space of possible solution for one that optimizes that
Yann LeCun (Chief AI Scientist) 107:39.070
score that you saw the AI OK now what they didn't know at the time is order complexity theory that basically every problem that is interesting is exponential or AP complete or whatever right and so oh we have to use heuristic programming you know kind of come up with heuristics
Yann LeCun (Chief AI Scientist) 107:55.510
for every new problem and basically you know there are general problem solver was not that general so like this idea somehow that the latest idea is going to take you to you know AGI or whatever you want to call it is very dangerous and a lot of very smart people fell into that
Yann LeCun (Chief AI Scientist) 108:14.110
trap many times over the last seven decades do
? (?) 108:17.270
you think that the field will ever figure out continual or incremental learning
Yann LeCun (Chief AI Scientist) 108:22.310
sure yeah that's sort of a technical problem well
? (?) 108:26.270
well i thought i thought catastrophic forgetting right because your weights that you trained so much money on get overwritten sure
Yann LeCun (Chief AI Scientist) 108:32.750
so you train just a little bit of it i mean we don't already do this with SSL right we train a foundation model like for video or something like vijay patu you know producers really good representations of video and then if you want to train the system for a particular task you
Yann LeCun (Chief AI Scientist) 108:46.990
train a small head on top of it and that head can be you know along continuously and even your word model can be trained continuously that's not an issue i don't see this as like a big a huge challenge frankly in fact raya heads to LPR simon and i and a few of our colleagues
Yann LeCun (Chief AI Scientist) 109:02.270
back in two thousand five two thousand six build a learning based navigation system for mobile robots that had this kind of idea so it was it was commercial net that was doing semantic segmentation from camera images and on the fly the top layers of that network would be adapted
Yann LeCun (Chief AI Scientist) 109:22.280
to the current environment so you do a good job and the labels came from short wrench uh traversability that were indicated by stereo vision essentially so yeah i mean you can do this it's particularly if you have multimodal yeah i don't see this as a big challenge
? (?) 109:46.180
it's been a pleasure to have you
Yann LeCun (Chief AI Scientist) 109:48.350
real pleasure to thank you so much thank you thank you