>> The next topic,
the last short talk in the series is on Deep Learning.
And I'd like to
invite our next speaker, Vineeth Bubramanian,
who's a hometown speak in the sense
that he's faculty at IT Hyderabad.
Vineeth finished his PhD at the University of Arizona,
after which he taught for a couple of years.
And now, he's come back
to take a faculty of bachelor in IT Hyderabad.
So over to you, Vineeth. Thank you.
>> All right. Okay. So, how many of you
in this room work on Deep Learning,
use Deep Learning, read Deep Learning?
Okay. Works for PJ on behalf of Deep Learning?
All right. Okay, let's get started.
We can go to the next slide, please. Thank you.
So, I think I don't need to talk about this,
I think there's been
explosive growth of Deep Learning in recent years.
I think we've all seen applications of
Deep Learning in vision, text,
speech and I think the number of publications in
Deep Learning has been growing over the years.
The bottom right graphic that you see there is
an interesting graphic there actually.
It's Google's graphic on the number of
folders on Google Servers,
which hold Deep Learning models.
I was just looking for Deep Learning statistics and
that's one of the ways people have
measured how Deep Learning has grown,
but that's an interesting statistic but that's
also going up fairly exponentially.
So, I think most of you are working in this space,
you don't need this background.
But, I'll quickly go over this and probably then try to
cover what I'm trying to cover on.
I think that I'll try to focus
on today's state of Deep Learning,
which is what I'm intended to be here for.
But, Deep Learning has a pretty deep history.
So, it started in 1940
with the McCulloch-Pitts model of the neuron.
Then came Rosenblatt with the perceptron.
Then came the Widrow-Hoff learning rule in 1960.
Unfortunately, Rosenblatt claimed
that the perceptron could
approximate any kind of a function.
And Minsky and Papert came in
1969 and showed the XOR example,
and showed that this is not possible.
And then came a kind of a dark age
for neural networks at that point in time until say,
the mid-80's when backpropagation came into the picture.
And again, neural networks took a spike.
But, probably late 90's and early 2000's,
it was dull again until 2006,
when Geoffrey Hinton and
Ruslan Salakhutdinov came up with their
Hierarchical Feature Learning using
restricted Boltzmann machines
to train deep-belief networks.
And since then, I think it's been an upward trend
and it's now effectively
the golden age for Deep Learning.
I'm not going to go more into this,
I think there's of course,
a lot of landmark achievements that were in
between like the neocognitron, and things like that.
But with just this background,
what I'll focus on for the rest of
my talk today is to actually talk about, I think.
Okay, one part that I missed, of course which, oops.
I think going back, is that still a problem?
Okay. So, I think all of us know by now that
AlexNet was one of the turning points
for Deep Learning that happened in 2012,
when Deep Learning models when the ImageNet challenge.
So, what I'll focus on for the rest of
the talk today is actually post 2012,
so we won't touch until 2012 what neural networks were.
So, I've tried to cover the thoughts in my head.
And probably, I should start with a disclaimer
that this talk obviously contains my own bias.
So, I'm sure each of you here can
come and give a talk on Deep Learning for 20 minutes.
So, bear with me if it disagrees,
and I'll be happy to talk offline if required.
So, I think I've tried to
cover the topics into three parts.
One is the consolidation
of successes of Deep Learning over the last few years.
So, it became successful for
computer vision and even before that for speech.
So, what would have happened in that space
and how was that success got consolidated?
I mean, in some sense establishing its territory,
something in that sense.
Then I'll talk about exploration of new frontiers.
What are the new directions and
emerging directions in this space?
And of course, I think no talk on Deep Learning
is complete without some Deep Learning bashing.
So, we will talk about some limitations,
which are challenges and opportunities
depending on how you look at it.
So, one of the main things that have happened in
the last four years in Deep Learning is obviously depth.
Okay, that's what the name stands for.
So, we also saw this in
Professor Joha's slides yesterday,
that I think the ImageNet challenge has been
solved by deeper and deeper networks.
Every other depth of the network
has been going higher and higher,
and it kind of stopped with the residual nets
at about 152 layers.
And today, I think most of the networks,
I should say, are in the order of 100's.
And very recently, I think last year there
was a paper on AAAI,
which talked about, When and Why Are
Deep Networks Better than Shallow Ones?
And they did an empirical study.
And I was actually personally surprised to conclude.
And in fact, I think it's the second paper on this list.
The first line of the abstract is, 'Yes, they do'.
That's the first line of the abstract.
Okay. Sorry, I think it's the first paper then,
Do Deep Convolutional Nets Really Need to be Deep?
I think the first line of
the abstract is, 'Yes, they do'.
So, they actually seem
to conclude that you do need depth,
because that's always a debate
in the Deep Learning community whether
you really need all the depth?
Can you manage this with some shallow networks?
So on and so forth.
So, that's been one of
the prime components of
developments over the last few years.
And the depth may not be only along one axis,
can also be at the other axis in case of LSTMs,
stack up a lot of LSTMs.
So, I think that's been another trend in
Deep Learning over the last few years.
I've probably referenced a paper
here that talks about this in more detail,
the understanding is that by stacking LSTMs,
you're probably learning
a more hierarchical feature representation
in the data space.
And the Google Neural Machine Translation,
which goes into Google Translate,
actually does use a stacked LSTM in its encoder.
So, that's another trend
that's been happening in Deep Learning.
The other thing that's happened over the last four years
is various kinds of hybridizing of architectures.
So, you take CNNs,
LSTMs, you mix-match them.
You probably take other components,
add them all together.
And take train layers from one,
untrained layers and mix them up.
I think that's been another trend that's actually
been happening over the last few years.
And that's led to various applications in say,
video captioning, image captioning, video classification,
object detection, and many other applications
where this has been one of the main themes.
But I think an interesting thing,
which probably reflects some of
the points covered yesterday too is,
I think in all of these efforts,
there has been an underlying theme to maintain
the end-to-end learning on these architectures.
Because as we all understand now,
that Deep Learning is representation learning.
So, doing the end-to-end
learning kind of helps you with that.
You don't want to again go back
into handcrafting the features.
So, doing the end-to-end learning helps you
with learning the features automatically.
So that's something that's maintained
despite all the modules that you try to
bring into whatever architecture that you're
trying to put together for Deep Learning.
So to some extent, it's become the mix-and-match,
plug-and-play kind of an approach at this point in time.
Probably, very loosely speaking,
I can say this is culminated in Capsule Nets,
which is Geoffrey Hinton's new development in October.
I think it's still not caught on much.
It has some interesting ideas but I think
it's probably not caught on as much.
So, the last four, five years,
we've also seen a significant amount of
development on the hyperparameter engineering space.
A lot of little, little developments that have taken say,
an idea like CNN from where it was four,
five years ago to where it is today
to solve various kinds of problems.
So, in terms of regularization,
you have DropOut, then DropConnect,
which is a generalization of DropOut,
batch normalization, or simply just data augmentation,
or adding noise to the data, label or gradient.
So, these have all been
different ways of doing regularization.
In terms of weight initialization,
there are again been a bunch of methods today.
In fact, I think there
was a paper a couple of years back called,
'All You Need Is A Good Init'.
They show that all you need is a good initialization
to get to a good local minimum.
But I think to this day, a lot of
practitioners use what is called
the Xavier's or the Glorot's initialization,
and then at times He's initialization.
Then, also choosing gradient descent parameters.
So, choosing the learning rate or
momentum in your gradient descent process.
So, that has led to a lot of methods like Adagrad,
RMSProp, Adam, Nesterov Momentum, so on and so forth.
Then on the activation functions front,
you have Rectified Linear Unit,
Exponential Linear Unit, which is more recent,
Parametric Rectified Linear Unit, so on and so forth.
And of course, a variety of loss functions
depending on what problem you're trying to solve.
And of course, this has a flip side.
A lot of this is hyperparemeter engineering
to make things work for a particular task,
and obviously that has a flip side,
that is perhaps the elephant in
the room for Deep Learning is how
to really choose which parameter to use when.
So, we'll revisit that towards the end.
And of course, I think the last four,
five years wouldn't be
complete without talking about the success of
Deep Reinforcement Learning and the success
of it has been reinforced over these years.
It started with Atari Breakout in 2013,
when not only did the model learn to play the game,
but it also learned strategies to build tunnels,
to go up and hit the roof and
bounce off the roof and hit the ceilings,
and get maximum points.
Then of course, in 2016 was AlphaGo,
much televised, much followed.
I think everybody knew that AlphaGo beat Lee Sedol,
the South Korean player four-one
on a series of five games.
But more recently, about a month back,
was AlphaZero, which is again DeepMind,
all of them are DeepMind's creations.
AlphaZero from DeepMind, which played the game of chess.
I think to put that in perspective,
I think all of you chess fanatics here,
all of you know that Magnus Carlsen is world number
one and he has an Elo rating of 2800.
So, Stockfish is one of
the best computer chess-playing systems
that has an Elo rating of 3,00.
So, which means if Stockfish played
Magnus Carlsen a series of 100 games,
Stockfish would beat Magnus Carlsen 95 times.
That's what the difference in 500 Elo points mean.
And, AlphaZero was trained by playing against itself.
Unlike AlphaGo, where there are also
heuristics put into the system,
AlphaZero is actually trained from
scratch by just playing against itself,
nothing else involved in the training.
And by training it just for nine hours,
they had AlphaZero play Stockfish a series of 100 games.
And AlphaZero beat Stockfish 28-0.
So, which has been like
a huge revelation for the community because
of the fact that it learned to
play completely by itself just by self-play,
nothing else, no strategies,
no rules given to it.
Of course, there's some fine print
on the kind of hardware that was
used to train these kinds of systems,
and that's of course a limitation
in academic environments.
That's something to be aware of.
In addition to these kinds of developments,
there's of course been a proliferation
of applications of Deep Learning.
It's everywhere now, especially in the vision,
text, and speech space.
Very broadly speaking, wherever there
is compositionality in your data,
Deep Learning seems to just work very, very well.
Like vision, text,
speech have some inherent compositionality in them,
and it just seems to learn features
in these kinds of domains really,
really well as against many other domains.
Then, obviously with the success
of Deep Learning in many applications,
we've also seen a proliferation
of Deep Learning frameworks.
So, now you have a lot of options coming from Google,
Facebook, Amazon, and Microsoft.
So, the graphic on the right,
kind of it's a little old,
it's about six months old,
it talks about the most searched framework.
I think as of six months back, it was TensorFlow.
But I think you'll take it with a pinch of salt,
these things keep changing with time.
And I think there are many other frameworks which are
popular and it depends
on the task that you're trying to achieve,
and what flexibility you
want while coding in that framework.
And of course, I think this is probably
an indication of how strong it is now.
It's in the business space, a lot of startups.
We're not going to get into naming any of them,
but just plenty of them at the space across the board,
so many application domains at this point.
So, let me take a few minutes to also talk about
the emerging directions in Deep Learning.
More at an algorithmic level,
a lot of these things are still in
development and have not gone
to a stage of
deployment or reaching the consumer I would say,
again, exceptions are always there but to a large extent.
And I think one of
the hottest areas in this space is GANs,
like deep generative models.
I think Generative Adversarial Networks have
been quite hot in the last couple of years.
Also, there are other frameworks
too Variational Autoencoders, pixel CNNs,
many other ways of generating images,
videos, and of late even text and documents.
So, in fact, I think there's a pretty popular post
on Quora by Yann LeCun
that says that adversarial training
is the coolest thing since sliced bread.
Okay. So, I'm quite sure it was an exaggeration,
but I think it was meant to say that it's
a pretty powerful training method.
So, many possible applications
in this space, art as an example.
And I think there are already some applications out there
which use GANs for esthetics and art.
But I think the broader capability of GANs is
perhaps the potential to contribute
to perhaps unsupervised or semi-supervised learning,
where you can probably
generate data similar to some dataset,
very heavily limited data and
probably train models which
can work reliably in a robust manner.
So, of course I think that space has
been hardly explored at this point in time.
The other thing that people have been trying
to do successful to a limited
extent is transfer learning.
How do you take Deep Learning to newer and newer tasks?
Okay. So, I think to a large extent within
similar domains between
very similar tasks it has worked very well.
I think everybody takes AlexNet and
modifies it to a new vision task and so on and so forth.
But I think what we mean by transfer learning
here is to take things to new domains,
where there's very little data,
very little annotation, very little expertise even.
So, can you actually use Deep Learning for
translating models into these kinds of spaces.
And of course, some examples here
would be Zero-shot learning,
One-shot learning, Few-shot learning,
where you have to classify data in one of
these classes but then you don't have
any data or have very little data from these classes.
So then, how do you transfer
models in these kinds of spaces?
Another important I think
algorithmic development that's come about
in Deep Learning the last few years is
the concept of attention and memory.
I think it started with natural language processing,
I think in 2015.
I think there are better experts here.
But since then it's been used for
images and video processing for captioning,
for visual question answering,
and also an interesting dimension is to go
into memory networks and Neural Turing Machines,
where attention on memory helps
you say read and write from memory,
and simulate a Turing Machine using Neural Networks.
So, there has been some work there about,
I mean probably I think if you google
Neural Turing Machines you can read more about it.
So, in fact, there's a paper in
late 2017 that say's that attention is all you need.
They say that you don't need RNNs at all,
you can kind of achieve
network similar to RNNs
with normal networks with just attention.
Okay. So that's one of the claims that they have
which probably makes sense
but hasn't really proven itself.
And of course I think this discussion would be
incomplete without other efforts that have
popped up in the last couple of years especially
on understanding the theory
behind Deep Learning and of course,
understanding the error surfaces.
Finally, Deep Learning is
all about navigating the error surface.
That's all Deep Learning is all
about in the training process.
So, I think the theory of
Deep Learning last year there was a paper based.
They said there they came up with something
called the information bottleneck principle.
I think I won't go into details today.
You can Google up if you're
interested by someone in Israel if I'm right.
By a group in Israel. And they say
that that's the reason why Deep Learning generalize well.
And more recently there was also a paper by Kawaguchi,
Bengio and others on
trying to study generalization and Deep Learning.
They actually came up with a bound, and
they said that the bound actually
depends on how many times you
validate your deep network on a different validation set.
Okay, which was interesting.
Which was an interesting way of
studying the generalization.
And there's also some work on understanding Deep Learning
using Random Matrix theory from Google Brain primarily.
In terms of understanding error surfaces,
I think there has been a lot
of interesting work in this space.
So, in fact, we do some work in this space
too and I'll be happy to
discuss offline about those things.
So, there was a work by Kawaguchi MIT on
Deep Learning with a local minima where
he claimed that under certain conditions,
when you have really very vast
parameter spaces such as Neural Networks,
all local minima are actually global minima.
Okay, they're all global minima. So you either
have saddle points or you
have global minima under
certain conditions and constraints of course.
So, but there have been various angles.
There's a group of Michael Jordan's group
has some papers on how to escape
saddle points efficiently using
portal gradient descent methods,
not directly tested on Deep Learning though.
More than the theoretical space at this point.
There's also some work on trying to understand
what kind of local minima are actually
good for Deep Learning and there's a paper
called Sharp Minima can Generalize for Deep Nets.
Where they say that how flat
a local minima is helps
in telling how generalizable it is.
More flat, more generalizable.
Okay, so that's what they claim.
And they have some interesting work
on how do you transform,
how do you go from a sharp minima to
a flat minima without changing
the cost function value, okay.
It's just an interesting thing again.
Of course, another thing here is efficient Deep Learning.
So, how do you make Deep Learning work on edge devices?
So, we all know AlexNet has
about 60 million parameters which boils
down to about 200 MB
plus and VGGNet takes about 500 MB plus.
So, how do you really make these things
efficient on edge devices?
Efficiency could mean storage wise,
efficiency could mean compute wise,
could be power wise, energy-wise,
whatever efficiently means.
So, there's lots of interest
growing in this space obviously
because I think many companies
want to take Deep Learning to hardware,
may make it a deployable on hardware directly.
I think one of the best works in
this space has been deep compression from
Stanford which actually won
the best paper award in ICLR 2016.
Where they just had a sequence of a pipeline
of simple things which brought down,
which got about 50X compression
on VGG with zero loss in performance.
Absolutely zero loss in performance.
Of course, since then there have been many other methods
more recent ones include knowledge distillation,
binarized neural nets and so on and so forth.
I think there's a pretty good survey that
came out recently on the various kinds of
deep model compression and acceleration methods
that if you're interested you can probably look at.
There's also now more recent work on doing Deep Learning,
but the input being a graph.
So far we've seen text,
we've seen speech, we've seen images.
One is the input to the Neural Network is a graph.
So, for the last couple
of years again there's been some
interesting traction in this space.
There's also a new sub-area matching
called Geometric Deep Learning,
which is about how do you do
Deep Learning on non-euclidean spaces?
So again probably I'll encourage you to go visit
that website to know more about it.
Of course, Deep Learning meets physics.
So, we'll probably covered this a little bit more in
probably the last few slides that I have.
So, the last few slides that I have is of course,
I think we've all seen it in
different forums over the last since yesterday,
as to what are the limitations,
why is Deep Learning not good?
So, we've covered a few of these things
in the next few slides and
probably I'll stop with that for the next stop.
As you've already seen interpretability and
explainability is one of the key limitations.
So,I think there are two issues here.
Why Deep Learning models work,
and how Deep Learning models work?
I think both are an issue.
I think why Deep Learning models work,
I think is more about the theory and trying to
study its generalization capabilities
and things like that.
But what we're talking about here is how
do Deep Learning models
really work. I mean how do they work?
I think so far most of the effort in this space has
been in trying to understand visualizing the weights,
probably trying to get a peek a little bit
more into some of the examples especially in
the image space has been trying to see if
a particular model classified
an image as a particular class.
Okay, what did it really see in
that image to give that particular class label?
Then you try to identify which portion of
the image the model was looking on.
So, you have some methods such as grad Cam.
I mean we have some work or plus plus in this space.
So, I think most of
the efforts in this space has been how
do you visualize the weights and understand them?
There's something called backprop to image
and you play around with the system to do it.
But I think what we mean by
interperability is something much much more.
I think there's a long way to go here.
I think we talked about it in
multiple sessions over the last couple of days.
Then what we really want from
Deep Learning Networks is
to give rationale for decisions.
It's not okay getting a peek into
the black boxes one step forward, which is great,
but I think what we're really looking for is for
the Deep Learning models to rationalized
their decisions and tell us why they did something.
And obviously that's when these models would go into
healthcare and other risks
sensitive domains where life
could be at stake when a decision is made.
Another important thing is the need for Causal Inference.
So there's a difference between
causality and correlation.
And it's important to understand
the Deep Learning models capture
correlation not causality, right?
So, this is just an example
here of a statistic where ice cream sales and
shark attacks have very similar statistics
between the months of say January to November.
Probably Siberia or someplace like that.
And obviously there's no causality here.
Okay, shark attacks is not
causing people to go eat ice cream.
Okay? So, obviously there's more to
understanding Causal Inference in Deep Learning.
I think there is some recent efforts from
Bernard Berofsky group in Max Planck.
So, in fact they had a paper last CVPR too.
But again, a long way to go.
A long, long, long way to go here in terms of
understanding causal relationships in
data automatically using Deep Learning.
And I think Professor Raj already spoke yesterday
about this, robustness and consistency.
Okay? So we already saw this yesterday,
where given an image which was
classified as the correct label,
a little bit of distortion of noise,
that's what you see here.
It gets classified as an ostrich.
So the third image is an ostrich in all those images.
Okay? So, in fact there's
another example that they have in the same paper.
So, the paper in CVPR of 2015.
Where all of these images have their corresponding labels
and they've actually classified with
99 percent confidence.
Okay, so you see. As you all can see,
you can see a cheetah there,
a robin there and a centipede there.
Okay? So, if you don't it's
your fault not Deep Learning's, okay?
But I mean this set
of images were generated by
taking just random noise images,
and adding a small component of what they
thought network learnt as
a representation of a cat let's say.
So, you can actually do simple things like back prop to
image for all cat images and try
to understand what is that base image
that the network thinks is a cat, okay?
And then you add a small portion of it to a noisy image,
and then boom the network thinks it's a cat now.
Okay? So, clearly this does not reflect human cognition.
So it seems to be learning some
discriminative features which it
thinks is a cat and
then these are the results that you actually see.
Okay, so, again a long way to go in terms of
robustness and consistency of
results across various kinds of data.
Integrating domain knowledge.
So, this is a recent work last year, 2016,
which talked about learning physical intuition
of block towers by example,
where they try to build
block towers and try to predict when it would fall.
Okay. So, an obvious question here is,
why don't we integrate physics laws?
Why should it be data driven?
Okay? So, why don't we integrate prior knowledge?
Why don't we integrate priors?
And obviously, this also connects to bringing together,
say, Bayesian approaches with Deep Learning.
So, that's another space that I think that's
reasonably open and needs progress.
Take one moment. Hyperparameters and Engineering.
Okay. I think I'll quickly finish.
I think all of us know this.
This is the elephant in the room, the hardest part of
deep learning is to find out what hyperperameters.
So, this is actually a
recent paper that's to be published,
where for a deep reinforcement learning
algorithm they take the same method,
two implementations, and get results and they found
that the results completely
vary with a huge amount of variance.
Okay. So, which clearly means questions,
the very reproducibility of these kinds of
models across just implementations.
Okay. Not even datasets.
Okay. I'll conclude with this last slide.
So, there's actually a paper called Deep Learning:
A Critical Appraisal that came up a couple of weeks back,
which seems to be reflecting some thoughts about,
is deep learning hitting a wall?
So, there's actually Francois Chollet's comment
out there, which says,
"For most problems where Deep Learning has enabled
transformationally better solutions in vision and speech,
we've entered diminishing returns territory
in the last year or so."
Okay. Of course, I think, in today's world,
at about three to six months,
if you don't get good results,
I mean, that's one generation gone, right?
So, that's how fast things
are going at this point in time.
But, things have slowed down at this point in time in
domains where we have known
good performance, so that, kind of,
begs the question, while we escape all kinds of
local saddle points and
probably hit better local minima while training,
has the field itself hit a local minima?
Of course, there Hinton
recently talked about moving beyond
backprop and I think that's
potentially another direction to look at.
I think I'll stop here.
I'm sorry if I exceeded time and I know we're
waiting for the big talk today.
>> Hello sir.
My question is more of a philosophical.
Why is it called Deep Learning?
I mean, we work so much and,
you know, combine and supervise the purchase also.
And GAN is handling
unsupervised problems and all and included meta-learning?
>> It's just another way of calling neural networks.
In fact, Hinton says in one of his interviews
that he just wanted to coin
a fancy word that makes people take notice.
So that's why we came up with Deep Learning.
So, I think, it's finely, just a name.
>> You discuss about the visual attention.
If I will apply Deep Learning for
making the saliency model, saliency maps.
So, how we will value it because most
of the complex this,
where people [inaudible] more
salient and another 50 percent
is telling region B is the most salient.
So, how will it be validated
because we don't have the real ground truth?
>> Sure. I mean,
I was talking about visual
attention in a different context.
I think saliency is a different way of
looking at visual attention.
In this case, what we're talking about is when we,
let's say, you do something, like, image captioning.Okay?
So, when you try to capture
an image, the typical, I mean, today,
one of the most popular approaches is to ask
the network to decide on
a particular part of the image to look at,
based on that you come up with a word.
Then, the network looks at another part
of the image, comes up with another word,
so on, and so forth, and then you
string the words together to make a caption.
Okay? So, the challenge there is to find out,
which part of the image should
my network look at, at a particular point in time?
So, I think there are methods for that.
There's something called hard attention,
soft attention, and things like that.
So, that's slightly different from saliency.
Saliency, I think, is a different thing,
I think in that case the labeling is subjective.
So, I mean, that's a separate problem by itself,
I think, maybe, we can talk offline about it.
I think in those kinds of cases,
you will have to do some kind of
crowdsource, vote aggregation,
or something like that, to come up with some kind
of a robust ground-truth
before you evaluate those kinds of models.
But today, I think, for a large part with
all of these methods and model saliency,
people just try to subjectively
evaluate and see how things work.
>> Thank you so much.
>> Hi. So, once you said
that this thing that Deep Learning is nothing,
but actually the does studies
related to neural networks only.
>> Yes. that's correct.
>> I just wanted to ask the question that,
if it's related to Neural Networks, then,
how much of the concepts or the theories established for
the Neural Networks are true
actually for the real brain as well,
where the extra neutrons exist?
Is anybody studying- are
any kind of studies being made related to that as well?
>> I think it's broadly understood today that
Neural Networks or Deep Learning as we
see are inspired by the human brain,
but they don't mimic or emulate or
simulate or anything in that context.
So, I think the model of the
neuron and all that is kind of
borrowed from the studies on the human brain,
but beyond that, I think,
there's not direct similarity at this point.
>> Are any scientific groups actually working to make
any kind of similarity or
any correlation between the two?
>> I mean there is-
>> There has to be a good mix
between the medical sciences
as well as the computer scientist as well.
>> I think, there is
a lot of work and personally, I think,
the one group of work that I'm personally aware of
is Tomaso Poggio's group in MIT,
I think they do some work in this space.
But at least, I have not seen
something convincing in that space,
to be honest, I mean, that's my understanding.
>> If you're available then I would
like to talk to you after break.
>> Sure. >> Thank you so much.
>> Yeah. Thank you.
>> Okay. Are there any more questions for Vineeth?
>> Okay. One last.
I think, people are trying to set it up.
So, maybe, we should use that time to have a question.
>> Hi. So, I want to ask like,
most of the part of your presentation was about,
like images, images, and images,.
>> Okay. All right. Okay.
>> So, how about text and other things?
>> Sure. Okay. So, in fact,
there is an interesting thing I wanted
to start with the disclaimer.
So, of late in Machine Learning,
there's been a lot of discussion
about bias in Machine Learning.
So, I have to disclaim that I work
a little more towards vision,
so that's why you had the bias in
these slides towards vision.
So, that was the only reason.
Okay. So, it's more, that I work in vision,
so most of the slides were towards vision,
but I think, maybe if I can,
if I have a couple of minutes.
So, if you look a few years back,
I think the fields of speech,
vision, NLP, all of them were very far away,
each of them had their own conferences.
I think, the feature extraction used to be
so different in these domains,
that there was not, of course,
the Machine Learning algorithms are similar,
but feature extraction used to be so different,
that there was not
much of a crosstalk between these domains.
But I think, with Deep Learning and
Feature Extraction being automated,
I think the methods that apply to all these domains,
are, kind of, coming together.
And that's the reason today you have so much
of crosstalk between these modalities.
Because the step that differentiated them,
which is Feature Extraction is now
gone out of the picture, right?
So, I think a lot of these methods are still relevant to
text but I think there are a lot of experts
here on text and speech,
I would definitely recommend
you to talk to them about it.
I'm sure they know more than that.
>> Sure. Thank you.
>> Okay. I think we will not, you know,
hold the next talk any longer.
With that, thank you again Vineeth.
Thank you Vineeth again. Yeah.
>> Thank you.

For more infomation >> News Conference: UMBC & Kansas St - Postgame - Duration: 45:49.
For more infomation >> News Conference: Florida State & Xavier - Postgame - Duration: 27:31. 


For more infomation >> EDLが解説!【YouTube】再生リストの説明欄を編集する方法 - Duration: 0:57. 


Không có nhận xét nào:
Đăng nhận xét