21 - Deep Learning - Regularization Part 5 [ID:15399]
50 von 67 angezeigt

Welcome everybody to deep learning. So today we want to conclude talking about different

regularization methods and we want to talk in particular about one more technique that

is called multitask learning. Also because I'm lazy, so you know. Multitask learning,

we want to extend the previous concept. So previously we only had one network for one

task and then we had transfer learning to reuse the network.

So that is transfer learning and it has been done in principle for many decades.

But the question is can we do more? Can we do it in a better way? And there are some

real world examples. For example, if you learn to play the piano and the violin, then in

both tasks you require good hearing, sense of rhythm, music notation and so on. So there

are some things that can be shared. Or also soccer and basketball training. Both require

stamina, speed, body awareness, body-eye coordination. So if you learn the one, then you typically

also have benefits for the other.

We're not starting from scratch because we've been doing this with humans for millennia.

So this would be even better than reusing. So you learn simultaneously and then provide

a better understanding of the shared underlying concepts.

So the idea now is that we train and network simultaneously on multiple related tasks.

So we adapt the loss function to assess performance for multiple tasks. And this then gives a

multitask learning that introduces a so-called inductive bias. We prefer a model that can

explain more than a single task.

The stuff that works best is really simple.

Also this reduces the risk of overfitting on one particular task and our model generalizes

better. So let's look at the setup. So we have some shared input layers. So these are

like the feature extraction layers and the representation layers. And then we split at

some point where we go into task specific layers and then evaluate it on task A, task

B, task C. And they may be very different but somehow related because otherwise it would

make sense to share the previous layers.

So several hidden layers are shared between all of the tasks. And as already shown by

Baxter in 97, multitask learning of n tasks reduces the chance of overfitting by an order

of n.

There is nothing more practical than a good theory.

Instead of hard sharing you can also do soft parameter sharing. Soft parameter sharing

would now introduce an additional loss. So you constrain the activations in the particular

layers to be similar. So each model has its own parameters but we somehow link them together

to perform similar extraction steps yet different extraction steps in the constrained layers.

And you can do that for example with an L2 norm or other norms that make them similar.

Now we still have to talk about the auxiliary tasks. So all of these tasks should have an

own purpose. You may also just include auxiliary tasks just because you want to create a more

stable network. So one example here is facial landmark detection by Zang. And they essentially

want to detect facial landmarks but this is impeded by occlusion and pose variances.

So they start simultaneously to learn landmarks and subtly related tasks like the face pose,

smiling, not smiling, glasses, no glasses, occlusion and gender. So they had this information

available and then you can set this up in a multitask learning framework as you see

here in the network architecture. And in the results you see that they then have the auxiliary

tasks here but they're actually interested in the facial landmarks and they compare this

to a CNN, a cascaded CNN and now they're multitask network with the auxiliary tasks and they

can show that also the landmark detection is improved by introduction of these auxiliary

tasks. So certain features may be difficult to learn for one task but it may be easier

for a related one. So the auxiliary tasks can help to steer the training in a specific

direction and we somehow include prior knowledge by choosing appropriate auxiliary tasks. And

of course then tasks can have different convergence rates so you can then also introduce task-based

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

00:06:54 Min

Aufnahmedatum

2020-05-09

Hochgeladen am

2020-05-10 00:56:13

Sprache

en-US

Deep Learning - Regularization Part 5

This video discusses multi-task learning.

Video References:
Lex Fridman's Channel

Further Reading:
A gentle Introduction to Deep Learning

Links:

Link - for details on Maximum A Posteriori estimation and the bias-variance decomposition
Link - for a comprehensive text about practical recommendations for regularization
Link - the paper about calibrating the variances

References:
[1] Sergey Ioffe and Christian Szegedy. “Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift”. In: Proceedings of The 32nd International Conference on Machine Learning. 2015, pp. 448–456.
[2] Jonathan Baxter. “A Bayesian/Information Theoretic Model of Learning to Learn via Multiple Task Sampling”. In: Machine Learning 28.1 (July 1997), pp. 7–39.
[3] Christopher M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Secaucus, NJ, USA: Springer-Verlag New York, Inc., 2006.
[4] Richard Caruana. “Multitask Learning: A Knowledge-Based Source of Inductive Bias”. In: Proceedings of the Tenth International Conference on Machine Learning. Morgan Kaufmann, 1993, pp. 41–48.
[5] Andre Esteva, Brett Kuprel, Roberto A Novoa, et al. “Dermatologist-level classification of skin cancer with deep neural networks”. In: Nature 542.7639 (2017), pp. 115–118.
[6] C. Ding, C. Xu, and D. Tao. “Multi-Task Pose-Invariant Face Recognition”. In: IEEE Transactions on Image Processing 24.3 (Mar. 2015), pp. 980–993.
[7] Li Wan, Matthew Zeiler, Sixin Zhang, et al. “Regularization of neural networks using drop connect”. In: Proceedings of the 30th International Conference on Machine Learning (ICML-2013), pp. 1058–1066.
[8] Nitish Srivastava, Geoffrey E Hinton, Alex Krizhevsky, et al. “Dropout: a simple way to prevent neural networks from overfitting.” In: Journal of Machine Learning Research 15.1 (2014), pp. 1929–1958.
[9] R. O. Duda, P. E. Hart, and D. G. Stork. Pattern Classification. John Wiley and Sons, inc., 2000.
[10] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. http://www.deeplearningbook.org. MIT Press, 2016.
[11] Yuxin Wu and Kaiming He. “Group normalization”. In: arXiv preprint arXiv:1803.08494 (2018).
[12] Kaiming He, Xiangyu Zhang, Shaoqing Ren, et al. “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification”. In: Proceedings of the IEEE international conference on computer vision. 2015, pp. 1026–1034.
[13] D Ulyanov, A Vedaldi, and VS Lempitsky. Instance normalization: the missing ingredient for fast stylization. CoRR abs/1607.0 [14] Günter Klambauer, Thomas Unterthiner, Andreas Mayr, et al. “Self-Normalizing Neural Networks”. In: Advances in Neural Information Processing Systems (NIPS). Vol. abs/1706.02515. 2017. arXiv: 1706.02515.
[15] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. “Layer normalization”. In: arXiv preprint arXiv:1607.06450 (2016).
[16] Nima Tajbakhsh, Jae Y Shin, Suryakanth R Gurudu, et al. “Convolutional neural networks for medical image analysis: Full training or fine tuning?” In: IEEE transactions on medical imaging 35.5 (2016), pp. 1299–1312.
[17] Yoshua Bengio. “Practical recommendations for gradient-based training of deep architectures”. In: Neural networks: Tricks of the trade. Springer, 2012, pp. 437–478.
[18] Chiyuan Zhang, Samy Bengio, Moritz Hardt, et al. “Understanding deep learning requires rethinking generalization”. In: arXiv preprint arXiv:1611.03530 (2016).
[19] Shibani Santurkar, Dimitris Tsipras, Andrew Ilyas, et al. “How Does Batch Normalization Help Optimization?” In: arXiv e-prints, arXiv:1805.11604 (May 2018), arXiv:1805.11604. arXiv: 1805.11604 [stat.ML].
[20] Tim Salimans and Diederik P Kingma. “Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks”. In: Advances in Neural Information Processing Systems 29. Curran Associates, Inc., 2016, pp. 901–909.
[21] Xavier Glorot and Yoshua Bengio. “Understanding the difficulty of training deep feedforward neural networks”. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence 2010, pp. 249–256.
[22] Zhanpeng Zhang, Ping Luo, Chen Change Loy, et al. “Facial Landmark Detection by Deep Multi-task Learning”. In: Computer Vision – ECCV 2014: 13th European Conference, Zurich, Switzerland, Cham: Springer International Publishing, 2014, pp. 94–108.

 

Tags

artificial intelligence deep learning machine learning pattern recognition Feedforward Networks overfitting meta-learning
Einbetten
Wordpress FAU Plugin
iFrame
Teilen