Top Banner
Convolutional Neural Network in Practice 2016.11 [email protected]
162

Convolutional neural network in practice

Jan 06, 2017

Download

Software

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Convolutional neural network in practice

Convolutional Neural Network in Practice

2016.11 [email protected]

Page 2: Convolutional neural network in practice

Preliminaries

Page 3: Convolutional neural network in practice

Buzz words nowadays

AIDeep

learning

Big dataMachine learning

Reinforcement Learning

???

Page 4: Convolutional neural network in practice

Glossary of AI terms

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

Page 5: Convolutional neural network in practice

Definitions

What is AI ?

“Artificial intelligence is that activity devoted to making machines intelligent, and intelligence is that quality that enables an entity to function appropriately and with foresight in its environment.”

Nils J. Nilsson, The Quest for Artificial Intelligence: A History of Ideas and Achievements (Cambridge, UK: Cambridge University Press, 2010).

“a computerized system that exhibits behavior that is commonly thought of as requiring intelligence”

Executive Office of the President National Science and Technology Council Committee on Technology: PREPARING FOR THE FUTURE OF ARTIFICIAL INTELLIGENCE (2016).

“any technique that enables computers to mimic human intelligence”

Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

Page 6: Convolutional neural network in practice

My diagram of AI terms

Environment

Data, Rules, Feedbacks ...

Teaching

Self-Learning,Engineering

...

AI

y = f(x)

Catf F18f

Page 7: Convolutional neural network in practice

Past, Present of AI

Page 8: Convolutional neural network in practice

Decades-old technology

● Long long history. From 1940s …

● But,

○ Before Oct. 2012.

○ After Oct. 2012.

Page 9: Convolutional neural network in practice

Venn diagram of AI terms

From Ian Goodfellow, Deep Learning (MIT press, 2016).

Page 10: Convolutional neural network in practice

Performance Hierarchy

Data

Features

Algorithms

Page 11: Convolutional neural network in practice

Flowcharts of AI

From Ian Goodfellow, Deep Learning (MIT press, 2016).

E2E(end-to-end)

Page 12: Convolutional neural network in practice

Image recognition error rate

From https://www.nervanasys.com/deep-learning-and-the-need-for-unified-tools/

2012

Page 13: Convolutional neural network in practice

Speech recognition error rate

2012

Page 14: Convolutional neural network in practice

5 Tribes of AI researchers

Symbolists(Rule, Logic-based)

Connectionists(PDP assumption)

Bayesians EvolutionistsAnalogizers

vs.

Page 15: Convolutional neural network in practice

Deep learning has had a long and rich history !

● 3 re-brandings.

○ Cybernetics ( 1940s ~ 1960s )

○ Artificial Neural Networks ( 1980s ~ 1990s)

○ Deep learning ( 2006 ~ )

Page 16: Convolutional neural network in practice

Nothing new !

● Alexnet 2012

○ based on CNN ( LeCunn, 1989 )

● Alpha Go

○ based on Reinforcement learning and

MCTS ( Sutton, 1998 )

Page 17: Convolutional neural network in practice

So, why now ?

● Computing Power

● Large labelled dataset

● Algorithm

Page 18: Convolutional neural network in practice

Size of neural networks

From Ian Goodfellow, Deep Learning (MIT press, 2016).

Singularity or Transcendence ?

Page 19: Convolutional neural network in practice

Depth is KING !

Page 20: Convolutional neural network in practice

Brief history of deep learning

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

1st Boom 2nd Boom1st Winter

Page 21: Convolutional neural network in practice

Brief history of deep learning

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

Page 22: Convolutional neural network in practice

Brief history of deep learning

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

2nd Winter

Page 23: Convolutional neural network in practice

Brief history of deep learning

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

3rd Boom

Page 24: Convolutional neural network in practice

Brief history of deep learning

From Roger Parloff, WHY DEEP LEARNING IS SUDDENLY CHANGING YOUR LIFE (Fortune, 2016).

Page 25: Convolutional neural network in practice

So, when 3rd winter ?

Nope !!!

● Features are mandatory in every AI problem.

● Deep learning is cheap learning! (Though someone can disprove the PDP assumptions, deep learning is the best practical tool in representation learning.)

Page 26: Convolutional neural network in practice

Biz trends after Oct.2012.

● 4 big players leading this sector.

● Bloody hiring war.○ Along the lines of NFL football players.

Page 27: Convolutional neural network in practice

Biz trend after Oct.2012.

● 2 leading research firms.

● 60+ startups

Page 28: Convolutional neural network in practice

Biz trend after Oct.2012.

Page 29: Convolutional neural network in practice

Future of AI

Page 30: Convolutional neural network in practice

Venn diagram of ML

From David silver, Reinforcement learning (UCL cource on RL, 2015).

Page 31: Convolutional neural network in practice

Unsupervised & Reinforcement Learning

● 2 leading research firms focus on:

○ Generative Models

○ Reinforcement Learning

Page 32: Convolutional neural network in practice

Towards General Artificial Intelligence

Page 33: Convolutional neural network in practice

Towards General Artificial Intelligence

Strong AI vs. Weak AIGeneral AI vs. Narrow AI

Page 34: Convolutional neural network in practice

Towards General Artificial Intelligence

Page 35: Convolutional neural network in practice

Towards General Artificial Intelligence

Page 36: Convolutional neural network in practice

Generative Adversarial Network

Xi Chen et al, InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets ( 2016 )

Page 37: Convolutional neural network in practice

Generative Adversarial Network

(From https://github.com/buriburisuri/supervised_infogan 2016)

Page 38: Convolutional neural network in practice

So what can we do with AI?

● Simply, it’s sophisticated software

writing software.

True personalization at scale!!!

Page 39: Convolutional neural network in practice

Is AI really necessary ?

“a lot of S&P 500 CEOs wished they had started thinking sooner than they did about their Internet strategy. I think five years from now there will be a number of S&P 500 CEOs that will wish they’d started thinking earlier about their AI strategy.”

“AI is the new electricity, just as 100 years ago electricity transformed industry after industry, AI will now do the same.”

Andrew Ng., chief scientist at Baidu Research.

Page 40: Convolutional neural network in practice

Conclusion

Computers have opened their eyes.

Page 41: Convolutional neural network in practice

Convolution Neural Network

Page 42: Convolutional neural network in practice

Convolution Neural Network

● Motivation

○ Sparse connectivity

■ smaller kernel size

○ Parameter sharing

■ shared kernel

○ Equivariant representation

■ convolution operation

Page 43: Convolutional neural network in practice

Fully Connected(Dense) Neural Network

● Typical 3-layer fully connected neural network

Page 44: Convolutional neural network in practice

Sparse connectivity vs.Dense connectivity

Sparse

Dense

From Ian Goodfellow, Deep Learning (MIT press, 2016).

Page 45: Convolutional neural network in practice

Parameter sharing

(x1, s1) ~ (x5, s5) share a single

parameter

From Ian Goodfellow, Deep Learning (MIT press, 2016).

Page 46: Convolutional neural network in practice

Equivariant representation

Convolution operation

satisfies equivariant property.

Page 50: Convolutional neural network in practice

Basic module of 2D CNN

Page 51: Convolutional neural network in practice

Pooling

● Average pooling = L1 pooling

● Max pooling = infinity norm pooling

Page 52: Convolutional neural network in practice

Max Pooling

● To improve translation invariance.

Page 53: Convolutional neural network in practice

Parameters of convolution

● Kernel size○ ( row, col, in_channel, out_channel)

● Padding

○ SAME, VALID, FULL

● Stride

○ if S > 1, use even kernel size F >

S * 2

Page 54: Convolutional neural network in practice

1 dimensional convolution

pad(P=1) pad(P=1) pad(P=1)

stride(S=1)

kernel(F=3)

stride(S=2)

● ‘SAME’(or ‘HALF’) pad size = (F - 1) * S / 2● ‘VALID’ pad size = 0● ‘FULL’ pad size : not used nowadays

Page 55: Convolutional neural network in practice

2 dimensional convolution

From : https://github.com/vdumoulin/conv_arithmetic

pad = ‘VALID’, F = 3, S = 1

Page 56: Convolutional neural network in practice

2 dimensional convolution

From : https://github.com/vdumoulin/conv_arithmetic

pad = ‘SAME’, F = 3, S = 1

Page 57: Convolutional neural network in practice

2 dimensional convolution

From : https://github.com/vdumoulin/conv_arithmetic

pad = ‘SAME’, F = 3, S = 2

Page 58: Convolutional neural network in practice

Artifacts of strides

From : http://distill.pub/2016/deconv-checkerboard/

F = 3, S = 2

Page 59: Convolutional neural network in practice

Artifacts of strides

F = 4, S = 2

From : http://distill.pub/2016/deconv-checkerboard/

Page 60: Convolutional neural network in practice

Artifacts of strides

From : http://distill.pub/2016/deconv-checkerboard/

F = 4, S = 2

Page 61: Convolutional neural network in practice

Pooling vs. Striding

● Same in the downsample aspect

● But, different in the location aspect

○ Location is lost in Pooling

○ Location is preserved in Striding

● Nowadays, striding is more popular

○ some kind of learnable pooling

Page 62: Convolutional neural network in practice

Kernel initialization

● Random number between -1 and 1

○ Orthogonality ( I.I.D. )

○ Uniform or Gaussian random

● Scale is paramount.

○ Adjust such that out(activation)

values have mean 0 and variance 1

○ If you encounter NaN, that may be

because of ill scale.

Page 63: Convolutional neural network in practice

Gabor Filter

Page 64: Convolutional neural network in practice

Activation results

Page 65: Convolutional neural network in practice

Initialization guide

● Xavier(or Glorot) initialization

○ http://jmlr.org/proceedings/papers/v9/glorot10a/glorot10a

.pdf

● He initialization

○ Good for RELU nonlinearity

○ https://arxiv.org/abs/1502.01852

● Use batch normalization if possible○ Immune to ill-scaled initialization

Page 66: Convolutional neural network in practice

Image classification

Page 67: Convolutional neural network in practice

Guide

● Start from robust baseline

○ 3 choices

■ VGG, Inception-v3, Resnet

● Smaller and deeper

● Towards getting rid of POOL and

final dense layer

● BN and skip connection are popular

Page 68: Convolutional neural network in practice

VGG

Page 69: Convolutional neural network in practice

VGG

● https://arxiv.org/abs/1409.1556

● VGG-16 is good start point.

○ apply BN if you train from scratch

● Image input : 224x224x3 ( -1 ~ 1 )

● Final outputs

○ conv5 : 7x7x512

○ fc2 : 4096

○ sm : 1000

Page 70: Convolutional neural network in practice

VGG practical tricks

● If gray image

○ divide all feature nums by 2

● Replace FCs with fully convolutional

layers

○ variable size input image

○ training/evaluation augmentation

○ read 4~5 pages in this paper

Page 71: Convolutional neural network in practice

Fully connected layer

● conv5 output : 7x7x512

● Fully connected layer

○ flatten : 1x25088

○ fc1 weight: 25088x4096

■ output : 1x4096

○ fc2 weight: 4096x4096

■ output : 1x4096

○ Fixed size image only

Page 72: Convolutional neural network in practice

Fully convolutional layer● conv5 output : 7x7x512

● Fully convolutional layer

○ fc1 ← conv 7x7@4096

■ output : (row-6)x(col-6)x4096

○ fc2 ← conv 1x1@4096

■ output : (row-6)x(col-6)x4096

○ Global average pooling

■ output : 1x1x4096

○ Variable sized images

Page 73: Convolutional neural network in practice

VGG Fully convolutional layer

From : https://github.com/buriburisuri/sugartensor/blob/master/sugartensor/sg_net.py

Page 74: Convolutional neural network in practice

Google Inception

Page 75: Convolutional neural network in practice

Google Inception● https://arxiv.org/pdf/1512.00567.pdf

● Bottlenecked architecture.

○ 1x1 conv

○ latest version : v5 ( v3 is popular )

● Image input : 224x224x3 ( -1 ~ 1 )

● Final output

○ conv5 : 7x7x1024 ( or 832 )

○ fc2 : 1024

○ sm : 1000

Page 76: Convolutional neural network in practice

Batch Normalization● https://arxiv.org/pdf/1502.03167.pdf

Page 77: Convolutional neural network in practice

Batch normalization

● Extremely powerful

○ Use everywhere possible

○ Absorb biases to BN’s shifts

Page 78: Convolutional neural network in practice

Resnet

Page 79: Convolutional neural network in practice

Resnet

● https://arxiv.org/pdf/1512.03385v1.pdf

● Residual block

○ skip connection + stride

○ bottleneck block

● Image input : 224x224x3 ( -1 ~ 1 )

● Final output

○ conv5 : 7x7x2048

○ fc2 : 1x1x2048 ( average pooling )

○ sm : 1000

Page 80: Convolutional neural network in practice

Resnet

● Very deep using skip connection○ Now, v2 - 1001 layer architecture

● Now, Resnet-152 v2 is the de-facto standard

Page 82: Convolutional neural network in practice

Summary

● Start from Resnet-50

● Use He’s initialization

● learning rate : 0.001 (with BN), 0.0001

(without BN)

● Use Adam ( should be alpha < beta ) optim

○ alpha=0.9, beta=0.999 (with easy training)

○ alpha=0.5, beta=0.95 (with hard training)

Page 83: Convolutional neural network in practice

Summary

● Minimize hyper-parameter tuning or

architecture modification.

○ Deep learning is highly nonlinear and

count-intuitive

○ Grid or random search is expensive

Page 84: Convolutional neural network in practice

Visualization

Page 85: Convolutional neural network in practice

Kernel visualization

Page 86: Convolutional neural network in practice

Feature visualization

Page 87: Convolutional neural network in practice

t-SNE visualization

https://lvdmaaten.github.io/tsne/

Page 88: Convolutional neural network in practice

Occlusion chart

https://arxiv.org/abs/1311.2901

Page 89: Convolutional neural network in practice

Activation chart

http://yosinski.com/deepvishttps://www.youtube.com/watch?v=AgkfIQ4IGaM

Page 90: Convolutional neural network in practice

CAM : Class Activation Map

http://cnnlocalization.csail.mit.edu/

Page 91: Convolutional neural network in practice

Saliency Maps

From : http://cs231n.stanford.edu/slides/winter1516_lecture9.pdf

Page 92: Convolutional neural network in practice

Deconvolution approach

From : http://cs231n.stanford.edu/slides/winter1516_lecture9.pdf

Page 93: Convolutional neural network in practice

Augmentation

Page 94: Convolutional neural network in practice

Augmentation

● 3 types of augmentation

○ Traing data augmentation

○ Evaluation augmentation

○ Label augmentation

● Augmentation is mandatory○ If you have really big data, then augment

data and increase model capacity

Page 95: Convolutional neural network in practice

Training Augmentation● Random crop/scale

○ random L in range [256, 480]

○ Resize training image, short side = L

○ Sample random 224x224 patch

Page 96: Convolutional neural network in practice

Training Augmentation● Random flip/rotate

● Color jitter

Page 97: Convolutional neural network in practice

Training Augmentation● Random flip/rotate

● Color jitter

● Random occlude

Page 98: Convolutional neural network in practice

Testing Augmentation● 10-crop testing ( VGG )

○ average(or max) scores

Page 99: Convolutional neural network in practice

Testing Augmentation

● Multi-scale testing

○ Fully convolutional layer is mandatory

○ Random L in range [224, 640]

○ Resize training image such that short side

= L

○ Average(or max) scores

● Used in Resnet

Page 100: Convolutional neural network in practice

Advanced Augmentation● Homography transform

○ https://arxiv.org/pdf/1606.03798v1.pdf

Page 101: Convolutional neural network in practice

Advanced Augmentation● Elastic transform for medical image

○ http://users.loni.usc.edu/~thompson/MAP/warp.html

Page 102: Convolutional neural network in practice

Augmentation in action

Page 103: Convolutional neural network in practice

Other Augmentation● Be aggressive and creative!

Page 104: Convolutional neural network in practice

Feature level Augmentation● Exploit equivariant property of CNN

○ Xu shen, “Transform-Invariant Convolutional Neural Networks for Image Classification and

Search”, 2016

○ Hyo-Eun Kim, “Semantic Noise Modeling for Better Representation Learning”, 2016

Page 105: Convolutional neural network in practice

Image Localization

Page 106: Convolutional neural network in practice

Localization and Detection

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 107: Convolutional neural network in practice

Classification + Localization

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 108: Convolutional neural network in practice

Simple recipe

CE loss

L2(MSE) loss

Joint-learning ( Multi-task learning )or

Separate learning

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 109: Convolutional neural network in practice

Regression head position

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 110: Convolutional neural network in practice

Multiple objects detection

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 111: Convolutional neural network in practice

R-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 112: Convolutional neural network in practice

Fast R-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 113: Convolutional neural network in practice

Faster R-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 114: Convolutional neural network in practice

Faster R-CNN

● https://arxiv.org/pdf/1506.01497.pdf

● de-facto standard

From : http://cs231n.stanford.edu/slides/winter1516_lecture8.pdf

Page 115: Convolutional neural network in practice

Segmentation

Page 116: Convolutional neural network in practice

Semantic Segmentation

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 117: Convolutional neural network in practice

Naive recipe

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 118: Convolutional neural network in practice

Fast recipe

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 119: Convolutional neural network in practice

Multi-scale refinement

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 120: Convolutional neural network in practice

Recurrent refinement

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 121: Convolutional neural network in practice

Upsampling

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 122: Convolutional neural network in practice

Deconvolution

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 123: Convolutional neural network in practice

Skip connection

Olaf, U-Net: Convolutional Networks for Biomedical Image Segmentation, 2015

Page 124: Convolutional neural network in practice

Instance Segmentation

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 125: Convolutional neural network in practice

R-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 126: Convolutional neural network in practice

Hypercolumns

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 127: Convolutional neural network in practice

Cascades

From : http://cs231n.stanford.edu/slides/winter1516_lecture13.pdf

Page 128: Convolutional neural network in practice

Deconvolution

● Learnable upsampling

○ resize * 2 + normal convolution

○ controversial names■ deconvolution, convolution transpose, upconvolution,

backward strided convolution, ½ strided convolution

○ Artifacts by strides and kernel sizes■ http://distill.pub/2016/deconv-checkerboard/

○ Restrict the freedom of architectures

Page 129: Convolutional neural network in practice

Convolution transposed

From : https://arxiv.org/abs/1609.07009

Page 130: Convolutional neural network in practice

½ strided(sub-pixel) convolution

From : https://arxiv.org/abs/1609.07009

Page 131: Convolutional neural network in practice

ESPCN ( Efficient Sub-pixel CNN)

Periodic shuffle

Wenzhe, Real-Time Single Image and Video Super-Resolution Using and Efficient Sub-Pixel Convolutional Neural Network, 2016

Page 132: Convolutional neural network in practice

L2 loss issue

Christian, Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network, 2016

Page 133: Convolutional neural network in practice

SRGAN

https://github.com/buriburisuri/SRGAN

Page 134: Convolutional neural network in practice

Videos

Page 135: Convolutional neural network in practice

ST-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture14.pdf

Page 136: Convolutional neural network in practice

ST-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture14.pdf

Page 137: Convolutional neural network in practice

Long-Time ST-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture14.pdf

Page 138: Convolutional neural network in practice

Long-Time ST-CNN

From : http://cs231n.stanford.edu/slides/winter1516_lecture14.pdf

Page 139: Convolutional neural network in practice

Summary

● Model temporal motion locally ( 3D CONV )

● Model temporal motion globally ( RNN )

● Hybrids of both

● IMHO, RNN will be replaced with 1D

convolution dilated (atrous convolution)

Page 140: Convolutional neural network in practice

Unsupervised learning

Page 141: Convolutional neural network in practice

Stacked Autoencoder

Page 142: Convolutional neural network in practice

Stacked Autoencoder

● Blurry artifacts caused by L2 loss

Page 143: Convolutional neural network in practice

Variational Autoencoder

● Generative model

● Blurry artifacts caused by L2 loss

Page 144: Convolutional neural network in practice

Variational Autoencoder

● SAE with mean and variance regularizer

● Bayesian meets deep learning

Page 145: Convolutional neural network in practice

Generative Model

● Find realistic generating function G(x) by deep learning !!!

y = G(x)

G : Generating functionx : Factors

y : Output data

Page 146: Convolutional neural network in practice

GAN(Generative Adversarial Networks)

Ian. J. Fellow et al. Generative Adverserial Networks. 2014. ( https://arxiv.org/abs/1406.2661)

Page 147: Convolutional neural network in practice

Discriminator

Page 148: Convolutional neural network in practice

Generator

Page 149: Convolutional neural network in practice

Adversarial Network

Page 150: Convolutional neural network in practice

Results

( From Ian. J. Fellow et al. Generative Adverserial Networks. 2014. )

( From P. Kingma et al. Auto-Encoding Variational Bayes. 2013. )

Page 151: Convolutional neural network in practice

Pitfalls of GAN

● Very difficult to train.

○ No guarantee to Nash Equilibrium.■ Tim Salimans et al, Improved Techniques for Training GANS, 2016.

■ Junbo Zhao et al, Energy-based Generative Adversarial Network,

2016.

● Cannot control generated data.

○ How can we condition generating

function G(x)?

Page 152: Convolutional neural network in practice

InfoGAN

Xi Chen et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets, 2016 ( https://arxiv.org/abs/1606.03657 )

● Add mutual Information regularizer for inducing latent codes to original GAN.

Page 153: Convolutional neural network in practice

InfoGAN

Page 154: Convolutional neural network in practice

Results

( From Xi Chen et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets)

Page 155: Convolutional neural network in practice

Results

Interpretable factors interfered on face dataset

Page 156: Convolutional neural network in practice

Supervised InfoGAN

Page 157: Convolutional neural network in practice

Results

(From https://github.com/buriburisuri/supervised_infogan)

Page 158: Convolutional neural network in practice

AC-GAN● Augustus, “Conditional Image Synthesis With Auxiliary Classifier GANs”,

2016

Page 159: Convolutional neural network in practice

Features of GAN

● Unsupervised

○ No labelled data used

● End-to-end

○ No human feature engineering

○ No prior nor assumption

● High fidelity

○ automatic highly non-linear pattern finding

⇒ Currently, SOTA in image generation.

Page 160: Convolutional neural network in practice

Skipped topics

● Ensemble & Distillation

● Attention + RNN

● Object Tracking

● And so many ...

Page 161: Convolutional neural network in practice

Computers have opened their eyes.

Page 162: Convolutional neural network in practice

Thanks