SlideShare a Scribd company logo
1


Deep Learning with Implicit Gradients
Shohei Taniguchi, Matsuo Lab (M1)
•
•
• 2
- Meta-Learning with Implicit Gradients
‣ MAML inner update 

iMAML
- RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing 

and Exploding Gradients?
‣ ERNN
2
Outline
1.
-
-
2.
- 1
‣ Implicit Reparameterization Gradients
3. Meta-Learning with Implicit Gradients
4. RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing and
Exploding Gradients?
3
4
• (e.g. 2 )
-
-
- NN
• (e.g. )
-
-
- 

( )
y = f (x) y = ax2
+ bx + c
f (x, y) = 0 x2
+ y2
= r2
x y 5
•
•
, ,


f(x, y) = 0
dy
dx
= −
∂f/∂x
∂f/∂y
= −
fx
fy
f(x, y) = 0 (x0, y0) fy (x0, y0)
x0 ∈ U y0 ∈ V g : U → V
{(x, g(x))|x ∈ U} = {(x, y) ∈ U × V| f(x, y) = 0}
6
• 1
-
- A
- B 

( )
• 2 Jacobian
f(x, y) = 0 (x0, y0)
fy (x0, y0)
x2
+ y2
− r2
= 0
y = r2
− x2
fy (r,0) = 2 × 0 = 0
y = ± r2
− x2
fy 7
( )
1.
- ( )
- iMAML
2.
-
- ERNN
8
( )
1.
- ( )
- iMAML
2.
-
- ERNN
9
Implicit Reparameterization Gradients
10
• NeurIPS 2018 accepted
•
- Michael Figurnov, Shakir Mohamed, Andriy Mnih
- DeepMind
• reparameterization
trick
•
iMAML ERNN
11
Reparameterization Trick
• VAE
•
reparameterization trick
•
𝔼q(z; ϕ) [log p (x|z)]−KL (q (z; ϕ)||p (z))
q ϵ = f (z; ϕ) =
z − μϕ
σϕ
ϵ ∼ 𝒩 (0,1)
ϕ ϵ
∇ϕ 𝔼q(z; ϕ) [log p (x|z)] = 𝔼p(ϵ)
[
∇ϕlog p (x|z)
z=f−1
(ϵ; ϕ)]
f f
12
Implicit Reparameterization Gradients
• 1 →
-
-
-
f
ϵ ∼ U (0,1) ϕ
z = f−1
(ϵ; ϕ)
∇ϕ 𝔼q(z; ϕ) [log p (x|z)] = 𝔼p(ϵ) [∇ϕlog p (x|z)]
= 𝔼p(ϵ) [∇zlog p (x|z)∇ϕz]
∇ϕz
13
Implicit Reparameterization Gradients
•
-
-
•
ϵ = f (z; ϕ) ⇔ f (z; ϕ) − ϵ = 0 z ϕ
∇ϕz = −
∇ϕ f (z; ϕ)
∇z f (z; ϕ)
= −
∇ϕ f (z; ϕ)
q (z; ϕ)
z q (z; ϕ) f−1
14
Meta-Learning with Implicit Gradients
15
• NeurIPS 2019 accepted
•
- Aravind Rajeswaran, Chelsea Finn, Sham Kakade, Sergey Levine
- MAML
• MAML
16
Model-Agnostic Meta-Learning (MAML)
•
-
- 1 (one-step adaptation)
θ*ML
:= argmin
θ∈Θ
F(θ),  where F(θ) =
1
M
M
∑
i=1
ℒ ( 𝒜lgi (θ), 𝒟test
i )
𝒜lgi (θ) = θ − α∇θℒ (θ, 𝒟tr
i )
17
MAML
•
-
• MAML
• 1 FOMAML
- FOMAML 

https://www.slideshare.net/DeepLearningJP2016/dl1maml
• iMAML
∇θF (θ) 𝒜lgi (θ)
18
Inner Loop
•
•
𝒜lg⋆
(θ) = argmin
ϕ′∈Φ
Gi (ϕ′, θ)
Gi (ϕ′, θ) = ̂ℒ (ϕ′)+
λ
2
ϕ′− θ
2
19
Outer Loop
• MAML outer loop
• inner loop
➡
θ ← θ − ηdθF(θ)
= θ − η
1
M
M
∑
i=1
d𝒜lgi(θ)
dθ
∇ϕℒi ( 𝒜lgi(θ))
(ϕ = 𝒜lgi(θ))
d𝒜lgi(θ)
dθ
20
Outer Loop
• inner loop
•
•
- adapt
ϕi ≡ 𝒜lg⋆
i (θ) = argmin
ϕ′∈Φ
Gi (ϕ′, θ)
∇ϕ′Gi (ϕ′, θ)
ϕ′=ϕi
= 0
∇ ̂ℒ(ϕi) + λ(𝒜lg⋆
i (θ) − θ) = 0
θ 𝒜lg⋆
(θ)
d𝒜lg⋆
(θ)
dθ
=
(
I +
1
λ
∇2 ̂ℒ (ϕi))
−1
ϕi 21
Outer Loop
• 2
① inner loop adapt
(SGD )
② 3
•
(
I +
1
λ
∇2 ̂ℒ (ϕi))
−1
ϕi
(
I +
1
λ
∇2 ̂ℒ (ϕi))
−1
∇ϕℒi ( 𝒜lgi(θ))
22
(CG )
•
•
Ax = b ⋯(1)
(1) f(x) =
1
2
xT
Ax − bT
x
x0 = 0,r0 = b − Ax0, p0 = r0
αk =
rT
k pk
pT
k Apk
xk+1 = xk + αk pk
rk+1 = rk − αkApk
pk+1 = rk+1 +
rT
k+1rk+1
rT
k rk
pk
23
(CG )
•
•
( 5 )
- (p22 ① )
‣ Appendix E
gi =
(
I +
1
λ
∇2 ̂ℒ (ϕi))
−1
∇ϕℒi ( 𝒜lgi(θ)) gi
(
I +
1
λ
∇2 ̂ℒ (ϕi))
gi = ∇ϕℒi ( 𝒜lgi(θ))
rk
𝒜lgi(θ)
24
iMAML
• inner loop
➡adapt
• outer loop inner loop
➡inner loop
‣ MAML 1
‣ iMAML Hessian-Free 2
adapt
25
•
- iMAML inner loop ( )
- FOMAML (CG )
- MAML 

(FOMAML ??)
O(1)
26
• Omniglot
- inner loop Hessian-Free iMAML
- iMAML way ( )
- FOMAML
27
• Mini-ImageNet
- Reptile (FOMAML )
-
??
28
iMAML
•
iMAML
• MAML
•
•
•
-
29
( )
1.
- ( )
- iMAML
2.
-
- ERNN
30
RNNs Evolving on an Equilibrium Manifold:
A Panacea for Vanishing and Exploding Gradients?
31
•
- Anil Kag, Ziming Zhang, Venkatesh Saligrama
- , MERL
• NeurIPS 2019 reject
•
RNN
•
•
32
RNN /
• RNN
- sigmoid tanh
• RNN /
- LSTM GRU
hk = ϕ (Uhk−1 + Wxk + b)
ϕ
∂hm
∂hn
=
∏
m≥k>n
∂hk
∂hk−1
=
∏
m≥k>n
∇ϕ (Uhk−1 + Wxk + b) U
33
RNN ODE
• RNN skip connection
(ODE)
• Neural ODE
- 

https://www.slideshare.net/DeepLearningJP2016/dlneural-ordinary-
differential-equations
dh(t)
dt
≜ h′(t) = ϕ (Uh(t) + Wxk + b)
⟹ hk = hk−1 + ηϕ (Uhk−1 + Wxk + b)
34
ODE
• ODE
• 1
➡ ( )
• ERNN
dh
dt
= f (h, x) f (h, x) = 0 ⋯(1)
(1) h x (h0, x0)
fh (h0, x0) (1)
h = g (x)
(h0, x0)
35
ERNN
• ERNN 



ODE
• 



➡ 

h′(t) = ϕ (U (h(t) + hk−1) + Wxk + b) − γ (h(t) + hk−1)
h′(t) = 0 hk
hk f (hk−1, h) = ϕ (U (h + hk−1) + Wxk + b) − γ (h + hk−1) = 0
∂h
∂hk−1
= −
∂f/∂hk−1
∂f/∂h
= − I
∂f/∂h
36
∂f/∂h
•
1. (sigmoid tanh OK)
2.
- 

( )
∂f
∂h
= ∇ϕ (U (h + hk−1) + Wxk + b) U
ϕ
U
37
•
• 5
•
-
h(0)
k
= 0
h(i+1)
k
= h(i)
k
+ η(i)
k [
ϕ
(
U (h(i)
k
+ hk−1) + Wxk + b
)
− γ (h(i)
k
+ hk−1)]
η(i)
k
38
HAR-2 RNN ERNN (log scale)
• RNN
• ERNN 1
∂hT
∂h1
39
• RNN 

ERNN
40
• SoTA
•
•
41
ERNN
• NN
1
RNN
•
• SoTA
• RNN
• accept
42
&
•
• iMAML ERNN
•
43

More Related Content

What's hot

「統計的学習理論」第1章
「統計的学習理論」第1章「統計的学習理論」第1章
「統計的学習理論」第1章
Kota Matsui
 
Sliced Wasserstein距離と生成モデル
Sliced Wasserstein距離と生成モデルSliced Wasserstein距離と生成モデル
Sliced Wasserstein距離と生成モデル
ohken
 
【DL輪読会】時系列予測 Transfomers の精度向上手法
【DL輪読会】時系列予測 Transfomers の精度向上手法【DL輪読会】時系列予測 Transfomers の精度向上手法
【DL輪読会】時系列予測 Transfomers の精度向上手法
Deep Learning JP
 
クラシックな機械学習の入門  9. モデル推定
クラシックな機械学習の入門  9. モデル推定クラシックな機械学習の入門  9. モデル推定
クラシックな機械学習の入門  9. モデル推定
Hiroshi Nakagawa
 
強化学習と逆強化学習を組み合わせた模倣学習
強化学習と逆強化学習を組み合わせた模倣学習強化学習と逆強化学習を組み合わせた模倣学習
強化学習と逆強化学習を組み合わせた模倣学習
Eiji Uchibe
 
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
Yahoo!デベロッパーネットワーク
 
PRML第6章「カーネル法」
PRML第6章「カーネル法」PRML第6章「カーネル法」
PRML第6章「カーネル法」
Keisuke Sugawara
 
勾配降下法の 最適化アルゴリズム
勾配降下法の最適化アルゴリズム勾配降下法の最適化アルゴリズム
勾配降下法の 最適化アルゴリズム
nishio
 
最適化計算の概要まとめ
最適化計算の概要まとめ最適化計算の概要まとめ
最適化計算の概要まとめ
Yuichiro MInato
 
数学で解き明かす深層学習の原理
数学で解き明かす深層学習の原理数学で解き明かす深層学習の原理
数学で解き明かす深層学習の原理
Taiji Suzuki
 
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
Yusuke Uchida
 
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
SSII
 
パターン認識と機械学習 §6.2 カーネル関数の構成
パターン認識と機械学習 §6.2 カーネル関数の構成パターン認識と機械学習 §6.2 カーネル関数の構成
パターン認識と機械学習 §6.2 カーネル関数の構成
Prunus 1350
 
[DL輪読会]近年のエネルギーベースモデルの進展
[DL輪読会]近年のエネルギーベースモデルの進展[DL輪読会]近年のエネルギーベースモデルの進展
[DL輪読会]近年のエネルギーベースモデルの進展
Deep Learning JP
 
強化学習その3
強化学習その3強化学習その3
強化学習その3
nishio
 
[DL輪読会]Meta Reinforcement Learning
[DL輪読会]Meta Reinforcement Learning[DL輪読会]Meta Reinforcement Learning
[DL輪読会]Meta Reinforcement Learning
Deep Learning JP
 
変分ベイズ法の説明
変分ベイズ法の説明変分ベイズ法の説明
変分ベイズ法の説明
Haruka Ozaki
 
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
Deep Learning JP
 
計算論的学習理論入門 -PAC学習とかVC次元とか-
計算論的学習理論入門 -PAC学習とかVC次元とか-計算論的学習理論入門 -PAC学習とかVC次元とか-
計算論的学習理論入門 -PAC学習とかVC次元とか-
sleepy_yoshi
 
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
Deep Learning JP
 

What's hot (20)

「統計的学習理論」第1章
「統計的学習理論」第1章「統計的学習理論」第1章
「統計的学習理論」第1章
 
Sliced Wasserstein距離と生成モデル
Sliced Wasserstein距離と生成モデルSliced Wasserstein距離と生成モデル
Sliced Wasserstein距離と生成モデル
 
【DL輪読会】時系列予測 Transfomers の精度向上手法
【DL輪読会】時系列予測 Transfomers の精度向上手法【DL輪読会】時系列予測 Transfomers の精度向上手法
【DL輪読会】時系列予測 Transfomers の精度向上手法
 
クラシックな機械学習の入門  9. モデル推定
クラシックな機械学習の入門  9. モデル推定クラシックな機械学習の入門  9. モデル推定
クラシックな機械学習の入門  9. モデル推定
 
強化学習と逆強化学習を組み合わせた模倣学習
強化学習と逆強化学習を組み合わせた模倣学習強化学習と逆強化学習を組み合わせた模倣学習
強化学習と逆強化学習を組み合わせた模倣学習
 
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
深層学習による自然言語処理入門: word2vecからBERT, GPT-3まで
 
PRML第6章「カーネル法」
PRML第6章「カーネル法」PRML第6章「カーネル法」
PRML第6章「カーネル法」
 
勾配降下法の 最適化アルゴリズム
勾配降下法の最適化アルゴリズム勾配降下法の最適化アルゴリズム
勾配降下法の 最適化アルゴリズム
 
最適化計算の概要まとめ
最適化計算の概要まとめ最適化計算の概要まとめ
最適化計算の概要まとめ
 
数学で解き明かす深層学習の原理
数学で解き明かす深層学習の原理数学で解き明かす深層学習の原理
数学で解き明かす深層学習の原理
 
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
Swin Transformer (ICCV'21 Best Paper) を完璧に理解する資料
 
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
SSII2021 [OS2-01] 転移学習の基礎:異なるタスクの知識を利用するための機械学習の方法
 
パターン認識と機械学習 §6.2 カーネル関数の構成
パターン認識と機械学習 §6.2 カーネル関数の構成パターン認識と機械学習 §6.2 カーネル関数の構成
パターン認識と機械学習 §6.2 カーネル関数の構成
 
[DL輪読会]近年のエネルギーベースモデルの進展
[DL輪読会]近年のエネルギーベースモデルの進展[DL輪読会]近年のエネルギーベースモデルの進展
[DL輪読会]近年のエネルギーベースモデルの進展
 
強化学習その3
強化学習その3強化学習その3
強化学習その3
 
[DL輪読会]Meta Reinforcement Learning
[DL輪読会]Meta Reinforcement Learning[DL輪読会]Meta Reinforcement Learning
[DL輪読会]Meta Reinforcement Learning
 
変分ベイズ法の説明
変分ベイズ法の説明変分ベイズ法の説明
変分ベイズ法の説明
 
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
[DL輪読会]Wasserstein GAN/Towards Principled Methods for Training Generative Adv...
 
計算論的学習理論入門 -PAC学習とかVC次元とか-
計算論的学習理論入門 -PAC学習とかVC次元とか-計算論的学習理論入門 -PAC学習とかVC次元とか-
計算論的学習理論入門 -PAC学習とかVC次元とか-
 
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
【DL輪読会】HyperTree Proof Search for Neural Theorem Proving
 

Similar to [DL輪読会]陰関数微分を用いた深層学習

Phase diagram at finite T & Mu in strong coupling limit of lattice QCD
Phase diagram at finite T & Mu in strong coupling limit of lattice QCDPhase diagram at finite T & Mu in strong coupling limit of lattice QCD
Phase diagram at finite T & Mu in strong coupling limit of lattice QCD
Benjamin Jaedon Choi
 
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
SEENET-MTP
 
Digital control systems (dcs) lecture 18-19-20
Digital control systems (dcs) lecture 18-19-20Digital control systems (dcs) lecture 18-19-20
Digital control systems (dcs) lecture 18-19-20
Ali Rind
 
勾配法
勾配法勾配法
勾配法
貴之 八木
 
Non Linear Dynamics Basics and Theory
Non Linear Dynamics Basics and TheoryNon Linear Dynamics Basics and Theory
Non Linear Dynamics Basics and Theory
Anupama Kate
 
Software Development for Space-group Analysis: Magnetic Space Group and Irred...
Software Development for Space-group Analysis: Magnetic Space Group and Irred...Software Development for Space-group Analysis: Magnetic Space Group and Irred...
Software Development for Space-group Analysis: Magnetic Space Group and Irred...
Kohei Shinohara
 
Singlevaropt
SinglevaroptSinglevaropt
Singlevaropt
sheetslibrary
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
inverse trigonometric function_1669522645.pptx
inverse trigonometric function_1669522645.pptxinverse trigonometric function_1669522645.pptx
inverse trigonometric function_1669522645.pptx
MasoudIbrahim3
 
Signal flow graph
Signal flow graphSignal flow graph
Signal flow graph
jani parth
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantification
Alexander Litvinenko
 
Ph ddefence
Ph ddefencePh ddefence
ALPSチュートリアル
ALPSチュートリアルALPSチュートリアル
Introducing Zap Q-Learning
Introducing Zap Q-Learning   Introducing Zap Q-Learning
Introducing Zap Q-Learning
Sean Meyn
 
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Gota Morota
 
May the Force NOT be with you
May the Force NOT be with youMay the Force NOT be with you
May the Force NOT be with you
Miguel Zuma
 
B. Sazdovic - Noncommutativity and T-duality
B. Sazdovic - Noncommutativity and T-dualityB. Sazdovic - Noncommutativity and T-duality
B. Sazdovic - Noncommutativity and T-duality
SEENET-MTP
 
Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...
Vladimir Bakhrushin
 
Quantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averagesQuantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averages
VjekoslavKovac1
 
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating HyperplaneESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
Shinichi Tamura
 

Similar to [DL輪読会]陰関数微分を用いた深層学習 (20)

Phase diagram at finite T & Mu in strong coupling limit of lattice QCD
Phase diagram at finite T & Mu in strong coupling limit of lattice QCDPhase diagram at finite T & Mu in strong coupling limit of lattice QCD
Phase diagram at finite T & Mu in strong coupling limit of lattice QCD
 
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
D. Vulcanov - On Cosmologies with non-Minimally Coupled Scalar Field and the ...
 
Digital control systems (dcs) lecture 18-19-20
Digital control systems (dcs) lecture 18-19-20Digital control systems (dcs) lecture 18-19-20
Digital control systems (dcs) lecture 18-19-20
 
勾配法
勾配法勾配法
勾配法
 
Non Linear Dynamics Basics and Theory
Non Linear Dynamics Basics and TheoryNon Linear Dynamics Basics and Theory
Non Linear Dynamics Basics and Theory
 
Software Development for Space-group Analysis: Magnetic Space Group and Irred...
Software Development for Space-group Analysis: Magnetic Space Group and Irred...Software Development for Space-group Analysis: Magnetic Space Group and Irred...
Software Development for Space-group Analysis: Magnetic Space Group and Irred...
 
Singlevaropt
SinglevaroptSinglevaropt
Singlevaropt
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
inverse trigonometric function_1669522645.pptx
inverse trigonometric function_1669522645.pptxinverse trigonometric function_1669522645.pptx
inverse trigonometric function_1669522645.pptx
 
Signal flow graph
Signal flow graphSignal flow graph
Signal flow graph
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantification
 
Ph ddefence
Ph ddefencePh ddefence
Ph ddefence
 
ALPSチュートリアル
ALPSチュートリアルALPSチュートリアル
ALPSチュートリアル
 
Introducing Zap Q-Learning
Introducing Zap Q-Learning   Introducing Zap Q-Learning
Introducing Zap Q-Learning
 
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
Application of Bayesian and Sparse Network Models for Assessing Linkage Diseq...
 
May the Force NOT be with you
May the Force NOT be with youMay the Force NOT be with you
May the Force NOT be with you
 
B. Sazdovic - Noncommutativity and T-duality
B. Sazdovic - Noncommutativity and T-dualityB. Sazdovic - Noncommutativity and T-duality
B. Sazdovic - Noncommutativity and T-duality
 
Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...Identification of the Mathematical Models of Complex Relaxation Processes in ...
Identification of the Mathematical Models of Complex Relaxation Processes in ...
 
Quantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averagesQuantitative norm convergence of some ergodic averages
Quantitative norm convergence of some ergodic averages
 
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating HyperplaneESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
ESL 4.4.3-4.5: Logistic Reression (contd.) and Separating Hyperplane
 

More from Deep Learning JP

【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
Deep Learning JP
 
【DL輪読会】事前学習用データセットについて
【DL輪読会】事前学習用データセットについて【DL輪読会】事前学習用データセットについて
【DL輪読会】事前学習用データセットについて
Deep Learning JP
 
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
Deep Learning JP
 
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
Deep Learning JP
 
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
Deep Learning JP
 
【DL輪読会】マルチモーダル LLM
【DL輪読会】マルチモーダル LLM【DL輪読会】マルチモーダル LLM
【DL輪読会】マルチモーダル LLM
Deep Learning JP
 
【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
 【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo... 【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
Deep Learning JP
 
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
Deep Learning JP
 
【DL輪読会】Can Neural Network Memorization Be Localized?
【DL輪読会】Can Neural Network Memorization Be Localized?【DL輪読会】Can Neural Network Memorization Be Localized?
【DL輪読会】Can Neural Network Memorization Be Localized?
Deep Learning JP
 
【DL輪読会】Hopfield network 関連研究について
【DL輪読会】Hopfield network 関連研究について【DL輪読会】Hopfield network 関連研究について
【DL輪読会】Hopfield network 関連研究について
Deep Learning JP
 
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
Deep Learning JP
 
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
Deep Learning JP
 
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
Deep Learning JP
 
【DL輪読会】"Language Instructed Reinforcement Learning for Human-AI Coordination "
【DL輪読会】"Language Instructed Reinforcement Learning  for Human-AI Coordination "【DL輪読会】"Language Instructed Reinforcement Learning  for Human-AI Coordination "
【DL輪読会】"Language Instructed Reinforcement Learning for Human-AI Coordination "
Deep Learning JP
 
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
Deep Learning JP
 
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
Deep Learning JP
 
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
Deep Learning JP
 
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
Deep Learning JP
 
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
Deep Learning JP
 
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
Deep Learning JP
 

More from Deep Learning JP (20)

【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
【DL輪読会】AdaptDiffuser: Diffusion Models as Adaptive Self-evolving Planners
 
【DL輪読会】事前学習用データセットについて
【DL輪読会】事前学習用データセットについて【DL輪読会】事前学習用データセットについて
【DL輪読会】事前学習用データセットについて
 
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
【DL輪読会】 "Learning to render novel views from wide-baseline stereo pairs." CVP...
 
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
【DL輪読会】Zero-Shot Dual-Lens Super-Resolution
 
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
【DL輪読会】BloombergGPT: A Large Language Model for Finance arxiv
 
【DL輪読会】マルチモーダル LLM
【DL輪読会】マルチモーダル LLM【DL輪読会】マルチモーダル LLM
【DL輪読会】マルチモーダル LLM
 
【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
 【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo... 【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
【 DL輪読会】ToolLLM: Facilitating Large Language Models to Master 16000+ Real-wo...
 
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
【DL輪読会】AnyLoc: Towards Universal Visual Place Recognition
 
【DL輪読会】Can Neural Network Memorization Be Localized?
【DL輪読会】Can Neural Network Memorization Be Localized?【DL輪読会】Can Neural Network Memorization Be Localized?
【DL輪読会】Can Neural Network Memorization Be Localized?
 
【DL輪読会】Hopfield network 関連研究について
【DL輪読会】Hopfield network 関連研究について【DL輪読会】Hopfield network 関連研究について
【DL輪読会】Hopfield network 関連研究について
 
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
【DL輪読会】SimPer: Simple self-supervised learning of periodic targets( ICLR 2023 )
 
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
【DL輪読会】RLCD: Reinforcement Learning from Contrast Distillation for Language M...
 
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
【DL輪読会】"Secrets of RLHF in Large Language Models Part I: PPO"
 
【DL輪読会】"Language Instructed Reinforcement Learning for Human-AI Coordination "
【DL輪読会】"Language Instructed Reinforcement Learning  for Human-AI Coordination "【DL輪読会】"Language Instructed Reinforcement Learning  for Human-AI Coordination "
【DL輪読会】"Language Instructed Reinforcement Learning for Human-AI Coordination "
 
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
 
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
【DL輪読会】"Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware"
 
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
【DL輪読会】Parameter is Not All You Need:Starting from Non-Parametric Networks fo...
 
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
【DL輪読会】Drag Your GAN: Interactive Point-based Manipulation on the Generative ...
 
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
【DL輪読会】Self-Supervised Learning from Images with a Joint-Embedding Predictive...
 
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
【DL輪読会】Towards Understanding Ensemble, Knowledge Distillation and Self-Distil...
 

Recently uploaded

GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...
ThomasParaiso2
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
mikeeftimakis1
 
Microsoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdfMicrosoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdf
Uni Systems S.M.S.A.
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Paige Cruz
 
A tale of scale & speed: How the US Navy is enabling software delivery from l...
A tale of scale & speed: How the US Navy is enabling software delivery from l...A tale of scale & speed: How the US Navy is enabling software delivery from l...
A tale of scale & speed: How the US Navy is enabling software delivery from l...
sonjaschweigert1
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
Neo4j
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
SOFTTECHHUB
 
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AI
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIEnchancing adoption of Open Source Libraries. A case study on Albumentations.AI
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AI
Vladimir Iglovikov, Ph.D.
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
名前 です男
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
Adtran
 
Free Complete Python - A step towards Data Science
Free Complete Python - A step towards Data ScienceFree Complete Python - A step towards Data Science
Free Complete Python - A step towards Data Science
RinaMondal9
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
Matthew Sinclair
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
Rohit Gautam
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
Safe Software
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
DianaGray10
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
Matthew Sinclair
 
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
SOFTTECHHUB
 

Recently uploaded (20)

GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
 
Microsoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdfMicrosoft - Power Platform_G.Aspiotis.pdf
Microsoft - Power Platform_G.Aspiotis.pdf
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
 
A tale of scale & speed: How the US Navy is enabling software delivery from l...
A tale of scale & speed: How the US Navy is enabling software delivery from l...A tale of scale & speed: How the US Navy is enabling software delivery from l...
A tale of scale & speed: How the US Navy is enabling software delivery from l...
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
 
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AI
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIEnchancing adoption of Open Source Libraries. A case study on Albumentations.AI
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AI
 
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
みなさんこんにちはこれ何文字まで入るの?40文字以下不可とか本当に意味わからないけどこれ限界文字数書いてないからマジでやばい文字数いけるんじゃないの?えこ...
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
 
Free Complete Python - A step towards Data Science
Free Complete Python - A step towards Data ScienceFree Complete Python - A step towards Data Science
Free Complete Python - A step towards Data Science
 
20240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 202420240607 QFM018 Elixir Reading List May 2024
20240607 QFM018 Elixir Reading List May 2024
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
 
UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5UiPath Test Automation using UiPath Test Suite series, part 5
UiPath Test Automation using UiPath Test Suite series, part 5
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
 
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
 

[DL輪読会]陰関数微分を用いた深層学習

  • 1. 1 
 Deep Learning with Implicit Gradients Shohei Taniguchi, Matsuo Lab (M1)
  • 2. • • • 2 - Meta-Learning with Implicit Gradients ‣ MAML inner update 
 iMAML - RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing 
 and Exploding Gradients? ‣ ERNN 2
  • 3. Outline 1. - - 2. - 1 ‣ Implicit Reparameterization Gradients 3. Meta-Learning with Implicit Gradients 4. RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing and Exploding Gradients? 3
  • 4. 4
  • 5. • (e.g. 2 ) - - - NN • (e.g. ) - - - 
 ( ) y = f (x) y = ax2 + bx + c f (x, y) = 0 x2 + y2 = r2 x y 5
  • 6. • • , , 
 f(x, y) = 0 dy dx = − ∂f/∂x ∂f/∂y = − fx fy f(x, y) = 0 (x0, y0) fy (x0, y0) x0 ∈ U y0 ∈ V g : U → V {(x, g(x))|x ∈ U} = {(x, y) ∈ U × V| f(x, y) = 0} 6
  • 7. • 1 - - A - B 
 ( ) • 2 Jacobian f(x, y) = 0 (x0, y0) fy (x0, y0) x2 + y2 − r2 = 0 y = r2 − x2 fy (r,0) = 2 × 0 = 0 y = ± r2 − x2 fy 7
  • 8. ( ) 1. - ( ) - iMAML 2. - - ERNN 8
  • 9. ( ) 1. - ( ) - iMAML 2. - - ERNN 9
  • 11. • NeurIPS 2018 accepted • - Michael Figurnov, Shakir Mohamed, Andriy Mnih - DeepMind • reparameterization trick • iMAML ERNN 11
  • 12. Reparameterization Trick • VAE • reparameterization trick • 𝔼q(z; ϕ) [log p (x|z)]−KL (q (z; ϕ)||p (z)) q ϵ = f (z; ϕ) = z − μϕ σϕ ϵ ∼ 𝒩 (0,1) ϕ ϵ ∇ϕ 𝔼q(z; ϕ) [log p (x|z)] = 𝔼p(ϵ) [ ∇ϕlog p (x|z) z=f−1 (ϵ; ϕ)] f f 12
  • 13. Implicit Reparameterization Gradients • 1 → - - - f ϵ ∼ U (0,1) ϕ z = f−1 (ϵ; ϕ) ∇ϕ 𝔼q(z; ϕ) [log p (x|z)] = 𝔼p(ϵ) [∇ϕlog p (x|z)] = 𝔼p(ϵ) [∇zlog p (x|z)∇ϕz] ∇ϕz 13
  • 14. Implicit Reparameterization Gradients • - - • ϵ = f (z; ϕ) ⇔ f (z; ϕ) − ϵ = 0 z ϕ ∇ϕz = − ∇ϕ f (z; ϕ) ∇z f (z; ϕ) = − ∇ϕ f (z; ϕ) q (z; ϕ) z q (z; ϕ) f−1 14
  • 16. • NeurIPS 2019 accepted • - Aravind Rajeswaran, Chelsea Finn, Sham Kakade, Sergey Levine - MAML • MAML 16
  • 17. Model-Agnostic Meta-Learning (MAML) • - - 1 (one-step adaptation) θ*ML := argmin θ∈Θ F(θ),  where F(θ) = 1 M M ∑ i=1 ℒ ( 𝒜lgi (θ), 𝒟test i ) 𝒜lgi (θ) = θ − α∇θℒ (θ, 𝒟tr i ) 17
  • 18. MAML • - • MAML • 1 FOMAML - FOMAML 
 https://www.slideshare.net/DeepLearningJP2016/dl1maml • iMAML ∇θF (θ) 𝒜lgi (θ) 18
  • 19. Inner Loop • • 𝒜lg⋆ (θ) = argmin ϕ′∈Φ Gi (ϕ′, θ) Gi (ϕ′, θ) = ̂ℒ (ϕ′)+ λ 2 ϕ′− θ 2 19
  • 20. Outer Loop • MAML outer loop • inner loop ➡ θ ← θ − ηdθF(θ) = θ − η 1 M M ∑ i=1 d𝒜lgi(θ) dθ ∇ϕℒi ( 𝒜lgi(θ)) (ϕ = 𝒜lgi(θ)) d𝒜lgi(θ) dθ 20
  • 21. Outer Loop • inner loop • • - adapt ϕi ≡ 𝒜lg⋆ i (θ) = argmin ϕ′∈Φ Gi (ϕ′, θ) ∇ϕ′Gi (ϕ′, θ) ϕ′=ϕi = 0 ∇ ̂ℒ(ϕi) + λ(𝒜lg⋆ i (θ) − θ) = 0 θ 𝒜lg⋆ (θ) d𝒜lg⋆ (θ) dθ = ( I + 1 λ ∇2 ̂ℒ (ϕi)) −1 ϕi 21
  • 22. Outer Loop • 2 ① inner loop adapt (SGD ) ② 3 • ( I + 1 λ ∇2 ̂ℒ (ϕi)) −1 ϕi ( I + 1 λ ∇2 ̂ℒ (ϕi)) −1 ∇ϕℒi ( 𝒜lgi(θ)) 22
  • 23. (CG ) • • Ax = b ⋯(1) (1) f(x) = 1 2 xT Ax − bT x x0 = 0,r0 = b − Ax0, p0 = r0 αk = rT k pk pT k Apk xk+1 = xk + αk pk rk+1 = rk − αkApk pk+1 = rk+1 + rT k+1rk+1 rT k rk pk 23
  • 24. (CG ) • • ( 5 ) - (p22 ① ) ‣ Appendix E gi = ( I + 1 λ ∇2 ̂ℒ (ϕi)) −1 ∇ϕℒi ( 𝒜lgi(θ)) gi ( I + 1 λ ∇2 ̂ℒ (ϕi)) gi = ∇ϕℒi ( 𝒜lgi(θ)) rk 𝒜lgi(θ) 24
  • 25. iMAML • inner loop ➡adapt • outer loop inner loop ➡inner loop ‣ MAML 1 ‣ iMAML Hessian-Free 2 adapt 25
  • 26. • - iMAML inner loop ( ) - FOMAML (CG ) - MAML 
 (FOMAML ??) O(1) 26
  • 27. • Omniglot - inner loop Hessian-Free iMAML - iMAML way ( ) - FOMAML 27
  • 28. • Mini-ImageNet - Reptile (FOMAML ) - ?? 28
  • 30. ( ) 1. - ( ) - iMAML 2. - - ERNN 30
  • 31. RNNs Evolving on an Equilibrium Manifold: A Panacea for Vanishing and Exploding Gradients? 31
  • 32. • - Anil Kag, Ziming Zhang, Venkatesh Saligrama - , MERL • NeurIPS 2019 reject • RNN • • 32
  • 33. RNN / • RNN - sigmoid tanh • RNN / - LSTM GRU hk = ϕ (Uhk−1 + Wxk + b) ϕ ∂hm ∂hn = ∏ m≥k>n ∂hk ∂hk−1 = ∏ m≥k>n ∇ϕ (Uhk−1 + Wxk + b) U 33
  • 34. RNN ODE • RNN skip connection (ODE) • Neural ODE - 
 https://www.slideshare.net/DeepLearningJP2016/dlneural-ordinary- differential-equations dh(t) dt ≜ h′(t) = ϕ (Uh(t) + Wxk + b) ⟹ hk = hk−1 + ηϕ (Uhk−1 + Wxk + b) 34
  • 35. ODE • ODE • 1 ➡ ( ) • ERNN dh dt = f (h, x) f (h, x) = 0 ⋯(1) (1) h x (h0, x0) fh (h0, x0) (1) h = g (x) (h0, x0) 35
  • 36. ERNN • ERNN 
 
 ODE • 
 
 ➡ 
 h′(t) = ϕ (U (h(t) + hk−1) + Wxk + b) − γ (h(t) + hk−1) h′(t) = 0 hk hk f (hk−1, h) = ϕ (U (h + hk−1) + Wxk + b) − γ (h + hk−1) = 0 ∂h ∂hk−1 = − ∂f/∂hk−1 ∂f/∂h = − I ∂f/∂h 36
  • 37. ∂f/∂h • 1. (sigmoid tanh OK) 2. - 
 ( ) ∂f ∂h = ∇ϕ (U (h + hk−1) + Wxk + b) U ϕ U 37
  • 38. • • 5 • - h(0) k = 0 h(i+1) k = h(i) k + η(i) k [ ϕ ( U (h(i) k + hk−1) + Wxk + b ) − γ (h(i) k + hk−1)] η(i) k 38
  • 39. HAR-2 RNN ERNN (log scale) • RNN • ERNN 1 ∂hT ∂h1 39