AAAI2023「Are Transformers Effective for Time Series Forecasting?」と、HuggingFace「Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)」の紹介です。
cvpaper.challenge の Meta Study Group 発表スライド
cvpaper.challenge はコンピュータビジョン分野の今を映し、トレンドを創り出す挑戦です。論文サマリ・アイディア考案・議論・実装・論文投稿に取り組み、凡ゆる知識を共有します。2019の目標「トップ会議30+本投稿」「2回以上のトップ会議網羅的サーベイ」
http://xpaperchallenge.org/cv/
AAAI2023「Are Transformers Effective for Time Series Forecasting?」と、HuggingFace「Yes, Transformers are Effective for Time Series Forecasting (+ Autoformer)」の紹介です。
cvpaper.challenge の Meta Study Group 発表スライド
cvpaper.challenge はコンピュータビジョン分野の今を映し、トレンドを創り出す挑戦です。論文サマリ・アイディア考案・議論・実装・論文投稿に取り組み、凡ゆる知識を共有します。2019の目標「トップ会議30+本投稿」「2回以上のトップ会議網羅的サーベイ」
http://xpaperchallenge.org/cv/
A presentation for meeting of statistics in Japan.
https://connpass.com/event/204931/
Topics:
Methods and their properties to measure dependencies between variables
Exploring Practices in Machine Learning and Machine Discovery for Heterogeneo...Ichigaku Takigawa
Video https://youtu.be/P4QogT8bdqY
ACS Spring 2023 Symposium on AI-Accelerated Scientific Workflow
https://acs.digitellinc.com/acs/sessions/526630/view
ACS SPRING 2023 ———— Crossroads of Chemistry
Indianapolis, IN & Hybrid, March 26-30
https://www.acs.org/meetings/acs-meetings/spring-2023.html
Slide PDF
https://itakigawa.page.link/acs2023spring
Our Paper
Accelerated discovery of multi-elemental reverse water-gas shift catalysts using extrapolative machine learning approach (2022, ChemRxiv)
https://doi.org/10.26434/chemrxiv-2022-695rj
Ichi Takigawa
https://itakigawa.github.io/
Machine Learning for Molecules: Lessons and Challenges of Data-Centric ChemistryIchigaku Takigawa
Perspectives on Artificial Intelligence and Machine Learning in Materials Science
February 4, 2022. – February 6, 2022.
https://joint.imi.kyushu-u.ac.jp/post-2698/
Machine Learning for Molecular Graph Representations and GeometriesIchigaku Takigawa
Dec 1, 2021, Pacifico Yokohama, Japan.
Symposium 1AS-17 "Data science and machine learning: Tackling the Noise and Heterogeneity of the Real World"
The 44th Annual Meetingn of the Molecular Biology Society of Japan
https://www2.aeplan.co.jp/mbsj2021/english/designation/index.html
43. https://icml.cc/Conferences/2019/ScheduleMultitrack?event=4343
A Tutorial on Attention in Deep Learning (ICML2019)
Alex Smola · Aston Zhang
The world's dumbest estimator
Better idea :
Watson-Nadaraya Estimator
(Watson, Nadaraya, 1964)
KeyQuery Value
Q: Query
K: Key
V: Value
Transformer
Transformer Scaled dot-product attention
•
• Pooling
https://d2l.ai
注意機構 (Attention)
53. 計量学習
Deep Face Recognition: A Survey. https://arxiv.org/abs/1804.06655
https://gombru.github.io/2019/04/03/ranking_loss/
Siamese Network (contrastive loss) Triplet Network (triplet loss) Angular margin loss
57. 陽な原理や知識の利⽤とデータ駆動の融合へ向けて
https://uclnlp.github.io/nampi/
Machine intelligence capable of learning complex procedural
behavior, inducing (latent) programs, and reasoning with these
programs is a key to solving artificial intelligence. Recently,
there have been a lot of success stories in the deep learning
community related to learning neural networks capable of
using trainable memory abstractions.
Neural Abstract Machines & Program Induction • Differentiable Neural Computers /
Neural Turing Machines (Graves+ 2014)
• Memory Networks (Weston+ 2014)
• Pointer Networks (Vinyals+ 2015)
• Neural Stacks (Grefenstette+ 2015, Joulin+ 2015)
• Hierarchical Attentive Memory
(Andrychowicz+ 2016)
• Neural Program Interpreters (Reed+ 2016)
• Neural Programmer (Neelakantan+ 2016)
• DeepCoder (Balog+ 2016)
:
⼿続き的・記号的操作も学習できるプログラムとして扱えるようになってきた
58. • Deep learning techniques thus far have proven to be data hungry, shallow,
brittle, and limited in their ability to generalize (Marcus, 2018)
• Current machine learning techniques are data-hungry and brittle—they can
only make sense of patterns they've seen before. (Chollet, 2020)
• A growing body of evidence shows that state-of-the-art models learn to exploit
spurious statistical patterns in datasets... instead of learning meaning in the
flexible and generalizable way that humans do. (Nie et al., 2019)
• Current machine learning methods seem weak when they are required to
generalize beyond the training distribution, which is what is often needed in
practice. (Bengio et al., 2019)
ただし適切な⽬的へ応⽤すれば極めて有効な技術だと実証されてきた!
59. • Deep learning techniques thus far have proven to be data hungry, shallow,
brittle, and limited in their ability to generalize (Marcus, 2018)
• Current machine learning techniques are data-hungry and brittle—they can
only make sense of patterns they've seen before. (Chollet, 2020)
• A growing body of evidence shows that state-of-the-art models learn to exploit
spurious statistical patterns in datasets... instead of learning meaning in the
flexible and generalizable way that humans do. (Nie et al., 2019)
• Current machine learning methods seem weak when they are required to
generalize beyond the training distribution, which is what is often needed in
practice. (Bengio et al., 2019)
ただし適切な⽬的へ応⽤すれば極めて有効な技術だと実証されてきた!
From AAAI-20 Oxford-Style Debate