SlideShare a Scribd company logo
자연어처리팀
김은희, 신동진, 황소현(발표자)
FinBERT
FinBERT:
A Pretrained Language Model for
Financial Communications
1. Introduction
1. Introduction
NLP in Financial Domain
- Real-time sentiment monitoring from online news/social media
⇒ Directional signal for trading
Sentiment 포트폴리오 > Index 펀드 (Bloomberg)
1. Introduction
1. Introduction
NLP in Financial Domain
- Finance 분야의 NLP 모델이 없었다!
- Unsupervised pretraining
- generic BERT 모델보다 강력
- corporate report / conference call transcript / analysis report 사용
+ = FinBERT
2. Related Work
Large NLP models
- BERT, ELMo, ULM-Fit, XLNet, GPT
- general domain corpora로 학습
(news, Wikipedia)
출처: https://medium.com/nwamaka-imasogie/clinicalbert-
using-deep-learning-transformer-model-to-predict-hospital-
readmission-c82ff0e4bb03
Domain-specific models
- BioBERT, ClinicalBERT, SciBERT, …
- Finance domain에 대해서는 FinBERT
가 최초
출처: https://github.com/thunlp/PLMpapers
2. Related Work
3. Financial Corpora
3. Financial Corpora
Overall Corpora Statistics
- 4.9 Billion tokens 확보
- BERT는 pre-training 3.3 Billion tokens
3. Financial Corpora
3. Financial Corpora
Corporate Reports 10-K & 10-Q
- 10-K : 연차보고서 / 10-Q : 분기보고서
- 금융 및 비즈니스에서 가장 중요한 텍스트 데이터
* 회사의 비즈니스 및 재무 상태(SEC 웹사이트에 공개 되어 있음)
-1994 ~ 2019년 사이의 60,490 10-K 와 142,622의 10-Q를 얻음
⚫ Item 1(Business)
⚫ Item 1A(Risk Factors)
⚫ Item 7(Managements Discusion and Analysis)
3. Financial Corpora
Earnings Call Transcripts
- 회사 성과에 대한 보고
- 2004 ~ 2019 사이의 7,740 개 회사의
136,578 얻음
- https://seekingalpha.com/earnings/earnings-call-
transcripts
Analyst Reports
- 기관 및 개별 투자자에게 유용한 정보를 제공
- 주식 추천, 수익 예측, 목표 가격을 포함한
여러 정량적인 요약 값을 제공
- 기관 투자자들은 매년 수백만 달러를 소비해서
Analyst Reports를 구매해서 읽음.
- 1995 ~ 2008년동안 S&P 에서 발행된 488,494
개 세트의 Analyst Reports를 구함.
3. Financial Corpora
FinBERT
Questions
4. FinBERT Training
Vocabulary
- SentenePiece library 사용 FinVocab 구축
- FinBERT VS BERT vocabulary
- FinBERT , BERT 교집합은 41%
FinBERT-Variants
- BERT-Base 와 동일한 환경으로 FinBERT
Corpora 사용
- 문장 길이를 128 tokens 로 설정하고 훈련 후,
512 tokens 를 허용하는 모델을 계속 training
- 4개의 다른 버전을 훈련
- FinBERT-Base Vocab, uncased/cased
- 250K iteration
- 2e^-5 learning rate
- FinBERT-FinVocab , uncased/cased
- 1M iteration
- Finvocab 사용
4. FinBERT Training
Cased Uncased
FinBERT 28,573 30,873
BERT 28,996 30,522
5. Financial Sentiment Experiments
5. Dataset
• Financial Phrase Bank
• AnalystTone Dataset
• FiQA Dataset
5. Financial Sentiment Experiments
5. Dataset
• Financial Phrase Bank (Malo et al., 2014)
- 금융 뉴스데이터에서 4,840 개의 문장 포함
- 16명의 전문지식을 갖춘 연구자들에 의해 수동 라벨링 하여 만들었음.
- 감정 라벨 : positive, neutral, negative
- Data Instances
-{ "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that
were hit by larger expenditures on R&D and marketing .",
- "label": "negative"
-}
5. Financial Sentiment Experiments
5. Dataset
• AnalystTone Dataset (Huang et al., 2014)
- 무작위로 10,000개의 문장이 포함
- 감정 라벨 : positive (3,580 개) , neutral (4,590 개), negative (1,830 개)
• FiQA Dataset (https://sites.google.com/view/fiqa/home)
- Financial Opinion Mining and Question Answering
- 1,111 개의 문장 포함 오픈 데이터
- numeric sentiment score, ranged from - 1 to 1 - > 이진 분류 작업
- split each dataset into 90% training and 10% testing 10 times and report the average.
5. Financial Sentiment Experiments
5. Experiment Results
FinBERT vs. BERT
5. Financial Sentiment Experiments
5. Experiment Results
FinVocab vs. BaseVocab
5. Financial Sentiment Experiments
5. Experiment Results
Cased vs. Uncased
5. Financial Sentiment Experiments
5. Experiment Results
Corpus Contribution
- FinBERT : financial-task oriented BERT 만들었어요!
- 세가지 financial sentiment classification tasks에서 BERT를 능가해요!
- FinBERT 릴리즈를 통해 실무자와 연구자들이 Fin-BERT를 활용할 수 있어요!
6. Conclusion
6. Conclusion
FinBERT
Questions

More Related Content

What's hot

[DL輪読会]1次近似系MAMLとその理論的背景
[DL輪読会]1次近似系MAMLとその理論的背景[DL輪読会]1次近似系MAMLとその理論的背景
[DL輪読会]1次近似系MAMLとその理論的背景
Deep Learning JP
 
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
PR-317: MLP-Mixer: An all-MLP Architecture for VisionPR-317: MLP-Mixer: An all-MLP Architecture for Vision
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
Jinwon Lee
 
KorQuAD v2.0 소개
KorQuAD v2.0 소개KorQuAD v2.0 소개
KorQuAD v2.0 소개
LGCNSairesearch
 
RoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position EmbeddingRoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position Embedding
taeseon ryu
 
LLM 모델 기반 서비스 실전 가이드
LLM 모델 기반 서비스 실전 가이드LLM 모델 기반 서비스 실전 가이드
LLM 모델 기반 서비스 실전 가이드
Tae Young Lee
 
[DL輪読会]When Does Label Smoothing Help?
[DL輪読会]When Does Label Smoothing Help?[DL輪読会]When Does Label Smoothing Help?
[DL輪読会]When Does Label Smoothing Help?
Deep Learning JP
 
Natural language processing and transformer models
Natural language processing and transformer modelsNatural language processing and transformer models
Natural language processing and transformer models
Ding Li
 
Thomas Wolf "Transfer learning in NLP"
Thomas Wolf "Transfer learning in NLP"Thomas Wolf "Transfer learning in NLP"
Thomas Wolf "Transfer learning in NLP"
Fwdays
 
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
Deep Learning JP
 
[DLHacks]Comet ML -機械学習のためのGitHub-
[DLHacks]Comet ML -機械学習のためのGitHub-[DLHacks]Comet ML -機械学習のためのGitHub-
[DLHacks]Comet ML -機械学習のためのGitHub-
Deep Learning JP
 
NLP using transformers
NLP using transformers NLP using transformers
NLP using transformers
Arvind Devaraj
 
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
Deep Learning JP
 
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search Engine Advert...
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search  Engine Advert...【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search  Engine Advert...
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search Engine Advert...
Deep Learning JP
 
Attention Is All You Need
Attention Is All You NeedAttention Is All You Need
Attention Is All You Need
Illia Polosukhin
 
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
Seongyun Byeon
 
[DL輪読会]MetaFormer is Actually What You Need for Vision
[DL輪読会]MetaFormer is Actually What You Need for Vision[DL輪読会]MetaFormer is Actually What You Need for Vision
[DL輪読会]MetaFormer is Actually What You Need for Vision
Deep Learning JP
 
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series DataToeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
Daiki Tanaka
 
[DL輪読会]Neural Ordinary Differential Equations
[DL輪読会]Neural Ordinary Differential Equations[DL輪読会]Neural Ordinary Differential Equations
[DL輪読会]Neural Ordinary Differential Equations
Deep Learning JP
 
Active Learning 入門
Active Learning 入門Active Learning 入門
Active Learning 入門Shuyo Nakatani
 
I3D and Kinetics datasets (Action Recognition)
I3D and Kinetics datasets (Action Recognition)I3D and Kinetics datasets (Action Recognition)
I3D and Kinetics datasets (Action Recognition)
Susang Kim
 

What's hot (20)

[DL輪読会]1次近似系MAMLとその理論的背景
[DL輪読会]1次近似系MAMLとその理論的背景[DL輪読会]1次近似系MAMLとその理論的背景
[DL輪読会]1次近似系MAMLとその理論的背景
 
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
PR-317: MLP-Mixer: An all-MLP Architecture for VisionPR-317: MLP-Mixer: An all-MLP Architecture for Vision
PR-317: MLP-Mixer: An all-MLP Architecture for Vision
 
KorQuAD v2.0 소개
KorQuAD v2.0 소개KorQuAD v2.0 소개
KorQuAD v2.0 소개
 
RoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position EmbeddingRoFormer: Enhanced Transformer with Rotary Position Embedding
RoFormer: Enhanced Transformer with Rotary Position Embedding
 
LLM 모델 기반 서비스 실전 가이드
LLM 모델 기반 서비스 실전 가이드LLM 모델 기반 서비스 실전 가이드
LLM 모델 기반 서비스 실전 가이드
 
[DL輪読会]When Does Label Smoothing Help?
[DL輪読会]When Does Label Smoothing Help?[DL輪読会]When Does Label Smoothing Help?
[DL輪読会]When Does Label Smoothing Help?
 
Natural language processing and transformer models
Natural language processing and transformer modelsNatural language processing and transformer models
Natural language processing and transformer models
 
Thomas Wolf "Transfer learning in NLP"
Thomas Wolf "Transfer learning in NLP"Thomas Wolf "Transfer learning in NLP"
Thomas Wolf "Transfer learning in NLP"
 
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
【DL輪読会】Llama 2: Open Foundation and Fine-Tuned Chat Models
 
[DLHacks]Comet ML -機械学習のためのGitHub-
[DLHacks]Comet ML -機械学習のためのGitHub-[DLHacks]Comet ML -機械学習のためのGitHub-
[DLHacks]Comet ML -機械学習のためのGitHub-
 
NLP using transformers
NLP using transformers NLP using transformers
NLP using transformers
 
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
[DL輪読会]BERT: Pre-training of Deep Bidirectional Transformers for Language Und...
 
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search Engine Advert...
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search  Engine Advert...【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search  Engine Advert...
【DL輪読会】Aspect-based Analysis of Advertising Appeals for Search Engine Advert...
 
Attention Is All You Need
Attention Is All You NeedAttention Is All You Need
Attention Is All You Need
 
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
[MLOps KR 행사] MLOps 춘추 전국 시대 정리(210605)
 
[DL輪読会]MetaFormer is Actually What You Need for Vision
[DL輪読会]MetaFormer is Actually What You Need for Vision[DL輪読会]MetaFormer is Actually What You Need for Vision
[DL輪読会]MetaFormer is Actually What You Need for Vision
 
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series DataToeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
Toeplitz Inverse Covariance-Based Clustering of Multivariate Time Series Data
 
[DL輪読会]Neural Ordinary Differential Equations
[DL輪読会]Neural Ordinary Differential Equations[DL輪読会]Neural Ordinary Differential Equations
[DL輪読会]Neural Ordinary Differential Equations
 
Active Learning 入門
Active Learning 入門Active Learning 入門
Active Learning 入門
 
I3D and Kinetics datasets (Action Recognition)
I3D and Kinetics datasets (Action Recognition)I3D and Kinetics datasets (Action Recognition)
I3D and Kinetics datasets (Action Recognition)
 

More from taeseon ryu

VoxelNet
VoxelNetVoxelNet
VoxelNet
taeseon ryu
 
OpineSum Entailment-based self-training for abstractive opinion summarization...
OpineSum Entailment-based self-training for abstractive opinion summarization...OpineSum Entailment-based self-training for abstractive opinion summarization...
OpineSum Entailment-based self-training for abstractive opinion summarization...
taeseon ryu
 
3D Gaussian Splatting
3D Gaussian Splatting3D Gaussian Splatting
3D Gaussian Splatting
taeseon ryu
 
JetsonTX2 Python
 JetsonTX2 Python  JetsonTX2 Python
JetsonTX2 Python
taeseon ryu
 
Hyperbolic Image Embedding.pptx
Hyperbolic  Image Embedding.pptxHyperbolic  Image Embedding.pptx
Hyperbolic Image Embedding.pptx
taeseon ryu
 
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
taeseon ryu
 
LLaMA Open and Efficient Foundation Language Models - 230528.pdf
LLaMA Open and Efficient Foundation Language Models - 230528.pdfLLaMA Open and Efficient Foundation Language Models - 230528.pdf
LLaMA Open and Efficient Foundation Language Models - 230528.pdf
taeseon ryu
 
YOLO V6
YOLO V6YOLO V6
YOLO V6
taeseon ryu
 
Dataset Distillation by Matching Training Trajectories
Dataset Distillation by Matching Training Trajectories Dataset Distillation by Matching Training Trajectories
Dataset Distillation by Matching Training Trajectories
taeseon ryu
 
RL_UpsideDown
RL_UpsideDownRL_UpsideDown
RL_UpsideDown
taeseon ryu
 
Packed Levitated Marker for Entity and Relation Extraction
Packed Levitated Marker for Entity and Relation ExtractionPacked Levitated Marker for Entity and Relation Extraction
Packed Levitated Marker for Entity and Relation Extraction
taeseon ryu
 
MOReL: Model-Based Offline Reinforcement Learning
MOReL: Model-Based Offline Reinforcement LearningMOReL: Model-Based Offline Reinforcement Learning
MOReL: Model-Based Offline Reinforcement Learning
taeseon ryu
 
Scaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language ModelsScaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language Models
taeseon ryu
 
Visual prompt tuning
Visual prompt tuningVisual prompt tuning
Visual prompt tuning
taeseon ryu
 
mPLUG
mPLUGmPLUG
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdfvariBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
taeseon ryu
 
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdfReinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
taeseon ryu
 
The Forward-Forward Algorithm
The Forward-Forward AlgorithmThe Forward-Forward Algorithm
The Forward-Forward Algorithm
taeseon ryu
 
Towards Robust and Reproducible Active Learning using Neural Networks
Towards Robust and Reproducible Active Learning using Neural NetworksTowards Robust and Reproducible Active Learning using Neural Networks
Towards Robust and Reproducible Active Learning using Neural Networks
taeseon ryu
 
BRIO: Bringing Order to Abstractive Summarization
BRIO: Bringing Order to Abstractive SummarizationBRIO: Bringing Order to Abstractive Summarization
BRIO: Bringing Order to Abstractive Summarization
taeseon ryu
 

More from taeseon ryu (20)

VoxelNet
VoxelNetVoxelNet
VoxelNet
 
OpineSum Entailment-based self-training for abstractive opinion summarization...
OpineSum Entailment-based self-training for abstractive opinion summarization...OpineSum Entailment-based self-training for abstractive opinion summarization...
OpineSum Entailment-based self-training for abstractive opinion summarization...
 
3D Gaussian Splatting
3D Gaussian Splatting3D Gaussian Splatting
3D Gaussian Splatting
 
JetsonTX2 Python
 JetsonTX2 Python  JetsonTX2 Python
JetsonTX2 Python
 
Hyperbolic Image Embedding.pptx
Hyperbolic  Image Embedding.pptxHyperbolic  Image Embedding.pptx
Hyperbolic Image Embedding.pptx
 
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
MCSE_Multimodal Contrastive Learning of Sentence Embeddings_변현정
 
LLaMA Open and Efficient Foundation Language Models - 230528.pdf
LLaMA Open and Efficient Foundation Language Models - 230528.pdfLLaMA Open and Efficient Foundation Language Models - 230528.pdf
LLaMA Open and Efficient Foundation Language Models - 230528.pdf
 
YOLO V6
YOLO V6YOLO V6
YOLO V6
 
Dataset Distillation by Matching Training Trajectories
Dataset Distillation by Matching Training Trajectories Dataset Distillation by Matching Training Trajectories
Dataset Distillation by Matching Training Trajectories
 
RL_UpsideDown
RL_UpsideDownRL_UpsideDown
RL_UpsideDown
 
Packed Levitated Marker for Entity and Relation Extraction
Packed Levitated Marker for Entity and Relation ExtractionPacked Levitated Marker for Entity and Relation Extraction
Packed Levitated Marker for Entity and Relation Extraction
 
MOReL: Model-Based Offline Reinforcement Learning
MOReL: Model-Based Offline Reinforcement LearningMOReL: Model-Based Offline Reinforcement Learning
MOReL: Model-Based Offline Reinforcement Learning
 
Scaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language ModelsScaling Instruction-Finetuned Language Models
Scaling Instruction-Finetuned Language Models
 
Visual prompt tuning
Visual prompt tuningVisual prompt tuning
Visual prompt tuning
 
mPLUG
mPLUGmPLUG
mPLUG
 
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdfvariBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
variBAD, A Very Good Method for Bayes-Adaptive Deep RL via Meta-Learning.pdf
 
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdfReinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
Reinforced Genetic Algorithm Learning For Optimizing Computation Graphs.pdf
 
The Forward-Forward Algorithm
The Forward-Forward AlgorithmThe Forward-Forward Algorithm
The Forward-Forward Algorithm
 
Towards Robust and Reproducible Active Learning using Neural Networks
Towards Robust and Reproducible Active Learning using Neural NetworksTowards Robust and Reproducible Active Learning using Neural Networks
Towards Robust and Reproducible Active Learning using Neural Networks
 
BRIO: Bringing Order to Abstractive Summarization
BRIO: Bringing Order to Abstractive SummarizationBRIO: Bringing Order to Abstractive Summarization
BRIO: Bringing Order to Abstractive Summarization
 

Fin bert paper review !

  • 1. 자연어처리팀 김은희, 신동진, 황소현(발표자) FinBERT FinBERT: A Pretrained Language Model for Financial Communications
  • 2. 1. Introduction 1. Introduction NLP in Financial Domain - Real-time sentiment monitoring from online news/social media ⇒ Directional signal for trading Sentiment 포트폴리오 > Index 펀드 (Bloomberg)
  • 3. 1. Introduction 1. Introduction NLP in Financial Domain - Finance 분야의 NLP 모델이 없었다! - Unsupervised pretraining - generic BERT 모델보다 강력 - corporate report / conference call transcript / analysis report 사용 + = FinBERT
  • 4. 2. Related Work Large NLP models - BERT, ELMo, ULM-Fit, XLNet, GPT - general domain corpora로 학습 (news, Wikipedia) 출처: https://medium.com/nwamaka-imasogie/clinicalbert- using-deep-learning-transformer-model-to-predict-hospital- readmission-c82ff0e4bb03 Domain-specific models - BioBERT, ClinicalBERT, SciBERT, … - Finance domain에 대해서는 FinBERT 가 최초 출처: https://github.com/thunlp/PLMpapers 2. Related Work
  • 5. 3. Financial Corpora 3. Financial Corpora Overall Corpora Statistics - 4.9 Billion tokens 확보 - BERT는 pre-training 3.3 Billion tokens
  • 6. 3. Financial Corpora 3. Financial Corpora Corporate Reports 10-K & 10-Q - 10-K : 연차보고서 / 10-Q : 분기보고서 - 금융 및 비즈니스에서 가장 중요한 텍스트 데이터 * 회사의 비즈니스 및 재무 상태(SEC 웹사이트에 공개 되어 있음) -1994 ~ 2019년 사이의 60,490 10-K 와 142,622의 10-Q를 얻음 ⚫ Item 1(Business) ⚫ Item 1A(Risk Factors) ⚫ Item 7(Managements Discusion and Analysis)
  • 7. 3. Financial Corpora Earnings Call Transcripts - 회사 성과에 대한 보고 - 2004 ~ 2019 사이의 7,740 개 회사의 136,578 얻음 - https://seekingalpha.com/earnings/earnings-call- transcripts Analyst Reports - 기관 및 개별 투자자에게 유용한 정보를 제공 - 주식 추천, 수익 예측, 목표 가격을 포함한 여러 정량적인 요약 값을 제공 - 기관 투자자들은 매년 수백만 달러를 소비해서 Analyst Reports를 구매해서 읽음. - 1995 ~ 2008년동안 S&P 에서 발행된 488,494 개 세트의 Analyst Reports를 구함. 3. Financial Corpora
  • 9. 4. FinBERT Training Vocabulary - SentenePiece library 사용 FinVocab 구축 - FinBERT VS BERT vocabulary - FinBERT , BERT 교집합은 41% FinBERT-Variants - BERT-Base 와 동일한 환경으로 FinBERT Corpora 사용 - 문장 길이를 128 tokens 로 설정하고 훈련 후, 512 tokens 를 허용하는 모델을 계속 training - 4개의 다른 버전을 훈련 - FinBERT-Base Vocab, uncased/cased - 250K iteration - 2e^-5 learning rate - FinBERT-FinVocab , uncased/cased - 1M iteration - Finvocab 사용 4. FinBERT Training Cased Uncased FinBERT 28,573 30,873 BERT 28,996 30,522
  • 10. 5. Financial Sentiment Experiments 5. Dataset • Financial Phrase Bank • AnalystTone Dataset • FiQA Dataset
  • 11. 5. Financial Sentiment Experiments 5. Dataset • Financial Phrase Bank (Malo et al., 2014) - 금융 뉴스데이터에서 4,840 개의 문장 포함 - 16명의 전문지식을 갖춘 연구자들에 의해 수동 라벨링 하여 만들었음. - 감정 라벨 : positive, neutral, negative - Data Instances -{ "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .", - "label": "negative" -}
  • 12. 5. Financial Sentiment Experiments 5. Dataset • AnalystTone Dataset (Huang et al., 2014) - 무작위로 10,000개의 문장이 포함 - 감정 라벨 : positive (3,580 개) , neutral (4,590 개), negative (1,830 개) • FiQA Dataset (https://sites.google.com/view/fiqa/home) - Financial Opinion Mining and Question Answering - 1,111 개의 문장 포함 오픈 데이터 - numeric sentiment score, ranged from - 1 to 1 - > 이진 분류 작업 - split each dataset into 90% training and 10% testing 10 times and report the average.
  • 13. 5. Financial Sentiment Experiments 5. Experiment Results FinBERT vs. BERT
  • 14. 5. Financial Sentiment Experiments 5. Experiment Results FinVocab vs. BaseVocab
  • 15. 5. Financial Sentiment Experiments 5. Experiment Results Cased vs. Uncased
  • 16. 5. Financial Sentiment Experiments 5. Experiment Results Corpus Contribution
  • 17. - FinBERT : financial-task oriented BERT 만들었어요! - 세가지 financial sentiment classification tasks에서 BERT를 능가해요! - FinBERT 릴리즈를 통해 실무자와 연구자들이 Fin-BERT를 활용할 수 있어요! 6. Conclusion 6. Conclusion