DARK HISTORY back in undergraduate year (2011). A naive and embarrassing effort to predict buy signals for stock index with price data analytics.
https://gitlab.com/tc-ying/cuhk/tree/master/CSCI4020-Financial-Data-Mining
논문 제목부터 재미있어 보이는 주제 입니다. 오늘 딥러닝 논문읽기 모임에서 소개드릴 논문은 DEAR: Deep Reinforcement Learning for Online Advertising Impression in Recommender Systems, 강화학습을 이용한 온라인 추천 시스템 입니다. 비공개 된 정보들이 몇가지가 있지만, 아이디어면에서 여러분들이 충분히 재밌게 들으실수 있습니다. 강화학습의 기본적인 개념부터,
논문에 대한 디테일하고 깊이 있는 리뷰를
펀디멘탈팀 김창연 님이 도와주셨습니다!
오늘도 많은 관심 미리 감사드립니다!
추가로 .. 딥러닝 논문읽기 모임은 청강방 오픈채팅 방을 운영하고 있습니다. 최근 악성 홍보 봇 계정이 늘어나 방을 비밀번호를 걸어두게 되었습니다
딥러닝 청강방도 많은 관심 부탁드립니다!
청강방 링크 : https://open.kakao.com/o/gp6GHMMc
청강방 비밀번호 : 0501
논문 제목부터 재미있어 보이는 주제 입니다. 오늘 딥러닝 논문읽기 모임에서 소개드릴 논문은 DEAR: Deep Reinforcement Learning for Online Advertising Impression in Recommender Systems, 강화학습을 이용한 온라인 추천 시스템 입니다. 비공개 된 정보들이 몇가지가 있지만, 아이디어면에서 여러분들이 충분히 재밌게 들으실수 있습니다. 강화학습의 기본적인 개념부터,
논문에 대한 디테일하고 깊이 있는 리뷰를
펀디멘탈팀 김창연 님이 도와주셨습니다!
오늘도 많은 관심 미리 감사드립니다!
추가로 .. 딥러닝 논문읽기 모임은 청강방 오픈채팅 방을 운영하고 있습니다. 최근 악성 홍보 봇 계정이 늘어나 방을 비밀번호를 걸어두게 되었습니다
딥러닝 청강방도 많은 관심 부탁드립니다!
청강방 링크 : https://open.kakao.com/o/gp6GHMMc
청강방 비밀번호 : 0501
발표자: 곽동현(서울대 박사과정, 현 NAVER Clova)
강화학습(Reinforcement learning)의 개요 및 최근 Deep learning 기반의 RL 트렌드를 소개합니다.
발표영상:
http://tv.naver.com/v/2024376
https://youtu.be/dw0sHzE1oAc
Intro to Reinforcement learning - part IIIMikko Mäkipää
Introduction to Reinforcement Learning, part III: Basic approximate methods
This is the final presentation in a three-part series covering the basics of Reinforcement Learning (RL).
In this presentation, we introduce value function approximation and cover three different approaches to generating features for linear models.
We then take a sidestep to cover stochastic gradient descent in some detail before we return to introduce semi-gradient descent for RL. We also briefly cover a batch method as an alternative for episodic methods.
We discuss the implementation of the RL algorithms. For further discussion and illustrating the simulation results, we refer to Github repositories with source code of the implementation as well as Jupyter notebooks visualizing the simulation results.
Two strategies for large-scale multi-label classification on the YouTube-8M d...Dalei Li
The project to participate in the Kaggle YouTube-8M video understanding competition. Four algorithms that can be run on a single machine are implemented, namely, multi-label k-nearest neighbor, multi-label radial basis function network (one-vs-rest), and multi-label logistic regression and on-vs-rest multi-layer neural network.
What am I going to get from this course?
Provides a basic conceptual understanding of how clustering works
Provides intuitive understanding of the mathematics behind various clustering algorithms
Walk through Python code examples on how to use various cluster algorithms
Show how clustering is applied in various industry applications
Check it on Experfy: https://www.experfy.com/training/courses/unsupervised-learning-clustering
발표자: 곽동현(서울대 박사과정, 현 NAVER Clova)
강화학습(Reinforcement learning)의 개요 및 최근 Deep learning 기반의 RL 트렌드를 소개합니다.
발표영상:
http://tv.naver.com/v/2024376
https://youtu.be/dw0sHzE1oAc
Intro to Reinforcement learning - part IIIMikko Mäkipää
Introduction to Reinforcement Learning, part III: Basic approximate methods
This is the final presentation in a three-part series covering the basics of Reinforcement Learning (RL).
In this presentation, we introduce value function approximation and cover three different approaches to generating features for linear models.
We then take a sidestep to cover stochastic gradient descent in some detail before we return to introduce semi-gradient descent for RL. We also briefly cover a batch method as an alternative for episodic methods.
We discuss the implementation of the RL algorithms. For further discussion and illustrating the simulation results, we refer to Github repositories with source code of the implementation as well as Jupyter notebooks visualizing the simulation results.
Two strategies for large-scale multi-label classification on the YouTube-8M d...Dalei Li
The project to participate in the Kaggle YouTube-8M video understanding competition. Four algorithms that can be run on a single machine are implemented, namely, multi-label k-nearest neighbor, multi-label radial basis function network (one-vs-rest), and multi-label logistic regression and on-vs-rest multi-layer neural network.
What am I going to get from this course?
Provides a basic conceptual understanding of how clustering works
Provides intuitive understanding of the mathematics behind various clustering algorithms
Walk through Python code examples on how to use various cluster algorithms
Show how clustering is applied in various industry applications
Check it on Experfy: https://www.experfy.com/training/courses/unsupervised-learning-clustering
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
4. Semester 1
• Finished data collection and pre-processing
• Binary classification
– Target label: “8-day bottom”
5. Semester 1
• K-Nearest Neighbour (k-NN)
– Low bias
• Sliding window testing
– Model changes through time
6. Data Visualization
• Label against various input attribute
• Label distribution through time
• Input Attributes:
• Prediction Target: 5-day bottom
Slope of SMA(22) and EMA(8, 13, 18)
Price Deviation from SMA(9, 22) and EMA(8, 16)
Percentage aggregate return for last 5 and 6 days
Daily gap open
10. Problem with fixed sliding window
• Given 3-NN majority voting
• Property of target label
– Continually positive
• Fail to learn
– From history
– From input variable
11. Function of time as attribute
• Recent labels are more important in trend-
dominant market
• New dimension in feature space:
Exponential function of time index
• Recent data vectors pulled closer
to current query point
• i.e. Memory decay
• controversial
12. Function of time as attribute
0.6
0.62
0.64
0.66
0.68
0.7
0.72
0.74
0 100 200 300 400 500 600
window size (days)
K = 3
Precision
Recall
13. Function of time as attribute
• New dimension in feature space:
Sine function of time index
• Data vectors at regular interval pulled closer
to each other
• Choice of period?
14. Function of time as attribute
0.65
0.66
0.67
0.68
0.69
0.7
0.71
0.72
0.73
0.74
0.75
0 100 200 300 400 500 600
window size (days)
K = 3
Precision
Recall
15. Result
• Looking back: 540 days
• K = 3 neighbours
• Error rate: 24%
• Precision: 0.731; Recall: 0.698
16. Limitation of label definition
• Consider a use case:
– A trader buys on first occurrence of positive
prediction?
– He needs a smart trading rule
18. Re-consider label definition
• N-bottom + r% gain after N days
– For each time point t, if its closing price is lower
than the low price by m% after N time points, then
time point t is a hold-N-bottom.
• Trading signal
19. Re-consider label definition
• N-bottom + m% stop-loss within N days
– For each time point t, if its closing price is lower
than the minimum low price within N time points
in the future by m%, then time point t is a stop-N-
bottom.
• More specific, more useful
• Fewer positive labels -> imbalanced dataset
20. New Result
• Looking back: 120 days
• K = 3 neighbours
• Error rate: 19.6%
• Precision: 0.601; Recall: 0.476
31. Limitations
• requires lots of supervised training, with lots
of input-output examples
• there is no guarantee that the system will
converge to an acceptable solution
~local minimum
32. Experiment
• The learning rate :0.25 to 0.5 ,and
• Momentum set to WEKA’s default:0.2
• the size of training set: 500 to 400000.
38. False Positive
RSI(14) Special FP rate
4 day bottom 0.216783216
8 day bottom 0.23656542
12 day bottom 0.221896955
16 day bottom 0.242370892
20 day bottom 0.251785714
Size of training set Learning rate = 0.35 Special FP rate
500 0.233870967
5000 0.175806451
10000 0.183870967
20000 0.182258064
30000 0.185483871
39. Discussion
• 100% accuracy is impossible in practice
• Effectiveness of classification model also
depends on how it is used, i.e. coupled with
trading rules