In this project work, a multi-step deep neural network is used to forecast power generation and load demand for a short-term time frame. The data or feature vectors that have been used to predict the target, is a sequential time series sequence. In this project, a Recurrent Neural Network has been used in combination with a convolutional neural network to have a better forecasting model for the Windpark, Solar park and Loadpark datasets. Moreover, the forecasting performance of Feedforward neural network and Long Short Term Memory also has been compared. The whole project work has divided into two parts, in the first approach the raw dataset has been divided into a train, test split and no previous step data have been used. In the second step whole raw dataset has been divided into test, train and validation split. Additionally, current and seven previous time steps data has been fed into the model.
Evaluating LLM Models for Production Systems Methods and Practices -alopatenko
This webinar is designed to offer a comprehensive understanding of the evaluation processes for LLMs, particularly in the context of preparing these models for deployment in production environments.
Key Highlights of the Seminar:
In-Depth Analysis of LLM Evaluation Methods: Gain insights into a variety of methods to evaluate LLM models, understanding their strengths and weaknesses.
End-to-End Evaluation Techniques: Explore how LLM augmented systems are assessed from a holistic perspective.
Pragmatic Approach to System Deployment: Learn practical strategies for applying these evaluation techniques to systems intended for real-world application.
Focused Overview on Critical LLM Aspects: Receive an overview of various evaluation techniques that are essential for assessing the most crucial elements of modern LLM systems.
Simplifying the Evaluation Process: Understand how to streamline the evaluation process, making the work of LLM scientists more efficient and productive.
Dr. Andrei Lopatenko is a seasoned expert and executive leader with over 15 years of experience in the tech industry, focusing on search engines, recommendation systems, and large-scale AI, ML, and NLP applications. He has contributed significantly to major companies like Google, Apple, Walmart, eBay, and Zillow, benefiting billions of customers. Dr. Lopatenko earned his PhD in Computer Science from the University of Manchester. He played a key role in developing Google's search engine, initiating Apple Maps, co-founding a Conversational AI startup acquired by Facebook/Meta, and leading Search, LLM, and Generative AI at Zillow.
Recurrent Neural Networks hold great promise as general sequence learning algorithms. As such, they are a very promising tool for text analysis. However, outside of very specific use cases such as handwriting recognition and recently, machine translation, they have not seen wide spread use. Why has this been the case?
In this presentation, we will first introduce RNNs as a concept. Then we will sketch how to implement them and cover the tricks necessary to make them work well. With the basics covered, we will investigate using RNNs as general text classification and regression models, examining where they succeed and where they fail compared to more traditional text analysis models. A straightforward open-source Python and Theano library for training RNNs with a scikit-learn style interface will be introduced and we’ll see how to use it through a tutorial on a real world text dataset
Deep Learning: Chapter 11 Practical MethodologyJason Tsai
Lecture for Deep Learning 101 study group to be held on June 9th, 2017.
Reference book: https://www.deeplearningbook.org/
Past video archives: https://goo.gl/hxermB
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/)
Evaluating LLM Models for Production Systems Methods and Practices -alopatenko
This webinar is designed to offer a comprehensive understanding of the evaluation processes for LLMs, particularly in the context of preparing these models for deployment in production environments.
Key Highlights of the Seminar:
In-Depth Analysis of LLM Evaluation Methods: Gain insights into a variety of methods to evaluate LLM models, understanding their strengths and weaknesses.
End-to-End Evaluation Techniques: Explore how LLM augmented systems are assessed from a holistic perspective.
Pragmatic Approach to System Deployment: Learn practical strategies for applying these evaluation techniques to systems intended for real-world application.
Focused Overview on Critical LLM Aspects: Receive an overview of various evaluation techniques that are essential for assessing the most crucial elements of modern LLM systems.
Simplifying the Evaluation Process: Understand how to streamline the evaluation process, making the work of LLM scientists more efficient and productive.
Dr. Andrei Lopatenko is a seasoned expert and executive leader with over 15 years of experience in the tech industry, focusing on search engines, recommendation systems, and large-scale AI, ML, and NLP applications. He has contributed significantly to major companies like Google, Apple, Walmart, eBay, and Zillow, benefiting billions of customers. Dr. Lopatenko earned his PhD in Computer Science from the University of Manchester. He played a key role in developing Google's search engine, initiating Apple Maps, co-founding a Conversational AI startup acquired by Facebook/Meta, and leading Search, LLM, and Generative AI at Zillow.
Recurrent Neural Networks hold great promise as general sequence learning algorithms. As such, they are a very promising tool for text analysis. However, outside of very specific use cases such as handwriting recognition and recently, machine translation, they have not seen wide spread use. Why has this been the case?
In this presentation, we will first introduce RNNs as a concept. Then we will sketch how to implement them and cover the tricks necessary to make them work well. With the basics covered, we will investigate using RNNs as general text classification and regression models, examining where they succeed and where they fail compared to more traditional text analysis models. A straightforward open-source Python and Theano library for training RNNs with a scikit-learn style interface will be introduced and we’ll see how to use it through a tutorial on a real world text dataset
Deep Learning: Chapter 11 Practical MethodologyJason Tsai
Lecture for Deep Learning 101 study group to be held on June 9th, 2017.
Reference book: https://www.deeplearningbook.org/
Past video archives: https://goo.gl/hxermB
Initiated by Taiwan AI Group (https://www.facebook.com/groups/Taiwan.AI.Group/)
Multiclass classification of imbalanced dataSaurabhWani6
Pydata Talk on Classification of imbalanced data.
It is an overview of concepts for better classification in imbalanced datasets.
Resampling techniques are introduced along with bagging and boosting methods.
List of top Machine Learning algorithms are making headway in the world of data science. Explained here are the top 10 of these machine learning algorithms - https://www.dezyre.com/article/top-10-machine-learning-algorithms/202
LLaMa 2 is a large language model (LLM) developed by Meta AI. It is a successor to the Llama model, and it is one of the most powerful LLMs available today. Llama 2 is trained on a massive dataset of text and code, and it can be used for a wide range of tasks, including:
Generating text, such as articles, poems, and code
Translating languages
Answering questions in a comprehensive and informative way
Following instructions and completing requests thoughtfully
LLama 2 is still under development, but it has already been shown to outperform other LLMs on many benchmarks. For example, Llama 2 outperforms other open source LLMs on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests.
Machine Learning based Hybrid Recommendation System
• Developed a Hybrid Movie Recommendation System using both Collaborative and Content-based methods
• Used linear regression framework for determining optimal feature weights from collaborative data
• Recommends movie with maximum similarity score of content-based data
This presentation provides an introduction to the artificial neural networks topic, its learning, network architecture, back propagation training algorithm, and its applications.
Welcome to the Supervised Machine Learning and Data Sciences.
Algorithms for building models. Support Vector Machines.
Classification algorithm explanation and code in Python ( SVM ) .
Multiclass classification of imbalanced dataSaurabhWani6
Pydata Talk on Classification of imbalanced data.
It is an overview of concepts for better classification in imbalanced datasets.
Resampling techniques are introduced along with bagging and boosting methods.
List of top Machine Learning algorithms are making headway in the world of data science. Explained here are the top 10 of these machine learning algorithms - https://www.dezyre.com/article/top-10-machine-learning-algorithms/202
LLaMa 2 is a large language model (LLM) developed by Meta AI. It is a successor to the Llama model, and it is one of the most powerful LLMs available today. Llama 2 is trained on a massive dataset of text and code, and it can be used for a wide range of tasks, including:
Generating text, such as articles, poems, and code
Translating languages
Answering questions in a comprehensive and informative way
Following instructions and completing requests thoughtfully
LLama 2 is still under development, but it has already been shown to outperform other LLMs on many benchmarks. For example, Llama 2 outperforms other open source LLMs on many external benchmarks, including reasoning, coding, proficiency, and knowledge tests.
Machine Learning based Hybrid Recommendation System
• Developed a Hybrid Movie Recommendation System using both Collaborative and Content-based methods
• Used linear regression framework for determining optimal feature weights from collaborative data
• Recommends movie with maximum similarity score of content-based data
This presentation provides an introduction to the artificial neural networks topic, its learning, network architecture, back propagation training algorithm, and its applications.
Welcome to the Supervised Machine Learning and Data Sciences.
Algorithms for building models. Support Vector Machines.
Classification algorithm explanation and code in Python ( SVM ) .
Performance prediction of PV & PV/T systems using Artificial Neural Networks ...Ali Al-Waeli
This presentation offers insight into use of ANN and machine learning for various applications in solar energy. Prepared and presented by Dr. Ali H. A. Alwaeli.
Explore how our student team leveraged data science to forecast power consumption, empowering smarter energy management and sustainability initiatives. visit for more: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
With the rise of containerization, as well as the established adoption of virtualization technologies, run-time power and energy management is becoming one of the key challenges in modern cloud computing. This is also fundamental as power consumption contributes to the 20% of the Total Cost of Ownership of a datacenter and energy costs will exceed hardware costs in the near future. In this context, several goals towards power optimization can be achieved. On the one hand, power capping can be enforced and on top of that the system should be able to maximize performance. On the other hand, when performance are critical, the system should be able to provide a minimum SLA and optimize power consumption without violating it. Within this context, we propose a common autonomic methodology based on the ODA control loop for containers and virtual machines. The proposed methodology is able to achieve 25% power savings for containers and can improve performance under a power cap for virtual machines.
Improving efficiency of Photovoltaic System with Neural Network Based MPPT Co...IJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
Keep Calm and React with Foresight: Strategies for Low-Latency and Energy-Eff...Tiziano De Matteis
This talk has been given at PPoPP 2016 (Barcelona)
The paper addresses the problem of designing control strategies for elastic stream processing applications. Elasticity allows applications to rapidly change their configuration (e.g. the number of used resources) on-the-fly, in response to fluctuations of their workload. In this work we face this problem by adopting the Model Predictive Control technique, a control-theoretic method aimed at finding the optimal application configuration along a limited prediction horizon by solving an online optimization problem. Our control strategies are designed to address latency constraints, by using Queueing Theory models, and energy consumption by changing the number of used cores and the CPU frequency through the Dynamic Voltage and Frequency Scaling (DVFS) function of modern multi-core CPUs. The proactive capabilities, in addition to the latency- and energy-awareness, represent the novel features of our approach. Experiments performed using a high-frequency trading application show the effectiveness compared with state-of-the-art techniques.
A full version of the slides (with transitions) is available at: https://docs.google.com/presentation/d/1VZ3y3RQDLFi_xA7Rl0Vj1iqBdoerxCMG4y53uMz9Ziw/edit?usp=sharing
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Design of c slotted microstrip antenna using artificial neural network modeleSAT Journals
Abstract In this paper, neural network model has been used to estimation of resonance frequency of a coaxial feed C-slotted Microstrip Antenna. The Multi-Layer Perceptron Feed forward back Propagation (MLPFFBP) and Radial basis function Artificial Neural Network (RBFANN) have been used to implement the neural network model. A relative performance analysis of the proposed neural network for different training algorithms. Number of neurons and number of hidden layer is also carried out for estimating the resonance frequency. The method of moment (MOM) based IE3D software was used to generate data dictionary for training and validation set of ANN. The results obtain using ANN are compared with simulation feeding and found quite satisfactory and also it is concluded that RBFANN network is more accurate and fast compared to MLPFFBP network algorithm. Index Terms: Artificial Neural Network, C slot, Microstrip Antenna, Multilayer Feed Forward Networks, Radial basis function Artificial Neural Network, Resonance frequency.
AI optimizing HPC simulations (presentation from 6th EULAG Workshop)byteLAKE
See our presentation from the 6th International EULAG Users Workshop. We talked about taking HPC to the "Industry 4.0" by implementing smart techniques to optimize the codes in terms of performance and energy consumption. It explains how Machine Learning can dynamically optimize HPC simulations and byteLAKE's software autotuning solution.
Find out more about byteLAKE at: www.byteLAKE.com
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Quantitative Data AnalysisReliability Analysis (Cronbach Alpha) Common Method...2023240532
Quantitative data Analysis
Overview
Reliability Analysis (Cronbach Alpha)
Common Method Bias (Harman Single Factor Test)
Frequency Analysis (Demographic)
Descriptive Analysis
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Discussion on Vector Databases, Unstructured Data and AI
https://www.meetup.com/unstructured-data-meetup-new-york/
This meetup is for people working in unstructured data. Speakers will come present about related topics such as vector databases, LLMs, and managing data at scale. The intended audience of this group includes roles like machine learning engineers, data scientists, data engineers, software engineers, and PMs.This meetup was formerly Milvus Meetup, and is sponsored by Zilliz maintainers of Milvus.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Test different neural networks models for forecasting of wind,solar and energy usage
1. Test different Neural Network
Models for forecasting of Wind,Solar Power
Generation and Energy usage within c/sells
Presented by:
Tonmoy Ibne Arif
Master's of Electrical and Communication Engineering
1
3. Motivation
Power forecasting of
renewable energy is an
active research field.
Smart grids requires
load forecasting.
Reduce operation cost.
Better load
scheduling.
Reduce dependency
on fossil fuel.
Test different NN
architecture on the
datasets
3
4. Dataset description
Dataset for EuropeWindFarm, German Solar farm and Load data from different
nodes.
Wind Dataset
• Day ahead forecast for 45 off & onshore wind farms.
• Time series for two years hourly averaged wind power generation
• Data features -time stamp of measurement, wind speed at different hub
height, air pressure, temperature and power generation.
• The original windfarm is masked using normalization.
4
Fig.1. Typical Wind turbine. [1]
5. Dataset description
Solar Dataset
• The dataset contains from 21 photovoltaic facilities from Germany
• The nominal power range from 100kW to 8500kW.
• The original solar farm is masked using normalization.
Load Dataset
• Load data from 89 different nodes.
• The data contains NWP. Which has three hour resolution weather prediction
• Load is taken for 12 months.
5
Fig.2. Hybrid greed system. [2]
6. Data pre-processing
• Cyclical continuous features are converted into two division features
sine and cosine.
• NaN and Null values are removed .
• Load data’s different blocks in h5 file can be accessed using different
keys
• Weather model and load model frame miss match has been dropped.
• Windfarm and Solar farm data was normalized.
• Redundancy is eliminated using normalization on load dataset.
• Drop features with no information
• For example:- forecasting time from wind park dataset.
6
Fig.3. Data pre-processing representation[3]
7. Data Split method
• In first approach, the whole dataset has been split into train and test dataset.
• In the second approach, the whole dataset has been split into train, validation and test set.
7
8. Feature selection
• Highly correlated feature(threshold 0.8 ) has been dropped
using correlation matrix. Wind speed at 10m height
dropped.
• To reduce the training cost, the best features is selected
using mutual info regression from scikit-learn feature
selection.
• This method measures dependency between the
variables.
• Zero means i.i.d.
• High value means higher dependency.
8
Fig.4. Mutula info regression fwature selection method
9. Feature selection
Wrapper method
• Sequential from MLX tend -greedy search algorithm
• D-dimensional to K-dimensional feature vector.(K<D)
• Reduce generalization error and improve computation efficiency by removing irrelevant features and
noise.
9Fig. 5. Performance with 17 features Fig. 6. Performance with 7 features Fig. 7. Performance with 14 features
10. Perceptron
where,
Xi=the input of the neuron.
wi =the weight of each connection to the neuron.
bi =the bias of the neuron.
f(…..) is the acivation function of the neuron.
10
Fig.8. Perceptron Learning Algorithm[8]
11. Feedforward neural network
• Initial weights are randomly initialized
• First observation of dataset feed into the input layer.
• Forward propagation from left to right.
• Measure the error of the prediction.
• Back propagation right to left.
• Weights updates after 100 batch observation.
• Training process finishes after 30 epochs.
11
Fig. 9. A simple Feedforward neural network[4]
12. ANN hyperparameter tuning
For ANN Adam optimizer with learning rate 0.001 and activation function ReLU is chosen.
12
Fig.10. Optimizer Adam lr 0.001 activation ReLU. Fig.11. Optimizer SGD lr.001 activation ReLU. Fig.12. Optimizer RMSprop lr 0.001 activation ReLU.
13. Recurrent Neural Network(RNN)
• Stacked LSTM -multiple LSTM layers.
• Stacked LSTM- 4 hidden layers.
• Return sequence true- One output per input time step
rather than one output time step for all input time steps.
• which Also provide 3D array.
• Each input requires 3 dimensional data
• Each layer provides a sequence output.
• Output single value as a 2d array.
13
Fig. 13. Implemented LSTM architecture.
14. Long Short Term Memory(LSTM)
14
Fig. 14. Typical LSTM architecture.[5] Fig. 15. The repeating module in an LSTM .[5]
15. Long Short Term Memory(LSTM)
• S is the weighted sum input from previous layer
and activated with sigmoid activation function.
• T is the weighted sum input from previous layer
and activated with tanh activation function.
• t- time step.
• X-input.
• h-hidden state, which act as a memory.
• Length of X- size/dimension of input.
• Length of h- no. of hidden state.
• C –cells state, act as a high way for the
sequence chain.
keras using state_size,units
15
Fig. 16. Animated LSTM architecture.[6]
16. LSTM hyperparameter tuning
16
Fig.17. Optimizer Adam lr 0.001 activation ReLU. Fig.18. Optimizer SGD lr.001 activation ReLU. Fig.19. Optimizer RMSprop lr 0.001 activation ReLU.
For RNN LSTM RMSprop optimizer with learning rate 0.001 and activation function ReLU is chosen.
17. Recurrent Convolutional Neural Network(RCNN)
• CNN LSTM –CNN + LSTM.
• CNN –efficiently extract and learn from sequential time series data.
• Each input requires 3 dimensional data
• CNN- to interpret subsequences of input.
• Conv1D- features from short fixed length.
• Automatically learn the salient features.
• Maxplooing 1 –stride size 2.
Simplifies the features maps by keeping ¼ of the values with the largest signal
• Flatten-multi dimensional vector to single dimensional vector.
The distilled features map from maxpooling layer are then flattened into one long
vector
• LSTM-Usage the previous layer output for decoding process.
• Dense- this layer produces output prediction.
17Fig. 20. Implemented CNN-LSTM architecture.
18. Recurrent Convolutional Neural Network(RCNN)
18Fig. 21. A Pooling Layer reducing a feature map by taking the largest value.[7]
19. CNN LSTM hyperparameter tuning
For CNN LSTM RMSprop optimizer with learning rate 0.001 and activation function ReLU is chosen.
19
Fig.22. Optimizer Adam lr 0.001 activation relu. Fig.23. Optimizer SGD lr.001 activation relu. Fig:24. Optimizer SGD lr 0.001 activation relu.
20. Results: ANN training on single dataset
20
Fig.25. ANN applied on single Solarpark dataset. Fig.27. ANN applied on single Loadpark dataset.Fig. 26. ANN applied on single Windpark dataset.
21. Results: LSTM forecasting accuracy on single
dataset
21
Fig.28. RNN LSTM with 1 hour ahead forecast resolution for single Windpark dataset . Fig.29. RNN LSTM with 1 hour ahead forecast resolution for single Windpark dataset .
22. Results: LSTM training on single dataset
22
Fig.30. LSTM applied on single Solarpark dataset. Fig.31. LSTM applied on single Windpark dataset. Fig.32. LSTM applied on single Loadpark dataset.
23. Results: CNN-LSTM training on single dataset.
23
Fig.33. CNN-LSTM applied on single Solarpark dataset. Fig.34. CNN-LSTM applied on single Windpark dataset. Fig.35. CNN-LSTM applied on single Loadpark dataset.
24. Results: NN models forecasting accuracy comparison.
24
Fig.36. NN models performance on whole Solarpark dataset. Fig.37. Boxplot measurement of NN models on whole Solarpark dataset.
25. Results: NN models forecasting accuracy comparison.
25
Fig.38. NN models performance on whole Windpark dataset. Fig.39. Boxplot measurement of NN models on whole Windpark dataset.
26. Results: NN models forecasting accuracy
comparison.
26
Fig.40. NN models performance on whole Loadpark dataset. Fig.41. Boxplot measurement of NN models on whole Loaddpark dataset.
27. Conclusion
and Outlook
27
The sole purpose of this
experiment is to
compare different NN
architecture for short
term forecasting.
Data post-processing
method applied to the
resultant data analysis.
As the Dataset is huge, it
took a lot's amount of
time to implement and
train different NN
models.
The future
improvements for this
project
Implementation of a
more fast learning
algorithm – Auto LSTM.
More robust feature
selection algorithm-
Auto Encoder.
Test more possible
neural networks and
compare their overall
performance.