Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
The Presentation answers various questions such as what is machine learning, how machine learning works, the difference between artificial intelligence, machine learning, deep learning, types of machine learning, and its applications.
The term Machine Learning was coined by Arthur Samuel in 1959, an american pioneer in the field of computer gaming and artificial intelligence and stated that “ it gives computers the ability to learn without being explicitly programmed” And in 1997, Tom Mitchell gave a “ well-Posed” mathematical and relational definition that “ A Computer Program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E”.
Machine learning is needed for tasks that are too complex for humans to code directly. So instead, we provide a large amount of data to a machine learning algorithm and let the algorithm work it out by exploring that data and searching for a model that will achieve what the programmers have set it out to achieve.
Uncertain Knowledge and Reasoning in Artificial IntelligenceExperfy
Learn how to take informed decisions based on probabilities and expert knowledge
Understand and explore one of the most exciting advances in AI in the last decades.
Many hands-on examples, including Python code.
Check it out: https://www.experfy.com/training/courses/uncertain-knowledge-and-reasoning-in-artificial-intelligence
Fairly Measuring Fairness In Machine LearningHJ van Veen
We look at a case and two research papers on measuring discrimination in machine learning models for extending credit. Presentation given as part of the Sao Paulo Machine Learning Meetup, theme "Ethics in Data Science".
Intuitive introduction with easy-to-understand explanation of fundamental concepts in machine learning and neural networks. No prior machine learning or computing experience required.
Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things.
The Presentation answers various questions such as what is machine learning, how machine learning works, the difference between artificial intelligence, machine learning, deep learning, types of machine learning, and its applications.
The term Machine Learning was coined by Arthur Samuel in 1959, an american pioneer in the field of computer gaming and artificial intelligence and stated that “ it gives computers the ability to learn without being explicitly programmed” And in 1997, Tom Mitchell gave a “ well-Posed” mathematical and relational definition that “ A Computer Program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E”.
Machine learning is needed for tasks that are too complex for humans to code directly. So instead, we provide a large amount of data to a machine learning algorithm and let the algorithm work it out by exploring that data and searching for a model that will achieve what the programmers have set it out to achieve.
Uncertain Knowledge and Reasoning in Artificial IntelligenceExperfy
Learn how to take informed decisions based on probabilities and expert knowledge
Understand and explore one of the most exciting advances in AI in the last decades.
Many hands-on examples, including Python code.
Check it out: https://www.experfy.com/training/courses/uncertain-knowledge-and-reasoning-in-artificial-intelligence
Fairly Measuring Fairness In Machine LearningHJ van Veen
We look at a case and two research papers on measuring discrimination in machine learning models for extending credit. Presentation given as part of the Sao Paulo Machine Learning Meetup, theme "Ethics in Data Science".
Intuitive introduction with easy-to-understand explanation of fundamental concepts in machine learning and neural networks. No prior machine learning or computing experience required.
Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things. Lecture related to machine learning. Here you can read multiple things.
This slide gives brief overview of supervised, unsupervised and reinforcement learning. Algorithms discussed are Naive Bayes, K nearest neighbour, SVM,decision tree, Markov model.
Difference between regression and classification. difference between supervised and reinforcement, iterative functioning of Markov model and machine learning applications.
This presentation is based on new trends and technology in computer science. In this presentation, we have depicted the basic principles of artificial intelligence. An attempt has been made to explain advanced technologies like machine learning and deep learning.
Machine Learning without the Math: An overview of Machine LearningArshad Ahmed
A brief overview of Machine Learning and its associated tasks from a high level. This presentation discusses key concepts without the maths.The more mathematically inclined are referred to Bishops book on Pattern Recognition and Machine Learning.
الموعد الإثنين 03 يناير 2022
143
مبادرة
#تواصل_تطوير
المحاضرة ال 143 من المبادرة
المهندس / محمد الرافعي طرباي
نقيب المبرمجين بالدقهلية
بعنوان
"IT INDUSTRY"
How To Getting Into IT With Zero Experience
وذلك يوم الإثنين 03 يناير2022
السابعة مساء توقيت القاهرة
الثامنة مساء توقيت مكة المكرمة
و الحضور من تطبيق زووم
https://us02web.zoom.us/meeting/register/tZUpf-GsrD4jH9N9AxO39J013c1D4bqJNTcu
علما ان هناك بث مباشر للمحاضرة على القنوات الخاصة بجمعية المهندسين المصريين
ونأمل أن نوفق في تقديم ما ينفع المهندس ومهمة الهندسة في عالمنا العربي
والله الموفق
للتواصل مع إدارة المبادرة عبر قناة التليجرام
https://t.me/EEAKSA
ومتابعة المبادرة والبث المباشر عبر نوافذنا المختلفة
رابط اللينكدان والمكتبة الالكترونية
https://www.linkedin.com/company/eeaksa-egyptian-engineers-association/
رابط قناة التويتر
https://twitter.com/eeaksa
رابط قناة الفيسبوك
https://www.facebook.com/EEAKSA
رابط قناة اليوتيوب
https://www.youtube.com/user/EEAchannal
رابط التسجيل العام للمحاضرات
https://forms.gle/vVmw7L187tiATRPw9
ملحوظة : توجد شهادات حضور مجانية لمن يسجل فى رابط التقيم اخر المحاضرة
ensemble methods use multiple learning algorithms to obtain better predictive performance than could be obtained from any of the constituent learning algorithms alone.
An ensemble is itself a supervised learning algorithm, because it can be trained and then used to make predictions. The trained ensemble, therefore, represents a single hypothesis. This hypothesis, however, is not necessarily contained within the hypothesis space of the models from which it is built.
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
2. Hey Y'all
Ayodele Odubela
I'm a Data Scientist at MINDBODY.
I have a Master's Degree in Data Science from Regis
University.
3 years of experience working with machine learning.
3. Today's Workshop
What's in an Algorithm?
ML vs AI
Machine Learning Applications
Math for Machine Learning
Types of ML
What We Can Predict
Decision Trees
Neural Networks
Natural Language Processing
Code Along
Wrap Up
7. ARTIFICIAL
INTELLIGENCE
“Artificial intelligence is the science
of making computers behave in
ways that we thought required
human intelligence.”
MACHINE
LEARNING
Machine learning algorithms can
figure out how to perform
important tasks by generalizing
from examples.
Andrew Moore
Carnegie Mellon University
Pedro Domingos
University of Washington
9. MACHINE LEARNING IN THE WILD
VIDEO GAME
ENEMIES
AI is the foundation of
many video games. From
controlling NPCs to
playing against the AI
trained on thousands of
past games.
PERSONAL
ASSISTANTS
Services like Siri, Alexa,
Google Assistant use
both audio processing
and natural language
processing to retrieve
results and help you
send texts with your
voice.
ROUTE
OPTIMIZATION
Google Maps and Waze
can run hundreds of
potential routes to get
you to your destination
quickest. Many search
algorithms used in route
optimization are
fundamental to artificial
intelligence.
MOVIE
RECCOMENDATIONS
Netflix's movie
reccomendaions
use collaborative
filtering which helps
suggest movies based
on our past views and
people similar to us.
10. WILD MACHINE LEARNING
POLICING
Police use systems to
predict where crimes will
happen and deploy more
officers to an at risk area.
This leads to an increase
of arrests and without
feedback they succumb
to confirmation bias.
CREDIT
SCORING
These algorithms asses
the risk a creditor takes
on by giving you a loan
or credit card. Inputs like
zip code can be useed as
a proxy for race.
90210 vs any South
Central.
RECIDIVISM
These algorithms are
used in counties across
the nation to predict
which incarcerated
people will commit
another crime after
release,
HR
SCREENINGS
Companies looking to
harness machine
learning in HR should be
weary of perpetuating
the same workplace bias
and hiring practices.
Does a candidate have to
look like successful
employees?.
11. Math Foundations
BIGGEST BARRIER TO ENTRY
• Linear Algebra
• Calculus
• Statistics
• Discrete Math
• Logical Operators
• Probability
• Statistics
WHAT'S USED FREQUENTLYWHAT THEY SAY YOU NEED
12.
13. SUPERVISED
The machine is shown data, but
there are labelled answers for if the
prediction is right or wrong.
This learning is supervised because it
requires input data to be properly
laabelled and often a binary
classification column is added as a
response variable.
LABELED OR NAH?
14. UNSUPERVISED
CREATES SIMILAR GROUPINGS
Unsu[pervised models don't have a
list of ground truth, but tends to have
adifferent goal.
Unsupervised learning methods
usually serve one of two purposes.
To cluster groups or to reduce
dimensionality.
15. REINFORCEMENTA way of letting a system learn by
navigating its surroundings without
guidance and improving with
performance the more it
understands its current state.
A RL model will calculate the value
of being in one state. There are
rewards for good actions (ie. roomba
in a tile and the state went from dirty
to clean) and penalties for bad ones.
WHAT MANY CONSIDER AI
17. CLASSIFICATIONIn this case a machine learning
model will predict the class of the
inputs.
Your model will output whether it
thinks someone has heart disease or
not (0 or 1) or what segment a
customer is in (multi-class)
DISCRETE/CATEGORICAL VARIABLE
18. REGRESSION
REAL/CONTINUOUS NUMBERS
The model predicts a value based on
past data. In a Linear dataset the
regression values will likely be
"through" the values it's trained on.
Your model will output what
temperature it will be tomorrow, the
price of Bitcoin, or the number of
people who will see the live action
Lion King movie.
19. DECISION TREES
• Breaks a dataset into small
subsets
• Tree structure includes
decision nodes and leaf nodes
• Root node is the best predictor
SUPERVISED
20. Entropy
• A measure of the degree of randomness in a variable
• "Good" Decision Trees have homogenous leaf nodes
• The higher the entropy, the harder to draw conclusions
21. Information Gain
• Used to decide which of the attributes are most relevant
• The purpose is to find the attribute that returns the most information gain
• Expected information gain = decrease in entropy
• The less random the variables, the more information is gained
22. Gini Index
• Measures how impure a node is
• Calculated per node
Gini index is used in CART
(Classification and Regression Trees)
IID3 search algorithm uses entropy and infromation gain
23. Pre-pruning
Involves setting the tree parameters before building it so it stops early without
completely being built.
Variables to tune:
⚬ Set max tree depth
⚬ Set max terminal nodes
⚬ Set max number of features
⚬ Set max samples for a node split
■ controls the size of terminal nodes
24. Post-pruning
• Validate the performance of the model on a test
• Cut back splits that seem to overfit the noise in the training set
• Removes a branch from a fully grown tree
• Available in R, but not Python's scikit-learn package
28. Forward Propogation
(aka making an
inference)
Calculate the weight input to the
hidden layer.
Mulitply weight by input and pass to
next layer.
Apply an activation function
Calculate this again to go from
hidden layer to output
29. Back Propogation
The output from forward
propogation is the predicted value.
We use a loss function to compare
the predicted value to the actual
value.
30. Learning Rate
• Gradient Descent is used to get
ideal weights for each neuron
• The learning rate is how fast or
slow you want the machine to
update weight values
WHAT YOU NEED TO KNOW:
• Learning rate should be high
enough to converge* in a
reasonable amount of time
• It should also be small enough to
find the local minima
Convergence is when the output
gets closer to the minima
31. Activation Functions
The job of an activation function is to convert an input to an output signal.
This is based on a mathematical threshold. The activation function tells
the node to activate once a critera has been met.
If a model thinks there's a 51% chance an image is a dog, the activation
function will output a prediction of dog (depending on your function)
32. Gradient Descent
• Gradient Descent is an iterative
machine learning optimization
algorithm to reduce the loss
function.
• Having a low loss function
means predicted values are
close to actual values
33. NATURAL
LANGUAGE
PROCESSING
Based on the field of lingustics,
natural language processing is aided
by new text analytics packages and
the abundance of sample text online.
Challenges:
• Thousands of languages with
hundreds of thousands of words.
• Complex syntax (varying words
per sentence, relative clauses)
• Many ambiguities (special
naames, sarcasm)
34. Word Embeddings
• Models that have mapped a set
of words or phrases to vectors
of numbers,
Most popular are:
• Word2Vec (Trained on Google
News}
• GloVe provided by Stanford
• FastText by Facebook
35. Sentiment Analysis
The polarity of a word of phrase on
how positive or negative it skews.
Types:
• Subjectivity classification
• Polarity classification
• Intent classification
Challenges:
• Biased to the dominant culture
• Many sentiment packages are
based off linguisitc work that is
not universal,
• Variance in individual speech
not a factor
36. Lexical Density
• The number of meaningful words
• After removing stop words like "the", "and",
"I", etc. lexical density is thenumber of
words that add content divided by the total
number of words.
She told him that she loved him
2 lexical words out of 7 total words
28.57% lexical density
37. Markov Chains for NLP
• First a dictionary is built based
on historical texts. They key is a
given word in a sentence and
the results are natural follow up
words.
• Next calculate the word most
likely to follow a given word.
39. Create-Your-Own Kanye Lyrics
GOAL
Generate rap verses that
almost sound like they
could belong to Kanye.
METHOD
Use a Markov Chain to
generate a new verse of
a Kany-AI song.
USE CASE
Perhaps you want to see
if your verses can fool
some fans.
42. Classification Accuracy
The most common metric you
might here will be accuracy
It's not always the best metric for
any given model
43. Confusion Matrices
Table that visualizes the
performance of a classification
algorithm.
Rows represent predicted class
Columns represent the actual
class
44. Normalization
• The process of getting all data
for predictions on the same
scale
• Most algorithms have trouble
performing well with data on
multiple scales.
• Usually between 0 and 1
• Step data pre-processing for
machine learning
45. Overfitting
When a model describes the pattern too
well.This means it has essentially
memorized the data without learning. We
say it hasn't learned because an overfit
model generalizes poorly to new data.
This is one of the biggest problems with
machine learning is that we taught our
machines what we told ourselves not to do.
Don't just memorize, learn.
46. Regularizarion
Adds a penalty to your model to avoid the
model becoming too complex and
overfitting the data.
Multiple methods to do this (L1 and L2
regularization)
If there is noise in the trianing data
regularization shrinks the learned "noise"
towards 0
Regularization reduces the variance of the
model without a major increase in bias
47. DATA IS VALUABLE
IF IT'S
PROTECTED
Does your favorite
website or app allow you
to use two-factor
authentication? How do
companies protect your
data?
WHO'S USING
IT
Our data has and will
continue to be used
against us. We question
the possibility of ever
having a free and fair
election.
WHO
COLLECTS IT
As we know from the
Cambridge Analytica
scandal, companies can
be dubious with our data.
and which developers
have access to it.
WHO OWNS IT
Recent outcry against
the privacy issues with
apps like FaceApp have
been at the forefront of
tech news.
48. USING ML IN YOUR PROJECTS
INVESTIGATE
DATA
TRAIN YOUR
MODEL
MODEL
EVALUATION
54. PICKING A MODEL
NO FREE LUNCH
• What type of data are my outputs?
• What am I trying to predict
• What is wrong with the data? (small sample size, imbalanced classes)
• What data cleaning did I do?
• Will this be a problem when runnning the model in the real world?
55. The Future: XAI
eXplainable AI
TCAV- Testing with Concept Activation Vectors is a new interpretability
method to understand what signals your neural networks models uses for
prediction.