際際滷

際際滷Share a Scribd company logo
Machine Learning
Neural Networks
際際滷s mostly adapted from Tom
Mithcell, Han and Kamber
Artificial Neural Networks
 Computational models inspired by the human
brain:
 Algorithms that try to mimic the brain.
 Massively parallel, distributed system, made up of
simple processing units (neurons)
 Synaptic connection strengths among neurons are
used to store the acquired knowledge.
 Knowledge is acquired by the network from its
environment through a learning process
History
 late-1800's - Neural Networks appear as an
analogy to biological systems
 1960's and 70's  Simple neural networks appear
 Fall out of favor because the perceptron is not
effective by itself, and there were no good algorithms
for multilayer nets
 1986  Backpropagation algorithm appears
 Neural Networks have a resurgence in popularity
 More computationally expensive
Applications of ANNs
 ANNs have been widely used in various domains
for:
 Pattern recognition
 Function approximation
 Associative memory
Properties
 Inputs are flexible
 any real values
 Highly correlated or independent
 Target function may be discrete-valued, real-valued, or
vectors of discrete or real values
 Outputs are real numbers between 0 and 1
 Resistant to errors in the training data
 Long training time
 Fast evaluation
 The function produced can be difficult for humans to
interpret
When to consider neural networks
 Input is high-dimensional discrete or raw-valued
 Output is discrete or real-valued
 Output is a vector of values
 Possibly noisy data
 Form of target function is unknown
 Human readability of the result is not important
Examples:
 Speech phoneme recognition
 Image classification
 Financial prediction
February 10, 2025
Data Mining: Concepts and
Techniques 7
A Neuron (= a perceptron)
 The n-dimensional input vector x is mapped into variable y by
means of the scalar product and a nonlinear function mapping
t
-
f
weighted
sum
Input
vector x
output y
Activation
function
weight
vector w

w0
w1
wn
x0
x1
xn
)
sign(
y
e
For Exampl
n
0
i
t
x
w i
i
Perceptron
 Basic unit in a neural network
 Linear separator
 Parts
 N inputs, x1 ... xn
 Weights for each input, w1 ... wn
 A bias input x0 (constant) and associated weight w0
 Weighted sum of inputs, y = w0x0 + w1x1 + ... + wnxn
 A threshold function or activation function,
i.e 1 if y > t, -1 if y <= t
Artificial Neural Networks (ANN)
 Model is an assembly of
inter-connected nodes
and weighted links
 Output node sums up
each of its input value
according to the weights
of its links
 Compare output node
against some threshold t

X1
X2
X3
Y
Black box
w1
t
Output
node
Input
nodes
w2
w3
)
( t
x
w
I
Y
i
i
i 
 
Perceptron Model
)
( t
x
w
sign
Y
i
i
i 
 
or
Types of connectivity
 Feedforward networks
 These compute a series of
transformations
 Typically, the first layer is the
input and the last layer is the
output.
 Recurrent networks
 These have directed cycles in their
connection graph. They can have
complicated dynamics.
 More biologically realistic.
hidden units
output units
input units
Different Network Topologies
 Single layer feed-forward networks
 Input layer projecting into the output layer
Input Output
layer layer
Single layer
network
Different Network Topologies
 Multi-layer feed-forward networks
 One or more hidden layers. Input projects only
from previous layers onto a layer.
Input Hidden Output
layer layer layer
2-layer or
1-hidden layer
fully connected
network
Different Network Topologies
 Multi-layer feed-forward networks
Input Hidden Output
layer layers layer
Different Network Topologies
 Recurrent networks
 A network with feedback, where some of its
inputs are connected to some of its outputs (discrete
time).
Input Output
layer layer
Recurrent
network
Algorithm for learning ANN
 Initialize the weights (w0, w1, , wk)
 Adjust the weights in such a way that the output
of ANN is consistent with class labels of training
examples
 Error function:
 Find the weights wis that minimize the above error
function
 e.g., gradient descent, backpropagation algorithm
 2
)
,
(
 

i
i
i
i X
w
f
Y
E
Optimizing concave/convex function
 Maximum of a concave function = minimum of a
convex function
Gradient ascent (concave) / Gradient descent (convex)
Gradient ascent rule
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Decision surface of a perceptron
 Decision surface is a hyperplane
 Can capture linearly separable classes
 Non-linearly separable
 Use a network of them
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Multi-layer Networks
 Linear units inappropriate
 No more expressive than a single layer
  Introduce non-linearity
 Threshold not differentiable
  Use sigmoid function
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial Intelligence
February 10, 2025
Data Mining: Concepts and
Techniques 31
Backpropagation
 Iteratively process a set of training tuples & compare the network's
prediction with the actual known target value
 For each training tuple, the weights are modified to minimize the mean
squared error between the network's prediction and the actual target
value
 Modifications are made in the backwards direction: from the output
layer, through each hidden layer down to the first hidden layer, hence
backpropagation
 Steps
 Initialize weights (to small random #s) and biases in the network
 Propagate the inputs forward (by applying activation function)
 Backpropagate the error (by updating weights and biases)
 Terminating condition (when error is very small, etc.)
Machine Learning Neural Networks Artificial Intelligence
February 10, 2025
Data Mining: Concepts and
Techniques 33
How A Multi-Layer Neural Network Works?
 The inputs to the network correspond to the attributes measured for
each training tuple
 Inputs are fed simultaneously into the units making up the input layer
 They are then weighted and fed simultaneously to a hidden layer
 The number of hidden layers is arbitrary, although usually only one
 The weighted outputs of the last hidden layer are input to units making
up the output layer, which emits the network's prediction
 The network is feed-forward in that none of the weights cycles back to
an input unit or to an output unit of a previous layer
 From a statistical point of view, networks perform nonlinear regression:
Given enough hidden units and enough training samples, they can
closely approximate any function
February 10, 2025
Data Mining: Concepts and
Techniques 34
Defining a Network Topology
 First decide the network topology: # of units in the input
layer, # of hidden layers (if > 1), # of units in each hidden
layer, and # of units in the output layer
 Normalizing the input values for each attribute measured in
the training tuples to [0.01.0]
 One input unit per domain value, each initialized to 0
 Output, if for classification and more than two classes, one
output unit per class is used
 Once a network has been trained and its accuracy is
unacceptable, repeat the training process with a different
network topology or a different set of initial weights
February 10, 2025
Data Mining: Concepts and
Techniques 35
Backpropagation and Interpretability
 Efficiency of backpropagation: Each epoch (one interation through the
training set) takes O(|D| * w), with |D| tuples and w weights, but # of
epochs can be exponential to n, the number of inputs, in the worst case
 Rule extraction from networks: network pruning
 Simplify the network structure by removing weighted links that have the
least effect on the trained network
 Then perform link, unit, or activation value clustering
 The set of input and activation values are studied to derive rules
describing the relationship between the input and hidden unit layers
 Sensitivity analysis: assess the impact that a given input variable has on a
network output. The knowledge gained from this analysis can be
represented in rules
February 10, 2025
Data Mining: Concepts and
Techniques 36
Neural Network as a Classifier
 Weakness
 Long training time
 Require a number of parameters typically best determined empirically,
e.g., the network topology or structure.
 Poor interpretability: Difficult to interpret the symbolic meaning behind
the learned weights and of hidden units in the network
 Strength
 High tolerance to noisy data
 Ability to classify untrained patterns
 Well-suited for continuous-valued inputs and outputs
 Successful on a wide array of real-world data
 Algorithms are inherently parallel
 Techniques have recently been developed for the extraction of rules
from trained neural networks
Machine Learning Neural Networks Artificial Intelligence
Artificial Neural Networks (ANN)
X1 X2 X3 Y
1 0 0 0
1 0 1 1
1 1 0 1
1 1 1 1
0 0 1 0
0 1 0 0
0 1 1 1
0 0 0 0

X1
X2
X3
Y
Black box
0.3
0.3
0.3 t=0.4
Output
node
Input
nodes









otherwise
0
true
is
if
1
)
(
where
)
0
4
.
0
3
.
0
3
.
0
3
.
0
( 3
2
1
z
z
I
X
X
X
I
Y
Learning Perceptrons
February 10, 2025
Data Mining: Concepts and
Techniques 40
A Multi-Layer Feed-Forward Neural Network
Output layer
Input layer
Hidden layer
Output vector
Input vector: X
wij
ij
k
i
i
k
j
k
j x
y
y
w
w )

( )
(
)
(
)
1
(
General Structure of ANN
Activation
function
g(Si
)
Si
Oi
I1
I2
I3
wi1
wi2
wi3
Oi
Neuron i
Input Output
threshold, t
Input
Layer
Hidden
Layer
Output
Layer
x1
x2
x3
x4
x5
y
Training ANN means learning
the weights of the neurons

More Related Content

Similar to Machine Learning Neural Networks Artificial Intelligence (20)

MNN
MNNMNN
MNN
International Islamic University
Multi Layer Network
Multi Layer NetworkMulti Layer Network
Multi Layer Network
International Islamic University
Classification by backpropacation
Classification by backpropacationClassification by backpropacation
Classification by backpropacation
Siva Priya
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagation
Krish_ver2
Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks
ShwethaShreeS
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
Data Mining: Concepts and techniques classification _chapter 9 :advanced methodsData Mining: Concepts and techniques classification _chapter 9 :advanced methods
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
Salah Amean
Chapter 09 classification advanced
Chapter 09 classification advancedChapter 09 classification advanced
Chapter 09 classification advanced
Houw Liong The
Chapter 09 class advanced
Chapter 09 class advancedChapter 09 class advanced
Chapter 09 class advanced
Houw Liong The
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cse
NaveenBhajantri1
AI-CH5 (ANN) - Artificial Neural Network
AI-CH5 (ANN) - Artificial Neural NetworkAI-CH5 (ANN) - Artificial Neural Network
AI-CH5 (ANN) - Artificial Neural Network
abrahadawit101
09 classadvanced
09 classadvanced09 classadvanced
09 classadvanced
JoonyoungJayGwak
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
csandit
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
cscpconf
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhvJyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
testankitkush
Artificial neural network by arpit_sharma
Artificial neural network by arpit_sharmaArtificial neural network by arpit_sharma
Artificial neural network by arpit_sharma
Er. Arpit Sharma
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithmBack_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
Chapter 9. Classification Advanced Methods.ppt
Chapter 9. Classification Advanced Methods.pptChapter 9. Classification Advanced Methods.ppt
Chapter 9. Classification Advanced Methods.ppt
Subrata Kumer Paul
artificial-neural-network-seminar-report.docx
artificial-neural-network-seminar-report.docxartificial-neural-network-seminar-report.docx
artificial-neural-network-seminar-report.docx
suriyakalavinoth
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTSNEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
hemasubbu08
Neural-Networks.ppt
Neural-Networks.pptNeural-Networks.ppt
Neural-Networks.ppt
RINUSATHYAN
Classification by backpropacation
Classification by backpropacationClassification by backpropacation
Classification by backpropacation
Siva Priya
2.5 backpropagation
2.5 backpropagation2.5 backpropagation
2.5 backpropagation
Krish_ver2
Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks
ShwethaShreeS
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
Data Mining: Concepts and techniques classification _chapter 9 :advanced methodsData Mining: Concepts and techniques classification _chapter 9 :advanced methods
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
Salah Amean
Chapter 09 classification advanced
Chapter 09 classification advancedChapter 09 classification advanced
Chapter 09 classification advanced
Houw Liong The
Chapter 09 class advanced
Chapter 09 class advancedChapter 09 class advanced
Chapter 09 class advanced
Houw Liong The
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cse
NaveenBhajantri1
AI-CH5 (ANN) - Artificial Neural Network
AI-CH5 (ANN) - Artificial Neural NetworkAI-CH5 (ANN) - Artificial Neural Network
AI-CH5 (ANN) - Artificial Neural Network
abrahadawit101
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
csandit
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
cscpconf
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhvJyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
testankitkush
Artificial neural network by arpit_sharma
Artificial neural network by arpit_sharmaArtificial neural network by arpit_sharma
Artificial neural network by arpit_sharma
Er. Arpit Sharma
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithmBack_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
Chapter 9. Classification Advanced Methods.ppt
Chapter 9. Classification Advanced Methods.pptChapter 9. Classification Advanced Methods.ppt
Chapter 9. Classification Advanced Methods.ppt
Subrata Kumer Paul
artificial-neural-network-seminar-report.docx
artificial-neural-network-seminar-report.docxartificial-neural-network-seminar-report.docx
artificial-neural-network-seminar-report.docx
suriyakalavinoth
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTSNEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
hemasubbu08
Neural-Networks.ppt
Neural-Networks.pptNeural-Networks.ppt
Neural-Networks.ppt
RINUSATHYAN

Recently uploaded (20)

Selzy: Simplifying Email Marketing for Maximum Growth
Selzy: Simplifying Email Marketing for Maximum GrowthSelzy: Simplifying Email Marketing for Maximum Growth
Selzy: Simplifying Email Marketing for Maximum Growth
Selzy
Data Management on AWS: from caos to centralized governance - 2025-03-26
Data Management on AWS: from caos to centralized governance - 2025-03-26Data Management on AWS: from caos to centralized governance - 2025-03-26
Data Management on AWS: from caos to centralized governance - 2025-03-26
Alessandra Bilardi
Know Your Nation In Numbers myIndia-2006
Know Your Nation In Numbers myIndia-2006Know Your Nation In Numbers myIndia-2006
Know Your Nation In Numbers myIndia-2006
sahimbarc
Digital Marketing Canvas for Charlotte Hornets
Digital Marketing Canvas for Charlotte HornetsDigital Marketing Canvas for Charlotte Hornets
Digital Marketing Canvas for Charlotte Hornets
DylanLee69
data compression.ppt tree structure vector
data compression.ppt tree structure vectordata compression.ppt tree structure vector
data compression.ppt tree structure vector
vidhyaminnalveeran29
Employee data login and attendance for region
Employee data login and attendance for regionEmployee data login and attendance for region
Employee data login and attendance for region
nagom47355
PostGIS Workshop: a comprehensive tutorial.ppt
PostGIS Workshop: a comprehensive tutorial.pptPostGIS Workshop: a comprehensive tutorial.ppt
PostGIS Workshop: a comprehensive tutorial.ppt
LonJames2
GenAI-powered assistants compared in a real case - 2025-03-18
GenAI-powered assistants compared in a real case - 2025-03-18GenAI-powered assistants compared in a real case - 2025-03-18
GenAI-powered assistants compared in a real case - 2025-03-18
Alessandra Bilardi
[3] Storing and backup - AIESEC Session 2
[3] Storing and backup - AIESEC Session 2[3] Storing and backup - AIESEC Session 2
[3] Storing and backup - AIESEC Session 2
mohamedepankolo
Lecture 2-DATABASE MODELS lecture 2.pptx
Lecture 2-DATABASE MODELS lecture 2.pptxLecture 2-DATABASE MODELS lecture 2.pptx
Lecture 2-DATABASE MODELS lecture 2.pptx
elvis24mutura
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Simone Spaccarotella
LITC-Living-in-the-IT-Era-for-CBA-Students.docx
LITC-Living-in-the-IT-Era-for-CBA-Students.docxLITC-Living-in-the-IT-Era-for-CBA-Students.docx
LITC-Living-in-the-IT-Era-for-CBA-Students.docx
JohnMark171
sterategicinformationsystem-250329162230-1990dc92.pptx
sterategicinformationsystem-250329162230-1990dc92.pptxsterategicinformationsystem-250329162230-1990dc92.pptx
sterategicinformationsystem-250329162230-1990dc92.pptx
EliasHaile7
Financial Ratios and CAMEL Presentation.ppt
Financial Ratios and CAMEL Presentation.pptFinancial Ratios and CAMEL Presentation.ppt
Financial Ratios and CAMEL Presentation.ppt
PrinceAyangbesanOlam
ARCH 2025: New Mexico Respite Provider Registry
ARCH 2025: New Mexico Respite Provider RegistryARCH 2025: New Mexico Respite Provider Registry
ARCH 2025: New Mexico Respite Provider Registry
Allen Shaw
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
hugoshan513
AI system mimicking human expert decision-making..pptx
AI system mimicking human expert decision-making..pptxAI system mimicking human expert decision-making..pptx
AI system mimicking human expert decision-making..pptx
ritikacompscience
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdfOrange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
gemmajfrancisco
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptxPLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
bhairamrohit948
Big-O notations, Algorithm and complexity analaysis
Big-O notations, Algorithm and complexity analaysisBig-O notations, Algorithm and complexity analaysis
Big-O notations, Algorithm and complexity analaysis
drsomya2019
Selzy: Simplifying Email Marketing for Maximum Growth
Selzy: Simplifying Email Marketing for Maximum GrowthSelzy: Simplifying Email Marketing for Maximum Growth
Selzy: Simplifying Email Marketing for Maximum Growth
Selzy
Data Management on AWS: from caos to centralized governance - 2025-03-26
Data Management on AWS: from caos to centralized governance - 2025-03-26Data Management on AWS: from caos to centralized governance - 2025-03-26
Data Management on AWS: from caos to centralized governance - 2025-03-26
Alessandra Bilardi
Know Your Nation In Numbers myIndia-2006
Know Your Nation In Numbers myIndia-2006Know Your Nation In Numbers myIndia-2006
Know Your Nation In Numbers myIndia-2006
sahimbarc
Digital Marketing Canvas for Charlotte Hornets
Digital Marketing Canvas for Charlotte HornetsDigital Marketing Canvas for Charlotte Hornets
Digital Marketing Canvas for Charlotte Hornets
DylanLee69
data compression.ppt tree structure vector
data compression.ppt tree structure vectordata compression.ppt tree structure vector
data compression.ppt tree structure vector
vidhyaminnalveeran29
Employee data login and attendance for region
Employee data login and attendance for regionEmployee data login and attendance for region
Employee data login and attendance for region
nagom47355
PostGIS Workshop: a comprehensive tutorial.ppt
PostGIS Workshop: a comprehensive tutorial.pptPostGIS Workshop: a comprehensive tutorial.ppt
PostGIS Workshop: a comprehensive tutorial.ppt
LonJames2
GenAI-powered assistants compared in a real case - 2025-03-18
GenAI-powered assistants compared in a real case - 2025-03-18GenAI-powered assistants compared in a real case - 2025-03-18
GenAI-powered assistants compared in a real case - 2025-03-18
Alessandra Bilardi
[3] Storing and backup - AIESEC Session 2
[3] Storing and backup - AIESEC Session 2[3] Storing and backup - AIESEC Session 2
[3] Storing and backup - AIESEC Session 2
mohamedepankolo
Lecture 2-DATABASE MODELS lecture 2.pptx
Lecture 2-DATABASE MODELS lecture 2.pptxLecture 2-DATABASE MODELS lecture 2.pptx
Lecture 2-DATABASE MODELS lecture 2.pptx
elvis24mutura
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...
Simone Spaccarotella
LITC-Living-in-the-IT-Era-for-CBA-Students.docx
LITC-Living-in-the-IT-Era-for-CBA-Students.docxLITC-Living-in-the-IT-Era-for-CBA-Students.docx
LITC-Living-in-the-IT-Era-for-CBA-Students.docx
JohnMark171
sterategicinformationsystem-250329162230-1990dc92.pptx
sterategicinformationsystem-250329162230-1990dc92.pptxsterategicinformationsystem-250329162230-1990dc92.pptx
sterategicinformationsystem-250329162230-1990dc92.pptx
EliasHaile7
Financial Ratios and CAMEL Presentation.ppt
Financial Ratios and CAMEL Presentation.pptFinancial Ratios and CAMEL Presentation.ppt
Financial Ratios and CAMEL Presentation.ppt
PrinceAyangbesanOlam
ARCH 2025: New Mexico Respite Provider Registry
ARCH 2025: New Mexico Respite Provider RegistryARCH 2025: New Mexico Respite Provider Registry
ARCH 2025: New Mexico Respite Provider Registry
Allen Shaw
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...
hugoshan513
AI system mimicking human expert decision-making..pptx
AI system mimicking human expert decision-making..pptxAI system mimicking human expert decision-making..pptx
AI system mimicking human expert decision-making..pptx
ritikacompscience
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdfOrange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
Orange County Tableau User Group 2025 Late Q1 2025-03-23.pdf
gemmajfrancisco
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptxPLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
PLAN_OF_WORK_PPT_BY_ROHIT_BHAIRAM_--2212020201003[1] new.pptx
bhairamrohit948
Big-O notations, Algorithm and complexity analaysis
Big-O notations, Algorithm and complexity analaysisBig-O notations, Algorithm and complexity analaysis
Big-O notations, Algorithm and complexity analaysis
drsomya2019

Machine Learning Neural Networks Artificial Intelligence

  • 1. Machine Learning Neural Networks 際際滷s mostly adapted from Tom Mithcell, Han and Kamber
  • 2. Artificial Neural Networks Computational models inspired by the human brain: Algorithms that try to mimic the brain. Massively parallel, distributed system, made up of simple processing units (neurons) Synaptic connection strengths among neurons are used to store the acquired knowledge. Knowledge is acquired by the network from its environment through a learning process
  • 3. History late-1800's - Neural Networks appear as an analogy to biological systems 1960's and 70's Simple neural networks appear Fall out of favor because the perceptron is not effective by itself, and there were no good algorithms for multilayer nets 1986 Backpropagation algorithm appears Neural Networks have a resurgence in popularity More computationally expensive
  • 4. Applications of ANNs ANNs have been widely used in various domains for: Pattern recognition Function approximation Associative memory
  • 5. Properties Inputs are flexible any real values Highly correlated or independent Target function may be discrete-valued, real-valued, or vectors of discrete or real values Outputs are real numbers between 0 and 1 Resistant to errors in the training data Long training time Fast evaluation The function produced can be difficult for humans to interpret
  • 6. When to consider neural networks Input is high-dimensional discrete or raw-valued Output is discrete or real-valued Output is a vector of values Possibly noisy data Form of target function is unknown Human readability of the result is not important Examples: Speech phoneme recognition Image classification Financial prediction
  • 7. February 10, 2025 Data Mining: Concepts and Techniques 7 A Neuron (= a perceptron) The n-dimensional input vector x is mapped into variable y by means of the scalar product and a nonlinear function mapping t - f weighted sum Input vector x output y Activation function weight vector w w0 w1 wn x0 x1 xn ) sign( y e For Exampl n 0 i t x w i i
  • 8. Perceptron Basic unit in a neural network Linear separator Parts N inputs, x1 ... xn Weights for each input, w1 ... wn A bias input x0 (constant) and associated weight w0 Weighted sum of inputs, y = w0x0 + w1x1 + ... + wnxn A threshold function or activation function, i.e 1 if y > t, -1 if y <= t
  • 9. Artificial Neural Networks (ANN) Model is an assembly of inter-connected nodes and weighted links Output node sums up each of its input value according to the weights of its links Compare output node against some threshold t X1 X2 X3 Y Black box w1 t Output node Input nodes w2 w3 ) ( t x w I Y i i i Perceptron Model ) ( t x w sign Y i i i or
  • 10. Types of connectivity Feedforward networks These compute a series of transformations Typically, the first layer is the input and the last layer is the output. Recurrent networks These have directed cycles in their connection graph. They can have complicated dynamics. More biologically realistic. hidden units output units input units
  • 11. Different Network Topologies Single layer feed-forward networks Input layer projecting into the output layer Input Output layer layer Single layer network
  • 12. Different Network Topologies Multi-layer feed-forward networks One or more hidden layers. Input projects only from previous layers onto a layer. Input Hidden Output layer layer layer 2-layer or 1-hidden layer fully connected network
  • 13. Different Network Topologies Multi-layer feed-forward networks Input Hidden Output layer layers layer
  • 14. Different Network Topologies Recurrent networks A network with feedback, where some of its inputs are connected to some of its outputs (discrete time). Input Output layer layer Recurrent network
  • 15. Algorithm for learning ANN Initialize the weights (w0, w1, , wk) Adjust the weights in such a way that the output of ANN is consistent with class labels of training examples Error function: Find the weights wis that minimize the above error function e.g., gradient descent, backpropagation algorithm 2 ) , ( i i i i X w f Y E
  • 16. Optimizing concave/convex function Maximum of a concave function = minimum of a convex function Gradient ascent (concave) / Gradient descent (convex) Gradient ascent rule
  • 24. Decision surface of a perceptron Decision surface is a hyperplane Can capture linearly separable classes Non-linearly separable Use a network of them
  • 27. Multi-layer Networks Linear units inappropriate No more expressive than a single layer Introduce non-linearity Threshold not differentiable Use sigmoid function
  • 31. February 10, 2025 Data Mining: Concepts and Techniques 31 Backpropagation Iteratively process a set of training tuples & compare the network's prediction with the actual known target value For each training tuple, the weights are modified to minimize the mean squared error between the network's prediction and the actual target value Modifications are made in the backwards direction: from the output layer, through each hidden layer down to the first hidden layer, hence backpropagation Steps Initialize weights (to small random #s) and biases in the network Propagate the inputs forward (by applying activation function) Backpropagate the error (by updating weights and biases) Terminating condition (when error is very small, etc.)
  • 33. February 10, 2025 Data Mining: Concepts and Techniques 33 How A Multi-Layer Neural Network Works? The inputs to the network correspond to the attributes measured for each training tuple Inputs are fed simultaneously into the units making up the input layer They are then weighted and fed simultaneously to a hidden layer The number of hidden layers is arbitrary, although usually only one The weighted outputs of the last hidden layer are input to units making up the output layer, which emits the network's prediction The network is feed-forward in that none of the weights cycles back to an input unit or to an output unit of a previous layer From a statistical point of view, networks perform nonlinear regression: Given enough hidden units and enough training samples, they can closely approximate any function
  • 34. February 10, 2025 Data Mining: Concepts and Techniques 34 Defining a Network Topology First decide the network topology: # of units in the input layer, # of hidden layers (if > 1), # of units in each hidden layer, and # of units in the output layer Normalizing the input values for each attribute measured in the training tuples to [0.01.0] One input unit per domain value, each initialized to 0 Output, if for classification and more than two classes, one output unit per class is used Once a network has been trained and its accuracy is unacceptable, repeat the training process with a different network topology or a different set of initial weights
  • 35. February 10, 2025 Data Mining: Concepts and Techniques 35 Backpropagation and Interpretability Efficiency of backpropagation: Each epoch (one interation through the training set) takes O(|D| * w), with |D| tuples and w weights, but # of epochs can be exponential to n, the number of inputs, in the worst case Rule extraction from networks: network pruning Simplify the network structure by removing weighted links that have the least effect on the trained network Then perform link, unit, or activation value clustering The set of input and activation values are studied to derive rules describing the relationship between the input and hidden unit layers Sensitivity analysis: assess the impact that a given input variable has on a network output. The knowledge gained from this analysis can be represented in rules
  • 36. February 10, 2025 Data Mining: Concepts and Techniques 36 Neural Network as a Classifier Weakness Long training time Require a number of parameters typically best determined empirically, e.g., the network topology or structure. Poor interpretability: Difficult to interpret the symbolic meaning behind the learned weights and of hidden units in the network Strength High tolerance to noisy data Ability to classify untrained patterns Well-suited for continuous-valued inputs and outputs Successful on a wide array of real-world data Algorithms are inherently parallel Techniques have recently been developed for the extraction of rules from trained neural networks
  • 38. Artificial Neural Networks (ANN) X1 X2 X3 Y 1 0 0 0 1 0 1 1 1 1 0 1 1 1 1 1 0 0 1 0 0 1 0 0 0 1 1 1 0 0 0 0 X1 X2 X3 Y Black box 0.3 0.3 0.3 t=0.4 Output node Input nodes otherwise 0 true is if 1 ) ( where ) 0 4 . 0 3 . 0 3 . 0 3 . 0 ( 3 2 1 z z I X X X I Y
  • 40. February 10, 2025 Data Mining: Concepts and Techniques 40 A Multi-Layer Feed-Forward Neural Network Output layer Input layer Hidden layer Output vector Input vector: X wij ij k i i k j k j x y y w w ) ( ) ( ) ( ) 1 (
  • 41. General Structure of ANN Activation function g(Si ) Si Oi I1 I2 I3 wi1 wi2 wi3 Oi Neuron i Input Output threshold, t Input Layer Hidden Layer Output Layer x1 x2 x3 x4 x5 y Training ANN means learning the weights of the neurons