This document provides an overview of artificial neural networks (ANNs). It discusses how ANNs are inspired by biological neural networks and are composed of interconnected nodes that mimic neurons. ANNs use a learning process to update synaptic connection weights between nodes based on training data to perform tasks like pattern recognition. The document outlines the history of ANNs and covers popular applications. It also describes common ANN properties, architectures, and the backpropagation algorithm used for training multilayer networks.
This document provides an overview of artificial neural networks (ANNs). It discusses how ANNs are inspired by biological neural networks and are composed of interconnected nodes that mimic neurons. ANNs use a learning process to update synaptic connection weights between nodes based on training data to perform tasks like pattern recognition. The document outlines the history of ANNs and covers popular applications. It also describes common ANN properties, architectures, and the backpropagation algorithm used for training multilayer networks.
ANNs have been widely used in various domains for: Pattern recognition Funct...vijaym148
油
The document discusses artificial neural networks (ANNs), which are computational models inspired by the human brain. ANNs consist of interconnected nodes that mimic neurons in the brain. Knowledge is stored in the synaptic connections between neurons. ANNs can be used for pattern recognition, function approximation, and associative memory. Backpropagation is an important algorithm for training multilayer ANNs by adjusting the synaptic weights based on examples. ANNs have been applied to problems like image classification, speech recognition, and financial prediction.
The document describes a multilayer neural network presentation. It discusses key concepts of neural networks including their architecture, types of neural networks, and backpropagation. The key points are:
1) Neural networks are composed of interconnected processing units (neurons) that can learn relationships in data through training. They are inspired by biological neural systems.
2) Common network architectures include multilayer perceptrons and recurrent networks. Backpropagation is commonly used to train multilayer feedforward networks by propagating errors backwards.
3) Neural networks have advantages like the ability to model complex nonlinear relationships, adapt to new data, and extract patterns from imperfect data. They are well-suited for problems like classification.
The document describes multilayer neural networks and their use for classification problems. It discusses how neural networks can handle continuous-valued inputs and outputs unlike decision trees. Neural networks are inherently parallel and can be sped up through parallelization techniques. The document then provides details on the basic components of neural networks, including neurons, weights, biases, and activation functions. It also describes common network architectures like feedforward networks and discusses backpropagation for training networks.
The document discusses neural networks and their use for pattern detection and machine learning. It describes how neural networks are modeled after the human nervous system and consist of connected input/output units with associated weights. The key points covered include:
- Neural networks can build highly accurate predictive models by training on large datasets.
- Backpropagation is a common neural network learning algorithm that adjusts weights to predict the correct class label of inputs.
- Neural networks have strengths like tolerance to noisy data and ability to classify untrained patterns, but also weaknesses like long training times and lack of interpretability.
The document discusses artificial neural networks and classification using backpropagation, describing neural networks as sets of connected input and output units where each connection has an associated weight. It explains backpropagation as a neural network learning algorithm that trains networks by adjusting weights to correctly predict the class label of input data, and how multi-layer feed-forward neural networks can be used for classification by propagating inputs through hidden layers to generate outputs.
An artificial neural network (ANN) is the piece of a computing system designed to simulate the way the human brain analyzes and processes information. It is the foundation of artificial intelligence (AI) and solves problems that would prove impossible or difficult by human or statistical standards. ANNs have self-learning capabilities that enable them to produce better results as more data becomes available.
Data Mining: Concepts and techniques classification _chapter 9 :advanced methodsSalah Amean
油
slides contain:
Bayesian Belief Networks,
Classification by Backpropagation,
Support Vector Machines,
Classification by Using Frequent Patterns,
Lazy Learners, (or Learning from Your Neighbors)
Other Classification Methods,
Additional Topics Regarding Classification,
Summary
by
Jiawei Han, Micheline Kamber, and Jian Pei,
University of Illinois at Urbana-Champaign &
Simon Fraser University,
息2013 Han, Kamber & Pei. All rights reserved.
Chapter 9 of the document discusses advanced classification methods including Bayesian belief networks, classification using backpropagation neural networks, support vector machines, classification with frequent patterns, lazy learning, and other techniques. It describes how these methods work, including how Bayesian networks are constructed, how backpropagation trains neural networks, how support vector machines find optimal separating hyperplanes, and considerations around efficiency and interpretability. The chapter also covers mathematical mappings of classification problems and discriminative versus generative classifiers.
Chapter 9 of the book discusses advanced classification methods including Bayesian belief networks, classification using backpropagation neural networks, support vector machines, frequent pattern-based classification, lazy learning, and other techniques. It describes how these methods work, including how to construct Bayesian networks, train neural networks using backpropagation, find optimal separating hyperplanes with support vector machines, and more. The chapter also covers topics like network topologies, training scenarios, efficiency and interpretability of different methods.
Artificial Neural Networks ppt.pptx for final sem cseNaveenBhajantri1
油
This document provides an overview of artificial neural networks. It discusses the biological inspiration from neurons in the brain and how artificial neural networks mimic this structure. The key components of artificial neurons and various network architectures are described, including fully connected, layered, feedforward, and modular networks. Supervised and unsupervised learning approaches are covered, with backpropagation highlighted as a commonly used supervised algorithm. Applications of neural networks are mentioned in areas like medicine, business, marketing and credit evaluation. Advantages include the ability to handle complex nonlinear problems and noisy data.
This chapter discusses advanced classification methods, including Bayesian belief networks, classification using backpropagation neural networks, support vector machines (SVM), and lazy learners. It describes Bayesian belief networks as probabilistic graphical models that represent conditional dependencies between variables. Backpropagation neural networks are introduced as a way to perform nonlinear regression to approximate functions through adjusting weights in a multi-layer feedforward network. SVM is covered as a method that transforms data into a higher dimensional space to find an optimal separating hyperplane, using support vectors.
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...csandit
油
The document describes an algorithm called X-TREPAN that extracts decision trees from trained neural networks. X-TREPAN is an enhancement of the TREPAN algorithm that allows it to handle both multi-class classification and multi-class regression problems. It can also analyze generalized feed forward networks. The algorithm was tested on several real-world datasets and was found to generate decision trees with good classification accuracy while also maintaining comprehensibility.
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...cscpconf
油
In this work, the TREPAN algorithm is enhanced and extended for extracting decision trees from neural networks. We empirically evaluated the performance of the algorithm on a set of databases from real world events. This benchmark enhancement was achieved by adapting Single-test TREPAN and C4.5 decision tree induction algorithms to analyze the datasets. The models are then compared with X-TREPAN for comprehensibility and classification accuracy. Furthermore, we validate the experimentations by applying statistical methods. Finally, the modified algorithm is extended to work with multi-class regression problems and the ability to comprehend generalized feed forward networks is achieved.
This document provides an introduction to artificial neural networks. It discusses biological neurons and how artificial neurons are modeled. The key components of a neural network including the network architecture, learning approaches, and the backpropagation algorithm for supervised learning are described. Applications and advantages of neural networks are also mentioned. Neural networks are modeled after the human brain and learn by modifying connection weights between nodes based on examples.
Jiawei Han, Micheline Kamber and Jian Pei
Data Mining: Concepts and Techniques, 3rd ed.
The Morgan Kaufmann Series in Data Management Systems
Morgan Kaufmann Publishers, July 2011. ISBN 978-0123814791
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTShemasubbu08
油
- Artificial neural networks are computational models inspired by the human brain that use algorithms to mimic brain functions. They are made up of simple processing units (neurons) connected in a massively parallel distributed system. Knowledge is acquired through a learning process that adjusts synaptic connection strengths.
- Neural networks can be used for pattern recognition, function approximation, and associative memory in domains like speech recognition, image classification, and financial prediction. They have flexible inputs, resistant to errors, and fast evaluation, though interpretation is difficult.
The document discusses different types of machine learning paradigms including supervised learning, unsupervised learning, and reinforcement learning. It then provides details on artificial neural networks, describing them as consisting of simple processing units that communicate through weighted connections, similar to neurons in the human brain. The document outlines key aspects of artificial neural networks like processing units, connections between units, propagation rules, and learning methods.
Selzy: Simplifying Email Marketing for Maximum GrowthSelzy
油
This presentation is about Selzy, an easy-to-use and affordable email marketing tool that helps businesses create and launch effective email campaigns with minimal effort. It highlights the challenges of traditional email marketing, showcases Selzys AI-powered email builder, fast setup, and 24/7 support, and demonstrates the tools impact through user growth and market potential. With a strong ROI and a rapidly expanding customer base, Selzy positions itself as a powerful yet simple solution for businesses looking to boost engagement and sales through email marketing.
The document describes a multilayer neural network presentation. It discusses key concepts of neural networks including their architecture, types of neural networks, and backpropagation. The key points are:
1) Neural networks are composed of interconnected processing units (neurons) that can learn relationships in data through training. They are inspired by biological neural systems.
2) Common network architectures include multilayer perceptrons and recurrent networks. Backpropagation is commonly used to train multilayer feedforward networks by propagating errors backwards.
3) Neural networks have advantages like the ability to model complex nonlinear relationships, adapt to new data, and extract patterns from imperfect data. They are well-suited for problems like classification.
The document describes multilayer neural networks and their use for classification problems. It discusses how neural networks can handle continuous-valued inputs and outputs unlike decision trees. Neural networks are inherently parallel and can be sped up through parallelization techniques. The document then provides details on the basic components of neural networks, including neurons, weights, biases, and activation functions. It also describes common network architectures like feedforward networks and discusses backpropagation for training networks.
The document discusses neural networks and their use for pattern detection and machine learning. It describes how neural networks are modeled after the human nervous system and consist of connected input/output units with associated weights. The key points covered include:
- Neural networks can build highly accurate predictive models by training on large datasets.
- Backpropagation is a common neural network learning algorithm that adjusts weights to predict the correct class label of inputs.
- Neural networks have strengths like tolerance to noisy data and ability to classify untrained patterns, but also weaknesses like long training times and lack of interpretability.
The document discusses artificial neural networks and classification using backpropagation, describing neural networks as sets of connected input and output units where each connection has an associated weight. It explains backpropagation as a neural network learning algorithm that trains networks by adjusting weights to correctly predict the class label of input data, and how multi-layer feed-forward neural networks can be used for classification by propagating inputs through hidden layers to generate outputs.
An artificial neural network (ANN) is the piece of a computing system designed to simulate the way the human brain analyzes and processes information. It is the foundation of artificial intelligence (AI) and solves problems that would prove impossible or difficult by human or statistical standards. ANNs have self-learning capabilities that enable them to produce better results as more data becomes available.
Data Mining: Concepts and techniques classification _chapter 9 :advanced methodsSalah Amean
油
slides contain:
Bayesian Belief Networks,
Classification by Backpropagation,
Support Vector Machines,
Classification by Using Frequent Patterns,
Lazy Learners, (or Learning from Your Neighbors)
Other Classification Methods,
Additional Topics Regarding Classification,
Summary
by
Jiawei Han, Micheline Kamber, and Jian Pei,
University of Illinois at Urbana-Champaign &
Simon Fraser University,
息2013 Han, Kamber & Pei. All rights reserved.
Chapter 9 of the document discusses advanced classification methods including Bayesian belief networks, classification using backpropagation neural networks, support vector machines, classification with frequent patterns, lazy learning, and other techniques. It describes how these methods work, including how Bayesian networks are constructed, how backpropagation trains neural networks, how support vector machines find optimal separating hyperplanes, and considerations around efficiency and interpretability. The chapter also covers mathematical mappings of classification problems and discriminative versus generative classifiers.
Chapter 9 of the book discusses advanced classification methods including Bayesian belief networks, classification using backpropagation neural networks, support vector machines, frequent pattern-based classification, lazy learning, and other techniques. It describes how these methods work, including how to construct Bayesian networks, train neural networks using backpropagation, find optimal separating hyperplanes with support vector machines, and more. The chapter also covers topics like network topologies, training scenarios, efficiency and interpretability of different methods.
Artificial Neural Networks ppt.pptx for final sem cseNaveenBhajantri1
油
This document provides an overview of artificial neural networks. It discusses the biological inspiration from neurons in the brain and how artificial neural networks mimic this structure. The key components of artificial neurons and various network architectures are described, including fully connected, layered, feedforward, and modular networks. Supervised and unsupervised learning approaches are covered, with backpropagation highlighted as a commonly used supervised algorithm. Applications of neural networks are mentioned in areas like medicine, business, marketing and credit evaluation. Advantages include the ability to handle complex nonlinear problems and noisy data.
This chapter discusses advanced classification methods, including Bayesian belief networks, classification using backpropagation neural networks, support vector machines (SVM), and lazy learners. It describes Bayesian belief networks as probabilistic graphical models that represent conditional dependencies between variables. Backpropagation neural networks are introduced as a way to perform nonlinear regression to approximate functions through adjusting weights in a multi-layer feedforward network. SVM is covered as a method that transforms data into a higher dimensional space to find an optimal separating hyperplane, using support vectors.
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...csandit
油
The document describes an algorithm called X-TREPAN that extracts decision trees from trained neural networks. X-TREPAN is an enhancement of the TREPAN algorithm that allows it to handle both multi-class classification and multi-class regression problems. It can also analyze generalized feed forward networks. The algorithm was tested on several real-world datasets and was found to generate decision trees with good classification accuracy while also maintaining comprehensibility.
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...cscpconf
油
In this work, the TREPAN algorithm is enhanced and extended for extracting decision trees from neural networks. We empirically evaluated the performance of the algorithm on a set of databases from real world events. This benchmark enhancement was achieved by adapting Single-test TREPAN and C4.5 decision tree induction algorithms to analyze the datasets. The models are then compared with X-TREPAN for comprehensibility and classification accuracy. Furthermore, we validate the experimentations by applying statistical methods. Finally, the modified algorithm is extended to work with multi-class regression problems and the ability to comprehend generalized feed forward networks is achieved.
This document provides an introduction to artificial neural networks. It discusses biological neurons and how artificial neurons are modeled. The key components of a neural network including the network architecture, learning approaches, and the backpropagation algorithm for supervised learning are described. Applications and advantages of neural networks are also mentioned. Neural networks are modeled after the human brain and learn by modifying connection weights between nodes based on examples.
Jiawei Han, Micheline Kamber and Jian Pei
Data Mining: Concepts and Techniques, 3rd ed.
The Morgan Kaufmann Series in Data Management Systems
Morgan Kaufmann Publishers, July 2011. ISBN 978-0123814791
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTShemasubbu08
油
- Artificial neural networks are computational models inspired by the human brain that use algorithms to mimic brain functions. They are made up of simple processing units (neurons) connected in a massively parallel distributed system. Knowledge is acquired through a learning process that adjusts synaptic connection strengths.
- Neural networks can be used for pattern recognition, function approximation, and associative memory in domains like speech recognition, image classification, and financial prediction. They have flexible inputs, resistant to errors, and fast evaluation, though interpretation is difficult.
The document discusses different types of machine learning paradigms including supervised learning, unsupervised learning, and reinforcement learning. It then provides details on artificial neural networks, describing them as consisting of simple processing units that communicate through weighted connections, similar to neurons in the human brain. The document outlines key aspects of artificial neural networks like processing units, connections between units, propagation rules, and learning methods.
Selzy: Simplifying Email Marketing for Maximum GrowthSelzy
油
This presentation is about Selzy, an easy-to-use and affordable email marketing tool that helps businesses create and launch effective email campaigns with minimal effort. It highlights the challenges of traditional email marketing, showcases Selzys AI-powered email builder, fast setup, and 24/7 support, and demonstrates the tools impact through user growth and market potential. With a strong ROI and a rapidly expanding customer base, Selzy positions itself as a powerful yet simple solution for businesses looking to boost engagement and sales through email marketing.
Beyond Metadata for BBC iPlayer: an autoencoder-driven approach for embedding...Simone Spaccarotella
油
I built a machine learning pipeline that generates metadata item embeddings for the calculation of item similarity in the context of Content-To-Content similarity recommendations for BBC iPlayer programmes.
INSIGHTS INTO STABILITY ASPECTS OF HYBRID SYSTEM; AN ENABLING TECHNOLOGY FOR ...hugoshan513
油
The development of a country is dependent on the per person energy consumption rate, which is very low
in Bangladesh. Bangladesh installed a capacity of 10416 MW electricity on June 2014 and three fourth of
which is considered to be accessible. Near about 45% people has no access to electricity. Therefore,
electricity shortage is an acute crisis in Bangladesh. As Saint Martin Island is far away from the main
land, it is almost impossible and cost ineffective to supply electricity from the national grid. For
connection of nearly 6000 peoples of Saint Martin to the main stream of development and to make this
island more attractive to the tourists, it is very essential to provide electricity for them. Power generation
by combining solar, wind and diesel, known as hybrid system can be the most efficient technique for the
electrification of these types of Island. Based on this principle, in this paper a hybrid system is designed
for electrification of Saint Martins Island. In the analysis, realistic data is used for load calculation and
optimization analysis for most effective solution. Hybrid Optimization Model for Electric Renewable
(HOMER) software is used to find out the final optimization and sensitive analysis of hybrid system. This
system satisfies the load demand and reduces carbon emission which will help to generate green energy.
2. Artificial Neural Networks
Computational models inspired by the human
brain:
Algorithms that try to mimic the brain.
Massively parallel, distributed system, made up of
simple processing units (neurons)
Synaptic connection strengths among neurons are
used to store the acquired knowledge.
Knowledge is acquired by the network from its
environment through a learning process
3. History
late-1800's - Neural Networks appear as an
analogy to biological systems
1960's and 70's Simple neural networks appear
Fall out of favor because the perceptron is not
effective by itself, and there were no good algorithms
for multilayer nets
1986 Backpropagation algorithm appears
Neural Networks have a resurgence in popularity
More computationally expensive
4. Applications of ANNs
ANNs have been widely used in various domains
for:
Pattern recognition
Function approximation
Associative memory
5. Properties
Inputs are flexible
any real values
Highly correlated or independent
Target function may be discrete-valued, real-valued, or
vectors of discrete or real values
Outputs are real numbers between 0 and 1
Resistant to errors in the training data
Long training time
Fast evaluation
The function produced can be difficult for humans to
interpret
6. When to consider neural networks
Input is high-dimensional discrete or raw-valued
Output is discrete or real-valued
Output is a vector of values
Possibly noisy data
Form of target function is unknown
Human readability of the result is not important
Examples:
Speech phoneme recognition
Image classification
Financial prediction
7. February 10, 2025
Data Mining: Concepts and
Techniques 7
A Neuron (= a perceptron)
The n-dimensional input vector x is mapped into variable y by
means of the scalar product and a nonlinear function mapping
t
-
f
weighted
sum
Input
vector x
output y
Activation
function
weight
vector w
w0
w1
wn
x0
x1
xn
)
sign(
y
e
For Exampl
n
0
i
t
x
w i
i
8. Perceptron
Basic unit in a neural network
Linear separator
Parts
N inputs, x1 ... xn
Weights for each input, w1 ... wn
A bias input x0 (constant) and associated weight w0
Weighted sum of inputs, y = w0x0 + w1x1 + ... + wnxn
A threshold function or activation function,
i.e 1 if y > t, -1 if y <= t
9. Artificial Neural Networks (ANN)
Model is an assembly of
inter-connected nodes
and weighted links
Output node sums up
each of its input value
according to the weights
of its links
Compare output node
against some threshold t
X1
X2
X3
Y
Black box
w1
t
Output
node
Input
nodes
w2
w3
)
( t
x
w
I
Y
i
i
i
Perceptron Model
)
( t
x
w
sign
Y
i
i
i
or
10. Types of connectivity
Feedforward networks
These compute a series of
transformations
Typically, the first layer is the
input and the last layer is the
output.
Recurrent networks
These have directed cycles in their
connection graph. They can have
complicated dynamics.
More biologically realistic.
hidden units
output units
input units
11. Different Network Topologies
Single layer feed-forward networks
Input layer projecting into the output layer
Input Output
layer layer
Single layer
network
12. Different Network Topologies
Multi-layer feed-forward networks
One or more hidden layers. Input projects only
from previous layers onto a layer.
Input Hidden Output
layer layer layer
2-layer or
1-hidden layer
fully connected
network
14. Different Network Topologies
Recurrent networks
A network with feedback, where some of its
inputs are connected to some of its outputs (discrete
time).
Input Output
layer layer
Recurrent
network
15. Algorithm for learning ANN
Initialize the weights (w0, w1, , wk)
Adjust the weights in such a way that the output
of ANN is consistent with class labels of training
examples
Error function:
Find the weights wis that minimize the above error
function
e.g., gradient descent, backpropagation algorithm
2
)
,
(
i
i
i
i X
w
f
Y
E
16. Optimizing concave/convex function
Maximum of a concave function = minimum of a
convex function
Gradient ascent (concave) / Gradient descent (convex)
Gradient ascent rule
24. Decision surface of a perceptron
Decision surface is a hyperplane
Can capture linearly separable classes
Non-linearly separable
Use a network of them
27. Multi-layer Networks
Linear units inappropriate
No more expressive than a single layer
Introduce non-linearity
Threshold not differentiable
Use sigmoid function
31. February 10, 2025
Data Mining: Concepts and
Techniques 31
Backpropagation
Iteratively process a set of training tuples & compare the network's
prediction with the actual known target value
For each training tuple, the weights are modified to minimize the mean
squared error between the network's prediction and the actual target
value
Modifications are made in the backwards direction: from the output
layer, through each hidden layer down to the first hidden layer, hence
backpropagation
Steps
Initialize weights (to small random #s) and biases in the network
Propagate the inputs forward (by applying activation function)
Backpropagate the error (by updating weights and biases)
Terminating condition (when error is very small, etc.)
33. February 10, 2025
Data Mining: Concepts and
Techniques 33
How A Multi-Layer Neural Network Works?
The inputs to the network correspond to the attributes measured for
each training tuple
Inputs are fed simultaneously into the units making up the input layer
They are then weighted and fed simultaneously to a hidden layer
The number of hidden layers is arbitrary, although usually only one
The weighted outputs of the last hidden layer are input to units making
up the output layer, which emits the network's prediction
The network is feed-forward in that none of the weights cycles back to
an input unit or to an output unit of a previous layer
From a statistical point of view, networks perform nonlinear regression:
Given enough hidden units and enough training samples, they can
closely approximate any function
34. February 10, 2025
Data Mining: Concepts and
Techniques 34
Defining a Network Topology
First decide the network topology: # of units in the input
layer, # of hidden layers (if > 1), # of units in each hidden
layer, and # of units in the output layer
Normalizing the input values for each attribute measured in
the training tuples to [0.01.0]
One input unit per domain value, each initialized to 0
Output, if for classification and more than two classes, one
output unit per class is used
Once a network has been trained and its accuracy is
unacceptable, repeat the training process with a different
network topology or a different set of initial weights
35. February 10, 2025
Data Mining: Concepts and
Techniques 35
Backpropagation and Interpretability
Efficiency of backpropagation: Each epoch (one interation through the
training set) takes O(|D| * w), with |D| tuples and w weights, but # of
epochs can be exponential to n, the number of inputs, in the worst case
Rule extraction from networks: network pruning
Simplify the network structure by removing weighted links that have the
least effect on the trained network
Then perform link, unit, or activation value clustering
The set of input and activation values are studied to derive rules
describing the relationship between the input and hidden unit layers
Sensitivity analysis: assess the impact that a given input variable has on a
network output. The knowledge gained from this analysis can be
represented in rules
36. February 10, 2025
Data Mining: Concepts and
Techniques 36
Neural Network as a Classifier
Weakness
Long training time
Require a number of parameters typically best determined empirically,
e.g., the network topology or structure.
Poor interpretability: Difficult to interpret the symbolic meaning behind
the learned weights and of hidden units in the network
Strength
High tolerance to noisy data
Ability to classify untrained patterns
Well-suited for continuous-valued inputs and outputs
Successful on a wide array of real-world data
Algorithms are inherently parallel
Techniques have recently been developed for the extraction of rules
from trained neural networks
38. Artificial Neural Networks (ANN)
X1 X2 X3 Y
1 0 0 0
1 0 1 1
1 1 0 1
1 1 1 1
0 0 1 0
0 1 0 0
0 1 1 1
0 0 0 0
X1
X2
X3
Y
Black box
0.3
0.3
0.3 t=0.4
Output
node
Input
nodes
otherwise
0
true
is
if
1
)
(
where
)
0
4
.
0
3
.
0
3
.
0
3
.
0
( 3
2
1
z
z
I
X
X
X
I
Y
40. February 10, 2025
Data Mining: Concepts and
Techniques 40
A Multi-Layer Feed-Forward Neural Network
Output layer
Input layer
Hidden layer
Output vector
Input vector: X
wij
ij
k
i
i
k
j
k
j x
y
y
w
w )
( )
(
)
(
)
1
(
41. General Structure of ANN
Activation
function
g(Si
)
Si
Oi
I1
I2
I3
wi1
wi2
wi3
Oi
Neuron i
Input Output
threshold, t
Input
Layer
Hidden
Layer
Output
Layer
x1
x2
x3
x4
x5
y
Training ANN means learning
the weights of the neurons