unit-3_ml[1]
unit-3_ml[1]
Neural Networks:
A neural network consists of connected units or nodes called artificial neurons, which loosely model
the neurons in the brain. Artificial neuron models that mimic biological neurons more closely have also
been recently investigated and shown to significantly improve performance. These are connected
by edges, which model the synapses in the brain. Each artificial neuron receives signals from connected
neurons, then processes them and sends a signal to other connected neurons. The "signal" is a real
number, and the output of each neuron is computed by some non-linear function of the sum of its inputs,
called the activation function. The strength of the signal at each connection is determined by a weight,
which adjusts during the learning process.
Typically, neurons are aggregated into layers. Different layers may perform different transformations on
their inputs. Signals travel from the first layer (the input layer) to the last layer (the output layer), possibly
passing through multiple intermediate layers (hidden layers). A network is typically called a deep neural
network if it has at least two hidden layers.
Artificial neural networks are used for various tasks, including predictive modeling, adaptive control, and
solving problems in artificial intelligence. They can learn from experience, and can derive conclusions from
a complex and seemingly unrelated set of information.
In summary, while both biological neurons and artificial neurons share the foundational concept of
transmitting and processing signals to achieve specific outcomes, they differ in terms of their physical
structure, signalling mechanisms, learning processes, and overall computational capabilities.
The process of training a neural network involves providing it with a labeled dataset, known as training
data, and adjusting the weights of the connections iteratively to minimize the difference between the
predicted outputs and the actual target outputs. This optimization process is typically achieved using
gradient descent algorithms.
Think of each individual node as its own linear regression model, composed of input data, weights, a bias
(or threshold), and an output. The formula would look something like this:
∑wixi + bias = w1x1 + w2x2 + w3x3 + bias
Output = f(x) = 1,
if ∑w1x1 + b >= 0
0 otherwise // i.e., if ∑w1x1 + b < 0
Perceptron:
A perceptron is one of the simplest types of artificial neural networks. It's a mathematical model that's
used for binary classification tasks, where the goal is to separate input data into two classes. The
perceptron takes a set of inputs, each multiplied by corresponding weights, and then applies an activation
function to produce an output. This output is used to classify the input into one of the two classes. The
perceptron is very useful for classifying data sets that are linearly separable.