Skip to the content.

Basics of Neural Networks

Perceptron

A simple type of artificial neuron that outputs either 0 or 1 based on a linear combination of inputs and a step function. It is the oldest model of a neural network.


Neuron (Artificial Neuron)

A mathematical unit that receives inputs, applies weights and bias, passes the result through an activation function, and produces an output.


Weight

A numeric value assigned to each input, representing its importance in the prediction.


Input Layer

The raw data or features provided to the neural network.


Output Layer

The final prediction made by the neural network after processing through all layers.


Hidden Layer

The layers between the input and output layers where the actual learning and pattern extraction happens. They are “hidden” because we do not directly see their outputs.


Multi-Layer Perceptron (MLP)

A neural network with multiple layers of neurons, including one or more hidden layers, enabling it to learn complex patterns.


Activation Function

A function applied to the neuron’s output that introduces non-linearity, allowing the network to learn complex relationships.


Loss Function

A function that measures the difference between the predicted output and the actual output. It guides how much the model’s prediction is wrong.


Gradient Descent

An optimization algorithm that updates the weights by minimizing the loss function using calculated gradients.


Optimizer

An algorithm that adjusts weights and biases during training (e.g., SGD, Adam), using gradient descent or its variations.


Epoch

One complete pass over the entire training dataset.


Batch

A small subset of the dataset used to train the model at one time.


Iteration

One update of the model’s parameters;

\[\Large \text{Iterations per Epoch} = \frac{\text{Dataset Size}}{\text{Batch Size}}\]

Overfitting

When the model learns the training data too well (including noise), but performs poorly on new, unseen data.


Underfitting

When the model is too simple to capture the patterns in the training data, leading to poor performance on both training and unseen data.


Summary of Basics of Neural Networks

Topic Description Importance
Perceptron Simplest neural network structure Foundation for understanding neural networks
Multilayer Perceptron Multi-layer networks capable of complex modeling Essential for real-world neural applications
Activation Functions Non-linear transformations enabling complex patterns Critical for neural network flexibility
Loss Functions Measure model accuracy and guide training Fundamental for model learning
Training Neural Networks Methods to adjust weights effectively Key to optimizing model performance
Overfitting & Underfitting Challenges affecting model generalization Necessary understanding for effective models
Validation and Testing Evaluating true model performance Ensures robust and reliable results

Summary

Understanding of these core concepts and techniques provides a strong foundation in neural networks, enabling further exploration of advanced deep learning topics and applications.