Techniques

Neural Network

A computing system of interconnected nodes inspired by biological brains, trained to recognise patterns.

Definition

An artificial neural network consists of layers of interconnected nodes (neurons) that process information using weighted connections. Each neuron applies a non-linear activation function (ReLU, sigmoid) to a weighted sum of its inputs, enabling the network to model complex non-linear relationships.

Networks are organised in layers: an input layer, one or more hidden layers, and an output layer. Deeper networks (more hidden layers) can represent increasingly complex functions. Training adjusts connection weights using backpropagation and gradient descent to minimise prediction error on training data.

Key architectures include feedforward networks (fully connected), convolutional networks (CNNs for images), recurrent networks (RNNs/LSTMs for sequences), and transformers (attention-based, dominating language and vision).

Examples

  • Image classifier
  • Speech recogniser
  • GPT-4 internal architecture

Want a deeper dive?

Read our full explainer with use cases, how-it-works, and FAQs.

Neural Networks concept guide