Neural Networks: Study Notes
General Science
July 28, 2025
4 min read
1. Definition
Neural Networks (NNs) are computational models inspired by biological brains. They consist of interconnected nodes (“neurons”) that process information using weighted connections. NNs are used for tasks such as image recognition, natural language processing, and predictive analytics.
2. Historical Context
- 1943: Warren McCulloch & Walter Pitts propose the first mathematical model of a neuron.
- 1958: Frank Rosenblatt invents the Perceptron, an early learning algorithm.
- 1986: Rumelhart, Hinton, and Williams introduce backpropagation, enabling multilayer NNs.
- 2012: AlexNet revolutionizes image classification, sparking modern deep learning.
- 2020s: Neural networks underpin advances in AI, such as GPT-3 and AlphaFold.
3. Structure of Neural Networks
Layers
- Input Layer: Receives raw data.
- Hidden Layers: Intermediate processing units.
- Output Layer: Produces final predictions.
Diagram

4. How Neural Networks Learn
- Forward Pass: Data flows from input to output.
- Loss Calculation: Measures prediction error.
- Backward Pass (Backpropagation): Adjusts weights to minimize error.
- Optimization: Uses algorithms (e.g., SGD, Adam) to update weights.
5. Types of Neural Networks
Type |
Description |
Typical Application |
Feedforward NN |
Data moves in one direction |
Classification |
Convolutional NN (CNN) |
Uses filters for spatial data |
Image recognition |
Recurrent NN (RNN) |
Loops for sequential data |
Language modeling |
Generative Adversarial NN |
Competing networks generate data |
Image synthesis |
Transformer |
Attention mechanism for sequence modeling |
Translation, chatbots |
6. Key Concepts
- Activation Function: Determines neuron output (e.g., ReLU, sigmoid).
- Weights: Numeric values that modulate connections.
- Bias: Shifts activation threshold.
- Epoch: One pass through the training data.
7. Example: Simple Neural Network
Python Code
# Python
import numpy as np
def sigmoid(x):
return 1 / (1 + np.exp(-x))
inputs = np.array([0.5, 0.3])
weights = np.array([0.4, 0.7])
bias = 0.1
output = sigmoid(np.dot(inputs, weights) + bias)
print(output)
8. Data Table: Neural Network Performance
Year |
Model |
Dataset |
Accuracy (%) |
Parameters (Millions) |
2012 |
AlexNet |
ImageNet |
84.7 |
60 |
2015 |
ResNet |
ImageNet |
96.4 |
25.6 |
2018 |
BERT |
GLUE |
82.1 |
110 |
2020 |
GPT-3 |
Various |
N/A |
175,000 |
2021 |
AlphaFold2 |
CASP14 |
92.4 |
~21 |
9. Common Misconceptions
- NNs mimic human intelligence: NNs are inspired by brains but lack consciousness or understanding.
- More layers always improve performance: Overly deep networks can overfit or become inefficient.
- NNs require huge data: Small networks can learn from limited data, but deep learning benefits from large datasets.
- NNs are “black boxes”: While complex, tools exist for interpreting their decisions (e.g., SHAP, LIME).
10. Surprising Facts
- Neural networks can outperform humans: In specific tasks, such as protein folding (AlphaFold2), NNs have surpassed expert performance.
- NNs can generate realistic synthetic data: GANs produce images and audio indistinguishable from real samples.
- NNs are used in astronomy: They help detect exoplanets and analyze cosmic phenomena.
11. Recent Research
- Citation: Jumper et al. (2021), “Highly accurate protein structure prediction with AlphaFold,” Nature, 596, 583–589. Link
- AlphaFold2, a neural network, predicted protein structures with unprecedented accuracy, revolutionizing computational biology.
12. Applications
- Healthcare: Disease prediction, drug discovery.
- Finance: Fraud detection, algorithmic trading.
- Robotics: Autonomous navigation.
- Language: Translation, sentiment analysis.
- Science: Protein folding, climate modeling.
13. The Great Barrier Reef
- The largest living structure on Earth.
- Visible from space.
- Example of complex patterns in nature, sometimes analyzed using neural networks for ecological monitoring.
14. Further Reading
- LeCun, Y., Bengio, Y., & Hinton, G. (2015). “Deep learning.” Nature, 521, 436–444.
- Brown, T.B. et al. (2020). “Language Models are Few-Shot Learners.” arXiv preprint.
15. Summary Table: NN Advantages & Limitations
Aspect |
Advantage |
Limitation |
Flexibility |
Adapts to many data types |
Needs tuning |
Performance |
High accuracy |
Computationally intensive |
Interpretability |
Tools exist |
Still challenging |
Scalability |
Handles large datasets |
May require special hardware |
16. Visualizing NN Training

17. Key Takeaways
- Neural networks are powerful, adaptable, and widely used in science and industry.
- They do not “think” like humans but can solve complex problems.
- Ongoing research continues to expand their capabilities and applications.