DeepLearning For Finance
  • Back to Main Website
  • Home
  • Introduction to Deep Learning
    • Introduction to Deep Learning
    • From Traditional Models to Deep Learning
    • The Multi-Layer Perceptron (MLP)
    • Automatic Differentiation: The Engine of Deep Learning
    • Computation Backends & Keras 3
    • GPUs and Deep Learning: When Hardware Matters
    • Keras Fundamentals: Models & Layers
    • Keras Matrix Operations: The Building Blocks
    • Activation Functions: Adding Non-linearity
    • Model Training Fundamentals

    • Travaux Pratiques
    • TP1 Corrected: Building Neural Networks - From Simple to Custom Implementations
    • TP1 Corrected: Building Neural Networks - From Simple to Custom Implementations
  • Recurrent Neural Networks
    • Recurrent Neural Networks
    • Sequential Data Processing: From MLPs to RNNs
    • Long Short-Term Memory Networks (LSTM)
    • Modern RNN Architectures
    • RNN Limitations: Computational Challenges

    • Travaux Pratiques
    • TP: Recurrent Neural Networks for Time Series Prediction
    • TP Corrected: Recurrent Neural Networks for Time Series Prediction
  • Training a Neural Network
    • Training a Neural Network
    • Understanding the Training Loop
    • Understanding Optimizers
    • Understanding Callbacks
    • Training Parameters and Practical Considerations

    • Travaux Pratiques
    • TP: Using Deep Learning Frameworks for General Optimization
    • tp_general_optimization_corrected.html
    • TP: Impact of Callbacks on Training
  • Essential Building Blocks of Modern Neural Networks
    • Essential Building Blocks of Modern Neural Networks
    • Residual Connections and Gating Mechanisms
    • Convolutional Layers: From Images to Time Series
    • Neural Network Embeddings: Learning Meaningful Representations
    • Attention Mechanisms: Learning What to Focus On
    • Encoder-Decoder Architectures

    • Travaux Pratiques
    • Practical Assignment: Building a Transformer-Based Architecture for Time Series Forecasting
    • Practical Assignment: Building a Transformer-Based Architecture for Time Series Forecasting
  • Projets
    • Projets
  • Code source

Essential Building Blocks of Modern Neural Networks

Remi Genet

2025-04-03

Les cours de cette partie sont:

 
Attention Mechanisms: Learning What to Focus On
Understanding attention mechanisms in neural networks: from basic self-attention to multi-head architectures.
Remi Genet
2025-04-03
 
Convolutional Layers: From Images to Time Series
Understanding convolution operations in neural networks and their applications beyond computer vision.
Remi Genet
2025-04-03

Encoder-Decoder Architectures
Understanding encoder-decoder architectures: principles, mathematics, and applications in sequence-to-sequence tasks.
Remi Genet
2025-04-03
 
Neural Network Embeddings: Learning Meaningful Representations
Understanding embeddings in neural networks: from discrete entities to continuous vector spaces.
Remi Genet
2025-04-03

Residual Connections and Gating Mechanisms
Understanding fundamental building blocks that enable training of deep neural networks: residual connections and gating mechanisms.
Remi Genet
2025-04-03
No matching items
Back to top
Residual Connections and Gating Mechanisms

Deep Learning For Finance, Rémi Genet.
Licence
Code source disponible sur Github

 

Site construit avec et Quarto
Inspiration pour la mise en forme du site ici
Code source disponible sur GitHub