Skip to content

Programming-Ocean-Academy/deep-learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Repository files navigation

Deep Learning Collection

Welcome to the Programming Ocean Academy's Deep Learning Repository.
This project is a comprehensive educational suite showcasing a variety of generative models implemented with PyTorch, ranging from foundational architectures to modern, cutting-edge designs.


Objective

This repository serves as an academic and teaching-oriented resource for understanding, building, and visualizing deep generative models. It is designed to help students, researchers, and practitioners explore the diversity of generative learning approaches in a modular and clear format.


Repository Structure

Each folder represents a specific category of generative or neural architecture:

Folder NameDescription
auto-regressive-modelsPixelCNN and related sequential density estimators
cnnBasic CNN models for image recognition
diffusionDenoising Diffusion Probabilistic Models (DDPM, DDIM)
dit-modelsDiffusion Transformers (DiT)
energy-based-modelsEBMs trained with Langevin dynamics
flow-based-modelsRealNVP, Glow, and other invertible models
gansGAN, DCGAN, WGAN, and conditional variants
latent-manifold-auto-encoderLatent space exploration with VAEs and AEs
multi-modelCross-modal tasks (e.g., text-to-image, image captioning)
restricted-boltzmann-machineContrastive Divergence and RBMs
rnnRecurrent networks (LSTM, GRU)
score-based-generative-convolutionScore-matching models with CNN backbones
score-based-generative-modelsLangevin and NCSN-style samplers
time-seriesForecasting models for temporal data
transformerSequence models and transformers (Vanilla, GPT)
variational-auto-encoderVAEs and conditional variants
vision-transformerViT for image understanding

Highlighted Projects

1. Diffusion Models

"A Concise Implementation of Denoising Diffusion Probabilistic Models for Generative Image Synthesis in PyTorch"

  • U-Net architecture with Gaussian noise scheduling
  • Reverse sampling with denoising

2. GANs

"Adversarial Image Synthesis with Generative Networks: A PyTorch Implementation of GANs on MNIST"

  • Generator and Discriminator loop
  • Real versus generated image comparison

3. VAEs

"Latent Variable Modeling and Image Generation with Variational Autoencoders: A PyTorch-Based Study on MNIST"

  • Reparameterization trick
  • Sampling and interpolation

4. Score-Based Models

"Unsupervised Image Synthesis via Score Matching and Langevin Dynamics: A Score-Based Generative Framework on MNIST"

  • Trainable score networks
  • MCMC sampling

5. Text-to-Image (Mini DALL·E)

"Learning Discrete Visual Representations from Textual Descriptions: A Simplified VQ-VAE Framework for Text-to-Image Generation"

  • VQ-VAE with Transformer
  • Captioned image generation (color, shape, objects)

6. Image Captioning

"Visual Grounding through Language: A Minimalist Encoder-Decoder Framework for Image Captioning with Attention in PyTorch"

  • ResNet encoder combined with LSTM and soft attention
  • Caption generation for synthetic scenes

Usage

All notebooks are written for clarity and modularity.

# Clone the repository https://github.com/Programming-Ocean-Academy/deep-learning.git Open any `.ipynb` file in JupyterLab, Google Colab, or VSCode and run directly. --- ## Contributing We welcome contributions to extend this educational repository: - Add new generative model examples - Improve visualizations or metrics - Refactor notebooks into scripts or modules --- ## License MIT License. Free for personal, educational, and research use. --- ## Acknowledgements Inspired by work from: - OpenAI, DeepMind, LucidRain - The PyTorch community and tutorials - Original research papers on DALL·E, VQ-VAE, and DDPM --- Enjoy exploring generative deep learning.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published