Deep Learning (with PyTorch)

Overview

Deep Learning (with PyTorch) Binder

This notebook repository now has a companion website, where all the course material can be found in video and textual format.

🇬🇧   🇨🇳   🇰🇷   🇪🇸   🇮🇹   🇹🇷   🇯🇵   🇸🇦   🇫🇷   🇮🇷   🇷🇺   🇻🇳   🇷🇸   🇵🇹   🇭🇺

Getting started

To be able to follow the exercises, you are going to need a laptop with Miniconda (a minimal version of Anaconda) and several Python packages installed. The following instruction would work as is for Mac or Ubuntu Linux users, Windows users would need to install and work in the Git BASH terminal.

Download and install Miniconda

Please go to the Anaconda website. Download and install the latest Miniconda version for Python 3.7 for your operating system.

wget <http:// link to miniconda>
sh <miniconda*.sh>

Check-out the git repository with the exercise

Once Miniconda is ready, checkout the course repository and proceed with setting up the environment:

git clone https://github.com/Atcold/pytorch-Deep-Learning

Create isolated Miniconda environment

Change directory (cd) into the course folder, then type:

# cd pytorch-Deep-Learning
conda env create -f environment.yml
source activate pDL

Start Jupyter Notebook or JupyterLab

Start from terminal as usual:

jupyter lab

Or, for the classic interface:

jupyter notebook

Notebooks visualisation

Jupyter Notebooks are used throughout these lectures for interactive data exploration and visualisation.

We use dark styles for both GitHub and Jupyter Notebook. You should try to do the same, or they will look ugly. JupyterLab has a built-in selectable dark theme, so you only need to install something if you want to use the classic notebook interface. To see the content appropriately in the classic interface install the following:

Comments
  • Chapter 5-2 docs

    Chapter 5-2 docs

    Optimization techniques II

    We discuss adaptive methods for SGD such as RMSprop and ADAM. We also talk about normalization layers and their effects on the neural network training process. Finally, we discuss a real-world example of neural nets being used in industry to make MRI scans faster and more efficient.

    Please let me know if any changes need to be made before merging.

    opened by guidopetri 16
  • Updates to current packages

    Updates to current packages

    This:

    • Moves PyTorch from 0.4 to 1.1 (one tiny code change)
    • Moves Python from 3.6 to 3.7 (no changes to code, just env)
    • Moves 1-2 requirements out of notebooks and into environment (potential nasty scipy pip install from librosa avoided!)
    • Uses conda kernels so the correct environment kernel is available (all notebooks rerun to pick up proper kernel)
    • Adds JuptyerLab (not required, but nice) - the interactive backend in the final notebook is still best in the classic interface. Try out built-in dark mode!

    All notebooks seem to run (except noted minor issue with JupyterLab)

    opened by henryiii 16
  • [FR & EN] YouTube subtitles

    [FR & EN] YouTube subtitles

    Hi Alf :wave:,

    As indicated in my last email, I can't afford to wait for Yann's return without a big delay on my side. So here are the subtitle files:

    • For English, it is the addition of the unicode. In practice:
    1. The list of files not modified during this review of the unicode: practinum1 (didn't need unicode), practinum4 (the file contains blocks of 3 instead of 2 for the others), for lecture 12 (the only file I didn't translate into French)

    2. The list of finished files (full English review + unicode) : lecture 6 & 9

    3. The list of about clean files (partial English review + unicode) : lecture 1-3,10,11 + practinum 1-3, 7-8, 10

    4. The list of not clean files (no English review + unicode): lecture 5-9,12-15 + practinum 5-6,9,11-15

    • For French, these are all the subtitles (except for lecture 12 where I have huge problems understanding Mike Lewis's accent and so I preferred not to put anything than to translate badly).

    I also added a disclamer for the V2 of the French translation of the website which should arrive this month. It should be my next and last PR closing the French translation work :boom:

    Loïck

    opened by lbourdois 13
  • Broken image links in 3.3. Properties of natural signals

    Broken image links in 3.3. Properties of natural signals

    The following image links are broken:

    • [x] Figure 2(a)
    • [x] Figure 2(b)
    • [x] Figure 3(a)
    • [x] Figure 3(b)

    See https://atcold.github.io/pytorch-Deep-Learning/en/week03/03-3/

    I think the images were originally obtained from this presentation: 02 - CNN.pdf

    See pages 10-11


    Also, small suggestions:

    • [x] Change Figure 4 to include R^7 and R^2 as in Slide 20 . This would better match the text for Figure 4.

    • [x] Include Figure (4b maybe?) with that on Slide 21 to show what Padding is doing

    opened by feedthebeat90 11
  • Portuguese translation

    Portuguese translation

    Hi @Atcold ! I would like to know how and where should I commit markdown files in Portuguese? I recall that you have commented something with @ebetica .

    opened by ricardobarroslourenco 11
  • [ZH] 13-3 Inline latex broken

    [ZH] 13-3 Inline latex broken

    Hi @JonathanSum ! Just for your info, There seems to be some inline latex broken on lecture 13-3:

    Screen Shot 2020-09-23 at 22 41 44

    The rest of the lectures I've checked seem to be fine.

    opened by xcastilla 9
  • Reorganize the website structure

    Reorganize the website structure

    This PR reorganizes the website structure, so we now have:

    en/
      index.md
      about.dm
      week01/
      week02/
      ...
    zh/
      index.md
      about.md
      week01/
      week02/
      ...
    ...
    

    Hopefully it's less messy and easier to work with.

    After this is merged, I will pull the images out into a global directory as well.

    Also fixes some broken links in zh/index.md

    opened by ebetica 9
  • Problem visualizing spanish translation on github.io

    Problem visualizing spanish translation on github.io

    I found an error visualizing on the github.io page the file /docs/es/week02/02-1.md.

    The english version of the file appears before some parts and the layout of the spanish parts after the english parts gets a bit messed up.

    grafik

    grafik

    opened by mt0rm0 8
  • [EN] Fix timers

    [EN] Fix timers

    A PR that fixes the timers of the sbv files that I couldn't correct in PR #660 to avoid conflicts.

    I also took the opportunity to correct the few errors I caught when translating the lecture10.

    I also noticed that the sbv files of the practinums of weeks 14 and 15 were missing.

    opened by lbourdois 8
  • [ZH] translation of 06-2 and 06.md

    [ZH] translation of 06-2 and 06.md

    I have translated the top 50% of the RNN(06-02) in Chinese.

    I passed the course on deep learning.ai and I also wrote a few notebooks to help students in the coursera Tensorflow time series seq2seq notebook.

    opened by JonathanSum 8
  • Vanishing gradient notebook

    Vanishing gradient notebook

    Poornima and I have compared an LSTM and RNN and visualized the gradients with respect to the input. We see that the gradients for the RNN are much smaller compared to the LSTM.

    We are able to train MNIST for a large input sequence with an LSTM and failed to do so with an RNN.

    Hope this is useful. If we need to make any chances, please let us know !

    opened by karanchahal 8
  • Software version update for 2023

    Software version update for 2023

    Hi there,

    I hope these tips can help you: Using Docker, with torchtext version 0.9.0 and PyTorch version 1.8.0.

    Please note that PyTorch 1.8 may not have good support for CUDA versions newer than 11. If you are using a newer version of CUDA, you may want to consider using the CPU instead.

    opened by wenxin-bupt 0
  • <Fix> evaluation dataset, printed samples

    evaluation dataset, printed samples

    Bunch of minor "theoretical" changes in the evaluation function:

    1. test_data_gen was used as the data generator in the evaluation, instead of data_generator, thereby evaluating the net on the test set used for training (not an actual issue here given the sequences are randomized and not sampled from existing datasets, but in principle would lead to a data leak in realistic scenarios);
    2. the correct sequences printed were a sampling (with reinsertion) of the first 10 evaluated, instead of 10 sampled from the whole set of correct ones;
    3. the condition for printing the incorrectly classified sequences would declare the absence of misclassifications if verbose==False, independently of their actual presence;
    opened by hypothe 1
  • Added controller trainer and improved truck class

    Added controller trainer and improved truck class

    Added

    • new truck methods for randomizing state within contraints
    • new truck methods for seeing if truck is at dock or offscreen
    • Training script for optimizing controller

    Note

    I currently have not successfully trained the controller to convergence. I have based the training off of this. On the website, they mention that the controller is hard to train. I have tried training it on the website with no success, so it seems like even their lessons are difficult to train. However, the code for training should be very similar to the code on the website. You may also alter the amount of lessons, max time steps, learning rate etc. to see if the model converges. I have been trying for over a week and have not succeeded yet.

    opened by dafaronbi 1
  • fix chinese version of 12-3

    fix chinese version of 12-3

    I found that the Chinese version was basically machine translated, which caused the latex syntax to be broken. Of course, there are a lot of unreasonable translation. This PR is mainly about fixing broken latex. I also did my best to fix some of the translations that were too much bullshit.

    opened by vipcxj 0
  • Russian translation (dictionary)

    Russian translation (dictionary)

    I would question some translations in the dictionary for Russian: I've graduated this year and we haven't really translated everything. For example, it will be more understandable if I say "one-hot" in Russian as it is, rather than "унитарный код". Basically, I've never heard anyone calling it "унитарный код", to be honest...

    So I guess there is a choice between being academically strict or being understood.

    opened by xufana 4
  • Use conda instead of source activate

    Use conda instead of source activate

    I think source activate is a few years old now and isn't supported anymore. https://stackoverflow.com/questions/49600611/python-anaconda-should-i-use-conda-activate-or-source-activate-in-linux

    opened by ebetica 0
Releases(dlsp19)
  • dlsp19(Jan 30, 2020)

    This is the notes for the Spring 2019 Deep Learning course at NYU. This course concerns the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional net and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.

    This is the initial draft of the course notes - they are based off of a course developed for the the African Masters of Machine Intelligence (AMMI). You can access that version here

    Source code(tar.gz)
    Source code(zip)
  • aims-fl18(Jan 30, 2020)

    The African Masters of Machine Intelligence (AMMI) is Africa's flagship program in machine intelligence led by The African Institute for Mathematical Sciences (AIMS). These lessons, developed during the course of several years while I've been teaching at Purdue and NYU, are here proposed for the AMMI (AIMS).

    Prior to this course delivered for AMMI (AIMS), an earlier version of this was delivered and video-recorded for the Computational and Data Science for High Energy Physics (CoDaS-HEP) summer school at Princeton University. Please refer to this version release here.

    Source code(tar.gz)
    Source code(zip)
  • v1.0.0(Nov 5, 2018)

    Click CoDaS-HEP_2018 to jump to this release.

    These lessons, developed during the course of several years while I've been teaching at Purdue and NYU, are here proposed for the Computational and Data Science for High Energy Physics (CoDaS-HEP) summer school at Princeton University. The whole course has been recorded and the playlist is made available here. Check the slides for drawings of better visual quality.

    Source code(tar.gz)
    Source code(zip)
Owner
Alfredo Canziani
Musician, math lover, cook, dancer, 🏳️‍🌈, and assistant professor of Computer Science at New York University
Alfredo Canziani
Lightweight plotting to the terminal. 4x resolution via Unicode.

Uniplot Lightweight plotting to the terminal. 4x resolution via Unicode. When working with production data science code it can be handy to have plotti

Olav Stetter 203 Dec 29, 2022
SoK: Vehicle Orientation Representations for Deep Rotation Estimation

SoK: Vehicle Orientation Representations for Deep Rotation Estimation Raymond H. Tu, Siyuan Peng, Valdimir Leung, Richard Gao, Jerry Lan This is the o

FIRE Capital One Machine Learning of the University of Maryland 12 Oct 07, 2022
SOTA easy to use PyTorch-based DL training library

Easily train or fine-tune SOTA computer vision models from one training repository. SuperGradients Introduction Welcome to SuperGradients, a free open

619 Jan 03, 2023
Implementation of popular SOTA self-supervised learning algorithms as Fastai Callbacks.

Self Supervised Learning with Fastai Implementation of popular SOTA self-supervised learning algorithms as Fastai Callbacks. Install pip install self-

Kerem Turgutlu 276 Dec 23, 2022
PyTorch implementation of DeepUME: Learning the Universal Manifold Embedding for Robust Point Cloud Registration (BMVC 2021)

DeepUME: Learning the Universal Manifold Embedding for Robust Point Cloud Registration [video] [paper] [supplementary] [data] [thesis] Introduction De

Natalie Lang 10 Dec 14, 2022
Companion repository to the paper accepted at the 4th ACM SIGSPATIAL International Workshop on Advances in Resilient and Intelligent Cities

Transfer learning approach to bicycle sharing systems station location planning using OpenStreetMap Companion repository to the paper accepted at the

Politechnika Wrocławska - repozytorium dla informatyków 4 Oct 24, 2022
[CVPR'2020] DeepDeform: Learning Non-rigid RGB-D Reconstruction with Semi-supervised Data

DeepDeform (CVPR'2020) DeepDeform is an RGB-D video dataset containing over 390,000 RGB-D frames in 400 videos, with 5,533 optical and scene flow imag

Aljaz Bozic 165 Jan 09, 2023
[AAAI-2022] Official implementations of MCL: Mutual Contrastive Learning for Visual Representation Learning

Mutual Contrastive Learning for Visual Representation Learning This project provides source code for our Mutual Contrastive Learning for Visual Repres

winycg 48 Jan 02, 2023
A Topic Modeling toolbox

Topik A Topic Modeling toolbox. Introduction The aim of topik is to provide a full suite and high-level interface for anyone interested in applying to

Anaconda, Inc. (formerly Continuum Analytics, Inc.) 93 Dec 01, 2022
Weakly Supervised End-to-End Learning (NeurIPS 2021)

WeaSEL: Weakly Supervised End-to-end Learning This is a PyTorch-Lightning-based framework, based on our End-to-End Weak Supervision paper (NeurIPS 202

Auton Lab, Carnegie Mellon University 131 Jan 06, 2023
Code base for NeurIPS 2021 publication titled Kernel Functional Optimisation (KFO)

KernelFunctionalOptimisation Code base for NeurIPS 2021 publication titled Kernel Functional Optimisation (KFO) We have conducted all our experiments

2 Jun 29, 2022
docTR by Mindee (Document Text Recognition) - a seamless, high-performing & accessible library for OCR-related tasks powered by Deep Learning.

docTR by Mindee (Document Text Recognition) - a seamless, high-performing & accessible library for OCR-related tasks powered by Deep Learning.

Mindee 1.5k Jan 01, 2023
CATE: Computation-aware Neural Architecture Encoding with Transformers

CATE: Computation-aware Neural Architecture Encoding with Transformers Code for paper: CATE: Computation-aware Neural Architecture Encoding with Trans

16 Dec 27, 2022
OpenMMLab Video Perception Toolbox. It supports Video Object Detection (VID), Multiple Object Tracking (MOT), Single Object Tracking (SOT), Video Instance Segmentation (VIS) with a unified framework.

English | 简体中文 Documentation: https://mmtracking.readthedocs.io/ Introduction MMTracking is an open source video perception toolbox based on PyTorch.

OpenMMLab 2.7k Jan 08, 2023
Training Confidence-Calibrated Classifier for Detecting Out-of-Distribution Samples / ICLR 2018

Training Confidence-Calibrated Classifier for Detecting Out-of-Distribution Samples This project is for the paper "Training Confidence-Calibrated Clas

168 Nov 29, 2022
Deep Learning agent of Starcraft2, similar to AlphaStar of DeepMind except size of network.

Introduction This repository is for Deep Learning agent of Starcraft2. It is very similar to AlphaStar of DeepMind except size of network. I only test

Dohyeong Kim 136 Jan 04, 2023
Implementation of SwinTransformerV2 in TensorFlow.

SwinTransformerV2-TensorFlow A TensorFlow implementation of SwinTransformerV2 by Microsoft Research Asia, based on their official implementation of Sw

Phan Nguyen 2 May 30, 2022
[WACV21] Code for our paper: Samuel, Atzmon and Chechik, "From Generalized zero-shot learning to long-tail with class descriptors"

DRAGON: From Generalized zero-shot learning to long-tail with class descriptors Paper Project Website Video Overview DRAGON learns to correct the bias

Dvir Samuel 25 Dec 06, 2022
A Multi-modal Model Chinese Spell Checker Released on ACL2021.

ReaLiSe ReaLiSe is a multi-modal Chinese spell checking model. This the office code for the paper Read, Listen, and See: Leveraging Multimodal Informa

DaDa 106 Dec 29, 2022
Weakly Supervised Text-to-SQL Parsing through Question Decomposition

Weakly Supervised Text-to-SQL Parsing through Question Decomposition The official repository for the paper "Weakly Supervised Text-to-SQL Parsing thro

14 Dec 19, 2022