Deep Learning (with PyTorch)

Overview

Deep Learning (with PyTorch) Binder

This notebook repository now has a companion website, where all the course material can be found in video and textual format.

🇬🇧   🇨🇳   🇰🇷   🇪🇸   🇮🇹   🇹🇷   🇯🇵   🇸🇦   🇫🇷   🇮🇷   🇷🇺   🇻🇳   🇷🇸   🇵🇹   🇭🇺

Getting started

To be able to follow the exercises, you are going to need a laptop with Miniconda (a minimal version of Anaconda) and several Python packages installed. The following instruction would work as is for Mac or Ubuntu Linux users, Windows users would need to install and work in the Git BASH terminal.

Download and install Miniconda

Please go to the Anaconda website. Download and install the latest Miniconda version for Python 3.7 for your operating system.

wget <http:// link to miniconda>
sh <miniconda*.sh>

Check-out the git repository with the exercise

Once Miniconda is ready, checkout the course repository and proceed with setting up the environment:

git clone https://github.com/Atcold/pytorch-Deep-Learning

Create isolated Miniconda environment

Change directory (cd) into the course folder, then type:

# cd pytorch-Deep-Learning
conda env create -f environment.yml
source activate pDL

Start Jupyter Notebook or JupyterLab

Start from terminal as usual:

jupyter lab

Or, for the classic interface:

jupyter notebook

Notebooks visualisation

Jupyter Notebooks are used throughout these lectures for interactive data exploration and visualisation.

We use dark styles for both GitHub and Jupyter Notebook. You should try to do the same, or they will look ugly. JupyterLab has a built-in selectable dark theme, so you only need to install something if you want to use the classic notebook interface. To see the content appropriately in the classic interface install the following:

Comments
  • Chapter 5-2 docs

    Chapter 5-2 docs

    Optimization techniques II

    We discuss adaptive methods for SGD such as RMSprop and ADAM. We also talk about normalization layers and their effects on the neural network training process. Finally, we discuss a real-world example of neural nets being used in industry to make MRI scans faster and more efficient.

    Please let me know if any changes need to be made before merging.

    opened by guidopetri 16
  • Updates to current packages

    Updates to current packages

    This:

    • Moves PyTorch from 0.4 to 1.1 (one tiny code change)
    • Moves Python from 3.6 to 3.7 (no changes to code, just env)
    • Moves 1-2 requirements out of notebooks and into environment (potential nasty scipy pip install from librosa avoided!)
    • Uses conda kernels so the correct environment kernel is available (all notebooks rerun to pick up proper kernel)
    • Adds JuptyerLab (not required, but nice) - the interactive backend in the final notebook is still best in the classic interface. Try out built-in dark mode!

    All notebooks seem to run (except noted minor issue with JupyterLab)

    opened by henryiii 16
  • [FR & EN] YouTube subtitles

    [FR & EN] YouTube subtitles

    Hi Alf :wave:,

    As indicated in my last email, I can't afford to wait for Yann's return without a big delay on my side. So here are the subtitle files:

    • For English, it is the addition of the unicode. In practice:
    1. The list of files not modified during this review of the unicode: practinum1 (didn't need unicode), practinum4 (the file contains blocks of 3 instead of 2 for the others), for lecture 12 (the only file I didn't translate into French)

    2. The list of finished files (full English review + unicode) : lecture 6 & 9

    3. The list of about clean files (partial English review + unicode) : lecture 1-3,10,11 + practinum 1-3, 7-8, 10

    4. The list of not clean files (no English review + unicode): lecture 5-9,12-15 + practinum 5-6,9,11-15

    • For French, these are all the subtitles (except for lecture 12 where I have huge problems understanding Mike Lewis's accent and so I preferred not to put anything than to translate badly).

    I also added a disclamer for the V2 of the French translation of the website which should arrive this month. It should be my next and last PR closing the French translation work :boom:

    Loïck

    opened by lbourdois 13
  • Broken image links in 3.3. Properties of natural signals

    Broken image links in 3.3. Properties of natural signals

    The following image links are broken:

    • [x] Figure 2(a)
    • [x] Figure 2(b)
    • [x] Figure 3(a)
    • [x] Figure 3(b)

    See https://atcold.github.io/pytorch-Deep-Learning/en/week03/03-3/

    I think the images were originally obtained from this presentation: 02 - CNN.pdf

    See pages 10-11


    Also, small suggestions:

    • [x] Change Figure 4 to include R^7 and R^2 as in Slide 20 . This would better match the text for Figure 4.

    • [x] Include Figure (4b maybe?) with that on Slide 21 to show what Padding is doing

    opened by feedthebeat90 11
  • Portuguese translation

    Portuguese translation

    Hi @Atcold ! I would like to know how and where should I commit markdown files in Portuguese? I recall that you have commented something with @ebetica .

    opened by ricardobarroslourenco 11
  • [ZH] 13-3 Inline latex broken

    [ZH] 13-3 Inline latex broken

    Hi @JonathanSum ! Just for your info, There seems to be some inline latex broken on lecture 13-3:

    Screen Shot 2020-09-23 at 22 41 44

    The rest of the lectures I've checked seem to be fine.

    opened by xcastilla 9
  • Reorganize the website structure

    Reorganize the website structure

    This PR reorganizes the website structure, so we now have:

    en/
      index.md
      about.dm
      week01/
      week02/
      ...
    zh/
      index.md
      about.md
      week01/
      week02/
      ...
    ...
    

    Hopefully it's less messy and easier to work with.

    After this is merged, I will pull the images out into a global directory as well.

    Also fixes some broken links in zh/index.md

    opened by ebetica 9
  • Problem visualizing spanish translation on github.io

    Problem visualizing spanish translation on github.io

    I found an error visualizing on the github.io page the file /docs/es/week02/02-1.md.

    The english version of the file appears before some parts and the layout of the spanish parts after the english parts gets a bit messed up.

    grafik

    grafik

    opened by mt0rm0 8
  • [EN] Fix timers

    [EN] Fix timers

    A PR that fixes the timers of the sbv files that I couldn't correct in PR #660 to avoid conflicts.

    I also took the opportunity to correct the few errors I caught when translating the lecture10.

    I also noticed that the sbv files of the practinums of weeks 14 and 15 were missing.

    opened by lbourdois 8
  • [ZH] translation of 06-2 and 06.md

    [ZH] translation of 06-2 and 06.md

    I have translated the top 50% of the RNN(06-02) in Chinese.

    I passed the course on deep learning.ai and I also wrote a few notebooks to help students in the coursera Tensorflow time series seq2seq notebook.

    opened by JonathanSum 8
  • Vanishing gradient notebook

    Vanishing gradient notebook

    Poornima and I have compared an LSTM and RNN and visualized the gradients with respect to the input. We see that the gradients for the RNN are much smaller compared to the LSTM.

    We are able to train MNIST for a large input sequence with an LSTM and failed to do so with an RNN.

    Hope this is useful. If we need to make any chances, please let us know !

    opened by karanchahal 8
  • Software version update for 2023

    Software version update for 2023

    Hi there,

    I hope these tips can help you: Using Docker, with torchtext version 0.9.0 and PyTorch version 1.8.0.

    Please note that PyTorch 1.8 may not have good support for CUDA versions newer than 11. If you are using a newer version of CUDA, you may want to consider using the CPU instead.

    opened by wenxin-bupt 0
  • <Fix> evaluation dataset, printed samples

    evaluation dataset, printed samples

    Bunch of minor "theoretical" changes in the evaluation function:

    1. test_data_gen was used as the data generator in the evaluation, instead of data_generator, thereby evaluating the net on the test set used for training (not an actual issue here given the sequences are randomized and not sampled from existing datasets, but in principle would lead to a data leak in realistic scenarios);
    2. the correct sequences printed were a sampling (with reinsertion) of the first 10 evaluated, instead of 10 sampled from the whole set of correct ones;
    3. the condition for printing the incorrectly classified sequences would declare the absence of misclassifications if verbose==False, independently of their actual presence;
    opened by hypothe 1
  • Added controller trainer and improved truck class

    Added controller trainer and improved truck class

    Added

    • new truck methods for randomizing state within contraints
    • new truck methods for seeing if truck is at dock or offscreen
    • Training script for optimizing controller

    Note

    I currently have not successfully trained the controller to convergence. I have based the training off of this. On the website, they mention that the controller is hard to train. I have tried training it on the website with no success, so it seems like even their lessons are difficult to train. However, the code for training should be very similar to the code on the website. You may also alter the amount of lessons, max time steps, learning rate etc. to see if the model converges. I have been trying for over a week and have not succeeded yet.

    opened by dafaronbi 1
  • fix chinese version of 12-3

    fix chinese version of 12-3

    I found that the Chinese version was basically machine translated, which caused the latex syntax to be broken. Of course, there are a lot of unreasonable translation. This PR is mainly about fixing broken latex. I also did my best to fix some of the translations that were too much bullshit.

    opened by vipcxj 0
  • Russian translation (dictionary)

    Russian translation (dictionary)

    I would question some translations in the dictionary for Russian: I've graduated this year and we haven't really translated everything. For example, it will be more understandable if I say "one-hot" in Russian as it is, rather than "унитарный код". Basically, I've never heard anyone calling it "унитарный код", to be honest...

    So I guess there is a choice between being academically strict or being understood.

    opened by xufana 4
  • Use conda instead of source activate

    Use conda instead of source activate

    I think source activate is a few years old now and isn't supported anymore. https://stackoverflow.com/questions/49600611/python-anaconda-should-i-use-conda-activate-or-source-activate-in-linux

    opened by ebetica 0
Releases(dlsp19)
  • dlsp19(Jan 30, 2020)

    This is the notes for the Spring 2019 Deep Learning course at NYU. This course concerns the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional net and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition.

    This is the initial draft of the course notes - they are based off of a course developed for the the African Masters of Machine Intelligence (AMMI). You can access that version here

    Source code(tar.gz)
    Source code(zip)
  • aims-fl18(Jan 30, 2020)

    The African Masters of Machine Intelligence (AMMI) is Africa's flagship program in machine intelligence led by The African Institute for Mathematical Sciences (AIMS). These lessons, developed during the course of several years while I've been teaching at Purdue and NYU, are here proposed for the AMMI (AIMS).

    Prior to this course delivered for AMMI (AIMS), an earlier version of this was delivered and video-recorded for the Computational and Data Science for High Energy Physics (CoDaS-HEP) summer school at Princeton University. Please refer to this version release here.

    Source code(tar.gz)
    Source code(zip)
  • v1.0.0(Nov 5, 2018)

    Click CoDaS-HEP_2018 to jump to this release.

    These lessons, developed during the course of several years while I've been teaching at Purdue and NYU, are here proposed for the Computational and Data Science for High Energy Physics (CoDaS-HEP) summer school at Princeton University. The whole course has been recorded and the playlist is made available here. Check the slides for drawings of better visual quality.

    Source code(tar.gz)
    Source code(zip)
Owner
Alfredo Canziani
Musician, math lover, cook, dancer, 🏳️‍🌈, and assistant professor of Computer Science at New York University
Alfredo Canziani
Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper

Ponder(ing) Transformer Implementation of a Transformer that learns to adapt the number of computational steps it takes depending on the difficulty of

Phil Wang 65 Oct 04, 2022
A basic neural network for image segmentation.

Unet_erythema_detection A basic neural network for image segmentation. 前期准备 1.在logs文件夹中下载h5权重文件,百度网盘链接在logs文件夹中 2.将所有原图 放置在“/dataset_1/JPEGImages/”文件夹

1 Jan 16, 2022
PyTorch GPU implementation of the ES-RNN model for time series forecasting

Fast ES-RNN: A GPU Implementation of the ES-RNN Algorithm A GPU-enabled version of the hybrid ES-RNN model by Slawek et al that won the M4 time-series

Kaung 305 Jan 03, 2023
Synthetic LiDAR sequential point cloud dataset with point-wise annotations

SynLiDAR dataset: Learning From Synthetic LiDAR Sequential Point Cloud This is official repository of the SynLiDAR dataset. For technical details, ple

78 Dec 27, 2022
A PyTorch implementation of Learning to learn by gradient descent by gradient descent

Intro PyTorch implementation of Learning to learn by gradient descent by gradient descent. Run python main.py TODO Initial implementation Toy data LST

Ilya Kostrikov 300 Dec 11, 2022
Tiny-NewsRec: Efficient and Effective PLM-based News Recommendation

Tiny-NewsRec The source codes for our paper "Tiny-NewsRec: Efficient and Effective PLM-based News Recommendation". Requirements PyTorch == 1.6.0 Tensor

Yang Yu 3 Dec 07, 2022
Python Environment for Bayesian Learning

Pebl is a python library and command line application for learning the structure of a Bayesian network given prior knowledge and observations. Pebl in

Abhik Shah 103 Jul 14, 2022
FindFunc is an IDA PRO plugin to find code functions that contain a certain assembly or byte pattern, reference a certain name or string, or conform to various other constraints.

FindFunc: Advanced Filtering/Finding of Functions in IDA Pro FindFunc is an IDA Pro plugin to find code functions that contain a certain assembly or b

213 Dec 17, 2022
This repo is duplication of jwyang/faster-rcnn.pytorch

Faster RCNN Pytorch This repo is duplication of jwyang/faster-rcnn.pytorch C/C++ code are removed and easier to study. Python 3.8.5 Ubuntu 20.04.1 LTS

Kim Jihwan 1 Jan 14, 2022
Genshin-assets - 👧 Public documentation & static assets for Genshin Impact data.

genshin-assets This repo provides easy access to the Genshin Impact assets, primarily for use on static sites. Sources Genshin Optimizer - An Artifact

Zerite Development 5 Nov 22, 2022
This repository is the offical Pytorch implementation of ContextPose: Context Modeling in 3D Human Pose Estimation: A Unified Perspective (CVPR 2021).

Context Modeling in 3D Human Pose Estimation: A Unified Perspective (CVPR 2021) Introduction This repository is the offical Pytorch implementation of

37 Nov 21, 2022
The official implementation of NeurIPS 2021 paper: Finding Optimal Tangent Points for Reducing Distortions of Hard-label Attacks

The official implementation of NeurIPS 2021 paper: Finding Optimal Tangent Points for Reducing Distortions of Hard-label Attacks

machen 11 Nov 27, 2022
UPSNet: A Unified Panoptic Segmentation Network

UPSNet: A Unified Panoptic Segmentation Network Introduction UPSNet is initially described in a CVPR 2019 oral paper. Disclaimer This repository is te

Uber Research 622 Dec 26, 2022
NeurIPS-2021: Neural Auto-Curricula in Two-Player Zero-Sum Games.

NAC Official PyTorch implementation of NAC from the paper: Neural Auto-Curricula in Two-Player Zero-Sum Games. We release code for: Gradient based ora

Xidong Feng 19 Nov 11, 2022
State-of-the-art language models can match human performance on many tasks

Status: Archive (code is provided as-is, no updates expected) Grade School Math [Blog Post] [Paper] State-of-the-art language models can match human p

OpenAI 259 Jan 08, 2023
The Deep Learning with Julia book, using Flux.jl.

Deep Learning with Julia DL with Julia is a book about how to do various deep learning tasks using the Julia programming language and specifically the

Logan Kilpatrick 67 Dec 25, 2022
Airborne Optical Sectioning (AOS) is a wide synthetic-aperture imaging technique

AOS: Airborne Optical Sectioning Airborne Optical Sectioning (AOS) is a wide synthetic-aperture imaging technique that employs manned or unmanned airc

JKU Linz, Institute of Computer Graphics 39 Dec 09, 2022
Data and code for ICCV 2021 paper Distant Supervision for Scene Graph Generation.

Distant Supervision for Scene Graph Generation Data and code for ICCV 2021 paper Distant Supervision for Scene Graph Generation. Introduction The pape

THUNLP 23 Dec 31, 2022
Official implementation of the paper 'Details or Artifacts: A Locally Discriminative Learning Approach to Realistic Image Super-Resolution' in CVPR 2022

LDL Paper | Supplementary Material Details or Artifacts: A Locally Discriminative Learning Approach to Realistic Image Super-Resolution Jie Liang*, Hu

150 Dec 26, 2022