============================================================================================================ `MILA will stop developing Theano <https://groups.google.com/d/msg/theano-users/7Poq8BZutbY/rNCIfvAEAwAJ>`_. The PyMC developers are continuing Theano development in a `fork <https://github.com/pymc-devs/theano-pymc>`_. ============================================================================================================ To install the package, see this page: http://deeplearning.net/software/theano/install.html For the documentation, see the project website: http://deeplearning.net/software/theano/ Related Projects: https://github.com/Theano/Theano/wiki/Related-projects It is recommended that you look at the documentation on the website, as it will be more current than the documentation included with the package. In order to build the documentation yourself, you will need sphinx. Issue the following command: :: python ./doc/scripts/docgen.py Documentation is built into ``html/`` The PDF of the documentation can be found at ``html/theano.pdf`` ================ DIRECTORY LAYOUT ================ ``Theano`` (current directory) is the distribution directory. * ``Theano/theano`` contains the package * ``Theano/theano`` has several submodules: * ``gof`` + ``compile`` are the core * ``scalar`` depends upon core * ``tensor`` depends upon ``scalar`` * ``sparse`` depends upon ``tensor`` * ``sandbox`` can depend on everything else * ``Theano/examples`` are copies of the example found on the wiki * ``Theano/benchmark`` and ``Theano/examples`` are in the distribution, but not in the Python package * ``Theano/bin`` contains executable scripts that are copied to the bin folder when the Python package is installed * Tests are distributed and are part of the package, i.e. fall in the appropriate submodules * ``Theano/doc`` contains files and scripts used to generate the documentation * ``Theano/html`` is where the documentation will be generated
Theano is a Python library that allows you to define, optimize, and evaluate mathematical expressions involving multi-dimensional arrays efficiently. It can use GPUs and perform efficient symbolic differentiation.
Overview
Applications using the GTN library and code to reproduce experiments in "Differentiable Weighted Finite-State Transducers"
gtn_applications An applications library using GTN. Current examples include: Offline handwriting recognition Automatic speech recognition Installing
Code for "Learning Canonical Representations for Scene Graph to Image Generation", Herzig & Bar et al., ECCV2020
Learning Canonical Representations for Scene Graph to Image Generation (ECCV 2020) Roei Herzig*, Amir Bar*, Huijuan Xu, Gal Chechik, Trevor Darrell, A
An evaluation toolkit for voice conversion models.
Voice-conversion-evaluation An evaluation toolkit for voice conversion models. Sample test pair Generate the metadata for evaluating models. The direc
PyTorch implementation of PP-LCNet
PP-LCNet-Pytorch Pre-Trained Models Google Drive p018 Accuracy Models Top1 Top5 PPLCNet_x0_25 0.5186 0.7565 PPLCNet_x0_35 0.5809 0.8083 PPLCNet_x0_5 0
Implementation of "Semi-supervised Domain Adaptive Structure Learning"
Semi-supervised Domain Adaptive Structure Learning - ASDA This repo contains the source code and dataset for our ASDA paper. Illustration of the propo
This porject is intented to build the most accurate model for predicting the porbability of loan default
Estimating-Loan-Default-Probability IBA ML2 Mid-project / Kaggle Competition This porject is intented to build the most accurate model for predicting
This is the official pytorch implementation for our ICCV 2021 paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering" on VQA Task
🌈 ERASOR (RA-L'21 with ICRA Option) Official page of "ERASOR: Egocentric Ratio of Pseudo Occupancy-based Dynamic Object Removal for Static 3D Point C
A PyTorch implementation of Mugs proposed by our paper "Mugs: A Multi-Granular Self-Supervised Learning Framework".
Mugs: A Multi-Granular Self-Supervised Learning Framework This is a PyTorch implementation of Mugs proposed by our paper "Mugs: A Multi-Granular Self-
The official PyTorch implementation of the paper: *Xili Dai, Xiaojun Yuan, Haigang Gong, Yi Ma. "Fully Convolutional Line Parsing." *.
F-Clip — Fully Convolutional Line Parsing This repository contains the official PyTorch implementation of the paper: *Xili Dai, Xiaojun Yuan, Haigang
My take on a practical implementation of Linformer for Pytorch.
Linformer Pytorch Implementation A practical implementation of the Linformer paper. This is attention with only linear complexity in n, allowing for v
particle tracking model, works with the ROMS output file(qck.nc, his.nc)
particle-tracking-model-for-ROMS particle tracking model, works with the ROMS output file(qck.nc, his.nc) description this is a 2-dimensional particle
Deep Text Search is an AI-powered multilingual text search and recommendation engine with state-of-the-art transformer-based multilingual text embedding (50+ languages).
Deep Text Search - AI Based Text Search & Recommendation System Deep Text Search is an AI-powered multilingual text search and recommendation engine w
Tutorial on active learning with the Nvidia Transfer Learning Toolkit (TLT).
Active Learning with the Nvidia TLT Tutorial on active learning with the Nvidia Transfer Learning Toolkit (TLT). In this tutorial, we will show you ho
Enhancing Knowledge Tracing via Adversarial Training
Enhancing Knowledge Tracing via Adversarial Training This repository contains source code for the paper "Enhancing Knowledge Tracing via Adversarial T
Deep Learning ❤️ OneFlow
Deep Learning with OneFlow made easy 🚀 ! Carefree? carefree-learn aims to provide CAREFREE usages for both users and developers. User Side Computer V
Efficient Training of Audio Transformers with Patchout
PaSST: Efficient Training of Audio Transformers with Patchout This is the implementation for Efficient Training of Audio Transformers with Patchout Pa
Parallel Latent Tree-Induction for Faster Sequence Encoding
FastTrees This repository contains the experimental code supporting the FastTrees paper by Bill Pung. Software Requirements Python 3.6, NLTK and PyTor
xitorch: differentiable scientific computing library
xitorch is a PyTorch-based library of differentiable functions and functionals that can be widely used in scientific computing applications as well as deep learning.
PyTorch code to run synthetic experiments.
Code repository for Invariant Risk Minimization Source code for the paper: @article{InvariantRiskMinimization, title={Invariant Risk Minimization}
Codes for CyGen, the novel generative modeling framework proposed in "On the Generative Utility of Cyclic Conditionals" (NeurIPS-21)
On the Generative Utility of Cyclic Conditionals This repository is the official implementation of "On the Generative Utility of Cyclic Conditionals"