Sections:
- Papers
- Languages
- Platforms
- Courses
- Live Sessions
- Books
- Cheat sheets
- Interview Questions
- DataSets
- Blogs and Community
- Online Competitions
- Other Sources
- About CNNs
- About GANs
Steps in approaching a Machine learning problem:
Below are the steps that I follow while approaching a ML problem.
- Defining and understanding the problem statement
- Gathering the Data
- Initial Exploration of Data
- In-depth EDA
- Building the model
- Analyzing the results with different models and shortlisting the ones which gives good performance measures
- Fine-tuning the selected model
- Document the code
- Deployment
- Monitoring the deployed model performance in real time.
- Towards a Human-like Open-Domain Chatbot
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators
- Reformer: The Efficient Transformer
- Longformer: The Long-Document Transformer
- You Impress Me: Dialogue Generation via Mutual Persona Perception
- Recipes for building an open-domain chatbot
- ToD-BERT: Pre-trained Natural Language Understanding for Task-Oriented Dialogues
- SOLOIST: Few-shot Task-Oriented Dialog with A Single Pre-trained Auto-regressive Model
- A Simple Language Model for Task-Oriented Dialogue
- FastBERT: a Self-distilling BERT with Adaptive Inference Time
- PoWER-BERT: Accelerating BERT Inference via Progressive Word-vector Elimination
- Data Augmentation using Pre-trained Transformer Models
- FLAT: Chinese NER Using Flat-Lattice Transformer
- End-to-End Object Detection with Transformers | code
- Objects as Points | code
- Acquisition of Localization Confidence for Accurate Object Detection | code
- Part-A^2 Net: 3D Part-Aware and Aggregation Neural Network for Object Detection from Point Cloud
- PointRCNN: 3D Object Proposal Generation and Detection from Point Cloud | code
- LaserNet: An Efficient Probabilistic 3D Object Detector for Autonomous Driving
- Probabilistic 3D Multi-Object Tracking for Autonomous Driving | code
- A Baseline for 3D Multi-Object Tracking | code
- Denoising Diffusion Probabilistic Models | code
- Joint Training of Variational Auto-Encoder and Latent Energy-Based Model
- Efficient and Scalable Bayesian Neural Nets with Rank-1 Factors | code
- BatchEnsemble: An Alternative Approach to Efficient Ensemble and Lifelong Learning | code
- Stable Neural Flows
- How Good is the Bayes Posterior in Deep Neural Networks Really
Knowing one language couldn't help to work in multidomain,so its good to learn two or more languages like python
, c++
, java
. For machine learning,one should be sound in any of the language with other language basics python
, R
, matlab
, julia
.
- Jet Brains
Jet Brains provides basics to advance training onJava
,Koltin
andPython
languages, can learn and apply in developing mini projects based on the level selected. - HackerRank
- HackerEarth
- LeetCode
- CodeChef
- CP algorithms
As we know that machine(deep) learning models need high computational power, it might not be possible for an individual for purchase addtional CPUs
or GPUs
, so we use cloud platforms.
Note: In general we use Jupyter Notebook
to run our models. For Jupyter Notebook we need to install the software on local machine, for windows, linux
Note: Google Products like Colab, Kaggle are free and best platform for an individual.
Most of the courses are available in Python
language. kindly re-check before taking the course of you have any language constraints.
Note: No organisation can be seen here, kindly pick the course based on your requirement.
Machine Learning
- Machine Learning (Stanford CS229)
- Python for Machine Learning (Great Learning)
- Statistics for Machine Learrning (Great Learning)
- Data Visualization using Python (Great Learning)
- Machine Learning Crash Course (Google)
- Machine Learning - Linear Regression (LEAPS)
- Getting started with Decision Trees (Analytics Vidhya)
- Machine Learning with Python: A Practical Introduction (Harvard EdX)
- Machine Learning Fundamentals (Harvard EdX)
- Machine Learning with Python: from Linear Models to Deep Learning (Harvard EdX)
- Machine Learning (Harvard EdX)
- Machine Learning with Python (IBM)
- Machine Learning – Dimensionality Reduction (IBM)
- Data Visualization with Python (IBM)
- Data Analysis with Python (IBM)
- Applied Machine Learning 2020 (Columbia)
DataScience
- Introduction to Python (Analytics Vidhya)
- Fundamentals of Data Analytics (LEAPS)
- Pandas for Data Analysis in Python (Analytics Vidhya)
- Tableau for Beginners (Analytics Vidhya)
- Top Data Science Projects for Analysts and Data Scientists (Analytics Vidhya)
- Machine Learning for Data Science and Analytics (Harvard EdX)
- A data science program for everyone (Harvard EdX)
- Data Science and Machine Learning Capstone Project (Harvard EdX)
- R Basics (Harvard EdX)
- Data Visualisation and EDA (Harvard EdX)
- Probability Theory (Harvard EdX)
- Data Science Inference and Modeling (Harvard EdX)
- Data Science Productivity Tools (Harvard EdX)
- Data Science Wrangling (Harvard EdX)
- Data Science Linear Regression (Harvard EdX)
- Capstone Project (Harvard EdX)
- Introduction to Data Science (IBM)
- Python for Data Science (IBM)
- SQL and Relational Databases 101 (IBM)
APIs
- Deep Learning Fundamentals (IBM)
- Deep Learning with TensorFlow (IBM)
- Accelerating Deep Learning with GPU (IBM)
- Convolutional Neural Networks for Visual Recognition (Stanford CS231n)
Artificial Intelligence
- Introduction to Artificial Intelligence with Python (Harvard EdX)
- Introduction to Artificial Intelligence (UC Berkeley CS188)
Natural Language Processing
- Introduction to Natural Language Processing
- Natural Language Processing (NLP) - Microsoft
- Natural Language Processing with Deep Learning (Stanford CS224N)
- Natural Language Understanding (Stanford CS224U)
Commputer Vision
Matlab
Note: In the below available sections if the required book/cheatsheet is not present or for more info check the more
.
Note: Few google links are getting crashing for few files, in such scenario, can direclty check more
(folder) .
- What is Machine Learning
- Tree Based Algorithms
- Natural Language Processing
- Data Cleaning with Numpy and Pandas
- Data Engineering CookBook
- Machine Learning Projects in Python (Compiled)
- Natural Langugae with Python
- Python
- R
- SQL
- Machine Learning
- Supervised learning
- Data Science
- Probability
- Statistics
- Data Engineering
- Git
- Python
- Data Science
- Variance in Data Science
- Interview question data science
- Self prepared Interview Questions
Note: These Self prepared Interview Questions will be updated weekly.
- Kaggle
- Meta-Sim: Learning to Generate Synthetic Datasets
- UCI
- Worldbank Data
- Tensorflow
- Movielens
- Huggingface
- Towards data science
- Analyticsvidhya
- Machinelearningmastery
- Medium
- PyTorch Discussion Forum
- StackOverflow PyTorch Tags
- Website
- Paperswithcode
- Google research
- Madewithml
- Deeplearning
- Huggingface
- Arcgis
- Pytorchdeeplearning
- Fastaibook
- EasyOCR
- Awesome Data Science
- NLP progress
- Recommenders
- Awesome pytorch list
- Free data science books
- Deeplearningdrizzle
- Data science ipython notebooks
- Julia
- Gym
- Regarding satellite images
- Monk Object Detection
- Awesome pytorch list
- Topdeeplearning
- NLPprogress
- Multi task NLP
- Opencv
- Transformers
- Code implementations for research papers
- Tool for visualizing attention in the Transformer model
- Powerful and efficient Computer Vision Annotation Tool (CVAT)
- Powerful and efficient Computer Vision Annotation Tool (CVAT)
- TransCoder
- Super Duper NLP
- Tessellate Imaging
- Machine Learning with Python
- GPT 2
- Data augmentation for NLP
- Mlops
- Reinforcement learning
- keras applications
- LegoNet: Efficient Convolutional Neural Networks with Lego Filters
- MeshCNN, a convolutional neural network designed specifically for triangular meshes
- Octave Convolution
- PyTorch Image Models, ResNet/ResNeXT, DPN, MobileNet-V3/V2/V1, MNASNet, Single-Path NAS, FBNet
- Deep Neural Networks with Box Convolutions
- Invertible Residual Networks
- Stochastic Downsampling for Cost-Adjustable Inference and Improved Regularization in Convolutional Networks
- Faster Faster R-CNN Implementation
- Faster R-CNN Another Implementation
- Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer
- Wide ResNet model in PyTorch
- DiracNets: Training Very Deep Neural Networks Without Skip-Connections
- An End-to-End Trainable Neural Network for Image-based Sequence Recognition and Its Application to Scene Text Recognition
- Efficient Densenet
- Video Frame Interpolation via Adaptive Separable Convolution
- Learning local feature descriptors with triplets and shallow convolutional neural networks
- Densely Connected Convolutional Networks
- Very Deep Convolutional Networks for Large-Scale Image Recognition
- SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size
- Deep Residual Learning for Image Recognition
- Training Wide ResNets for CIFAR-10 and CIFAR-100 in PyTorch
- Deformable Convolutional Network
- Convolutional Neural Fabrics
- Deformable Convolutional Networks in PyTorch
- Dilated ResNet combination with Dilated Convolutions
- Striving for Simplicity: The All Convolutional Net
- Convolutional LSTM Network
- Big collection of pretrained classification models
- PyTorch Image Classification with Kaggle Dogs vs Cats Dataset
- CIFAR-10 on Pytorch with VGG, ResNet and DenseNet
- Base pretrained models and datasets in pytorch (MNIST, SVHN, CIFAR10, CIFAR100, STL10, AlexNet, VGG16, VGG19, ResNet, Inception, SqueezeNet)
- NVIDIA/unsupervised-video-interpolation
- Mimicry, PyTorch Library for Reproducibility of GAN Research
- Clean Readable CycleGAN
- StarGAN
- Block Neural Autoregressive Flow
- High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
- A Style-Based Generator Architecture for Generative Adversarial Networks
- GANDissect, PyTorch Tool for Visualizing Neurons in GANs
- Learning deep representations by mutual information estimation and maximization
- Variational Laplace Autoencoders
- VeGANS, library for easily training GANs
- Progressive Growing of GANs for Improved Quality, Stability, and Variation
- Conditional GAN
- Wasserstein GAN
- Adversarial Generator-Encoder Network
- Image-to-Image Translation with Conditional Adversarial Networks
- Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks
- On the Effects of Batch and Weight Normalization in Generative Adversarial Networks
- Improved Training of Wasserstein GANs
- Collection of Generative Models with PyTorch
- Generative Adversarial Nets (GAN)
- Variational Autoencoder (VAE)
- Improved Training of Wasserstein GANs
- CycleGAN and Semi-Supervised GAN
- Improving Variational Auto-Encoders using Householder Flow and using convex combination linear Inverse Autoregressive Flow
- PyTorch GAN Collection
- Generative Adversarial Networks, focusing on anime face drawing
- Simple Generative Adversarial Networks
- Adversarial Auto-encoders
- torchgan: Framework for modelling Generative Adversarial Networks in Pytorch
Thank You for visiting the repo, hope it helped you! I would like to hear suggestions from you!!