Hiroki11x
PhD Candidate at Université de Montréal, Mila / Student Research at Google DeepMind / HPC, Deep Learning, LLM / ex-Tokyo Tech, Microsoft Research, IBM Research
Mila, Université de MontréalMontreal, QC, Canada
Hiroki11x's Stars
jadore801120/attention-is-all-you-need-pytorch
A PyTorch implementation of the Transformer model in "Attention is All You Need".
intel/intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
lblaoke/EMCMC
PyTorch implementation of the paper "Entropy-MCMC: Sampling from Flat Basins with Ease" (ICLR 2024)
thuml/CLIPood
About Code Release for "CLIPood: Generalizing CLIP to Out-of-Distributions" (ICML 2023), https://arxiv.org/abs/2302.00864
ryoungj/optdom
[ICLR'22] Self-supervised learning optimally robust representations for domain shift.
yorkerlin/StructuredNGD-DL
Matrix-multiplication-only KFAC; Code for ICML 2023 paper on Simplifying Momentum-based Positive-definite Submanifold Optimization with Applications to Deep Learning
AiMl-hub/UPLM
Uncertainty-Guided Pseudo-Labelling with Model Averaging
garrettj403/SciencePlots
Matplotlib styles for scientific plotting
ZZhangsm/OOKD
Revisiting Knowledge Distillation under Distribution Shift: https://arxiv.org/abs/2312.16242
MarlonBecker/MSAM
qiaoruiyt/NoiseRobustDG
hitachi-nlp/ensemble-metrics
lolapriego/coursework
Checklist of videos and exercises that you can follow if you want to get a good baseline for Data Structures and Algorithms.
Nutlope/aicommits
A CLI that writes your git commit messages for you with AI
rethread-studio/algorithmic-art-course
Collection of resources for the algorithmic art course at the Université de Montréal
qingsongedu/Awesome-TimeSeries-SpatioTemporal-LM-LLM
A professional list on Large (Language) Models and Foundation Models (LLM, LM, FM) for Time Series, Spatiotemporal, and Event Data.
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
IntelLabs/academic-budget-bert
Repository containing code for "How to Train BERT with an Academic Budget" paper
sebastianbergner/ExploringCLFD
Code for the paper Exploring the Properties of Hypernetworks for Continual Learning in Robotics.
adityatelange/hugo-PaperMod
A fast, clean, responsive Hugo theme.
TrustAIoT/CR-SAM
Code for AAAI 2024 paper: CR-SAM: Curvature Regularized Sharpness-Aware Minimization
d-li14/mobilenetv2.pytorch
72.8% MobileNetV2 1.0 model on ImageNet and a spectrum of pre-trained MobileNetV2 models
llm-jp/modelwg
LLM-jp Model-WG Working Directory
openai/weak-to-strong
microsoft/mttl
Building modular LMs with parameter-efficient fine-tuning.
r-three/mats
filipbasara0/simple-convnext
Simple implementation of the ConvNext architecture in PyTorch
namkoong-lab/whyshift
A python package providing a benchmark with various specified distribution shift patterns.
VirtuosoResearch/Robust-Fine-Tuning
Measuring generalization properties of fine-tuning using Hessian
mlcommons/algorithmic-efficiency
MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvements in both training algorithms and models.