Pinned Repositories
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Gaudi-tutorials
Tutorials for running models on First-gen Gaudi and Gaudi2 for Training and Inference. The source files for the tutorials on https://developer.habana.ai/
Habana_Custom_Kernel
Provides the examples to write and build Habana custom kernels using the HabanaTools
hccl_demo
Megatron-DeepSpeed
Intel Gaudi's Megatron DeepSpeed Large Language Models for training
Model-References
Reference models for Intel(R) Gaudi(R) AI Accelerator
Setup_and_Install
Setup and Installation Instructions for Habana binaries, docker image creation
SynapseAI_Core
SynapseAI Core is a reference implementation of the SynapseAI API running on Habana Gaudi
tpc_llvm
TPC-CLANG compiler that compiles a TPC C programming language which is used in HabanaLabs Deep-Learning Accelerators
vllm-fork
A high-throughput and memory-efficient inference and serving engine for LLMs
Intel® Gaudi® AI Accelerator 's Repositories
HabanaAI/Model-References
Reference models for Intel(R) Gaudi(R) AI Accelerator
HabanaAI/Gaudi-tutorials
Tutorials for running models on First-gen Gaudi and Gaudi2 for Training and Inference. The source files for the tutorials on https://developer.habana.ai/
HabanaAI/vllm-fork
A high-throughput and memory-efficient inference and serving engine for LLMs
HabanaAI/tpc_llvm
TPC-CLANG compiler that compiles a TPC C programming language which is used in HabanaLabs Deep-Learning Accelerators
HabanaAI/Setup_and_Install
Setup and Installation Instructions for Habana binaries, docker image creation
HabanaAI/Habana_Custom_Kernel
Provides the examples to write and build Habana custom kernels using the HabanaTools
HabanaAI/hccl_demo
HabanaAI/Megatron-DeepSpeed
Intel Gaudi's Megatron DeepSpeed Large Language Models for training
HabanaAI/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
HabanaAI/Gaudi-solutions
Full End-to-End examples showing how to use First-gen Gaudi and Gaudi2 in common use cases
HabanaAI/HCL
HabanaAI/Gaudi2-Workshop
HabanaAI/deepspeed_old
HabanaAI/hl-thunk-open
Thunk library for HabanaLabs kernel driver
HabanaAI/Intel_Gaudi3_Software
Intel® Gaudi® Software is an implementation of the runtime and graph compiler for Gaudi3
HabanaAI/Megatron-LM
Ongoing research training transformer models at scale
HabanaAI/habanalabs-k8s-device-plugin
HABANA device plugin for Kubernetes
HabanaAI/Fairseq
HabanaAI/optimum-habana-fork
Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)
HabanaAI/TOWL
HabanaAI/vllm-hpu-extension
HabanaAI/hccl_ofi_wrapper
HabanaAI/slurm
Slurm: A Highly Scalable Workload Manager
HabanaAI/drivers.accel.habanalabs.kernel
HabanaAI/papers
Academic papers by Habana research team
HabanaAI/pytorch-fork
Tensors and Dynamic neural networks in Python with strong GPU acceleration
HabanaAI/rdma-core
RDMA core userspace libraries and daemons
HabanaAI/AutoGPTQ
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
HabanaAI/drivers.gpu.linux-nic.kernel
NIC drivers (Ethernet, IBverbs and common) for the NIC IP that is inside Intel's data-center GPU
HabanaAI/neural-compressor
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime