xla
There are 49 repositories under xla topic.
elixir-nx/nx
Multi-dimensional arrays (tensors) and numerical definitions for Elixir
pytorch/xla
Enabling PyTorch on XLA Devices (e.g. Google TPU)
zml/zml
Any model. Any hardware. Zero compromise. Built with @ziglang / @openxla / MLIR / @bazelbuild
n2cholas/awesome-jax
JAX - A curated list of resources https://github.com/google/jax
gomlx/gomlx
GoMLX: An Accelerated Machine Learning Framework For Go
gordicaleksa/get-started-with-JAX
The purpose of this repo is to make it easy to get started with JAX, Flax, and Haiku. It contains my "Machine Learning with JAX" series of tutorials (YouTube videos and Jupyter Notebooks) as well as the content I found useful while learning about the JAX ecosystem.
felafax/felafax
Felafax is building AI infra for non-NVIDIA GPUs
mpi4jax/mpi4jax
Zero-copy MPI communication of JAX arrays, for turbo-charged HPC applications in Python :zap:
dfm/extending-jax
Extending JAX with custom C++ and CUDA code
JuliaGPU/XLA.jl
Julia on TPUs
kamalkraj/ALBERT-TF2.0
ALBERT model Pretraining and Fine Tuning using TF2.0
HomebrewML/revlib
Simple and efficient RevNet-Library for PyTorch with XLA and DeepSpeed support and parameter offload
inoryy/tensorflow-optimized-wheels
TensorFlow wheels built for latest CUDA/CuDNN and enabled performance flags: SSE, AVX, FMA; XLA
flaport/sax
S + Autograd + XLA :: S-parameter based frequency domain circuit simulations and optimizations using JAX.
HuiResearch/tfbert
基于tensorflow1.x的预训练模型调用,支持单机多卡、梯度累积,XLA加速,混合精度。可灵活训练、验证、预测。
AlibabaPAI/torchacc
PyTorch distributed training acceleration framework
sayakpaul/keras-xla-benchmarks
Presents comprehensive benchmarks of XLA-compatible pre-trained models in Keras.
DifferentiableUniverseInitiative/jaxDecomp
JAX bindings for the NVIDIA cuDecomp library
scala-network/scala-pool
Official scala pool repository
bahremsd/tmmax
A fast transfer matrix method written in jax for modelling optical multilayer thin films
onnx/onnx-xla
XLA integration of Open Neural Network Exchange (ONNX)
sseung0703/TF2-jit-compile-on-multi-gpu
Tensorflow2 training code with jit compiling on multi-GPU.
gottingen/tf-reading
tensorflow code reading
InikoPro/mineveruscoinonarm
Mine verus coin on ARM like Pi, Tablet, Mobile & Other.
AlibabaPAI/FlashModels
Fast and easy distributed model training examples.
mzguntalan/neptune
[WIP] Neptune: JAX iterop-able library in Haskell.
sayakpaul/you-dont-know-tensorflow
Contains materials for my talk "You don't know TensorFlow".
sayakpaul/xla-benchmark-sd
Provides code to serialize the different models involved in Stable Diffusion as SavedModels and to compile them with XLA.
jhn-nt/data-snax
Versatile Data Ingestion Pipelines for Jax
kmkolasinski/tensorflow-nanoGPT
Example how to train GPT-2 (XLA + AMP), export to SavedModel and serve with Tensorflow Serving
scala-network/StellitePay-API
DEPRECATED ⛔️
googleinterns/paksha
Compiling JAX to WebAssembly for exploring client-side machine learning
mugithi/google-terraform-pytorch-tpu
Automated provisioner of a Google Cloud TPU environment for training in PyTorch
wcxve/xspex
Access Xspec models and corresponding JAX/XLA ops.
jhashekhar/multilingual-clf
Classification of multilingual dataset trained only on English training data using pre-trained models. Model is trained on TPUs using PyTorch and torch_xla library.
ReturnToFirst/FastTFWorkflow
Tutorial about How to change your slow tensorflow training faster