Ongoing research training transformer models at scale
Making large AI models cheaper, faster and more accessible
A unified framework for scalable computing
Implementation of model parallel autoregressive transformers on GPUs
Distributed Deep learning with Keras & Spark
A TensorFlow implementation of Scalable Distributed Deep-RL
Intel® Nervana™ reference deep learning framework