With the Intel Gaudi SynapseAI 1.13.0 release, users can run Fine Tune the Llama2 70B model using only 8 Gaudi2 Accelerators.| Intel Gaudi Developers
With Habana’s SynapseAI 1.8.0 release support of DeepSpeed Inference, users can run inference on large language models, including BLOOM 176B.| Habana Developers
In this post, we show you how to run Habana’s DeepSpeed enabled BERT1.5B model from our Model-References repository.| Habana Developers
Fine tuning GPT2 with Hugging Face and Habana Gaudi. In this tutorial, we will demonstrate fine tuning a GPT2 model on Habana Gaudi AI processors using Hugging Face optimum-habana library with DeepSpeed.| Habana Developers
Optimize your deep learning with data parallel processes on Intel Gaudi using DeepSpeed. Enhance efficiency in training with our expert insights.| Intel Gaudi Developers
One of the main challenges in training Large Language Models (LLMs) is that they are often too large to fit on a single node or even if they fit, the training may be too slow. To address this issue, their training can be parallelized across multiple Gaudi accelerators (HPUs).| Habana Developers
If you want to train a large model using Megatron-DeepSpeed, but the model you want is not included in the implementation, you can port it to the Megatron-DeepSpeed package. Assuming your model is transformer-based, you can add your implementation easily, basing it on existing code.| Habana Developers
We have optimized additional Large Language Models on Hugging Face using the Optimum Habana library.| Intel Gaudi Developers