Fine-Tuning Llama2-70B with DeepSpeed ZeRO-3 and Low-Rank Adaptation (LoRA) on Intel® Gaudi®2 AI Accelerator
With Habana’s SynapseAI 1.13.0 release, users can run Fine Tune the Llama2 70B model using only 8 Gaudi2 Accelerators.
DeepSpeed, Fine Tuning, Llama, LoRA
We’re excited to introduce the release of Habana® SynapseAI® Software version 1.13.0
Bringing forth numerous enhancements and updates for an improved user experience.
Training Llama and Bloom 13 Billion Parameter LLMs with 3D Parallelism on Habana® Gaudi2®
One of the main challenges in training Large Language Models (LLMs) is that they are often too large to fit on a single node or even if they fit, the training may be too slow. To address this issue, their training can be parallelized across multiple Gaudi accelerators (HPUs).
3D-Parallelism, DeepSpeed, GenAI, Large Language Models
Porting a model to Megatron-DeepSpeed with Habana Gaudi
If you want to train a large model using Megatron-DeepSpeed, but the model you want is not included in the implementation, you can port it to the Megatron-DeepSpeed package. Assuming your model is transformer-based, you can add your implementation easily, basing it on existing code.
3D-Parallelism, DeepSpeed, GenAI, Large Language Models
Optimizing Large Language Model Inference on Gaudi2 with Hugging Face Optimum-Habana
We have optimized additional Large Language Models on Hugging Face using the Optimum Habana library.
DeepSpeed, Hugging Face, Inference
The Habana team is happy to announce the release of Habana® SynapseAI® Software version 1.11.0.
In this release, we’ve upgraded versions of several libraries, including DeepSpeed 0.9.4, PyTorch Lightning 2.0.4 and TensorFlow 2.12.1.
New MLCommons Results Highlight Impressive Competitive AI Gains for Intel
This Intel NewsByte was originally published in the Intel Newsroom. Habana Gaudi2 and 4th Gen ...
Gaudi2, MLPERF
Habana® Gaudi®2 Powers Deep Learning Instances on Genesis Cloud at Collision 2023
Habana Labs, an Intel company, and Genesis Cloud are collaborating to deliver a new class ...
Gaudi2
Habana Showcases Gaudi2 Performance on Large Language and Generative AI Models at ISC
We’re excited to participate in this year’s ISC High Performance Compute 2023 event in Hamburg Germany.
Equus Lab-as-a-Service with Habana Gaudi2 Processors Eases Testing and Deployment of Deep Learning Systems
Equus and Habana have teamed up to simplify the process of testing, implementing and deploying ...
Gaudi2
New Habana Autonomous Driving Use Case Enabled with Gaudi Processors
Announcing a new End-to-End use case showing Training of a semantic segmentation model for Autonomous Driving
Gaudi2, Training
The Habana team is happy to announce the release of SynapseAI® Software version 1.9.0.
In the 1.9 release, we’ve upgraded versions of several libraries, including PyTorch Lightning 1.9.4, DeepSpeed 0.7.7, fairseq 0.12.3, and Horovod v0.27.0.
synapseai
Fast Inference on Large Language Models: BLOOMZ on Habana Gaudi2 Accelerator
In this article, you'll learn how to easily deploy multi-billion parameter language models on Habana Gaudi2 and get a view into the Hugging Face performance evaluation of Gaudi2 and A100 on BLOOMZ.
Accelerating Distributed Training Performance using EFA Peer Direct on Gaudi-Based AWS EC2 DL1 Instances
AWS and Habana collaborated to enable EFA Peer Direct support on the Gaudi-based AWS DL1 ...
Gaudi2
New Habana AI Retail Use Case Enables Automated Store Shelf Management with Gaudi Processors
AI is becoming increasingly important for retail use cases. It can provide retailers with advanced ...
Gaudi2, Retail
Faster Training and Inference: Habana Gaudi®-2 vs Nvidia A100 80GB
In this article, you will learn how to use Habana® Gaudi®2 to accelerate model training and inference, and train bigger models with 🤗 Optimum Habana.
developer, Gaudi2, Hugging Face
BLOOM 176B Inference on Habana Gaudi2
With Habana’s SynapseAI 1.8.0 release support of DeepSpeed Inference, users can run inference on large language models, including BLOOM 176B.
BLOOM, DeepSpeed, Inference
White Paper: Riken Accelerates Medical and Drug Research with Habana Gaudi AI Processor
Our blog today features a Riken white paper, initially prepared and published by the Intel ...
The Habana team is happy to announce the release of SynapseAI® Software version 1.8.0
We have upgraded versions of several libraries with SynapseAI 1.8.0, including PyTorch 1.13.1, PyTorch Lightning 1.8.6 and TensorFlow 2.11.0 & 2.8.4.
developer, synapseai
Pre-Training the BERT 1.5B model with DeepSpeed
In this post, we show you how to run Habana’s DeepSpeed enabled BERT1.5B model from our Model-References repository.
BERT, DeepSpeed, developer, Gaudi, Gaudi2, pytorch, synapseai
Road sign detection using Transfer Learning with TensorFlow EfficientDet-D0
In this paper we’ll show how Transfer Learning is an efficient way to train an existing model on a new and unique dataset with equivalent accuracy and significantly less training time.
developer, EfficientDet, Transfer Learning
Eitan Medina – Video Interview – Cambrian AI Research
Habana’s Gaudi2 delivers amazing deep learning performance and price advantage for both training as well ...
Large Model usage with minGPT
This tutorial provides example training scripts to demonstrate different DeepSpeed optimization technologies on HPU. This tutorial will focus on the memory optimization technologies, including Zero Redundancy Optimizer(ZeRO) and Activation Checkpointing.
developer
Training Causal Language Models on SDSC’s Gaudi-based Voyager Supercomputing Cluster
The SDSC Voyager supercomputer is an innovative AI system designed specifically for science and engineering research at scale.
developer