A blog about my thoughts on ML Sys and LLMs| Biao's Blog
We’re on a journey to advance and democratize artificial intelligence through open source and open science.| huggingface.co
Using Fully Sharded Data Parallel (FSDP) with Intel Gaudi¶| docs.habana.ai
Training large language models (LLMs) costs less than you think. Using the MosaicML platform, we show how fast, cheap, and easy it is to train these models at scale (1B -> 70B parameters). With new training recipes and infrastructure designed for large workloads, we enable you to train LLMs while maintaining total customizability over your model and dataset.| Databricks
We benchmarked LLM training on a multi-node AMD MI250 cluster and found near-linear scaling on up to 128 GPUs, demonstrating a compelling option for multi-node LLM training.| Databricks