Skip to content

Latest commit

 

History

History
68 lines (39 loc) · 3.69 KB

README.md

File metadata and controls

68 lines (39 loc) · 3.69 KB

A PyTorch Native LLM Training Framework

An Industrial-Level Framework for Easy-of-Use

  • 🔥 PyTorch Native: veScale is rooted in PyTorch-native data structures, operators, and APIs, enjoying the ecosystem of PyTorch that dominates the ML world.

  • 🛡 Zero Model Code Change: veScale decouples distributed system design from model architecture, requiring near-zero or zero modification on the model code of users.

  • 🚀 Single Device Abstraction: veScale provides single-device semantics to users, automatically distributing and orchestrating model execution in a cluster of devices.

  • 🎯 Automatic Parallelism Planning: veScale parallelizes model execution with a synergy of strategies (tensor, sequence, data, ZeRO, pipeline parallelism) under semi- or full-automation [coming soon].

  • Eager & Compile Mode: veScale supports not only Eager-mode automation for parallel training and inference but also Compile-mode for ultimate performance [coming soon].

  • 📀 Automatic Checkpoint Resharding: veScale manages distributed checkpoints automatically with online resharding across different cluster sizes and different parallelism strategies.

Latest News

Coming Soon

veScale is still in its early phase. We are refactoring our internal LLM training system components to meet open source standard. The tentative timeline is as follows:

  • High-level nD parallel api for extreme ease of use

  • Power-user plan api for easy customization of nD parallel training

  • End-to-end vescale/examples with 5D parallel training (TP, SP, DP, ZeRO, PP)

Table of Content (web view)

Introduction

Quick Start

DTensor

Parallel

Plan

Checkpoint

The veScale Project is under the Apache License v2.0.