Skip to main content

12.10 Introduction to Mindformers

The goal of MindSpore Transformers is to build a full-flow development suite for large model training, fine-tuning, evaluation, inference, and deployment: providing the industry's mainstream Transformer class of pre-training models and SOTA downstream task applications, covering a wealth of parallel features. It is expected to help users easily realize large model training and innovative research and development. MindSpore Transformers Suite is based on MindSpore's built-in parallel technology and componentized design, with the following features:

● Seamlessly switch from single card to large-scale cluster training with a single line of code;

● Provides flexible and easy-to-use personalized parallel configuration;

●Ability to automatically perform topology sensing and efficiently fuse data parallelism and model parallelism strategies;

● One-click initiation of single/multi-card training, fine-tuning, evaluation, and inference processes for any task;

●Support users to perform componentized configuration of any module, such as optimizer, learning strategy, network assembly, etc.;

● Provide high-level ease-of-use interfaces such as Trainer, pipeline, and AutoClass;

●Provide automatic download and loading of preset SOTA weights;

● Support seamless migration and deployment of AI computing centers;

The list of currently supported models is as follows:

modelmodel name
LLama2llama2_7b, llama2_13b, llama2_7b_lora, llama2_13b_lora, llama2_70b
GLM2glm2_6b, glm2_6b_lora
CodeGeex2codegeex2_6b
LLamallama_7b, llama_13b, llama_7b_lora
GLMglm_6b, glm_6b_lora
Bloombloom_560m, bloom_7.1b
GPT2gpt2, gpt2_13b
PanGuAlphapangualpha_2_6_b, pangualpha_13b
BLIP2blip2_stage1_vit_g
CLIPclip_vit_b_32,clip_vit_b_16,clip_vit_l_14,clip_vit_l_14@336
T5t5_small
samsam_vit_b, sam_vit_l, sam_vit_h
MAEmae_vit_base_p16
VITvit_base_p16
Swinswin_base_p4w7
skyworkskywork_13b
Baichuan2baichuan2_7b,baichuan2_13b,baichuan2_7b_lora, baichuan2_13b_lora
Baichuanbaichuan_7b, baichuan_13b
Qwenqwen_7b, qwen_14b, qwen_7b_lora, qwen_14b_lora
Wizardcoderwizardcoder_15b
Internlminternlm_7b,internlm_20b,internlm_7b_lora
ziyaziya_13b
VisualGLMvisualglm

Model Support List

user web-1

user web-1

user web-1

user web-1

user web-1