Detailed Explanation of GLM-130B: An Open Bilingual Pre-trained Model

Detailed Explanation of GLM-130B: An Open Bilingual Pre-trained Model

Source: Contribution Author: Mao Huaqing Editor: Xuejie Table of Contents Related Knowledge GPT BERT T5 Summary Background Introduction Main Contributions and Innovations GLM 6B Custom Mask Model Quantization 1TB Bilingual Instruction Fine-tuning RLHF PEFT Training Strategy Model Parameters Six Metrics Other Evaluation Results Environment Preparation Running Invocation Code Invocation Web Service Command Line Invocation Model … Read more