BlueLM 7B — это мощная открытая языковая модель от vivo AI Lab, обученная на колоссальном массиве из 2,6 трлн токенов. ИИ отлично справляется с генерацией кода и сложными диалогами, предлагая версии с контекстным окном до 32K для глубокой работы с данными.
BlueLM is a large-scale open-source language model independently developed by the vivo AI Lab. This release includes 2K and 32K context length versions for both Base and Chat models. High-quality Data: BlueLM is trained on a high-quality data with 2.6 trillion tokens. Our train corpus mainly consists of Chinese and English data, with a small amount of Japanese and Korean data. Stronger Performance: BlueLM-7B-Chat achieves a strong competitive performance in C-Eval and CMMLU benchmarks of the same size. Longer Context: We have extended the context length of both BlueLM-7B-Base-32K and BlueLM-7B-Chat-32K models from 2K to 32K. The models can support longer context understanding while maintaining the same basic capabilities. Model License: BlueLM weights are open for academic research and commercial use.