We present InternLM, a multilingual foundational language model with 104B parameters. InternLM is pre-trained on a large corpora with 1.6T tokens with a multi-phase progressive process, and then fine-tuned to align with human preferences. We also developed a training system called Uniscale-LLM for efficient large language model training. The evaluation on a number of benchmarks shows that InternLM achieves state-of-the-art performance in multiple aspects, including knowledge understanding, reading comprehension, mathematics, and coding. With such well-rounded capabilities, InternLM achieves outstanding performances on comprehensive exams, including MMLU, AGIEval, C-Eval and GAOKAO-Bench, without resorting to external tools. On these benchmarks, InternLM not only significantly outperforms open-source models, but also obtains superior performance compared to ChatGPT. Also, InternLM demonstrates excellent capability of understanding Chinese language and Chinese culture, which makes it a suitable foundation model to support Chinese-oriented language applications. This manuscript gives a detailed study of our results, with benchmarks and examples across a diverse set of knowledge domains and tasks.
What was learned yesterday is already outdated today?Follow "Playground", Our goal is to provide a central hub for the community to share their innovative solutions and explore the edge of AI technologies.
- Support text-to-image synthesis empowered by Disco-Diffusion.
- Support 3D generation model EG3D (CVPR 2022).
- Support image restoration model NAFNet (ECCV 2022).
Join us to make it better! Try at https://github.com/open-mmlab/mmediting/tree/1.x
The first coherent rotating object detection toolbox
Efficient benchmark model
Modular design and flexible configuration files
Welcome to use and contribute!
- Support different algorithms: NAS, Pruning and KD
- Quickly applied to different CV tasks
- Decouple model and compression algorithm
- Flexible and modular design
If I may make a suggestion, I would encourage you to add one or two lines to your Readme explaining what the project does in clear, non-technical language for people like me.
What does one do with a "Human Parametric Model Toolbox"? Do I give parameters and get an animated person as output? Does it capture my movements from video and turns them into 3D coordinates? If so, how many cameras are needed?
I think the video answers these questions, but nonetheless it could be a good idea to add it to the text.