Announcing MPT-7B-8K: 8K Context Length for Document Understanding


AI summary: The MPT-7B-8K, a 7B parameter open-source language learning model (LLM) with an 8k context length, has been released. Trained on the MosaicML platform, it specializes in document summarization and question-answering. The release includes three models: MPT-7B-8k, MPT-7B-8k-Instruct, and MPT-7B-8k-Chat, each designed for specific tasks. The models are optimized for faster training and inference, and can be fine-tuned on domain-specific data. They outperform other open-source 8K context length models in in-context learning evaluations.
Read more…