5 open-source LLMs #2

We collected the links to all research papers and repositories with the code.

  • NLLB is a series of open-source models capable of delivering high-quality translations directly between any pair of 200+ languages

  • GLM-130B is an open bilingual (English & Chinese) bidirectional dense model with 130 billion parameters.

  • RWKV is an RNN with Transformer-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable).

  • Flan-T5

  • Galactica is a general-purpose scientific language model. It is trained on a large corpus of scientific text and data.

Every day we post helpful lists and bite-sized explanations on our Twitter. Please join us there!

Join the conversation

or to participate.