Just-Go-Parallel (Parallel Distributed)

The model repository for the "Parallel Distributed" setting of the following paper:

Just Go Parallel: Improving the Multilingual Capabilities of Large Language Models

Muhammad Reza Qorib, Junyi Li, and Hwee Tou Ng

The 63rd Annual Meeting of the Association for Computational Linguistics (to appear)

We use the architecture and tokenizer of TinyLlama v1.1. Please use transformers>=4.35.

Models

The main branch of the repository contains the best-performing model that was evaluated in the paper. Other checkpoints produced during training are also hosted in this repository under different branch names (also called "revisions" in HuggingFace), with each branch name referring to the number of training steps.

Downloads last month
5
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support