Skip to content

Commit

Permalink
add the pretrain code
Browse files Browse the repository at this point in the history
  • Loading branch information
Chengyue Wu committed May 8, 2024
1 parent b658e8c commit a2dc8fb
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,6 @@
📃 <a href="https://arxiv.org/abs/2401.02415" target="_blank">Paper</a> • 🤗 <a href="https://huggingface.co/TencentARC/LLaMA-Pro-8B" target="_blank">Demo & Model</a>
</p>

## TODO List
* Add the pretrain code of [Mistral-Pro-8B-v0.1](https://huggingface.co/TencentARC/Mistral_Pro_8B_v0.1)

## News
* [2024/01/06] We open source the [LLaMA-Pro repository](https://github.com/TencentARC/LLaMA-Pro) and [Demo & Model](https://huggingface.co/TencentARC/LLaMA-Pro-8B).
* [2024/01/07] Add how to run gradio demo locally in [demo](./demo/app.py)
Expand All @@ -14,6 +11,9 @@
![assets/mistral_pro_performance.png](assets/mistral_pro_performance.png)
* [2024/02/23] We release the evaluation code of [Mistral-Pro-8B-v0.1](https://huggingface.co/TencentARC/Mistral_Pro_8B_v0.1) in [lm-evaluation-harness](https://github.com/hills-code/lm-evaluation-harness).
* [2024/02/23] We release [MetaMath-Mistral-Pro](https://huggingface.co/TencentARC/MetaMath-Mistral-Pro) that surpasses previous MetaMath series 7B models at both GSM8k and MATH. The evaluation is following [the official MetaMath repo](https://github.com/meta-math/MetaMath).
* [2024/05/08] Add the pre-train tokenizer in [open-instruct](https://github.com/hills-code/open-instruct/tree/llama-pro).



🔥 Comprehensive Results

Expand Down

0 comments on commit a2dc8fb

Please sign in to comment.