Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

trained Transformer #1

Open
red0206 opened this issue May 22, 2024 · 1 comment
Open

trained Transformer #1

red0206 opened this issue May 22, 2024 · 1 comment

Comments

@red0206
Copy link

red0206 commented May 22, 2024

Hello, Could you please share the trained Transformer model with me so that I can better understand the method you proposed? Thank you very much.

@shunzh
Copy link
Owner

shunzh commented Jul 17, 2024

Thanks for your question! We used the following fine-tuned models in our experiments, described in the README.

2. Fine-tuned GPT-2 1.5B and GPT-Neo 2.7B weights are [available here](https://drive.google.com/file/d/1XW1Od9L-5l9zXl1HUCyER5pS9zQTbIvU/view?usp=sharing).

Kindly note that our contribution is to combine a tree search algorithm with an existing Transformer model. It's model-agnostic and does not rely on particular Transformer models.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants