Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

minicpm-llama3-v2_5模型微调 model_max_length参数 #290

Open
todaydeath opened this issue Jun 21, 2024 · 1 comment
Open

minicpm-llama3-v2_5模型微调 model_max_length参数 #290

todaydeath opened this issue Jun 21, 2024 · 1 comment

Comments

@todaydeath
Copy link

请教下,minicpm-llama3-v2_5模型在微调的时候,图片也算到token里面么?我看微调模型里面model_max_length默认值是2048,如果图片算到token里面且图片比较大的时候,这个值会不会不够啊

@todaydeath todaydeath changed the title 模型微调 minicpm-llama3-v2_5模型微调 model_max_length参数 Jun 21, 2024
@LDLINGLINGLING
Copy link

首先图片将要计算token,但是一般不会超过2048,因为我们的max_slice_num是9,num_query等于96,因此图片最大占用不会超过1000token。但是如果你的文本信息也很大的时候,就会超过这个值。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants