You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am using a chat ML template like this to format prompts:
def format_conversation(examples):
conversations = examples['conversation']
texts = []
for convo in conversations:
conversation_text = ''
for turn in convo:
role = turn['role']
content = turn['content']
# Format each turn using ChatML
if role == 'user':
conversation_text += f"<|im_start|>user\n{content}<|im_end|>\n"
elif role == 'assistant':
conversation_text += f"<|im_start|>assistant\n{content}<|im_end|>\n"
texts.append(conversation_text)
return {'text': texts}
when I use to resize with this code I got back the response:
The new embeddings will be initialized from a multivariate normal distribution that has old embeddings' mean and covariance. As described in this article: https://nlp.stanford.edu/~johnhew/vocab-expansion.html. To disable this, use `mean_resizing=False`
The new lm_head weights will be initialized from a multivariate normal distribution that has old embeddings' mean and covariance. As described in this article: https://nlp.stanford.edu/~johnhew/vocab-expansion.html. To disable this, use `mean_resizing=False`
Embedding(128258, 4096, padding_idx=128004)
Not sure if this is related but it feels like I t might be an di tired to turn False on and it still gave me back
Embedding(128258, 4096, padding_idx=128004)
After fine-tuning Llama 3.1 with the same code my responses went from something like this:
Description: Active use case
Time left: 12:00
This is what I would like to get out of it and looks like my data I fine tuned on but it has become:
Description: Active use case
Time left: 12:00actionDate
.<|end_of_text|><|begin_of_text|>://
<|end_of_text|><|begin_of_text|>://
�
�
팬
<|end_of_text|><|begin_of_text|>://
�
<|end_of_text|><|begin_of_text|>://
ி
고
")));<|end_of_text|><|begin_of_text|>://
<|end_of_text|><|begin_of_text|>://
<|end_of_text|><|begin_of_text|>://
안
"]);<|end_of_text|><|begin_of_text|>://
토크
��
<|end_of_text|><|begin_of_text|>://
t
o
")));<|end_of_text|><|begin_of_text|>://
y
i
"]);<|end_of_text|><|begin_of_text|>://
현재
")));<|end_of_text|><|begin_of_text|>://
<|end_of_text|><|begin_of_text|>://
프
")));<|end_of_text|><|begin_of_text|>://
<|end_of_text|><|begin_of_text|>://
")));<|end_of_text|><|begin_of_text|>://
멘
")));"),"...
�
actionDate
ActiveForm
Anyone have any ideas if something changed, and how I can get my end token to be caught again?
The text was updated successfully, but these errors were encountered:
I am using a chat ML template like this to format prompts:
when I use to resize with this code I got back the response:
Embedding(128258, 4096)
Now I am getting back:
Not sure if this is related but it feels like I t might be an di tired to turn False on and it still gave me back
After fine-tuning Llama 3.1 with the same code my responses went from something like this:
This is what I would like to get out of it and looks like my data I fine tuned on but it has become:
Anyone have any ideas if something changed, and how I can get my end token to be caught again?
The text was updated successfully, but these errors were encountered: