Skip to content

fix: checkpoint crash and sequence truncation in forward()#4

Open
thesohamdatta wants to merge 1 commit intoOpenVanguard:mainfrom
thesohamdatta:fix/checkpoint-signature-and-sequence-truncation
Open

fix: checkpoint crash and sequence truncation in forward()#4
thesohamdatta wants to merge 1 commit intoOpenVanguard:mainfrom
thesohamdatta:fix/checkpoint-signature-and-sequence-truncation

Conversation

@thesohamdatta
Copy link
Copy Markdown

save_checkpoint() was missing the path param that train_custom.py passes. TransformerModel.forward() clamped T but never sliced x, so tok_emb and pos_emb shapes didn't match on inputs longer than block_size.

save_checkpoint() was missing the path param that train_custom.py passes.
TransformerModel.forward() clamped T but never sliced x, so tok_emb and
pos_emb shapes didn't match on inputs longer than block_size.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant