-
Notifications
You must be signed in to change notification settings - Fork 29
Issues: aws-neuron/transformers-neuronx
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Llama and Mistral model inputs change after using different computation
#104
opened Dec 20, 2024 by
kahfizulkifli
Loading compiled fails:
model_type=bert -> transformers
being used in compiled config.
#102
opened Dec 2, 2024 by
michaelfeil
Gibberish output for princeton-nlp/Sheared-LLaMA-1.3B with continuous batching
#94
opened Jul 15, 2024 by
pinak-p
Latest changes introduced for continuous batching break Mixtral model
bug
Something isn't working
#84
opened Apr 15, 2024 by
dacorvo
Backward compatibility with saved llama 2 compiled artifacts
enhancement
New feature or request
#78
opened Jan 18, 2024 by
dacorvo
User feedback when compiling and reloading a large model
enhancement
New feature or request
#76
opened Jan 17, 2024 by
dacorvo
Infering logits from Improvements or additions to documentation
model.forward
for the entire batch instead of the last forward's output.
documentation
#73
opened Jan 10, 2024 by
michaelfeil
Any solution to save the converted model?
enhancement
New feature or request
#29
opened Aug 14, 2023 by
aliseyfi
Discrepancies Between GPU and Neuron-based Outputs for GPTJ Model on inf2.24xlarge
#28
opened Aug 13, 2023 by
ho4040
ProTip!
Updated in the last three days: updated:>2024-12-22.