Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix(test): replace jackfram llama with smollm #2370

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

NanoCode012
Copy link
Collaborator

Description

Replace the old jackfram llama 68m with the smollm 135m due to breaking tests.

CI broken: https://github.com/axolotl-ai-cloud/axolotl/actions/runs/13533778077

Motivation and Context

How has this been tested?

Screenshots (if appropriate)

Types of changes

Social Handles (Optional)

@winglian
Copy link
Collaborator

need to add

                "special_tokens": {
                    "pad_token": "<|endoftext|>",
                },

and remove "model_type": "AutoModelForCausalLM", and "tokenizer_type": "LlamaTokenizer", from the configs

"sequence_len": 1024,
"load_in_8bit": True,
"adapter": "lora",
"lora_r": 8,
"lora_alpha": 16,
"lora_dropout": 0.05,
"lora_target_linear": True,
"lora_modules_to_save": ["lm_head", "embed_tokens"],
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this isn't needed if you correct the pad_token. I'm not a fan of changing the test case for a model swap, but rather we should add a new test case to include lora_modules_to_save. otherwise we risk having regressions down the road

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should also remove the unk_token, bos_tokenm, and eos_token from the configs where you swap this. Would also not swap in an instruct model since the original model you're replacing isn't an instruct model

@NanoCode012 NanoCode012 marked this pull request as draft March 7, 2025 11:54
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants