Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

minor fix in mixtral_moe.py for apple silicon #170

Merged
merged 3 commits into from
Feb 23, 2024

Conversation

hannibalhuang
Copy link
Contributor

There are two fixes that address errors encountered when ran on Apple M2 macbook with Rosetta enabled, not sure if it is a universal problem or specific to my environment (therefore I didn't file an issue due to this uncertainty).

  • Fix 1 : cast str type for tokenizer.bos_token, this aims for a fix of persistent weird error "typeError: unsupported operand type(s) for +: 'NoneType' and 'str'" although bos_token is indeed str type.

  • Fix 2 : add pad_token value, per suggestion of padding_strategy raised ValueError "Asking to pad but the tokenizer does not have a padding token. " (Line 2708 in transformers/tokenization_utils_base.py)

There are two fixes that address errors encountered when ran on Apple M2 macbook with Rosetta enabled
Copy link
Collaborator

@cg123 cg123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the PR - I appreciate it! Put some minor comments up.

@@ -113,7 +113,7 @@ def tokenize_prompts(
prompts: List[str], tokenizer: transformers.PreTrainedTokenizerBase
):
return tokenizer(
[tokenizer.bos_token + p for p in prompts],
[str(tokenizer.bos_token) + p for p in prompts],
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It probably would be better to do tokenizer.bos_token or "" here - otherwise it'll be prepending the string "None" to your prompts if the BOS token isn't set.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ok done

@@ -366,6 +366,7 @@ def build(
)
tokenizer.padding_side = "left"
tokenizer.pad_token_id = tokenizer.bos_token_id
tokenizer.pad_token = tokenizer.eos_token
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It'd be great to make this conditional - if the pad token is already set it'd be good to not overwrite it.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

@cg123 cg123 merged commit 5040ac7 into arcee-ai:main Feb 23, 2024
4 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants