Skip to content

Support for loading in FP8 mode#138

Open
Yahweasel wants to merge 1 commit intoByteDance-Seed:mainfrom
Yahweasel:main
Open

Support for loading in FP8 mode#138
Yahweasel wants to merge 1 commit intoByteDance-Seed:mainfrom
Yahweasel:main

Conversation

@Yahweasel
Copy link
Copy Markdown

Unfortunately, the 7900XTX and other consumer-grade AMD GPUs don't really support bitsandbytes. Or dfloat11. But, FP8 support works fine. With this patch, I can load and run BAGEL on a 7900XTX. Unfortunately, it still takes about 17GB at max, so it's too much for a 16GB card, but it does work, and it's a heck of a lot faster than CPU.

Use --mode 4 to load in FP8 mode. Loads the FP16 model with on-the-fly quantization. Because most math isn't supported in FP8, the weights are changed dynamically as needed, so a few checks are added to perform the upconversion.

Loads the FP16 model with on-the-fly quantization. Because most math
isn't supported in FP8, the weights are changed dynamically as needed,
so a few checks are added to perform the upconversion.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant