Add support for Llama 3.1 8B#655
Conversation
|
👋 Hi! Thank you for contributing to vLLM support on Spyre. We also recommend installing prek and configuring it to check your code before every local commit. |
Signed-off-by: AnishPahilajani <anishhp13@gmail.com>
ec15912 to
de87017
Compare
|
|
||
| # Log once upfront that we detected the model | ||
| logger.info( | ||
| "Llama 3.1 8b dense model with tensor parallel size 4 detected. " |
There was a problem hiding this comment.
Is this a dense model?
Signed-off-by: AnishPahilajani <anishhp13@gmail.com>
| ) | ||
|
|
||
| @classmethod | ||
| def configure_llama_3_1_8b(cls, vllm_config: VllmConfig): |
There was a problem hiding this comment.
It looks like this is all copy-pasted, which I would rather not do. I think @tjohnson31415 has been working on cleaning this up a bit to be more reusable, we should sync up on that
|
RE: #655 (comment) After the configuration refactor is complete and merged from #669, adding new model support should just be a bit of YAML: TODO: could add other CB configs or tune numbers |
Description
Added support for Llama 3.1 8B model