I think we can provide users with clear guidance on where to find the best-optimized and quantized versions of open weights large language models (LLMs) on Hugging Face. Many developers are releasing quantized versions of popular LLMs to improve performance and efficiency, and it's crucial to help users find these resources easily.
Not Many have powerful GPU's to run even fp16 8b models so they will be truly disappointed if they dont know about quantized models and what and from where to download exactly so you can guide them to either some video plus suggest some popular profiles to use for downloading quantized models from hugginface like i use majorly bartowski quantized models - https://huggingface.co/bartowski