-
Notifications
You must be signed in to change notification settings - Fork 26.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
🌐 [i18n-KO] Translated main_classes/quantization.md
to Korean
#33959
base: main
Are you sure you want to change the base?
Conversation
|
||
# 양자화[[quantization]] | ||
|
||
Quantization techniques reduce memory and computational costs by representing weights and activations with lower-precision data types like 8-bit integers (int8). This enables loading larger models you normally wouldn't be able to fit into memory, and speeding up inference. Transformers supports the AWQ and GPTQ quantization algorithms and it supports 8-bit and 4-bit quantization with bitsandbytes. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Quantization techniques reduce memory and computational costs by representing weights and activations with lower-precision data types like 8-bit integers (int8). This enables loading larger models you normally wouldn't be able to fit into memory, and speeding up inference. Transformers supports the AWQ and GPTQ quantization algorithms and it supports 8-bit and 4-bit quantization with bitsandbytes. |
아래 번역 해두신것 같아 원문부분 삭제했습니다~!
|
||
Quantization techniques reduce memory and computational costs by representing weights and activations with lower-precision data types like 8-bit integers (int8). This enables loading larger models you normally wouldn't be able to fit into memory, and speeding up inference. Transformers supports the AWQ and GPTQ quantization algorithms and it supports 8-bit and 4-bit quantization with bitsandbytes. | ||
|
||
Quantization techniques that aren't supported in Transformers can be added with the [`HfQuantizer`] class. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Quantization techniques that aren't supported in Transformers can be added with the [`HfQuantizer`] class. |
아래 번역 해두신것 같아 원문부분 삭제했습니다~!
|
||
<Tip> | ||
|
||
이 [양자화](../quantization) 가이드를 통해서 모델을 양자화하는 방법을 배울 수 있습니다. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
이 [양자화](../quantization) 가이드를 통해서 모델을 양자화하는 방법을 배울 수 있습니다. | |
모델을 양자화하는 방법은 이 [양자화](../quantization) 가이드를 통해 배울 수 있습니다. |
좀더 가독성 있어보이게 바꾸었습니다!
What does this PR do?
Translated the
main_classes/quantization.md
file of the documentation to Korean.Thank you in advance for your review.
Part of #20179
Before reviewing
[[lowercased-header]]
)Who can review? (Initial)
@chhaewxn, @ahnjj, @jun048098, @fabxoe, @nuatmochoi, @heuristicwave
Before submitting
Pull Request section?
to it if that's the case.
documentation guidelines, and
here are tips on formatting docstrings.
Who can review? (Final)