Skip to content

Commit

Permalink
Info on models
Browse files Browse the repository at this point in the history
  • Loading branch information
sgolebiewski-intel committed Nov 12, 2024
1 parent ffd8de4 commit 6dd8f78
Showing 1 changed file with 1 addition and 1 deletion.
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ from about 25GB to 4GB using 4-bit weight compression.
compression may result in more accuracy reduction than with larger models.
Therefore, weight compression is recommended for use with LLMs only.

LLMs and other models that require
LLMs and other GenAI models that require
extensive memory to store the weights during inference can benefit
from weight compression as it:

Expand Down

0 comments on commit 6dd8f78

Please sign in to comment.