You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
With the tensor subclass guide published in #391, FP6-LLM implementation, which is currently a Linear module replacement, should migrate to tensor subclass too.
The text was updated successfully, but these errors were encountered:
* Revert "Revert "Embedding quantization per backend (pytorch#402)" (pytorch#411)"
This reverts commit 8b35acdff4fded779799ab8a419e55f885dd8918.
* 4b and 8b embedding table quantization
* minor changes
* remove extra et workflow
yanbing-j
pushed a commit
to yanbing-j/ao
that referenced
this issue
Dec 9, 2024
With the tensor subclass guide published in #391, FP6-LLM implementation, which is currently a Linear module replacement, should migrate to tensor subclass too.
The text was updated successfully, but these errors were encountered: