Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
wenhua cheng's picture
19 6 22

wenhua cheng

wenhuach
serajsreepur's profile picture Pent's profile picture Ravishka26's profile picture
Β·
  • wenhuach21

AI & ML interests

Model Compression, CV

Recent Activity

updated a model 4 days ago
Intel/Qwen3-8B-GGUF-Q2KS-AS-AutoRound
published a model 4 days ago
Intel/Qwen3-8B-GGUF-Q2KS-AS-AutoRound
reacted to their post with πŸš€ 4 days ago
AutoRound keeps evolving its LLM quantization algorithm! πŸš€ After enhancing W2A16 quantization, we now offer a fast algorithm to generate mixed bits/data-type schemes (~2mins for 8B models), great for MXFP4 and W2A16. Learn more: https://github.com/intel/auto-round/blob/main/docs/step_by_step.md#autoscheme
View all activity

Organizations

Intel's profile picture Need4Speed's profile picture Qwen's profile picture

authored 2 papers about 2 years ago

TEQ: Trainable Equivalent Transformation for Quantization of LLMs

Paper β€’ 2310.10944 β€’ Published Oct 17, 2023 β€’ 10

Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs

Paper β€’ 2309.05516 β€’ Published Sep 11, 2023 β€’ 10
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs