From 7b80c2ea5d1dd4d5d690694bff80d8426b639811 Mon Sep 17 00:00:00 2001 From: chensuyue Date: Fri, 14 Mar 2025 16:53:35 +0800 Subject: [PATCH 1/3] update publication_list.md Signed-off-by: chensuyue --- docs/source/publication_list.md | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/source/publication_list.md b/docs/source/publication_list.md index 014ee45d6fa..a3a7cef7134 100644 --- a/docs/source/publication_list.md +++ b/docs/source/publication_list.md @@ -1,7 +1,8 @@ -Full Publications/Events (87) +Full Publications/Events (88) ========== -## 2025 (1) +## 2025 (2) * Blog on SqueezeBits: [[Intel Gaudi] #4. FP8 Quantization](https://blog.squeezebits.com/intel-gaudi-4-fp8-quantization--40269) (Jan 2025) +* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (March 2025) ## 2024 (7) * Blog by Microsoft: [Phi-4 quantization and inference speedup](https://techcommunity.microsoft.com/blog/machinelearningblog/phi-4-quantization-and-inference-speedup/4360047) (Dec 2024) From c2fb2d3b0e74d088067ded734b6832119fc81a6e Mon Sep 17 00:00:00 2001 From: chensuyue Date: Mon, 17 Mar 2025 11:54:05 +0800 Subject: [PATCH 2/3] update publication_list.md Signed-off-by: chensuyue --- docs/source/publication_list.md | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/docs/source/publication_list.md b/docs/source/publication_list.md index a3a7cef7134..790b9170fdc 100644 --- a/docs/source/publication_list.md +++ b/docs/source/publication_list.md @@ -1,8 +1,9 @@ -Full Publications/Events (88) +Full Publications/Events (89) ========== -## 2025 (2) +## 2025 (3) +* arXiv: [Faster Inference of LLMs using FP8 on the Intel Gaudi](https://arxiv.org/abs/2503.09975) (Mar 2025) +* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (Mar 2025) * Blog on SqueezeBits: [[Intel Gaudi] #4. FP8 Quantization](https://blog.squeezebits.com/intel-gaudi-4-fp8-quantization--40269) (Jan 2025) -* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (March 2025) ## 2024 (7) * Blog by Microsoft: [Phi-4 quantization and inference speedup](https://techcommunity.microsoft.com/blog/machinelearningblog/phi-4-quantization-and-inference-speedup/4360047) (Dec 2024) From 6af1f76d9a593a439a135dca4a4b2f1cacac4458 Mon Sep 17 00:00:00 2001 From: chensuyue Date: Mon, 17 Mar 2025 11:56:52 +0800 Subject: [PATCH 3/3] update publication in README.md Signed-off-by: chensuyue --- README.md | 8 +++----- 1 file changed, 3 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index 9f31f24fae6..1d0f0c35580 100644 --- a/README.md +++ b/README.md @@ -173,12 +173,10 @@ model = load( ## Selected Publications/Events +* arXiv: [Faster Inference of LLMs using FP8 on the Intel Gaudi](https://arxiv.org/abs/2503.09975) (Mar 2025) +* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (Mar 2025) +* Blog on SqueezeBits: [[Intel Gaudi] #4. FP8 Quantization](https://blog.squeezebits.com/intel-gaudi-4-fp8-quantization--40269) (Jan 2025) * EMNLP'2024: [Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs](https://arxiv.org/abs/2309.05516) (Sep 2024) -* Blog on Medium: [Quantization on Intel Gaudi Series AI Accelerators](https://medium.com/intel-analytics-software/intel-neural-compressor-v3-0-a-quantization-tool-across-intel-hardware-9856adee6f11) (Aug 2024) -* Blog by Intel: [Neural Compressor: Boosting AI Model Efficiency](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Neural-Compressor-Boosting-AI-Model-Efficiency/post/1604740) (June 2024) -* Blog by Intel: [Optimization of Intel AI Solutions for Alibaba Cloud’s Qwen2 Large Language Models](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) (June 2024) -* Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html) (Apr 2024) -* EMNLP'2023 (Under Review): [TEQ: Trainable Equivalent Transformation for Quantization of LLMs](https://openreview.net/forum?id=iaI8xEINAf&referrer=%5BAuthor%20Console%5D) (Sep 2023) * arXiv: [Efficient Post-training Quantization with FP8 Formats](https://arxiv.org/abs/2309.14592) (Sep 2023) * arXiv: [Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs](https://arxiv.org/abs/2309.05516) (Sep 2023)