New Case Study:See how Anthropic automated 95% of dependency reviews with Socket.Learn More
Socket
Sign inDemoInstall
Socket

intel-extension-for-pytorch

Package Overview
Dependencies
Maintainers
3
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

intel-extension-for-pytorch

Intel® Extension for PyTorch*

  • 2.6.0
  • PyPI
  • Socket score

Maintainers
3

Intel® Extension for PyTorch*

CPU 💻main branch   |   🌱Quick Start   |   📖Documentations   |   🏃Installation   |   💻LLM Example
GPU 💻main branch   |   🌱Quick Start   |   📖Documentations   |   🏃Installation   |   💻LLM Example

Intel® Extension for PyTorch* extends PyTorch* with up-to-date features optimizations for an extra performance boost on Intel hardware. Optimizations take advantage of Intel® Advanced Vector Extensions 512 (Intel® AVX-512) Vector Neural Network Instructions (VNNI) and Intel® Advanced Matrix Extensions (Intel® AMX) on Intel CPUs as well as Intel Xe Matrix Extensions (XMX) AI engines on Intel discrete GPUs. Moreover, Intel® Extension for PyTorch* provides easy GPU acceleration for Intel discrete GPUs through the PyTorch* xpu device.

ipex.llm - Large Language Models (LLMs) Optimization

In the current technological landscape, Generative AI (GenAI) workloads and models have gained widespread attention and popularity. Large Language Models (LLMs) have emerged as the dominant models driving these GenAI applications. Starting from 2.1.0, specific optimizations for certain LLM models are introduced in the Intel® Extension for PyTorch*. Check LLM optimizations for details.

Optimized Model List

MODEL FAMILYMODEL NAME (Huggingface hub)FP32BF16Static quantization INT8Weight only quantization INT8Weight only quantization INT4
LLAMAmeta-llama/Llama-2-7b-hf
LLAMAmeta-llama/Llama-2-13b-hf
LLAMAmeta-llama/Llama-2-70b-hf
LLAMAmeta-llama/Meta-Llama-3-8B
LLAMAmeta-llama/Meta-Llama-3-70B
LLAMAmeta-llama/Meta-Llama-3.1-8B-Instruct
LLAMAmeta-llama/Llama-3.2-3B-Instruct
LLAMAmeta-llama/Llama-3.2-11B-Vision-Instruct
GPT-JEleutherAI/gpt-j-6b
GPT-NEOXEleutherAI/gpt-neox-20b
DOLLYdatabricks/dolly-v2-12b
FALCONtiiuae/falcon-7b
FALCONtiiuae/falcon-11b
FALCONtiiuae/falcon-40b
OPTfacebook/opt-30b
OPTfacebook/opt-1.3b
Bloombigscience/bloom-1b7
CodeGenSalesforce/codegen-2B-multi
Baichuanbaichuan-inc/Baichuan2-7B-Chat
Baichuanbaichuan-inc/Baichuan2-13B-Chat
Baichuanbaichuan-inc/Baichuan-13B-Chat
ChatGLMTHUDM/chatglm3-6b
ChatGLMTHUDM/chatglm2-6b
GPTBigCodebigcode/starcoder
T5google/flan-t5-xl
MPTmosaicml/mpt-7b
Mistralmistralai/Mistral-7B-v0.1
Mixtralmistralai/Mixtral-8x7B-v0.1
Stablelmstabilityai/stablelm-2-1_6b
QwenQwen/Qwen-7B-Chat
QwenQwen/Qwen2-7B
LLaVAliuhaotian/llava-v1.5-7b
GITmicrosoft/git-base
YuanIEITYuan/Yuan2-102B-hf
Phimicrosoft/phi-2
Phimicrosoft/Phi-3-mini-4k-instruct
Phimicrosoft/Phi-3-mini-128k-instruct
Phimicrosoft/Phi-3-medium-4k-instruct
Phimicrosoft/Phi-3-medium-128k-instruct
Whisperopenai/whisper-large-v2
Mairamicrosoft/maira-2
Jambaai21labs/Jamba-v0.1
DeepSeekdeepseek-ai/DeepSeek-V2.5-1210

Note: The above verified models (including other models in the same model family, like "codellama/CodeLlama-7b-hf" from LLAMA family) are well supported with all optimizations like indirect access KV cache, fused ROPE, and customized linear kernels. We are working in progress to better support the models in the tables with various data types. In addition, more models will be optimized in the future.

In addition, Intel® Extension for PyTorch* introduces module level optimization APIs (prototype feature) since release 2.3.0. The feature provides optimized alternatives for several commonly used LLM modules and functionalities for the optimizations of the niche or customized LLMs. Please read LLM module level optimization practice to better understand how to optimize your own LLM and achieve better performance.

Support

The team tracks bugs and enhancement requests using GitHub issues. Before submitting a suggestion or bug report, search the existing GitHub issues to see if your issue has already been reported.

License

Apache License, Version 2.0. As found in LICENSE file.

Security

See Intel's Security Center for information on how to report a potential security issue or vulnerability.

See also: Security Policy

FAQs


Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc