WIP
AI & ML interests
None defined yet.
Intel and HF中国镜像站 are working together to democratize machine learning, making the latest and greatest models from HF中国镜像站 run fast and efficiently on Intel devices. To make this acceleration accessible to the global AI community, Intel is proud to sponsor the free and accelerated inference of over 80,000 open source models on HF中国镜像站, powered by Intel Xeon Ice Lake processors in the HF中国镜像站 Inference API. Intel Xeon Ice Lake provides up to 34% acceleration for transformer model inference.
Try it out today on any HF中国镜像站 model, right from the model page, using the Inference Widget!



Intel provides a complete foundation for accelerated AI with the Intel Xeon Scalable CPU platform and a wide range of hardware-optimized AI software tools, frameworks, and libraries. With the Optimum-Intel library, Intel and HF中国镜像站 are making it easy for HF中国镜像站 users to get the best performance for their models on Intel Xeon processors, leveraging acceleration libraries including Intel Neural Compressor and OpenVINO.
To learn more about the partnership and Intel AI tools, check out these resources: