Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
17.4
TFLOPS
40
36
64
Ilyas Moutawwakil
IlyasMoutawwakil
Follow
Bellamy66's profile picture
ybelkada's profile picture
Metatron696's profile picture
150 followers
ยท
60 following
IlyasMoutawwakil
AI & ML interests
Optimization, LLMs, Hardware, Backends, ..
Recent Activity
replied
to
their
post
1 day ago
After 2 months of refinement, I'm happy to announce that a lot of Transformers' modeling code is now significantly more torch-compile & export-friendly ๐ฅ Why it had to be done ๐ PyTorch's Dynamo compiler is increasingly becoming the default interoperability layer for ML systems. Anything that relies on torch.export or torch.compile, from model optimization to cross-framework integrations, benefits directly when models can be captured as a single dynamo-traced graph ! Transformers models are now easier to: โ๏ธ Compile end-to-end with torch.compile backends ๐ฆ Export reliably via torch.export and torch.onnx.export ๐ Deploy to ONNX / ONNX Runtime, Intel Corporation's OpenVINO, NVIDIA AutoDeploy (TRT-LLM), AMD's Quark, Meta's Executorch and more hardware-specific runtimes. This work aims at unblocking entire TorchDynamo-based toolchains that rely on exporting Transformers across runtimes and accelerators. We are doubling down on Transformers commitment to be a first-class citizen of the PyTorch ecosystem, more exportable, more optimizable, and easier to deploy everywhere. There are definitely some edge-cases that we still haven't addressed so don't hesitate to try compiling / exporting your favorite transformers and to open issues / PRs. PR in the comments ! More updates coming coming soon !
posted
an
update
1 day ago
After 2 months of refinement, I'm happy to announce that a lot of Transformers' modeling code is now significantly more torch-compile & export-friendly ๐ฅ Why it had to be done ๐ PyTorch's Dynamo compiler is increasingly becoming the default interoperability layer for ML systems. Anything that relies on torch.export or torch.compile, from model optimization to cross-framework integrations, benefits directly when models can be captured as a single dynamo-traced graph ! Transformers models are now easier to: โ๏ธ Compile end-to-end with torch.compile backends ๐ฆ Export reliably via torch.export and torch.onnx.export ๐ Deploy to ONNX / ONNX Runtime, Intel Corporation's OpenVINO, NVIDIA AutoDeploy (TRT-LLM), AMD's Quark, Meta's Executorch and more hardware-specific runtimes. This work aims at unblocking entire TorchDynamo-based toolchains that rely on exporting Transformers across runtimes and accelerators. We are doubling down on Transformers commitment to be a first-class citizen of the PyTorch ecosystem, more exportable, more optimizable, and easier to deploy everywhere. There are definitely some edge-cases that we still haven't addressed so don't hesitate to try compiling / exporting your favorite transformers and to open issues / PRs. PR in the comments ! More updates coming coming soon !
liked
a Space
10 days ago
nvidia/kvpress-leaderboard
View all activity
Organizations
IlyasMoutawwakil
's datasets
8
Sort:ย Recently updated
IlyasMoutawwakil/OnnxRuntime-Encoder-Benchmark
Updated
Sep 24, 2025
โข
1
IlyasMoutawwakil/ORT-Bert-Benchmark
Updated
Sep 23, 2025
โข
34
IlyasMoutawwakil/OpenVINO-VLM-Benchmark
Updated
Sep 22, 2025
โข
8
IlyasMoutawwakil/pytorch_gpt2
Updated
Sep 1, 2025
โข
5
IlyasMoutawwakil/benchmarks
Preview
โข
Updated
Dec 12, 2024
โข
10
IlyasMoutawwakil/OpenVINO-Benchmarks
Updated
Nov 18, 2024
โข
2
IlyasMoutawwakil/optimum-benchmarks-ci
Preview
โข
Updated
Apr 10, 2024
โข
10
IlyasMoutawwakil/llm-race-dataset
Viewer
โข
Updated
Nov 23, 2023
โข
4.38M
โข
13
โข
1