This is no longer true.
If you use NV’s TensorRT plugin with the A1111 development branch, TensorRT works very well with SDXL (it’s actually much less painful to use than SD1.5 TensorRT was initially).
The big constraint is VRAM capacity. I can use it for 1024x1024 (and similar-total-pixel-count) SDXL generations on my 4090, but can’t go much beyond that without tiling (though that is generally what you do anyway for larger resolutions).
Just like for SD1.5, TensorRT speeds up generation by almost a factor of 2 for SDXL (compared to an “optimized” baseline using SDP).
This is no longer true.
If you use NV’s TensorRT plugin with the A1111 development branch, TensorRT works very well with SDXL (it’s actually much less painful to use than SD1.5 TensorRT was initially).
The big constraint is VRAM capacity. I can use it for 1024x1024 (and similar-total-pixel-count) SDXL generations on my 4090, but can’t go much beyond that without tiling (though that is generally what you do anyway for larger resolutions).
Just like for SD1.5, TensorRT speeds up generation by almost a factor of 2 for SDXL (compared to an “optimized” baseline using SDP).
Alright thanks. This stuff is moving very fast, and I was only looking at the master branch.