by artagnon on 12/17/23, 7:18 PM with 14 comments
by brucethemoose2 on 12/18/23, 2:12 AM
By that time we will have a good number of MI300 hosts. AMD Strix Halo (and the Intel equivalent?) will be out for high memory jobs locally. Intel Falcon Shores and who knows will finally be coming out, and from the looks of it the software ecosystem will be at least a little more hardware agnostic.
by CalChris on 12/18/23, 5:33 AM
How is that an abstraction? It sounds more like a representation.
by viksit on 12/18/23, 4:33 AM
(have worked extensively with tf / pytorch)
by JonChesterfield on 12/18/23, 3:58 PM
by westurner on 12/17/23, 11:40 PM
> XeGPU dialect models a subset of Xe GPU’s unique features focusing on GEMM performance. The operations include 2d load, dpas, atomic, scattered load, 1d load, named barrier, mfence, and compile-hint. These operations provide a minimum set to support high-performance MLIR GEMM implementation for a wide range of GEMM shapes. XeGPU dialect complements Arith, Math, Vector, and Memref dialects. This allows XeGPU based MLIR GEMM implementation fused with other operations lowered through existing MLIR dialects.
by KingLancelot on 12/17/23, 8:06 PM
Accelerators already have a common middle layer.
https://discourse.llvm.org/t/rfc-introducing-llvm-project-of...
by gardenfelder on 12/17/23, 7:37 PM