Proposal to add stream/queue as an optional argument to few GPU dialect ops

I wasn’t necessarily suggesting to punt completely but to isolate. Those aspects are far from standard in all of the programming models, and in MLIR, we usually model such differences with dialects vs lowest common denominator. There’s also substantial priors for these items just coming from a related runtime library with some form of handle interop.

Not very well, I suspect, but these are already at two different levels of abstraction. Would need to check with the vulkan folks regarding how much of the GPU dialect really matches that (vs just happening to have some correspondence that can be used in simple cases).