You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently we have cinm.compute with attributes for workgroup shape and DPU memory size.
We assume this specification is correct, that is, the lowering pass cannot change them.
They should be obtained through the cost model.
Implement a simple Samsung dialect
Implement a pass that annotates Samsung and Upmem kernels with their time estimation
Implement the upmem cost estimator in C++
Lower priority
Add verifier for shape of scatter map in UPMEM
Fix the GPU lowering, was probably broken by recent changes to CNM
Optimization
Hoist buffer alloc and free outside of loops
Malloc avoidance
Avoid tensor reshapes that do a copy (Especially for VA that's a problem)
Unify buffers across loop iterations
Affine map simplification with dimension sizes
The text was updated successfully, but these errors were encountered:
High priority
upmem.scatter
is legal, that is, all scattered elements must be contiguous in memory #6Cost model (deadline 04.08)
Currently we have
cinm.compute
with attributes for workgroup shape and DPU memory size.We assume this specification is correct, that is, the lowering pass cannot change them.
They should be obtained through the cost model.
Lower priority
Optimization
The text was updated successfully, but these errors were encountered: