/

/

Announcing cuTile.jl v0.2 With New Features and Improved Performance

Announcing cuTile.jl v0.2 With New Features and Improved Performance

Announcing cuTile.jl v0.2 With New Features and Improved Performance

Date Published

Contributors

Share

Date Published

Contributors

Share

cuTile.jl v0.2, the first major update of the Julia package for writing GPU kernels using NVIDIA's CUDA model is available now. Writing high-performance GPU kernels has traditionally required deep expertise and low-level optimization. But with tools like cuTile.jl, that process is becoming more accessible, expressive, and powerful for Julia developers.

In its latest release, cuTile.jl v0.2 marks a major step forward, bringing significant performance improvements, expanded language support, and a more intuitive programming experience for developers working with GPU workloads.

One of the most notable highlights is the introduction of native Julia for loops, eliminating the need for workarounds and making kernel code more natural to write and read. The update also introduces a new ct.@fpmode macro, giving developers fine-grained control over floating-point behavior.

Beyond syntax improvements, the release pushes usability further with keyword-based APIs, enhanced debugging using standard print and println, and experimental host-level abstractions that can automatically generate tiled GPU kernels, reducing the need to write explicit kernel code.

Performance is another major focus. A redesigned multi-pass optimization pipeline dramatically improves the generated code, enabling better instruction-level parallelism, smarter memory handling, and more efficient arithmetic transformations. 

To showcase these capabilities, the release includes additional advanced examples such as Mixture of Experts kernels and Flash Multi-Head Attention. Performance of these examples is on par with cuTile Python, demonstrating how cuTile.jl can handle modern AI and HPC workloads.

If you're working with GPU computing in Julia or exploring more efficient ways to build high-performance kernels, this release is worth a closer look.

Read the full blog to dive deeper into the features, benchmarks, and what’s next for cuTile.jl.

Authors

JuliaHub, formerly Julia Computing, was founded in 2015 by the four co-creators of Julia (Dr. Viral Shah, Prof. Alan Edelman, Dr. Jeff Bezanson and Stefan Karpinski) together with Deepak Vinchhi and Keno Fischer. Julia is the fastest and easiest high productivity language for scientific computing. Julia is used by over 10,000 companies and over 1,500 universities. Julia’s creators won the prestigious James H. Wilkinson Prize for Numerical Software and the Sidney Fernbach Award.

Authors

JuliaHub, formerly Julia Computing, was founded in 2015 by the four co-creators of Julia (Dr. Viral Shah, Prof. Alan Edelman, Dr. Jeff Bezanson and Stefan Karpinski) together with Deepak Vinchhi and Keno Fischer. Julia is the fastest and easiest high productivity language for scientific computing. Julia is used by over 10,000 companies and over 1,500 universities. Julia’s creators won the prestigious James H. Wilkinson Prize for Numerical Software and the Sidney Fernbach Award.

Authors

JuliaHub, formerly Julia Computing, was founded in 2015 by the four co-creators of Julia (Dr. Viral Shah, Prof. Alan Edelman, Dr. Jeff Bezanson and Stefan Karpinski) together with Deepak Vinchhi and Keno Fischer. Julia is the fastest and easiest high productivity language for scientific computing. Julia is used by over 10,000 companies and over 1,500 universities. Julia’s creators won the prestigious James H. Wilkinson Prize for Numerical Software and the Sidney Fernbach Award.

Learn about Dyad

Get Dyad Studio – Download and install the IDE to start building hardware like software.

Read the Dyad Documentation – Dive into the language, tools, and workflow.

Join the Dyad Community – Connect with fellow engineers, ask questions, and share ideas.

Contact Us

Want to get enterprise support, schedule a demo, or learn about how we can help build a custom solution? We are here to help.