kernel optimization

  • AMD iGPUs Use 128GB Memory on Linux via GTT


    TIL you can allocate 128 GB of unified memory to normal AMD iGPUs on Linux via GTTAMD's integrated GPUs (iGPUs) on Linux can leverage up to 128 GB of system memory as VRAM through a feature called Graphics Translation Table (GTT). This dynamic allocation allows developers to utilize iGPUs for tasks like kernel optimization without impacting the CPU's memory pool until needed. While iGPUs are slower for inference tasks, they offer a cost-effective solution for development and profiling, especially when used alongside a main GPU. This capability is particularly beneficial for those working on hybrid CPU/GPU architectures, enabling efficient memory management and development of large memory AMD GPU kernels. This matters because it opens up new possibilities for affordable and efficient computational development on standard hardware.

    Read Full Article: AMD iGPUs Use 128GB Memory on Linux via GTT

  • Wafer: Streamlining GPU Kernel Optimization in VSCode


    Wafer: VSCode extension to help you develop, profile, and optimize GPU kernelsWafer is a new VS Code extension designed to streamline GPU performance engineering by integrating several tools directly into the development environment. It aims to simplify the process of developing, profiling, and optimizing GPU kernels, which are crucial for improving training and inference speeds in deep learning applications. Traditionally, this workflow involves using multiple fragmented tools and tabs, but Wafer consolidates these functionalities, allowing developers to work more efficiently within a single interface. The extension offers several key features to enhance the development experience. It integrates Nsight Compute directly into the editor, enabling users to run performance analysis and view results alongside their code. Additionally, Wafer includes a CUDA compiler explorer that allows developers to inspect PTX and SASS code mapped back to their source, facilitating quicker iteration on kernel changes. Furthermore, a GPU documentation search feature is embedded within the editor, providing detailed optimization guidance and context to assist developers in making informed decisions. Wafer is particularly beneficial for those involved in training and inference performance work, as it consolidates essential tools and resources into the familiar environment of VS Code. By reducing the need to switch between different applications and tabs, Wafer enhances productivity and allows developers to focus on optimizing their GPU kernels more effectively. This matters because improving GPU performance can significantly impact the efficiency and speed of deep learning models, leading to faster and more cost-effective AI solutions.

    Read Full Article: Wafer: Streamlining GPU Kernel Optimization in VSCode