local AGI

  • Deepseek v3.2 on 16 AMD MI50 GPUs: Efficient AI Setup


    16x AMD MI50 32GB at 10 t/s (tg) & 2k t/s (pp) with Deepseek v3.2 (vllm-gfx906)Deepseek v3.2 has been optimized to run on a setup of 16 AMD MI50 32GB GPUs, achieving a token generation speed of 10 tokens per second and prompt processing speed of 2000 tokens per second. This configuration is designed to be cost-effective, with a power draw of 550W when idle and 2400W at peak inference, offering a viable alternative to expensive CPU hardware as RAM prices increase. The setup aims to facilitate the development of local artificial general intelligence (AGI) without incurring costs exceeding $300,000. The open-source community has been instrumental in this endeavor, and future plans include expanding the setup to 32 GPUs for enhanced performance. Why this matters: This development provides a more affordable and efficient approach to running advanced AI models, potentially democratizing access to powerful computational resources.

    Read Full Article: Deepseek v3.2 on 16 AMD MI50 GPUs: Efficient AI Setup