gpt-oss

  • Advancements in Llama AI Technology 2025-2026


    39C3 - 51 Ways to Spell the Image Giraffe: The Hidden Politics of Token Languages in Generative AIIn 2025 and early 2026, significant advancements in Llama AI technology have been marked by the maturation of open-source Vision-Language Models (VLMs), which are anticipated to be widely productized by 2026. Mixture of Experts (MoE) models have gained popularity, with users now operating models with 100-120 billion parameters, a significant increase from the previous year's 30 billion. Z.ai has emerged as a key player with models optimized for inference, while OpenAI's GPT-OSS has been lauded for its tool-calling capabilities. Additionally, Alibaba has expanded its offerings with a variety of models, and coding agents have demonstrated the undeniable potential of generative AI. This matters because these advancements reflect the rapid evolution and diversification of AI technologies, influencing a wide range of applications and industries.

    Read Full Article: Advancements in Llama AI Technology 2025-2026

  • Optimizing LLMs for Efficiency and Performance


    My opinion on some trending topics about LLMsLarge Language Models (LLMs) are being optimized for efficiency and performance across various hardware setups. The best model sizes for running high-quality, fast responses are 7B-A1B, 20B-A3B, and 100-120B MoEs, which are compatible with a range of GPUs. While the "Mamba" model design saves context space, it does not match the performance of fully transformer-based models in agentic tasks. The MXFP4 architecture, supported by mature software like GPT-OSS, offers a cost-effective way to train models by allowing direct distillation and efficient use of resources. This approach can lead to models that are both fast and intelligent, providing an optimal balance of performance and cost. This matters because it highlights the importance of model architecture and software maturity in achieving efficient and effective AI solutions.

    Read Full Article: Optimizing LLMs for Efficiency and Performance

  • Llama.cpp: Native mxfp4 Support Boosts Speed


    llama.cpp, experimental native mxfp4 support for blackwell (25% preprocessing speedup!)The recent update to llama.cpp introduces experimental native mxfp4 support for Blackwell, resulting in a 25% preprocessing speedup compared to the previous version. While this update is currently 10% slower than the master version, it shows significant promise, especially for gpt-oss models. To utilize this feature, compiling with the flag -DCMAKE_CUDA_ARCHITECTURES="120f" is necessary. Although there are some concerns about potential correctness issues due to the quantization of activation to mxfp4 instead of q8, initial tests indicate no noticeable quality degradation in models like gpt-oss-120b. This matters because it enhances processing efficiency, potentially leading to faster and more efficient AI model training and deployment.

    Read Full Article: Llama.cpp: Native mxfp4 Support Boosts Speed