AI & Technology Updates

  • Plamo3 Support Merged into llama.cpp


    Plamo3 (2B/8B/31B) support has been merged into llama.cppPLaMo 3 NICT 31B Base is a sophisticated language model developed through a collaboration between Preferred Networks, Inc. and the National Institute of Information and Communications Technology (NICT). It is pre-trained on both English and Japanese datasets, showcasing a hybrid architecture that combines Sliding Window Attention (SWA) with traditional attention layers. This integration into llama.cpp signifies an advancement in multilingual model capabilities, enhancing the potential for more nuanced and context-aware language processing. This matters because it represents a significant step forward in creating more versatile and powerful language models that can handle complex linguistic tasks across multiple languages.


  • Advancements in Local LLMs and MoE Models


    original KEFv3.2 link, v4.1 with mutation parameter , test it , puplic domain, freewareSignificant advancements in the local Large Language Model (LLM) landscape have emerged in 2025, with notable developments such as the dominance of llama.cpp due to its superior performance and integration with Llama models. The rise of Mixture of Experts (MoE) models has allowed for efficient running of large models on consumer hardware, balancing performance and resource usage. New local LLMs with enhanced vision and multimodal capabilities are expanding the range of applications, while Retrieval-Augmented Generation (RAG) is being used to simulate continuous learning by integrating external knowledge bases. Additionally, investments in high-VRAM hardware are enabling the use of larger and more complex models on consumer-grade machines. This matters as it highlights the rapid evolution of AI technology and its increasing accessibility to a broader range of users and applications.


  • Farewell to ChatGPT After Two Years


    After almost 2 years, it's time to say goodbye 😢After nearly two years of use, the decision has been made to discontinue the subscription to OpenAI's ChatGPT due to the inability to justify the monthly fee. Despite a positive experience and gratitude towards OpenAI and ChatGPT, the availability of superior products from competitors has influenced the decision to switch, even at a higher cost. The farewell is heartfelt, with appreciation for the contributions made by ChatGPT, but the current landscape necessitates moving on. This matters as it highlights the competitive nature of AI services and the importance of evolving to meet user needs and preferences.


  • Lovable Integration in ChatGPT: A Developer’s Aid


    The new Lovable integration in ChatGPT is the closest thing to "Agent Mode" I’ve seen yetThe new Lovable integration in ChatGPT represents a significant advancement in the model's ability to handle complex tasks autonomously. Unlike previous iterations that simply provided code, this integration allows the model to act more like a developer, making decisions such as creating an admin dashboard for lead management without explicit prompts. It demonstrates improved reasoning capabilities, integrating features like property filters and map sections seamlessly. However, the process requires transitioning to the Lovable editor for detailed adjustments, as updates cannot be directly communicated back into the live build from the GPT interface. This development compresses the initial stages of a development project significantly, showcasing a promising step towards more autonomous AI-driven workflows. This matters because it enhances the efficiency and capability of AI in handling complex, multi-step tasks, potentially transforming how development projects are initiated and managed.


  • NVIDIA’s NitroGen: AI Model for Gaming Agents


    NVIDIA AI Researchers Release NitroGen: An Open Vision Action Foundation Model For Generalist Gaming AgentsNVIDIA's AI research team has introduced NitroGen, a groundbreaking vision action foundation model designed for generalist gaming agents. NitroGen learns to play commercial games directly from visual data and gamepad actions, utilizing a vast dataset of 40,000 hours of gameplay from over 1,000 games. The model employs a sophisticated action extraction pipeline to convert video data into actionable insights, enabling it to achieve significant task completion rates across various gaming genres without reinforcement learning. NitroGen's unified controller action space allows for seamless policy transfer across multiple games, demonstrating improved performance when fine-tuned on new titles. This advancement matters because it showcases the potential of AI to autonomously learn complex tasks from large-scale, diverse data sources, paving the way for more versatile and adaptive AI systems in gaming and beyond.