AI & Technology Updates

  • FCC Halts Smart Home Security Certification Plan


    The FCC has probably killed a plan to improve smart home securityThe US Cyber Trust Mark Program, designed to certify smart home devices for cybersecurity standards, is facing uncertainty after UL Solutions, its lead administrator, stepped down. This decision follows an investigation by the Federal Communications Commission (FCC) into the program's connections with China. The program, which was intended to provide a recognizable certification similar to the Energy Star label, has not yet been officially terminated but remains in a state of limbo. This development is part of a broader trend of the FCC rolling back cybersecurity initiatives, including recent changes to telecom regulations and the decertification of certain testing labs. Why this matters: The potential demise of the US Cyber Trust Mark Program highlights challenges in establishing robust cybersecurity standards for smart home devices, which are increasingly integral to daily life.


  • VCs Predict AI Spending Shift in 2026


    VCs predict enterprises will spend more on AI in 2026 — through fewer vendorsEnterprises are expected to significantly increase their AI budgets by 2026, but this spending will be focused on fewer vendors and specific AI products that demonstrate clear results. Investors predict a shift from experimentation with multiple AI tools to a consolidation of investments in proven technologies, with enterprises concentrating on strengthening data foundations, optimizing models, and consolidating tools. This trend may lead to a narrowing of the enterprise AI landscape, where only a few vendors capture a large share of the market, while many startups face challenges unless they offer unique, hard-to-replicate solutions. As enterprises prioritize AI tools that ensure safety and deliver measurable ROI, startups with proprietary data and distinct products may still thrive, but those similar to large suppliers might struggle. This matters because it signals a major shift in enterprise AI investment strategies, potentially reshaping the competitive landscape and impacting the viability of many AI startups.


  • Exploring Direct Preference Optimization (DPO)


    Following up on my PPO derivation – I worked through DPO (Direct Preference Optimization) from first principlesDirect Preference Optimization (DPO) offers a streamlined and efficient method for aligning large language models (LLMs) with human preferences, bypassing the complexities of traditional reinforcement learning approaches like PPO (Proximal Policy Optimization). Unlike PPO, which involves a multi-component objective and a complex loop of reward modeling and sampling, DPO simplifies the process by directly optimizing a supervised objective on preference pairs through gradient descent. This approach eliminates the need for separate reward model training and the intricate PPO clipping process, making it a more approachable and computationally lightweight alternative. Understanding DPO is crucial as it provides a more straightforward and efficient way to enhance AI models' alignment with human values and preferences.


  • Softbank’s $40B Investment in OpenAI


    Softbank has fully funded $40 billion investment in OpenAI, sources tell CNBCSoftbank has reportedly completed a $40 billion investment in OpenAI, a significant move that underscores the growing interest and financial backing in artificial intelligence technologies. This investment aims to bolster OpenAI's development and deployment of cutting-edge AI systems, potentially accelerating advancements in the field. The funding highlights the strategic importance placed on AI by major global investors, reflecting the transformative potential AI holds for various industries. This matters as it showcases the increasing commitment of financial giants to AI, which could drive innovation and shape the future of technology.


  • Training AI Co-Scientists with Rubric Rewards


    Training AI Co-Scientists using Rubric RewardsMeta has introduced a scalable method to train AI systems to aid scientists in reaching their research objectives by leveraging large language models (LLMs) to extract research goals and grading rubrics from scientific literature. These rubrics are then used in reinforcement learning (RL) training, where the AI self-grades its progress to bridge the generator-verifier gap. Fine-tuning the Qwen3-30B model with this self-grading approach has shown to enhance research plans for 70% of machine learning goals, achieving results comparable to Grok-4-Thinking, though GPT-5-Thinking remains superior. This approach also demonstrates significant cross-domain generalization, supporting the potential of AI as versatile co-scientists. This matters because it highlights the potential for AI to significantly enhance scientific research processes across various domains.