developer tools

  • Pagesource: CLI Tool for Web Dev with LLM Context


    Pagesource - CLI tool to dump website runtime sources for local LLM contextPagesource is a command-line tool designed to capture and dump the runtime sources of a website, providing a more accurate representation of the site's structure for local language model (LLM) context. Unlike the traditional "Save As" feature in browsers that flattens the webpage into a single HTML file, Pagesource preserves the actual file structure, including separate JavaScript modules, CSS files, and lazy-loaded resources. Built on Playwright, it allows developers to access all dynamically loaded JS modules and maintain the original directory structure, making it particularly useful for web developers who need to replicate or analyze website components effectively. This matters because it enhances the ability to work with LLMs by providing them with a more detailed and accurate context of web resources.

    Read Full Article: Pagesource: CLI Tool for Web Dev with LLM Context

  • Enhance Streaming, Coding & Browsing with Chrome Extensions


    I Built 4 Chrome Extensions to Improve Streaming, Coding & BrowsingNikaOrvion has developed four innovative Chrome extensions aimed at enhancing streaming, coding, and browsing experiences while maintaining user privacy. The Auto High Quality extension ensures the highest video quality on platforms like YouTube and Netflix, while DevFontX allows developers to customize coding fonts directly in the browser. The Global Loading Progress Bar provides a customizable loading bar for all websites, and Seamless PDF converts Jupyter Notebooks into high-quality PDFs. These tools focus on performance, privacy, and usability, offering valuable enhancements for productivity and web experiences. Why this matters: These extensions provide practical solutions for improving digital workflows, enhancing both user experience and productivity while prioritizing privacy.

    Read Full Article: Enhance Streaming, Coding & Browsing with Chrome Extensions

  • Git-aware File Tree & Search in Jupyter Lab


    Modern Git-aware File Tree and global search/replace in JupyterA new extension for Jupyter Lab enhances its functionality by adding a Git-aware file tree and a global search/replace feature. The file explorer sidebar now includes Git status colors and icons, marking files based on their Git status such as uncommitted modifications or ignored files. Additionally, the global search and replace tool works across all file types, including Jupyter notebooks, while automatically skipping ignored files like virtual environments or node modules. This matters because it brings Jupyter Lab closer to the capabilities of modern editors like VSCode, improving workflow efficiency for developers.

    Read Full Article: Git-aware File Tree & Search in Jupyter Lab

  • TensorFlow Lite Plugin for Flutter Released


    The TensorFlow Lite Plugin for Flutter is Officially AvailableThe TensorFlow Lite plugin for Flutter has been officially released, now maintained by the Google team after its successful creation by a Google Summer of Code contributor. This plugin allows developers to integrate TensorFlow Lite models into Flutter apps, enhancing mobile app capabilities with features like object detection through a live camera feed. TensorFlow Lite offers cross-platform support and on-device performance optimizations, making it ideal for mobile, embedded, web, and edge devices. Developers can find pre-trained models or create custom ones, and the plugin's GitHub repository provides examples for various machine learning tasks, including image classification. This development is significant as it simplifies the integration of advanced machine learning models into Flutter applications, broadening the scope of what developers can achieve on mobile platforms.

    Read Full Article: TensorFlow Lite Plugin for Flutter Released

  • Apple Pauses App Store Changes in Texas


    Apple pauses app store changes in Texas after court blocks age-assurance lawApple has decided to pause its planned changes to the App Store in Texas following a federal judge's decision to block a new age-verification law, citing First Amendment concerns. The law, known as the App Store Accountability Act, would have required app stores to verify user ages and obtain parental consent for users under 18, while also sharing age data with developers. Apple had announced new requirements for apps in Texas to comply with the law, including the use of Family Sharing groups for minors and updated APIs for developers. Despite the legal setback for Texas lawmakers, Apple continues to offer developer tools for age assurance, emphasizing privacy concerns over the collection of sensitive information. This matters because it highlights the ongoing tension between tech companies and lawmakers over privacy and user data protection.

    Read Full Article: Apple Pauses App Store Changes in Texas

  • Sketch to HTML with Qwen3-VL


    Creating a Sketch to HTML Application with Qwen3-VLQwen3-VL is showcased as a powerful tool for developing a sketch-to-HTML application, highlighting its practical application in creating real-world solutions. The process involves using Qwen3-VL to convert hand-drawn sketches into functional HTML code, demonstrating the model's capability to bridge the gap between design and development. This approach not only streamlines the workflow for designers and developers but also exemplifies how advanced machine learning models can be harnessed to automate and enhance creative processes. Understanding and implementing such technology can significantly improve efficiency in web development projects, making it a valuable asset for both individual developers and teams.

    Read Full Article: Sketch to HTML with Qwen3-VL

  • Training a Model for Code Edit Predictions


    A deep dive into how I trained an edit model to show highly relevant code suggestions while programmingDeveloping a coding agent like NES, designed to predict the next change needed in a code file, is a complex task that requires understanding how developers write and edit code. The model considers the entire file and recent edit history to predict where and what the next change should be. Capturing real developer intent is challenging due to the messy nature of real commits, which often include unrelated changes and skip incremental steps. To train the edit model effectively, special edit tokens were used to define editable regions, cursor positions, and intended edits, allowing the model to predict the next code edit within a specified region. Data sources like CommitPackFT and Zeta were utilized, and the dataset was normalized into a unified format with filtering to remove non-sequential edits. The choice of base model for fine-tuning was crucial, with Gemini 2.5 Flash Lite selected for its ease of use and operational efficiency. This managed model avoids the overhead of running an open-source model and uses LoRA for lightweight fine-tuning, ensuring the model remains stable and cost-effective. Flash Lite enhances user experience by providing faster responses and lower compute costs, enabling frequent improvements without significant downtime or version drift. Evaluation of the edit model was conducted using the LLM-as-a-Judge metric, which assesses the semantic correctness and logical consistency of predicted edits. This approach is more aligned with human judgment than simple token-level comparisons, allowing for scalable and sensitive evaluation processes. To make the Next Edit Suggestions responsive, the model receives more than just the current file snapshot at inference time; it also includes the user's recent edit history and additional semantic context. This comprehensive input helps the model understand user intent and predict the next edit accurately. This matters because it enhances coding efficiency and accuracy, offering developers a more intuitive and reliable tool for code editing.

    Read Full Article: Training a Model for Code Edit Predictions

  • Wafer: Streamlining GPU Kernel Optimization in VSCode


    Wafer: VSCode extension to help you develop, profile, and optimize GPU kernelsWafer is a new VS Code extension designed to streamline GPU performance engineering by integrating several tools directly into the development environment. It aims to simplify the process of developing, profiling, and optimizing GPU kernels, which are crucial for improving training and inference speeds in deep learning applications. Traditionally, this workflow involves using multiple fragmented tools and tabs, but Wafer consolidates these functionalities, allowing developers to work more efficiently within a single interface. The extension offers several key features to enhance the development experience. It integrates Nsight Compute directly into the editor, enabling users to run performance analysis and view results alongside their code. Additionally, Wafer includes a CUDA compiler explorer that allows developers to inspect PTX and SASS code mapped back to their source, facilitating quicker iteration on kernel changes. Furthermore, a GPU documentation search feature is embedded within the editor, providing detailed optimization guidance and context to assist developers in making informed decisions. Wafer is particularly beneficial for those involved in training and inference performance work, as it consolidates essential tools and resources into the familiar environment of VS Code. By reducing the need to switch between different applications and tabs, Wafer enhances productivity and allows developers to focus on optimizing their GPU kernels more effectively. This matters because improving GPU performance can significantly impact the efficiency and speed of deep learning models, leading to faster and more cost-effective AI solutions.

    Read Full Article: Wafer: Streamlining GPU Kernel Optimization in VSCode

  • Quint: Interactive Buttons for Chatbots


    I created interactive buttons for chatbots (opensource)Quint is an innovative open-source library designed to enhance chatbot interactions by moving beyond the traditional command-line interface (CLI) approach. Developed as a React library, Quint allows developers to create structured and deterministic interactions on top of large language models (LLMs). By enabling explicit choices through interactive buttons, users can reveal information or send structured input back to the model, with full control over the output display. This separation of model input, user interface, and output rendering helps make interactions like multiple-choice questions, explanations, and role-play scenarios more predictable and less reliant on workaround solutions. One of Quint's key features is its flexibility in terms of presentation, as it only manages the state and behavior of interactions, leaving the design and styling to the developers. This means that developers can fully customize the buttons and user interface elements to fit their specific needs and aesthetic preferences. Additionally, Quint is independent of any specific AI provider, as it operates through callbacks, allowing for integration with various models such as OpenAI, Gemini, Claude, or even mock functions. This versatility ensures that Quint can be used effectively regardless of the underlying AI technology. Currently in its early stages (version 0.1.0), Quint offers a stable core abstraction that promises to evolve into a more comprehensive solution for interactive chatbot interfaces. The creator is seeking feedback to refine and improve the library, aiming to eventually render entire UI elements through LLMs, simplifying interactions for the average end user. This development matters because it represents a significant step forward in making chatbot interactions more intuitive and accessible, potentially transforming how users engage with AI-driven systems.

    Read Full Article: Quint: Interactive Buttons for Chatbots