AI & Technology Updates

  • Transcribe: Local Audio Transcription with Whisper


    Transcribe: local Whisper transcription (GUI + CLI) with diarization, timestamps, optional OllamaTranscribe (tx) is a free desktop and CLI tool designed for local audio transcription using Whisper, capable of capturing audio from files, microphones, or system audio to produce timestamped transcripts with speaker diarization. It offers multiple modes, including file mode for WAV file transcription, mic mode for live microphone capture, and speaker mode for capturing system audio with optional microphone input. The tool is offline-friendly, running locally after the initial model download, and supports optional summaries via Ollama models. It is cross-platform, working on Windows, macOS, and Linux, and is automation-friendly with CLI support for batch processing and repeatable workflows. This matters as it provides a versatile, privacy-focused solution for audio transcription and analysis without relying on cloud services.


  • Lár: Open-Source Framework for Transparent AI Agents


    I built a "Glass Box" agent framework because I was tired of debugging magic black boxes. (Apache 2.0)Lár v1.0.0 is an open-source framework designed to build deterministic and auditable AI agents, addressing the challenges of debugging opaque systems. Unlike existing tools, Lár offers transparency through auditable logs that provide a detailed JSON record of an agent's operations, allowing developers to understand and trust the process. Key features include easy local support with minimal changes, IDE-friendly setup, standardized core patterns for common agent flows, and an integration builder for seamless tool creation. The framework is air-gap ready, ensuring security for enterprise deployments, and remains simple with its node and router-based architecture. This matters because it empowers developers to create reliable AI systems with greater transparency and security.


  • Reddit’s AI Content Cycle


    It's happening right in front of usReddit's decision to charge for large-scale API access in July 2023 was partly due to companies using its data to train large language models (LLMs). As a result, Reddit is now experiencing an influx of AI-generated content, creating a cycle where AI companies pay to train their models on this content, which then influences future AI-generated content on the platform. This self-reinforcing loop is likened to a "snake eating its tail," highlighting the potential for an unprecedented cycle of AI content generation and training. Understanding this cycle is crucial as it may significantly impact the quality and authenticity of online content.


  • Exploring Human Perception with DCGAN and Flower Images


    I trained a DCGAN on 2k+ flower images to test human perception limits. Here are the results (Live Demo included)Training a DCGAN (Deep Convolutional Generative Adversarial Network) on over 2,000 flower images aimed to explore the boundaries of human perception in distinguishing between real and generated images. The project highlights the effectiveness of Python as the primary programming language for machine learning due to its ease of use, rich ecosystem of libraries like TensorFlow and PyTorch, and strong community support. Other languages such as R, Julia, C++, Scala, Rust, and Kotlin also offer unique advantages, particularly in statistical analysis, performance, and big data processing. Understanding the strengths of different programming languages can significantly enhance the development and performance of machine learning models.


  • Instagram’s Challenge: Authenticity in an AI World


    You can’t trust your eyes to tell you what’s real anymore, says the head of InstagramInstagram faces the challenge of adapting to a rapidly changing world where authenticity is becoming infinitely reproducible through advancements in AI and deepfake technology. As AI-generated content becomes increasingly indistinguishable from real media, the platform must focus on identifying and verifying authentic content while providing context about the creators behind it. The shift from polished, professional-looking images to raw, unfiltered content signals a demand for authenticity, as people seek content that feels real and personal. To maintain trust and relevance, Instagram and similar platforms need to develop tools that label AI content, verify real media, and highlight credibility signals about content creators. This matters because the ability to discern authenticity in digital media is crucial for maintaining trust in the information we consume.