embedding generation

  • VectorDBZ: Local GUI for Vector Databases


    I built a local GUI for vector DBs (pgvector, Qdrant, Chroma, Milvus, Weaviate)VectorDBZ is a desktop application designed to facilitate the exploration and debugging of vector databases like Qdrant, Weaviate, Milvus, Chroma, and pgvector in local and self-hosted environments. It addresses the challenge of inspecting vector stores without relying on cloud-based tools or cumbersome scripts by providing features such as browsing collections, running vector similarity searches, generating embeddings, and visualizing data using techniques like PCA, t-SNE, or UMAP. By storing all configurations and API keys locally, VectorDBZ enhances privacy and is particularly useful for debugging local RAG pipelines and semantic search setups. This matters because it empowers developers working with vector databases to efficiently manage and analyze data in a secure, local environment.

    Read Full Article: VectorDBZ: Local GUI for Vector Databases

  • LLMeQueue: Efficient LLM Request Management


    LLMeQueue: let me queue LLM requests from my GPU - local or over the internetLLMeQueue is a proof-of-concept project designed to efficiently handle large volumes of requests for generating embeddings and chat completions using a locally available NVIDIA GPU. The setup involves a lightweight public server that receives requests, which are then processed by a local worker connected to the server. This worker, capable of concurrent processing, uses the GPU to execute tasks in the OpenAI API format, with llama3.2:3b as the default model, although other models can be specified if available in the worker’s Ollama environment. LLMeQueue aims to streamline the process of managing and processing AI requests by leveraging local resources effectively. This matters because it offers a scalable solution for developers needing to handle high volumes of AI tasks without relying solely on external cloud services.

    Read Full Article: LLMeQueue: Efficient LLM Request Management