VesperAI's picture
addede a Production Branch
179cb94
metadata
title: Visual Product Matcher
emoji: ๐Ÿ”
colorFrom: blue
colorTo: green
sdk: docker
app_port: 7860
pinned: false

Visual Product Search ๐Ÿ”

An intelligent visual search engine that revolutionizes product discovery using state-of-the-art AI technology. This application combines CLIP (Contrastive Language-Image Pre-Training) with Qdrant vector database to enable semantic search across image collections, making it perfect for e-commerce, inventory management, and content discovery.

๐ŸŒŸ Key Features

  • ๐ŸŽฏ Multi-Modal Search: Search using text descriptions, uploaded images, or image URLs
  • ๐Ÿ–ผ๏ธ Smart Indexing: Automatically indexes and monitors image folders with real-time updates
  • ๐Ÿ” Semantic Understanding: Uses OpenAI's CLIP model for deep image-text comprehension
  • ๏ฟฝ Similarity Scoring: Provides percentage-based similarity scores for accurate results
  • โšก Real-time Processing: WebSocket-powered live progress updates during indexing
  • ๐ŸŽจ Modern UI: Clean, responsive interface with advanced search capabilities
  • ๐ŸŒ URL Support: Direct image search from web URLs
  • ๐Ÿ“ฑ Mobile Responsive: Works seamlessly across all devices

๐Ÿง  Technical Approach & Solution

Problem Statement

Traditional image search relies on metadata and filenames, which often fail to capture the actual visual content. Users struggle to find specific products or images without knowing exact file names or having perfect tagging systems.

Our Solution Architecture

1. Multi-Modal Embedding Generation

Text Query โ†’ CLIP Text Encoder โ†’ 512D Vector
Image Input โ†’ CLIP Vision Encoder โ†’ 512D Vector
URL Image โ†’ Download โ†’ CLIP Vision Encoder โ†’ 512D Vector

2. Vector Similarity Search

  • Database: Qdrant cloud vector database for scalable similarity search
  • Indexing: Real-time folder monitoring with automatic embedding generation
  • Storage: Hybrid approach - embeddings in Qdrant, metadata in SQLite

3. Semantic Matching Pipeline

User Input โ†’ Feature Extraction โ†’ Vector Search โ†’ Similarity Ranking โ†’ Results

๏ฟฝ๏ธ Architecture Components

Backend (FastAPI)

  • Image Processing: PIL + CLIP for feature extraction
  • Vector Operations: Qdrant client for similarity search
  • File Management: Automatic folder monitoring and indexing
  • API Endpoints: RESTful APIs for all search operations

Frontend (Modern Web UI)

  • Framework: Vanilla JavaScript with Bootstrap 5
  • Styling: Custom CSS with modern design principles
  • Real-time Updates: WebSocket connections for live progress
  • Responsive Design: Mobile-first approach

Database Layer

  • Vector Storage: Qdrant cloud for embeddings and similarity search
  • Metadata Storage: SQLite for image metadata and file information
  • Caching: Thumbnail generation and caching for performance

๐Ÿš€ Quick Start

Prerequisites

  • Python 3.8+
  • CUDA-compatible GPU (optional, recommended for performance)
  • Qdrant Cloud account (free tier available)

Installation

  1. Clone the repository:
git clone https://github.com/itsfuad/SnapSeek
cd SnapSeek
  1. Create virtual environment:
python -m venv venv
source venv/bin/activate  # Windows: venv\Scripts\activate
  1. Install dependencies:
pip install -r requirements.txt
  1. Configure environment: Create a .env file:
QDRANT_API_KEY=your_qdrant_api_key
QDRANT_URL=your_qdrant_cluster_url
  1. Launch the application:
python app.py
  1. Access the interface: Open http://localhost:8000 in your browser

๐ŸŽฏ Usage Guide

1. Index Your Images

  • Click "Add Folder" to select image directories
  • Watch real-time indexing progress
  • Images are automatically monitored for changes

2. Search Methods

Text Search

"red sports car"
"woman wearing blue dress"
"modern kitchen design"

Image Upload Search

  • Click the image icon
  • Upload a reference image
  • Get visually similar results

URL Search

  • Click the link icon
  • Paste any image URL
  • Find similar images in your collection

3. Results & Insights

  • Similarity percentages for each match
  • High-resolution image previews
  • Metadata and file information

๐Ÿญ Production Deployment

Recommended Platforms

1. Railway (Recommended)

  • Why: Best for AI/ML applications with generous free tier
  • Resources: 512MB RAM, 1GB storage
  • Benefits: No sleep mode, automatic GitHub deployments
# Dockerfile
FROM python:3.9-slim

WORKDIR /app
COPY requirements.txt .
RUN pip install --no-cache-dir -r requirements.txt
COPY . .
EXPOSE 8000
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "8000"]

2. Render

  • Resources: 512MB RAM, 1GB storage
  • Benefits: Free SSL, auto-deploy, no cold starts

3. Fly.io

  • Resources: 256MB RAM, 3GB storage volume
  • Benefits: Global edge deployment, persistent volumes

Environment Variables for Production

QDRANT_API_KEY=your_production_key
QDRANT_URL=your_production_cluster
PORT=8000
DATA_DIR=/app/data

๐Ÿ› ๏ธ Development & Testing

Project Structure

SnapSeek/
โ”œโ”€โ”€ app.py                 # FastAPI application
โ”œโ”€โ”€ image_indexer.py       # Image processing and indexing
โ”œโ”€โ”€ image_search.py        # Search logic and CLIP integration
โ”œโ”€โ”€ image_database.py      # Database operations
โ”œโ”€โ”€ folder_manager.py      # Folder monitoring and management
โ”œโ”€โ”€ qdrant_singleton.py    # Qdrant client management
โ”œโ”€โ”€ requirements.txt       # Dependencies
โ”œโ”€โ”€ .env                   # Environment configuration
โ”œโ”€โ”€ templates/
โ”‚   โ””โ”€โ”€ index.html        # Main UI template
โ”œโ”€โ”€ static/
โ”‚   โ”œโ”€โ”€ js/
โ”‚   โ”‚   โ””โ”€โ”€ script.js     # Frontend JavaScript
โ”‚   โ””โ”€โ”€ image.png         # Application icon
โ”œโ”€โ”€ config/
โ”‚   โ””โ”€โ”€ folders.json      # Folder configuration
โ””โ”€โ”€ tests/
    โ””โ”€โ”€ test_*.py         # Test files

Running Tests

pip install -r requirements-test.txt
pytest tests/ -v

Development Setup

# Install development dependencies
pip install -r requirements-test.txt

# Run with auto-reload
uvicorn app:app --reload --host 0.0.0.0 --port 8000

๐Ÿ”ง Performance Optimization

Model Selection

# For production (smaller, faster)
MODEL_NAME = "openai/clip-vit-base-patch16"

# For development (balance)
MODEL_NAME = "openai/clip-vit-base-patch32"

Hardware Recommendations

  • CPU: 4+ cores for concurrent processing
  • RAM: 8GB+ for model loading and image processing
  • Storage: SSD recommended for faster I/O
  • GPU: Optional, CUDA-compatible for faster inference

Scaling Considerations

  • Batch Processing: Process multiple images simultaneously
  • Caching: Implement Redis for frequent queries
  • Load Balancing: Use multiple instances for high traffic
  • Database Sharding: Split collections by categories

๐Ÿ› Troubleshooting

Common Issues

1. Model Loading Errors

# Clear cache and reinstall
pip uninstall torch torchvision transformers
pip install torch torchvision transformers --no-cache-dir

2. Qdrant Connection Issues

  • Verify API key and URL in .env
  • Check network connectivity
  • Ensure Qdrant cluster is active

3. Memory Issues

  • Reduce batch size in processing
  • Use CPU-only mode: device="cpu"
  • Close unused applications

4. Slow Performance

  • Enable GPU acceleration
  • Optimize image sizes
  • Implement result caching

Performance Monitoring

# Add logging for performance tracking
import time
import logging

logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)

# Time search operations
start_time = time.time()
results = await searcher.search_by_text(query)
logger.info(f"Search completed in {time.time() - start_time:.2f}s")

๐Ÿค Contributing

  1. Fork the repository
  2. Create a feature branch: git checkout -b feature-name
  3. Make your changes and add tests
  4. Run tests: pytest tests/
  5. Commit changes: git commit -m "Add feature"
  6. Push to branch: git push origin feature-name
  7. Create a Pull Request

Code Standards

  • Follow PEP 8 style guidelines
  • Add docstrings to all functions
  • Include type hints where appropriate
  • Write tests for new features

๐Ÿ“Š Use Cases & Applications

E-commerce

  • Product recommendation systems
  • Visual search for online stores
  • Inventory management
  • Duplicate product detection

Content Management

  • Digital asset organization
  • Stock photo searching
  • Brand consistency checking
  • Content moderation

Research & Education

  • Academic image databases
  • Scientific data analysis
  • Historical archive searches
  • Educational content discovery

๐Ÿ”ฎ Future Enhancements

  • Multi-language Support: Extend text search to multiple languages
  • Advanced Filters: Add size, color, and metadata filters
  • Batch Operations: Upload and search multiple images at once
  • API Integration: RESTful API for external applications
  • Machine Learning: Custom fine-tuned models for specific domains
  • Analytics Dashboard: Search metrics and usage statistics
  • Mobile App: Native mobile applications
  • Cloud Storage: Integration with AWS S3, Google Drive, etc.

๐Ÿ“„ License

This project is licensed under the Mozilla Public License 2.0 - see the LICENSE file for details.

๐Ÿ™ Acknowledgments

  • OpenAI: For the CLIP model and research
  • Qdrant: For the excellent vector database
  • FastAPI: For the modern web framework
  • Transformers: For the model implementation
  • Bootstrap: For the UI components

๐Ÿ“ž Support & Contact


Made with โค๏ธ by itsfuad

Revolutionizing visual search with AI technology