Skip to content

gmh5225/DIPC

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Document Intelligence & Parsing Center (DIPC)

Overview

DIPC is an enterprise-grade document processing system that leverages multi-modal LLMs to extract structured information from various document formats. Built with a microservices architecture, it provides high-performance, scalable document intelligence capabilities.

✨ Key Features

  • 📄 Multi-Format Support - PDFs, images (JPG/PNG/GIF/WEBP), text files, and ZIP archives
  • 🚀 High Performance - Asynchronous processing with distributed workers
  • 💰 Cost Management - Built-in estimation and usage limiting
  • 🔍 Vector Search - Optional semantic search capabilities
  • 🌐 Multi-Architecture - Native support for x86_64 and ARM64 (including Apple Silicon)
  • 🔒 Enterprise Ready - Production-grade security and monitoring

🚀 Quick Start (3 minutes)

Prerequisites

Deploy with Pre-built Images

# 1. Create project directory
mkdir dipc && cd dipc

# 2. Download configuration
curl -O https://raw.githubusercontent.com/nociex/DIPC/main/docker-compose.yml
curl -O https://raw.githubusercontent.com/nociex/DIPC/main/.env.example

# 3. Configure API key
cp .env.example .env
# Edit .env and add your OpenAI/OpenRouter API key

# 4. Start services
docker compose up -d

# 5. Access the application
# Frontend: http://localhost:38110
# API Docs: http://localhost:38100/docs

🏗️ Architecture

graph LR
    A[Next.js Frontend] --> B[FastAPI Gateway]
    B --> C[Redis Queue]
    C --> D[Celery Workers]
    D --> E[LLM Providers]
    D --> F[Vector DB]
    B --> G[(PostgreSQL)]
    D --> G
Loading

📊 Performance

  • Processing Speed: 10-50 pages/minute (depending on complexity)
  • Concurrent Jobs: Up to 100 simultaneous documents
  • API Latency: <100ms for most endpoints
  • Storage: S3-compatible object storage support

🔧 Configuration

Basic Configuration

# Required: Choose one LLM provider
OPENAI_API_KEY=sk-your-key
# OR
OPENROUTER_API_KEY=sk-or-your-key

# Optional: Advanced settings
MAX_WORKERS=4
PROCESSING_TIMEOUT=300
VECTOR_DB_ENABLED=true

Advanced Deployment

For production deployments, Kubernetes configurations, and advanced settings, see the Deployment Guide.

📚 Documentation

🤝 Contributing

We welcome contributions! Please see our Contributing Guide for details.

📄 License

This project is licensed under the MIT License - see the LICENSE file for details.

🙏 Acknowledgments

Built with excellent open-source technologies including FastAPI, Next.js, Celery, and more.


Made with ❤️ by the DIPC Team

About

Document Intelligence & Parsing Center - A comprehensive document processing system using multi-modal LLMs

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 50.3%
  • TypeScript 47.4%
  • Shell 1.1%
  • Batchfile 0.4%
  • CSS 0.3%
  • JavaScript 0.2%
  • Other 0.3%