Skip to content

v0.3.19

Latest

Choose a tag to compare

@rstrahan rstrahan released this 03 Oct 22:17
· 110 commits to main since this release

[0.3.19]

Added

  • Error Analyzer (Troubleshooting Tool) for AI-Powered Failure Diagnosis

    • Introduced intelligent AI-powered troubleshooting agent that automatically diagnoses document processing failures using Claude Sonnet 4 with the Strands agent framework
    • Key Capabilities: Natural language query interface, intelligent routing between document-specific and system-wide analysis, multi-source data correlation (CloudWatch Logs, DynamoDB, Step Functions), root cause identification with actionable recommendations, evidence-based analysis with collapsible log details
    • Web UI Integration: Accessible via "Troubleshoot" button on failed documents with real-time job status, progress tracking, automatic job resumption, and formatted results (Root Cause, Recommendations, Evidence sections)
    • Tool Ecosystem: 8 specialized tools including analyze_errors (main router), analyze_document_failure, analyze_recent_system_errors, CloudWatch log search tools, DynamoDB integration tools, and Lambda context retrieval - additional tools will be added as the feature evolves.
    • Configuration: Configurable via Web UI including model selection (Claude Sonnet 4 recommended), system prompt customization, max_log_events (default: 5), and time_range_hours_default (default: 24)
    • Documentation: Comprehensive guide in docs/error-analyzer.md with architecture diagrams, usage examples, best practices, troubleshooting guide.
  • Claude Sonnet 4.5 Model Support

    • Added support for Claude Sonnet 4.5 and Claude Sonnet 4.5 - Long Context models
    • Available for configuration across all document processing steps

Fixed

  • Problem with setting correctly formatted WAF IPv4 CIDR range - #73

  • Duplicate Step Functions Executions on Document Reprocess - GitHub Issue #66

    • Eliminated duplicate workflow executions when reprocessing large documents (>40MB, 500+ pages)
    • Root Cause: S3 copy_object operations were triggering multiple "Object Created" events for large files, causing queue_sender to create duplicate document entries and workflow executions
    • Solution: Refactored reprocess_document_resolver to directly create fresh Document objects and queue to SQS, completely bypassing S3 event notifications
    • Benefits: Eliminates unnecessary S3 copy operations (cost savings)