An art institution implemented a sophisticated multimodal search system for their collection of 40 million art assets using vector databases and LLMs. The system combines text and image-based search capabilities, allowing users to find artworks based on various attributes including style, content, and visual similarity. The solution evolved from using basic cloud services to a more cost-effective and flexible approach, reducing infrastructure costs to approximately $1,000 per region while maintaining high search accuracy.
This case study explores the implementation of a sophisticated multimodal search system for a large art collection, highlighting the challenges and solutions in deploying LLMs and vector databases in a production environment.
The project's core challenge was managing and making searchable a vast collection of approximately 40 million art assets. The institution needed to enable complex search capabilities that could handle both text and image-based queries, while maintaining reasonable costs and system performance.
**Initial Architecture and Challenges**
The initial system architecture was built around several key components:
* Vector databases for storing and querying embeddings
* AWS Bedrock for model serving
* Image transformation pipelines for vector generation
* Multiple language models for different aspects of the search functionality
The team faced several significant challenges in their initial implementation:
* Cost Management: The initial solution proved expensive, with significant costs associated with image transformation to vectors. While the infrastructure cost was approximately $1,000 per month per region, the additional processing costs were substantial.
* Scale Challenges: With 40 million assets to process, the team had to carefully consider their approach to batch processing and index updates.
* Search Quality: Traditional search methods weren't sufficient for art-specific queries, requiring a more sophisticated approach combining multiple search modalities.
**Evolution and Optimization**
The team made several strategic improvements to optimize their system:
*Infrastructure Optimization*
* Reduced redundant processing by implementing smarter indexing strategies
* Implemented a more selective approach to when images needed to be re-processed
* Developed a more efficient storage pipeline
*Search Quality Improvements*
* Implemented vector normalization techniques to improve search results
* Added support for combination searches mixing different modalities
* Developed specialized ranking algorithms for art-specific use cases
*Cost Reduction Strategies*
* Moved away from certain expensive cloud services to more cost-effective alternatives
* Implemented batch processing optimizations
* Reduced unnecessary re-processing of stable assets
**Technical Implementation Details**
The system employs several sophisticated technical approaches:
*Vector Processing and Storage*
* Custom vector normalization pipelines to ensure consistent search quality
* Integration with OpenSearch for efficient vector storage and retrieval
* Specialized ranking algorithms for art-specific search scenarios
*Model Integration*
* Multiple specialized models for different aspects of art analysis
* Integration with AWS Bedrock for reliable model serving
* Custom connectors for efficient model pipeline management
*Search Implementation*
* Multimodal search capabilities combining text and image inputs
* Support for similarity search across different artistic attributes
* Flexible query processing supporting various search modalities
**Reliability and Redundancy**
The team implemented several reliability measures:
* Secondary infrastructure deployment for failover
* Redundant processing pipelines
* Backup systems for critical components
**Results and Outcomes**
The implemented system successfully achieved several key objectives:
* Enabled sophisticated art search capabilities across a massive collection
* Maintained reasonable operational costs after optimization
* Provided high-quality search results across different search modalities
* Supported complex queries combining multiple search attributes
**Lessons Learned and Best Practices**
The project revealed several important insights for similar LLMOps implementations:
*Cost Management*
* Careful evaluation of cloud service costs is essential for large-scale implementations
* Batch processing strategies can significantly impact operational costs
* Regular review and optimization of processing pipelines is crucial
*Technical Implementation*
* Vector normalization is crucial for consistent search quality
* Multiple search modalities can provide better results than single-approach methods
* Careful consideration of infrastructure requirements can prevent overprovisioning
*Operational Considerations*
* Regular monitoring of model performance is essential
* Backup systems and redundancy are crucial for production systems
* Careful documentation of system behavior helps in ongoing optimization
The case study demonstrates the complexity of implementing LLM-based systems in production environments, particularly for specialized domains like art collections. The team's iterative approach to optimization and their focus on both technical performance and cost management provides valuable insights for similar implementations.
The solution shows how combining multiple AI technologies - including LLMs, vector databases, and custom ranking algorithms - can create sophisticated search capabilities while maintaining reasonable operational costs. The emphasis on optimization and careful system design demonstrates the importance of thoughtful LLMOps practices in production environments.
Start your new ML Project today with ZenML Pro
Join 1,000s of members already deploying models with ZenML.