Currently implementing a Data Quality framework using the DQX framework with a metadata-driven architecture. The solution incorporates various data quality checks such as null checks, duplicate detection, date validation, and numeric validations.
Could you please share your recommendations on the overall architecture and solution approach for designing and scaling such a framework? Specifically, I would be interested in guidance on:
1. Structuring metadata/configuration tables for flexible rule management
2. Designing a reusable and scalable validation engine
3. Handling rule execution, logging, and auditability
4. Best practices for integrating with data pipelines and workflows