
•Design an AI-powered data quality monitoring framework for Azure Data Lake within LexisNexis Risk Solutions’ Insurance Technology division
•Apply machine learning and time-series methods to detect anomalies, missing values, formatting issues, and other data quality problems
•Evaluate Azure-native AI/ML tools and third-party solutions for data quality monitoring
•Build dashboards to visualize data trends and quality signals
•Deliver a proof-of-concept monitoring system with clear design and operational recommendations

•Automate the Bridger Watchlist research process for LexisNexis Risk Solutions
•Analyze assigned URLs to understand data formats and identify anomalies
•Document findings and develop Python scripts for web scraping and data extraction
•Work with HTML, JSON, and XML data structures and apply regex- based extraction techniques
•Build repeatable, automated workflows that improve operational efficiency

•Research and organize metadata from multiple internal sources to support a ChatFlow-based knowledge system at LexisNexis Risk Solutions
•Interview data engineering teams to understand metadata needs and current workflows
•Gather, analyze, and document metadata to identify key patterns and insights
•Contribute recommendations that improve metadata accessibility and internal processes
•Support the development of a unified, searchable internal knowledge base

•Optimize PowerBI data refresh processes by integrating Azure Data Factory (ADF) with the PowerBI REST API
•Build secure API authentication and improve pipeline reliability to reduce refresh failures
•Enhance error handling to address timing conflicts between ADF and Synapse updates
•Design scalable ADF pipelines and implement automated PowerBI refresh triggers
•Collaborate with EDI and PowerBI engineering teams to deliver a more stable, efficient reporting workflow

•Develop an AI-powered framework for real- time data quality monitoring in Azure Data Lake for LexisNexis Risk Solutions
•Design and implement anomaly detection and time-series models to identify missing values, formatting issues, duplicates, and unusual data trends
•Analyze large-scale insurance datasets to detect and categorize data quality problems
•Evaluate Azure-native machine learning tools alongside third-party solutions
•Recommend the most effective approach for automated, scalable data quality management
•Design and deliver a performance- monitoring dashboard for AES using Power BI
•Gather workflow and performance data to identify key metrics and visualization needs
•Apply UI/UX dashboard design principles and create iterative mock-ups
•Develop and refine the dashboard based on client requirements and ongoing feedback
•Test the solution across devices and produce final documentation, a project report, and a client presentation
.png)
•Develop a digital customer success toolkit for Georgia- Pacific to automate Quarterly Business Reviews (QBRs)
•Design standardized QBR templates and build ROI calculation models
•Create dashboards that visualize operational efficiency, labor savings, and performance trends
•Integrate data from IoT dispenser systems to generate meaningful, data- driven insights
•Develop customer success playbooks and engagement strategies
•Deliver a scalable, automated framework that enhances customer engagement and demonstrates measurable business impact