case study

On-prem Oracle Exadata data warehouse migration to BigQuery

Insurance

Business Impacts

60%

Infrastructure & Maintenance cost savings

3x

Growth in the analytics business

150 TB

Data migrated

21k

Data Mart tables added

Customer Key Facts

  • Country : America
  • Industry : Insurance
  • About : The client is a leading U.S. - based global commercial property and casualty insurance

Challenges

  • The client wanted a model orchestrator that will integrate with the Claim Center/Rules Engine and Support the deployment of clients’ flag library & deploying predictive models in Vertex AI.
  • Client also wanted a ML model that will give scoring to the claims based on severity (High, medium, low), once the FNOL notification has came
  • Preparation of input data for ML model – combining data points from uploaded structured and unstructured documents with legacy data from client database
  • Integration with multiple modules

Technologies Used

Cloud Storage

Cloud Storage

BigQuery

BigQuery

Data Fusion

Data Fusion

Dataflow

Dataflow

Composer

Composer

Looker

Looker

Pub/sub

Pub/sub

DLP API

DLP API

Solution

  • Quantiphi migrated 150 terabytes of data from Oracle Exadata to Google’s BigQuery.
  • Built a claim center data warehouse on GCP
  • Developed an enterprise-grade DLP solution

Results

Significant Cost Savings

60% Reduction in Infrastructure & Maintenance Costs: The migration from Oracle Exadata to Google’s BigQuery resulted in substantial cost savings in infrastructure and maintenance, reducing overall operational expenses.

Improved Claim Processing

  • Integrated Claim Center Data Warehouse: Built a comprehensive claim center data warehouse on GCP, enabling better integration with the Claim Center/Rules Engine.
  • Predictive Model Deployment: Developed and deployed predictive models in Vertex AI for scoring claims based on severity (high, medium, low) upon First Notice of Loss (FNOL).

Expanded Data Infrastructure

  • 21,000 Data Mart Tables Added: The addition of 21,000 data mart tables expanded the client's data infrastructure, providing a robust foundation for advanced analytics and reporting.

Data Integration and Preparation

  • Combining Structured and Unstructured Data: Prepared input data for machine learning models by integrating data points from uploaded structured and unstructured documents with legacy data from the client's database.
  • Seamless Module Integration: Achieved seamless integration with multiple modules, enhancing the overall functionality and efficiency of the system.

Enhanced Analytical Capabilities

  • 3x Growth in Analytics Business: The migration facilitated a threefold increase in the analytics business, allowing the client to leverage enhanced data processing and analytical capabilities.
  • 150 TB Data Migrated: Successfully migrated 150 terabytes of data, ensuring seamless continuity and accessibility of historical data.

Enterprise-Grade Data Security

Enterprise-Grade DLP Solution: Developed a robust Data Loss Prevention (DLP) solution, ensuring the security and compliance of sensitive data throughout the migration and beyond.

Thank you for reaching out to us!

Our experts will be in touch with you shortly.

In the meantime, explore our insightful blogs and case studies.

Something went wrong!

Please try it again.

Share