Our client is a rapidly growing Insuretech company developing innovative products in the insurance industry. We use advanced machine learning to drive optimal claims outcomes.
Key Responsibilities:
- Design and optimize Data Lakehouse architecture for high performance and scalability.
- Manage data integration from various sources.
- Collaborate with teams to align data models with business needs.
- Enforce data governance, quality, and security standards.
- Automate processes using scripting and CI/CD tools.
- Troubleshoot and improve system performance.
- Work with big data tools like Apache Spark and Delta Lake.
- Maintain Data Lakehouse stability and reliability.
- Stay updated on big data and cloud technology trends.
Skills and Expertise:
- 5+ years with cloud-based data solutions (Redshift, Snowflake, BigQuery).
- Experience with AWS Lake Formation.
- 3+ years in ELT/ETL development.
- Proficient in Python, Java, or SQL.
- Strong SQL programming and data modeling skills.
- Experience with data pipeline and orchestration tools (Hive, Spark, Airflow).
- Knowledge of containerization (Docker, Kubernetes).
- Agile development experience.
- Understanding of machine learning workflows and data visualization tools is a plus.
- Strong communication and collaboration skills.
Nice to Have:
- AWS certifications.
- Knowledge of commercial claims management systems.