- Responsibility
- Collaborate with data scientists to phase in statistical, predictive machine learning and AI models to production scale and continuously optimizing performance.
- Design, build, optimize, launch and support new and existing data models and ETL processes in production based on data products and stakeholder needs.
- Define and manage SLA and accuracy for all data sets in allocated areas of ownership.
- Design and continuously improve data infrastructure and identify infra issues and drive to resolution.
- Support software development team to build and maintain data collectors in Migo application ecosystem based on data warehouse and analytics user requirements.
- Basic Qualification:
- Bachelor's degree in Computer Science, Information Management or related field.
- 2+ years hands-on experience in the data warehouse space, custom ETL design, implementation and maintenance.
- 2+ years hands-on experience in SQL or similar languages and development experience in at least one scripting language (Python preferred).
- Strong data architecture, data modeling, schema design and effective project management skills.
- Excellent communication skills and proven experience in leading data driven projects from definition through interpretation and execution.
- Experience with large data sets and data profiling techniques.
- Ability to initiate and drive projects, and communicate data warehouse plans to internal clients/stakeholders.
- Preferred Qualification:
- Experience with big data and distributed computing technology such as Hive, Spark, Presto, Parquet
- Experience building and maintaining production level data lake with Hadoop Cluster or AWS S3.
- Experience with batch processing and streaming data pipeline/architecture design patterns such as lambda architecture or kappa architecture.
by via developer jobs - Stack Overflow
No comments:
Post a Comment