Description:
Design and build data structures on MPP platform like AWS RedShift and or Druid.io
Design and build highly scalable data pipelines using AWS tools like Glue (Spark based), Data Pipeline, Lambda.
Translate complex business requirements into scalable technical solutions.
Strong understanding of analytics needs
Collaborate with the team on building dashboards, using Self-Service tools like Apache Superset or Tableau, and data analysis to support business
Collaborate with multiple cross-functional teams and work on solutions which have a large impact on Xapo's business.
Requirements:
In-depth understanding of data structures and algorithms
Experience in designing and building dimensional data models to improve accessibility, efficiency, and quality of data
Experience in designing and developing ETL data pipelines.
Should be proficient in writing Advanced SQLs, Expertise in performance tuning of SQLs
Programming experience in building high quality software. Skills with Python or Scala preferred
Strong analytical and communication skills
Work/project experience with big data and advanced programming languages is a plus. Experience using java, Spark, Hive, oozie, Kafka, and Map Reduce is a huge plus!
Work experience with AWS tools to process data (Glue, Pipeline, Kinesis, Lambda, etc) is a plus
Experience with or advanced courses on data science and machine learning is a plus
Bachelor’s Degree in computer science or a related field
Fluent in English
Speak a second language, ideally but not a requirement.
by via developer jobs - Stack Overflow
No comments:
Post a Comment