Duties and Responsibilities
- Work in an fast-paced agile development environment architecting and developing Hadoop applications
 - Provide technology recommendations for potential product application development
 - Gather and analyze requirements from product owners ensuring products meet business requirements
 - Collaborate with other software engineers and team leads in designing and developing software solutions which meet high quality standards
 - Quickly prototype and develop Python/Java/Scala applications in diverse operating environments capable of interfacing with NoSQL datastores such as Accumulo and HBase
 - Write efficient code to extract, transform, load, and query very large datasets to include both structured and unstructured datasets
 - Develop standards and new design patterns for Big Data applications and master the tools and technology components within the Hadoop and Cloudera environments
 - Design and implement REST API applications provide web application connectivity to backend datastores
 
Skills & Requirements
- 3 years of building Java applications including framework experience (J2EE, Spring, etc.)
 - Experience using traditional ETL tools & RDBMS
 - Experience developing REST web services
 - Demonstrated ability to quickly learn and apply new technologies
 - Demonstrated effective and successful verbal and written communication skills
 - Bachelor degree in Computer Science or related technological degree
 - U.S. citizen
 
Desired Qualifications
- Experience building and coding applications using Hadoop components – HDFS, HBase, Hive, Sqoop, Flume, Spark, etc
 - Experience building and maintaining Cloudera-based clusters
 - Full life cycle software application development experience
 - Experience with unstructured datasets, such as log files, email, text
 - Experience with geospatial datasets and datastores
 
by via developer jobs - Stack Overflow
No comments:
Post a Comment