Placement papers | Freshers Walkin | Jobs daily: Principal Data Engineer at USA TODAY NETWORK (McLean, VA) (allows remote)


Search jobs and placement papers

Principal Data Engineer at USA TODAY NETWORK (McLean, VA) (allows remote)

The API Services team is responsible for engineering and delivering cutting-edge services to aide in content delivery to end customers. These services support 110 news brands, and more than 110 million unique monthly visitors.


The Principal Data Engineer will play a key role in architecting, developing and maintaining the data architecture for Gannett's new Content Platform that supports the content production & delivery systems that are consumed by both our network of 3000 journalists & our customer facing products. You will be expected to design & consume large scale, fault tolerant and highly available architectures. A large part of your role will be forward looking, with an emphasis on optimizing content structures & relationships.If you have a passion for rapid development, automation, learning, challenging and bettering your peers, with a strong desire to operate in a full stack environment, you'd probably fit in well here.


Responsibilities:



  • Collaborate with stakeholders & developers to identify data needs & ideal implementation.

  • Contribute to the architecture and vision of Gannett's content data pipeline.

  • Track record of evolving complex data environments.

  • Continuously evaluate data usage patterns and identify areas of improvement.

  • Interface closely with data scientists and engineering to ensure reliability and scalability of data environment.

  • Drive future state technologies, designs and ideas across the organization.

  • Provide planning for two-week sprints.

  • Provide day to day operational support for our applications.

  • Improve and establish best practice around our application and infrastructure monitoring.


Automate everything:



  • Containerizing applications with Docker

  • Scripting new solutions/APIs/services to reduce toil

  • Research new tools to optimize cost, deployment speed and resource usage

  • Assist in improving our onboarding structure and documentation.


Responsibility Breakdown:



  • 30% - Data architecture design / review

  • 20% - Mentoring

  • 15% - Application Support

  • 15% - Planning / Documentation

  • 10% - Design applications / recommendations / poc

  • 10% - New Technology Evaluation


Technologies:


Systems:



  • Linux

  • Couchbase

  • Elastic Search

  • Solr

  • Neo4j

  • Other NoSQL Databases


Exciting things you get to do:



  • Engineering high-performant applications with an emphasis on concurrency

  • Agile

  • Amazon Web Services, Google Compute Engine

  • Google DataStore, Spanner, DynamoDB

  • Docker, Kubernetes

  • Database testing

  • GraphQL

  • Fastly

  • Terraform

  • Monitoring with NewRelic


Minimum Qualifications:



  • Deep experience in ETL design, schema design and dimensional data modeling.

  • Ability to match business requirements to technical ETL design and data infrastructure needs.

  • Experience using search technologies like Elasticsearch and Solr and designing the integration of search with a persistent data store.

  • Deep understanding of data normalization methodologies.

  • Deep understanding of both Relational and NoSQL databases.

  • Experience with data solutions like Hadoop, Teradata, Oracle.

  • Proven expertise with query languages such as SQL, T-SQL, NRQL, solr querying.

  • Self-Starter that can operate in a remote-friendly environment.

  • Experience with Agile (Scrum) and test driven development, continuous integration and version control (GIT).

  • Experience deploying to Cloud compute or container hosting.

  • Experience working with data modeling tools.

  • Basic understanding of REST APIs, SDKs and CLI toolsets.

  • Understanding of web technologies.

  • Experience with Data in a media industry is a plus.


by via developer jobs - Stack Overflow
 

No comments:

Post a Comment