What's the role and how do you fit in?
BenevolentAI harnesses artificial intelligence to accelerate scientific discovery by making sense of highly fragmented information to develop new medicines for hard to treat diseases, using AI as a force for good. Valued at $2bn in 2018, we are the largest independent AI company in the world.
Our software engineers work in multi-functional squads working alongside the best scientists, machine learning gurus, world-renown biologists, and extraordinary scientists to make the world a better place.
Our tech stack includes Java, Python, variety of graph databases, ElasticSearch, Linux, Ansible, Kubernetes, Mesos, Puppet, Chef, Docker, GitLab, Bamboo, Travis, Terraform, AWS, EC2, Jenkins, Sonar, Git, VMWare, Saltstack, Atlassian, MySQL, NoSQL, Hadoop, Spark, Nginx, Apache, Cassandra and more.
We don't stop there; we empower our engineers to look at improving our engineering stack wherever possible. You will be encouraged to bring in the newest advancements and cutting-edge solutions, and to collaborate on the ever evolving needs of our technology.
To learn more about our Squads and what you'll be working on follow this link or scroll down.
What will you be accountable for?
- As a core team member, you are able to develop the platform to provide a scalable infrastructure and own the intelligent search technology to facilitate ground-breaking discoveries.
- You are responsible for Agile best-practices to ensure the software is resilient enough for our scientists to rely upon.
- You will pioneer the creation of leading database servers, graph data structures, distributed computing and data repositories (mixture of local development servers and cloud).
- You will join our Agile multi-functional squads and become an advocate of high-quality engineering and best-practice in production software, automation as well as rapid prototypes.
- We trust you - you will have the autonomy to make decisions within this fast paced environment while working in cross-functional squads with great engineering, AI and big data minds.
The role also calls for microservice-style web services, database servers, graph data structures, all of which are deployed onto our Kubernetes cluster.
What skills, experience and qualifications do you need to be successful in this role?
- You are an expert in either Java or Python.
- You have solid programming experience from either commercial environment.
- You have a degree (BSc/MSc) in a relevant field.
- Bonus points for experience with API programming and Docker/Kubernetes. Additional points for experience with: SQL, graph databases, GraphQL, ElasticSearch, Redis, MongoDB, TensorFlow, PyTorch, etc.
- You have some experience of Agile scrum, fast iterative environments, processing of large volume of data, and version control.
- You are a strong advocate of test driven development, automated testing, and other practices for producing resilient software.
- You are a strong communicator who can explain what is required to build and deliver great software products while collaboratively working with the teams.
- You have a problem-solving mind set with an agnostic approach to languages and technology.
- Knowledge of relevant technologies and opinions on what to use such as Java, Python, graph databases, ElasticSearch, Linux, Ansible, Puppet, Chef, Docker, Bamboo, Travis, Terraform, AWS, EC2, Jenkins, Sonar, Git, VMWare, Saltstack, Atlassian, MySQL, NoSQL, Hadoop, Spark, Nginx, Apache, Cassandra, Redis, MongoDB, TitanDB, OrientDB, Kubernetes
Benefits
- Your own MacBook and the newest tech
- $1300 learning budget - when you grow, we grow
- Attendance at industry conferences to stay up to date with newest tech/AI developments
- Performance bonus for some extra $$$
- Fun internal events (boat parties, karting, Oktober Fest, laser tags, bake-offs, bowling, concerts and more!)
- 25 days of holidays (up to 30)
- Health Insurance, Dental and Vision
- Share options
- Relocation/visa assistance
- 401k
We do not offer gym subsidies, kiddi vouchers, Monday breakfasts, nor transportation assistance in our NY office - YET! We are continually working hard on providing you with the best environment, just give us some time to catch up.
What are the Squads?
Depending on your experience and what you are passionate about, you can expect to work across the following:
Backend Squads
Our data pipelines are primarily Java-based, with documents making their way through Cassandra, Spark, before arriving into graph databases and Elastic Search. We also ingest significant amounts of structured data on chemistry and biology from structured databases, such as ChEMBL.
There’s some significant complexity in our data-processing pipeline, so we’re interested in measuring, monitoring, acceptance testing and performance tuning.
We’re enthusiastic adopters of GraphQL which ties all our data and APIs together.
Machine Learning Squads
We primarily write in Python, using libraries such as TensorFlow, PyTorch and Keras for machine learning. We implement cutting-edge research from recent papers with these libraries, as well as producing our own novel research through relationship extraction (RE), named entity recognition (NER) and data pipelining.
NER and RE, you can also say Information Extraction (IE) turns documents from external sources (articles, patents, abstracts etc.) into new edges in our unstructured graph that represent a relation between two entities. Connecting the dots and identifying relationships powers our AI tools.
In particular we apply good engineering practices to the data-preprocessing stages of our machine learning stages, using PySpark for data processing, and systems such as dvc for storing versioned data and models.
Frontend Squads
We have built various interesting applications for our chemical and biological users which include:
- Presenting information in our knowledge graph.
- Triggering AI models and triaging predictions.
- Gathering scientific feedback.
- Designing new molecules.
Sometimes we need to build production-quality frontends, other times we quickly prototype interesting ideas to develop. We build unified API to access our Knowledge Graph and other services.
We make the most of modern frontend tech, including Angular6, Typescript/ES6, RxJS, Flux Pattern, SASS and Material Design. We have in-house UX and Design to make our applications useful and breathtaking.
Deployment Squads
We deploy everything to Kubernetes and RiseML. We manage our own clusters, as well as spinning up additional compute capacity on AWS.
Our in-house-cluster is also home to our Nvidia DGX-1 Deep Learning Supercomputer, which we use to analyse billions of currently theoretical molecules.
by via developer jobs - Stack Overflow
No comments:
Post a Comment