Data EngineerUS Citizen/Green CardW2 ONLY
There is an enormous need and potential here to do something that has never been done before. We need great people to help transform these complex scientific datasets into innovative software that is deployed across the pipeline, accelerating the pace and quality of all crop system development decisions to unbelievable levels.
What you will do is why you should join us:Be a critical senior member of a data engineering team focused on creating distributed analysis capabilities around a large variety of datasetsTake pride in software craftsmanship, apply a deep knowledge of algorithms and data structures to continuously improve and innovateWork with other top-level talent solving a wide range of complex and unique challenges that have real world impactExplore relevant technology stacks to find the best fit for each datasetPursue opportunities to present our work at relevant technical conferencesGoogle Cloud Next 2019: https://www.youtube.com/watch?v=fqvuyOID6v4GraphConnect 2015: https://www.youtube.com/watch?v=6KEvLURBenMGoogle Cloud Blog: https://cloud.google.com/blog/products/containers-kubernetes/google-kubernetes-engine-clusters-can-have-up-to-15000-nodesProject your talent into relevant projects. Strength of ideas trumps position on an org chart
RequiredAt least 7 years experience in software engineeringAt least 2 years experience with GoProven experience (2 years) building and maintaining data-intensive APIs using a RESTful approachExperience with stream processing using Apache KafkaA level of comfort with Unit Testing and Test Driven Development methodologiesFamiliarity with creating and maintaining containerized application deployments with a platform like DockerA proven ability to build and maintain cloud based infrastructure on a major cloud provider like AWS, Azure or Google Cloud PlatformExperience data modeling for large scale databases, either relational or NoSQL
Great to have:Experience with protocol buffers and gRPCExperience with: Google Cloud Platform, Apache Beam and or Google Cloud Dataflow, Google Kubernetes Engine or KubernetesExperience working with scientific datasets, or a background in the application of quantitative science to business problemsBioinformatics experience, especially large scale storage and data mining of variant data, variant annotation, and genotype to phenotype correlation
Location: Creve Coeur, MO or Remote from another location.