Create, maintain, and deploy Python packages for existing research code repositories.
Implement scalable data-intensive processing pipelines that applies geospatial data to ML/DL models.
Develop new data visualization tooling and notebooks to be used for exploratory data analysis.
Develop Spark functionality into existing Python tooling.
Conduct reviews and walkthroughs of existing code repositories with team members.
Write beautiful, robust, well-documented code.
Transition exploratory research into scalable processing pipelines for further development.
Collaborate with engineering teams (internal stakeholders and external partners) working with geospatial platforms to implement models and algorithms at scale for production.
Basic Qualifications:
B.S./B.A. in Computer Science, Engineering or related field.
Experience working and visualization of geospatial datasets and maps.
Excellent communication and collaboration skills.
Strong development experience in Python, Git, and Docker.
Ability to write queries either in SQL, NoSQL and Big Query to perform ETL on datasets.
Preferred Qualifications:
Diverse data structure experience.
Experience working with cloud tools such as AWS, Azure, or Domino.
Experience with PySpark and handling large datasets.
Experience with web services, api and database design.
Ability to handle multiple, competing priorities in a fast-paced development environment.