Job description

The Data Engineer will be a part of an international team that designs, develops and delivers new applications for Koch Industries. Koch Industries is a privately held global organization with over 120,000 employees around the world, with subsidiaries involved in manufacturing, trading, and investments. Koch Global Services (KGS) is being developed in India to as a Shared Service Operations, as well as act as a hub for innovation across functions. As KGS rapidly scales up its operations in India, it’s employees will get opportunities to carve out a career path for themselves within the organization. This role will have the opportunity to join on the ground floor and will play a critical part in helping build out the Koch Global Services (KGS) over the next several years. Working closely with global colleagues would provide significant international exposure to the employees. We are seeking a Data Engineer expert to join KGS Analytics capability. We love passionate, forward thinking individuals who are driven to innovate. You will have the opportunity to engage with Business Analysts, Analytics Consultants, and internal customers to implement ideas, optimize existing data pieplines, and create data products using powerful, contemporary tools. This opportunity engages diverse types of business applications and data sets at a rapid pace, and our ideal candidate gets excited when they are faced with a challenge. What value does this role provide to KGS? Being a part of data engineering team, focus on designing ELT solutions for data consumers in the business. Create value through provide consumable data, by measuring progress against goals, and through data grounded situational awareness. Why would a candidate want this role? If a candidate is entrepreneurial in the way they approach ideas, Koch is among the most fulfilling organizations they could join. We are growing an analytics capability and looking for entrepreneurial minded innovators who can help us further develop this service of exceptionally high value to our business. Due to the diversity of companies and work within Koch, we are frequently working in new and interesting global business spaces, with data and analytics applications that are unique relative to opportunities from other employers in the marketplace. Define what success looks like Bringing forward innovative and valuable ideas relevant to our mission of creating value through analytics; developing complex analysis products at scale, with high accuracy. Developing valuable, relevant, and accurate analytics products that internal customers use in their decision-making processes. How will this role prepare the candidate for career growth? There are few better places to be able to work with world class tools and unique, value-oriented problem types at this scale. As a result, the successful candidate will develop advanced knowledge of best-in-class tools and techniques, while expanding their capability for analytic, entrepreneurial thinking that will benefit them in achieving their career aspirations. Describe your top performer. Enthusiastically collaborative, value seeking engineer whose exceptional technical skills are only surpassed by their appetite for learning and innovation. Open to challenge and be challenged with new ideas and established approaches. Rapidly prototypes analytic approaches that explore for opportunity or meet customer needs, and readily translate into production.

A Day In The Life Could Include:

(job responsibilities)

What You Will Need To Bring With You:

(experience & Education Required)

  • Work with data partners to understand key business drivers and use that knowledge to experiment and transform enterprise data platform to capture the value of logistic data.
  • Design, build and maintain optimal data pipeline architecture, assemble large, complex data sets that meet functional / non-functional business requirements.
  • Improve data pipeline reliability, scalability, and security existing solutions.
  • Work closely with the Product Owners and stake holders to design the Technical Architecture for data


platform to meet the requirements of the proposed solution.

  • Design & document technical solutions in Big Data space (Stack like Spark, M/R, HDFS,NoSQL stores like
  • At least 5+ years of Data Engineering experience (AWS) in delivering Advance Analytics solution, Data Warehousing, Big Data or Cloud.
  • Should have strong knowledge in SQL & python, developing, deploying, and modelling DWH and data pipelines on AWS cloud or similar other cloud environments.
  • 5+ years of experience with business and technical requirements analysis, elicitation, data modeling, verification, and methodology development with a good hold of communicating complex technical ideas to technical and non-technical team members MongoDB, HBase etc, Databased like Snowflake, Redshift etc)
  • Assist in developing, documenting & implementing consistent processes for data modeling, mining, and till production implementation
  • Experience in architecting data engineering solution on public cloud.
  • Design & implement quality check guidelines for data validation.
  • Help the Data Engineering team produce high-quality code that allows us to put solutions into production
  • Creating reusable and scalable data pipelines solutions
  • Focus on implementing development processes and tools that allow for the collection of metadata, access to metadata, and completed in a way that allows for widespread code reuse (e.g., utilization of ETL Frameworks, Generic Metadata driven Tools, shared data dimensions, etc.) that will enable impact analysis as well as source to target tracking and reporting
  • Manage data related requests, analyze issues and provide efficient resolution. Design all program specifications and perform required tests.
  • Experience in authoring or reviewing system design documents for enterprise solutions.
  • Knowledge of Big Data technologies, such as Spark, Hadoop/MapReduce.
  • Demonstrated experience using git-based source control management platforms (Gitlab, GitHub, DevOps, etc.).
  • Experience of working in Agile delivery
  • Experience in Data Harmonization, Master Data Management & Critical Data Elements Management


What Will Put You Ahead:

(experience & education preferred) Other Considerations: (physical demands/ unusual working conditions)

  • 5+ years’ experience in the Amazon Web Services stack experience including S3, Athena, Redshift, Glue
  • 5+ years’ experience with cloud data warehousing solutions including Snowflake with developing in and implementation of dimensional modeling
  • Experience with enterprise data tools & integration with AWS platform will be preferred
  • Certified as data engineer from a reputed public cloud.
  • Development experience with docker and a


Kubernetes environment (would be a plus)

  • Understanding of infrastructure (including hosting, container-based deployments and storage architectures) would be advantageous.

Please let the company know that you found this position on this Job Board as a way to support us, so we can keep posting cool jobs.