For companies
  • Hire developers
  • Hire designers
  • Hire marketers
  • Hire product managers
  • Hire project managers
  • Hire assistants
  • How Arc works
  • How much can you save?
  • Case studies
  • Pricing
    • Remote dev salary explorer
    • Freelance developer rate explorer
    • Job description templates
    • Interview questions
    • Remote work FAQs
    • Team bonding playbooks
    • Employer blog
For talent
  • Overview
  • Remote jobs
  • Remote companies
    • Resume builder and guide
    • Talent career blog
Cummins India
Cummins India

Data Engineer - Senior

Location

Remote restrictions apply
See all remote locations

Salary Estimate

N/AIconOpenNewWindows

Seniority

Senior

Tech stacks

Data
Database
Testing
+53

Permanent role
8 days ago
Apply now

Description

GPP Database Link (https://cummins365.sharepoint.com/sites/CS38534/)

Job Summary

Leads projects for the design, development, and maintenance of a data and analytics platform. Effectively and efficiently processes, stores, and makes data available to analysts and other consumers. Works with key business stakeholders, IT experts, and subject-matter experts to plan, design, and deliver optimal analytics and data science solutions. Works on one or many product teams at a time. Though the role category is generally listed as Remote, this specific position is designated as Hybrid.

Key Responsibilities

  • Business Alignment & Collaboration – Partner with the Product Owner to align data solutions with strategic goals and business requirements.
  • Data Pipeline Development & Management – Design, develop, test, and deploy scalable data pipelines for efficient data transport into Cummins Digital Core (Azure DataLake, Snowflake) from various sources (ERP, CRM, relational, event-based, unstructured).
  • Architecture & Standardization – Ensure compliance with AAI Digital Core and AAI Solutions Architecture standards for data pipeline design and implementation.
  • Automation & Optimization – Design and automate distributed data ingestion and transformation systems, integrating ETL/ELT tools and scripting languages to ensure scalability, efficiency, and quality.
  • Data Quality & Governance – Implement data governance processes, including metadata management, access control, and retention policies, while continuously monitoring and troubleshooting data integrity issues.
  • Performance & Storage Optimization – Develop and implement physical data models, optimize database performance (indexing, table relationships), and operate large-scale distributed/cloud-based storage solutions (Data Lakes, Hadoop, HBase, Cassandra, MongoDB, Accumulo, DynamoDB).
  • Innovation & Tool Evaluation – Conduct proof-of-concept (POC) initiatives, evaluate new data tools, and provide recommendations for improvements in data management and integration.
  • Documentation & Best Practices – Maintain standard operating procedures (SOPs) and data engineering documentation to support consistency and efficiency.
  • Agile Development & Automation – Use Agile methodologies (DevOps, Scrum, Kanban) to drive automation in data integration, preparation, and infrastructure management, reducing manual effort and errors.
  • Coaching & Team Development – Provide guidance and mentorship to junior team members, fostering skill development and knowledge sharing.

Responsibilities

Competencies:

  • System Requirements Engineering: Translates stakeholder needs into verifiable requirements, tracks status, and assesses impact changes.
  • Collaborates: Builds partnerships and works collaboratively with others to meet shared objectives.
  • Communicates Effectively: Delivers multi-mode communications tailored to different audiences.
  • Customer Focus: Builds strong customer relationships and provides customer-centric solutions.
  • Decision Quality: Makes good and timely decisions that drive the organization forward.
  • Data Extraction: Performs ETL activities from various sources using appropriate tools and technologies.
  • Programming: Develops, tests, and maintains code using industry standards, version control, and automation tools.
  • Quality Assurance Metrics: Measures and assesses solution effectiveness using IT Operating Model (ITOM) standards.
  • Solution Documentation: Documents knowledge gained and communicates solutions for improved productivity.
  • Solution Validation Testing: Validates configurations and solutions to meet customer requirements using SDLC best practices.
  • Data Quality: Identifies, corrects, and manages data flaws to support effective governance and decision-making.
  • Problem Solving: Uses systematic analysis to determine root causes and implement robust solutions.
  • Values Differences: Recognizes and leverages the value of diverse perspectives and cultures.

Education, Licenses, Certifications

  • Bachelor's degree in a relevant technical discipline, or equivalent experience required.
  • This position may require licensing for compliance with export controls or sanctions regulations.

Qualifications

Preferred Experience:

  • Technical Expertise – Intermediate experience in data engineering with hands-on knowledge of SPARK, Scala/Java, MapReduce, Hive, HBase, Kafka, and SQL.
  • Big Data & Cloud Solutions – Proven ability to design and develop Big Data platforms, manage large datasets, and implement clustered compute solutions in cloud environments.
  • Data Processing & Movement – Experience developing applications requiring large-scale file movement and utilizing various data extraction tools in cloud-based environments.
  • Business & Industry Knowledge – Familiarity with analyzing complex business systems, industry requirements, and data regulations to ensure compliance and efficiency.
  • Analytical & IoT Solutions – Experience building analytical solutions with exposure to IoT technology and its integration into data engineering processes.
  • Agile Development – Strong understanding of Agile methodologies, including Scrum and Kanban, for iterative development and deployment.
  • Technology Trends – Awareness of emerging technologies and trends in data engineering, with a proactive approach to innovation and continuous learning.

Technical Skills

  • Programming Languages: Proficiency in Python, Java, and/or Scala.
  • Database Management: Expertise in SQL and NoSQL databases.
  • Big Data Technologies: Hands-on experience with Hadoop, Spark, Kafka, and similar frameworks.
  • Cloud Services: Experience with Azure, Databricks, and AWS platforms.
  • ETL Processes: Strong understanding of Extract, Transform, Load (ETL) processes.
  • Data Replication: Working knowledge of replication technologies like Qlik Replicate is a plus.
  • API Integration: Experience working with APIs to consume data from ERP and CRM systems.

Job Systems/Information Technology

Organization Cummins Inc.

Role Category Remote

Job Type Exempt - Experienced

ReqID 2410681

Relocation Package No

About Cummins India

🔗Website
Visit company profileIconOpenNewWindows

Unlock all Arc benefits!

  • Browse remote jobs in one place
  • Land interviews more quickly
  • Get hands-on recruiter support
PRODUCTS
Arc

The remote career platform for talent

Codementor

Find a mentor to help you in real time

LINKS
About usPricingArc Careers - Hiring Now!Remote Junior JobsRemote jobsCareer Success StoriesTalent Career BlogArc Newsletter
JOBS BY EXPERTISE
Remote Front End Developer JobsRemote Back End Developer JobsRemote Full Stack Developer JobsRemote Mobile Developer JobsRemote Data Scientist JobsRemote Game Developer JobsRemote Data Engineer JobsRemote Programming JobsRemote Design JobsRemote Marketing JobsRemote Product Manager JobsRemote Project Manager JobsRemote Administrative Support Jobs
JOBS BY TECH STACKS
Remote AWS Developer JobsRemote Java Developer JobsRemote Javascript Developer JobsRemote Python Developer JobsRemote React Developer JobsRemote Shopify Developer JobsRemote SQL Developer JobsRemote Unity Developer JobsRemote Wordpress Developer JobsRemote Web Development JobsRemote Motion Graphic JobsRemote SEO JobsRemote AI Jobs
© Copyright 2025 Arc
Cookie PolicyPrivacy PolicyTerms of Service