Job Title: Sr Python Developer & Lead
3;Location: Auburn Hills, MI- Onsite
3;Remote work: No
3;
3;Mandatory Skills: Data Engineering, Python, PySpark, CI/CD, Airflow, Workflow Orchestration
3;
3;Key Responsibilities
3;1. Data Engineering: Design, develop, and optimize scalable data pipelines using Python and PySpark for batch and streaming workloads.
3;2. Workflow Orchestration: Build, schedule, and monitor complex workflows using Airflow, ensuring reliability and maintainability.
3;3. CI/CD Pipeline Development: Architect and implement CI/CD pipelines for data engineering projects using GitHub, Docker, and cloud-native solutions.
3;4. Testing & Quality: Apply test-driven development (TDD) practices and automate unit/integration tests for data pipelines.
3;5. Secure Development: Implement secure coding best practices and design patterns throughout the development lifecycle.
3;6. Collaboration: Work closely with Data Architects, QA teams, and business stakeholders to translate requirements into technical solutions.
3;7. Documentation: Create and maintain technical documentation, including process/data flow diagrams and system design artifacts.
3;8. Mentorship: Lead and mentor junior engineers, providing guidance on coding, testing, and deployment best practices.
3;9. Troubleshooting: Analyze and resolve technical issues across the data stack, including pipeline failures and performance bottlenecks.
3;
3;
3;Technical Experience:
3;1. Hands-on Data Engineering: Minimum 5+ yearsof practical experience building production-grade data pipelines using Python and PySpark.
3;2. Airflow Expertise: Proven track record of designing, deploying, and managing Airflow DAGs in enterprise environments.
3;3. CI/CD for Data Projects: Ability to build and maintain CI/CD pipelinesfor data engineering workflows, including automated testing and deployment**.
3;4. Cloud & Containers: Experience with containerization (Docker and cloud platforms (GCP) for data engineering workloads. Appreciation for twelve-factor design principles
3;5. Python Fluency: Ability to write object-oriented Python code manage dependencies, and follow industry best practices
3;6. Version Control: Proficiency with **Git** for source code management and collaboration (commits, branching, merging, GitHub/GitLab workflows).
3;7. Unix/Linux: Strong command-line skills** in Unix-like environments.
3;8. SQL: Solid understanding of SQL for data ingestion and analysis.
3;9. Collaborative Development: Comfortable with code reviews, pair programming and usingremote collaboration tools effectively.
3;10. Engineering Mindset: Writes code with an eye for maintainability and testability; excited to build production-grade software
3;11. Education: Bachelor’s or graduate degree in Computer Science, Data Analytics or related field, or equivalent work experience.
3;
3;• A high tolerance for OpenShift, Cloudera, Tableau, Confluence, Jira, and other enterprise tools.
3;Location: Auburn Hills, MI- Onsite
3;Remote work: No
3;
3;Mandatory Skills: Data Engineering, Python, PySpark, CI/CD, Airflow, Workflow Orchestration
3;
3;Key Responsibilities
3;1. Data Engineering: Design, develop, and optimize scalable data pipelines using Python and PySpark for batch and streaming workloads.
3;2. Workflow Orchestration: Build, schedule, and monitor complex workflows using Airflow, ensuring reliability and maintainability.
3;3. CI/CD Pipeline Development: Architect and implement CI/CD pipelines for data engineering projects using GitHub, Docker, and cloud-native solutions.
3;4. Testing & Quality: Apply test-driven development (TDD) practices and automate unit/integration tests for data pipelines.
3;5. Secure Development: Implement secure coding best practices and design patterns throughout the development lifecycle.
3;6. Collaboration: Work closely with Data Architects, QA teams, and business stakeholders to translate requirements into technical solutions.
3;7. Documentation: Create and maintain technical documentation, including process/data flow diagrams and system design artifacts.
3;8. Mentorship: Lead and mentor junior engineers, providing guidance on coding, testing, and deployment best practices.
3;9. Troubleshooting: Analyze and resolve technical issues across the data stack, including pipeline failures and performance bottlenecks.
3;
3;
3;Technical Experience:
3;1. Hands-on Data Engineering: Minimum 5+ yearsof practical experience building production-grade data pipelines using Python and PySpark.
3;2. Airflow Expertise: Proven track record of designing, deploying, and managing Airflow DAGs in enterprise environments.
3;3. CI/CD for Data Projects: Ability to build and maintain CI/CD pipelinesfor data engineering workflows, including automated testing and deployment**.
3;4. Cloud & Containers: Experience with containerization (Docker and cloud platforms (GCP) for data engineering workloads. Appreciation for twelve-factor design principles
3;5. Python Fluency: Ability to write object-oriented Python code manage dependencies, and follow industry best practices
3;6. Version Control: Proficiency with **Git** for source code management and collaboration (commits, branching, merging, GitHub/GitLab workflows).
3;7. Unix/Linux: Strong command-line skills** in Unix-like environments.
3;8. SQL: Solid understanding of SQL for data ingestion and analysis.
3;9. Collaborative Development: Comfortable with code reviews, pair programming and usingremote collaboration tools effectively.
3;10. Engineering Mindset: Writes code with an eye for maintainability and testability; excited to build production-grade software
3;11. Education: Bachelor’s or graduate degree in Computer Science, Data Analytics or related field, or equivalent work experience.
3;
3;• A high tolerance for OpenShift, Cloudera, Tableau, Confluence, Jira, and other enterprise tools.