Location
Chennai | India
Job description
- As a Data Engineer, you will play a key role in designing, developing, and maintaining our data infrastructure and pipelines
- You will collaborate closely with the rest of our standalone Data team (comprising analysts, analytics engineers, and data governance experts), and with engineering and operations teams across Arcadia,, to ensure the smooth flow and availability of high-quality data for analysis and reporting purposes
- Your expertise will be essential in optimizing data workflows, ensuring data integrity, and scaling our data infrastructure to support our company's growth
- This is an exceptional opportunity for someone who relishes the chance to engage with cutting-edge technology, influence the development of a world-class data ecosystem and work in a fast-paced environment on a high-morale, tightly knit team
- Our core data stack makes heavy use of Snowflake and dbt, orchestrated in Prefect in our broader AWS-based ecosystem
- Most of our wide range of data sources are loaded with Fivetran, but we use custom Python when it s the right tool for the job
What you'll do:
- Design, develop, and maintain scalable and efficient data pipelines in an AWS environment, centered on our Snowflake instance and using Fivetran, Prefect and dbt.
- Collaborate with data analysts, analytics engineers, and software engineers to understand data requirements and deliver reliable solutions.
- Design, build and maintain the tooling that the wider Data team (and Arcadia as a whole) uses to interact with our data platform, including CI/CD pipelines for our data lakehouse, unit/integration/validation testing frameworks for our data pipelines, and command-line tools for ad-hoc data evaluation.
- Identify and implement best practices for data ingestion, transformation, and storage to ensure data integrity and accuracy.
- Optimize and tune data pipelines for improved performance, scalability, and reliability.
- Monitor data pipelines and proactively address any issues or bottlenecks to ensure uninterrupted data flow.
- Develop and maintain documentation for data pipelines, ensuring knowledge sharing and smooth onboarding of new team members.
What will help you succeed
Must-haves:
- 3+ years as a Data Engineer, data-adjacent Software Engineer, or a did-everything small data team member with a focus on building and maintaining data pipelines.
- Strong Python skills, especially in the context of data orchestration.
- Proficiency in SQL
- Familiarity with data integration patterns, ETL/ELT processes, and data warehousing concepts.
- Experience with Prefect, Airflow, or similar data orchestration tools.
- Excellent problem-solving and analytical skills with a strong attention to detail.
- Ability to bring a customer-oriented and empathetic approach to understanding how data is used to drive the business.
- Strong communication skills.
Nice-to-haves:
- Undergraduate and/or graduate degree in math, statistics, engineering, computer science, or related technical field
- Experience with our stack-AWS, Snowflake, Fivetran, Prefect, dbt, and Github Actions, along with some ancillary tools
- Experience with DevOps practices, especially CI/CD
- Previous experience managing enterprise-level data pipelines and working with large datasets
- Strong understanding of database management and design, including experience with Snowflake or an equivalent platform.
- Experience in the energy sector
Benefits:
- Competitive compensation based on market standards
- We are working on a hybrid model
Apart from Fixed Base Salary , the candidates are eligible for following benefits
- Flexible Leave Policy
- Office Located in the heart of the City in case you need to step in for any purpose
- Medical Insurance (1+5 Family Members)
- Quarterly team engagement activities and rewards & recognitions
- L&D programs to foster professional growth
Job tags
Salary