Data Engineer(Remote)

Posted 2025-10-26
Remote, USA Full Time Immediate Start
Summary Python + Airflow + Snowflake Data Engineer We are seeking a highly skilled Python / Airflow Data Engineer with 7 to 8 years of experience, specifically with a strong background in AWS technologies. The ideal candidate will have a deep understanding of Apache Airflow and its integration within AWS ecosystem, enabling efficient data pipeline orchestration and management. Responsibilities • Design, develop, and maintain complex data pipelines using Python for efficient data processing and orchestration. • Collaborate with cross-functional teams to understand data requirements and architect robust solutions within the AWS environment. • Implement data integration and transformation processes to ensure optimal performance and reliability of data pipelines. • Optimize and fine-tune existing data pipelines / Airflow to improve efficiency, scalability, and maintainability. • Troubleshoot and resolve issues related to data pipelines, ensuring smooth operation and minimal downtime. • Work closely with AWS services like S3, Glue, EMR, Redshift, and other related technologies to design and optimize data infrastructure. • Develop and maintain documentation for data pipelines, processes, and system architecture. • Stay updated with the latest industry trends and best practices related to data engineering and AWS services. Requirements • Bachelor’s degree in Computer Science, Engineering, or a related field. • Proficiency in Python, PySpark and SQL for data processing and manipulation. • Min 5 years of experience in data engineering, specifically working with Apache Airflow and AWS technologies. • Strong knowledge of AWS services, particularly S3, Glue, EMR, Redshift, and AWS Lambda. • Understanding of Snowflake Data Lake is preferred. • Experience with optimizing and scaling data pipelines for performance and efficiency. • Good understanding of data modeling, ETL processes, and data warehousing concepts. • Excellent problem-solving skills and ability to work in a fast-paced, collaborative environment. • Effective communication skills and the ability to articulate technical concepts to non-technical stakeholders. Preferred Qualifications: • AWS certification(s) related to data engineering or big data. • Experience working with big data technologies like Snowflake, Spark, Hadoop, or related frameworks. • Familiarity with other data orchestration tools in addition to Apache Airflow. • Knowledge of version control systems like Bitbucket, Git. Job Types: Full-time, Contractual / Temporary Pay: ₹500,911.64 - ₹1,841,614.32 per year Work Location: Remote Apply tot his job Apply To this Job
Back to Job Board