Senior Python ETL Engineer

Описание вакансии
EPAM is on the lookout for a talented Python ETL Engineer to join the team. In this role, you'll be the wizard behind the curtain, designing and implementing ETL processes that weave data magic. Extracting, transforming and loading data from various sources into our data warehouse will be your domain. Your expertise will be pivotal in understanding business requirements, crafting data models, and ensuring the accuracy and quality of our data.

Got a passion for Python? Love the thrill of creating ETL pipelines that run like clockwork? Perfect! Your solid Python development background, advanced programming knowledge and expertise with Airflow or Luigi will drive our data-driven success. And let's not forget your experience in data modeling, database design, and SQL – a true game-changer.

So, if you're passionate about ETL development and crave the excitement of cutting-edge technologies, we can't wait to welcome you aboard!
Требования
  • Bachelor's degree in Computer Science, Information Technology or a related field
  • 3+ years of experience in Python development
  • Advanced Python programming language knowledge with an understanding of internal mechanics (interpreter, memory management, concurrency, etc.)
  • Ability to develop scalable, maintainable and reliable ETL code
  • Proficiency with ETL frameworks such as Airflow or Luigi
  • Familiarity with data extraction from diverse sources such as databases, APIs, and flat files and transforming them to meet business requirements
  • Strong understanding of data modeling and database design, including experience with dimensional modeling and star schemas
  • Knowledge of SQL and relational databases such as PostgreSQL, SQL Server, and Oracle, including experience with database optimization techniques
  • Showcase of working with cloud technologies such as AWS, Azure or Google Cloud Platform English – B1+, both verbal and written
  • Competency in maintaining CI/CD pipelines for ETL applications
  • Expertise in web applications for data ingestion using Flask or FastAPI
  • Knowledge of Agile development methodologies (SCRUM/Kanban)
  • Availability of Cloud certification (AWS, GCP, Azure)
  • Familiarity with Docker and Docker Compose
  • Flexibility to use big data technologies such as Hadoop, Spark
Обязанности
  • Design and implement ETL processes
  • Understand business requirements and translate them into technical specifications
  • Develop custom ETL scripts using Python
  • Maintain ETL pipelines using frameworks such as Airflow or Luigi
  • Ensure data accuracy and quality by implementing ETL best practices such as data cleansing, data validation, and error handling
  • Work with data warehouses such as Redshift, Snowflake, or BigQuery to store and manage large data sets
  • Run SQL and relational databases such as PostgreSQL, SQL Server, and Oracle to design and optimize database schemas and queries
  • Operate cloud technologies such as AWS, Azure, or Google Cloud Platform, including cloud-based ETL services such as AWS Glue or Azure Data Factory
  • Collaborate with cross-functional teams to ensure the successful delivery of projects
  • Stay up-to-date with industry trends and best practices in ETL development
Добавлено 12 дней назад
Для связи с работодателем или просмотра контактов нажмите на кнопку