Senior Data Engineer

Yonkers, NY, USA

Ranked as #12 on Forbes’ List of 25 Fastest Growing Public Tech Companies for 2017, EPAM is committed to providing our global team of 25,900+ EPAMers with inspiring careers from day one. EPAMers lead with passion and honesty and think creatively. Our people are the source of our success and we value collaboration, try to always understand our customers’ business, and strive for the highest standards of excellence. No matter where you are located, you’ll join a dedicated, diverse community that will help you discover your fullest potential.


You are curious, persistent, logical and clever – a true techie at heart. You enjoy living by the code of your craft and developing elegant solutions for complex problems. If this sounds like you, this could be the perfect opportunity to join EPAM as a Senior Data Engineer. Scroll down to learn more about the position’s responsibilities and requirements.

We are looking for an engineer with a software engineering background and experience with data projects with Strong Knowledge and Hands-on experience in building ETL systems and data-pipelines with a combination of below skill sets:

What You’ll Do

  • Design, deploy, provide performance tune and monitor analytical databases;
  • Design, build, and launch new data models, data pipelines and ETLs that ingest, map, aggregate, and broadcast data feeds to internal and external groups;
  • Design, develop, and deploy various data visualizations;
  • Analyze business requirements, provide development estimates and feedback, and determine implementation/execution approaches;
  • Work with other software engineers, database developers, infrastructure, operations and QA teams to design and implement quality software solutions;
  • Support, troubleshoot, monitor and optimize existing data engineering systems;
  • Help streamline software development by contributing to the implementation of a continuous integration process, automating manual processes, and eliminating defects.

What You Have

  • Programming skills - Python, PySpark, and/or Java/JEE;
  • Experience with cloud technologies on AWS used for data processing - EC2, S3, Glue, Athena, EMR, Lambda etc;
  • Experience with Notebooks (Zeppelin, Jupyter);
  • Good knowledge of DBMS concepts with hands-on experience with Nosql Databases (MongoDB), and Relational Databases (Oracle);
  • Experience with technologies such as Apache Spark, Kafka, Flume, Sqoop would be a plus;
  • Experience with data visualizations software (Tableau/Qlikview) would be a plus;
  • Experience working in projects involving building Data Warehouse, Data Lakes, Big data ecosystem (Hadoop, HDFS, Hive, and Hbase ) and Data Science would be a plus.

What We Offer

  • Medical, Dental and Vision Insurance (Subsidized);
  • Health Savings Account;
  • Flexible Spending Accounts (Healthcare, Dependent Care, Commuter);
  • Short-Term and Long-Term Disability (Company Provided);
  • Life and AD&D Insurance (Company Provided);
  • Matched 401(k) Retirement Savings Plan;
  • Paid Time Off;
  • Legal Plan and Identity Theft Protection;
  • Accident Insurance;
  • Employee Discounts;
  • Pet Insurance.