Data Engineer

Posted on March 20, 2025

Apply Now

Job Description

  • *Need very quality candite who can perform coding as well in technical round*
  • LinkedIn and Email id needed after screening
  • Job Title - Data Engineer
  • Exp- 6 to 9 years
  • Location = Remote
  • Duration - 6 Months
  • Skills - Apache Spark AWS Big Data (Hadoop/Kafka/Spark)
  • ETL pipelines Programing Language (JAVA/Python/Scala)
  • Key Responsibilities:
  • Design, develop, and maintain ETL pipelines to process and aggregate large
  • volumes of data efficiently.
  • Optimize ETL jobs for performance and scalability.
  • Work with Spark, Python, and SQL to build and maintain data transformation
  • processes.
  • Implement data schema design and database creation best practices.
  • Manage data storage solutions such as S3, Data Lakes, and Data Warehouses.
  • Ensure logical access control and user restrictions on data usage.
  • Collaborate with the architecture team to understand the platform and its data
  • management strategies.
  • Perform data extraction, transformation, and loading operations from multiple
  • sources.
  • Implement best practices for data governance, security, and access control in the
  • database.
  • Required Skills and Experience:
  • Strong experience with ETL tools and frameworks.
  • Proficiency in Spark and Python for data processing and analytics.
  • Strong SQL skills and experience working with relational and non-relational
  • databases.
  • Hands-on experience with cloud storage solutions such as AWS S3 and Data
  • Lakes.
  • Understanding of data schema design, database creation, and role-based access
  • control.
  • Experience in optimizing ETL jobs for performance and efficiency.
  • Knowledge of access control mechanisms and security best practices.
  • Strong problem-solving skills and the ability to work with large-scale data
  • platforms.
  • Preferred Qualifications:
  • Experience with distributed computing and big data technologies.
  • Knowledge of data warehousing concepts and architecture.
  • Familiarity with real-time data processing frameworks.
  • Prior experience working in a cloud-based environment (AWS, Azure, or GCP).
  • Interview Approach:
  • The interview will focus on understanding the candidate's approach to working
  • within a product-based environment. Candidates will be expected to demonstrate
  • their knowledge of ETL processes, data architecture, optimization techniques, and
  • access control strategies.

Required Skills

apache spark aws big data (hadoop/kafka/spark) etl pipelines programing language (java/python/scala)