Data Engineer
Posted on March 20, 2025
Job Description
- *Need very quality candite who can perform coding as well in technical round*
- LinkedIn and Email id needed after screening
- Job Title - Data Engineer
- Exp- 6 to 9 years
- Location = Remote
- Duration - 6 Months
- Skills - Apache Spark AWS Big Data (Hadoop/Kafka/Spark)
- ETL pipelines Programing Language (JAVA/Python/Scala)
- Key Responsibilities:
- Design, develop, and maintain ETL pipelines to process and aggregate large
- volumes of data efficiently.
- Optimize ETL jobs for performance and scalability.
- Work with Spark, Python, and SQL to build and maintain data transformation
- processes.
- Implement data schema design and database creation best practices.
- Manage data storage solutions such as S3, Data Lakes, and Data Warehouses.
- Ensure logical access control and user restrictions on data usage.
- Collaborate with the architecture team to understand the platform and its data
- management strategies.
- Perform data extraction, transformation, and loading operations from multiple
- sources.
- Implement best practices for data governance, security, and access control in the
- database.
- Required Skills and Experience:
- Strong experience with ETL tools and frameworks.
- Proficiency in Spark and Python for data processing and analytics.
- Strong SQL skills and experience working with relational and non-relational
- databases.
- Hands-on experience with cloud storage solutions such as AWS S3 and Data
- Lakes.
- Understanding of data schema design, database creation, and role-based access
- control.
- Experience in optimizing ETL jobs for performance and efficiency.
- Knowledge of access control mechanisms and security best practices.
- Strong problem-solving skills and the ability to work with large-scale data
- platforms.
- Preferred Qualifications:
- Experience with distributed computing and big data technologies.
- Knowledge of data warehousing concepts and architecture.
- Familiarity with real-time data processing frameworks.
- Prior experience working in a cloud-based environment (AWS, Azure, or GCP).
- Interview Approach:
- The interview will focus on understanding the candidate's approach to working
- within a product-based environment. Candidates will be expected to demonstrate
- their knowledge of ETL processes, data architecture, optimization techniques, and
- access control strategies.
Required Skills
apache spark aws big data (hadoop/kafka/spark) etl pipelines programing language (java/python/scala)