Data Engineer | Hybrid | 6-15 years

Posted on August 5, 2025

Apply Now

Job Description

  • Data Engineer
  • Budget- 90k
  • Note: Only vendors with an ISO:27001 certificate are eligible to apply and they have Datbrick certified candidates.
  • Job description:
  • We are looking for people who are self-motivated, responsive individuals who are passionate about data. You will build data solutions to address complex business questions. You�ll take data through its lifecycle, from the pipeline for data processing, data infrastructure, and create dataset data products.
  • Core responsibilities:
  • Monitor and optimize the performance of data pipelines and queries, addressing any bottlenecks and improving efficiency.
  • Design, develop, and maintain robust data pipelines using SQL, Python, and PySpark.
  • Partner with business team(s) to understand the business requirements, understand the impact to existing systems and design and implement new data provisioning pipeline process for Finance/ External reporting domains.
  • Monitor and troubleshoot operational or data issues in the data pipelines
  • Drive architectural plans and implementation for future data storage, reporting, and analytic solutions
  • Work closely with data analysts, data scientists, and other stakeholders to understand data needs and deliver high-quality data solutions.
  • Identify, troubleshoot, and resolve data-related issues, ensuring high availability and reliability of data systems.
  • Qualifications:
  • 5+ years of relevant work experience in data engineering or software engineering equivalent
  • 3+ years of experience in implementing big data processing technology: AWS / Azure / GCP, Apache Spark, Python
  • Experience writing and optimizing SQL queries in a business environment with large-scale, complex datasets
  • Hands-on experience with PySpark for big data processing, including data frame operations, Spark SQL, and Spark Streaming.
  • Proven experience with SQL, Python, and PySpark in a production environment.
  • Experience with big data technologies such as Hadoop, Spark, and Kafka are a plus.
  • Monitor and troubleshoot data pipeline issues to ensure seamless data flow.
  • working experience in a DevOps CI/CD environment
  • Data Brick Certificate is imortent t.o apply for this profile

Required Skills

datbrick certified