Python Developer

Posted on July 10, 2025

Apply Now

Job Description

  • Required Skill set:
  • 1. SQL Databases
  • 2. Databricks
  • 3. AWS
  • 4. High Volume data stream
  • About the Role
  • * Build Data lake and Data warehouse infrastructure to enable the users to execute business intelligence, data analytics, data science and ML/AI workloads
  • * Design, develop and maintain data pipelines for extraction, transformation, and loading of data from a wide variety of data sources.
  • * Build workflows to run periodic jobs on our data infrastructure
  • * Build frameworks for data processing jobs to solve data ingestion , data quality and feature engineering
  • * Building observability over the data pipelines to maintain consistent data with an agreed-upon latency
  • * Ideate processes to maintain the health of the infrastructure such as CPU, memory monitoring, query monitoring, workload monitoring and educate the teams to follow the best Practices.
  • * Work closely with data engineering leads and Data Architects
  • * Keep the data secure from the external world and have fine-grained access control within the organization to make sure that only relevant data is available to every user.
  • * Create or integrate with tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  • Responsibilities of the role
  • * Proven experience in a Data Engineering role with at least 4+ years of experience
  • * Proficient in Python, Java or Scala
  • * Good understanding of SQL and NoSQL databases
  • * Strong analytical skills to break down complex problems and build the right solutions
  • * Hands-on working experience with Python, ETL pipelines, advanced SQL
  • * Experience in any Cloud DW Redshift/Snowflake/BigQuery
  • * Experience building and optimizing big data' data pipelines, architectures and datasets.
  • * Strong understanding of AWS Services - Redshift, Lambda, Glue, Athena, security protocols
  • * Experience with cloud platforms like AWS/Azure/GCP
  • * Strong data Modeling and database design experience with Redshift or other relational databases
  • * Good knowledge of big data tools such as Hadoop, Apache Spark, Apache Druid,S3, Glue, Athena, Flink, Airflow, etc.
  • * Good exposure to streaming technologies like Kafka, SQS, Kinesis, Apache Spark Structured Streaming
  • * Experience in gathering and processing raw data at scale including writing scripts, web scraping, calling APIs
  • * Experience in supporting and working with data analysts, data scientists, quality analysts dynamic environment

Required Skills

No specific skills listed.