Python Developer
Posted on July 10, 2025
Job Description
- Required Skill set:
- 1. SQL Databases
- 2. Databricks
- 3. AWS
- 4. High Volume data stream
- About the Role
- * Build Data lake and Data warehouse infrastructure to enable the users to execute business intelligence, data analytics, data science and ML/AI workloads
- * Design, develop and maintain data pipelines for extraction, transformation, and loading of data from a wide variety of data sources.
- * Build workflows to run periodic jobs on our data infrastructure
- * Build frameworks for data processing jobs to solve data ingestion , data quality and feature engineering
- * Building observability over the data pipelines to maintain consistent data with an agreed-upon latency
- * Ideate processes to maintain the health of the infrastructure such as CPU, memory monitoring, query monitoring, workload monitoring and educate the teams to follow the best Practices.
- * Work closely with data engineering leads and Data Architects
- * Keep the data secure from the external world and have fine-grained access control within the organization to make sure that only relevant data is available to every user.
- * Create or integrate with tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
- Responsibilities of the role
- * Proven experience in a Data Engineering role with at least 4+ years of experience
- * Proficient in Python, Java or Scala
- * Good understanding of SQL and NoSQL databases
- * Strong analytical skills to break down complex problems and build the right solutions
- * Hands-on working experience with Python, ETL pipelines, advanced SQL
- * Experience in any Cloud DW Redshift/Snowflake/BigQuery
- * Experience building and optimizing big data' data pipelines, architectures and datasets.
- * Strong understanding of AWS Services - Redshift, Lambda, Glue, Athena, security protocols
- * Experience with cloud platforms like AWS/Azure/GCP
- * Strong data Modeling and database design experience with Redshift or other relational databases
- * Good knowledge of big data tools such as Hadoop, Apache Spark, Apache Druid,S3, Glue, Athena, Flink, Airflow, etc.
- * Good exposure to streaming technologies like Kafka, SQS, Kinesis, Apache Spark Structured Streaming
- * Experience in gathering and processing raw data at scale including writing scripts, web scraping, calling APIs
- * Experience in supporting and working with data analysts, data scientists, quality analysts dynamic environment
Required Skills
No specific skills listed.