Job Location : Bangalore
Experience : 7 Yr
CTC Budget : 1000000 to 2400000
Posted At : 15-Oct-2025
We are Looking for Lead Data Engineer
Required Skills-
Who have 7+ years of Experience in Data engineering
5+ years in Data analysis
4+ years in Data Structure
5+ years in Data Warehousing
5+ years in AWS Cloud data warehousing
4+ years in Python
Detailed JD-
Cloud Data Architecture & Design (AWS-focused) Scalable ETL/ELT Pipelines & Big Data Engineering Data Lakes, Lakehouses, and Warehousing (AWS & Hybrid) Distributed Systems & Parallel Data Processing DevOps, CI/CD, and Container Orchestration
End-to-End Project Ownership:
From POC to Production Team Leadership, Mentorship, and Stakeholder Engagement
Technical Skills Summary
Cloud & Infrastructure Cloud Platforms:
AWS (expert), Azure (intermediate) AWS Services (extensive hands-on): S3, Redshift, RDS, Athena, Glue (incl. Spark), Lambda, Step Functions, EC2, EMR, DMS, Data Catalog, CloudWatch, EKS, API Gateway, SNS, MWAA Azure Services: Blob Storage, Data Factory, VMs, App Services
Containerization & Orchestration: Docker, EKS, Kubernetes, ECS Fargate Infrastructure as Code & CI/CD: GitHub, Bitbucket, Jenkins, CircleCI, Terraform (basic), GitOps best practices Data Engineering & Big Data
Data Pipelines:
Architected and deployed large-scale pipelines with Spark (Glue, EMR), Python, Airflow, Dagster, Luigi ETL/ELT Tools: AWS Glue (Spark, PySpark), SSIS, Informatica Power Center Streamlined ingestion from on-prem, SaaS, and RDBMS sources to modern lakehouse architecture
Performance Tuning: - Expert in optimizing queries, storage, and compute resources across Redshift, SQL Server, and Snowflake
Data Modeling: - Dimensional modeling (Kimball), data marts, warehouse optimization
Data Governance: Implemented metadata management, quality rules, and lineage (Glue Catalog, custom solutions) Programming & Scripting Programming Languages: Python (expert), Shell scripting, PHP, R, .NET (basic), JavaScript
Data Engineering Tools: PySpark, Pandas, SQL (advanced), Shell scripts, API integrations
Automation & Monitoring: Custom scripts, ing, log analysis, performance dashboards Databases & Storage
RDBMS: PostgreSQL, SQL Server, MySQL, Oracle, DB2, SQLite
Cloud DWH: AWS Redshift, Snowflake
NoSQL & Semi-Structured: MongoDB, Elasticsearch, Cassandra, Redis
Object Storage: S3, Azure Blob Big
Data Tools: Hadoop, Hive, Spark (Glue, EMR)
Data Visualization & BI Dashboards & Insights: Tableau (advanced), Power BI, SAP HANA Reporting KPI/Metric
Frameworks: Built executive dashboards tied to business metrics and SLAs Experience delivering reporting solutions to support Pharma, Insurance, Telecom, and Publishing sectors Web & Backend Development (Supportive Skills)
Backend & Web Dev: PHP, JavaScript, .NET, Excel Macros (VBA), SharePoint API & Integration: API Gateway, REST services, backend data exposure and enrichment Workflow Design: Automating tasks and pipelines using scripting, Python, and platform-native schedulers