Role:
- Develop a new Datawarehouse (Postgres/Redshift) by creating a workflow to systemize manual user data daily
- Build simple UI for CRUD applications
- Build scalable open-source ETL data pipelines using Pyspark & SQL(Postgres & Redshift) to automate manual Data Reconciliation and derived tables
- Schedule Jobs to trigger emails with planned queries and attend ad hoc data queries from different Team
- Build and maintaint large no. of Tableau Dashboards for Daily/Monthly MIS Reporting
- Troubleshoot dowtime issues in datawarehouse and the scheduled ETL Scipts.
- Integrate complete life cycle data of a customer from multiple sources
- Undertake handover of complete data maintenance of related projects in the Database
Skill
Mandatory:
- 5+ yrs experience in working in database management and analysis
- Experience in building data pipelines using big data technologies like Spark, Hadoop etc.
- Java/Flask to build simpe CRUD UI
- Knowledge on datawarehouse concepts,database design, etc. and experience working Postgres and Redshift DB
- Proficiency in Advanced SQL,Python,PySpark,any ETL Tool
Good to have:
- Apache Airflow/ any Email orchestration
- Experience in working with AWS services like S3,Redshift,EMR,Lambda,SQS etc
Persona:
- Bachelor's in CSE from Tier-I,II institutions
- Have direct exposure to Data Engineering tasks and big data analytics
- Have handed ETL projects to systemize manual data
- Fluent data handling and strong sense of exploratory data analysis
- Ability to work with data in any tool Excel Vs Database Tools
- Managed a team of Data Engineers/Analysts
Location: Bangalore
Didn’t find the job appropriate? Report this Job