Responsibilities: Experience with the system development lifecycle, such as gathering and refining requirements, data analysing and data profiling, designing and implementing solution, handling sensitive data sets, production deployment, troubleshooting and data conversion, etc.
- Create and maintain optimal data pipelines using ETL tools like Pentaho, Streamsets.
- Assemble large, complex data sets that meet functional / non-functional business requirements.
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc
- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using big data technologies.
- Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency, and other key business performance metrics.
- Build processes supporting data transformation, data structures, metadata, dependency, and workload management.
- Work with stakeholders including the internal teams to assist with data-related technical issues, support their data infrastructure needs, to strive for greater functionality in our data systems.
- Being vigilant round-the-clock to provide post-implementation support
- To provide training, troubleshoot and upgrade the system as the need arises
- Responsible for identifying the performance bottlenecks and providing the solution for implementing the same.
- Should have strong analytical skills
Qualifications Bachelor’s degree in computer science or a related technical discipline.