Manuh Technologies

Lead AWS Data Engineer - ETL

Click Here to Apply

Job Location

hyderabad, India

Job Description

Position Lead : AWS Data Engineer. Location : Pune/ Gurgaon/Hyderabad Hybrid. Experience : 9. Job Description : - 9 years of experience as a Data Engineer on the AWS Stack PySpark. - Solid experience with AWS services such as Cloud Formation, S3, Athena, Glue, Glue Data Brew, EMR/Spark, RDS, Redshift, Data Sync, DMS, DynamoDB, Lambda, Step Functions, IAM, KMS, SM, Event Bridge, EC2, SQS, SNS, Lake Formation, Cloud Watch, Cloud Trail. - Programming experience with Python, Shell scripting, and SQL. - Responsible for building, test, QA & UAT environments using Cloud Formation. - Build & implement CI/CD pipelines for the EDP Platform using Cloud Formation and Jenkins. Good to Have : - Implement high-velocity streaming solutions and orchestration using Amazon Kinesis, AWS Managed Airflow, and AWS Managed Kafka (preferred. - Solid experience building solutions on AWS data lake/data warehouse. - Analyze, design, Development , and implement data ingestion pipeline in AWS. - Knowledge of implementing ETL/ELT for data solutions end to eng. - Ingest data from Rest APIs to AWS data lake (S3) and relational databases such as Amazon RDS, Aurora, and Redshift. - Perform the Peer Code Review and, perform code quality analysis, and associated tools end-to-end for Prudential's platform. - Create detailed, comprehensive, and well-structured test cases that follow best practices and techniques. - Estimate, prioritize, plan & coordinate quality testing activities. - Understanding requirements, and data solutions (ingest, storage, integration, processing, access) on AWS. - Knowledge of implementing RBAC strategy/solutions using AWS IAM and Redshift RBAC model. - Knowledge of analyzing data using SQL Stored procedure. - Build automated data pipelines to ingest data from relational database systems, file system , and NAS shares to AWS relational databases such as Amazon RDS, Aurora, and Redshift. - Build Automated data pipelines to develop test plans, execute manual and automated test cases, help to identify the root causes, and articulate defects clearly. - Recreate production issues to help determine the issue and verify any fixes. - Conducting End to End verification and validation for the entire application. - Creating Jenkins CI pipelines to integrate Sonar/Security scans and test automation scripts. - Using Git/bitbucket for efficient remote team working, storing framework, and developing test scripts. - Part of DevOps QA and AWS team focusing on building CI/CD pipeline. - Part of the release/build team and mainly worked on release management, CI/CD pipeline. - Deploy multiple instances by using cloud formation templates. Responsibilities : - Designing, building, and maintaining efficient, reusable, and reliable code. - Ensure the best possible performance and quality of high-scale data applications and services. - Participate in system design discussions. - Independently perform hands-on Development and unit testing of the applications. - Collaborate with the development team and build individual components into the enterprise data platform. - Work in a team environment with the product, QE/QA, and cross-functional teams to deliver a project throughout the whole software development cycle. - Responsible to identify and resolve any performance issues. - Keep up to date with new technology development and implementation. - Participate in code review to make sure standards and best practices are met. (ref:hirist.tech)

Location: hyderabad, IN

Posted Date: 11/27/2024
Click Here to Apply
View More Manuh Technologies Jobs

Contact Information

Contact Human Resources
Manuh Technologies

Posted

November 27, 2024
UID: 4892965250

AboutJobs.com does not guarantee the validity or accuracy of the job information posted in this database. It is the job seeker's responsibility to independently review all posting companies, contracts and job offers.