Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Resume Format 1

Download as pdf or txt
Download as pdf or txt
You are on page 1of 5

Narendra Gujiri

Data Engineer

Email: narendra.gujiri@wissen.com

Results driven, Innovative and Progressive Data Engineer with around 12 years of Industry Experience in Data space (Data
Modelling, Data warehouse/ ETL/ Big Data/ machine learning). Highly skilled in multitasking on multiple vendor technologies, across
multiple Business domain challenges. Well versed in reliability theory with hands on experience in performing reliability
engineering for complex systems. Always on toes to upskill and stay aware of latest ongoings in Technical and Businessspace.

EDUCATION
Bachelor of Engineering in Electronics & Telecommunication ► Mumbai University Aug 2007 - Mar 2012

SKILLS
Spark SQL, Hive, Impala, Sqoop, Snowflake | Scala, Python, Shell Scripting | Cloudera (CDH), Streamsets, Kafka | SAP Data
Services, Talend, Cognos Datamanger, Atscale, Informatica| Oracle SQL, PL/SQL, Microsoft SQL Server, Sybase | Jenkins, MS
Azure Git, Intellij, Visual Studio | GCP, Azure | Tableau | Machine learning | Technical Lead | Salesforce

EXPERIENCE
Principal Engineer – Wissen Technology ► Morgan Stanley., Mumbai May 2022 – till present
Business Functions worked in - Sales and Marketing Investment Mangement (Mutual Funds)
Responsibilities -
Requirements/Blue Printing/Sprint Planning -
• Involved in nascent stage of project planning and defining roadmaps and goals.
• Lift and shift the planned roadmap into Sprint epics, features, or stories and define priorities for the same.
Discovery/Designing/Documentation -
• Involved in Designing and Understanding business requirements from Product Owners/Business Analysts.
• Discovering the sources, objects, and plan towards the architecture for ingestion pipelines.
Development -
• Involved in the implementation of ingestion pipelines using Salesforce, Sybase, Informatica, Snowflake, python
• Building datalake from scratch in Snowflake, setting up development standards and best practices for fellow
teammates.
• Automated ETL process by building python frameworks from legacy systems to Snowflake and vice versa along with
Informatica.
• Built testing frameworks in python for Unit testing and Data validation. Ensured data correctness and completeness
thus setting data governance standards.
• I was responsible for Data cataloging of the metadata for end user, business stakeholders access.
• Automated machine learning flows using Databricks for product recommendation to users using collaborative and
content based filtering.
• Integrating data from diverse sources like flat files, Legacy systems, External Data and building a meaningful facts and
dimensions out of it for consumable purpose by reporting team and for data science team.
Code Maintenance, Deployment, Scheduling -
• Bit Bucket, Git Hub, was used for code maintenance was used.
• Jenkins and Azure pipelines was used for build and deployment service.
• Autosys job for scheduling.
Mentorship -
• Onboarded new team members and Knowledge Transfer created standard document of the onboarding process
• when team upsized, so that things are robust and smooth.
• Organizing Trainings for new team members, cross functional trainings etc.

Assistant Consultant ► Tata Consultancy Services, Mumbai Dec 2021 – Apr 2022
Business Functions worked in - Retail (Nielsen Assorted Space optimization)
Responsibilities-
Data Engineer - Involved in Designing and Understanding business requirements for Migration of on prem Hadoop
objects into Azure Databricks
Development -
• Involved in the implementation of ingestion pipelines using Azure Databricks, Spark, Scala, Hive
• Responsible for setting up operational Spark Applications in production which involves deploying packaged Scala jar,
• Spark Sql, other app conf in production.
• HQL queries developed for both batch loads and unit testing during development (For unit testing Impala was used for
quick data checks).
• Code Maintenance, Deployment, Scheduling -
• Bitbucket was used for code maintenance.
• Maven was used for build and deployment service.
• Airflow was used for scheduling all the Hadoop jobs.

Data Engineer - Consultant ► General Mills India Pvt. Ltd., Mumbai Jun 2015 - Nov 2021
Business Functions worked in - Supply Chain, Marketing, Strategic Revenue Management, Finance (Business
Performance Management)
Responsibilities -
Requirements/Blue Printing/Sprint Planning -
• Involved in nascent stage of project planning and defining roadmaps and goals.
• Quarterly reviewing Project scope definition and making necessary changes to roadmap defined.
• Lift and shift the planned roadmap into Sprint epics, features, or stories and define priorities for the same.
Discovery/Designing/Documentation -
• Involved in Designing and Understanding business requirements from Product Owners/Business Analysts.
• Discovering the sources, objects, and plan towards the architecture for ingestion pipelines.
Development -
• Involved in the implementation of ingestion pipelines using Hadoop Ecosystem tools using Spark, Scala, Hive, Impala,
Sqoop.
• Defining Quality rules using Spark Scala for ensuring correctness, completeness in data in pre model layers with the
help of Dataframe and Dataset APIs and Spark Transformations.
• Responsible for setting up operational Spark Applications in production which involves deploying packaged Scala jar,
Spark Sql, other app conf in production edge node.
• HQL queries developed for both batch loads and unit testing during development (For unit testing Impala was used
for quick data checks).
• Sqooped data from and into RDMS source, SAP HANA into Hadoop Datalake.
• Building WebCrawler using Python to pull data from standard APIs
• OLAP model building on Atscale for data virtualization.

Code Maintenance, Deployment, Scheduling -


• Azure TFS Git was used for code maintenance was used.
• Jenkins was used for build and deployment service.
• Airflow and Tidal was used for scheduling all the Hadoop jobs.
Non Hadoop -
• Experience working on Oracle as a Datawarehouse and SAP Data Services ETL tool.
• Responsible for building complex views using windows function for calculation, transformation, Package,
• Procedures, Tables, Materialized View.
• Performance tuning and Optimization.
Mentorship -
• Onboarded new team members and Knowledge Transfer created standard document of the onboarding process
when team upsized, so that things are robust and smooth.
• Organizing Trainings for new team members, cross functional trainings etc.
• Served as Interview Panel
• Worked as Technical lead in certain projects.

Senior Software Engineer - 2 ► Aon Hewitt Nov 2014 - May 2015


Business Functions worked in - Human Capital Management, Payroll
Responsibilities -
• Responsible working on building Front End Transactional Application using Oracle, PL/SQL objects like Packages,
Procedures, Tables, Views.
• Performance Tuning and Optimization.
• Implemented complex business transformations and calculations
• Building ETL pipelines using Oracle DB links and shellscripting.
• Worked on reporting tools like Cognos Studio.

Senior Software Engineer ► Capgemini India Pvt. Ltd. Apr 2012 - Nov 2014
Business Functions worked in - Travel Domain
Responsibilities-
• Responsible working on Oracle Frontend and Backend Database.
• Implemented Business Logic for Ticket Booking, Coupon Printing, Payment/Discounts, Aftersales, Refund logic in
Oracle Pl/SQL database.
• Implemented Oracle Backend Datawarehouse which fetched data from Oracle frontend database via ETL tool,
Cognos Datamanager and Shell Scripting
• Responsible for Oracle Admin activities like maintenance taking backups and creating new databases, new users,
roles etc.
• End-to-End Unit testing and documentation for end users.
• Metadata maintenance in Knowledge management system.
• Oracle PL/SQL code tuning, SQL tuning for optimum performance.

You might also like