R01 - 1 (1480)
R01 - 1 (1480)
R01 - 1 (1480)
PROFESSIONAL SUMMARY
Around 7 years of professional experience in Data Analysis, Data Conversion, Data Validation, Data Profiling,
UAT Testing and Report Creation and working experience in Tableau, AWS Redshift, AWS S3, Python, Unix and
Oracle.
Expert level understanding of Data Cleansing, Statistical Modeling, Exploratory Data Analysis and Predictive
Analysis.
Good understanding of Data Warehousing Components, Architecture and Tools such as Teradata, SAS,
PostgreSQL and Tableau.
Hands-on experience with development and maintenance of Relational Database Management Systems such
as Microsoft Access, MySQL, Teradata and PostgreSQL and ability to design efficient queries.
Knowledge of the Software Development Life Cycle (SDLC) methodologies like Agile and Waterfall.
Smart project management skills and ability to identify processes that can be automated to save man-hours
and resources.
Strong knowledge of stored procedures, data analysis and data validation using SQL.
Expertise in customer segmentation, targeting, recommendation and personalization strategies.
Professional working experience in end-to-end data science workflows – Exploratory Data Analysis, design &
build model, validate using Cross-Validation techniques and maintenance of these machine learning based
prediction models.
Expert level understanding of e-commerce projects and involved process flows including inventory reporting,
website optimization, hands on A/B experimentation and marketing.
Worked closely with project stakeholders, SMEs, and staff to understand the requirements and specifications
for new applications along with enhancements to the existing applications.
Elicited requirements from stakeholders, vendors, consultants, subject matter experts, Product Owner’s and
existing system users using various techniques like JAD sessions, brain storming, one on one.
Exceptional ability to research, analyze and convey complex technical information to diverse end-users at all
levels. Solutions-driven strategist who consistently improves efficiency, productivity and the bottom line.
Hands-on experience with Supervised and Unsupervised Machine Learning.
Effective Communicator with good Interpersonal and Client Relational skills.
Used project management tools like JIRA for Status reporting and planning.
TECHNICAL EXPERTISE
GUI & Reporting Tools: SSRS, Business Objects12, Tableau, Power BI
ETL / SSIS: Business Analytics
RDMS: SQL Server, MS Access 7.0
Programming: SQL, Python
Methodologies: Waterfall, Agile methodology (SCRUM)
Data Modeling: Star-Schema Modeling, FACT and dimension tables, Pivot Tables, Erwin
EDUCATION
Course Year of Passing Board/University
Master of Science (Computer Science) 2000 Nagarjuna University
Bachelor of Science (Computer Science) 1998 Nagarjuna University
WORK EXPERIENCE
Responsibilities:
Performed Data wrangling, Preprocessed and transformed the data into useful formats, including joining
multiple tables using Pandas.
Performed data pre-processing tasks like Normalization, Scaling, treating the missing values, outliers and thus,
preparing it for statistical analysis like multivariate and correlation analysis
Implemented numerous Feature Engineering techniques to generate new features from the existing ones and
tested their performance.
Performed a complete data analysis, extracted meaningful insights from the data.
Used python for Exploratory Data Analysis, A/B testing, HQL, VQL, Data Lake, oozie, Dataiku, test and
Hypothesis test to compare and identify the effectiveness of Creative Campaigns.
Developed multiple classification models to predict the potential defaulters and compared with the
performance of the current working model.
Implemented Predictive analytics and algorithms to forecast key metrics in the form of designed for the
company's core business.
Data wrangling to clean, transform and reshape the data utilizing NumPy and Pandas library.
Contribute to data mining architectures, modeling standards, reporting, and data analysis methodologies.
Conduct research and make recommendations on data mining products, services, protocols, and standards in
support of procurement and development efforts.
Involved in defining the source to target data mappings, business rules, data definitions.
Worked with different data science teams and provided respective data as required on an ad-hoc request
basis
Assisted both application engineering and data scientist teams in mutual agreements/provisions of data.
Used Python, SQL to create Statistical algorithms involving Linear Regression, Logistic Regression, Random
Forest, Decision trees for estimating the risks.
Responsibilities:
Responsible for Retrieving data using SQL/Hive Queries from the database and performing Analysis
enhancements.
Worked as an RLC (Regulatory and Legal Compliance) team member and undertook user stories (tasks) with
critical deadlines in Agile environment.
Applied Regression in identifying the probability of the Agent's location regarding the insurance policies sold.
Used advanced Microsoft Excel functions such as Pivot tables and VLOOKUP in order to analyze the data and
prepare programs.
Performed various statistical tests for clear understanding to the client.
Actively involved in Analysis, development and unit testing of the data and delivery assurance of the user
story in the Agile environment.
Worked with Data frames and other data interfaces in Python for retrieving and storing the data.
Applied various machine learning algorithms such as decision trees, K-Means, Random Forests and regression
in Python with the required packages installed.
Applied K-Means algorithm in determining the position of an agent based on the data collected.
Read data from various files including HTML, CSV and sas7bdat file etc. using Python.
Involved in Analyzing system failures, identifying root causes, and recommended course of actions.
Coded, tested, debugged, implemented and documented data using Python.
Researched multi-layer classification algorithms as well as building natural language processing models
through ensemble.
Worked with SQL, PL/SQL procedures and functions, stored procedures and packages within the mappings.
Tuned Informatica mappings and sessions for optimum performance.
Responsibilities:
Responsible for Retrieving data using SQL/Hive Queries from the database and performing analysis
developments.
Used Python/SQL to manipulate data and develop and validate quantitative methods.
Worked with SQL, PL/SQL, Procedures and Functions, Stored Procedures and Packages with mapping.
Used Advanced Microsoft Excel functions such as Pivot tables and VLOOKUP to analyze the data and prepare
programs.
Skilled in Advanced Regression Modelling, Correlation, Multivariate Analysis, Model Building, Business
Intelligence tools and application of Statistical Concepts.
Importing and exporting the data into HDFS and Hive using Sqoop. Worked on NOSQL databases like
MangoDB, Cassandra.
Cleaned the data by analyzing and removing the duplicate and inaccurate data using Python with Data Frames
and other data interfaces in Python for storing and retrieving the data.
Coded, Tested, Debugged, implemented, and documented data using Python many tasks from collecting the
data and exploring the data and interpreting the statistical information.
Used pruning algorithms to cut away the connections and perceptron to significantly improve the
performance of back-propagation algorithms.
Conducting studies, rapid plots and using advanced data mining and statistical modeling techniques to build a
solution that optimizes the quality and performance of data.
Worked with several outlier algorithms like Z-score, PCA, LMS, and DBSCAN to better process the data for
higher accuracy.
Worked with ETL source to specification documents and understood the business requirements and
performed in extraction, transformation and loading the data into the applications.
Environment: Python, NOSQL, MANGODB, CASSANDRA, SQL