Syed Ahmed
Syed Ahmed
Syed Ahmed
Email: syed.engg123@gmail.com
(914) 236-0661
Summary
Overall 8+ years of experience as Data Engineer and Data Analyst including designing, developing and implementation of
data models for enterprise-level applications and systems.
Excellent understanding of Hadoop architecture and underlying framework including storage management.
Excellent working experience in Scrum / Agile framework and Waterfall project execution methodologies.
Good working on analysis tool like Tableau for regression analysis, pie charts, and bar graphs.
Experience in Data transformation, Data mapping from source to target database schemas, Data Cleansing procedures.
Experience in Worked on NoSQL databases - HBase, Cassandra & MongoDB, database performance tuning & data
modeling.
Expertise in Developing Big data solutions using Data ingestion, Data Storage.
Highly motivated to work on Python scripts for statistics analytics for generating reports for Data Quality.
Good experience in using Sqoop for traditional RDBMS data pulls.
Good Knowledge with cloud technologies like Azure and AWS (EMR, S3, Red Shift, EC2).
Extensive experience in loading and analyzing large datasets with Hadoop framework (HDFS, PIG, HIVE, Flume,
Sqoop).
Hands on experience in Normalization and De-moralization techniques for effective and optimum performance in OLTP
and OLAP environments.
Good experience in Data Modeling and Data Analysis as a Proficient in gathering business requirements and handling
requirements management.
Solid knowledge of Dimensional Data Modeling with Ralph Kimball Methodology (Star Schema Modeling, Snow-Flake
Modeling for FACT and Dimensions Tables) using Analysis Services.
Responsible for troubleshooting issues in the execution of MapReduce jobs by inspecting and reviewing log files.
Strong experience in migrating data warehouses and databases into Hadoop/NoSQL platforms.
Expertise in Data Migration, Data Profiling, Data Cleansing, Transformation, Integration, Data Import
Extensive experience in using ER modeling tools such as Erwin and ER/Studio, Teradata, and MDM.
Knowledge about using Data Bricks Platform, Cloudera Manager and Hortonworks Distribution to monitor and manage
clusters.
Good Knowledge with cloud technologies like Azure and AWS (EMR, S3, RedShift, EC2, DynamoDB).
Experience in configuring and administering the Hadoop Cluster using major Hadoop Distributions like Apache Hadoop
and Cloudera.
Experience in NoSQL databases - HBase, Cassandra & MongoDB, database performance tuning & data modeling.
Excellent experience in development of Big Data projects using Hadoop, Hive, HDP, Pig, Flume, Storm and MapReduce
open source tools/technologies.
Working on the ad-hoc queries, Indexing, Replication, Load balancing, and Aggregation in MongoDB.
Technical Skills
Hadoop Ecosystem: MapReduce, Spark 2.3, HBase 1.2, Hive 2.3, Pig 0.17, Solr 7.2, Flume 1.8, Sqoop 1.4, Kafka 1.0.1,
Oozie 4.3, Hue, Cloudera Manager, Stream sets, Neo4j, Hadoop 3.0, Apache Nifi 1.6, Cassandra 3.11
Cloud Management: Microsoft Azure , Amazon Web Services(AWS)
OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9
Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED
RDBMS Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server 2016/2014, DB2.
NoSQL Database: Cassandra, HBase, Mongo DB, Dynamo DB, Cosmos DB
Testing and defect tracking Tools: HP/Mercury, Quality Center, Win Runner, MS Visio 2016 & Visual Source Safe
Operating System: Windows 7/8/10, Unix, Sun Solaris
ETL/Data warehouse Tools: Informatica v10, SAP Business Objects Business Intelligence 4.2 Service Pack 03, Talend,
Tableau, and Pentaho.
Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Agile, Waterfall Model.
Work Experience