Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Nilesh Kumar: Mail: Mobile No:+91-9008630725

Download as pdf or txt
Download as pdf or txt
You are on page 1of 5

Nilesh Kumar

Mail : nilesh.it447@gmail.com Mobile No :+91-9008630725

Technology Professional with over 10+ Years of experience in software development


comprising:

• Experience in Java, REST API, Kafka, Spark and related Technologies


• Knowledge on Cloud platform like Oracle Cloud, PCF, GAIA, AWS
• Big Data Technologies (Spark, Kafka, Hadoop, HIVE, IMPALA)
• Having good experience in Micro Service Architecture.
• Java, Scala, REST API, Hibernate, Oracle, Spring, Spring Boot
• CI and Code quality tools (Hudson, Sonar, Jules, Jira, AIM)
• Version control- GIT, Bitbucket
• Have extensive experience in the SDLC process of analysis, design, development
• Good in problem solving using Data Structures and Algorithms.
• Using tools to achieve AGILE goals.
• Mentoring juniors, code reviews
• Excellent and fast at self-learning of any Technologies.
• Good hands-on experience in deployment tools like Docker and Kubernetes

Work Experience:

JPMorgan Chase & Co.


Associate Software Engineer

Project: RCDR FEB 2020 to till

Rcdr (Risk Control and data reservoir) is big data service for migrating different legacy
data source, finding, organizing, enriching and reporting data on GAIA cloud and beyond
in order to track risk related data its status and resolution in existing organization.

Roles and Responsibilities:

• Worked on design and development for RCDR by adding or enhancing new


features, onboard legacy data source new RCDR platform.
• Envolve in technical and architectural discussion to enhance existing system.
• Taking care of release cycle of the project fixing bugs, security scans etc.
• Upgrade existing deployment environment during CDP upgrades.

Technologies used: Kafka, spark, Hadoop, HIVE, IMPALA, REST, JAVA, Maven, Git.
L&T Technology Services (client: Philips)
Project Lead

Project: Philips HSDP May 2019 to FEB 2020

Alcon HSDP service is used for reading/processing the


cataract surgery devices data using API gateways and provide the integrated solution to
support new FHIR health format and cloud solution

Roles and Responsibilities:

• Working on design and development of CDR services


to connect read the data from legacy database and convert to new FHIR health
format and persist to Philips cloud CDR.
• Design and architects the proxy service and token services which will be
deployable to Cloud foundry.

Technologies used: Spring boot, REST, JAVA, Maven, Git, cloud foundry, Docker, HSDP
CLOUD services.

ORACLE India Private Ltd. October 2015–May 2019


Senior Member Technical Staff

PROJECT: DATACATALOG

Duration: 1.5 year

Data Catalog is cloud service for discovering, finding, organizing, enriching and tracing data
assets on Oracle cloud and beyond, in order to provide efficient and governed data
exploitation.

Roles and Responsibilities:

• Worked on design and development of metadata Harvest module to connect to different


data sources system like databases, object storage, Kafka, cloud storage and read the
metadata from different files systems.
• Develop wrapper scheduling framework on top of cron job. To trigger schedule job for
metadata harvest.

Technologies used: Kafka, spark, Hadoop, REST, JAVA, Maven, Git, Docker, Kubernetes ,
IntelliJ

PROJECT: DataflowML

Duration: 2 year

DFML is a cloud platform that provides easy to use services to define manage and
execute smart pipelines to process and transform data. It's a 'Lambda Application
Platform with Real-time Analytics' as a Service based on Apache Spark, Machine
Learning, Mesos, Kafka, YARN, DockerContainers, and written in Scala .
Roles and Responsibilities:

• Worked on runtime and runtime framework and actively involved in


coding/development of ingestion service, execution of runtime pipeline and managing
the lifecycle of the pipelines using spark and Kafka.
• Worked on Development of scheduling framework REST wrapper on top of
Chronos/Mesos scheduling framework which will handle the error/exception reporting.

Technologies used: Kafka, Spark, REST , Scala , Java, Chronos, Mesos, Maven, Git, Docker

EMC Data Storage Systems, India September 2010 – 2015 (5 yrs. exp.)
Associate consultant II

PROJECT: CeTi LifeCare


Client: Apollo , Manipal
Duration: 8 Month

Life Care Project is the Cloud based enterprise application, which is for checking of disease of
patient and generate the alert if he/she require consulting the doctor based on measurement.
A Medhub device (Bluetooth enabled) which takes the measurement from different medical
devices and process the measurement and give the alert if it’s required.

Roles and Responsibilities:

• Worked on processing module to process the different type of vital, and call the CDSS
web services for generating the alert.
• Design the architecture of processing module and mediation flow based on spring
integration.

Technologies : Spring Integration, Spring data JPA, Spring web Services, Maven, Git

PROJECT: SGX
Client: Singapore Stock Exchange
Duration: 8 Month

SGX Project is the Cloud based enterprise application, which manages the stock
Related activity in fast and more reliable by uses of best architecture and best practice which
helps and supports the clients to make use of Java and spring related application
development.

Roles and Responsibilities:

• Worked on Audit trail framework, User, Customer and Spring Security module.
• Utilizing and applying best practices of all spring related technology.

Technologies used: vFabric SQLFire, Spring, JIRA, Maven Github, Sonar.

Achievement: Got Silver Prize from company for good contribution in the project.
PROJECT: EMC Cloud
Client: EMC Initiative/ Product
Duration: 6 Month

EMC Cloud competency makes the technology Cloud enabled by using EMC cloud products like
Cloud foundry and applying best practices of Platform as a Service using Cloud foundry. It
helps and supports EMC internal clients to make use of Cloud platform and Java application
development.

Roles and Responsibilities:

• Ramp up on vFabric SQLFire.


• POCs on Spring Data, SQLFire, Spring MVC, Spring Security
• JavaScript frameworks including Bootstrap.js
• Utilizing, and applying best practices of Platform as a Service using Cloud foundry.

Technologies used: vFabric SQLFire, Spring, JIRA, Maven, Github.

PROJECT: Data As a Service (DAAS)


Client: EMC Initiative/ Product
Duration: 6 Month

DAAS is providing end user the application that has capability to provide on time data with
minimal time and maximum accuracy. The data is for business people that do not understand
the database technologies but want to access data for their analysis and reporting work.

Roles and Responsibilities:

• Set up Activiti BPM process and create the process diagram to decide the flow of entire
application.
• Integrate Activiti with GlobalIds and Attivio for access data at application level, sending
request to create index for Attivio and finally create external table in GP.
• Full flow testing and Activiti BPM work flow management.

Technologies used: Activiti BPM, Spring MVC, GreenPlum, REST, Attivio, GlobalIds

PROJECT: Paetec Network – EMC, Bangalore


Client: Paetec (wind stream), USA.
Duration: 6 Month

Objective of this project is to migrate the Existing legacy database ‘AS400’ to Oracle database.
PAETEC is telecom domain project which will maintain wired network. Each module will
provide new scope for maintaining Paetec wired network and its components like Site,
Shelf, Slot, and Port. And migrating the network database ‘AS400’ to oracle database.

Roles and Responsibilities:

• Develop codes to automate all the manual process for high performance.
• Worked on analysis, design and feature development of the project
• Working with SQL, PL/SQL code to fetch the proper data from Paetec Databases i.e.
used by Paetec to manage network data.
PROJECT: Lightspeed (Wired Network Management)
Client: ATnT LABS/ EMC Consulting Integrated Product
Duration: 1.8 years

LIGHTSPEED is telecom domain project which will maintain ATnT wired network. Each module
will provide new scope for maintaining AT&T wired network and its components like Site,
Shelf, Slot, and Port. Also manages Inventory system that contains the data i.e. Granite
Inventory.

Roles and Responsibilities:

• Involved in end to end feature development and writing unit test for each feature.
• Working with SQL PLUS to fetch the proper data from Granite telcodia inventory i.e.
used by AT&T to manage network data. Resolving the issues about project
requirements (Software, Hardware, Resources).

Technologies used: Core JAVA, SOAP, XML, SQL, Junit

Awards and Recognitions: Got Silver award as individual contributor in Java based
Project. Appreciated For the good knowledge of Core JAVA, Recognized as fast learner of core
Spring.

EDUCATION: Institute: Gandhi Institute of Engineering and Technology Orissa.


University: Biju Patnaik University of Technology (BPUT) Raurkela
Grade: B. TECH (IT) in 2006-2010 with CGPA 7.48

CERTIFICATIONS: Oracle Certified java professional (java-1.6).


Oracle Certified Associates (Oracle9i) (SQL, PL/SQL).

You might also like