Vous êtes sur la page 1sur 3

B indiya Sharma

Current Location: Pune . Looking for positions in Pune, INDIA.


Phone # 9011050949. Email : bindiyamsharma@gmail.com

A result-oriented hands-on Big Data Professional having 12+ years of experience in diversified technologies and domains; strong work ethic with proven
track record to implement technologies that decrease costs, increase performance and positively affect the bottom line.

Expertise in complete product life cycle development, for implementation and integration from requirement analysis to deliver the complete project.
Engaged with Cybage; experiencing growth from software engineer, analyst, lead to architect.
Strong background in Database architecture, design, data modeling and Database development.
Good communication skills with ability to work as lead, independently as well as collaboratively in team environment to contribute problem-solving
strategies. At ease in interacting with people at all levels.
Aiming for a Technical position that utilizes my architectural, design, technical, analytical skills and leadership skills.

Technical Summary
Bigdata: Hadoop, Map Reduce, Sqoop, Hive, Oozie , Zookeeper, Pig, Impala, Kafka, Spark, MQTT, HBase, Cassandra
Bigdata Fabric: Cloudera
Operating System :Linux, Windows
Database :SQL Server, MySQL, Oracle
Programming Languages:SQL , Java, C#, Java Script
Cloud: Azure IaaS for CDH Setup, P2S set up, Google VM(IAAS), Awareness of AWS, EMR, Data Pipeline
BI:SSIS, SSAS, Powerpivot , Power View Reports, Qlickview Reports, Crystal Reports
Framework:.Net 4, ASP.Net Web Form, MVC, Web API, Entity framework
Domain :Enterprise Management Information System, Logistics, Awareness of Machine Learning
Methodology/Concepts :Agile, Sprint, Waterfall, UML/OOPs, XML, JSON
Tools
Project Management: TFS, Microsoft Project Plan (MPP)
Version Control : TFS,CVS,SVN
Compiler / Debugger / Editor : VSTS, Eclipse, Hue, SSMS, Toad

Professional Engagements

Engaged with Cybage from Feb 2006 till date.


Worked in vCustomer from Nov 2004 till Aug 2005.

Project Highlights

Project Architect and Lead, Big Data, Hitachi,Cybage

Architectural evaluation of the data ingestion pipeline provided by customer.


Business need was to do power generation forecasting by sensor data generated from wind turbine sites. Required to ingest the data
into different components to serve multiple purposes like data analytics, historical and real time data monitoring.
Understood the developed pieces of system, identified futuristic requirements of the project, evaluated current state of system and
suggested recommendations on design and architecture.
Implemented Kafka pipeline as per recommendations and latency expected. The data sent from turbine sites over Internet is published to Kafka
brokers which in-turn is consumed by Cassandra for historical monitoring, MQTT server for real time monitoring and Postgres database which acts as
data analytics source.
Development of new requirement where persistent store of data was required by Customer. Used Kafka source of Flume and HDFS sink to do the data
transfer into HDFS. Did POC and proposed alternative of implementing the same using HDFS connector to Customer.
Proposed solution for processing near real time streams and iterative processing. Explored spark streaming & machine learning and suggested spark
for the same.
Deployment of the application on Google VM.

Project Architect and Lead, Big Data, Technology Group,Cybage

Understanding needs of Organization, create solution and map it with the technology stack right from data source till presentation.
Business has user activity data which is being recorded via different systems like tool installed on each client machine or the internal systems used
in Cybage which is provided to server in some frequency.
Business need is to convert this data into useful information like software usage, skills used, machine time spent etc. which would help
organization in effective decision making.

Contact at: bindiyamsharma@gmail.com;9011050949


Worked as single point of contact for all teams like the data source teams, processing team and the presentation team for any query/confirmation
related to all execution phases.
Solution to different business problems like skill identification, machine time, software usage etc. by user activity logs recorded in internal monitoring
system. Identified data points, which needs to be captured for effective solution and communicated the details of the same to the team.
Continuous tools evaluation to solve problem statement; Evaluated different tools available to execute the need and did the appropriate tool
selection. Evaluated strategy like stages of data storage, partitions of data, processing frequencies, archival stages, data format selection as per
storage and processing needs.
Designed data pipeline from data ingestion, covering data processing till provisioning the data for presentation. Build a common framework to read
the configurations from file and log the failures in a common format, which can be represented as to-do list in UI.
Development environment setup with multi-node clusters. Capacity planning for production setup of on premise Hadoop cluster in terms of machines
and the capacity for both storage, compute and buffer incase of machine failures.
Explored other tools than used as well primarily Spark, HBase, Flume, Kafka.
Explored Machine learning to understand where and how to apply machine learning in our solutions.
Processing: User Skills
Requirement: Identify user skills in organization along with proficiency level by the user activity logs recorded for effective decision making while
resource management activities. Few of the consumer applications of this processing are resource proposal automation, search of resources of a
particular skill, primary skill automation etc.
Here data input was 10 GB per day stored in sql server database having different user activity logs like tools used, browsed sites, files worked upon
etc. which needs to be processed to identify skills.
We used Sqoop with HCatalog to import the data from sql server into Hive in ORC format with dynamic partition on a daily frequency .Processing
for skill identification was done in Hive. The scheduler used was Oozie to schedule Jobs.
The data was queried from Hadoop environment for presentation via Impala.
Processing: User Software Usage
Requirement: Identify the utilization of software installed on machine by users and decide on the renewal of licenses, assignment of licenses to
project/machines, generate software un-installation alerts or suggest solutions wherever optimum utilization is not happening.
Here data sources were log files of user software usage like tools used, frequency of tools usage, duration, softwares installed on machine each
date with a zip size of 7 GB.
Data ingestion is done via custom code written in java with Hadoop commands. Processing is done in Hive and Oozie is used for scheduling the
jobs.
Data is exported to Sql server database for further analysis and reporting.
POC: Deploy Hadoop Application on Azure(IAAS)
Worked on proof of concept to deploy the on premise development environment on Azure IAAS where the need is to have a scheduled start and
shutdown of Hadoop cluster.
We created virtual network with the VMs to set up Hadoop environment. Installed CDH over it and created data pipeline for above processing.
For integration between on premise and cloud environment, we had set up point to site VPN for using Sqoop exports for data transfer. We wrote
PowerShell scripts to start and stop cluster and its services on schedule.
Contributed in other processing solutions like remote user time spent, user machine time spent, creating real time reporting POC of incidents using
device logs like router, firewall etc.

Technology Consultant, Big Data, Technology Group,Cybage

Technology consultant to different project related to System Architecture, Technology Roadmap, Architecture Evaluation and non-functional
requirements.
Capability building on technologies and proof of concepts.
Analyzed application scenarios for system architecture.
Identifying technology planning needs through assessment and identification of critical focus areas.
Hands-on support to different projects during implementation and deployment phase.

Lead, .Net/Sql Server, MIS,Cybage

Technology lead: Started contribution as developer cum lead from single module of the MIS program.
MIS works on automating processes/activities in the organization. MIS is an integrated product where operational data is stored for different entities
across different departments, which in turn is utilized for effective decision-making.
Contributed in all SDLC phases for most of the projects as a lead where responsibilities were to convert requirement into features, bridge gaps ,
propose solutions, design components , database and mentor team from both technical &solution aspects of project.
Contributed in following various projects. Key project descriptions is provided below:
Project Management System stores customer and their projects information along with stakeholders from both sides, point of contact details,
contract dates etc. This provides data to evaluate relationship management status with customers which helps in proactive approach. Integrated with
all other modules like leave, resource management, compliance, task tracking, invoicing, appraisal etc.
Designed and developed common component for workflow, which can be utilized for any workflow like leave, resource request, project creation
etc.
Designed the system flow and integrated workflow for approval actions.
Contact at: bindiyamsharma@gmail.com;9011050949
Report Builder is a data warehouse containing data for all entities present in MIS for analytics and reporting purpose. Central place which provides
services to all modules for processed/de-normalized data.
Designed and created data warehouse using snowflake schema having facts and dimensions. Used surrogate key concept to store history of slow
changing dimensions.
Built batches to populate data and created framework to schedule jobs for different batch types.
Explored SSIS, SSAS, power pivot and power view reports. Reports creation as a POC using SSAS , power pivot and power view reports.

Resource Management automates job of requesting, searching and proposing resources more intelligently and transparently using the requirement
specified and data present about employees in system.
Designed and developed common component for request logging which can be utilized for different purpose like certification request, helpdesk
request etc.
Designed and developed common component for communication logging for all the comments logged at each step of workflow to have complete
view of communication.
Designed resource management system flow and integrated request, workflow and communication modules with it.
Acted as horizontal lead and performed activities like performance tuning of pages, database objects. Contributed on the data integration points
analysis to resolve the errors and the data goes to multiple analytic systems from here.

Developer,Java/Oracle, Catlogic,Cybage
Joined Catlogic team as a software developer;Project was about integrated solution for Transport Management Solution in logistics domain which
caters to transport like Road, Rail, Container, Warehouse and Terminal Operators. This integrated solution includes on-board computers for real-time
tracking and tracing of vehicles.
The project was developed using custom tool for front end, which was developed in java and backend in Oracle 9i.
Contributed to developing database objects including tables, indexes, views, sequences, packages, cursors and procedures for transport management
solution(TMS)in Oracle9i.Specialized in business logic creation using PL/SQL(Cursors, packages, triggers, procedures etc). Did master tables
identification along with the dependencies present in different sub components , resolved the dependencies and created central master (metadata)
component.
Learned groovy and contributed in design of common component for conversion of custom code into groovy.
Learned crystal reports and did creation and maintenance of the reports in project.
Execution of migration of database from Sql server to Oracle9i.
Got opportunity to perform lead role within a year; Was involved in estimations , design, reviews and communication with Client.
Certified as Oracle Certified Associate (9i)

Tech Support Engineer, vCustomer

Worked as technical support engineer for Linksys products.


Specialized in Wired and Wireless networking support.

Post graduate diploma in IT, Symbiosis Distance Learning,2009

Diploma in advance computing, CDAC,Pune,2006

Bachelors in Engineering, NMU,2004

Contact at: bindiyamsharma@gmail.com;9011050949

Vous aimerez peut-être aussi