Se ha denunciado esta presentación.
Utilizamos tu perfil de LinkedIn y tus datos de actividad para personalizar los anuncios y mostrarte publicidad más relevante. Puedes cambiar tus preferencias de publicidad en cualquier momento.


266 visualizaciones

Publicado el

  • I’ve personally never heard of companies who can produce a paper for you until word got around among my college groupmates. My professor asked me to write a research paper based on a field I have no idea about. My research skills are also very poor. So, I thought I’d give it a try. I chose a writer who matched my writing style and fulfilled every requirement I proposed. I turned my paper in and I actually got a good grade. I highly recommend ⇒ ⇐
    ¿Estás seguro?    No
    Tu mensaje aparecerá aquí
  • Sé el primero en recomendar esto


  1. 1. Robin David Mobile: +919952447654 email: Objective: Intend to build a career with leading corporate of hi-tech environment with committed and dedicated people. Willing to work as a key player in challenging and creative environment A position in which my technical abilities, education and past work experience will be utilized to best benefit the organization. Experience:  Around 9 years of total experience in IT with 3+ years of Experience in Hadoop  Professional experience of 4 months with Polaris, as Senior consultant mainly focusing on building Reference Data Management (RDM) solution within Hadoop  Professional experience of 18 months with iGATE Global Solutions, as Technical lead mainly focusing on Data Processing with Hadoop, Developing IV3 engine (iGATE’S Proprietary Big Data Platform) and Hadoop cluster Administration  Professional experience of 5.7 Years with Cindrel Info Tech, as Technology Specialist mainly focusing on ADS, LDAP, IIS, FTP, DW/BI and Hadoop  Professional experience of 11 months with Purple Info Tech, as Software Engineer mainly focusing on Routers and Switches Achievements in Hadoop • Having experience to build own plug-in with web user interface for o HDFS data encryption/decryption o Column based data masking on Hive o Hive Benchmarking o Sqoop automation for data ingestion on HDFS o HDD space issues alert automation on Hadoop cluster environments • Integrated Revolution R with cloudera distribution • Integrated Informatica - 9.5 and HDFS for Data processing in Apache and cloudera Cluster
  2. 2. Professional Summary • Expertise in creating data lake environment in HDFS within secure manner • Having experience to pull the data from RDBMS and various staging area to HDFS using Sqoop, Flume and Shell scripts • Providing solution and technical architecture to migrate existing DWH to Hadoop platform • Using the data Integration tool Pentaho for designing ETL jobs in the process of building Data lake • Having experience to build and processing data within the Datastax Cassandra cluster • Having good experience to handle data within advance hive • Experienced in Installation, Configuration and Management Pivotal, Cloudera, Hortonworks and Apache Hadoop Clusters • Designed and built IV3 IGATE proprietary Big Data platform • Good experience in Hadoop Distributed File System [HDFS] management • Having experience in shell scripting for various HDFS operations • Installation, Configuration and Management on EMC tools (gemfireXD, Green plum DB, and HAWQ) • Configure Hadoop clusters on amazon cloud • Recovery, Hadoop cluster, and name node or data node failures • End-to-end performance tuning of Hadoop clusters and Hadoop Map/Reduce routines against very large data sets • Kerberos implementation with Hadoop clusters • Configuring HA on Hadoop clusters • Integrate Splunk server to HDFS • Installing Hadoop cluster Monitoring tools (Pivotal Command Center, Ambari, Cloudera manager and Ganglia) • Cluster health monitoring and fixing performance issues • Data balancing on cluster and Commissioning/Decommissioning data nodes in existing Hadoop cluster • Having experience on HDFS create directory structure/permission for project specific needs and Set access permission for groups and users as required for project specific needs • Understand/Analyze specific job (project) and the run process • Understand/Analyze the scripts, map reduce codes, and input/output files/data for operations support • Build an archiving platform in Hadoop environment • Good exposure on understanding on Hadoop services and quick problem resolution skills • Having good experience in ADS, LDAP, DNS, DHCP, IIS, GPO, User Administration, Patch Maintenance, SSH, SUDO ,Configuring RPM through YUM, FTP and NFS Technical Skills Operating System RHEL 5.x/6.x, Centos 5.x/6.x, UBUNTU, Windows servers and Client family Hardware Dell , IBM and HP
  3. 3. Database MySQL, Postgresql, MSSQL and ORACLE Tools Command Center, Check_MK, Ambari, Ganglia, Cloudera Manager and GIT LAB. Languages Shell Scripting and Core Java Cloud Computing Framework AWS Hadoop Ecosystem Hadoop, ZooKeeper, Pig, Hive, Sqoop, Flume, Hue and Spark. Certifications:  Microsoft Certified IT Professional (MCITP - 2008)  Microsoft Certified Professional (MCP-2003)  CCNA Educational Qualifications: Bachelor of Science (Computer) St. Joseph’s College, Bharthidasan University - Trichy Major Assignments: Project1 Market Reference Data Management is a pure play reference data management solution built using big data platform and RDBMS to be used in the securities market industry. As part of this project data is collected from different market data vendors such as Reuters, interactive data etc. for different types of asset classes such as equity, fixed income and derivatives. The entire solution is built using pentaho as the ETL tool and the final tables are stored in Hive. The different downstream application will access data from hive as and when required. • Responsible for provide an architecture plan to implement entire RDM solution • Understanding the securities master data model • Create an API for getting data from various sources. • Create Hive data models • Design ETL jobs with pentaho for Data cleansing, Data identification and Load the data to Hive tables • Create a Hive (HQL) scripts to process the data (12 – 2015) To (till date) Project: Reference Data Management Domain Solution Building in Finance Domain Environment: CDH 5.0 Role: Senior Consultant
  4. 4. Project 2 GE purchases parts from different vendor across the world within all the business units. There is no central repository to monitor the vendors across the business units. Parts purchased were charged on different scale within the vendor and its subsidiary and other vendors. To identify the purchase price difference and to build a master list of vendors, GE software COE and IGATE together designed a data lake on Pivotal Hadoop consisting of all PO (Purchase order) And invoice data imported from multiple SAP/ERP sources. The data in the data lake is cleansed, integrated with DNB to build a master list of vendors and the data is analyzed to identify anomaly behavior in PO. Job Responsibilities: • Monitoring POD and IPS Hadoop clusters - Each environment has Sandbox, Development and Production division. • Having experience in CHECK_MK tool for monitoring Hadoop cluster environment. • Provide the solution for all Hadoop ecosystem and EMC tools. • Having experience in working together with EMC support. • Shell scripting for various Hadoop operations • User creation and quota allocation on Hadoop cluster and GPDB environment. • Talend Support • Having experience in GIT-LAB tool • Provide the solution for performance issues in Green plum DB environment • Bring back failure segments to active in GPDB environment Project 3 Retail Omni channel Solution leverages Cross Channel Analytics (Web, Mobile & Store) along with Bluetooth LE technology in the store to deliver superior customer experience. The target messages / personalized promotions are delivered at the right time (Moment of Truth) to maximize sales conversions. Job Responsibilities: • Create a Hive SQL Script for Data Processing and Data merging from multiple tables (03 – 2015) To (09 - 2015) Project: Data Lake Support Client GE - Software COE Environment: Pivotal and EMC Tools Role: Technical Lead (01 – 2014) To (06 - 2014) Project: Retail Omni channel Solution Client Retail Giant in US Environment: Cloudera Distribution CDH (4X) Role: Technical Lead
  5. 5. • Loading data to HDFS • Export Data from HDFS to RDBMS (MySQL) using Sqoop • Create a script for IROCS automation Project 4 The principle motivation for IV3 is to provide a turnkey Big Data platform that abstracts the complexities of technology implementation and frees up bandwidth to focus on creating differentiated business value. IV3 is software based big data analytics platform designed to work with enterprise class Hadoop distributions providing an open architecture and big data specific software engineering processes. IV3 is power-packed with components and enablers covering the life cycle of Big Data implementation starting from Data Ingestion, storage & transformation to various analytical models. It aims to marshal the three Vs of Big Data (Volume x Velocity x Variety) to deliver the maximum business impact. Job Responsibilities: • Implement data ingestion (RDBMS to HDFS) in IV3 Platform • Testing IV3 Tools on different Hadoop Distribution • Configure auto Yarn-Memory Calculator on IV3 Platform • HDFS data encryption/decryption • Column based data masking on Hive • Hive Benchmarking • Sqoop automation for data ingestion on HDFS • Create a automation script for detecting HDD space issues on Hadoop cluster environments Project 5 (06- 2014) – (09-2014) Project: Predictive Fleet Maintenance Client Penske Environment: Cloudera (CDH 4.6) - Hive Role: Technical Lead The business requirement of Penske is about collecting data from repository of un tapped data – Vehicle Diagnostics, Maintenance & Repair and this data potentially be leveraged to generate economic value. Penske want to create a future ready Big Data platform to efficiently store, process and analyze the data in consonance with their strategic initiatives. Penske engaged IGATE to partner with them in this strategic initiative to tap insights hidden in -Diagnosis, Maintenance & Repair data. IGATE would be leveraging its state of the art Big Data Engineering lab to implement the data engineering and data science part of this project. (12 - 2013) To (09 - 2015) Project: IV3 (Proprietary Big Data Platform) Client IGATE Environment: CDH, HDP and Pivotal Role: Technical Lead
  6. 6. Job Responsibilities: • Understand project scope, business requirement and current business processes. • Map business requirements with use cases. • Implemented use cases using with Hive. Project 6 (01 – 2012) To (11-2013) Project: WA-Insights and Analytics Client Watenmal Group – Global Environment: Apache Hadoop - HIVE, Map Reduce, Sqoop Role: Technology Specialist WAIA is intended to support all retail business segments involved in sale of goods and supporting services. WAIA Retail store integrated data model addresses 3 major aspects of a store business. (1) The physical flow and control of merchandise into, through and out of the store. (2)The selling process where the products and services offered for sale are transformed into tender and sales revenue is recognized. (3)The control and tracking of tender from the point of sale where it is received through its deposit into a bank or other depository Job Responsibilities: • Understanding Hadoop main components and architecture • Data migration between RDBMS to HDFS by using Sqoop • Understanding nuance map/reduce program in UDF • Data merging and optimization in hive • Adding new data nodes in existing Hadoop cluster • Safely decommissioning failure data nodes • Monitoring Hadoop cluster with ganglia tool Project 7: (5 – 2008) to (12 – 2011) Project: eyeTprofit and Maginss Client SI and Arjun Chemicals Environment: Windows – sql server, .Net Framework and LDAP Role: Technology Specialist EyeTprofit and Magniss enables businesses to easily analyze profitability, budget versus actual, revenue, inventory, and cash requirements etc instantaneously especially when the information is spread across multiple applications. EyeTprofit and Magniss is a non-invasive reporting system that facilitates information from multiple functions to be culled out and presented in an appropriate form to enable informed decisions.
  7. 7. Job Responsibilities: • LDAP integration with BI Tool • Administering Active directory, DHCP and DNS Servers • Managing group policies • Distributed file system management (DFS) • Administering FTP servers and IIS servers • Patch Maintenance • Administering File Shares, Disk Quotas • Providing access to the share drive users • Remote Support • Configure virtual machines Project 8: (5 – 2007) to (04 - 2008) Project: TRMS Client TTP Environment: Windows – Storage server , Router and Layer 3 Switches Role: Software Engineer A State of the art Traffic Management system, the first of its kind in India. It helps regulate and enforce law with the efficiency, expediency and accuracy of technology. Job Responsibilities: • Expertise in handling Wireless Communication • Maintaining with hand-held computer devices • Expertise in handling storage server • Responsibility of designing the topology of network in the client places implementing the infrastructure with high security and with hierarchy • Handling the issues on the Network related problems on the client places, mainly debugged issues that oriented on Network peripherals