Alice Jones

123 Any St., Apt. 102
LinkedIn ▪ GitHub ▪ Google+East Brunswick, NJ 00000

Profile: Hadoop Developer and Administrator

Purveyor of competitive intelligence and holistic, timely analyses of big data made possible by the successful installation, configuration, and administration of Hadoop ecosystem components and architecture.

  • Two years’ experience installing, configuring, and testing Hadoop ecosystem components.
  • Capable of processing large sets of structured, semi-structured, and unstructured data and supporting systems application architecture.
  • Able to assess business rules, collaborate with stakeholders, and perform source-to-target data mapping, design, and review.
  • Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning, and advanced data processing. Experience optimizing ETL workflows.
  • Hortonworks Certified Hadoop Developer, Cloudera Certified Hadoop Developer and Certified Hadoop Administrator.

Areas of Expertise:

  • Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig, Sqoop, Cassandra, Oozie, Flume, Chukwa, Pentaho Kettle, and Talend
  • Programming Languages: Java, C/C++, eVB, Assembly Language (8085/8086)
  • Scripting Languages: JSP and Servlets, PHP, JavaScript, XML, HTML, Python, and Bash
  • Databases: NoSQL, Oracle
  • UNIX Tools: Apache, Yum, RPM
  • Tools: Eclipse, JDeveloper, JProbe, CVS, Ant, MS Visual Studio
  • Platforms: Windows(2000/XP), Linux, Solaris, AIX, HPUX
  • Application Servers: Apache Tomcat 5.x 6.0, Jboss 4.0
  • Testing Tools: NetBeans, Eclipse, WSAD, RAD
  • Methodologies: Agile, UML, Design Patterns

Professional Experience:

Hadoop Developer
Investor Online Network, Englewood Cliff, New Jersey2013 to present
Facilitated insightful daily analyses of 60GB to 80GB of website data collected by external sources. Spawning recommendations and tips that increased traffic 38% and advertising revenue 16% for this online provider of financial market intelligence.

  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Managed and reviewed Hadoop log files.
  • Tested raw data and executed performance scripts.
  • Shared responsibility for administration of Hadoop, Hive and Pig.

Hadoop Developer/Administrator
Bank of the East, Yonkers, New York2012–2013
Helped this regional bank streamline business processes by developing, installing, and configuring Hadoop ecosystem components that moved data from individual servers to HDFS.

  • Installed and configured MapReduce, HIVE, and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Created HBase tables to load large sets of structured, semi-structured, and unstructured data coming from UNIX, NoSQL, and a variety of portfolios.
  • Supported code/design analysis, strategy development, and project planning.
  • Created reports for the BI team using Sqoop to export data into HDFS and Hive.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application, and BI teams to ensure data quality and availability.
  • Administrator for Pig, Hive, and Hbase installing updates, patches, and upgrades.

Java Developer
New York Bank, New York, New York2010–2012
Improved user satisfaction and adoption rates by designing, coding, debugging, documenting, maintaining, and modifying a number of apps and programs for ATM and online banking. Participated in Hadoop training and development as part of a cross-training program.

  • Led the migration of monthly statements from UNIX platform to MVC web-based Windows application using Java, JSP, and Struts technology.
  • Prepared use cases, designed, and developed object models and class diagrams.
  • Developed SQL statements to improve back-end communications.
  • Incorporated custom logging mechanism for tracing errors, resolving all issues and bugs before deploying the application in the WebSphere Server.
  • Received praise from users, shareholders, and analysts for developing a highly interactive and intuitive UI using JSP, AJAX, JSF, and JQuery techniques.
  • View samples at

Education, Training, and Professional Development
Old Jersey Institute of Technology, BS Computer Science

Hadoop Training
Accelebrate: “Hadoop Administration Training”
Cloudera University Courses: “Hadoop Essentials” and “Hadoop Fundamentals I & II”
MapReduce Courses: “Introduction to Apache MapReduce and HDFS,” “Writing MapReduce Applications” and “Intro to Cluster Administration”
Nitesh Jain: “Become a Certified Hadoop Developer”

Member, Hadoop Users Group of New Jersey