This site uses cookies. To find out more, see our Cookies Policy

Big Data Engineer in Woodbridge Township, NJ at AccruePartners

Date Posted: 3/3/2018

Job Snapshot

Job Description

AccruePartners values our contract and consulting employees. We strive to provide comprehensive, quality and an affordable suite benefits. Our employees are eligible for Medical, Vision, Dental, Short Term Disability and 401K and other benefits. 


  • Fortune 100 Financial Services Company
  • 100-year history of dedication to customer satisfaction, success and growth
  • Tremendous growth and new business strategy leading to the need for new talent
  • Significant investments in cutting-edge technology


  • Culture: Excellent work environment that fosters collaboration
  • Growth: Ability to make an impact on the direction of the organization
  • Opportunity: Gain hands-on experience working with cutting-edge technology
  • Stability: Recent financial performance of the company has reported record profits


  • Iselin, NJ or Charlotte, NC ONSITE


  • Implement and on-go administration of Hadoop infrastructure including the installation, configuration and upgrading of Cloudera distribution of Hadoop
  • File system, cluster monitoring, and performance tuning of Hadoop ecosystem
  • Resolve issues involving map reduce, yarn, sqoop job failures; Analyze multi-tenancy job execution issues and resolve
  • Design and manage backup and disaster recovery solution for Hadoop clusters
  • Work on Unix operating systems to efficiently handle system administration tasks related to Hadoop clusters
  • Manage the Apache Kafka and Apache NIFI environments
  • Participate and manage the data lakes data movements involving Hadoop, NO-SQL databases like HBase, Cassandra and Mongodb
  • Work with data delivery teams to setup new Hadoop users. Includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and Map Reduce access for the new users. Configure Hadoop security aspects including Kerberos setup and RBAC authorization using Apache Sentry
  • Create and document best practices for Hadoop and big data environment
  • Participate in new data product or new technology evaluations; manage the certification process and evaluate and implement new initiatives in technology and process improvements
  • Interact with Security Engineering to design solutions, tools, testing and validation for controls
  • Evaluate the database administration and operational practices, and evolve automation procedures (Using scripting languages such as Shell, Python, Chef, Puppet, CFEngine, Ruby etc.)
  • Advance the cloud architecture for data stores; Work with Cloud engineering team with automation; Help operationalize Cloud usage for databases and for the Hadoop platform
  • Engage vendors for feasibility of new tools, concepts and features, understand their pros and cons and prepare the team for rollout
  • Analyze vendor suggestions/recommendations for applicability of the environment and design implementation details
  • Perform short and long term system/database planning and analysis as well as capacity planning
  • Integrate/collaborate with application development and support teams on various IT projects


  • Bachelor’s degree; Preferably in Computer Science or Information Systems
  • Ten or more years of overall IT/DBMS/Data Store experience
  • Three or more years of experience in, big data, data caching, data federation and data virtualization management including experience in leveraging Hadoop
  • Two or more years of expertise and in-depth knowledge of SAN, system administration, VmWare, backups, restores, data partitioning, database clustering and performance management
  • Experience writing shell scripts, and automating tasks.
  • Experience in the implementation details of Hadoop Clusters, Impala, and HBase and other emerging data techniques
  • Experience monitoring technologies for databases
  • Experience with orchestration techniques, infrastructure automation and cloud deployments
  • Understating of Linux, Windows, Dockers / containers
  • Familiarity with “IaaS” and “DBaaS” Service oriented concepts
  • Familiarity of Cloud Architecture (Public and Private clouds) – AWS , AZURE
  • Working knowledge of VMware and VMware vCloud Automation Center (vCAC)