Job Description: Implement data processing pipelines using Big data tools – Spark, Hive, Talend, Java, Cassandra and Sqoop to analyze large volumes of data on various Hadoop distributed environments like Cloudera and Hortonworks. Responsible for getting requirements from business and implementing, testing and deployment to production environment. Develop and implement training programs for new employees. Perform internal and external code reviews on various modules. Prepare test cases to perform verification and validation of software and data. Implement various performance optimizations in querying large volumes of data to achieve less latency. Collaborate with multiple teams for design and development features involving integration with multiple systems. Perform problem resolution and root cause analysis along with recommending and implementing preventive techniques. Logical and Physical database design to support new and existing applications. Create functional specification document pertaining to applications. Analyze user requirements and propose systems. Provide full life cycle support, implement reporting and planning processes, create reports and perform optimizations. Technologies used to perform jobs includes Spark, Scala, Hive, Talend, Sqoop, Cassandra, SQL, Kafka, Java, AWS, AZURE, MySQL, Oracle, GIT, Linux.
Position: FT (Full-time)
Jobsite: Herndon, VA or may change to unanticipated job sites throughout the US as required.
Send resume to Gurus Infotech, Inc. 704 Pine Street, Herndon, VA 20170.
ATTN: #003.
Requirements
New and exciting developments in business