Hi, Hope doing great…!!!
I have an urgent requirement below, please go through it and share your updated profile ASAP at *a <dee...@sysmind.com>m...@zenithtechsolutions.com <m...@zenithtechsolutions.com>* 1 - Hadoop PLATFORM ENGINEERS: Looking for seasoned Hadoop Admins & Engineers, exp building and managing these platforms, tuning, optimization, security, need hands on exp Questions to ask the candidate - Have they setup and built something large scale? Or just learned from books. How do you integrate X tool into the Hadoop ecosystems? What are the steps Most of initial needs are here · Seasoned Hadoop administrators – Experience building large Hadoop platforms, system tuning, YARN setting and optimization, integration with kerberos (security), high availability/ resiliency/ DR of the platform and services that run on Hadoop · Tool integrators – Integrating tools such as Python, Spark , R, Casandra with Hadoop ecosystem · Solution architects / engineers : Piece together various data technology/ components · Platforms experience – either built from scratch preferably, or managing/administration at least 2 - CODING and TUNING EXPERTISE : They have a platform on Teradata and SAS, need to migrate to new platform. So need coders with this experience, how to role from SAS to Python or R, How to move Teradata SQL to Hive? Refactor legacy code · Expertise in coding in Python, Hive, R, Spark, Java with emphasis on tuning / optimization and re-factoring legacy code that runs on traditional SQL engines (such as Oracle, Teradata) to optimized Hive queries (or) moving existing SAS code to R and Python 3 - NOSQL administrators / engineers : · Experience managing NOSQL databases such HBASE , Casandra, MongoDB 4 - Real-time data platforms : · Experience engineering/ building and managing real-time platforms/ frameworks such as SPARK / FLINK/ APEX · Kafka administrators · NIFI experience with knowledge of creating new NIFI processors and workflows Soft skills : · Demonstrate the ability to learn new Open source frameworks / tools / technologies · Self-learner · Good communication skills Need people who can learn new tools, good communication skills Worried about people who are googling answers on the interview, or who just know things from books without hands on experience Need to highlight which skillsets they are a fit for in the resume Finding less than 5-10% of what they need in resumes – maybe 1 out of 10 resumes are passing muster Only 1 candidate is moving forward out of 7 interviews, numerable submittals Need mostly Platform Engineers, most resumes were more Data Engineers *How to sell this:* They are fundamentally changing and modernizing the data ecosystem, influence the Discover org, huge initiative, state of the art system They will have influence on the direction Discover is going, versus being 1 engineer in 1000 at any other tech company *Questions*: · How important is NIFI? *It’s a nice to have, can be trained. Hadoop, Python, Spark, R, Hive are the hard skillsets required* · Industry background is not important, just nice to have · Will be getting specific feedback on interviews · · Hadoop Developer is not a real thing…they would be a Hive Developer, etc. Hadoop is a general ecosystem · Can administer small coding tests to include when submitting · They will be creating a panel interview process moving forward · Open to sponsorship? *YES…need to give Liz the heads up* · Hortonworks Data platform is the specific Hadoop distribution they are using. Hortonworks' product named Hortonworks Data Platform (HDP) includes Apache Hadoop and is used for storing, processing, and analyzing large volumes of data. The platform is designed to deal with data from many sources and formats. The platform includes various Apache Hadoop <https://en.wikipedia.org/wiki/Apache_Hadoop> projects including the Hadoop Distributed File System, MapReduce <https://en.wikipedia.org/wiki/MapReduce>, Pig, Hive, HBase <https://en.wikipedia.org/wiki/HBase> and Zookeeper and additional components Advanced Analytics - Data Engineer The Position • Develop data driven solutions with current and next gen “Big data” technologies to meet evolving business needs • Develop greenfield capabilities leveraging Open source next-gen technologies • Code and integrate open source solutions into an enterprise Hadoop ecosystem • Utilize multiple development languages/tools such as Python, SPARK, Hive, R, Java • Fast prototype solutions by integrating various open source components • Operationalize open source data-analytic tools for enterprise use • Develop real-time data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark, NIFI, Python, HBase and Hadoop What We Are Looking For • Problem Solver who likes to tackle challenging issues and pursue them through resolution. • Passionate learner who enjoys education through class room training and self-discovery on a variety of emerging technologies • Someone who enjoys challenges, excels under pressure and delivers innovative business solutions • Collaborative individual who excels in working within a team and with business partners to identify, develop and deliver solutions to big data challenges Preferred Qualifications (for Paygrade 5): • Bachelor's Degree or higher level of education • 2+ years experience developing Java solutions • 2+ years of experience with Spark, Python, Hive, HBase and/or Hadoop • 2+ years of experience with SQL, Linux • Experience with Agile development practices *Mandatory:* Hadoop Python Spark *Ideal Candidate:* NIFI Kafka Amith Singh * Sr. Recruitment Manager* *Zenith tech Solutions* * Desk: **518-621-004* <518-621-0048>*6* * Fax: **518-244-4977* <518-244-4977> * 3 park HillAlbany, NY 12204 **ami...@zenithtechsolutions.com <a...@zenithtechsolutions.com>* -- You received this message because you are subscribed to the Google Groups "Hot List" group. To unsubscribe from this group and stop receiving emails from it, send an email to directclienteq+unsubscr...@googlegroups.com. To post to this group, send email to directclienteq@googlegroups.com. Visit this group at https://groups.google.com/group/directclienteq. For more options, visit https://groups.google.com/d/optout.