*Data Engineer*
*Irvine, CA*
*Ingram Micro*
*7-10 Years*
*Best Rate on C2C*
*Very Long term Contract*
*Need H1B/H1 Transfer, USC/GCOPT & CPT 1990 & Below Year Of Birth*
*===*
*sek...@chabeztech.com <sek...@chabeztech.com>*
*sekhar.salesrecrui...@gmail.com <sekhar.salesrecrui...@gmail.com> *
*(No Phone Calls Please Message me In Hangout)*
*434-322-084*

Detailed Requirement:

Data Engineer
·         Data flow design and implementation:
·         Create and maintain optimal data pipeline architecture.
·         Gather and process large, complex, raw data sets at scale
(including writing scripts, web scraping, calling APIs, write SQL queries,
etc.) that meet functional / non-functional business requirements.
·         Identify, design, and implement internal process improvements:
automating manual processes, optimizing data delivery, re-designing
infrastructure for greater scalability, etc.
·         Build the infrastructure required for optimal extraction,
transformation, and loading of data from a wide variety of data sources
using SQL and ‘big data’ technologies.
·         Work with stakeholders including the Business, Product, Data and
Design teams to assist with data-related technical issues and support their
data infrastructure needs.
·         Create data tools for analytics and engineering team members that
assist them in building and optimizing our product into an innovative
industry leader.
·         Be a data expert to strive for greater functionality in our data
systems.

Data solutions development:
·         Employing your skills in designing, developing and delivering
world class data algorithmic artifacts, including documentation and coding;
coordinate data algorithmic development with infrastructural development
·         Work closely with our engineering team to integrate your amazing
innovations and algorithms into our products.
·         Research and apply advanced algorithms and methods involving data
mining, statistical analysis and machine learning techniques
·         Process unstructured data into a form suitable for analysis – and
then do the analysis.
·         Support business decisions with ad hoc analysis as needed.
·         Master third party systems and interfaces, including: data
available by the parties, API to be used for obtaining the data,
limitations related to these interfaces
·         Excellent subject matter expertise in designing algorithms,
business logics to automate commerce process flows.
·         Apply your broad-based data development expertise to create
practical and innovative solutions
·         Efficiently implement clean, maintainable, and testable data
solutions with high availability, blazing speed in performance and fault
tolerant.
·         Participate in agile project execution and provide accurate work
effort estimates
·         Apply excellent communications skills, creativity and practical
knowledge to benefit our customers
·         What you bring to the role:
·         Bachelor's degree in Computer Science, Engineering, Science and
Math or related technical discipline is required
·         Preferred: an MBA (or equivalent) from a top-tier institution, or
equivalent business experience preferred
·         7-10 years of technical experience, with at least 5+ years of
experience with web services development and middleware applications or
Master’s degree plus 5-7 years of technical experience.
·         Advanced working SQL knowledge and experience working with
relational databases, query authoring (SQL) as well as working familiarity
with a variety of databases.
·         Experience building and optimizing ‘big data’ data pipelines,
architectures and data sets.
·         Experience performing root cause analysis on internal and
external data and processes to answer specific business questions and
identify opportunities for improvement.
·         Strong analytic skills related to working with unstructured
datasets.
·         Build processes supporting data transformation, data structures,
metadata, dependency and workload management.
·         A successful history of manipulating, processing and extracting
value from large disconnected datasets.
·         Working knowledge of message queuing, stream processing, and
highly scalable ‘big data’ data stores.
·         Experience with big data tools: Hadoop, Spark, Kafka, etc.
·         Experience with relational SQL and NoSQL databases, including
Postgres and Cassandra.
·         Experience with data pipeline and workflow management tools:
Azkaban, Luigi, Airflow, etc.
·         Experience with stream-processing systems: Storm,
Spark-Streaming, etc.
·         Experience with object-oriented/object function scripting
languages: Java, C# etc.

-- 
*Respectfully Regards,*
*N . Sekhar*
*Sr. Technical Recruiter*
*Hangouts : sekhar.salesrecrui...@gmail.com
<sekhar.salesrecrui...@gmail.com>*
*Linkedin : linkedin.com/in/sekhar-nallamelli
<http://linkedin.com/in/sekhar-nallamelli>*
*Skype : sekhar.nallamelli*
This e-mail and any files transmitted with it are for the sole use of the
intended recipient(s) and may contain confidential and privileged
information. If you are not the intended recipient(s), please reply to the
sender and destroy all copies of the original message. Any unauthorized
review, use, disclosure, dissemination, forwarding, printing or copying of
this email, and/or any action taken in reliance on the contents of this
e-mail is strictly prohibited and may be unlawful. Where permitted by
applicable law, this e-mail and other e-mail communications sent to and
from Augusta e-mail addresses may be monitored.

-- 
You received this message because you are subscribed to the Google Groups 
"CorptoCorp" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to corptocorp+unsubscr...@googlegroups.com.
To post to this group, send email to corptocorp@googlegroups.com.
Visit this group at https://groups.google.com/group/corptocorp.
For more options, visit https://groups.google.com/d/optout.

Reply via email to