Hello,

We are looking for a developer to help us with a small ETL project using Spark 
and Kubernetes. Here are some of the requirements:

1. We need a REST API to run and schedule jobs. We would prefer this done in 
Node.js but can be done using Java. The REST API will not be available to the 
public.
2. We need an easy way to create new jobs in Java without deploying the whole 
server again.
3. We want jobs deployed/ran using Kubernetes.
4. Must be able to scale to 1000s of ETL jobs.
5. Source for data will be one REST API.
6. Destination for data will be one Couchbase Database cluster. (Couchbase also 
uses a REST API)
7. I am not sure how many records will be processed per job.
8. The data is mostly sales related data.

I know there are commercial ETL solutions that do everything I want. We are 
looking for something simple and do not need a fancy UI to describe our ETL. We 
want to use Spark and Java to programmatically describe out ETL jobs.

Please let me know if you are interested.

Thanks,

Warren Bell
-- 


**********************************************************************

This
 email and any files transmitted with it are confidential and intended 

solely for the use of the individual or entity to whom they are 

addressed. They may not be disseminated or distributed to persons or 

entities other than the ones intended without the authority of the 
sender. 
If you have received this email in error or are not the
 intended 
recipient, you may not use, copy, disseminate or distribute 
it. Delete it 
immediately from your system and notify the sender 
promptly
 by email that 
you have done so. This footnote also confirms that this 
email message has 
been scanned for the presence of computer viruses.


********************************************************************** 

 
Please consider the environment before printing

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to