This is done using the Jet API programmatically from your Java application. Jet is meant to be used operationally with developed and deployed applications.<p>We can use HDFS as a source or a sink.
See <a href="https://github.com/hazelcast/hazelcast-jet-code-samples/blob/0.6-maintenance/batch/wordcount-hadoop/src/main/java/HadoopWordCount.java" rel="nofollow">https://github.com/hazelcast/hazelcast-jet-code-samples/blob...</a> for a HDFS Wordcount example.<p>Jet jobs run in an isolated class loader, which is distributed to the cluster when the Job is started. You do this by adding classes/jars to JobConfig. See <a href="http://docs.hazelcast.org/docs/jet/0.6/manual/#practical-" rel="nofollow">http://docs.hazelcast.org/docs/jet/0.6/manual/#practical-</a> considerations for details.
If I already have a Hadoop cluster, which can run a spark job in a jar file on HDFS with spark-submit, how can I install Hazelcast Jet so that I can do the same as with Spark?