You can now work on data directly imported from HDFS (CSV / Parquet / Avro) and publish them again on your cluster. The authentication to the HDFS cluster is secured via Kerberos.
The tDataprepRun component can now be used to apply one of your preparations in a Spark batch Job designed in Talend Studio.