1) We consume from kafka using structured streaming and write the processed data set to s3. We also want to write the processed data to kafka, moving forward, is it possible to do this from the same streaming request? (spark version 2.1.1)
2) In the logs, I see the result of the streaming request, and I have the duration of the JSON sample from the log, can someone please explain what is the difference between addBatch and getBatch ?
3) TriggerExecution - does it take time to process the extracted data and write to the receiver?
"durationMs" : { "addBatch" : 2263426, "getBatch" : 12, "getOffset" : 273, "queryPlanning" : 13, "triggerExecution" : 2264288, "walCommit" : 552 },
Regards aravias
source share