hadoop - Spark Streaming: What are things we should monitor to keep the streaming running? -
4 core is a spark project running on 16 GB (both Master / Worker) example, now someone can tell me what is i to monitor things so that my cluster / jobs never go down?
I have created a small list that contains the following items, please expand the list if you know more:
- Network Connectivity Monitor for Master / Employee
- Monitor Spark Jobs
-
This is a good list but in addition to those people I actually have some of the data for the receiver's status of streaming apps (assuming you -acdifs source) will be monitored, whether they are connected or not. Well, to be honest, it was difficult to do this with previous versions of Spark Streaming because the existence of the devices did not exist in order to obtain the status of the receiver. However, with SPARC 1.0 (to be released soon), you can use the org.apache.spark.streaming.StreamingListener interface to get events regarding the status of the receiver.
There is a quietly top-to-recorded spark 1.0 docs
Comments
Post a Comment