Я использую Docker Toolbox для использования Spark в Windows 10.
Образ докера sequenceiq/spark
.https://hub.docker.com/r/sequenceiq/spark/
Тем не менее, отображается много «логов».
scala> dataRDD.collect()
18/12/07 03:59:48 INFO mapred.FileInputFormat: Total input paths to process : 1
18/12/07 03:59:49 INFO spark.SparkContext: Starting job: collect at <console>:32
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Got job 0 (collect at <console>:32) with 1 output partitions
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Final stage: ResultStage 0 (collect at <console>:32)
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Parents of final stage: List()
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Missing parents: List()
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Submitting ResultStage 0 (MapPartitionsRDD[5] at map at <console>:29), which has no missing parents
18/12/07 03:59:49 INFO storage.MemoryStore: Block broadcast_2 stored as values in memory (estimated size 3.2 KB, free 357.5 KB)
18/12/07 03:59:49 INFO storage.MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 1870.0 B, free 359.3 KB)
18/12/07 03:59:49 INFO storage.BlockManagerInfo: Added broadcast_2_piece0 in memory on localhost:39093 (size: 1870.0 B, free: 517.4 MB)
18/12/07 03:59:49 INFO spark.SparkContext: Created broadcast 2 from broadcast at DAGScheduler.scala:1006
18/12/07 03:59:49 INFO scheduler.DAGScheduler: Submitting 1 missing tasks from ResultStage 0 (MapPartitionsRDD[5] at map at <console>:29)
...
Могу ли я скрыть их благодаря опции во время запуска, может быть?
Спасибо,