spark.conf 1.6 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647
  1. # Spark home
  2. hibench.spark.home /spark
  3. # Spark master
  4. # standalone mode: spark://xxx:7077
  5. # YARN mode: yarn-client
  6. hibench.spark.master k8s://https://172.9.0.240:6443
  7. # executor number and cores when running on Yarn
  8. hibench.yarn.executor.num 1
  9. hibench.yarn.executor.cores 4
  10. # executor and driver memory in standalone & YARN mode
  11. spark.executor.memory 4g
  12. spark.driver.memory 4g
  13. # set spark parallelism property according to hibench's parallelism value
  14. spark.default.parallelism ${hibench.default.map.parallelism}
  15. # set spark sql's default shuffle partitions according to hibench's parallelism value
  16. spark.sql.shuffle.partitions ${hibench.default.shuffle.parallelism}
  17. #======================================================
  18. # Spark Streaming
  19. #======================================================
  20. # Spark streaming Batchnterval in millisecond (default 100)
  21. hibench.streambench.spark.batchInterval 100
  22. # Number of nodes that will receive kafka input (default: 4)
  23. hibench.streambench.spark.receiverNumber 4
  24. # Indicate RDD storage level. (default: 2)
  25. # 0 = StorageLevel.MEMORY_ONLY
  26. # 1 = StorageLevel.MEMORY_AND_DISK_SER
  27. # other = StorageLevel.MEMORY_AND_DISK_SER_2
  28. hibench.streambench.spark.storageLevel 2
  29. # indicate whether to test the write ahead log new feature (default: false)
  30. hibench.streambench.spark.enableWAL false
  31. # if testWAL is true, this path to store stream context in hdfs shall be specified. If false, it can be empty (default: /var/tmp)
  32. hibench.streambench.spark.checkpointPath /var/tmp
  33. # whether to use direct approach or not (dafault: true)
  34. hibench.streambench.spark.useDirectMode true