| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647 |
- # Spark home
- hibench.spark.home /spark
- # Spark master
- # standalone mode: spark://xxx:7077
- # YARN mode: yarn-client
- hibench.spark.master k8s://https://172.9.0.240:6443
- # executor number and cores when running on Yarn
- hibench.yarn.executor.num 1
- hibench.yarn.executor.cores 4
- # executor and driver memory in standalone & YARN mode
- spark.executor.memory 4g
- spark.driver.memory 4g
- # set spark parallelism property according to hibench's parallelism value
- spark.default.parallelism ${hibench.default.map.parallelism}
- # set spark sql's default shuffle partitions according to hibench's parallelism value
- spark.sql.shuffle.partitions ${hibench.default.shuffle.parallelism}
- #======================================================
- # Spark Streaming
- #======================================================
- # Spark streaming Batchnterval in millisecond (default 100)
- hibench.streambench.spark.batchInterval 100
- # Number of nodes that will receive kafka input (default: 4)
- hibench.streambench.spark.receiverNumber 4
- # Indicate RDD storage level. (default: 2)
- # 0 = StorageLevel.MEMORY_ONLY
- # 1 = StorageLevel.MEMORY_AND_DISK_SER
- # other = StorageLevel.MEMORY_AND_DISK_SER_2
- hibench.streambench.spark.storageLevel 2
- # indicate whether to test the write ahead log new feature (default: false)
- hibench.streambench.spark.enableWAL false
- # if testWAL is true, this path to store stream context in hdfs shall be specified. If false, it can be empty (default: /var/tmp)
- hibench.streambench.spark.checkpointPath /var/tmp
- # whether to use direct approach or not (dafault: true)
- hibench.streambench.spark.useDirectMode true
|