hcbo / rfs

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Usage

配置

配置文件位于src/main/resources/hcbConfig.properties

#spark任务带状态操作符个数
ops=3
#数据来源个数
sources=10
#Zookeeper servers
zkServers=192.168.225.6:2181,192.168.225.6:2182,192.168.225.6:2183
#Kafka brokers
kafkaServers=192.168.225.6:9092,192.168.225.6:9093,192.168.225.6:9094
#如果spark检查点文件partition个数为默认未做修改的话,不需要修改此配置
sparkPartitions=200
#运行日志路径
logPath=/Users/hcb/Documents/logs2/log4j/neu2025.log

Package

执行ex.sh,将生成的target/hcbMfs-1.0.0.jar放在spark任务的依赖中.

使用示例代码

val dataStreamWriter = wordcount
      .writeStream
      .queryName("kafka_test2")
			// localhost替换为kafka的一个broker ip
      .option("checkpointLocation","mfs://localhost:8888/china")
      .outputMode(OutputMode.Complete())
      .format("console")

val query = dataStreamWriter.start()
query.awaitTermination()

About


Languages

Language:Java 99.7%Language:Shell 0.3%