Apache Hadoop生态组件部署分享-Kafka

zookeeper: Apache Hadoop生态组件部署分享-zookeeper

hadoop:Apache Hadoop生态组件部署分享-Hadoop

hive: Apache Hadoop生态组件部署分享-Hive

hbase: Apache Hadoop生态组件部署分享-Hbase

impala:Apache Hadoop生态组件部署分享-Impala

spark: Apache Hadoop生态组件部署分享-Spark

sqoop: Apache Hadoop生态组件部署分享-Sqoop

下载地址: https://kafka.apache.org/downloads

文档地址: https://kafka.apache.org/documentation/#java

说明: kafaka4.1.0 已经脱离了zk的依赖,因此你可以把它当成独立的组件去使用.(这里需要提高java版本,在启动脚本加上export JAVA_HOME=/opt/module/jdk-17.0.12即可,这样也不会影响其他组件)

1、上传并解压并分发

apache 复制代码
tar -xf kafka_2.13-4.1.0.tgz -C /opt/apache/scp -rp kafka_2.13-4.1.0/ 192.168.242.231:/opt/apache/scp -rp kafka_2.13-4.1.0/ 192.168.242.232:/opt/apache/

2、修改配置

A. apache230.hadoop.com 配置内容

ini 复制代码
process.roles=broker,controller
node.id=1#controller.quorum.bootstrap.servers=apache230.hadoop.com:9093,apache231.hadoop.com:9093,apache232.hadoop.com:9093controller.quorum.voters=1@apache230.hadoop.com:9093,2@apache231.hadoop.com:9093,3@apache232.hadoop.com:9093############################# Socket Server Settings #############################advertised.listeners=PLAINTEXT://apache230.hadoop.com:9092controller.listener.names=CONTROLLERlistener.security.protocol.map=CONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSLnum.network.threads=3num.io.threads=8socket.send.buffer.bytes=102400socket.receive.buffer.bytes=102400socket.request.max.bytes=104857600

############################# Log Basics #############################log.dirs=/opt/apache/kafka_2.13-4.1.0/kraft-combined-logs_1
num.partitions=1num.recovery.threads.per.data.dir=1
############################# Internal Topic Settings  #############################offsets.topic.replication.factor=1share.coordinator.state.topic.replication.factor=1share.coordinator.state.topic.min.isr=1transaction.state.log.replication.factor=1transaction.state.log.min.isr=1
log.retention.hours=168log.segment.bytes=1073741824log.retention.check.interval.ms=300000

B. apache231.hadoop.com 配置内容

ruby 复制代码
process.roles=broker,controllernode.id=2#controller.quorum.bootstrap.servers=apache230.hadoop.com:9093,apache231.hadoop.com:9093,apache232.hadoop.com:9093controller.quorum.voters=1@apache230.hadoop.com:9093,2@apache231.hadoop.com:9093,3@apache232.hadoop.com:9093############################# Socket Server Settings #############################advertised.listeners=PLAINTEXT://apache231.hadoop.com:9092controller.listener.names=CONTROLLERlistener.security.protocol.map=CONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSLnum.network.threads=3num.io.threads=8socket.send.buffer.bytes=102400socket.receive.buffer.bytes=102400socket.request.max.bytes=104857600############################# Log Basics #############################log.dirs=/opt/apache/kafka_2.13-4.1.0/kraft-combined-logs_1num.partitions=1num.recovery.threads.per.data.dir=1############################# Internal Topic Settings  #############################offsets.topic.replication.factor=1share.coordinator.state.topic.replication.factor=1share.coordinator.state.topic.min.isr=1transaction.state.log.replication.factor=1transaction.state.log.min.isr=1log.retention.hours=168log.segment.bytes=1073741824log.retention.check.interval.ms=300000

C. apache232.hadoop.com 配置内容

ini 复制代码
process.roles=broker,controller
node.id=3#controller.quorum.bootstrap.servers=apache230.hadoop.com:9093,apache231.hadoop.com:9093,apache232.hadoop.com:9093controller.quorum.voters=1@apache230.hadoop.com:9093,2@apache231.hadoop.com:9093,3@apache232.hadoop.com:9093############################# Socket Server Settings #############################advertised.listeners=PLAINTEXT://apache232.hadoop.com:9092controller.listener.names=CONTROLLERlistener.security.protocol.map=CONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSLnum.network.threads=3num.io.threads=8socket.send.buffer.bytes=102400socket.receive.buffer.bytes=102400socket.request.max.bytes=104857600

############################# Log Basics #############################log.dirs=/opt/apache/kafka_2.13-4.1.0/kraft-combined-logs_1
num.partitions=1num.recovery.threads.per.data.dir=1
############################# Internal Topic Settings  #############################offsets.topic.replication.factor=1share.coordinator.state.topic.replication.factor=1share.coordinator.state.topic.min.isr=1transaction.state.log.replication.factor=1transaction.state.log.min.isr=1
log.retention.hours=168log.segment.bytes=1073741824log.retention.check.interval.ms=300000

3、生成UUI并格式化kafka

python 复制代码
[root@apache230 kafka_2.13-4.1.0]# ./bin/kafka-storage.sh random-uuid2025-10-10 09:54:31,580 INFO utils.Log4jControllerRegistration$: Registered `kafka:type=kafka.Log4jController` MBeaniDUiq4ziSGC65Y38y9fOvA
三台节点都执行bin/kafka-storage.sh format --cluster-id iDUiq4ziSGC65Y38y9fOvA --config config/server.properties

4、启动kafka

bash 复制代码
bin/kafka-server-start.sh config/server.properties

5、kafka相关操作验证

python 复制代码
[root@apache232 kafka_2.13-4.1.0]# bin/kafka-topics.sh --create --topic quickstart-events --bootstrap-server apache232.hadoop.com:9092Created topic quickstart-events.

bin/kafka-topics.sh --describe --topic quickstart-events --bootstrap-server apache232.hadoop.com:9092[root@apache232 kafka_2.13-4.1.0]# bin/kafka-topics.sh --describe --topic quickstart-events --bootstrap-server apache232.hadoop.com:9092Topic: quickstart-events        TopicId: Y3um9focTcmQkzRd5q8SaA PartitionCount: 1       ReplicationFactor: 1    Configs: min.insync.replicas=1,segment.bytes=1073741824        Topic: quickstart-events        Partition: 0    Leader: 2       Replicas: 2     Isr: 2  Elr:    LastKnownElr:
#生产者bin/kafka-console-producer.sh --topic quickstart-events --bootstrap-server apache232.hadoop.com:9092
#消费者bin/kafka-console-consumer.sh --topic quickstart-events --from-beginning --bootstrap-server apache232.hadoop.com:9092

消费截图

相关推荐
武子康17 小时前
大数据-243 离线数仓 - 实战电商核心交易增量导入(DataX - HDFS - Hive 分区
大数据·后端·apache hive
代码匠心2 天前
从零开始学Flink:Flink SQL四大Join解析
大数据·flink·flink sql·大数据处理
武子康3 天前
大数据-242 离线数仓 - DataX 实战:MySQL 全量/增量导入 HDFS + Hive 分区(离线数仓 ODS
大数据·后端·apache hive
SelectDB4 天前
易车 × Apache Doris:构建湖仓一体新架构,加速 AI 业务融合实践
大数据·agent·mcp
武子康5 天前
大数据-241 离线数仓 - 实战:电商核心交易数据模型与 MySQL 源表设计(订单/商品/品类/店铺/支付)
大数据·后端·mysql
茶杯梦轩5 天前
从零起步学习RabbitMQ || 第三章:RabbitMQ的生产者、Broker、消费者如何保证消息不丢失(可靠性)详解
分布式·后端·面试
IvanCodes5 天前
一、消息队列理论基础与Kafka架构价值解析
大数据·后端·kafka
武子康6 天前
大数据-240 离线数仓 - 广告业务 Hive ADS 实战:DataX 将 HDFS 分区表导出到 MySQL
大数据·后端·apache hive
回家路上绕了弯6 天前
深入解析Agent Subagent架构:原理、协同逻辑与实战落地指南
分布式·后端
初次攀爬者6 天前
Kafka的Rebalance基础介绍
后端·kafka