kafka单机版安装测试
2016-11-19 14:55
429 查看
1.kafka
1.1可以让你发布订阅你的消息,类似于消息系统
1.2可以让你以容错方式存储记录流
1.3可以让你处理记录流当它们发生时。
kafka的优势
构建可靠地在系统或应用程序之间获取数据的实时流数据流水线
构建变换或响应数据流的实时流应用程序
2.安装
下载地址:http://mirrors.cnnic.cn/apache/kafka/0.10.1.0/kafka_2.11-0.10.1.0.tgz
安装略
3.启动测试
3.1启动自带的zookeeper
3.2启动server
3.3创建topic
3.4测试topic是否创建成功
查看topic的详细描述
./kafka-topics.sh --describe --zookeeper localhost:2181 --topic test
3.5命令行启动生产者发送hello kafka的消息
3.6命令行启动消费者收到消息
4.java api (基于2.11-0.10.1.0 新版本api)
producer生产者
/**
* caicongyang.com Inc.
* Copyright (c) 2004-2016 All Rights Reserved.
*/
package com.caicongyang.kafka;
import java.util.Properties;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.Producer;
import org.apache.kafka.clients.producer.ProducerRecord;
/**
* @author caicongyang1
* @version id: producer, v 0.1 16/11/14 下午4:50 caicongyang1 Exp $$
*/
public class producer {
public static void main(String[] args) {
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
//all,-1:所有副本都同步完成,才确认消息发送成功
//1:一个副本写入完成既确认消息发送成功
//0:无需等待broker确认消息写入,不触发重试机制
props.put("acks", "all");
props.put("retries", 3);
props.put("batch.size", 16384);
props.put("linger.ms", 1);
props.put("buffer.memory", 33554432);
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
Producer<String, String> producer = new KafkaProducer<>(props);
for(int i = 0; i < 100; i++){
producer.send(new ProducerRecord<String, String>("test", Integer.toString(i), Integer.toString(i)));
}
producer.close();
}
}
consumer消费者
/**
* caicongyang.com Inc.
* Copyright (c) 2004-2016 All Rights Reserved.
*/
package com.caicongyang.kafka;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import java.util.Arrays;
import java.util.Properties;
/**
* @author caicongyang1
* @version id: consumer, v 0.1 16/11/14 下午5:15 caicongyang1 Exp $$
*/
public class consumer {
public static void main(String[] args) {
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
props.put("group.id", "test");
props.put("enable.auto.commit", "true");
props.put("auto.commit.interval.ms", "1000");
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
consumer.subscribe(Arrays.asList("test"));
while (true) {
ConsumerRecords<String, String> records = consumer.poll(10);
for (ConsumerRecord<String, String> record : records)
System.out.printf("offset = %d, key = %s, value = %s%n", record.offset(), record.key(), record.value());
}
}
}
1.1可以让你发布订阅你的消息,类似于消息系统
1.2可以让你以容错方式存储记录流
1.3可以让你处理记录流当它们发生时。
kafka的优势
构建可靠地在系统或应用程序之间获取数据的实时流数据流水线
构建变换或响应数据流的实时流应用程序
2.安装
下载地址:http://mirrors.cnnic.cn/apache/kafka/0.10.1.0/kafka_2.11-0.10.1.0.tgz
安装略
3.启动测试
3.1启动自带的zookeeper
nohup ./zookeeper-server-start.sh ../config/zookeeper.properties &
3.2启动server
nohup ./kafka-server-start.sh ../config/server.properties &
3.3创建topic
./kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic test
3.4测试topic是否创建成功
./kafka-topics.sh --list --zookeeper localhost:2181
查看topic的详细描述
./kafka-topics.sh --describe --zookeeper localhost:2181 --topic test
3.5命令行启动生产者发送hello kafka的消息
./kafka-console-producer.sh --broker-list localhost:9092 --topic test hello kafka
3.6命令行启动消费者收到消息
./kafka-console-consumer.sh --zookeeper localhost:2181 --topic test --from-beginning hello kafka
4.java api (基于2.11-0.10.1.0 新版本api)
producer生产者
/**
* caicongyang.com Inc.
* Copyright (c) 2004-2016 All Rights Reserved.
*/
package com.caicongyang.kafka;
import java.util.Properties;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.Producer;
import org.apache.kafka.clients.producer.ProducerRecord;
/**
* @author caicongyang1
* @version id: producer, v 0.1 16/11/14 下午4:50 caicongyang1 Exp $$
*/
public class producer {
public static void main(String[] args) {
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
//all,-1:所有副本都同步完成,才确认消息发送成功
//1:一个副本写入完成既确认消息发送成功
//0:无需等待broker确认消息写入,不触发重试机制
props.put("acks", "all");
props.put("retries", 3);
props.put("batch.size", 16384);
props.put("linger.ms", 1);
props.put("buffer.memory", 33554432);
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
Producer<String, String> producer = new KafkaProducer<>(props);
for(int i = 0; i < 100; i++){
producer.send(new ProducerRecord<String, String>("test", Integer.toString(i), Integer.toString(i)));
}
producer.close();
}
}
consumer消费者
/**
* caicongyang.com Inc.
* Copyright (c) 2004-2016 All Rights Reserved.
*/
package com.caicongyang.kafka;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import java.util.Arrays;
import java.util.Properties;
/**
* @author caicongyang1
* @version id: consumer, v 0.1 16/11/14 下午5:15 caicongyang1 Exp $$
*/
public class consumer {
public static void main(String[] args) {
Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
props.put("group.id", "test");
props.put("enable.auto.commit", "true");
props.put("auto.commit.interval.ms", "1000");
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
consumer.subscribe(Arrays.asList("test"));
while (true) {
ConsumerRecords<String, String> records = consumer.poll(10);
for (ConsumerRecord<String, String> record : records)
System.out.printf("offset = %d, key = %s, value = %s%n", record.offset(), record.key(), record.value());
}
}
}
相关文章推荐
- Kafka 单机安装、测试和监控
- kafka安装配置及测试-单机版(转)
- kafka单机简单安装与测试
- kafka windows单机安装测试
- kafka windows单机安装测试
- kafka windows单机安装测试
- kafka 单机的安装和测试
- kafka 安装 单机测试
- kafka安装和测试
- Kafka-0.8.0单机配置安装
- [bigdata] kafka集群安装及测试
- [Hadoop]单机尝试安装并测试Hadoop2.7.1(附带注释脚本)
- Linux下部署Kafka分布式集群,安装与测试
- kafka单机安装
- Storm和Kafka的集成安装和测试
- 集群安装完毕,该如何测试和使用集群-hadoop单机(伪分布)
- 单机安装kafka
- Kafka单机安装
- kafka2.9.2的分布式集群安装和demo(java api)测试
- kafka 集群安装与安装测试