热门标签 | HotTags
当前位置:  开发笔记 > 编程语言 > 正文

基于Web的Kafka管理器工具之Kafkamanager的编译部署详细安装(支持kafka0.8、0.9和0.10以后版本)(图文详解)(默认端口或任意自定义端口)

问题详情无奈于,在kafka里没有一个较好自带的webui。启动后无法观看,并且不友好。所以,需安装一个第三方的kafka管理工具为了简化开发者和服务工程师维护

问题详情

无奈于,在kafka里没有一个较好自带的web ui。启动后无法观看,并且不友好。所以,需安装一个第三方的kafka管理工具

为了简化开发者和服务工程师维护Kafka集群的工作,yahoo构建了一个叫做Kafka管理器的基于Web工具,叫做 Kafka Manager。这个管理工具可以很容易地发现分布在集群中的哪些topic分布不均匀,或者是分区在整个集群分布不均匀的的情况。

它支持管理多个集群、选择副本、副本重新分配以及创建Topic。同时,这个管理工具也是一个非常好的可以快速浏览这个集群的工具。

有如下功能:

管理多个kafka集群
便捷的检查kafka集群状态(topics,brokers,备份分布情况,分区分布情况)
选择你要运行的副本
基于当前分区状况进行
可以选择topic配置并创建topic(0.8.1.1和0.8.2的配置不同)
删除topic(只支持0.8.2以上的版本并且要在broker配置中设置delete.topic.enable=true)
Topic list会指明哪些topic被删除(在0.8.2以上版本适用)
为已存在的topic增加分区
为已存在的topic更新配置
在多个topic上批量重分区
在多个topic上批量重分区(可选partition broker位置)

大家编译的步骤,可以参考

或者

我这里不多说。大家去看看这个流程就好

直接采用下面这位博主已经编译好之后分享的。

谢谢他!

下载地址: 

  若此连接失效,则大家可以在我这篇博客下方留言评论,我将无偿发送给你们。

步骤:

1、解压kafka-manager-1.3.2.1.zip

lrwxrwxrwx. hadoop hadoop Apr : hadoop - hadoop-2.6.
drwxr-xr-x. hadoop hadoop Apr : hadoop-2.6.
lrwxrwxrwx. hadoop hadoop Apr : hbase - hbase-0.98.
drwxrwxr-x. hadoop hadoop Apr : hbase-0.98.
lrwxrwxrwx. hadoop hadoop Apr : hive - hive-1.0.
drwxrwxr-x. hadoop hadoop Jul : hive-1.0.
lrwxrwxrwx. hadoop hadoop Apr : jdk - jdk1..0_79
drwxr-xr-x. hadoop hadoop Apr jdk1..0_79
drwxr-xr-x. hadoop hadoop Aug jdk1..0_60
lrwxrwxrwx hadoop hadoop Jul : kafka - kafka_2.-0.9.0.1/
drwxr-xr-x hadoop hadoop Jul : kafka_2.-0.9.0.1
drwxr-xr-x hadoop hadoop May : kafka_2.-0.8.2.2
-rw-r--r-- hadoop hadoop Jul : kafka-manager-1.3.2.1.zip
lrwxrwxrwx hadoop hadoop Apr : kibana - kibana-4.6.-linux-x86_64/
drwxrwxr-x hadoop hadoop Nov kibana-4.6.-linux-x86_64
lrwxrwxrwx hadoop hadoop Jul : phoenix - apache-phoenix-4.10.-HBase-0.98-bin/
lrwxrwxrwx hadoop hadoop Jun : scala - scala-2.11./
drwxrwxr-x hadoop hadoop Feb scala-2.10.
drwxrwxr-x hadoop hadoop Mar scala-2.11.
lrwxrwxrwx hadoop hadoop May : snappy - snappy-1.1.
drwxr-xr-x hadoop hadoop May : snappy-1.1.
lrwxrwxrwx hadoop hadoop Jun : spark - spark-1.6.-bin-hadoop2./
drwxr-xr-x hadoop hadoop Feb spark-1.6.-bin-hadoop2.
lrwxrwxrwx. hadoop hadoop Apr : sqoop - sqoop-1.4.
drwxr-xr-x. hadoop hadoop May : sqoop-1.4.
lrwxrwxrwx hadoop hadoop May : storm - apache-storm-1.0./
lrwxrwxrwx hadoop hadoop Jun : zeppelin - zeppelin-0.5.-incubating-bin-all/
drwxr-xr-x hadoop hadoop Jun : zeppelin-0.5.-incubating-bin-all
lrwxrwxrwx. hadoop hadoop Apr : zookeeper - zookeeper-3.4.
drwxr-xr-x. hadoop hadoop Apr : zookeeper-3.4.
[hadoop@master app]$ unzip kafka-manager-1.3.2.1.zip

2、cd kafka-manager-1.3.2.1

lrwxrwxrwx. hadoop hadoop Apr : hive - hive-1.0.
drwxrwxr-x. hadoop hadoop Jul : hive-1.0.
lrwxrwxrwx. hadoop hadoop Apr : jdk - jdk1..0_79
drwxr-xr-x. hadoop hadoop Apr jdk1..0_79
drwxr-xr-x. hadoop hadoop Aug jdk1..0_60
lrwxrwxrwx hadoop hadoop Jul : kafka - kafka_2.-0.9.0.1/
drwxr-xr-x hadoop hadoop Jul : kafka_2.-0.9.0.1
drwxr-xr-x hadoop hadoop May : kafka_2.-0.8.2.2
drwxrwxr-x hadoop hadoop Jul : kafka-manager-1.3.2.1
-rw-r--r-- hadoop hadoop Jul : kafka-manager-1.3.2.1.zip
lrwxrwxrwx hadoop hadoop Apr : kibana - kibana-4.6.-linux-x86_64/
drwxrwxr-x hadoop hadoop Nov kibana-4.6.-linux-x86_64
lrwxrwxrwx hadoop hadoop Jul : phoenix - apache-phoenix-4.10.-HBase-0.98-bin/
lrwxrwxrwx hadoop hadoop Jun : scala - scala-2.11./
drwxrwxr-x hadoop hadoop Feb scala-2.10.
drwxrwxr-x hadoop hadoop Mar scala-2.11.
lrwxrwxrwx hadoop hadoop May : snappy - snappy-1.1.
drwxr-xr-x hadoop hadoop May : snappy-1.1.
lrwxrwxrwx hadoop hadoop Jun : spark - spark-1.6.-bin-hadoop2./
drwxr-xr-x hadoop hadoop Feb spark-1.6.-bin-hadoop2.
lrwxrwxrwx. hadoop hadoop Apr : sqoop - sqoop-1.4.
drwxr-xr-x. hadoop hadoop May : sqoop-1.4.
lrwxrwxrwx hadoop hadoop May : storm - apache-storm-1.0./
lrwxrwxrwx hadoop hadoop Jun : zeppelin - zeppelin-0.5.-incubating-bin-all/
drwxr-xr-x hadoop hadoop Jun : zeppelin-0.5.-incubating-bin-all
lrwxrwxrwx. hadoop hadoop Apr : zookeeper - zookeeper-3.4.
drwxr-xr-x. hadoop hadoop Apr : zookeeper-3.4.
[hadoop@master app]$ rm kafka-manager-1.3.2.1.zip
[hadoop@master app]$ cd kafka-manager-1.3.2.1/

3、 修改conf/application.conf文件,特别是kafka-manager.zkhosts的配置

[hadoop@master kafka-manager-1.3.2.1]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1
[hadoop@master kafka-manager-1.3.2.1]$ ll
total
drwxrwxr-x hadoop hadoop Jul : bin
drwxrwxr-x hadoop hadoop Jul : conf
drwxrwxr-x hadoop hadoop Jul : lib
-rw-r--r-- hadoop hadoop Feb : README.md
drwxrwxr-x hadoop hadoop Jul : share
[hadoop@master kafka-manager-1.3.2.1]$ cd conf/
[hadoop@master conf]$
[hadoop@master conf]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1/conf
[hadoop@master conf]$ ll
total
-rw-r--r-- hadoop hadoop Feb : application.conf
-rw-r--r-- hadoop hadoop Feb : consumer.properties
-rw-r--r-- hadoop hadoop Feb : logback.xml
-rw-r--r-- hadoop hadoop Feb : logger.xml
-rw-r--r-- hadoop hadoop Feb : routes
[hadoop@master conf]$ vim application.conf

以下是默认,我贴出来,大家学习学习

# Copyright Yahoo Inc. Licensed under the Apache License, Version 2.0
# See accompanying LICENSE file.
# This is the main configuration file for the application.
# ~~~~~
# Secret key
# ~~~~~
# The secret key is used to secure cryptographics functions.
# If you deploy your application to several instances be sure to use the same key!
play.crypto.secret="^ csmm5Fx4d=r2HEX8pelM3iBkFVv?k[mc;IZE _Qoq8EkX_/7@Zt6dP05Pzea3U"
play.crypto.secret=${?APPLICATION_SECRET}
# The application languages
# ~~~~~
play.i18n.langs=["en"]
play.http.requestHandler = "play.http.DefaultHttpRequestHandler"
play.http.cOntext= "/"
play.application.loader=loader.KafkaManagerLoader
kafka-manager.zkhosts="localhost:2181"
kafka-manager.zkhosts=${?ZK_HOSTS}
pinned-dispatcher.type="PinnedDispatcher"
pinned-dispatcher.executor="thread-pool-executor"
application.features=["KMClusterManagerFeature","KMTopicManagerFeature","KMPreferredReplicaElectionFeature","KMReassignPartitionsFeature"]
akka {
loggers = ["akka.event.slf4j.Slf4jLogger"]
loglevel = "INFO"
basicAuthentication.enabled=false
basicAuthentication.username="admin"
basicAuthentication.password="password"
basicAuthentication.realm="Kafka-Manager"
kafka-manager.consumer.properties.file=${?CONSUMER_PROPERTIES_FILE}

修改

我的是master、slave1和slave2,大家根据自己的机器情况对应进行修改即可。

kafka-manager.zkhosts="master:2181,slave1:2181,slave2:2181"

4、 运行kafka manager

注意:默认启动端口为9000。

要到大家kafka-manager的安装目录下来执行。

如我这里是在/home/hadoop/app/kafka-manager-1.3.2.1

bin/kafka-manager -Dconfig.file=conf/application.conf

或者

nohup bin/kafka-manager -Dconfig.file=conf/application.conf (后台运行)

当然,大家可以以这个端口为所用,大家也可以在启动的时候,开启另一个端口,比如我这里开启10000端口。

最好使用绝对路径。

要到大家kafka-manager的安装目录下来执行。

如我这里是在/home/hadoop/app/kafka-manager-1.3.2.1

nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=

5、打开浏览器,访问http://IP:10000

[hadoop@master kafka-manager-1.3.2.1]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1
[hadoop@master kafka-manager-1.3.2.1]$ ll
total
drwxrwxr-x hadoop hadoop Jul : bin
drwxrwxr-x hadoop hadoop Jul : conf
drwxrwxr-x hadoop hadoop Jul : lib
-rw-r--r-- hadoop hadoop Feb : README.md
drwxrwxr-x hadoop hadoop Jul : share
[hadoop@master kafka-manager-1.3.2.1]$ nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=
[]
[hadoop@master kafka-manager-1.3.2.1]$ nohup: ignoring input and appending output to `nohup.out'
[]+ Exit nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=
[hadoop@master kafka-manager-1.3.2.1]$

[hadoop@master kafka-manager-1.3.2.1]$ cat nohup.out
This application is already running (Or delete /home/hadoop/app/kafka-manager-1.3.2.1/RUNNING_PID file).
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback.groovy]
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback-test.xml]
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Found resource [logback.xml] at [file:/home/hadoop/app/kafka-manager-1.3.2.1/conf/logback.xml]
::, |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - debug attribute not set
::, |-INFO in ch.qos.logback.core.joran.action.ConversionRuleAction - registering conversion word coloredLevel with class [play.api.Logger$ColoredLevel]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [FILE]
::, |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
::, |-ERROR in ch.qos.logback.core.joran.spi.Interpreter@: - no applicable action for [totalSizeCap], current ElementPath is [[configuration][appender][rollingPolicy][totalSizeCap]]
::, |-INFO in c.q.l.core.rolling.TimeBasedRollingPolicy - No compression will be used
::, |-INFO in c.q.l.core.rolling.TimeBasedRollingPolicy - Will use the pattern application.home_IS_UNDEFINED/logs/application.%d{yyyy-MM-dd}.log for the active file
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - The date pattern is 'yyyy-MM-dd' from file name pattern 'application.home_IS_UNDEFINED/logs/application.%d{yyyy-MM-dd}.log'.
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - Roll-over at midnight.
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - Setting initial period to Fri Jul :: CST
::, |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[FILE] - Active log file name: application.home_IS_UNDEFINED/logs/application.log
::, |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[FILE] - File property is set to [application.home_IS_UNDEFINED/logs/application.log]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.ConsoleAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [STDOUT]
::, |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.classic.AsyncAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [ASYNCFILE]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [FILE] to ch.qos.logback.classic.AsyncAppender[ASYNCFILE]
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCFILE] - Attaching appender named [FILE] to AsyncAppender.
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCFILE] - Setting discardingThreshold to
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.classic.AsyncAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [ASYNCSTDOUT]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT]
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT] - Attaching appender named [STDOUT] to AsyncAppender.
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT] - Setting discardingThreshold to
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [play] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [application] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [kafka.manager] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebean.config.PropertyMapLoader] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebeaninternal.server.core.XmlConfigLoader] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebeaninternal.server.lib.BackgroundThread] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.gargoylesoftware.htmlunit.Javascript] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.zookeeper] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to WARN
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [ASYNCFILE] to Logger[ROOT]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [ASYNCSTDOUT] to Logger[ROOT]
::, |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - End of configuration.
::, |-INFO in ch.qos.logback.classic.joran.JoranConfigurator@18cf1e03 - Registering current configuration as safe fallback point
[warn] o.a.c.r.ExponentialBackoffRetry - maxRetries too large (). Pinning to
[info] k.m.a.KafkaManagerActor - Starting curator...
[info] o.a.z.ZooKeeper - Client environment:zookeeper.version=3.4.--, built on // : GMT
[info] o.a.z.ZooKeeper - Client environment:host.name=master
[info] o.a.z.ZooKeeper - Client environment:java.version=1.8.0_60
[info] o.a.z.ZooKeeper - Client environment:java.vendor=Oracle Corporation
[info] o.a.z.ZooKeeper - Client environment:java.home=/home/hadoop/app/jdk1..0_60/jre
[info] o.a.z.ZooKeeper - Client environment:java.class.path=/home/hadoop/app/kafka-manager-1.3.2.1/lib/../conf/:/home/hadoop/app/kafka-manager-1.3.2.1/lib/kafka-manager.kafka-manager-1.3.2.1-sans-externalized.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.scala-library-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.twirl-api_2.-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.commons.commons-lang3-3.4.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-server_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.build-link-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-exceptions-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.javassist.javassist-3.19.-GA.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-iteratees_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-stm.scala-stm_2.-0.7.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.config-1.3..jar:/home/hadoop/app/
kafka-manager-1.3.2.1/lib/com.typesafe.play.play-json_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-functional_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-datacommons_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/joda-time.joda-time-2.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.joda.joda-convert-1.7.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.datatype.jackson-datatype-jdk8-2.5..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.datatype.jackson-datatype-jsr310-2.5..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-netty-utils-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.jul-to-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.jcl-over-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/ch.qos.logback.logback-core-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/ch.qos.logback.logback-classic-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.akka.akka-actor_2.-2.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.akka.akka-slf4j_2.-2.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/commons-codec.commons-codec-1.10.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/xerces.xercesImpl-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/xml-apis.xml-apis-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/javax.transaction.jta-1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.inject.guice-4.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/javax.inject.javax.inject-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/aopalliance.aopalliance-1.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.guava.guava-16.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.inject.extensions.guice-assistedinject-4.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-netty-server_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/io.netty.netty-3.10..Final.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.netty.netty-http-pipelining-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.code.findbugs.jsr305-2.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-play_2.-2.4.-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.requirejs-2.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-locator-0.28.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-locator-core-0.27.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.commons.commons-compress-1.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.npm.validate.js-0.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.bootstrap-3.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.jquery-2.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.backbonejs-1.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.underscorejs-1.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.dustjs-linkedin-2.6.-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.json--.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-framework-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-client-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/jline.jline-0.9..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-recipes-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-jackson_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-core_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-ast_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-scalap_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.thoughtworks.paranamer.paranamer-2.8.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.modules.scala-xml_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-databind-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-annotations-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-core-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-scalaz_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scalaz.scalaz-core_2.-7.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.log4j-over-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.adrianhurt.play-bootstrap3_2.-0.4.-P24.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.clapper.grizzled-slf4j_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.kafka.kafka_2.-0.10.1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.kafka.kafka-clients-0.10.1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/net.jpountz.lz4.lz4-1.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.xerial.snappy.snappy-java-1.1.2.6.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.slf4j-api-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/net.sf.jopt-simple.jopt-simple-4.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.yammer.metrics.metrics-core-2.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.101tec.zkclient-0.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.zookeeper.zookeeper-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.modules.scala-parser-combinators_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.beachape.enumeratum_2.-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.beachape.enumera
tum-macros_2.-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.scala-reflect-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/kafka-manager.kafka-manager-1.3.2.1-assets.jar
[info] o.a.z.ZooKeeper - Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
[info] o.a.z.ZooKeeper - Client environment:java.io.tmpdir=/tmp
[info] o.a.z.ZooKeeper - Client environment:java.compiler= NA
[info] o.a.z.ZooKeeper - Client environment:os.name=Linux
[info] o.a.z.ZooKeeper - Client environment:os.arch=amd64
[info] o.a.z.ZooKeeper - Client environment:os.version=2.6.-.el6.x86_64
[info] o.a.z.ZooKeeper - Client environment:user.name=hadoop
[info] o.a.z.ZooKeeper - Client environment:user.home=/home/hadoop
[info] o.a.z.ZooKeeper - Client environment:user.dir=/home/hadoop/app/kafka-manager-1.3.2.1
[info] o.a.z.ZooKeeper - Initiating client connection, cOnnectString=master:2181,slave1:2181,slave2:2181 sessiOnTimeout= watcher=org.apache.curator.ConnectionState@1dc0b758
[info] o.a.z.ClientCnxn - Opening socket connection to server slave1/192.168.80.146:. Will not attempt to authenticate using SASL (unknown error)
[info] k.m.a.KafkaManagerActor - zk=master:,slave1:,slave2:
[info] k.m.a.KafkaManagerActor - baseZkPath=/kafka-manager
[info] o.a.z.ClientCnxn - Socket connection established to slave1/192.168.80.146:, initiating session
[info] o.a.z.ClientCnxn - Session establishment complete on server slave1/192.168.80.146:, sessiOnid= 0x25d88012f990002, negotiated timeout =
[info] k.m.a.DeleteClusterActor - Started actor akka://kafka-manager-system/user/kafka-manager/delete-cluster
[info] k.m.a.DeleteClusterActor - Starting delete clusters path cache...
[info] k.m.a.KafkaManagerActor - Started actor akka://kafka-manager-system/user/kafka-manager
[info] k.m.a.KafkaManagerActor - Starting delete clusters path cache...
[info] k.m.a.DeleteClusterActor - Adding kafka manager path cache listener...
[info] k.m.a.DeleteClusterActor - Scheduling updater for seconds
[info] k.m.a.KafkaManagerActor - Starting kafka manager path cache...
[info] k.m.a.KafkaManagerActor - Adding kafka manager path cache listener...
[info] k.m.a.KafkaManagerActor - Updating internal state...
[info] play.api.Play - Application started (Prod)
[info] p.c.s.NettyServer - Listening for HTTP on /0:0:0:0:0:0:0:0:10000
[info] k.m.a.KafkaManagerActor - Updating internal state...
[info] k.m.a.KafkaManagerActor - Updating internal state...

也许,大家在这一步启动的时候,出现如下错误

Kafka-manager的进程是

[hadoop@master ~]$ jps
ProdServerStart

比如kill掉它,则

然后,再次打开,即可

http://192.168.80.145:10000/

或者

http://master:10000/

欢迎大家,加入我的微信公众号:大数据躺过的坑        人工智能躺过的坑
 
 
 

同时,大家可以关注我的个人博客:

 http://www.cnblogs.com/zlslch/ 和   http://www.cnblogs.com/lchzls/   http://www.cnblogs.com/sunnyDream/   

 详情请见:http://www.cnblogs.com/zlslch/p/7473861.html

人生苦短,我愿分享。本公众号将秉持活到老学到老学习无休止的交流分享开源精神,汇聚于互联网和个人学习工作的精华干货知识,一切来于互联网,反馈回互联网。
目前研究领域:大数据、机器学习、深度学习、人工智能、数据挖掘、数据分析。 语言涉及:Java、Scala、Python、Shell、Linux等 。同时还涉及平常所使用的手机、电脑和互联网上的使用技巧、问题和实用软件。 只要你一直关注和呆在群里,每天必须有收获

对应本平台的讨论和答疑QQ群:大数据和人工智能躺过的坑(总群)(161156071) 


推荐阅读
  • Hadoop入门与核心组件详解
    本文详细介绍了Hadoop的基础知识及其核心组件,包括HDFS、MapReduce和YARN。通过本文,读者可以全面了解Hadoop的生态系统及应用场景。 ... [详细]
  • 本文详细分析了Hive在启动过程中遇到的权限拒绝错误,并提供了多种解决方案,包括调整文件权限、用户组设置以及环境变量配置等。 ... [详细]
  • Hadoop发行版本选择指南:技术解析与应用实践
    本文详细介绍了Hadoop的不同发行版本及其特点,帮助读者根据实际需求选择最合适的Hadoop版本。内容涵盖Apache Hadoop、Cloudera CDH等主流版本的特性及应用场景。 ... [详细]
  • 全面解析运维监控:白盒与黑盒监控及四大黄金指标
    本文深入探讨了白盒和黑盒监控的概念,以及它们在系统监控中的应用。通过详细分析基础监控和业务监控的不同采集方法,结合四个黄金指标的解读,帮助读者更好地理解和实施有效的监控策略。 ... [详细]
  • EasyMock实战指南
    本文介绍了如何使用EasyMock进行单元测试,特别是当测试对象的合作者依赖于外部资源或尚未实现时。通过具体的示例,展示了EasyMock在模拟对象行为方面的强大功能。 ... [详细]
  • window下kafka的安装以及测试
    目录一、安装JDK(需要安装依赖javaJDK)二、安装Kafka三、测试参考在Windows系统上安装消息队列kafka一、安装JDKÿ ... [详细]
  • 本文详细介绍了如何在Linux系统上安装和配置Smokeping,以实现对网络链路质量的实时监控。通过详细的步骤和必要的依赖包安装,确保用户能够顺利完成部署并优化其网络性能监控。 ... [详细]
  • 本文详细介绍了 Dockerfile 的编写方法及其在网络配置中的应用,涵盖基础指令、镜像构建与发布流程,并深入探讨了 Docker 的默认网络、容器互联及自定义网络的实现。 ... [详细]
  • 基于KVM的SRIOV直通配置及性能测试
    SRIOV介绍、VF直通配置,以及包转发率性能测试小慢哥的原创文章,欢迎转载目录?1.SRIOV介绍?2.环境说明?3.开启SRIOV?4.生成VF?5.VF ... [详细]
  • 深入探讨CPU虚拟化与KVM内存管理
    本文详细介绍了现代服务器架构中的CPU虚拟化技术,包括SMP、NUMA和MPP三种多处理器结构,并深入探讨了KVM的内存虚拟化机制。通过对比不同架构的特点和应用场景,帮助读者理解如何选择最适合的架构以优化性能。 ... [详细]
  • Netflix利用Druid实现高效实时数据分析
    本文探讨了全球领先的在线娱乐公司Netflix如何通过采用Apache Druid,实现了高效的数据采集、处理和实时分析,从而显著提升了用户体验和业务决策的准确性。文章详细介绍了Netflix在系统架构、数据摄取、管理和查询方面的实践,并展示了Druid在大规模数据处理中的卓越性能。 ... [详细]
  • 字节跳动夏季招聘面试经验分享
    本文详细记录了字节跳动夏季招聘的面试经历,涵盖了一、二、三轮面试的技术问题及项目讨论,旨在为准备类似面试的求职者提供参考。 ... [详细]
  • Kafka Topic 数据管理与清理策略
    本文探讨了在生产环境中如何有效管理和定期清理Kafka Topic中的数据。介绍了基于时间、日志大小和日志起始偏移量三种清除方式,并重点讲解了基于时间的清除策略及其配置方法。 ... [详细]
  • 深入理解Kafka架构
    本文将详细介绍Kafka的内部工作机制,包括其工作流程、文件存储机制、生产者与消费者的具体实现,以及如何通过高效读写技术和Zookeeper支持来确保系统的高性能和稳定性。 ... [详细]
  • 利用GitHub热门资源,成功斩获阿里、京东、腾讯三巨头Offer
    Spring框架作为Java生态系统中的重要组成部分,因其强大的功能和灵活的扩展性,被广泛应用于各种规模的企业级应用开发中。本文将通过一份在GitHub上获得极高评价的Spring全家桶文档,探讨如何掌握Spring框架及其相关技术,助力职业发展。 ... [详细]
author-avatar
miya的发现王国sGA_998
这个家伙很懒,什么也没留下!
PHP1.CN | 中国最专业的PHP中文社区 | DevBox开发工具箱 | json解析格式化 |PHP资讯 | PHP教程 | 数据库技术 | 服务器技术 | 前端开发技术 | PHP框架 | 开发工具 | 在线工具
Copyright © 1998 - 2020 PHP1.CN. All Rights Reserved | 京公网安备 11010802041100号 | 京ICP备19059560号-4 | PHP1.CN 第一PHP社区 版权所有