热门标签 | HotTags
当前位置:  开发笔记 > 编程语言 > 正文

基于Web的Kafka管理器工具之Kafkamanager的编译部署详细安装(支持kafka0.8、0.9和0.10以后版本)(图文详解)(默认端口或任意自定义端口)

问题详情无奈于,在kafka里没有一个较好自带的webui。启动后无法观看,并且不友好。所以,需安装一个第三方的kafka管理工具为了简化开发者和服务工程师维护

问题详情

无奈于,在kafka里没有一个较好自带的web ui。启动后无法观看,并且不友好。所以,需安装一个第三方的kafka管理工具

为了简化开发者和服务工程师维护Kafka集群的工作,yahoo构建了一个叫做Kafka管理器的基于Web工具,叫做 Kafka Manager。这个管理工具可以很容易地发现分布在集群中的哪些topic分布不均匀,或者是分区在整个集群分布不均匀的的情况。

它支持管理多个集群、选择副本、副本重新分配以及创建Topic。同时,这个管理工具也是一个非常好的可以快速浏览这个集群的工具。

有如下功能:

管理多个kafka集群
便捷的检查kafka集群状态(topics,brokers,备份分布情况,分区分布情况)
选择你要运行的副本
基于当前分区状况进行
可以选择topic配置并创建topic(0.8.1.1和0.8.2的配置不同)
删除topic(只支持0.8.2以上的版本并且要在broker配置中设置delete.topic.enable=true)
Topic list会指明哪些topic被删除(在0.8.2以上版本适用)
为已存在的topic增加分区
为已存在的topic更新配置
在多个topic上批量重分区
在多个topic上批量重分区(可选partition broker位置)

大家编译的步骤,可以参考

或者

我这里不多说。大家去看看这个流程就好

直接采用下面这位博主已经编译好之后分享的。

谢谢他!

下载地址: 

  若此连接失效,则大家可以在我这篇博客下方留言评论,我将无偿发送给你们。

步骤:

1、解压kafka-manager-1.3.2.1.zip

lrwxrwxrwx. hadoop hadoop Apr : hadoop - hadoop-2.6.
drwxr-xr-x. hadoop hadoop Apr : hadoop-2.6.
lrwxrwxrwx. hadoop hadoop Apr : hbase - hbase-0.98.
drwxrwxr-x. hadoop hadoop Apr : hbase-0.98.
lrwxrwxrwx. hadoop hadoop Apr : hive - hive-1.0.
drwxrwxr-x. hadoop hadoop Jul : hive-1.0.
lrwxrwxrwx. hadoop hadoop Apr : jdk - jdk1..0_79
drwxr-xr-x. hadoop hadoop Apr jdk1..0_79
drwxr-xr-x. hadoop hadoop Aug jdk1..0_60
lrwxrwxrwx hadoop hadoop Jul : kafka - kafka_2.-0.9.0.1/
drwxr-xr-x hadoop hadoop Jul : kafka_2.-0.9.0.1
drwxr-xr-x hadoop hadoop May : kafka_2.-0.8.2.2
-rw-r--r-- hadoop hadoop Jul : kafka-manager-1.3.2.1.zip
lrwxrwxrwx hadoop hadoop Apr : kibana - kibana-4.6.-linux-x86_64/
drwxrwxr-x hadoop hadoop Nov kibana-4.6.-linux-x86_64
lrwxrwxrwx hadoop hadoop Jul : phoenix - apache-phoenix-4.10.-HBase-0.98-bin/
lrwxrwxrwx hadoop hadoop Jun : scala - scala-2.11./
drwxrwxr-x hadoop hadoop Feb scala-2.10.
drwxrwxr-x hadoop hadoop Mar scala-2.11.
lrwxrwxrwx hadoop hadoop May : snappy - snappy-1.1.
drwxr-xr-x hadoop hadoop May : snappy-1.1.
lrwxrwxrwx hadoop hadoop Jun : spark - spark-1.6.-bin-hadoop2./
drwxr-xr-x hadoop hadoop Feb spark-1.6.-bin-hadoop2.
lrwxrwxrwx. hadoop hadoop Apr : sqoop - sqoop-1.4.
drwxr-xr-x. hadoop hadoop May : sqoop-1.4.
lrwxrwxrwx hadoop hadoop May : storm - apache-storm-1.0./
lrwxrwxrwx hadoop hadoop Jun : zeppelin - zeppelin-0.5.-incubating-bin-all/
drwxr-xr-x hadoop hadoop Jun : zeppelin-0.5.-incubating-bin-all
lrwxrwxrwx. hadoop hadoop Apr : zookeeper - zookeeper-3.4.
drwxr-xr-x. hadoop hadoop Apr : zookeeper-3.4.
[hadoop@master app]$ unzip kafka-manager-1.3.2.1.zip

2、cd kafka-manager-1.3.2.1

lrwxrwxrwx. hadoop hadoop Apr : hive - hive-1.0.
drwxrwxr-x. hadoop hadoop Jul : hive-1.0.
lrwxrwxrwx. hadoop hadoop Apr : jdk - jdk1..0_79
drwxr-xr-x. hadoop hadoop Apr jdk1..0_79
drwxr-xr-x. hadoop hadoop Aug jdk1..0_60
lrwxrwxrwx hadoop hadoop Jul : kafka - kafka_2.-0.9.0.1/
drwxr-xr-x hadoop hadoop Jul : kafka_2.-0.9.0.1
drwxr-xr-x hadoop hadoop May : kafka_2.-0.8.2.2
drwxrwxr-x hadoop hadoop Jul : kafka-manager-1.3.2.1
-rw-r--r-- hadoop hadoop Jul : kafka-manager-1.3.2.1.zip
lrwxrwxrwx hadoop hadoop Apr : kibana - kibana-4.6.-linux-x86_64/
drwxrwxr-x hadoop hadoop Nov kibana-4.6.-linux-x86_64
lrwxrwxrwx hadoop hadoop Jul : phoenix - apache-phoenix-4.10.-HBase-0.98-bin/
lrwxrwxrwx hadoop hadoop Jun : scala - scala-2.11./
drwxrwxr-x hadoop hadoop Feb scala-2.10.
drwxrwxr-x hadoop hadoop Mar scala-2.11.
lrwxrwxrwx hadoop hadoop May : snappy - snappy-1.1.
drwxr-xr-x hadoop hadoop May : snappy-1.1.
lrwxrwxrwx hadoop hadoop Jun : spark - spark-1.6.-bin-hadoop2./
drwxr-xr-x hadoop hadoop Feb spark-1.6.-bin-hadoop2.
lrwxrwxrwx. hadoop hadoop Apr : sqoop - sqoop-1.4.
drwxr-xr-x. hadoop hadoop May : sqoop-1.4.
lrwxrwxrwx hadoop hadoop May : storm - apache-storm-1.0./
lrwxrwxrwx hadoop hadoop Jun : zeppelin - zeppelin-0.5.-incubating-bin-all/
drwxr-xr-x hadoop hadoop Jun : zeppelin-0.5.-incubating-bin-all
lrwxrwxrwx. hadoop hadoop Apr : zookeeper - zookeeper-3.4.
drwxr-xr-x. hadoop hadoop Apr : zookeeper-3.4.
[hadoop@master app]$ rm kafka-manager-1.3.2.1.zip
[hadoop@master app]$ cd kafka-manager-1.3.2.1/

3、 修改conf/application.conf文件,特别是kafka-manager.zkhosts的配置

[hadoop@master kafka-manager-1.3.2.1]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1
[hadoop@master kafka-manager-1.3.2.1]$ ll
total
drwxrwxr-x hadoop hadoop Jul : bin
drwxrwxr-x hadoop hadoop Jul : conf
drwxrwxr-x hadoop hadoop Jul : lib
-rw-r--r-- hadoop hadoop Feb : README.md
drwxrwxr-x hadoop hadoop Jul : share
[hadoop@master kafka-manager-1.3.2.1]$ cd conf/
[hadoop@master conf]$
[hadoop@master conf]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1/conf
[hadoop@master conf]$ ll
total
-rw-r--r-- hadoop hadoop Feb : application.conf
-rw-r--r-- hadoop hadoop Feb : consumer.properties
-rw-r--r-- hadoop hadoop Feb : logback.xml
-rw-r--r-- hadoop hadoop Feb : logger.xml
-rw-r--r-- hadoop hadoop Feb : routes
[hadoop@master conf]$ vim application.conf

以下是默认,我贴出来,大家学习学习

# Copyright Yahoo Inc. Licensed under the Apache License, Version 2.0
# See accompanying LICENSE file.
# This is the main configuration file for the application.
# ~~~~~
# Secret key
# ~~~~~
# The secret key is used to secure cryptographics functions.
# If you deploy your application to several instances be sure to use the same key!
play.crypto.secret="^ csmm5Fx4d=r2HEX8pelM3iBkFVv?k[mc;IZE _Qoq8EkX_/7@Zt6dP05Pzea3U"
play.crypto.secret=${?APPLICATION_SECRET}
# The application languages
# ~~~~~
play.i18n.langs=["en"]
play.http.requestHandler = "play.http.DefaultHttpRequestHandler"
play.http.cOntext= "/"
play.application.loader=loader.KafkaManagerLoader
kafka-manager.zkhosts="localhost:2181"
kafka-manager.zkhosts=${?ZK_HOSTS}
pinned-dispatcher.type="PinnedDispatcher"
pinned-dispatcher.executor="thread-pool-executor"
application.features=["KMClusterManagerFeature","KMTopicManagerFeature","KMPreferredReplicaElectionFeature","KMReassignPartitionsFeature"]
akka {
loggers = ["akka.event.slf4j.Slf4jLogger"]
loglevel = "INFO"
basicAuthentication.enabled=false
basicAuthentication.username="admin"
basicAuthentication.password="password"
basicAuthentication.realm="Kafka-Manager"
kafka-manager.consumer.properties.file=${?CONSUMER_PROPERTIES_FILE}

修改

我的是master、slave1和slave2,大家根据自己的机器情况对应进行修改即可。

kafka-manager.zkhosts="master:2181,slave1:2181,slave2:2181"

4、 运行kafka manager

注意:默认启动端口为9000。

要到大家kafka-manager的安装目录下来执行。

如我这里是在/home/hadoop/app/kafka-manager-1.3.2.1

bin/kafka-manager -Dconfig.file=conf/application.conf

或者

nohup bin/kafka-manager -Dconfig.file=conf/application.conf (后台运行)

当然,大家可以以这个端口为所用,大家也可以在启动的时候,开启另一个端口,比如我这里开启10000端口。

最好使用绝对路径。

要到大家kafka-manager的安装目录下来执行。

如我这里是在/home/hadoop/app/kafka-manager-1.3.2.1

nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=

5、打开浏览器,访问http://IP:10000

[hadoop@master kafka-manager-1.3.2.1]$ pwd
/home/hadoop/app/kafka-manager-1.3.2.1
[hadoop@master kafka-manager-1.3.2.1]$ ll
total
drwxrwxr-x hadoop hadoop Jul : bin
drwxrwxr-x hadoop hadoop Jul : conf
drwxrwxr-x hadoop hadoop Jul : lib
-rw-r--r-- hadoop hadoop Feb : README.md
drwxrwxr-x hadoop hadoop Jul : share
[hadoop@master kafka-manager-1.3.2.1]$ nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=
[]
[hadoop@master kafka-manager-1.3.2.1]$ nohup: ignoring input and appending output to `nohup.out'
[]+ Exit nohup bin/kafka-manager -Dconfig.file=/home/hadoop/app/kafka-manager-1.3.2.1/conf/application.conf -Dhttp.port=
[hadoop@master kafka-manager-1.3.2.1]$

[hadoop@master kafka-manager-1.3.2.1]$ cat nohup.out
This application is already running (Or delete /home/hadoop/app/kafka-manager-1.3.2.1/RUNNING_PID file).
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback.groovy]
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Could NOT find resource [logback-test.xml]
::, |-INFO in ch.qos.logback.classic.LoggerContext[default] - Found resource [logback.xml] at [file:/home/hadoop/app/kafka-manager-1.3.2.1/conf/logback.xml]
::, |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - debug attribute not set
::, |-INFO in ch.qos.logback.core.joran.action.ConversionRuleAction - registering conversion word coloredLevel with class [play.api.Logger$ColoredLevel]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.rolling.RollingFileAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [FILE]
::, |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
::, |-ERROR in ch.qos.logback.core.joran.spi.Interpreter@: - no applicable action for [totalSizeCap], current ElementPath is [[configuration][appender][rollingPolicy][totalSizeCap]]
::, |-INFO in c.q.l.core.rolling.TimeBasedRollingPolicy - No compression will be used
::, |-INFO in c.q.l.core.rolling.TimeBasedRollingPolicy - Will use the pattern application.home_IS_UNDEFINED/logs/application.%d{yyyy-MM-dd}.log for the active file
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - The date pattern is 'yyyy-MM-dd' from file name pattern 'application.home_IS_UNDEFINED/logs/application.%d{yyyy-MM-dd}.log'.
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - Roll-over at midnight.
::, |-INFO in c.q.l.core.rolling.DefaultTimeBasedFileNamingAndTriggeringPolicy - Setting initial period to Fri Jul :: CST
::, |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[FILE] - Active log file name: application.home_IS_UNDEFINED/logs/application.log
::, |-INFO in ch.qos.logback.core.rolling.RollingFileAppender[FILE] - File property is set to [application.home_IS_UNDEFINED/logs/application.log]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.core.ConsoleAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [STDOUT]
::, |-INFO in ch.qos.logback.core.joran.action.NestedComplexPropertyIA - Assuming default type [ch.qos.logback.classic.encoder.PatternLayoutEncoder] for [encoder] property
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.classic.AsyncAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [ASYNCFILE]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [FILE] to ch.qos.logback.classic.AsyncAppender[ASYNCFILE]
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCFILE] - Attaching appender named [FILE] to AsyncAppender.
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCFILE] - Setting discardingThreshold to
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - About to instantiate appender of type [ch.qos.logback.classic.AsyncAppender]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderAction - Naming appender as [ASYNCSTDOUT]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [STDOUT] to ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT]
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT] - Attaching appender named [STDOUT] to AsyncAppender.
::, |-INFO in ch.qos.logback.classic.AsyncAppender[ASYNCSTDOUT] - Setting discardingThreshold to
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [play] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [application] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [kafka.manager] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebean.config.PropertyMapLoader] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebeaninternal.server.core.XmlConfigLoader] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.avaje.ebeaninternal.server.lib.BackgroundThread] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [com.gargoylesoftware.htmlunit.Javascript] to OFF
::, |-INFO in ch.qos.logback.classic.joran.action.LoggerAction - Setting level of logger [org.apache.zookeeper] to INFO
::, |-INFO in ch.qos.logback.classic.joran.action.RootLoggerAction - Setting level of ROOT logger to WARN
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [ASYNCFILE] to Logger[ROOT]
::, |-INFO in ch.qos.logback.core.joran.action.AppenderRefAction - Attaching appender named [ASYNCSTDOUT] to Logger[ROOT]
::, |-INFO in ch.qos.logback.classic.joran.action.ConfigurationAction - End of configuration.
::, |-INFO in ch.qos.logback.classic.joran.JoranConfigurator@18cf1e03 - Registering current configuration as safe fallback point
[warn] o.a.c.r.ExponentialBackoffRetry - maxRetries too large (). Pinning to
[info] k.m.a.KafkaManagerActor - Starting curator...
[info] o.a.z.ZooKeeper - Client environment:zookeeper.version=3.4.--, built on // : GMT
[info] o.a.z.ZooKeeper - Client environment:host.name=master
[info] o.a.z.ZooKeeper - Client environment:java.version=1.8.0_60
[info] o.a.z.ZooKeeper - Client environment:java.vendor=Oracle Corporation
[info] o.a.z.ZooKeeper - Client environment:java.home=/home/hadoop/app/jdk1..0_60/jre
[info] o.a.z.ZooKeeper - Client environment:java.class.path=/home/hadoop/app/kafka-manager-1.3.2.1/lib/../conf/:/home/hadoop/app/kafka-manager-1.3.2.1/lib/kafka-manager.kafka-manager-1.3.2.1-sans-externalized.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.scala-library-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.twirl-api_2.-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.commons.commons-lang3-3.4.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-server_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.build-link-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-exceptions-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.javassist.javassist-3.19.-GA.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-iteratees_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-stm.scala-stm_2.-0.7.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.config-1.3..jar:/home/hadoop/app/
kafka-manager-1.3.2.1/lib/com.typesafe.play.play-json_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-functional_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-datacommons_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/joda-time.joda-time-2.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.joda.joda-convert-1.7.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.datatype.jackson-datatype-jdk8-2.5..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.datatype.jackson-datatype-jsr310-2.5..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-netty-utils-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.jul-to-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.jcl-over-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/ch.qos.logback.logback-core-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/ch.qos.logback.logback-classic-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.akka.akka-actor_2.-2.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.akka.akka-slf4j_2.-2.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/commons-codec.commons-codec-1.10.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/xerces.xercesImpl-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/xml-apis.xml-apis-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/javax.transaction.jta-1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.inject.guice-4.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/javax.inject.javax.inject-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/aopalliance.aopalliance-1.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.guava.guava-16.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.inject.extensions.guice-assistedinject-4.0.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.play.play-netty-server_2.-2.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/io.netty.netty-3.10..Final.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.typesafe.netty.netty-http-pipelining-1.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.google.code.findbugs.jsr305-2.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-play_2.-2.4.-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.requirejs-2.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-locator-0.28.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.webjars-locator-core-0.27.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.commons.commons-compress-1.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.npm.validate.js-0.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.bootstrap-3.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.jquery-2.1..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.backbonejs-1.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.underscorejs-1.8..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.dustjs-linkedin-2.6.-.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.webjars.json--.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-framework-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-client-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/jline.jline-0.9..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.curator.curator-recipes-2.10..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-jackson_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-core_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-ast_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-scalap_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.thoughtworks.paranamer.paranamer-2.8.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.modules.scala-xml_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-databind-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-annotations-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.fasterxml.jackson.core.jackson-core-2.6..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.json4s.json4s-scalaz_2.-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scalaz.scalaz-core_2.-7.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.log4j-over-slf4j-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.adrianhurt.play-bootstrap3_2.-0.4.-P24.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.clapper.grizzled-slf4j_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.kafka.kafka_2.-0.10.1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.kafka.kafka-clients-0.10.1.1.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/net.jpountz.lz4.lz4-1.3..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.xerial.snappy.snappy-java-1.1.2.6.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.slf4j.slf4j-api-1.7..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/net.sf.jopt-simple.jopt-simple-4.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.yammer.metrics.metrics-core-2.2..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.101tec.zkclient-0.9.jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.apache.zookeeper.zookeeper-3.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.modules.scala-parser-combinators_2.-1.0..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.beachape.enumeratum_2.-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/com.beachape.enumera
tum-macros_2.-1.4..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/org.scala-lang.scala-reflect-2.11..jar:/home/hadoop/app/kafka-manager-1.3.2.1/lib/kafka-manager.kafka-manager-1.3.2.1-assets.jar
[info] o.a.z.ZooKeeper - Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
[info] o.a.z.ZooKeeper - Client environment:java.io.tmpdir=/tmp
[info] o.a.z.ZooKeeper - Client environment:java.compiler= NA
[info] o.a.z.ZooKeeper - Client environment:os.name=Linux
[info] o.a.z.ZooKeeper - Client environment:os.arch=amd64
[info] o.a.z.ZooKeeper - Client environment:os.version=2.6.-.el6.x86_64
[info] o.a.z.ZooKeeper - Client environment:user.name=hadoop
[info] o.a.z.ZooKeeper - Client environment:user.home=/home/hadoop
[info] o.a.z.ZooKeeper - Client environment:user.dir=/home/hadoop/app/kafka-manager-1.3.2.1
[info] o.a.z.ZooKeeper - Initiating client connection, cOnnectString=master:2181,slave1:2181,slave2:2181 sessiOnTimeout= watcher=org.apache.curator.ConnectionState@1dc0b758
[info] o.a.z.ClientCnxn - Opening socket connection to server slave1/192.168.80.146:. Will not attempt to authenticate using SASL (unknown error)
[info] k.m.a.KafkaManagerActor - zk=master:,slave1:,slave2:
[info] k.m.a.KafkaManagerActor - baseZkPath=/kafka-manager
[info] o.a.z.ClientCnxn - Socket connection established to slave1/192.168.80.146:, initiating session
[info] o.a.z.ClientCnxn - Session establishment complete on server slave1/192.168.80.146:, sessiOnid= 0x25d88012f990002, negotiated timeout =
[info] k.m.a.DeleteClusterActor - Started actor akka://kafka-manager-system/user/kafka-manager/delete-cluster
[info] k.m.a.DeleteClusterActor - Starting delete clusters path cache...
[info] k.m.a.KafkaManagerActor - Started actor akka://kafka-manager-system/user/kafka-manager
[info] k.m.a.KafkaManagerActor - Starting delete clusters path cache...
[info] k.m.a.DeleteClusterActor - Adding kafka manager path cache listener...
[info] k.m.a.DeleteClusterActor - Scheduling updater for seconds
[info] k.m.a.KafkaManagerActor - Starting kafka manager path cache...
[info] k.m.a.KafkaManagerActor - Adding kafka manager path cache listener...
[info] k.m.a.KafkaManagerActor - Updating internal state...
[info] play.api.Play - Application started (Prod)
[info] p.c.s.NettyServer - Listening for HTTP on /0:0:0:0:0:0:0:0:10000
[info] k.m.a.KafkaManagerActor - Updating internal state...
[info] k.m.a.KafkaManagerActor - Updating internal state...

也许,大家在这一步启动的时候,出现如下错误

Kafka-manager的进程是

[hadoop@master ~]$ jps
ProdServerStart

比如kill掉它,则

然后,再次打开,即可

http://192.168.80.145:10000/

或者

http://master:10000/

欢迎大家,加入我的微信公众号:大数据躺过的坑        人工智能躺过的坑
 
 
 

同时,大家可以关注我的个人博客:

 http://www.cnblogs.com/zlslch/ 和   http://www.cnblogs.com/lchzls/   http://www.cnblogs.com/sunnyDream/   

 详情请见:http://www.cnblogs.com/zlslch/p/7473861.html

人生苦短,我愿分享。本公众号将秉持活到老学到老学习无休止的交流分享开源精神,汇聚于互联网和个人学习工作的精华干货知识,一切来于互联网,反馈回互联网。
目前研究领域:大数据、机器学习、深度学习、人工智能、数据挖掘、数据分析。 语言涉及:Java、Scala、Python、Shell、Linux等 。同时还涉及平常所使用的手机、电脑和互联网上的使用技巧、问题和实用软件。 只要你一直关注和呆在群里,每天必须有收获

对应本平台的讨论和答疑QQ群:大数据和人工智能躺过的坑(总群)(161156071) 


推荐阅读
  • Hadoop的分布式架构改进与应用
    nsitionalENhttp:www.w3.orgTRxhtml1DTDxhtml1-transitional.dtd ... [详细]
  • Hbase 的伪分布部署、shell基本操作及hbase相关理念
    1,HBase的的的的伪分布式配置-对zookeeper的配置,这个前面配置过,修改zoo.cfg文件,指定zookeeper的主入口-配置的HBase的的:进入optmo ... [详细]
  • 前期Linux环境准备1.修改Linux主机名2.修改IP3.修改主机名和IP的映射关系4.关闭防火墙5.ssh免登陆6.安装JDK,配置环境变量等集群规划主机 IP安装软件运行进 ... [详细]
  • 安装hadoop2.9.2jdk1.8centos7
    安装JDK1.8查看JDK1.8的安装https:www.cnblogs.comTJ21p13208514.html安装hadoop上传hadoop下载hadoop地址http:m ... [详细]
  • Ubuntu下Nginx启动、停止等常用命令
    nsitionalENhttp:www.w3.orgTRxhtml1DTDxhtml1-transitional.dtd ... [详细]
  • 转载:https:blog.csdn.nethigh2011articledetails70155431清华大学镜像网:https:mirrors.tu ... [详细]
  • Hadoop组件具有机架感知功能。例如,通过将一个块的分片放在不同的机架上,HDFS块放置将使用机架感知来实现容错。这可以在群集中发生网络切换故障或分区 ... [详细]
  • APMpinpoint(分布式集群环境下的性能监控)
    [官方参考](https:github.comnaverpinpointblobmasterdocinstallation.md”pinpoint官方文档”)中文文档参考环境要求: ... [详细]
  • nsitionalENhttp:www.w3.orgTRxhtml1DTDxhtml1-transitional.dtd ... [详细]
  • 一、Hadoop来历Hadoop的思想来源于Google在做搜索引擎的时候出现一个很大的问题就是这么多网页我如何才能以最快的速度来搜索到,由于这个问题Google发明 ... [详细]
  • 本文由编程笔记#小编为大家整理,主要介绍了StartingzookeeperFAILEDTOSTART相关的知识,希望对你有一定的参考价值。下载路径:https://ar ... [详细]
  • 本文介绍了在Linux下安装和配置Kafka的方法,包括安装JDK、下载和解压Kafka、配置Kafka的参数,以及配置Kafka的日志目录、服务器IP和日志存放路径等。同时还提供了单机配置部署的方法和zookeeper地址和端口的配置。通过实操成功的案例,帮助读者快速完成Kafka的安装和配置。 ... [详细]
  • 一次上线事故,30岁+的程序员踩坑经验之谈
    本文主要介绍了一位30岁+的程序员在一次上线事故中踩坑的经验之谈。文章提到了在双十一活动期间,作为一个在线医疗项目,他们进行了优惠折扣活动的升级改造。然而,在上线前的最后一天,由于大量数据请求,导致部分接口出现问题。作者通过部署两台opentsdb来解决问题,但读数据的opentsdb仍然经常假死。作者只能查询最近24小时的数据。这次事故给他带来了很多教训和经验。 ... [详细]
  • 本文总结了初学者在使用dubbo设计架构过程中遇到的问题,并提供了相应的解决方法。问题包括传输字节流限制、分布式事务、序列化、多点部署、zk端口冲突、服务失败请求3次机制以及启动时检查。通过解决这些问题,初学者能够更好地理解和应用dubbo设计架构。 ... [详细]
  • Hudi是一种数据湖的存储格式,在Hadoop文件系统之上提供了更新数据和删除数据的能力以及流式消费变化数据的能力。应用场景近实时数据摄取Hudi支持插入、更新和删除数据的能力。您 ... [详细]
author-avatar
miya的发现王国sGA_998
这个家伙很懒,什么也没留下!
PHP1.CN | 中国最专业的PHP中文社区 | DevBox开发工具箱 | json解析格式化 |PHP资讯 | PHP教程 | 数据库技术 | 服务器技术 | 前端开发技术 | PHP框架 | 开发工具 | 在线工具
Copyright © 1998 - 2020 PHP1.CN. All Rights Reserved | 京公网安备 11010802041100号 | 京ICP备19059560号-4 | PHP1.CN 第一PHP社区 版权所有