I wanted to see if I can connect Spring Cloud Stream Kafka with the help of docker-compose in Docker containers, but I’m stuck and I didn’t find a solution yet, please help me.
I’m working from Spring Microservices In Action; I didn’t find any help by now.
Docker-compose with Kafka and Zookeeper:
version: '2' services: zookeeper: image: wurstmeister/zookeeper container_name: zookeeper ports: - "2181:2181" kafkaserver: image: wurstmeister/kafka container_name: kafka ports: - "9092:9092" environment: - KAFKA_ADVERTISED_HOST_NAME=192.168.99.100 #kafka - KAFKA_ADVERTISED_PORT=9092 - KAFKA_ZOOKEEPER_CONNECT=zookeeper:2181 - KAFKA_CREATE_TOPICS=dresses:1:1, ratings:1:1 - KAFKA_SOCKET_REQUEST_MAX_BYTES=2000000000 - KAFKA_HEAP_OPTS=-Xmx512M -Xmx5g volumes: - "/var/run/docker.sock:/var/run/docker.sock" depends_on: - zookeeper
Docker-Compose with my Spring services:
version: '2' services: ... informatii: container_name: informatii build: C:UsersmariusIdeaProjectstototoinformatii-service restart: on-failure ports: - 1000:1000 environment: #SPRING_CLOUD_STREAM_KAFKA_BINDER_BROKERS: 192:168:99:100:9092 #SPRING_CLOUD_STREAM_KAFKA_BINDER_DEFAULTBROKERPORT: 9092 #KAFKA_ADVERTISED_LISTENERS=PLAINTEXT: //kafka:9092
App.properties for my services:
server.port=2000 spring.cloud.stream.bindings.output.destination=orgChangeTopic spring.cloud.stream.bindings.output.content-type=application/json spring.cloud.stream.kafka.binder.zkNodes=kafka # zookeeper:2181 #localhost>? zipkin spring.cloud.stream.kafka.binder.brokers=kafka spring.zipkin.baseUrl=http://zipkin:9411 -------- spring.cloud.stream.bindings.inboundOrgChanges.destination=orgChangeTopic spring.cloud.stream.bindings.inboundOrgChanges.content-type=application/json spring.cloud.stream.bindings.inboundOrgChanges.group=informatiiGroup spring.cloud.stream.kafka.binder.brokers=kafka spring.cloud.stream.kafka.binder.zkNodes=kafka spring.zipkin.baseUrl=http://zipkin:9411
Logs from docker containers:
kafka | waiting for kafka to be ready kafka | [Configuring] 'advertised.port' in '/opt/kafka/config/server.properties' kafka | Excluding KAFKA_HOME from broker config kafka | [Configuring] 'advertised.host.name' in '/opt/kafka/config/server.properties' kafka | [Configuring] 'port' in '/opt/kafka/config/server.properties' kafka | [Configuring] 'socket.request.max.bytes' in '/opt/kafka/config/server.properties' kafka | [Configuring] 'broker.id' in '/opt/kafka/config/server.properties' kafka | Excluding KAFKA_VERSION from broker config kafka | [Configuring] 'zookeeper.connect' in '/opt/kafka/config/server.properties' kafka | [Configuring] 'log.dirs' in '/opt/kafka/config/server.properties' kafka | waiting for kafka to be ready kafka | waiting for kafka to be ready zookeeper | ZooKeeper JMX enabled by default zookeeper | Using config: /opt/zookeeper-3.4.13/bin/../conf/zoo.cfg zookeeper | 2021-03-27 00:05:54,081 [myid:] - INFO [main:QuorumPeerConfig@136] - Reading configuration from: /opt/zookeeper-3.4.13/bin/../conf/zoo.cfg zookeeper | 2021-03-27 00:05:54,182 [myid:] - INFO [main:DatadirCleanupManager@78] - autopurge.snapRetainCount set to 3 zookeeper | 2021-03-27 00:05:54,183 [myid:] - INFO [main:DatadirCleanupManager@79] - autopurge.purgeInterval set to 1 zookeeper | 2021-03-27 00:05:54,190 [myid:] - WARN [main:QuorumPeerMain@116] - Either no config or no quorum defined in config, running in standalone mode zookeeper | 2021-03-27 00:05:54,196 [myid:] - INFO [PurgeTask:DatadirCleanupManager$PurgeTask@138] - Purge task started. zookeeper | 2021-03-27 00:05:54,350 [myid:] - INFO [main:QuorumPeerConfig@136] - Reading configuration from: /opt/zookeeper-3.4.13/bin/../conf/zoo.cfg zookeeper | 2021-03-27 00:05:54,366 [myid:] - INFO [main:ZooKeeperServerMain@98] - Starting server zookeeper | 2021-03-27 00:05:54,409 [myid:] - INFO [PurgeTask:DatadirCleanupManager$PurgeTask@144] - Purge task completed. kafka | waiting for kafka to be ready kafka | waiting for kafka to be ready kafka | [2021-03-27 00:06:03,740] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) kafka | waiting for kafka to be ready kafka | [2021-03-27 00:06:17,450] INFO Client environment:user.name=root (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,456] INFO Client environment:user.home=/root (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,459] INFO Client environment:user.dir=/ (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,464] INFO Client environment:os.memory.free=20MB (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,467] INFO Client environment:os.memory.max=5120MB (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,470] INFO Client environment:os.memory.total=32MB (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,534] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@333291e3 (org.apache.zookeeper.ZooKeeper) kafka | [2021-03-27 00:06:17,756] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) kafka | [2021-03-27 00:06:18,005] INFO zookeeper.request.timeout value is 0. feature enabled= (org.apache.zookeeper.ClientCnxn) kafka | [2021-03-27 00:06:18,053] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) kafka | [2021-03-27 00:06:18,182] INFO Opening socket connection to server zookeeper/172.24.0.3:2181. Will not attempt to authenticate using SASL (unknown error) (org.apache.zookeeper.ClientCnxn) zookeeper | 2021-03-27 00:05:55,247 [myid:] - INFO [main:NIOServerCnxnFactory@89] - binding to port 0.0.0.0/0.0.0.0:2181 zookeeper | 2021-03-27 00:06:18,709 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@215] - Accepted socket connection from /172.24.0.4:54606 zookeeper | 2021-03-27 00:06:18,796 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@949] - Client attempting to establish new session at /172.24.0.4:54606 zookeeper | 2021-03-27 00:06:18,945 [myid:] - INFO [SyncThread:0:FileTxnLog@213] - Creating new log file: log.1 kafka | [2021-03-27 00:06:18,683] INFO Socket connection established, initiating session, client: /172.24.0.4:54606, server: zookeeper/172.24.0.3:2181 (org.apache.zookeeper.ClientCnxn) kafka | [2021-03-27 00:06:19,190] INFO Session establishment complete on server zookeeper/172.24.0.3:2181, sessionid = 0x100033564dc0000, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) kafka | [2021-03-27 00:06:19,267] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) zookeeper | 2021-03-27 00:06:19,133 [myid:] - INFO [SyncThread:0:ZooKeeperServer@694] - Established session 0x100033564dc0000 with negotiated timeout 18000 for client /172.24.0.4:54606 zookeeper | 2021-03-27 00:06:21,262 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@653] - Got user-level KeeperException when processing sessionid:0x100033564dc0000 type:create cxid:0x2 zxid:0x3 txntype:-1 reqpath:n/a Error Path:/brokers Error:KeeperErrorCode = NoNode fo r /brokers zookeeper | 2021-03-27 00:06:21,326 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@653] - Got user-level KeeperException when processing sessionid:0x100033564dc0000 type:create cxid:0x6 zxid:0x7 txntype:-1 reqpath:n/a Error Path:/config Error:KeeperErrorCode = NoNode for /config kafka | waiting for kafka to be ready kafka | [2021-03-27 00:06:22,473] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) kafka | [2021-03-27 00:06:22,684] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) zookeeper | 2021-03-27 00:06:21,520 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@653] - Got user-level KeeperException when processing sessionid:0x100033564dc0000 type:create cxid:0x9 zxid:0xa txntype:-1 reqpath:n/a Error Path:/admin Error:KeeperErrorCode = NoNode for /admin kafka | [2021-03-27 00:06:27,109] INFO KafkaConfig values: kafka | advertised.host.name = 192.168.99.100 kafka | advertised.listeners = null kafka | advertised.port = 9092 ... kafka | listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL kafka | listeners = null ... kafka | (kafka.server.KafkaConfig) kafka | [2021-03-27 00:06:27,256] INFO KafkaConfig values: kafka | advertised.host.name = 192.168.99.100 kafka | advertised.listeners = null kafka | advertised.port = 9092 ... PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL kafka | listeners = null kafka | log.cleaner.backoff.ms = 15000 kafka | log.cleaner.dedupe.buffer.size = 134217728 kafka | log.cleaner.delete.retention.ms = 86400000 kafka | [2021-03-27 00:06:42,648] INFO Created ConnectionAcceptRate-PLAINTEXT sensor, quotaLimit=2147483647 (kafka.network.ConnectionQuotas) kafka | [2021-03-27 00:06:42,675] INFO Updated PLAINTEXT max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) kafka | [2021-03-27 00:06:42,709] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.Acceptor) kafka | [2021-03-27 00:06:47,696] INFO Creating /brokers/ids/1001 (is it secure? false) (kafka.zk.KafkaZkClient) kafka | [2021-03-27 00:06:48,288] INFO Stat of the created znode at /brokers/ids/1001 is: 25,25,1616803608051,1616803608051,1,0,0,72061121898217472,212,0,25 kafka | (kafka.zk.KafkaZkClient) kafka | [2021-03-27 00:06:48,317] INFO Registered broker 1001 at path /brokers/ids/1001 with addresses: PLAINTEXT://192.168.99.100:9092, czxid (broker epoch): 25 (kafka.zk.KafkaZkClient) kafka | [2021-03-27 00:06:50,511] INFO [ExpirationReaper-1001-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) kafka | creating topics: dresses:1:1 kafka | creating topics: ratings:1:1 kafka | [2021-03-27 00:06:52,268] INFO [GroupCoordinator 1001]: Starting up. (kafka.coordinator.group.GroupCoordinator) kafka | [2021-03-27 00:06:52,312] INFO [GroupCoordinator 1001]: Startup complete. (kafka.coordinator.group.GroupCoordinator) zookeeper | 2021-03-27 00:06:58,423 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@596] - Got user-level KeeperException when processing sessionid:0x100033564dc0000 type:multi cxid:0x3e zxid:0x1e txntype:-1 reqpath:n/a aborting remaining multi ops. Error Path:/admin/pref erred_replica_election Error:KeeperErrorCode = NoNode for /admin/preferred_replica_election zookeeper | 2021-03-27 00:07:18,457 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@215] - Accepted socket connection from /172.24.0.4:54612 zookeeper | 2021-03-27 00:07:18,492 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@215] - Accepted socket connection from /172.24.0.4:54614 zookeeper | 2021-03-27 00:07:18,521 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@949] - Client attempting to establish new session at /172.24.0.4:54612 zookeeper | 2021-03-27 00:07:18,534 [myid:] - INFO [SyncThread:0:ZooKeeperServer@694] - Established session 0x100033564dc0001 with negotiated timeout 30000 for client /172.24.0.4:54612 zookeeper | 2021-03-27 00:07:18,566 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@949] - Client attempting to establish new session at /172.24.0.4:54614 zookeeper | 2021-03-27 00:07:18,577 [myid:] - INFO [SyncThread:0:ZooKeeperServer@694] - Established session 0x100033564dc0002 with negotiated timeout 30000 for client /172.24.0.4:54614 zookeeper | 2021-03-27 00:07:25,621 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@653] - Got user-level KeeperException when processing sessionid:0x100033564dc0001 type:setData cxid:0x4 zxid:0x21 txntype:-1 reqpath:n/a Error Path:/config/topics/ratings Error:KeeperError Code = NoNode for /config/topics/ratings kafka | [2021-03-27 00:06:59,948] INFO [broker-1001-to-controller-send-thread]: Recorded new controller, from now on will use broker 1001 (kafka.server.BrokerToControllerRequestThread) zookeeper | 2021-03-27 00:07:25,917 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@653] - Got user-level KeeperException when processing sessionid:0x100033564dc0002 type:setData cxid:0x4 zxid:0x22 txntype:-1 reqpath:n/a Error Path:/config/topics/dresses Error:KeeperError Code = NoNode for /config/topics/dresses zookeeper | 2021-03-27 00:07:26,569 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@487] - Processed session termination for sessionid: 0x100033564dc0001 kafka | Created topic ratings. zookeeper | 2021-03-27 00:07:26,606 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1056] - Closed socket connection for client /172.24.0.4:54612 which had sessionid 0x100033564dc0001 zookeeper | 2021-03-27 00:07:28,347 [myid:] - INFO [ProcessThread(sid:0 cport:2181)::PrepRequestProcessor@487] - Processed session termination for sessionid: 0x100033564dc0002 kafka | Created topic dresses. kafka | [2021-03-27 00:07:31,500] INFO [ReplicaFetcherManager on broker 1001] Removed fetcher for partitions Set(ratings-0) (kafka.server.ReplicaFetcherManager) kafka | [2021-03-27 00:07:46,393] INFO [Log partition=ratings-0, dir=/kafka/kafka-logs-0d39e2b35aa6] Loading producer state till offset 0 with message format version 2 (kafka.log.Log) kafka | [2021-03-27 00:07:47,707] INFO Created log for partition ratings-0 in /kafka/kafka-logs-0d39e2b35aa6/ratings-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.7-IV2, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1048588, min.compaction.lag.ms -> 0, message.ti mestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segm ent.index.bytes -> 10485760}. (kafka.log.LogManager) kafka | [2021-03-27 00:07:47,836] INFO [Partition ratings-0 broker=1001] No checkpointed highwatermark is found for partition ratings-0 (kafka.cluster.Partition) kafka | [2021-03-27 00:07:47,932] INFO [Partition ratings-0 broker=1001] Log loaded for partition ratings-0 with initial high watermark 0 (kafka.cluster.Partition) kafka | [2021-03-27 00:07:53,030] INFO [ReplicaFetcherManager on broker 1001] Removed fetcher for partitions Set(dresses-0) (kafka.server.ReplicaFetcherManager) kafka | [2021-03-27 00:07:53,591] INFO [Log partition=dresses-0, dir=/kafka/kafka-logs-0d39e2b35aa6] Loading producer state till offset 0 with message format version 2 (kafka.log.Log) kafka | [2021-03-27 00:07:53,864] INFO Created log for partition dresses-0 in /kafka/kafka-logs-0d39e2b35aa6/dresses-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [delete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.7-IV2, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1048588, min.compaction.lag.ms -> 0, message.ti mestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 9223372036854775807, segm ent.index.bytes -> 10485760}. (kafka.log.LogManager) kafka | [2021-03-27 00:07:54,058] INFO [Partition dresses-0 broker=1001] No checkpointed highwatermark is found for partition dresses-0 (kafka.cluster.Partition) zookeeper | 2021-03-27 00:07:28,394 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1056] - Closed socket connection for client /172.24.0.4:54614 which had sessionid 0x100033564dc0002 zookeeper | 2021-03-27 00:08:11,558 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@215] - Accepted socket connection from /192.168.99.1:56469 zookeeper | 2021-03-27 00:08:11,574 [myid:] - WARN [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@383] - Exception causing close of session 0x0: Len error 1195725856 kafka | [2021-03-27 00:07:54,097] INFO [Partition dresses-0 broker=1001] Log loaded for partition dresses-0 with initial high watermark 0 (kafka.cluster.Partition) zookeeper | 2021-03-27 00:08:11,578 [myid:] - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1056] - Closed socket connection for client /192.168.99.1:56469 (no session established for client) kafka | [2021-03-27 00:24:07,405] INFO Creating topic orgChangeTopic with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(1001)) (kafka.zk.AdminZkClient) kafka | [2021-03-27 00:24:08,389] INFO [ReplicaFetcherManager on broker 1001] Removed fetcher for partitions Set(orgChangeTopic-0) (kafka.server.ReplicaFetcherManager) kafka | [2021-03-27 00:24:08,494] INFO [Log partition=orgChangeTopic-0, dir=/kafka/kafka-logs-0d39e2b35aa6] Loading producer state till offset 0 with message format version 2 (kafka.log.Log) kafka | [2021-03-27 00:24:08,584] INFO Created log for partition orgChangeTopic-0 in /kafka/kafka-logs-0d39e2b35aa6/orgChangeTopic-0 with properties {compression.type -> producer, message.downconversion.enable -> true, min.insync.replicas -> 1, segment.jitter.ms -> 0, cleanup.policy -> [del ete], flush.ms -> 9223372036854775807, segment.bytes -> 1073741824, retention.ms -> 604800000, flush.messages -> 9223372036854775807, message.format.version -> 2.7-IV2, file.delete.delay.ms -> 60000, max.compaction.lag.ms -> 9223372036854775807, max.message.bytes -> 1048588, min.compaction.lag.ms -> 0, message.timestamp.type -> CreateTime, preallocate -> false, min.cleanable.dirty.ratio -> 0.5, index.interval.bytes -> 4096, unclean.leader.election.enable -> false, retention.bytes -> -1, delete.retention.ms -> 86400000, segment.ms -> 604800000, message.timestamp.difference.max.ms -> 92233720368 54775807, segment.index.bytes -> 10485760}. (kafka.log.LogManager) kafka | [2021-03-27 00:24:08,617] INFO [Partition orgChangeTopic-0 broker=1001] No checkpointed highwatermark is found for partition orgChangeTopic-0 (kafka.cluster.Partition)
Docker Logs for my services…:
informatii | 2021-03-27 00:23:29.604 INFO 1 --- [ main] o.s.s.c.ThreadPoolTaskScheduler : Initializing ExecutorService 'taskScheduler' informatii | 2021-03-27 00:23:31.037 INFO 1 --- [ main] onConfiguration$FunctionBindingRegistrar : Functional binding is disabled due to the presense of @EnableBinding annotation in your configuration informatii | 2021-03-27 00:23:39.353 INFO 1 --- [ main] o.s.c.s.m.DirectWithAttributesChannel : Channel 'informatii-service-1.input' has 1 subscriber(s). informatii | 2021-03-27 00:23:39.448 INFO 1 --- [ main] o.s.i.endpoint.EventDrivenConsumer : Adding {logging-channel-adapter:_org.springframework.integration.errorLogger} as a subscriber to the 'errorChannel' channel informatii | 2021-03-27 00:23:39.457 INFO 1 --- [ main] o.s.i.channel.PublishSubscribeChannel : Channel 'informatii-service-1.errorChannel' has 1 subscriber(s). informatii | 2021-03-27 00:23:39.462 INFO 1 --- [ main] o.s.i.endpoint.EventDrivenConsumer : started bean '_org.springframework.integration.errorLogger' informatii | 2021-03-27 00:24:20.088 INFO 1 --- [ main] o.s.c.s.binder.DefaultBinderFactory : Creating binder: kafka divizie | 2021-03-27 00:24:22.519 INFO 1 --- [ main] c.n.d.provider.DiscoveryJerseyProvider : Using JSON decoding codec LegacyJacksonJson divizie | 2021-03-27 00:24:29.323 INFO 1 --- [ main] c.n.d.provider.DiscoveryJerseyProvider : Using XML encoding codec XStreamXml informatii | 2021-03-27 00:24:27.450 INFO 1 --- [ main] b.c.PropertySourceBootstrapConfiguration : Located property source: [BootstrapPropertySource {name='bootstrapProperties-configClient'}, BootstrapPropertySource {name='bootstrapProperties-https://github. nformatii-service.yml'}, BootstrapPropertySource {name='bootstrapProperties-https://github.com/hideyourname/kubes.git/application.yml'}] informatii | 2021-03-27 00:24:37.060 INFO 1 --- [ main] o.s.c.s.binder.DefaultBinderFactory : Caching the binder: kafka informatii | 2021-03-27 00:24:37.079 INFO 1 --- [ main] o.s.c.s.binder.DefaultBinderFactory : Retrieving cached binder: kafka informatii | 2021-03-27 00:44:45.335 WARN 1 --- [ask-scheduler-3] org.apache.kafka.clients.ClientUtils : Couldn't resolve server 192:168:99:100:9092 from bootstrap.servers as DNS resolution failed for 192:168:99:100 gateway | 2021-03-27 00:40:00.105 INFO 1 --- [trap-executor-0] c.n.d.s.r.aws.ConfigClusterResolver : Resolving eureka endpoints via configuration server | 2021-03-27 00:44:42.136 INFO 1 --- [a-EvictionTimer] c.n.e.registry.AbstractInstanceRegistry : Running the evict task with compensationTime 0ms informatii | 2021-03-27 00:44:45.343 ERROR 1 --- [ask-scheduler-3] o.s.cloud.stream.binding.BindingService : Failed to create consumer binding; retrying in 30 seconds informatii | informatii | org.springframework.cloud.stream.binder.BinderException: Exception thrown while starting consumer: informatii | at org.springframework.cloud.stream.binder.AbstractMessageChannelBinder.doBindConsumer(AbstractMessageChannelBinder.java:462) ~[spring-cloud-stream-3.0.11.RELEASE.jar!/:3.0.11.RELEASE] informatii | at org.springframework.cloud.stream.binder.AbstractMessageChannelBinder.doBindConsumer(AbstractMessageChannelBinder.java:91) ~[spring-cloud-stream-3.0.11.RELEASE.jar!/:3.0.11.RELEASE] informatii | at org.springframework.cloud.stream.binder.AbstractBinder.bindConsumer(AbstractBinder.java:143) ~[spring-cloud-stream-3.0.11.RELEASE.jar!/:3.0.11.RELEASE] informatii | at org.springframework.cloud.stream.binding.BindingService.lambda$rescheduleConsumerBinding$1(BindingService.java:201) ~[spring-cloud-stream-3.0.11.RELEASE.jar!/:3.0.11.RELEASE] informatii | at org.springframework.scheduling.support.DelegatingErrorHandlingRunnable.run(DelegatingErrorHandlingRunnable.java:54) ~[spring-context-5.2.13.RELEASE.jar!/:5.2.13.RELEASE] informatii | at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) ~[na:na] informatii | at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[na:na] informatii | at java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[na:na] informatii | at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) ~[na:na] informatii | at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) ~[na:na] informatii | at java.base/java.lang.Thread.run(Thread.java:835) ~[na:na] informatii | Caused by: org.apache.kafka.common.KafkaException: Failed to create new KafkaAdminClient informatii | at org.apache.kafka.clients.admin.KafkaAdminClient.createInternal(KafkaAdminClient.java:479) ~[kafka-clients-2.5.1.jar!/:na] informatii | at org.apache.kafka.clients.admin.Admin.create(Admin.java:71) ~[kafka-clients-2.5.1.jar!/:na] informatii | at org.apache.kafka.clients.admin.AdminClient.create(AdminClient.java:49) ~[kafka-clients-2.5.1.jar!/:na] informatii | at org.springframework.cloud.stream.binder.kafka.provisioning.KafkaTopicProvisioner.createAdminClient(KafkaTopicProvisioner.java:259) ~[spring-cloud-stream-binder-kafka-core-3.0.11.RELEASE.jar!/:3.0.11. informatii | at org.springframework.cloud.stream.binder.kafka.provisioning.KafkaTopicProvisioner.doProvisionConsumerDestination(KafkaTopicProvisioner.java:229) ~[spring-cloud-stream-binder-kafka-core-3.0.11.RELEASE. informatii | at org.springframework.cloud.stream.binder.kafka.provisioning.KafkaTopicProvisioner.provisionConsumerDestination(KafkaTopicProvisioner.java:196) ~[spring-cloud-stream-binder-kafka-core-3.0.11.RELEASE.ja informatii | at org.springframework.cloud.stream.binder.kafka.provisioning.KafkaTopicProvisioner.provisionConsumerDestination(KafkaTopicProvisioner.java:86) ~[spring-cloud-stream-binder-kafka-core-3.0.11.RELEASE.jar informatii | at org.springframework.cloud.stream.binder.AbstractMessageChannelBinder.doBindConsumer(AbstractMessageChannelBinder.java:403) ~[spring-cloud-stream-3.0.11.RELEASE.jar!/:3.0.11.RELEASE] informatii | ... 10 common frames omitted informatii | Caused by: org.apache.kafka.common.config.ConfigException: No resolvable bootstrap urls given in bootstrap.servers informatii | at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:89) ~[kafka-clients-2.5.1.jar!/:na] informatii | at org.apache.kafka.clients.ClientUtils.parseAndValidateAddresses(ClientUtils.java:48) ~[kafka-clients-2.5.1.jar!/:na] informatii | at org.apache.kafka.clients.admin.KafkaAdminClient.createInternal(KafkaAdminClient.java:439) ~[kafka-clients-2.5.1.jar!/:na] informatii | ... 17 common frames omitted informatii | informatii | 2021-03-27 00:45:15.354 INFO 1 --- [ask-scheduler-6] o.a.k.clients.admin.AdminClientConfig : AdminClientConfig values: informatii | bootstrap.servers = [192:168:99:100:9092] informatii | client.dns.lookup = default informatii | client.id = ... informatii | 2021-03-27 00:45:15.364 WARN 1 --- [ask-scheduler-6] org.apache.kafka.clients.ClientUtils : Couldn't resolve server 192:168:99:100:9092 from bootstrap.servers as DNS resolution failed for 192:168:99:100
Advertisement
Answer
You need to change back
KAFKA_ADVERTISED_HOST_NAME=kafka
Then, use same compose file for your Zookeeper, Kafka, and your app to ensure they are in the same network.
Also, the zkNodes
addresses should point to Zookeeper container, if you plan on using that property.