I solved this problem by using Confluent's Kafka REST proxy image.
https://hub.docker.com/r/confluentinc/cp-kafka-rest/
Documentation of the REST Proxy is here:
http://docs.confluent.io/3.1.2/kafka-rest/docs/index.html
Step A: Build a Kafka broker docker image using latest Kafka version
I used a custom built Kafka broker image based on the same image you used. You basically just need to update cloudtrackinc's image to use Kafka version 0.10.1.0 or otherwise it won't work. Just update the Dockerfile from cloudertrackinc's image to use the latest wurstmeister kafka image and rebuild the docker image.
- FROM wurstmeister/kafka:0.10.1.0
I set the ADVERTISED_HOST_NAME for each Kafka broker to POD's IP so each broker gets an unique URL.
- name: ADVERTISED_HOST_NAME
valueFrom:
fieldRef:
fieldPath: status.podIP
Step B: Setup cp-kafka-rest proxy to use your Kafka broker cluster
Kafka Rest Proxy must be running within the same cluster as your Kafka broker cluster.
You need to provide two environment variables to the cp-kafka-rest image at the minimum for it to run. KAFKA_REST_HOST_NAME and KAFKA_REST_ZOOKEEPER_CONNECT. You can set KAFKA_REST_HOST_NAME to use POD's IP.
- name: KAFKA_REST_HOST_NAME
valueFrom:
fieldRef:
fieldPath: status.podIP
- name: KAFKA_REST_ZOOKEEPER_CONNECT
value: "zookeeper-svc-1:2181,zookeeper-svc-2:2181,zookeeper-svc-3:2181"
Step C: Expose the Kafka REST proxy as a service
spec:
type: NodePort or LoadBalancer
ports:
- name: kafka-rest-port
port: 8082
protocol: TCP
You can use NodePort or LoadBalancer to utilize single or multiple Kafka REST Proxy pods.
Pros and Cons of using Kafka REST proxy
Pros:
Cons:
So if you can live with the issues above, then give Kafka Rest Proxy a try.
This seems not to be possible at the moment, the network architecture of kafka is pretty poor regarding to this topic. The new consumer uses a list of brokers, which return the host of the zookeeper, but unfortunately this is in a different network, so it is not possible to reach it from your local client. The poor part of kafka is, that is not possible to specify the brokers AND the zookeeper servers. This prevents clients accessing the system from outside.
We worked around this for the moment using a busybox, where we installed tools to interact with kafka. In our case plunger
I had the same problem with accessing kafka from outside of k8s cluster on AWS. I manage to solve this issue by using kafka listeners feature which from version 0.10.2 supports multiple interfaces.
here is how I configured kafka container.
ports:
- containerPort: 9092
- containerPort: 9093
env:
- name: KAFKA_ZOOKEEPER_CONNECT
value: "zookeeper:2181"
- name: KAFKA_LISTENER_SECURITY_PROTOCOL_MAP
value: "INTERNAL_PLAINTEXT:PLAINTEXT,EXTERNAL_PLAINTEXT:PLAINTEXT"
- name: KAFKA_ADVERTISED_LISTENERS
value: "INTERNAL_PLAINTEXT://kafka-internal-service:9092,EXTERNAL_PLAINTEXT://123.us-east-2.elb.amazonaws.com:9093"
- name: KAFKA_LISTENERS
value: "INTERNAL_PLAINTEXT://0.0.0.0:9092,EXTERNAL_PLAINTEXT://0.0.0.0:9093"
- name: KAFKA_INTER_BROKER_LISTENER_NAME
value: "INTERNAL_PLAINTEXT"
Apart from that I configured two Services. One for internal(Headless) & one for external(LoadBalancer) communication.
Hopefully this will save people's time.
I was able to solve my problem by doing the following changes -
Using NodeSelector in YML to make kafka pod run on a particular node of kube cluster.
Set KAFKA_ADVERTISED_HOST_NAME
to Kube hostName where this Kafka POD has been configured to run on ( as configured in step 1 )
Expose Kafka Service using NodePort and set POD port same as that of exposed NodePort as shown below -
spec:
ports:
- name: broker-2
port: **30031**
targetPort: 9092
nodePort: **30031**
protocol: TCP
selector:
app: kafka-2
broker_id: "2"
type: NodePort
Now, you can access Kafka brokers from outside of kube cluster using host:exposedPort