Best practice for reading data from Kafka to AWS Redshift
问题 What is the best practice for moving data from a Kafka cluster to a Redshift table? We have continuous data arriving on Kafka and I want to write it to tables in Redshift (it doesn't have to be in real time). Should I use Lambda function? Should I write a Redshift connector (consumer) that will run on a dedicated EC2 instance? (downside is that I need to handle redundancy) Is there some AWS pipeline service for that? 回答1: Kafka Connect is commonly used for streaming data from Kafka to (and