Open Issues Need Help
View All on GitHubAI Summary: The task is to modify the `automq-cli`'s cluster deployment functionality to accept `AWS_ACCESS_KEY_ID` and `AWS_SECRET_ACCESS_KEY` environment variables in addition to the existing `KAFKA_S3_ACCESS_KEY` and `KAFKA_S3_SECRET_KEY`. This will allow users to deploy AutoMQ clusters using standard AWS credentials.
AutoMQ is a stateless/diskless Kafka on S3. 10x Cost-Effective. No Cross-AZ Traffic Cost. Autoscale in seconds. Single-digit ms latency. Multi-AZ Availability.
AI Summary: The task involves assessing the feasibility and operational stability of a single AutoMQ cluster spanning three geographically diverse data centers (IDCs), each with its own Ceph storage cluster, while maintaining a unified metadata management through a Kraft cluster. The goal is to ensure seamless failover and continued operation in case of IDC failure, maintaining data writing and consumption without interruption.
AutoMQ is a stateless/diskless Kafka on S3. 10x Cost-Effective. No Cross-AZ Traffic Cost. Autoscale in seconds. Single-digit ms latency. Multi-AZ Availability.
AI Summary: The task is to fix a bug in the AutoMQ project where the `AwsObjectStorage#doList` method only lists the first 1000 objects when it should support listing more than 1000 objects using a prefix. This involves modifying the existing code to utilize the AWS S3 API's `ListObjectsV2` functionality correctly to handle pagination and retrieve all objects matching the prefix.
AutoMQ is a stateless/diskless Kafka on S3. 10x Cost-Effective. No Cross-AZ Traffic Cost. Autoscale in seconds. Single-digit ms latency. Multi-AZ Availability.