
- Two S3 storage buckets for gazette and druid data.
- Buckets can be configured to use AES256, KMS, or no encryption.
- An EKS cluster with a minimum of two node pools: base pool and druid pool.
-
The base node pool should be labeled with
arize=trueandarize-base=true. -
The druid node pool should be labeled with
arize=trueanddruid-historical=true. -
Storage classes
gp2is preferred and used by default. - An ECR or docker registry is optional as Arize pulls images from Arize AX’s central image registry by default.
-
Namespaces
arize,arize-operator, andarize-sparkcan be pre-existing or created later by the helm chart. -
If deployed on a private VPC, these endpoints must be accessible from the cluster:
- com.amazonaws..s3
- com.amazonaws..ecr.api
- com.amazonaws..ecr.dkr
- com.amazonaws..ec2
- com.amazonaws..elasticloadbalancing
- com.amazonaws..sts
- com.amazonaws..ebs
-
An IAM role with the following policy actions on the Arize druid and gazette buckets:
- s3:ListBucket
- s3:*Object
- kms:Encrypt
- kms:Decrypt
- kms:ReEncrypt*
- kms:GenerateDataKey*
- kms:DescribeKey
- bedrock:InvokeModel
-
If using IAM roles for service accounts (IRSA):
-
The roles must have a trust policy that allows these service accounts to assume the role:
- system:serviceaccount:arize:*
- system:serviceaccount:arize-spark:*
- system:serviceaccount:arize-operator:*
-
The roles must have a trust policy that allows these service accounts to assume the role:
-
If not using IAM roles for service accounts (IRSA):
- The policy actions should be added to the role attached to the nodes.
- Pods should be able to discover the node role through instance metadata.
small1b or medium2b.
values.yaml: