Configure inferenceService to Access AWS Services from KServe
Configuration for accessing AWS services for inference services such as pulling images from private ECR and downloading models from S3 bucket.
Access AWS Service from Kserve with IAM Roles for ServiceAccount(IRSA)
-
Export env values:
export CLUSTER_NAME="<>" export CLUSTER_REGION="<>" export PROFILE_NAMESPACE=kubeflow-user-example-com export SERVICE_ACCOUNT_NAME=aws-sa # 123456789.dkr.ecr.us-west-2.amazonaws.com/kserve/sklearnserver:v0.8.0 export ECR_IMAGE_URL="<>" # s3://your-s3-bucket/model export S3_BUCKET_URL="<>"
-
Create Service Account with IAM Role using IRSA. The following command attaches both
AmazonEC2ContainerRegistryReadOnly
andAmazonS3ReadOnlyAccess
IAM policies:eksctl create iamserviceaccount --name ${SERVICE_ACCOUNT_NAME} --namespace ${PROFILE_NAMESPACE} --cluster ${CLUSTER_NAME} --region ${CLUSTER_REGION} --attach-policy-arn=arn:aws:iam::aws:policy/AmazonEC2ContainerRegistryReadOnly --attach-policy-arn=arn:aws:iam::aws:policy/AmazonS3ReadOnlyAccess --override-existing-serviceaccounts --approve
NOTE: You can use ECR (
AmazonEC2ContainerRegistryReadOnly
) and S3 (AmazonS3ReadOnlyAccess
) ReadOnly managed policies. We recommend creating fine grained policy for production usecase.
Deploy models from S3 Bucket
- Create Secret:
cat <<EOF > secret.yaml
apiVersion: v1
kind: Secret
metadata:
name: aws-secret
namespace: ${PROFILE_NAMESPACE}
annotations:
serving.kserve.io/s3-endpoint: s3.amazonaws.com
serving.kserve.io/s3-usehttps: "1"
serving.kserve.io/s3-region: ${CLUSTER_REGION}
type: Opaque
EOF
kubectl apply -f secret.yaml
- Attach secret to IRSA in your profile namespace:
kubectl patch serviceaccount ${SERVICE_ACCOUNT_NAME} -n ${PROFILE_NAMESPACE} -p '{"secrets": [{"name": "aws-secret"}]}'
Create an InferenceService
- Specify the service account in the model server spec :
NOTE: make sure you have workable image in
${ECR_IMAGE_URL}
and model in${S3_BUCKET_URL}
for the inferenceService to work. Versioning of model and image must be consistent: eg. you can not use a v1 model then a v2 image.
cat <<EOF > inferenceService.yaml
apiVersion: serving.kserve.io/v1beta1
kind: InferenceService
metadata:
name: "sklearn-iris"
namespace: ${PROFILE_NAMESPACE}
annotations:
sidecar.istio.io/inject: "false"
spec:
predictor:
serviceAccountName: ${SERVICE_ACCOUNT_NAME}
model:
modelFormat:
name: sklearn
image: ${ECR_IMAGE_URL}
storageUri: ${S3_BUCKET_URL}
EOF
kubectl apply -f inferenceService.yaml
- Check the InferenceService status:
kubectl get inferenceservices sklearn-iris -n ${PROFILE_NAMESPACE}
NAME URL READY PREV LATEST PREVROLLEDOUTREVISION LATESTREADYREVISION AGE
sklearn-iris http://sklearn-iris.kubeflow-user-example-com.example.com True 100 sklearn-iris-predictor-default-00001 105s
Last modified September 1, 2023: v1.7.0-aws-b1.0.3 website changes (#791) (7faf1a5)