Somehow I am unable to attach further links. So I attached logs on below links. Please have a look.
@hyagi
https://textdoc.co/LRfwZMhGkP1ySxcT
https://textdoc.co/hfbWc90QVN15lasP
Somehow I am unable to attach further links. So I attached logs on below links. Please have a look.
@hyagi
https://textdoc.co/LRfwZMhGkP1ySxcT
https://textdoc.co/hfbWc90QVN15lasP
hum … from the operator logs, the test-s3
Secret seems to be missing fields:
{"Secret.Namespace": "test", "Secret.Name": "test-s3", "error": "could not find "s3-access-key-id" key in test-s3 secret"}
when you tried to run
kubectl -ntest apply -f- <<EOF
apiVersion: v1
kind: Secret
metadata:
name: 'test-s3'
stringData:
s3-access-key-id: "<S3_ACCESS_KEY>"
s3-secret-access-key: "<S3_SECRET_ACCESS_KEY>"
s3-bucket-name: "<S3_BUCKET_NAME>"
s3-region: "<S3_REGION>"
EOF
did you get any errors?
you can double-check if the Secret has been created:
kubectl -n test get secrets test-s3
NAME TYPE DATA AGE
test-s3 Opaque 4 5h42m
and you can also check the content through:
kubectl -ntest get secrets test-s3 -ojsonpath='{.data}'|jq
{
"s3-access-key-id": "...",
"s3-bucket-name": "...",
"s3-region": "...",
"s3-secret-access-key": "..."
}
Great! I just updated the config with s3-secret-access-key and s3-access-key-id and it started working. Thank you so much @hyagi. Initially I did not add it as EKS & S3 are In AWS and I assumed that keys are not needed.
Now, do we have a way to use external PostgreSQL & Redis instead of using a containerised PostgreSQL & Redis?
Nice! Glad to know that it worked
Initially I did not add it as EKS & S3 are In AWS and I assumed that keys are not needed.
Unfortunately, we don’t have such a level of integration yet so we need to pass the credentials manually.
Now, do we have a way to use external PostgreSQL & Redis instead of using a containerised PostgreSQL & Redis?
Yes, we do. You can create a secret pointing to the external databases in this case.
kubectl -ntest create secret generic external-database \
--from-literal=POSTGRES_HOST=my-postgres-host.example.com \
--from-literal=POSTGRES_PORT=5432 \
--from-literal=POSTGRES_USERNAME=pulp-admin \
--from-literal=POSTGRES_PASSWORD=password \
--from-literal=POSTGRES_DB_NAME=pulp \
--from-literal=POSTGRES_SSLMODE=prefer
kubectl -ntest create secret generic external-redis \
--from-literal=REDIS_HOST=my-redis-host.example.com \
--from-literal=REDIS_PORT=6379 \
--from-literal=REDIS_PASSWORD="" \
--from-literal=REDIS_DB=""
kubectl -ntest patch pulp test --type=merge -p '{"spec": {"database": {"external_db_secret": "external-database"}, "cache": {"enabled": true, "external_cache_secret": "external-redis"} }}'
pulp-operator should notice the new settings and start to redeploy the pods.
After all the pods get into a READY state you can verify the status through:
kubectl -ntest exec deployment/test-api -- curl -sL localhost:24817/pulp/api/v3/status | jq '{"postgres": .database_connection, "redis": .redis_connection}'
{
"postgres": {
"connected": true
},
"redis": {
"connected": true
}
}
Here are the docs for more information:
https://docs.pulpproject.org/pulp_operator/configuring/database/#configuring-pulp-operator-to-use-an-external-postgresql-installation
https://docs.pulpproject.org/pulp_operator/configuring/cache/#configuring-pulp-operator-to-use-an-external-redis-installation
Awesome, surely I will give it a try. One last question I have about exposing the UI using a Kubernetes Load Balancer instead of listening on port localhost. Do we have a way for this as well?
Awesome, surely I will give it a try.
Cool! We would like to hear about your test results.
One last question I have about exposing the UI using a Kubernetes Load Balancer instead of listening on port localhost. Do we have a way for this as well?
Yes. Since this is an EKS cluster I can think of 3 ways to do this:
.spec.ingress_type: loadbalancer
the operator will create a k8s Service of type LoadBalancer
(the k8s integration with the cloud provider will create the lb for you)
.spec.ingress_type: ingress
the operator will deploy a k8s Ingress resource with the defined IngressController (more info in Reverse Proxy - Pulp Operator).
.spec.ingress_type: nodeport
the operator will create a k8s Service type NodePort
and you will need to manually create the AWS LoadBalancer to point to the k8s <nodes>:<node-port>
Definitely. I will keep you posted.
While I was going through my artifacts on S3, I got a question about protocol used to upload/download artifacts on S3 bucket. Does it use S3FS in backend. I am asking this because S3FS is slower than actual S3 api calls. S3fs does a mounting and uses for upload/download files. Can you please clarify on this as well.
Pulp is using the s3 apis via django-storages backend.
And if not otherwise configured, will redirect client access to the artifacts to s3 directly using presigned redirect urls.
Does this answer your question?
Yep, thank you so much.
To install Pulp on an AWS EKS cluster with S3 storage, you’ll need to follow these general steps:
Set up AWS EKS Cluster:
Create an EKS cluster using the AWS Management Console or AWS CLI.
Ensure your IAM roles and policies grant necessary permissions for EKS, S3, and other AWS services you'll be using.
Prepare S3 Bucket:
Create an S3 bucket to store your Pulp content.
Set appropriate permissions and access controls for the bucket.
Deploy Pulp on EKS:
Create Kubernetes manifests or Helm charts for deploying Pulp on your EKS cluster.
Configure Pulp to use S3 storage by providing the S3 bucket details in the configuration.
Configure Networking:
Set up network policies and ensure proper communication between your EKS cluster and the S3 bucket.
Test and Verify:
Deploy Pulp resources to your EKS cluster.
Verify that Pulp is running correctly and can access the S3 storage.
Monitor and Maintain:
Implement monitoring and logging for your Pulp deployment.
Regularly update and maintain your EKS cluster and Pulp installation.
Backup and Disaster Recovery:
Implement backup strategies for your Pulp content stored in S3.
Plan for disaster recovery scenarios and ensure data integrity.
Remember to refer to the official documentation of Pulp, AWS EKS, and S3 for detailed instructions specific to your setup and requirements.
Do we have a option to attach the service account (that will have access to s3) to the pod.
Do we have a option to attach the service account (that will have access to s3) to the pod.
Unfortunately, no, we don’t. Right now, we are configuring the pods with a ServiceAccount
based on the Pulp CR
name, and we don’t provide a way to modify the SA
for the pods.
As a workaround, it is possible to put the operator in an unmanaged state and manually config the Deployments
with the expected ServiceAccount/ServiceAccountName
.
Thanks for helping with the info. Will try it when the rest of the setup is sorted.
I was trying to for now configure pulp and check the flow of deb packages with s3 as backend storage (using the creds itself). I encountered the below error while accessing the content
File “/usr/local/lib/python3.9/site-packages/django/core/files/storage/base.py”, line 132, in path raise NotImplementedError(“This backend doesn’t support absolute paths.”)
Any clue on this error would help.
To have a better understanding of this issue, would you mind providing the following outputs in https://pastebin.centos.org?
kubectl get pulp -ojson
kubectl get pods -ojson
kubectl logs deployment/pulp-content
Thanks for your time. Somehow reinstalling pulp helped to solve the problem.
I faced another issue while installing the pulpoperator from git directly. The operator installation through helm work fine.
Have raised a separate issue as well Pulp operator directly from git isnt installing the rest of the pods. The same steps had worked a few days back though.
Any guidance on this would be helpful.
Currently, while configuring the S3 as only storage for the Pulp in eks, the only way to provide access for pods to s3 is through access keys specified in a k8 secret.
I did give a try with giving necessary permissions to a service account role to access s3, and attaching that service account role arn to the pods using spec.sa_annotations option in Pulp CR . But the pulp cr always expects spec.object_storage_s3_secret. Thus wasnt able to use only the service account option (without a s3 secret) with pulpcr. As mentioned earlier probably its not supported currently, my request is, could this support be added to the Pulp CR, so that access key as secret can be avoided. Especially when the pulp deployment is taken care using git, the storing of the access keys in git is slightly challenging, and cant be fully automated.
As mentioned earlier probably its not supported currently, my request is, could this support be added to the Pulp CR, so that access key as secret can be avoided.
Yes, it seems like a good idea to support S3 installation providing a SA instead of a Secret. Would you mind opening an issue in https://github.com/pulp/pulp-operator/issues/new/choose?
Thanks @hyagi for looking into this.
As suggested, have opened an issue https://github.com/pulp/pulp-operator/issues/1424 for the same.