-
-
Notifications
You must be signed in to change notification settings - Fork 233
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
POD recovery, Can't Sync pods IP to nodes.conf #20
Comments
Can you please share the logs and screenshot, also are you using the latest version? |
the operator resources with master branch sourcecode. when i apply the redis.yaml(ref: example/redis-cluster-example.yaml), and the redis master&slave success; After delete the redis cluster with the redis.yaml, In a few minutes, i apply the redis.yaml again to create redis-cluster. redis.yaml as below: apiVersion: redis.opstreelabs.in/v1alpha1
kind: Redis
metadata:
name: redis
spec:
mode: cluster
size: 3
global:
image: opstree/redis:v2.0
imagePullPolicy: Always
password: "N1A8mhMAVqxx"
resources:
requests:
cpu: 100m
memory: 128Mi
limits:
cpu: 100m
memory: 128Mi
master:
service:
type: ClusterIP
slave:
service:
type: ClusterIP
redisExporter:
enabled: true
image: quay.io/opstree/redis-exporter:1.0
imagePullPolicy: Always
resources:
requests:
cpu: 100m
memory: 128Mi
limits:
cpu: 100m
memory: 128Mi
storage:
VolumeClaimTemplates:
spec:
accessModes:
- ReadWriteOnce
storageClassName: dev-ceph-block
resources:
requests:
storage: 500M
selector: {} |
i have same problem sometime. apiVersion: redis.opstreelabs.in/v1alpha1
kind: Redis
metadata:
name: redis
spec:
global:
image: 'quay.io/opstree/redis:v2.0'
imagePullPolicy: IfNotPresent
password: Opstree@12345
resources:
limits:
cpu: 100m
memory: 128Mi
requests:
cpu: 100m
memory: 128Mi
master:
service:
type: ClusterIP
mode: cluster
redisExporter:
enabled: true
image: 'quay.io/opstree/redis-exporter:1.0'
imagePullPolicy: Always
resources:
limits:
cpu: 100m
memory: 128Mi
requests:
cpu: 100m
memory: 128Mi
size: 3
slave:
service:
type: ClusterIP
storage:
volumeClaimTemplate:
selector: {}
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 1Gi
|
Experiencing the same problem. Stale entries in the nodes.conf is redirecting redis clients to IPs that no longer exist. I've got around it quickly by extending the redis image and amending the start_redis() command
A more elegant solution is referenced here redis/redis#4289 but was easier for me to amend the image than the operator. |
Seems this is partially fixed in github but not on operator hub |
Fixed in #26 |
/reopen |
This is still an issue for new clusters: old version of nodes.conf (i.e. after cluster re-build) are retained in PVC. Suggestion: add functionality to RedisCluster reconciler that will run CLUSTER RESET HARD on new nodes if cluster state is "Bootstrap" |
In my environment, this issue only occurs when all Redis nodes are recreated simultaneously, and they cannot distinguish themselves from one another. This does not happen during a rolling update. |
That's exactly what I'm talking about: when deploying new cluster and not getting all configuration right (ESPECIALLY SSL/TLS, these are notoriously error-prone), full cluster restart is required, and the fastest way to achieve that is to kill all running pods. Which leads directly to this problem. |
when one POD is fault, the POD IP(my) is invalid in the Persistence nodes.conf;
I suggest container start-up to refresh.
The text was updated successfully, but these errors were encountered: