kubernetes 更新后,kubernetes 上的 mongodb StatefulSet 不再工作

标签 mongodb kubernetes azure-aks

我已将我的 AKS Azure Kubernetes 集群更新到版本 1.11.5,在此集群中运行 MongoDB Statefulset:

statefulset 是用这个文件创建的:

---
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
  name: default-view
roleRef:
  apiGroup: rbac.authorization.k8s.io
  kind: ClusterRole
  name: view
subjects:
  - kind: ServiceAccount
    name: default
    namespace: default
---
apiVersion: v1
kind: Service
metadata:
  name: mongo
  labels:
    name: mongo
spec:
  ports:
  - port: 27017
    targetPort: 27017
  clusterIP: None
  selector:
    role: mongo
---
apiVersion: apps/v1beta1
kind: StatefulSet
metadata:
  name: mongo
spec:
  serviceName: "mongo"
  replicas: 2
  template:
    metadata:
      labels:
        role: mongo
        environment: test
    spec:
      terminationGracePeriodSeconds: 10
      containers:
        - name: mongo
          image: mongo
          command:
            - mongod
            - "--replSet"
            - rs0
            - "--bind_ip"
            - 0.0.0.0            
            - "--smallfiles"
            - "--noprealloc"
          ports:
            - containerPort: 27017
          volumeMounts:
            - name: mongo-persistent-storage
              mountPath: /data/db
        - name: mongo-sidecar
          image: cvallance/mongo-k8s-sidecar
          env:
            - name: MONGO_SIDECAR_POD_LABELS
              value: "role=mongo,environment=test"
  volumeClaimTemplates:
  - metadata:
      name: mongo-persistent-storage
      annotations:
        volume.beta.kubernetes.io/storage-class: "managed-premium"
    spec:
      accessModes: [ "ReadWriteOnce" ]
      resources:
        requests:
          storage: 32Gi

在提到将集群更新到新的 k8s 版本后,我收到此错误:

mongo-0                        1/2     CrashLoopBackOff   6          9m
mongo-1                        2/2     Running            0          1h

来自 pod 的详细日志如下:

2018-12-18T14:28:44.281+0000 W STORAGE  [initandlisten] Detected configuration for non-active storage engine mmapv1 when current storage engine is wiredTiger
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten]
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten] ** WARNING: Access control is not enabled for the database.
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten] **          Read and write access to data and configuration is unrestricted.
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten] ** WARNING: You are running this process as the root user, which is not recommended.
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten]
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten]
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten] ** WARNING: /sys/kernel/mm/transparent_hugepage/enabled is 'always'.
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten] **        We suggest setting it to 'never'
2018-12-18T14:28:44.281+0000 I CONTROL  [initandlisten]
2018-12-18T14:28:44.477+0000 I FTDC     [initandlisten] Initializing full-time diagnostic data capture with directory '/data/db/diagnostic.data'
2018-12-18T14:28:44.478+0000 I REPL     [initandlisten] Rollback ID is 7
2018-12-18T14:28:44.479+0000 I REPL     [initandlisten] Recovering from stable timestamp: Timestamp(1545077719, 1) (top of oplog: { ts: Timestamp(1545077349, 1), t: 5 }, appliedThrough: { ts: Timestamp(1545077719, 1), t: 6 }, TruncateAfter: Timestamp(0, 0))
2018-12-18T14:28:44.480+0000 I REPL     [initandlisten] Starting recovery oplog application at the stable timestamp: Timestamp(1545077719, 1)
2018-12-18T14:28:44.480+0000 F REPL     [initandlisten] Applied op { : Timestamp(1545077719, 1) } not found. Top of oplog is { : Timestamp(1545077349, 1) }.
2018-12-18T14:28:44.480+0000 F -        [initandlisten] Fatal Assertion 40313 at src/mongo/db/repl/replication_recovery.cpp 361
2018-12-18T14:28:44.480+0000 F -        [initandlisten]

***aborting after fassert() failure

这两个实例似乎不同步并且无法恢复。有人可以帮忙吗?

最佳答案

我有解决这个问题的方法:

  1. 将 MongoDB 容器添加到集群以转储和恢复 MongoDB 数据
  2. 转储当前数据库
  3. 删除 MongoDB 实例
  4. 重新创建一个新的 MongoDB 实例
  5. 将数据恢复到新实例

是的,不幸的是,这会导致停机

关于kubernetes 更新后,kubernetes 上的 mongodb StatefulSet 不再工作,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/53835384/

相关文章:

node.js - 如何在 Mean Stack (Node.js + Mongodb) 上的用户身份验证中检查当前登录的用户?

MongoDB AddToSet 如何防止特定字段重复?

kubernetes - OpenShift 3.11中的运行状况检查修改

azure-aks - Azure Kubernetes服务上的垂直Pod自动缩放

ssl - 在 NGINX-Ingress 上使用 session 关联性 (Cookie) 和 SSL 直通

image - Tomcat 中未调用 Servlet

mongodb - 如何在 mongodb 中聚合查询

kubernetes - 无法在GKE(us-east1-d)上创建k8s 1.10集群

amazon-web-services - 停止并启动实例后,如何在AWS实例中重新启动minikube kubernetes集群?

azure - 自托管 Azure DevOps Pipeline Agent 失败并出现错误“ token 受众无效”