写点什么

ceph-csi 源码分析(4)-rbd driver-controllerserver 分析

用户头像
良凯尔
关注
发布于: 2021 年 04 月 30 日
ceph-csi源码分析(4)-rbd driver-controllerserver分析

kubernetes ceph-csi 分析 - 目录导航:

https://xie.infoq.cn/article/4b1d3e32f124307a49cd9c1e3


当 ceph-csi 组件启动时指定的 driver type 为 rbd 时,会启动 rbd driver 相关的服务。然后再根据controllerservernodeserver的参数配置,决定启动ControllerServerIdentityServer,或NodeServerIdentityServer


基于 tag v3.0.0


https://github.com/ceph/ceph-csi/releases/tag/v3.0.0



rbd driver 分析将分为 4 个部分,分别是服务入口分析、controllerserver 分析、nodeserver 分析与 IdentityServer 分析。


这节进行 controllerserver 分析,controllerserver 主要包括了 CreateVolume(创建存储)、DeleteVolume(删除存储)、ControllerExpandVolume(存储扩容)、CreateSnapshot(创建存储快照)、DeleteSnapshot(删除存储快照)操作,这里主要对 CreateVolume(创建存储)、DeleteVolume(删除存储)、ControllerExpandVolume(存储扩容)进行分析。

controllerserver 分析

(1)CreateVolume

简介

调用 ceph 存储后端,创建存储(rbd image)。


CreateVolume creates the volume in backend.


大致步骤:


(1)检查该 driver 是否支持存储的创建,不支持则直接返回错误;


(2)构建 ceph 请求凭证,生成 volume ID;


(3)调用 ceph 存储后端,创建存储。


三种创建来源:


(1)从已有 image 拷贝;


(2)根据快照创建 image;


(3)直接创建新的 image。

CreateVolume

主体流程:


(1)检查该 driver 是否支持存储的创建,不支持则直接返回错误;


(2)根据 secret 构建 ceph 请求凭证(secret 由 external-provisioner 组件传入);


(3)将请求参数转换为 rbdVol 结构体;


(4)创建与 Ceph 集群的连接;


(5)生成 RbdImageName 与 ReservedID,并生成 volume ID;


(6)调用 createBackingImage 去创建 image。


(暂时跳过从已有 image 拷贝与根据快照创建 image 的分析)


//ceph-csi/internal/rbd/controllerserver.go
func (cs *ControllerServer) CreateVolume(ctx context.Context, req *csi.CreateVolumeRequest) (*csi.CreateVolumeResponse, error) { // (1)校验请求参数; if err := cs.validateVolumeReq(ctx, req); err != nil { return nil, err }
// (2)根据secret构建ceph请求凭证; cr, err := util.NewUserCredentials(req.GetSecrets()) if err != nil { return nil, status.Error(codes.Internal, err.Error()) } defer cr.DeleteCredentials() // (3)将请求参数转换为rbdVol结构体; rbdVol, err := cs.parseVolCreateRequest(ctx, req) if err != nil { return nil, err } defer rbdVol.Destroy() // Existence and conflict checks if acquired := cs.VolumeLocks.TryAcquire(req.GetName()); !acquired { klog.Errorf(util.Log(ctx, util.VolumeOperationAlreadyExistsFmt), req.GetName()) return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, req.GetName()) } defer cs.VolumeLocks.Release(req.GetName()) // (4)创建与Ceph集群的连接; err = rbdVol.Connect(cr) if err != nil { klog.Errorf(util.Log(ctx, "failed to connect to volume %v: %v"), rbdVol.RbdImageName, err) return nil, status.Error(codes.Internal, err.Error()) }
parentVol, rbdSnap, err := checkContentSource(ctx, req, cr) if err != nil { return nil, err }
found, err := rbdVol.Exists(ctx, parentVol) if err != nil { return nil, getGRPCErrorForCreateVolume(err) } if found { if rbdSnap != nil { // check if image depth is reached limit and requires flatten err = checkFlatten(ctx, rbdVol, cr) if err != nil { return nil, err } } return buildCreateVolumeResponse(ctx, req, rbdVol) }
err = validateRequestedVolumeSize(rbdVol, parentVol, rbdSnap, cr) if err != nil { return nil, err }
err = flattenParentImage(ctx, parentVol, cr) if err != nil { return nil, err } // (5)生成rbd image name并生成volume ID; err = reserveVol(ctx, rbdVol, rbdSnap, cr) if err != nil { return nil, status.Error(codes.Internal, err.Error()) } defer func() { if err != nil { if !errors.Is(err, ErrFlattenInProgress) { errDefer := undoVolReservation(ctx, rbdVol, cr) if errDefer != nil { klog.Warningf(util.Log(ctx, "failed undoing reservation of volume: %s (%s)"), req.GetName(), errDefer) } } } }() // (6)调用createBackingImage去创建image。 err = cs.createBackingImage(ctx, cr, rbdVol, parentVol, rbdSnap) if err != nil { if errors.Is(err, ErrFlattenInProgress) { return nil, status.Error(codes.Aborted, err.Error()) } return nil, err }
volumeContext := req.GetParameters() volumeContext["pool"] = rbdVol.Pool volumeContext["journalPool"] = rbdVol.JournalPool volumeContext["imageName"] = rbdVol.RbdImageName volume := &csi.Volume{ VolumeId: rbdVol.VolID, CapacityBytes: rbdVol.VolSize, VolumeContext: volumeContext, ContentSource: req.GetVolumeContentSource(), } if rbdVol.Topology != nil { volume.AccessibleTopology = []*csi.Topology{ { Segments: rbdVol.Topology, }, } } return &csi.CreateVolumeResponse{Volume: volume}, nil}
复制代码


1 reserveVol


用于生成 RbdImageName 与 ReservedID,并生成 volume ID。调用 j.ReserveName 获取 imageName,调用 util.GenerateVolID 生成 volume ID。


//ceph-csi/internal/rbd/rbd_journal.go
// reserveVol is a helper routine to request a rbdVolume name reservation and generate the// volume ID for the generated name.func reserveVol(ctx context.Context, rbdVol *rbdVolume, rbdSnap *rbdSnapshot, cr *util.Credentials) error { var ( err error )
err = updateTopologyConstraints(rbdVol, rbdSnap) if err != nil { return err }
journalPoolID, imagePoolID, err := util.GetPoolIDs(ctx, rbdVol.Monitors, rbdVol.JournalPool, rbdVol.Pool, cr) if err != nil { return err }
kmsID := "" if rbdVol.Encrypted { kmsID = rbdVol.KMS.GetID() }
j, err := volJournal.Connect(rbdVol.Monitors, cr) if err != nil { return err } defer j.Destroy() // 生成rbd image name rbdVol.ReservedID, rbdVol.RbdImageName, err = j.ReserveName( ctx, rbdVol.JournalPool, journalPoolID, rbdVol.Pool, imagePoolID, rbdVol.RequestName, rbdVol.NamePrefix, "", kmsID) if err != nil { return err } // 生成volume ID rbdVol.VolID, err = util.GenerateVolID(ctx, rbdVol.Monitors, cr, imagePoolID, rbdVol.Pool, rbdVol.ClusterID, rbdVol.ReservedID, volIDVersion) if err != nil { return err }
util.DebugLog(ctx, "generated Volume ID (%s) and image name (%s) for request name (%s)", rbdVol.VolID, rbdVol.RbdImageName, rbdVol.RequestName)
return nil}
复制代码


1.1 GenerateVolID


调用 vi.ComposeCSIID()来生成 volume ID。


//ceph-csi/internal/util/util.go
func GenerateVolID(ctx context.Context, monitors string, cr *Credentials, locationID int64, pool, clusterID, objUUID string, volIDVersion uint16) (string, error) { var err error
if locationID == InvalidPoolID { locationID, err = GetPoolID(monitors, cr, pool) if err != nil { return "", err } }
// generate the volume ID to return to the CO system vi := CSIIdentifier{ LocationID: locationID, EncodingVersion: volIDVersion, ClusterID: clusterID, ObjectUUID: objUUID, }
volID, err := vi.ComposeCSIID()
return volID, err}
复制代码


vi.ComposeCSIID()为生成 volume ID 的方法,volume ID 由csi_id_version + length of clusterID + clusterID + poolID + ObjectUUID组成,共 64bytes。


//ceph-csi/internal/util/volid.go
/*ComposeCSIID composes a CSI ID from passed in parameters.Version 1 of the encoding scheme is as follows, [csi_id_version=1:4byte] + [-:1byte] [length of clusterID=1:4byte] + [-:1byte] [clusterID:36bytes (MAX)] + [-:1byte] [poolID:16bytes] + [-:1byte] [ObjectUUID:36bytes]
Total of constant field lengths, including '-' field separators would hence be, 4+1+4+1+1+16+1+36 = 64*/func (ci CSIIdentifier) ComposeCSIID() (string, error) { buf16 := make([]byte, 2) buf64 := make([]byte, 8)
if (knownFieldSize + len(ci.ClusterID)) > maxVolIDLen { return "", errors.New("CSI ID encoding length overflow") }
if len(ci.ObjectUUID) != uuidSize { return "", errors.New("CSI ID invalid object uuid") }
binary.BigEndian.PutUint16(buf16, ci.EncodingVersion) versionEncodedHex := hex.EncodeToString(buf16)
binary.BigEndian.PutUint16(buf16, uint16(len(ci.ClusterID))) clusterIDLength := hex.EncodeToString(buf16)
binary.BigEndian.PutUint64(buf64, uint64(ci.LocationID)) poolIDEncodedHex := hex.EncodeToString(buf64)
return strings.Join([]string{versionEncodedHex, clusterIDLength, ci.ClusterID, poolIDEncodedHex, ci.ObjectUUID}, "-"), nil}
复制代码


2 createBackingImage


主要流程:


(1)调用 createImage 创建 image;


(2)调用 j.StoreImageID 存储 image ID 等信息进 omap。


//ceph-csi/internal/rbd/controllerserver.go
func (cs *ControllerServer) createBackingImage(ctx context.Context, cr *util.Credentials, rbdVol, parentVol *rbdVolume, rbdSnap *rbdSnapshot) error { var err error
var j = &journal.Connection{} j, err = volJournal.Connect(rbdVol.Monitors, cr) if err != nil { return status.Error(codes.Internal, err.Error()) } defer j.Destroy()
// nolint:gocritic // this ifElseChain can not be rewritten to a switch statement if rbdSnap != nil { if err = cs.OperationLocks.GetRestoreLock(rbdSnap.SnapID); err != nil { klog.Error(util.Log(ctx, err.Error())) return status.Error(codes.Aborted, err.Error()) } defer cs.OperationLocks.ReleaseRestoreLock(rbdSnap.SnapID)
err = cs.createVolumeFromSnapshot(ctx, cr, rbdVol, rbdSnap.SnapID) if err != nil { return err } util.DebugLog(ctx, "created volume %s from snapshot %s", rbdVol.RequestName, rbdSnap.RbdSnapName) } else if parentVol != nil { if err = cs.OperationLocks.GetCloneLock(parentVol.VolID); err != nil { klog.Error(util.Log(ctx, err.Error())) return status.Error(codes.Aborted, err.Error()) } defer cs.OperationLocks.ReleaseCloneLock(parentVol.VolID) return rbdVol.createCloneFromImage(ctx, parentVol) } else { err = createImage(ctx, rbdVol, cr) if err != nil { klog.Errorf(util.Log(ctx, "failed to create volume: %v"), err) return status.Error(codes.Internal, err.Error()) } }
util.DebugLog(ctx, "created volume %s backed by image %s", rbdVol.RequestName, rbdVol.RbdImageName)
defer func() { if err != nil { if !errors.Is(err, ErrFlattenInProgress) { if deleteErr := deleteImage(ctx, rbdVol, cr); deleteErr != nil { klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s with error: %v"), rbdVol, deleteErr) } } } }() err = rbdVol.getImageID() if err != nil { klog.Errorf(util.Log(ctx, "failed to get volume id %s: %v"), rbdVol, err) return status.Error(codes.Internal, err.Error()) }
err = j.StoreImageID(ctx, rbdVol.JournalPool, rbdVol.ReservedID, rbdVol.ImageID, cr) if err != nil { klog.Errorf(util.Log(ctx, "failed to reserve volume %s: %v"), rbdVol, err) return status.Error(codes.Internal, err.Error()) }
if rbdSnap != nil { err = rbdVol.flattenRbdImage(ctx, cr, false, rbdHardMaxCloneDepth, rbdSoftMaxCloneDepth) if err != nil { klog.Errorf(util.Log(ctx, "failed to flatten image %s: %v"), rbdVol, err) return err } } if rbdVol.Encrypted { err = rbdVol.ensureEncryptionMetadataSet(rbdImageRequiresEncryption) if err != nil { klog.Errorf(util.Log(ctx, "failed to save encryption status, deleting image %s: %s"), rbdVol, err) return status.Error(codes.Internal, err.Error()) } } return nil}
复制代码


2.1 StoreImageID


调用 setOMapKeys 将 ImageID 与 ReservedID 存入 omap 中。


//ceph-csi/internal/journal/voljournal.go
// StoreImageID stores the image ID in omap.func (conn *Connection) StoreImageID(ctx context.Context, pool, reservedUUID, imageID string, cr *util.Credentials) error { err := setOMapKeys(ctx, conn, pool, conn.config.namespace, conn.config.cephUUIDDirectoryPrefix+reservedUUID, map[string]string{conn.config.csiImageIDKey: imageID}) if err != nil { return err } return nil}
复制代码


//ceph-csi/internal/journal/omap.go
func setOMapKeys( ctx context.Context, conn *Connection, poolName, namespace, oid string, pairs map[string]string) error { // fetch and configure the rados ioctx ioctx, err := conn.conn.GetIoctx(poolName) if err != nil { return omapPoolError(err) } defer ioctx.Destroy()
if namespace != "" { ioctx.SetNamespace(namespace) }
bpairs := make(map[string][]byte, len(pairs)) for k, v := range pairs { bpairs[k] = []byte(v) } err = ioctx.SetOmap(oid, bpairs) if err != nil { klog.Errorf( util.Log(ctx, "failed setting omap keys (pool=%q, namespace=%q, name=%q, pairs=%+v): %v"), poolName, namespace, oid, pairs, err) return err } util.DebugLog(ctx, "set omap keys (pool=%q, namespace=%q, name=%q): %+v)", poolName, namespace, oid, pairs) return nil}
复制代码
imageName/volUUID(ReservedID)/VolumeID 示例
apiVersion: v1kind: PersistentVolumemetadata:  annotations:    pv.kubernetes.io/provisioned-by: rbd.csi.ceph.com  creationTimestamp: "2020-08-31T08:23:14Z"  finalizers:  - kubernetes.io/pv-protection  name: pvc-c61bb423-609b-49c0-bbd7-e3db1ad7ba8d  resourceVersion: "66339437"  selfLink: /api/v1/persistentvolumes/pvc-c61bb423-609b-49c0-bbd7-e3db1ad7ba8d  uid: 37f1cbde-deaa-4daf-a8f0-4e61bef47a61spec:  accessModes:  - ReadWriteOnce  capacity:    storage: 1Gi  claimRef:    apiVersion: v1    kind: PersistentVolumeClaim    name: zjl-test2    namespace: test    resourceVersion: "66339424"    uid: c61bb423-609b-49c0-bbd7-e3db1ad7ba8d  csi:    controllerExpandSecretRef:      name: csi-rbd-secret      namespace: test    driver: rbd.csi.ceph.com    fsType: ext4    nodeStageSecretRef:      name: csi-rbd-secret      namespace: test    volumeAttributes:      clusterID: 0bba3be9-0a1c-41db-a619-26ffea20161e      imageFeatures: layering      imageName: csi-vol-36eaa70e-eb63-11ea-8e79-246e96907f74      journalPool: kubernetes      pool: kubernetes      storage.kubernetes.io/csiProvisionerIdentity: 1598236777786-8081-rbd.csi.ceph.com    volumeHandle: 0001-0024-0bba3be9-0a1c-41db-a619-26ffea20161e-0000000000000004-36eaa70e-eb63-11ea-8e79-246e96907f74  mountOptions:  - discard  persistentVolumeReclaimPolicy: Delete  storageClassName: csi-rbd-sc  volumeMode: Filesystemstatus:  phase: Bound
复制代码


imageName: csi-vol-36eaa70e-eb63-11ea-8e79-246e96907f74


volUUID/ReservedID: 36eaa70e-eb63-11ea-8e79-246e96907f74


VolumeID: 0001-0024-0bba3be9-0a1c-41db-a619-26ffea20161e-0000000000000004-36eaa70e-eb63-11ea-8e79-246e96907f74


setOMap(ReservedID/ImageID)

ceph-csi 组件日志示例

操作:CreateVolume


来源:deployment:csi-rbdplugin-provisioner,container:csi-rbdplugin


I0827 07:24:54.293449       1 utils.go:159] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 GRPC call: /csi.v1.Controller/CreateVolumeI0827 07:24:54.296490       1 utils.go:160] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 GRPC request: {"capacity_range":{"required_bytes":1073741824},"name":"pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423","parameters":{"clusterID":"0bba3be9-0a1c-41db-a619-26ffea20161e","imageFeatures":"layering","pool":"kubernetes"},"secrets":"***stripped***","volume_capabilities":[{"AccessType":{"Mount":{"fs_type":"ext4","mount_flags":["discard"]}},"access_mode":{"mode":1}}]}I0827 07:24:54.296666       1 rbd_util.go:722] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 setting disableInUseChecks on rbd volume to: falseI0827 07:24:54.298803       1 omap.go:74] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 got omap values: (pool="kubernetes", namespace="", name="csi.volumes.default"): map[]I0827 07:24:54.332458       1 omap.go:140] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 set omap keys (pool="kubernetes", namespace="", name="csi.volumes.default"): map[csi.volume.pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423:670cc2a2-e836-11ea-8e79-246e96907f74])I0827 07:24:54.342465       1 omap.go:140] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 set omap keys (pool="kubernetes", namespace="", name="csi.volume.670cc2a2-e836-11ea-8e79-246e96907f74"): map[csi.imagename:csi-vol-670cc2a2-e836-11ea-8e79-246e96907f74 csi.volname:pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423])I0827 07:24:54.342511       1 rbd_journal.go:445] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 generated Volume ID (0001-0024-0bba3be9-0a1c-41db-a619-26ffea20161e-0000000000000004-670cc2a2-e836-11ea-8e79-246e96907f74) and image name (csi-vol-670cc2a2-e836-11ea-8e79-246e96907f74) for request name (pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423)I0827 07:24:54.342594       1 rbd_util.go:191] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 rbd: create kubernetes/csi-vol-670cc2a2-e836-11ea-8e79-246e96907f74 size 1024M (features: [layering]) using mon 10.248.32.13:6789,10.248.32.14:6789,10.248.32.15:6789I0827 07:24:54.395928       1 controllerserver.go:446] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 created volume pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 backed by image csi-vol-670cc2a2-e836-11ea-8e79-246e96907f74I0827 07:24:54.479605       1 omap.go:140] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 set omap keys (pool="kubernetes", namespace="", name="csi.volume.670cc2a2-e836-11ea-8e79-246e96907f74"): map[csi.imageid:e5832a10d20c])I0827 07:24:54.481188       1 utils.go:165] ID: 5013 Req-ID: pvc-6eb10d8d-7c2c-4792-9ed6-df17996b1423 GRPC response: {"volume":{"capacity_bytes":1073741824,"volume_context":{"clusterID":"0bba3be9-0a1c-41db-a619-26ffea20161e","imageFeatures":"layering","imageName":"csi-vol-670cc2a2-e836-11ea-8e79-246e96907f74","journalPool":"kubernetes","pool":"kubernetes"},"volume_id":"0001-0024-0bba3be9-0a1c-41db-a619-26ffea20161e-0000000000000004-670cc2a2-e836-11ea-8e79-246e96907f74"}}
复制代码

(2)DeleteVolume

简介

调用 ceph 存储后端,删除存储(rbd image)。


DeleteVolume deletes the volume in backend and removes the volume metadata from store


大致步骤:


(1)检查该 driver 是否支持存储的删除,不支持则直接返回错误;


(2)判断 image 是否还在使用(是否有 watchers);


(3)没被使用则调用 ceph 存储后端删除该 image。

DeleteVolume

主体流程:


(1)检查该 driver 是否支持存储的删除,不支持则直接返回错误;


(2)根据 secret 构建 ceph 请求凭证(secret 由 external-provisioner 组件传入);


(3)从请求参数获取 volumeID;


(4)将请求参数转换为 rbdVol 结构体;


(5)判断 rbd image 是否还在被使用,被使用则返回错误;


(6)删除 temporary rbd image(clone 操作的临时 image);


(7)删除 rbd image。


//ceph-csi/internal/rbd/controllerserver.go
func (cs *ControllerServer) DeleteVolume(ctx context.Context, req *csi.DeleteVolumeRequest) (*csi.DeleteVolumeResponse, error) { // (1)校验请求参数; if err := cs.Driver.ValidateControllerServiceRequest(csi.ControllerServiceCapability_RPC_CREATE_DELETE_VOLUME); err != nil { klog.Errorf(util.Log(ctx, "invalid delete volume req: %v"), protosanitizer.StripSecrets(req)) return nil, err } // (2)根据secret构建ceph请求凭证; cr, err := util.NewUserCredentials(req.GetSecrets()) if err != nil { return nil, status.Error(codes.Internal, err.Error()) } defer cr.DeleteCredentials() // (3)从请求参数获取volumeID; // For now the image get unconditionally deleted, but here retention policy can be checked volumeID := req.GetVolumeId() if volumeID == "" { return nil, status.Error(codes.InvalidArgument, "empty volume ID in request") }
if acquired := cs.VolumeLocks.TryAcquire(volumeID); !acquired { klog.Errorf(util.Log(ctx, util.VolumeOperationAlreadyExistsFmt), volumeID) return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, volumeID) } defer cs.VolumeLocks.Release(volumeID)
// lock out volumeID for clone and expand operation if err = cs.OperationLocks.GetDeleteLock(volumeID); err != nil { klog.Error(util.Log(ctx, err.Error())) return nil, status.Error(codes.Aborted, err.Error()) } defer cs.OperationLocks.ReleaseDeleteLock(volumeID)
rbdVol := &rbdVolume{} defer rbdVol.Destroy() // (4)将请求参数转换为rbdVol结构体; rbdVol, err = genVolFromVolID(ctx, volumeID, cr, req.GetSecrets()) if err != nil { if errors.Is(err, util.ErrPoolNotFound) { klog.Warningf(util.Log(ctx, "failed to get backend volume for %s: %v"), volumeID, err) return &csi.DeleteVolumeResponse{}, nil }
// if error is ErrKeyNotFound, then a previous attempt at deletion was complete // or partially complete (image and imageOMap are garbage collected already), hence return // success as deletion is complete if errors.Is(err, util.ErrKeyNotFound) { klog.Warningf(util.Log(ctx, "Failed to volume options for %s: %v"), volumeID, err) return &csi.DeleteVolumeResponse{}, nil }
// All errors other than ErrImageNotFound should return an error back to the caller if !errors.Is(err, ErrImageNotFound) { return nil, status.Error(codes.Internal, err.Error()) }
// If error is ErrImageNotFound then we failed to find the image, but found the imageOMap // to lead us to the image, hence the imageOMap needs to be garbage collected, by calling // unreserve for the same if acquired := cs.VolumeLocks.TryAcquire(rbdVol.RequestName); !acquired { klog.Errorf(util.Log(ctx, util.VolumeOperationAlreadyExistsFmt), rbdVol.RequestName) return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, rbdVol.RequestName) } defer cs.VolumeLocks.Release(rbdVol.RequestName)
if err = undoVolReservation(ctx, rbdVol, cr); err != nil { return nil, status.Error(codes.Internal, err.Error()) } return &csi.DeleteVolumeResponse{}, nil } defer rbdVol.Destroy()
// lock out parallel create requests against the same volume name as we // cleanup the image and associated omaps for the same if acquired := cs.VolumeLocks.TryAcquire(rbdVol.RequestName); !acquired { klog.Errorf(util.Log(ctx, util.VolumeOperationAlreadyExistsFmt), rbdVol.RequestName) return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, rbdVol.RequestName) } defer cs.VolumeLocks.Release(rbdVol.RequestName) // (5)判断rbd image是否还在被使用,被使用则返回错误; inUse, err := rbdVol.isInUse() if err != nil { klog.Errorf(util.Log(ctx, "failed getting information for image (%s): (%s)"), rbdVol, err) return nil, status.Error(codes.Internal, err.Error()) } if inUse { klog.Errorf(util.Log(ctx, "rbd %s is still being used"), rbdVol) return nil, status.Errorf(codes.Internal, "rbd %s is still being used", rbdVol.RbdImageName) } // (6)删除temporary rbd image; // delete the temporary rbd image created as part of volume clone during // create volume tempClone := rbdVol.generateTempClone() err = deleteImage(ctx, tempClone, cr) if err != nil { // return error if it is not ErrImageNotFound if !errors.Is(err, ErrImageNotFound) { klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s with error: %v"), tempClone, err) return nil, status.Error(codes.Internal, err.Error()) } }
// (7)删除rbd image. // Deleting rbd image util.DebugLog(ctx, "deleting image %s", rbdVol.RbdImageName) if err = deleteImage(ctx, rbdVol, cr); err != nil { klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s with error: %v"), rbdVol, err) return nil, status.Error(codes.Internal, err.Error()) }
if err = undoVolReservation(ctx, rbdVol, cr); err != nil { klog.Errorf(util.Log(ctx, "failed to remove reservation for volume (%s) with backing image (%s) (%s)"), rbdVol.RequestName, rbdVol.RbdImageName, err) return nil, status.Error(codes.Internal, err.Error()) }
if rbdVol.Encrypted { if err = rbdVol.KMS.DeletePassphrase(rbdVol.VolID); err != nil { klog.Warningf(util.Log(ctx, "failed to clean the passphrase for volume %s: %s"), rbdVol.VolID, err) } }
return &csi.DeleteVolumeResponse{}, nil}
复制代码


isInUse


通过列出 image 的 watcher 来判断该 image 是否在被使用。当 image 没有 watcher 时,isInUse 返回 false,否则返回 true。


//ceph-csi/internal/rbd/rbd_util.go
func (rv *rbdVolume) isInUse() (bool, error) { image, err := rv.open() if err != nil { if errors.Is(err, ErrImageNotFound) || errors.Is(err, util.ErrPoolNotFound) { return false, err } // any error should assume something else is using the image return true, err } defer image.Close()
watchers, err := image.ListWatchers() if err != nil { return false, err }
// because we opened the image, there is at least one watcher return len(watchers) != 1, nil}
复制代码

(3)ControllerExpandVolume

简介

调用 ceph 存储后端,扩容存储(rbd image)。


ControllerExpandVolume expand RBD Volumes on demand based on resizer request.


大致步骤:


(1)检查该 driver 是否支持存储扩容,不支持则直接返回错误;


(2)调用 ceph 存储后端扩容该 image。


实际上,存储扩容分为两大步骤,第一步是 external-provisioner 组件调用 rbdType-ControllerServer 类型的 ceph-csi 进行 ControllerExpandVolume,主要负责将底层存储扩容;第二步是 kubelet 调用 rbdType-NodeServer 类型的 ceph-csi 进行 NodeExpandVolume,主要负责将底层 rbd image 的扩容信息同步到 rbd/nbd device,对 xfs/ext 文件系统进行扩展(当 volumemode 是 block,则不用进行 node 端扩容操作)。

ControllerExpandVolume

主体流程:


(1)检查该 driver 是否支持存储扩容,不支持则直接返回错误;


(2)根据 secret 构建 ceph 请求凭证(secret 由 external-provisioner 组件传入);


(3)将请求参数转换为 rbdVol 结构体;


(4)调用 rbdVol.resize 去扩容 image。


func (cs *ControllerServer) ControllerExpandVolume(ctx context.Context, req *csi.ControllerExpandVolumeRequest) (*csi.ControllerExpandVolumeResponse, error) {  if err := cs.Driver.ValidateControllerServiceRequest(csi.ControllerServiceCapability_RPC_EXPAND_VOLUME); err != nil {    klog.Errorf(util.Log(ctx, "invalid expand volume req: %v"), protosanitizer.StripSecrets(req))    return nil, err  }
volID := req.GetVolumeId() if volID == "" { return nil, status.Error(codes.InvalidArgument, "volume ID cannot be empty") }
capRange := req.GetCapacityRange() if capRange == nil { return nil, status.Error(codes.InvalidArgument, "capacityRange cannot be empty") }
// lock out parallel requests against the same volume ID if acquired := cs.VolumeLocks.TryAcquire(volID); !acquired { klog.Errorf(util.Log(ctx, util.VolumeOperationAlreadyExistsFmt), volID) return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, volID) } defer cs.VolumeLocks.Release(volID)
// lock out volumeID for clone and delete operation if err := cs.OperationLocks.GetExpandLock(volID); err != nil { klog.Error(util.Log(ctx, err.Error())) return nil, status.Error(codes.Aborted, err.Error()) } defer cs.OperationLocks.ReleaseExpandLock(volID)
cr, err := util.NewUserCredentials(req.GetSecrets()) if err != nil { return nil, status.Error(codes.Internal, err.Error()) } defer cr.DeleteCredentials()
rbdVol := &rbdVolume{} defer rbdVol.Destroy()
rbdVol, err = genVolFromVolID(ctx, volID, cr, req.GetSecrets()) if err != nil { // nolint:gocritic // this ifElseChain can not be rewritten to a switch statement if errors.Is(err, ErrImageNotFound) { err = status.Errorf(codes.NotFound, "volume ID %s not found", volID) } else if errors.Is(err, util.ErrPoolNotFound) { klog.Errorf(util.Log(ctx, "failed to get backend volume for %s: %v"), volID, err) err = status.Errorf(codes.NotFound, err.Error()) } else { err = status.Errorf(codes.Internal, err.Error()) } return nil, err } defer rbdVol.Destroy()
if rbdVol.Encrypted { return nil, status.Errorf(codes.InvalidArgument, "encrypted volumes do not support resize (%s)", rbdVol) }
// always round up the request size in bytes to the nearest MiB/GiB volSize := util.RoundOffBytes(req.GetCapacityRange().GetRequiredBytes())
// resize volume if required nodeExpansion := false if rbdVol.VolSize < volSize { util.DebugLog(ctx, "rbd volume %s size is %v,resizing to %v", rbdVol, rbdVol.VolSize, volSize) rbdVol.VolSize = volSize nodeExpansion = true err = rbdVol.resize(ctx, cr) if err != nil { klog.Errorf(util.Log(ctx, "failed to resize rbd image: %s with error: %v"), rbdVol, err) return nil, status.Error(codes.Internal, err.Error()) } }
return &csi.ControllerExpandVolumeResponse{ CapacityBytes: rbdVol.VolSize, NodeExpansionRequired: nodeExpansion, }, nil}
复制代码


rbdVol.resize


拼接 resize 命令并执行。


func (rv *rbdVolume) resize(ctx context.Context, cr *util.Credentials) error {  mon := rv.Monitors  volSzMiB := fmt.Sprintf("%dM", util.RoundOffVolSize(rv.VolSize))
args := []string{"resize", rv.String(), "--size", volSzMiB, "--id", cr.ID, "-m", mon, "--keyfile=" + cr.KeyFile} _, stderr, err := util.ExecCommand(ctx, "rbd", args...)
if err != nil { return fmt.Errorf("failed to resize rbd image (%w), command output: %s", err, stderr) }
return nil}
复制代码


至此,rbd driver-controllerserver 的分析已经完成,下面做个总结。

rbd driver-controllerserver 分析总结

(1)controllerserver 主要包括了 CreateVolume(创建存储)、DeleteVolume(删除存储)、ControllerExpandVolume(存储扩容)、CreateSnapshot(创建存储快照)、DeleteSnapshot(删除存储快照)操作,这里主要对 CreateVolume(创建存储)、DeleteVolume(删除存储)、ControllerExpandVolume(存储扩容)进行分析,作用分别如下:


CreateVolume:调用 ceph 存储后端,创建存储(rbd image)。


DeleteVolume:调用 ceph 存储后端,删除存储(rbd image)。


ControllerExpandVolume:调用 ceph 存储后端,扩容存储(rbd image)。


(2)存储扩容分为两大步骤,第一步是 external-provisioner 组件调用 rbdType-ControllerServer 类型的 ceph-csi 进行 ControllerExpandVolume,主要负责将底层存储扩容;第二步是 kubelet 调用 rbdType-NodeServer 类型的 ceph-csi 进行 NodeExpandVolume,主要负责将底层 rbd image 的扩容信息同步到 rbd/nbd device,对 xfs/ext 文件系统进行扩展(当 volumemode 是 block,则不用进行 node 端扩容操作)。

发布于: 2021 年 04 月 30 日阅读数: 15
用户头像

良凯尔

关注

热爱的力量 2020.01.10 加入

kubernetes开发者

评论

发布
暂无评论
ceph-csi源码分析(4)-rbd driver-controllerserver分析