Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

implement Ephemeral Inline Volumes support #198

Closed
andyzhangx opened this issue Jan 8, 2021 · 36 comments
Closed

implement Ephemeral Inline Volumes support #198

andyzhangx opened this issue Jan 8, 2021 · 36 comments
Labels
kind/feature Categorizes issue or PR as related to a new feature. lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.

Comments

@andyzhangx
Copy link
Member

Is your feature request related to a problem?/Why is this needed

Describe the solution you'd like in detail

refer to
https://kubernetes-csi.github.io/docs/ephemeral-local-volumes.html#which-feature-should-my-driver-support

Describe alternatives you've considered

Additional context

@andyzhangx andyzhangx added kind/feature Categorizes issue or PR as related to a new feature. help wanted Denotes an issue that needs help from a contributor. Must meet "help wanted" guidelines. labels Jan 8, 2021
@andyzhangx
Copy link
Member Author

cc @rajibmitra

@rajibmitra
Copy link

thank you @andyzhangx, looking at it.

@fejta-bot
Copy link

Issues go stale after 90d of inactivity.
Mark the issue as fresh with /remove-lifecycle stale.
Stale issues rot after an additional 30d of inactivity and eventually close.

If this issue is safe to close now please do so with /close.

Send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label May 2, 2021
@fejta-bot
Copy link

Stale issues rot after 30d of inactivity.
Mark the issue as fresh with /remove-lifecycle rotten.
Rotten issues close after an additional 30d of inactivity.

If this issue is safe to close now please do so with /close.

Send feedback to sig-contributor-experience at kubernetes/community.
/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Jun 1, 2021
@rajibmitra
Copy link

/remove-lifecycle rotten

@k8s-ci-robot k8s-ci-robot removed the lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. label Jun 19, 2021
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Sep 17, 2021
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Oct 17, 2021
@omerozery
Copy link

any news?

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

@k8s-ci-robot
Copy link
Contributor

@k8s-triage-robot: Closing this issue.

In response to this:

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@GuyPaddock
Copy link

/reopen

I believe something like this could solve some problems for us. We want pods to be able to connect to existing, pre-allocated storage accounts on a CIFS/SMB or NFS server without having to pre-declare a PV and manage a PVC for each one.

One question -- how would mount options like UID, GID, and authentication credentials be handled with an inline volume? I believe that the docs for inline say that mount options aren't supported.

@k8s-ci-robot
Copy link
Contributor

@GuyPaddock: You can't reopen an issue/PR unless you authored it or you are a collaborator.

In response to this:

/reopen

I believe something like this could solve some problems for us. We want pods to be able to connect to existing, pre-allocated storage accounts on a CIFS/SMB or NFS server without having to pre-declare a PV and manage a PVC for each one.

One question -- how would mount options like UID, GID, and authentication credentials be handled with an inline volume? I believe that the docs for inline say that mount options aren't supported.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@GuyPaddock
Copy link

@andyzhangx would you consider re-opening?

@andyzhangx
Copy link
Member Author

@GuyPaddock
there are concerns from community that we don't want to support this feature, details: kubernetes-csi/csi-driver-nfs#148
cc @jsafrane

@andyzhangx andyzhangx reopened this Dec 31, 2021
@andyzhangx andyzhangx removed the lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. label Jan 9, 2022
@andyzhangx
Copy link
Member Author

andyzhangx commented Jan 9, 2022

/reopen

I believe something like this could solve some problems for us. We want pods to be able to connect to existing, pre-allocated storage accounts on a CIFS/SMB or NFS server without having to pre-declare a PV and manage a PVC for each one.

One question -- how would mount options like UID, GID, and authentication credentials be handled with an inline volume? I believe that the docs for inline say that mount options aren't supported.

@GuyPaddock I re-evaluated this inline volume requirement, this would requires that pod could read any secret in same namespace (to get username and password), there is security risk to support such functionality. It would be ok for NFS driver since nfs does not use password auth, thus does not require reading secret.

So we don't want to implement Ephemeral Inline Volumes support for this SMB driver, pls use PV/PVC way, it can also connect to existing, pre-allocated storage accounts.

connect to existing, pre-allocated storage accounts on a CIFS/SMB or NFS server

BTW, are you referring Azure storage accounts?

@GuyPaddock
Copy link

@andyzhangx Thank you for the context! That makes sense. We were using Azure Files but are now moving to Qumulo.

@andyzhangx
Copy link
Member Author

Qumulo

@GuyPaddock do you have any suggestions why you are using Qumulo instead of Azure Files? thanks.

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle stale
  • Mark this issue or PR as rotten with /lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Jun 12, 2022
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Close this issue or PR with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Jul 12, 2022
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

@k8s-ci-robot
Copy link
Contributor

@k8s-triage-robot: Closing this issue.

In response to this:

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues and PRs according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue or PR with /reopen
  • Mark this issue or PR as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@kfox1111
Copy link

/remove-lifecycle stale

@kfox1111
Copy link

/remove-lifecycle rotten

@kfox1111
Copy link

I think this is still a valid use case. The same functionality was implemented optionally with the nfs csi driver.

@k8s-ci-robot k8s-ci-robot removed the lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. label Aug 22, 2022
@kfox1111
Copy link

/reopen

@k8s-ci-robot
Copy link
Contributor

@kfox1111: Reopened this issue.

In response to this:

/reopen

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@k8s-ci-robot k8s-ci-robot reopened this Aug 22, 2022
@raafatseif
Copy link

raafatseif commented Dec 28, 2022

Hi @andyzhangx! Regarding reading secrets for username and password, would it deviate a lot from the current implementation? I know it currently reads from the secret provided in the storage class or PV. Would it be possible to use the same implementation only difference would be providing the secret through the pod inline?

  csi.storage.k8s.io/node-stage-secret-name: "smbcreds"
  csi.storage.k8s.io/node-stage-secret-namespace: "default"

@kfox1111
Copy link

csi.storage.k8s.io/node-stage-secret-namespace should be ignored in inline mode. only the pods namespace should be honored.

@raafatseif
Copy link

Gotcha, I think that might actually simplify things. Was just looking at the implementation from azurefile inline volume example.

@andyzhangx andyzhangx removed the help wanted Denotes an issue that needs help from a contributor. Must meet "help wanted" guidelines. label Dec 29, 2022
andyzhangx added a commit to andyzhangx/csi-driver-smb that referenced this issue Feb 19, 2023
aa61bfd0 Merge pull request kubernetes-csi#218 from xing-yang/update_csi_driver
7563d196 Update CSI_PROW_DRIVER_VERSION to v1.11.0
a2171bef Merge pull request kubernetes-csi#216 from msau42/process
cb987826 Merge pull request kubernetes-csi#217 from msau42/owners
a11216e4 add new reviewers and remove inactive reviewers
dd986754 Add step for checking builds
b66c0824 Merge pull request kubernetes-csi#214 from pohly/junit-fixes
b9b6763b filter-junit.go: fix loss of testcases when parsing Ginkgo v2 JUnit
d4277839 filter-junit.go: preserve system error log
38e11468 prow.sh: publish individual JUnit files as separate artifacts
78c0fb71 Merge pull request kubernetes-csi#208 from jsafrane/skip-selinux
36e433e2 Skip SELinux tests in CI by default
348d4a92 Merge pull request kubernetes-csi#207 from RaunakShah/reviewers
1efc2724 Merge pull request kubernetes-csi#206 from RaunakShah/update-prow
7d410d88 Changes to csi prow to run e2e tests in sidecars
cfa5a75c Merge pull request kubernetes-csi#203 from humblec/test-vendor
4edd1d8a Add RaunakShah to CSI reviewers group
7ccc9594 release tools update to 1.19
d24254f6 Merge pull request kubernetes-csi#202 from xing-yang/kind_0.14.0
0faa3fc7 Update to Kind v0.14.0 images
ef4e1b2b Merge pull request kubernetes-csi#201 from xing-yang/add_1.24_image
4ddce251 Add 1.24 Kind image
7fe51491 Merge pull request kubernetes-csi#200 from pohly/bump-kubernetes-version
70915a8e prow.sh: update snapshotter version
31a3f38b Merge pull request kubernetes-csi#199 from pohly/bump-kubernetes-version
7577454a prow.sh: bump Kubernetes to v1.22.0
d29a2e75 Merge pull request kubernetes-csi#198 from pohly/csi-test-5.0.0
41cb70d3 prow.sh: sanity testing with csi-test v5.0.0
c85a63fb Merge pull request kubernetes-csi#197 from pohly/fix-alpha-testing
b86d8e94 support Kubernetes 1.25 + Ginkgo v2
ab0b0a3d Merge pull request kubernetes-csi#192 from andyzhangx/patch-1
7bbab24e Merge pull request kubernetes-csi#196 from humblec/non-alpha
e51ff2cc introduce control variable for non alpha feature gate configuration
ca19ef52 Merge pull request kubernetes-csi#195 from pohly/fix-alpha-testing
3948331e fix testing with latest Kubernetes
e4dab7ff Merge pull request kubernetes-csi#194 from yselkowitz/registry-k8s-io
84a4d5a1 Move from k8s.gcr.io to registry.k8s.io
9a0260c5 fix boilerplate header
37d1104 Merge pull request kubernetes-csi#191 from pohly/go-1.18
db917f5 update to Go 1.18

git-subtree-dir: release-tools
git-subtree-split: aa61bfd0c1a80460aba7cb0feddc8cdee03622a4
andyzhangx added a commit that referenced this issue Feb 20, 2023
aa61bfd0 Merge pull request #218 from xing-yang/update_csi_driver
7563d196 Update CSI_PROW_DRIVER_VERSION to v1.11.0
a2171bef Merge pull request #216 from msau42/process
cb987826 Merge pull request #217 from msau42/owners
a11216e4 add new reviewers and remove inactive reviewers
dd986754 Add step for checking builds
b66c0824 Merge pull request #214 from pohly/junit-fixes
b9b6763b filter-junit.go: fix loss of testcases when parsing Ginkgo v2 JUnit
d4277839 filter-junit.go: preserve system error log
38e11468 prow.sh: publish individual JUnit files as separate artifacts
78c0fb71 Merge pull request #208 from jsafrane/skip-selinux
36e433e2 Skip SELinux tests in CI by default
348d4a92 Merge pull request #207 from RaunakShah/reviewers
1efc2724 Merge pull request #206 from RaunakShah/update-prow
7d410d88 Changes to csi prow to run e2e tests in sidecars
cfa5a75c Merge pull request #203 from humblec/test-vendor
4edd1d8a Add RaunakShah to CSI reviewers group
7ccc9594 release tools update to 1.19
d24254f6 Merge pull request #202 from xing-yang/kind_0.14.0
0faa3fc7 Update to Kind v0.14.0 images
ef4e1b2b Merge pull request #201 from xing-yang/add_1.24_image
4ddce251 Add 1.24 Kind image
7fe51491 Merge pull request #200 from pohly/bump-kubernetes-version
70915a8e prow.sh: update snapshotter version
31a3f38b Merge pull request #199 from pohly/bump-kubernetes-version
7577454a prow.sh: bump Kubernetes to v1.22.0
d29a2e75 Merge pull request #198 from pohly/csi-test-5.0.0
41cb70d3 prow.sh: sanity testing with csi-test v5.0.0
c85a63fb Merge pull request #197 from pohly/fix-alpha-testing
b86d8e94 support Kubernetes 1.25 + Ginkgo v2
ab0b0a3d Merge pull request #192 from andyzhangx/patch-1
7bbab24e Merge pull request #196 from humblec/non-alpha
e51ff2cc introduce control variable for non alpha feature gate configuration
ca19ef52 Merge pull request #195 from pohly/fix-alpha-testing
3948331e fix testing with latest Kubernetes
e4dab7ff Merge pull request #194 from yselkowitz/registry-k8s-io
84a4d5a1 Move from k8s.gcr.io to registry.k8s.io
9a0260c5 fix boilerplate header
37d1104 Merge pull request #191 from pohly/go-1.18
db917f5 update to Go 1.18

git-subtree-dir: release-tools
git-subtree-split: aa61bfd0c1a80460aba7cb0feddc8cdee03622a4
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Mar 29, 2023
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle rotten
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle rotten

@k8s-ci-robot k8s-ci-robot added lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed. and removed lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. labels Apr 28, 2023
@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue with /reopen
  • Mark this issue as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close not-planned

@k8s-ci-robot
Copy link
Contributor

@k8s-triage-robot: Closing this issue, marking it as "Not Planned".

In response to this:

The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs.

This bot triages issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Reopen this issue with /reopen
  • Mark this issue as fresh with /remove-lifecycle rotten
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/close not-planned

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@k8s-ci-robot k8s-ci-robot closed this as not planned Won't fix, can't repro, duplicate, stale May 28, 2023
@kfox1111
Copy link

Still useful

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/feature Categorizes issue or PR as related to a new feature. lifecycle/rotten Denotes an issue or PR that has aged beyond stale and will be auto-closed.
Projects
None yet
Development

No branches or pull requests

10 participants