Dave Chinner | 0b61f8a | 2018-06-05 19:42:14 -0700 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0 |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 2 | /* |
| 3 | * Copyright (c) 2008, Christoph Hellwig |
| 4 | * All Rights Reserved. |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 5 | */ |
| 6 | #include "xfs.h" |
Darrick J. Wong | 5467b34 | 2019-06-28 19:25:35 -0700 | [diff] [blame] | 7 | #include "xfs_shared.h" |
Dave Chinner | a4fbe6a | 2013-10-23 10:51:50 +1100 | [diff] [blame] | 8 | #include "xfs_format.h" |
Dave Chinner | 6943283 | 2013-08-12 20:49:23 +1000 | [diff] [blame] | 9 | #include "xfs_log_format.h" |
Dave Chinner | 7fd36c4 | 2013-08-12 20:49:32 +1000 | [diff] [blame] | 10 | #include "xfs_trans_resv.h" |
Dave Chinner | 0a8aa19 | 2013-06-05 12:09:10 +1000 | [diff] [blame] | 11 | #include "xfs_mount.h" |
Dave Chinner | a4fbe6a | 2013-10-23 10:51:50 +1100 | [diff] [blame] | 12 | #include "xfs_inode.h" |
Dave Chinner | a4fbe6a | 2013-10-23 10:51:50 +1100 | [diff] [blame] | 13 | #include "xfs_attr.h" |
Christoph Hellwig | 0b1b213 | 2009-12-14 23:14:59 +0000 | [diff] [blame] | 14 | #include "xfs_trace.h" |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 15 | #include "xfs_error.h" |
Darrick J. Wong | 5f213dd | 2019-11-06 17:19:33 -0800 | [diff] [blame] | 16 | #include "xfs_acl.h" |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 17 | #include "xfs_da_format.h" |
| 18 | #include "xfs_da_btree.h" |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 19 | #include "xfs_trans.h" |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 20 | |
Darrick J. Wong | 5f213dd | 2019-11-06 17:19:33 -0800 | [diff] [blame] | 21 | #include <linux/posix_acl_xattr.h> |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 22 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 23 | /* |
| 24 | * Locking scheme: |
| 25 | * - all ACL updates are protected by inode->i_mutex, which is taken before |
| 26 | * calling into this file. |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 27 | */ |
| 28 | |
| 29 | STATIC struct posix_acl * |
Dave Chinner | 0a8aa19 | 2013-06-05 12:09:10 +1000 | [diff] [blame] | 30 | xfs_acl_from_disk( |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 31 | struct xfs_mount *mp, |
Andreas Gruenbacher | 86a21c7 | 2015-11-03 12:41:59 +1100 | [diff] [blame] | 32 | const struct xfs_acl *aclp, |
| 33 | int len, |
| 34 | int max_entries) |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 35 | { |
| 36 | struct posix_acl_entry *acl_e; |
| 37 | struct posix_acl *acl; |
Andreas Gruenbacher | 86a21c7 | 2015-11-03 12:41:59 +1100 | [diff] [blame] | 38 | const struct xfs_acl_entry *ace; |
Xi Wang | 093019c | 2011-12-12 21:55:52 +0000 | [diff] [blame] | 39 | unsigned int count, i; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 40 | |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 41 | if (len < sizeof(*aclp)) { |
| 42 | XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, aclp, |
| 43 | len); |
Andreas Gruenbacher | 86a21c7 | 2015-11-03 12:41:59 +1100 | [diff] [blame] | 44 | return ERR_PTR(-EFSCORRUPTED); |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 45 | } |
| 46 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 47 | count = be32_to_cpu(aclp->acl_cnt); |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 48 | if (count > max_entries || XFS_ACL_SIZE(count) != len) { |
| 49 | XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, aclp, |
| 50 | len); |
Christoph Hellwig | fa8b18e | 2011-11-20 15:35:32 +0000 | [diff] [blame] | 51 | return ERR_PTR(-EFSCORRUPTED); |
Darrick J. Wong | a5155b8 | 2019-11-02 09:40:53 -0700 | [diff] [blame] | 52 | } |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 53 | |
| 54 | acl = posix_acl_alloc(count, GFP_KERNEL); |
| 55 | if (!acl) |
| 56 | return ERR_PTR(-ENOMEM); |
| 57 | |
| 58 | for (i = 0; i < count; i++) { |
| 59 | acl_e = &acl->a_entries[i]; |
| 60 | ace = &aclp->acl_entry[i]; |
| 61 | |
| 62 | /* |
| 63 | * The tag is 32 bits on disk and 16 bits in core. |
| 64 | * |
| 65 | * Because every access to it goes through the core |
| 66 | * format first this is not a problem. |
| 67 | */ |
| 68 | acl_e->e_tag = be32_to_cpu(ace->ae_tag); |
| 69 | acl_e->e_perm = be16_to_cpu(ace->ae_perm); |
| 70 | |
| 71 | switch (acl_e->e_tag) { |
| 72 | case ACL_USER: |
Christoph Hellwig | ba8adad | 2020-02-21 08:31:27 -0800 | [diff] [blame] | 73 | acl_e->e_uid = make_kuid(&init_user_ns, |
| 74 | be32_to_cpu(ace->ae_id)); |
Dwight Engen | 288bbe0 | 2013-08-15 14:07:59 -0400 | [diff] [blame] | 75 | break; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 76 | case ACL_GROUP: |
Christoph Hellwig | ba8adad | 2020-02-21 08:31:27 -0800 | [diff] [blame] | 77 | acl_e->e_gid = make_kgid(&init_user_ns, |
| 78 | be32_to_cpu(ace->ae_id)); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 79 | break; |
| 80 | case ACL_USER_OBJ: |
| 81 | case ACL_GROUP_OBJ: |
| 82 | case ACL_MASK: |
| 83 | case ACL_OTHER: |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 84 | break; |
| 85 | default: |
| 86 | goto fail; |
| 87 | } |
| 88 | } |
| 89 | return acl; |
| 90 | |
| 91 | fail: |
| 92 | posix_acl_release(acl); |
| 93 | return ERR_PTR(-EINVAL); |
| 94 | } |
| 95 | |
| 96 | STATIC void |
| 97 | xfs_acl_to_disk(struct xfs_acl *aclp, const struct posix_acl *acl) |
| 98 | { |
| 99 | const struct posix_acl_entry *acl_e; |
| 100 | struct xfs_acl_entry *ace; |
| 101 | int i; |
| 102 | |
| 103 | aclp->acl_cnt = cpu_to_be32(acl->a_count); |
| 104 | for (i = 0; i < acl->a_count; i++) { |
| 105 | ace = &aclp->acl_entry[i]; |
| 106 | acl_e = &acl->a_entries[i]; |
| 107 | |
| 108 | ace->ae_tag = cpu_to_be32(acl_e->e_tag); |
Dwight Engen | 288bbe0 | 2013-08-15 14:07:59 -0400 | [diff] [blame] | 109 | switch (acl_e->e_tag) { |
| 110 | case ACL_USER: |
Christoph Hellwig | ba8adad | 2020-02-21 08:31:27 -0800 | [diff] [blame] | 111 | ace->ae_id = cpu_to_be32( |
| 112 | from_kuid(&init_user_ns, acl_e->e_uid)); |
Dwight Engen | 288bbe0 | 2013-08-15 14:07:59 -0400 | [diff] [blame] | 113 | break; |
| 114 | case ACL_GROUP: |
Christoph Hellwig | ba8adad | 2020-02-21 08:31:27 -0800 | [diff] [blame] | 115 | ace->ae_id = cpu_to_be32( |
| 116 | from_kgid(&init_user_ns, acl_e->e_gid)); |
Dwight Engen | 288bbe0 | 2013-08-15 14:07:59 -0400 | [diff] [blame] | 117 | break; |
| 118 | default: |
| 119 | ace->ae_id = cpu_to_be32(ACL_UNDEFINED_ID); |
| 120 | break; |
| 121 | } |
| 122 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 123 | ace->ae_perm = cpu_to_be16(acl_e->e_perm); |
| 124 | } |
| 125 | } |
| 126 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 127 | struct posix_acl * |
| 128 | xfs_get_acl(struct inode *inode, int type) |
| 129 | { |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 130 | struct xfs_inode *ip = XFS_I(inode); |
| 131 | struct xfs_mount *mp = ip->i_mount; |
| 132 | struct posix_acl *acl = NULL; |
| 133 | struct xfs_da_args args = { |
| 134 | .dp = ip, |
Christoph Hellwig | d5f0f49 | 2020-02-26 17:30:42 -0800 | [diff] [blame] | 135 | .attr_filter = XFS_ATTR_ROOT, |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 136 | .valuelen = XFS_ACL_MAX_SIZE(mp), |
| 137 | }; |
| 138 | int error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 139 | |
Christoph Hellwig | 4e34e71 | 2011-07-23 17:37:31 +0200 | [diff] [blame] | 140 | trace_xfs_get_acl(ip); |
| 141 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 142 | switch (type) { |
| 143 | case ACL_TYPE_ACCESS: |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 144 | args.name = SGI_ACL_FILE; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 145 | break; |
| 146 | case ACL_TYPE_DEFAULT: |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 147 | args.name = SGI_ACL_DEFAULT; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 148 | break; |
| 149 | default: |
Al Viro | 1cbd20d | 2009-06-09 13:29:39 -0400 | [diff] [blame] | 150 | BUG(); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 151 | } |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 152 | args.namelen = strlen(args.name); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 153 | |
Christoph Hellwig | d49db18 | 2020-02-26 17:30:35 -0800 | [diff] [blame] | 154 | /* |
| 155 | * If the attribute doesn't exist make sure we have a negative cache |
| 156 | * entry, for any other error assume it is transient. |
| 157 | */ |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 158 | error = xfs_attr_get(&args); |
Christoph Hellwig | d49db18 | 2020-02-26 17:30:35 -0800 | [diff] [blame] | 159 | if (!error) { |
Christoph Hellwig | e5171d7 | 2020-02-26 17:30:34 -0800 | [diff] [blame] | 160 | acl = xfs_acl_from_disk(mp, args.value, args.valuelen, |
| 161 | XFS_ACL_MAX_ENTRIES(mp)); |
Christoph Hellwig | d49db18 | 2020-02-26 17:30:35 -0800 | [diff] [blame] | 162 | } else if (error != -ENOATTR) { |
| 163 | acl = ERR_PTR(error); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 164 | } |
Christoph Hellwig | d49db18 | 2020-02-26 17:30:35 -0800 | [diff] [blame] | 165 | |
| 166 | kmem_free(args.value); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 167 | return acl; |
| 168 | } |
| 169 | |
Jan Kara | 8ba3587 | 2017-06-26 08:48:18 -0700 | [diff] [blame] | 170 | int |
| 171 | __xfs_set_acl(struct inode *inode, struct posix_acl *acl, int type) |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 172 | { |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 173 | struct xfs_inode *ip = XFS_I(inode); |
| 174 | struct xfs_da_args args = { |
| 175 | .dp = ip, |
Christoph Hellwig | d5f0f49 | 2020-02-26 17:30:42 -0800 | [diff] [blame] | 176 | .attr_filter = XFS_ATTR_ROOT, |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 177 | }; |
| 178 | int error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 179 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 180 | switch (type) { |
| 181 | case ACL_TYPE_ACCESS: |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 182 | args.name = SGI_ACL_FILE; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 183 | break; |
| 184 | case ACL_TYPE_DEFAULT: |
| 185 | if (!S_ISDIR(inode->i_mode)) |
| 186 | return acl ? -EACCES : 0; |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 187 | args.name = SGI_ACL_DEFAULT; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 188 | break; |
| 189 | default: |
| 190 | return -EINVAL; |
| 191 | } |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 192 | args.namelen = strlen(args.name); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 193 | |
| 194 | if (acl) { |
Christoph Hellwig | ed02d13 | 2020-02-26 17:30:44 -0800 | [diff] [blame] | 195 | args.valuelen = XFS_ACL_SIZE(acl->a_count); |
Carlos Maiolino | 8ca79df | 2020-09-01 11:47:12 -0700 | [diff] [blame] | 196 | args.value = kvzalloc(args.valuelen, GFP_KERNEL); |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 197 | if (!args.value) |
Dave Chinner | fdd3cce | 2013-09-02 20:53:00 +1000 | [diff] [blame] | 198 | return -ENOMEM; |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 199 | xfs_acl_to_disk(args.value, acl); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 200 | } |
| 201 | |
Christoph Hellwig | a254462 | 2020-02-26 17:30:33 -0800 | [diff] [blame] | 202 | error = xfs_attr_set(&args); |
| 203 | kmem_free(args.value); |
Christoph Hellwig | 0eb81a5 | 2020-02-26 17:30:29 -0800 | [diff] [blame] | 204 | |
| 205 | /* |
| 206 | * If the attribute didn't exist to start with that's fine. |
| 207 | */ |
| 208 | if (!acl && error == -ENOATTR) |
| 209 | error = 0; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 210 | if (!error) |
Al Viro | 1cbd20d | 2009-06-09 13:29:39 -0400 | [diff] [blame] | 211 | set_cached_acl(inode, type, acl); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 212 | return error; |
| 213 | } |
| 214 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 215 | static int |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 216 | xfs_acl_set_mode( |
| 217 | struct inode *inode, |
| 218 | umode_t mode) |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 219 | { |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 220 | struct xfs_inode *ip = XFS_I(inode); |
| 221 | struct xfs_mount *mp = ip->i_mount; |
| 222 | struct xfs_trans *tp; |
| 223 | int error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 224 | |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 225 | error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp); |
| 226 | if (error) |
| 227 | return error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 228 | |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 229 | xfs_ilock(ip, XFS_ILOCK_EXCL); |
| 230 | xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL); |
| 231 | inode->i_mode = mode; |
| 232 | inode->i_ctime = current_time(inode); |
| 233 | xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 234 | |
Dave Chinner | 0560f31 | 2021-08-18 18:46:52 -0700 | [diff] [blame^] | 235 | if (xfs_has_wsync(mp)) |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 236 | xfs_trans_set_sync(tp); |
| 237 | return xfs_trans_commit(tp); |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 238 | } |
| 239 | |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 240 | int |
Christian Brauner | 549c729 | 2021-01-21 14:19:43 +0100 | [diff] [blame] | 241 | xfs_set_acl(struct user_namespace *mnt_userns, struct inode *inode, |
| 242 | struct posix_acl *acl, int type) |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 243 | { |
Dave Chinner | 67f2ffe | 2017-10-09 11:37:23 -0700 | [diff] [blame] | 244 | umode_t mode; |
| 245 | bool set_mode = false; |
Christoph Hellwig | 431547b | 2009-11-13 09:52:56 +0000 | [diff] [blame] | 246 | int error = 0; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 247 | |
Christoph Hellwig | 2401dc2 | 2013-12-20 05:16:50 -0800 | [diff] [blame] | 248 | if (!acl) |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 249 | goto set_acl; |
| 250 | |
Jie Liu | 4ae69fe | 2014-02-07 15:26:11 +1100 | [diff] [blame] | 251 | error = -E2BIG; |
Dave Chinner | 0a8aa19 | 2013-06-05 12:09:10 +1000 | [diff] [blame] | 252 | if (acl->a_count > XFS_ACL_MAX_ENTRIES(XFS_M(inode->i_sb))) |
Christoph Hellwig | 2401dc2 | 2013-12-20 05:16:50 -0800 | [diff] [blame] | 253 | return error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 254 | |
| 255 | if (type == ACL_TYPE_ACCESS) { |
Christoph Hellwig | f736d93 | 2021-01-21 14:19:58 +0100 | [diff] [blame] | 256 | error = posix_acl_update_mode(mnt_userns, inode, &mode, &acl); |
Jan Kara | 0739310 | 2016-09-19 17:39:09 +0200 | [diff] [blame] | 257 | if (error) |
| 258 | return error; |
Dave Chinner | 67f2ffe | 2017-10-09 11:37:23 -0700 | [diff] [blame] | 259 | set_mode = true; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 260 | } |
| 261 | |
| 262 | set_acl: |
Dave Chinner | 67f2ffe | 2017-10-09 11:37:23 -0700 | [diff] [blame] | 263 | /* |
| 264 | * We set the mode after successfully updating the ACL xattr because the |
| 265 | * xattr update can fail at ENOSPC and we don't want to change the mode |
| 266 | * if the ACL update hasn't been applied. |
| 267 | */ |
Christoph Hellwig | 5d24ec4c | 2020-12-10 20:00:39 -0800 | [diff] [blame] | 268 | error = __xfs_set_acl(inode, acl, type); |
| 269 | if (!error && set_mode && mode != inode->i_mode) |
| 270 | error = xfs_acl_set_mode(inode, mode); |
Dave Chinner | 67f2ffe | 2017-10-09 11:37:23 -0700 | [diff] [blame] | 271 | return error; |
Christoph Hellwig | ef14f0c | 2009-06-10 17:07:47 +0200 | [diff] [blame] | 272 | } |
Christoph Hellwig | 5a3930e | 2020-02-26 17:30:41 -0800 | [diff] [blame] | 273 | |
| 274 | /* |
| 275 | * Invalidate any cached ACLs if the user has bypassed the ACL interface. |
| 276 | * We don't validate the content whatsoever so it is caller responsibility to |
| 277 | * provide data in valid format and ensure i_mode is consistent. |
| 278 | */ |
| 279 | void |
| 280 | xfs_forget_acl( |
| 281 | struct inode *inode, |
| 282 | const char *name) |
| 283 | { |
| 284 | if (!strcmp(name, SGI_ACL_FILE)) |
| 285 | forget_cached_acl(inode, ACL_TYPE_ACCESS); |
| 286 | else if (!strcmp(name, SGI_ACL_DEFAULT)) |
| 287 | forget_cached_acl(inode, ACL_TYPE_DEFAULT); |
| 288 | } |