David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0-only |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 2 | /* |
| 3 | * |
| 4 | * Copyright (C) 2011 Novell Inc. |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 5 | */ |
| 6 | |
| 7 | #include <linux/fs.h> |
| 8 | #include <linux/slab.h> |
| 9 | #include <linux/cred.h> |
| 10 | #include <linux/xattr.h> |
| 11 | #include <linux/posix_acl.h> |
| 12 | #include <linux/ratelimit.h> |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 13 | #include <linux/fiemap.h> |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 14 | #include "overlayfs.h" |
| 15 | |
| 16 | |
| 17 | int ovl_setattr(struct dentry *dentry, struct iattr *attr) |
| 18 | { |
| 19 | int err; |
| 20 | bool full_copy_up = false; |
| 21 | struct dentry *upperdentry; |
| 22 | const struct cred *old_cred; |
| 23 | |
| 24 | err = setattr_prepare(dentry, attr); |
| 25 | if (err) |
| 26 | return err; |
| 27 | |
| 28 | err = ovl_want_write(dentry); |
| 29 | if (err) |
| 30 | goto out; |
| 31 | |
| 32 | if (attr->ia_valid & ATTR_SIZE) { |
| 33 | struct inode *realinode = d_inode(ovl_dentry_real(dentry)); |
| 34 | |
| 35 | err = -ETXTBSY; |
| 36 | if (atomic_read(&realinode->i_writecount) < 0) |
| 37 | goto out_drop_write; |
| 38 | |
| 39 | /* Truncate should trigger data copy up as well */ |
| 40 | full_copy_up = true; |
| 41 | } |
| 42 | |
| 43 | if (!full_copy_up) |
| 44 | err = ovl_copy_up(dentry); |
| 45 | else |
| 46 | err = ovl_copy_up_with_data(dentry); |
| 47 | if (!err) { |
| 48 | struct inode *winode = NULL; |
| 49 | |
| 50 | upperdentry = ovl_dentry_upper(dentry); |
| 51 | |
| 52 | if (attr->ia_valid & ATTR_SIZE) { |
| 53 | winode = d_inode(upperdentry); |
| 54 | err = get_write_access(winode); |
| 55 | if (err) |
| 56 | goto out_drop_write; |
| 57 | } |
| 58 | |
| 59 | if (attr->ia_valid & (ATTR_KILL_SUID|ATTR_KILL_SGID)) |
| 60 | attr->ia_valid &= ~ATTR_MODE; |
| 61 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 62 | /* |
| 63 | * We might have to translate ovl file into real file object |
| 64 | * once use cases emerge. For now, simply don't let underlying |
| 65 | * filesystem rely on attr->ia_file |
| 66 | */ |
| 67 | attr->ia_valid &= ~ATTR_FILE; |
| 68 | |
| 69 | /* |
| 70 | * If open(O_TRUNC) is done, VFS calls ->setattr with ATTR_OPEN |
| 71 | * set. Overlayfs does not pass O_TRUNC flag to underlying |
| 72 | * filesystem during open -> do not pass ATTR_OPEN. This |
| 73 | * disables optimization in fuse which assumes open(O_TRUNC) |
| 74 | * already set file size to 0. But we never passed O_TRUNC to |
| 75 | * fuse. So by clearing ATTR_OPEN, fuse will be forced to send |
| 76 | * setattr request to server. |
| 77 | */ |
| 78 | attr->ia_valid &= ~ATTR_OPEN; |
| 79 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 80 | inode_lock(upperdentry->d_inode); |
| 81 | old_cred = ovl_override_creds(dentry->d_sb); |
| 82 | err = notify_change(upperdentry, attr, NULL); |
| 83 | revert_creds(old_cred); |
| 84 | if (!err) |
| 85 | ovl_copyattr(upperdentry->d_inode, dentry->d_inode); |
| 86 | inode_unlock(upperdentry->d_inode); |
| 87 | |
| 88 | if (winode) |
| 89 | put_write_access(winode); |
| 90 | } |
| 91 | out_drop_write: |
| 92 | ovl_drop_write(dentry); |
| 93 | out: |
| 94 | return err; |
| 95 | } |
| 96 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 97 | static int ovl_map_dev_ino(struct dentry *dentry, struct kstat *stat, int fsid) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 98 | { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 99 | bool samefs = ovl_same_fs(dentry->d_sb); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 100 | unsigned int xinobits = ovl_xino_bits(dentry->d_sb); |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 101 | unsigned int xinoshift = 64 - xinobits; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 102 | |
| 103 | if (samefs) { |
| 104 | /* |
| 105 | * When all layers are on the same fs, all real inode |
| 106 | * number are unique, so we use the overlay st_dev, |
| 107 | * which is friendly to du -x. |
| 108 | */ |
| 109 | stat->dev = dentry->d_sb->s_dev; |
| 110 | return 0; |
| 111 | } else if (xinobits) { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 112 | /* |
| 113 | * All inode numbers of underlying fs should not be using the |
| 114 | * high xinobits, so we use high xinobits to partition the |
| 115 | * overlay st_ino address space. The high bits holds the fsid |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 116 | * (upper fsid is 0). The lowest xinobit is reserved for mapping |
| 117 | * the non-peresistent inode numbers range in case of overflow. |
| 118 | * This way all overlay inode numbers are unique and use the |
| 119 | * overlay st_dev. |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 120 | */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 121 | if (likely(!(stat->ino >> xinoshift))) { |
| 122 | stat->ino |= ((u64)fsid) << (xinoshift + 1); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 123 | stat->dev = dentry->d_sb->s_dev; |
| 124 | return 0; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 125 | } else if (ovl_xino_warn(dentry->d_sb)) { |
| 126 | pr_warn_ratelimited("inode number too big (%pd2, ino=%llu, xinobits=%d)\n", |
| 127 | dentry, stat->ino, xinobits); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 128 | } |
| 129 | } |
| 130 | |
| 131 | /* The inode could not be mapped to a unified st_ino address space */ |
| 132 | if (S_ISDIR(dentry->d_inode->i_mode)) { |
| 133 | /* |
| 134 | * Always use the overlay st_dev for directories, so 'find |
| 135 | * -xdev' will scan the entire overlay mount and won't cross the |
| 136 | * overlay mount boundaries. |
| 137 | * |
| 138 | * If not all layers are on the same fs the pair {real st_ino; |
| 139 | * overlay st_dev} is not unique, so use the non persistent |
| 140 | * overlay st_ino for directories. |
| 141 | */ |
| 142 | stat->dev = dentry->d_sb->s_dev; |
| 143 | stat->ino = dentry->d_inode->i_ino; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 144 | } else { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 145 | /* |
| 146 | * For non-samefs setup, if we cannot map all layers st_ino |
| 147 | * to a unified address space, we need to make sure that st_dev |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 148 | * is unique per underlying fs, so we use the unique anonymous |
| 149 | * bdev assigned to the underlying fs. |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 150 | */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 151 | stat->dev = OVL_FS(dentry->d_sb)->fs[fsid].pseudo_dev; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 152 | } |
| 153 | |
| 154 | return 0; |
| 155 | } |
| 156 | |
| 157 | int ovl_getattr(const struct path *path, struct kstat *stat, |
| 158 | u32 request_mask, unsigned int flags) |
| 159 | { |
| 160 | struct dentry *dentry = path->dentry; |
| 161 | enum ovl_path_type type; |
| 162 | struct path realpath; |
| 163 | const struct cred *old_cred; |
| 164 | bool is_dir = S_ISDIR(dentry->d_inode->i_mode); |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 165 | int fsid = 0; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 166 | int err; |
| 167 | bool metacopy_blocks = false; |
| 168 | |
| 169 | metacopy_blocks = ovl_is_metacopy_dentry(dentry); |
| 170 | |
| 171 | type = ovl_path_real(dentry, &realpath); |
| 172 | old_cred = ovl_override_creds(dentry->d_sb); |
| 173 | err = vfs_getattr(&realpath, stat, request_mask, flags); |
| 174 | if (err) |
| 175 | goto out; |
| 176 | |
| 177 | /* |
| 178 | * For non-dir or same fs, we use st_ino of the copy up origin. |
| 179 | * This guaranties constant st_dev/st_ino across copy up. |
| 180 | * With xino feature and non-samefs, we use st_ino of the copy up |
| 181 | * origin masked with high bits that represent the layer id. |
| 182 | * |
| 183 | * If lower filesystem supports NFS file handles, this also guaranties |
| 184 | * persistent st_ino across mount cycle. |
| 185 | */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 186 | if (!is_dir || ovl_same_dev(dentry->d_sb)) { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 187 | if (!OVL_TYPE_UPPER(type)) { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 188 | fsid = ovl_layer_lower(dentry)->fsid; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 189 | } else if (OVL_TYPE_ORIGIN(type)) { |
| 190 | struct kstat lowerstat; |
| 191 | u32 lowermask = STATX_INO | STATX_BLOCKS | |
| 192 | (!is_dir ? STATX_NLINK : 0); |
| 193 | |
| 194 | ovl_path_lower(dentry, &realpath); |
| 195 | err = vfs_getattr(&realpath, &lowerstat, |
| 196 | lowermask, flags); |
| 197 | if (err) |
| 198 | goto out; |
| 199 | |
| 200 | /* |
| 201 | * Lower hardlinks may be broken on copy up to different |
| 202 | * upper files, so we cannot use the lower origin st_ino |
| 203 | * for those different files, even for the same fs case. |
| 204 | * |
| 205 | * Similarly, several redirected dirs can point to the |
| 206 | * same dir on a lower layer. With the "verify_lower" |
| 207 | * feature, we do not use the lower origin st_ino, if |
| 208 | * we haven't verified that this redirect is unique. |
| 209 | * |
| 210 | * With inodes index enabled, it is safe to use st_ino |
| 211 | * of an indexed origin. The index validates that the |
| 212 | * upper hardlink is not broken and that a redirected |
| 213 | * dir is the only redirect to that origin. |
| 214 | */ |
| 215 | if (ovl_test_flag(OVL_INDEX, d_inode(dentry)) || |
| 216 | (!ovl_verify_lower(dentry->d_sb) && |
| 217 | (is_dir || lowerstat.nlink == 1))) { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 218 | fsid = ovl_layer_lower(dentry)->fsid; |
| 219 | stat->ino = lowerstat.ino; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 220 | } |
| 221 | |
| 222 | /* |
| 223 | * If we are querying a metacopy dentry and lower |
| 224 | * dentry is data dentry, then use the blocks we |
| 225 | * queried just now. We don't have to do additional |
| 226 | * vfs_getattr(). If lower itself is metacopy, then |
| 227 | * additional vfs_getattr() is unavoidable. |
| 228 | */ |
| 229 | if (metacopy_blocks && |
| 230 | realpath.dentry == ovl_dentry_lowerdata(dentry)) { |
| 231 | stat->blocks = lowerstat.blocks; |
| 232 | metacopy_blocks = false; |
| 233 | } |
| 234 | } |
| 235 | |
| 236 | if (metacopy_blocks) { |
| 237 | /* |
| 238 | * If lower is not same as lowerdata or if there was |
| 239 | * no origin on upper, we can end up here. |
| 240 | */ |
| 241 | struct kstat lowerdatastat; |
| 242 | u32 lowermask = STATX_BLOCKS; |
| 243 | |
| 244 | ovl_path_lowerdata(dentry, &realpath); |
| 245 | err = vfs_getattr(&realpath, &lowerdatastat, |
| 246 | lowermask, flags); |
| 247 | if (err) |
| 248 | goto out; |
| 249 | stat->blocks = lowerdatastat.blocks; |
| 250 | } |
| 251 | } |
| 252 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 253 | err = ovl_map_dev_ino(dentry, stat, fsid); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 254 | if (err) |
| 255 | goto out; |
| 256 | |
| 257 | /* |
| 258 | * It's probably not worth it to count subdirs to get the |
| 259 | * correct link count. nlink=1 seems to pacify 'find' and |
| 260 | * other utilities. |
| 261 | */ |
| 262 | if (is_dir && OVL_TYPE_MERGE(type)) |
| 263 | stat->nlink = 1; |
| 264 | |
| 265 | /* |
| 266 | * Return the overlay inode nlinks for indexed upper inodes. |
| 267 | * Overlay inode nlink counts the union of the upper hardlinks |
| 268 | * and non-covered lower hardlinks. It does not include the upper |
| 269 | * index hardlink. |
| 270 | */ |
| 271 | if (!is_dir && ovl_test_flag(OVL_INDEX, d_inode(dentry))) |
| 272 | stat->nlink = dentry->d_inode->i_nlink; |
| 273 | |
| 274 | out: |
| 275 | revert_creds(old_cred); |
| 276 | |
| 277 | return err; |
| 278 | } |
| 279 | |
| 280 | int ovl_permission(struct inode *inode, int mask) |
| 281 | { |
| 282 | struct inode *upperinode = ovl_inode_upper(inode); |
| 283 | struct inode *realinode = upperinode ?: ovl_inode_lower(inode); |
| 284 | const struct cred *old_cred; |
| 285 | int err; |
| 286 | |
| 287 | /* Careful in RCU walk mode */ |
| 288 | if (!realinode) { |
| 289 | WARN_ON(!(mask & MAY_NOT_BLOCK)); |
| 290 | return -ECHILD; |
| 291 | } |
| 292 | |
| 293 | /* |
| 294 | * Check overlay inode with the creds of task and underlying inode |
| 295 | * with creds of mounter |
| 296 | */ |
| 297 | err = generic_permission(inode, mask); |
| 298 | if (err) |
| 299 | return err; |
| 300 | |
| 301 | old_cred = ovl_override_creds(inode->i_sb); |
| 302 | if (!upperinode && |
| 303 | !special_file(realinode->i_mode) && mask & MAY_WRITE) { |
| 304 | mask &= ~(MAY_WRITE | MAY_APPEND); |
| 305 | /* Make sure mounter can read file for copy up later */ |
| 306 | mask |= MAY_READ; |
| 307 | } |
| 308 | err = inode_permission(realinode, mask); |
| 309 | revert_creds(old_cred); |
| 310 | |
| 311 | return err; |
| 312 | } |
| 313 | |
| 314 | static const char *ovl_get_link(struct dentry *dentry, |
| 315 | struct inode *inode, |
| 316 | struct delayed_call *done) |
| 317 | { |
| 318 | const struct cred *old_cred; |
| 319 | const char *p; |
| 320 | |
| 321 | if (!dentry) |
| 322 | return ERR_PTR(-ECHILD); |
| 323 | |
| 324 | old_cred = ovl_override_creds(dentry->d_sb); |
| 325 | p = vfs_get_link(ovl_dentry_real(dentry), done); |
| 326 | revert_creds(old_cred); |
| 327 | return p; |
| 328 | } |
| 329 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 330 | bool ovl_is_private_xattr(struct super_block *sb, const char *name) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 331 | { |
| 332 | return strncmp(name, OVL_XATTR_PREFIX, |
| 333 | sizeof(OVL_XATTR_PREFIX) - 1) == 0; |
| 334 | } |
| 335 | |
| 336 | int ovl_xattr_set(struct dentry *dentry, struct inode *inode, const char *name, |
| 337 | const void *value, size_t size, int flags) |
| 338 | { |
| 339 | int err; |
| 340 | struct dentry *upperdentry = ovl_i_dentry_upper(inode); |
| 341 | struct dentry *realdentry = upperdentry ?: ovl_dentry_lower(dentry); |
| 342 | const struct cred *old_cred; |
| 343 | |
| 344 | err = ovl_want_write(dentry); |
| 345 | if (err) |
| 346 | goto out; |
| 347 | |
| 348 | if (!value && !upperdentry) { |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 349 | old_cred = ovl_override_creds(dentry->d_sb); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 350 | err = vfs_getxattr(realdentry, name, NULL, 0); |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 351 | revert_creds(old_cred); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 352 | if (err < 0) |
| 353 | goto out_drop_write; |
| 354 | } |
| 355 | |
| 356 | if (!upperdentry) { |
| 357 | err = ovl_copy_up(dentry); |
| 358 | if (err) |
| 359 | goto out_drop_write; |
| 360 | |
| 361 | realdentry = ovl_dentry_upper(dentry); |
| 362 | } |
| 363 | |
| 364 | old_cred = ovl_override_creds(dentry->d_sb); |
| 365 | if (value) |
| 366 | err = vfs_setxattr(realdentry, name, value, size, flags); |
| 367 | else { |
| 368 | WARN_ON(flags != XATTR_REPLACE); |
| 369 | err = vfs_removexattr(realdentry, name); |
| 370 | } |
| 371 | revert_creds(old_cred); |
| 372 | |
| 373 | /* copy c/mtime */ |
| 374 | ovl_copyattr(d_inode(realdentry), inode); |
| 375 | |
| 376 | out_drop_write: |
| 377 | ovl_drop_write(dentry); |
| 378 | out: |
| 379 | return err; |
| 380 | } |
| 381 | |
| 382 | int ovl_xattr_get(struct dentry *dentry, struct inode *inode, const char *name, |
| 383 | void *value, size_t size) |
| 384 | { |
| 385 | ssize_t res; |
| 386 | const struct cred *old_cred; |
| 387 | struct dentry *realdentry = |
| 388 | ovl_i_dentry_upper(inode) ?: ovl_dentry_lower(dentry); |
| 389 | |
| 390 | old_cred = ovl_override_creds(dentry->d_sb); |
| 391 | res = vfs_getxattr(realdentry, name, value, size); |
| 392 | revert_creds(old_cred); |
| 393 | return res; |
| 394 | } |
| 395 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 396 | static bool ovl_can_list(struct super_block *sb, const char *s) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 397 | { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 398 | /* Never list private (.overlay) */ |
| 399 | if (ovl_is_private_xattr(sb, s)) |
| 400 | return false; |
| 401 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 402 | /* List all non-trusted xatts */ |
| 403 | if (strncmp(s, XATTR_TRUSTED_PREFIX, XATTR_TRUSTED_PREFIX_LEN) != 0) |
| 404 | return true; |
| 405 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 406 | /* list other trusted for superuser only */ |
| 407 | return ns_capable_noaudit(&init_user_ns, CAP_SYS_ADMIN); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 408 | } |
| 409 | |
| 410 | ssize_t ovl_listxattr(struct dentry *dentry, char *list, size_t size) |
| 411 | { |
| 412 | struct dentry *realdentry = ovl_dentry_real(dentry); |
| 413 | ssize_t res; |
| 414 | size_t len; |
| 415 | char *s; |
| 416 | const struct cred *old_cred; |
| 417 | |
| 418 | old_cred = ovl_override_creds(dentry->d_sb); |
| 419 | res = vfs_listxattr(realdentry, list, size); |
| 420 | revert_creds(old_cred); |
| 421 | if (res <= 0 || size == 0) |
| 422 | return res; |
| 423 | |
| 424 | /* filter out private xattrs */ |
| 425 | for (s = list, len = res; len;) { |
| 426 | size_t slen = strnlen(s, len) + 1; |
| 427 | |
| 428 | /* underlying fs providing us with an broken xattr list? */ |
| 429 | if (WARN_ON(slen > len)) |
| 430 | return -EIO; |
| 431 | |
| 432 | len -= slen; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 433 | if (!ovl_can_list(dentry->d_sb, s)) { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 434 | res -= slen; |
| 435 | memmove(s, s + slen, len); |
| 436 | } else { |
| 437 | s += slen; |
| 438 | } |
| 439 | } |
| 440 | |
| 441 | return res; |
| 442 | } |
| 443 | |
| 444 | struct posix_acl *ovl_get_acl(struct inode *inode, int type) |
| 445 | { |
| 446 | struct inode *realinode = ovl_inode_real(inode); |
| 447 | const struct cred *old_cred; |
| 448 | struct posix_acl *acl; |
| 449 | |
| 450 | if (!IS_ENABLED(CONFIG_FS_POSIX_ACL) || !IS_POSIXACL(realinode)) |
| 451 | return NULL; |
| 452 | |
| 453 | old_cred = ovl_override_creds(inode->i_sb); |
| 454 | acl = get_acl(realinode, type); |
| 455 | revert_creds(old_cred); |
| 456 | |
| 457 | return acl; |
| 458 | } |
| 459 | |
| 460 | int ovl_update_time(struct inode *inode, struct timespec64 *ts, int flags) |
| 461 | { |
| 462 | if (flags & S_ATIME) { |
| 463 | struct ovl_fs *ofs = inode->i_sb->s_fs_info; |
| 464 | struct path upperpath = { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 465 | .mnt = ovl_upper_mnt(ofs), |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 466 | .dentry = ovl_upperdentry_dereference(OVL_I(inode)), |
| 467 | }; |
| 468 | |
| 469 | if (upperpath.dentry) { |
| 470 | touch_atime(&upperpath); |
| 471 | inode->i_atime = d_inode(upperpath.dentry)->i_atime; |
| 472 | } |
| 473 | } |
| 474 | return 0; |
| 475 | } |
| 476 | |
| 477 | static int ovl_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, |
| 478 | u64 start, u64 len) |
| 479 | { |
| 480 | int err; |
| 481 | struct inode *realinode = ovl_inode_real(inode); |
| 482 | const struct cred *old_cred; |
| 483 | |
| 484 | if (!realinode->i_op->fiemap) |
| 485 | return -EOPNOTSUPP; |
| 486 | |
| 487 | old_cred = ovl_override_creds(inode->i_sb); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 488 | err = realinode->i_op->fiemap(realinode, fieinfo, start, len); |
| 489 | revert_creds(old_cred); |
| 490 | |
| 491 | return err; |
| 492 | } |
| 493 | |
| 494 | static const struct inode_operations ovl_file_inode_operations = { |
| 495 | .setattr = ovl_setattr, |
| 496 | .permission = ovl_permission, |
| 497 | .getattr = ovl_getattr, |
| 498 | .listxattr = ovl_listxattr, |
| 499 | .get_acl = ovl_get_acl, |
| 500 | .update_time = ovl_update_time, |
| 501 | .fiemap = ovl_fiemap, |
| 502 | }; |
| 503 | |
| 504 | static const struct inode_operations ovl_symlink_inode_operations = { |
| 505 | .setattr = ovl_setattr, |
| 506 | .get_link = ovl_get_link, |
| 507 | .getattr = ovl_getattr, |
| 508 | .listxattr = ovl_listxattr, |
| 509 | .update_time = ovl_update_time, |
| 510 | }; |
| 511 | |
| 512 | static const struct inode_operations ovl_special_inode_operations = { |
| 513 | .setattr = ovl_setattr, |
| 514 | .permission = ovl_permission, |
| 515 | .getattr = ovl_getattr, |
| 516 | .listxattr = ovl_listxattr, |
| 517 | .get_acl = ovl_get_acl, |
| 518 | .update_time = ovl_update_time, |
| 519 | }; |
| 520 | |
| 521 | static const struct address_space_operations ovl_aops = { |
| 522 | /* For O_DIRECT dentry_open() checks f_mapping->a_ops->direct_IO */ |
| 523 | .direct_IO = noop_direct_IO, |
| 524 | }; |
| 525 | |
| 526 | /* |
| 527 | * It is possible to stack overlayfs instance on top of another |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 528 | * overlayfs instance as lower layer. We need to annotate the |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 529 | * stackable i_mutex locks according to stack level of the super |
| 530 | * block instance. An overlayfs instance can never be in stack |
| 531 | * depth 0 (there is always a real fs below it). An overlayfs |
| 532 | * inode lock will use the lockdep annotaion ovl_i_mutex_key[depth]. |
| 533 | * |
| 534 | * For example, here is a snip from /proc/lockdep_chains after |
| 535 | * dir_iterate of nested overlayfs: |
| 536 | * |
| 537 | * [...] &ovl_i_mutex_dir_key[depth] (stack_depth=2) |
| 538 | * [...] &ovl_i_mutex_dir_key[depth]#2 (stack_depth=1) |
| 539 | * [...] &type->i_mutex_dir_key (stack_depth=0) |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 540 | * |
| 541 | * Locking order w.r.t ovl_want_write() is important for nested overlayfs. |
| 542 | * |
| 543 | * This chain is valid: |
| 544 | * - inode->i_rwsem (inode_lock[2]) |
| 545 | * - upper_mnt->mnt_sb->s_writers (ovl_want_write[0]) |
| 546 | * - OVL_I(inode)->lock (ovl_inode_lock[2]) |
| 547 | * - OVL_I(lowerinode)->lock (ovl_inode_lock[1]) |
| 548 | * |
| 549 | * And this chain is valid: |
| 550 | * - inode->i_rwsem (inode_lock[2]) |
| 551 | * - OVL_I(inode)->lock (ovl_inode_lock[2]) |
| 552 | * - lowerinode->i_rwsem (inode_lock[1]) |
| 553 | * - OVL_I(lowerinode)->lock (ovl_inode_lock[1]) |
| 554 | * |
| 555 | * But lowerinode->i_rwsem SHOULD NOT be acquired while ovl_want_write() is |
| 556 | * held, because it is in reverse order of the non-nested case using the same |
| 557 | * upper fs: |
| 558 | * - inode->i_rwsem (inode_lock[1]) |
| 559 | * - upper_mnt->mnt_sb->s_writers (ovl_want_write[0]) |
| 560 | * - OVL_I(inode)->lock (ovl_inode_lock[1]) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 561 | */ |
| 562 | #define OVL_MAX_NESTING FILESYSTEM_MAX_STACK_DEPTH |
| 563 | |
| 564 | static inline void ovl_lockdep_annotate_inode_mutex_key(struct inode *inode) |
| 565 | { |
| 566 | #ifdef CONFIG_LOCKDEP |
| 567 | static struct lock_class_key ovl_i_mutex_key[OVL_MAX_NESTING]; |
| 568 | static struct lock_class_key ovl_i_mutex_dir_key[OVL_MAX_NESTING]; |
| 569 | static struct lock_class_key ovl_i_lock_key[OVL_MAX_NESTING]; |
| 570 | |
| 571 | int depth = inode->i_sb->s_stack_depth - 1; |
| 572 | |
| 573 | if (WARN_ON_ONCE(depth < 0 || depth >= OVL_MAX_NESTING)) |
| 574 | depth = 0; |
| 575 | |
| 576 | if (S_ISDIR(inode->i_mode)) |
| 577 | lockdep_set_class(&inode->i_rwsem, &ovl_i_mutex_dir_key[depth]); |
| 578 | else |
| 579 | lockdep_set_class(&inode->i_rwsem, &ovl_i_mutex_key[depth]); |
| 580 | |
| 581 | lockdep_set_class(&OVL_I(inode)->lock, &ovl_i_lock_key[depth]); |
| 582 | #endif |
| 583 | } |
| 584 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 585 | static void ovl_next_ino(struct inode *inode) |
| 586 | { |
| 587 | struct ovl_fs *ofs = inode->i_sb->s_fs_info; |
| 588 | |
| 589 | inode->i_ino = atomic_long_inc_return(&ofs->last_ino); |
| 590 | if (unlikely(!inode->i_ino)) |
| 591 | inode->i_ino = atomic_long_inc_return(&ofs->last_ino); |
| 592 | } |
| 593 | |
| 594 | static void ovl_map_ino(struct inode *inode, unsigned long ino, int fsid) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 595 | { |
| 596 | int xinobits = ovl_xino_bits(inode->i_sb); |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 597 | unsigned int xinoshift = 64 - xinobits; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 598 | |
| 599 | /* |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 600 | * When d_ino is consistent with st_ino (samefs or i_ino has enough |
| 601 | * bits to encode layer), set the same value used for st_ino to i_ino, |
| 602 | * so inode number exposed via /proc/locks and a like will be |
| 603 | * consistent with d_ino and st_ino values. An i_ino value inconsistent |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 604 | * with d_ino also causes nfsd readdirplus to fail. |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 605 | */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 606 | inode->i_ino = ino; |
| 607 | if (ovl_same_fs(inode->i_sb)) { |
| 608 | return; |
| 609 | } else if (xinobits && likely(!(ino >> xinoshift))) { |
| 610 | inode->i_ino |= (unsigned long)fsid << (xinoshift + 1); |
| 611 | return; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 612 | } |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 613 | |
| 614 | /* |
| 615 | * For directory inodes on non-samefs with xino disabled or xino |
| 616 | * overflow, we allocate a non-persistent inode number, to be used for |
| 617 | * resolving st_ino collisions in ovl_map_dev_ino(). |
| 618 | * |
| 619 | * To avoid ino collision with legitimate xino values from upper |
| 620 | * layer (fsid 0), use the lowest xinobit to map the non |
| 621 | * persistent inode numbers to the unified st_ino address space. |
| 622 | */ |
| 623 | if (S_ISDIR(inode->i_mode)) { |
| 624 | ovl_next_ino(inode); |
| 625 | if (xinobits) { |
| 626 | inode->i_ino &= ~0UL >> xinobits; |
| 627 | inode->i_ino |= 1UL << xinoshift; |
| 628 | } |
| 629 | } |
| 630 | } |
| 631 | |
| 632 | void ovl_inode_init(struct inode *inode, struct ovl_inode_params *oip, |
| 633 | unsigned long ino, int fsid) |
| 634 | { |
| 635 | struct inode *realinode; |
| 636 | |
| 637 | if (oip->upperdentry) |
| 638 | OVL_I(inode)->__upperdentry = oip->upperdentry; |
| 639 | if (oip->lowerpath && oip->lowerpath->dentry) |
| 640 | OVL_I(inode)->lower = igrab(d_inode(oip->lowerpath->dentry)); |
| 641 | if (oip->lowerdata) |
| 642 | OVL_I(inode)->lowerdata = igrab(d_inode(oip->lowerdata)); |
| 643 | |
| 644 | realinode = ovl_inode_real(inode); |
| 645 | ovl_copyattr(realinode, inode); |
| 646 | ovl_copyflags(realinode, inode); |
| 647 | ovl_map_ino(inode, ino, fsid); |
| 648 | } |
| 649 | |
| 650 | static void ovl_fill_inode(struct inode *inode, umode_t mode, dev_t rdev) |
| 651 | { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 652 | inode->i_mode = mode; |
| 653 | inode->i_flags |= S_NOCMTIME; |
| 654 | #ifdef CONFIG_FS_POSIX_ACL |
| 655 | inode->i_acl = inode->i_default_acl = ACL_DONT_CACHE; |
| 656 | #endif |
| 657 | |
| 658 | ovl_lockdep_annotate_inode_mutex_key(inode); |
| 659 | |
| 660 | switch (mode & S_IFMT) { |
| 661 | case S_IFREG: |
| 662 | inode->i_op = &ovl_file_inode_operations; |
| 663 | inode->i_fop = &ovl_file_operations; |
| 664 | inode->i_mapping->a_ops = &ovl_aops; |
| 665 | break; |
| 666 | |
| 667 | case S_IFDIR: |
| 668 | inode->i_op = &ovl_dir_inode_operations; |
| 669 | inode->i_fop = &ovl_dir_operations; |
| 670 | break; |
| 671 | |
| 672 | case S_IFLNK: |
| 673 | inode->i_op = &ovl_symlink_inode_operations; |
| 674 | break; |
| 675 | |
| 676 | default: |
| 677 | inode->i_op = &ovl_special_inode_operations; |
| 678 | init_special_inode(inode, mode, rdev); |
| 679 | break; |
| 680 | } |
| 681 | } |
| 682 | |
| 683 | /* |
| 684 | * With inodes index enabled, an overlay inode nlink counts the union of upper |
| 685 | * hardlinks and non-covered lower hardlinks. During the lifetime of a non-pure |
| 686 | * upper inode, the following nlink modifying operations can happen: |
| 687 | * |
| 688 | * 1. Lower hardlink copy up |
| 689 | * 2. Upper hardlink created, unlinked or renamed over |
| 690 | * 3. Lower hardlink whiteout or renamed over |
| 691 | * |
| 692 | * For the first, copy up case, the union nlink does not change, whether the |
| 693 | * operation succeeds or fails, but the upper inode nlink may change. |
| 694 | * Therefore, before copy up, we store the union nlink value relative to the |
| 695 | * lower inode nlink in the index inode xattr trusted.overlay.nlink. |
| 696 | * |
| 697 | * For the second, upper hardlink case, the union nlink should be incremented |
| 698 | * or decremented IFF the operation succeeds, aligned with nlink change of the |
| 699 | * upper inode. Therefore, before link/unlink/rename, we store the union nlink |
| 700 | * value relative to the upper inode nlink in the index inode. |
| 701 | * |
| 702 | * For the last, lower cover up case, we simplify things by preceding the |
| 703 | * whiteout or cover up with copy up. This makes sure that there is an index |
| 704 | * upper inode where the nlink xattr can be stored before the copied up upper |
| 705 | * entry is unlink. |
| 706 | */ |
| 707 | #define OVL_NLINK_ADD_UPPER (1 << 0) |
| 708 | |
| 709 | /* |
| 710 | * On-disk format for indexed nlink: |
| 711 | * |
| 712 | * nlink relative to the upper inode - "U[+-]NUM" |
| 713 | * nlink relative to the lower inode - "L[+-]NUM" |
| 714 | */ |
| 715 | |
| 716 | static int ovl_set_nlink_common(struct dentry *dentry, |
| 717 | struct dentry *realdentry, const char *format) |
| 718 | { |
| 719 | struct inode *inode = d_inode(dentry); |
| 720 | struct inode *realinode = d_inode(realdentry); |
| 721 | char buf[13]; |
| 722 | int len; |
| 723 | |
| 724 | len = snprintf(buf, sizeof(buf), format, |
| 725 | (int) (inode->i_nlink - realinode->i_nlink)); |
| 726 | |
| 727 | if (WARN_ON(len >= sizeof(buf))) |
| 728 | return -EIO; |
| 729 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 730 | return ovl_do_setxattr(OVL_FS(inode->i_sb), ovl_dentry_upper(dentry), |
| 731 | OVL_XATTR_NLINK, buf, len); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 732 | } |
| 733 | |
| 734 | int ovl_set_nlink_upper(struct dentry *dentry) |
| 735 | { |
| 736 | return ovl_set_nlink_common(dentry, ovl_dentry_upper(dentry), "U%+i"); |
| 737 | } |
| 738 | |
| 739 | int ovl_set_nlink_lower(struct dentry *dentry) |
| 740 | { |
| 741 | return ovl_set_nlink_common(dentry, ovl_dentry_lower(dentry), "L%+i"); |
| 742 | } |
| 743 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 744 | unsigned int ovl_get_nlink(struct ovl_fs *ofs, struct dentry *lowerdentry, |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 745 | struct dentry *upperdentry, |
| 746 | unsigned int fallback) |
| 747 | { |
| 748 | int nlink_diff; |
| 749 | int nlink; |
| 750 | char buf[13]; |
| 751 | int err; |
| 752 | |
| 753 | if (!lowerdentry || !upperdentry || d_inode(lowerdentry)->i_nlink == 1) |
| 754 | return fallback; |
| 755 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 756 | err = ovl_do_getxattr(ofs, upperdentry, OVL_XATTR_NLINK, |
| 757 | &buf, sizeof(buf) - 1); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 758 | if (err < 0) |
| 759 | goto fail; |
| 760 | |
| 761 | buf[err] = '\0'; |
| 762 | if ((buf[0] != 'L' && buf[0] != 'U') || |
| 763 | (buf[1] != '+' && buf[1] != '-')) |
| 764 | goto fail; |
| 765 | |
| 766 | err = kstrtoint(buf + 1, 10, &nlink_diff); |
| 767 | if (err < 0) |
| 768 | goto fail; |
| 769 | |
| 770 | nlink = d_inode(buf[0] == 'L' ? lowerdentry : upperdentry)->i_nlink; |
| 771 | nlink += nlink_diff; |
| 772 | |
| 773 | if (nlink <= 0) |
| 774 | goto fail; |
| 775 | |
| 776 | return nlink; |
| 777 | |
| 778 | fail: |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 779 | pr_warn_ratelimited("failed to get index nlink (%pd2, err=%i)\n", |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 780 | upperdentry, err); |
| 781 | return fallback; |
| 782 | } |
| 783 | |
| 784 | struct inode *ovl_new_inode(struct super_block *sb, umode_t mode, dev_t rdev) |
| 785 | { |
| 786 | struct inode *inode; |
| 787 | |
| 788 | inode = new_inode(sb); |
| 789 | if (inode) |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 790 | ovl_fill_inode(inode, mode, rdev); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 791 | |
| 792 | return inode; |
| 793 | } |
| 794 | |
| 795 | static int ovl_inode_test(struct inode *inode, void *data) |
| 796 | { |
| 797 | return inode->i_private == data; |
| 798 | } |
| 799 | |
| 800 | static int ovl_inode_set(struct inode *inode, void *data) |
| 801 | { |
| 802 | inode->i_private = data; |
| 803 | return 0; |
| 804 | } |
| 805 | |
| 806 | static bool ovl_verify_inode(struct inode *inode, struct dentry *lowerdentry, |
| 807 | struct dentry *upperdentry, bool strict) |
| 808 | { |
| 809 | /* |
| 810 | * For directories, @strict verify from lookup path performs consistency |
| 811 | * checks, so NULL lower/upper in dentry must match NULL lower/upper in |
| 812 | * inode. Non @strict verify from NFS handle decode path passes NULL for |
| 813 | * 'unknown' lower/upper. |
| 814 | */ |
| 815 | if (S_ISDIR(inode->i_mode) && strict) { |
| 816 | /* Real lower dir moved to upper layer under us? */ |
| 817 | if (!lowerdentry && ovl_inode_lower(inode)) |
| 818 | return false; |
| 819 | |
| 820 | /* Lookup of an uncovered redirect origin? */ |
| 821 | if (!upperdentry && ovl_inode_upper(inode)) |
| 822 | return false; |
| 823 | } |
| 824 | |
| 825 | /* |
| 826 | * Allow non-NULL lower inode in ovl_inode even if lowerdentry is NULL. |
| 827 | * This happens when finding a copied up overlay inode for a renamed |
| 828 | * or hardlinked overlay dentry and lower dentry cannot be followed |
| 829 | * by origin because lower fs does not support file handles. |
| 830 | */ |
| 831 | if (lowerdentry && ovl_inode_lower(inode) != d_inode(lowerdentry)) |
| 832 | return false; |
| 833 | |
| 834 | /* |
| 835 | * Allow non-NULL __upperdentry in inode even if upperdentry is NULL. |
| 836 | * This happens when finding a lower alias for a copied up hard link. |
| 837 | */ |
| 838 | if (upperdentry && ovl_inode_upper(inode) != d_inode(upperdentry)) |
| 839 | return false; |
| 840 | |
| 841 | return true; |
| 842 | } |
| 843 | |
| 844 | struct inode *ovl_lookup_inode(struct super_block *sb, struct dentry *real, |
| 845 | bool is_upper) |
| 846 | { |
| 847 | struct inode *inode, *key = d_inode(real); |
| 848 | |
| 849 | inode = ilookup5(sb, (unsigned long) key, ovl_inode_test, key); |
| 850 | if (!inode) |
| 851 | return NULL; |
| 852 | |
| 853 | if (!ovl_verify_inode(inode, is_upper ? NULL : real, |
| 854 | is_upper ? real : NULL, false)) { |
| 855 | iput(inode); |
| 856 | return ERR_PTR(-ESTALE); |
| 857 | } |
| 858 | |
| 859 | return inode; |
| 860 | } |
| 861 | |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 862 | bool ovl_lookup_trap_inode(struct super_block *sb, struct dentry *dir) |
| 863 | { |
| 864 | struct inode *key = d_inode(dir); |
| 865 | struct inode *trap; |
| 866 | bool res; |
| 867 | |
| 868 | trap = ilookup5(sb, (unsigned long) key, ovl_inode_test, key); |
| 869 | if (!trap) |
| 870 | return false; |
| 871 | |
| 872 | res = IS_DEADDIR(trap) && !ovl_inode_upper(trap) && |
| 873 | !ovl_inode_lower(trap); |
| 874 | |
| 875 | iput(trap); |
| 876 | return res; |
| 877 | } |
| 878 | |
| 879 | /* |
| 880 | * Create an inode cache entry for layer root dir, that will intentionally |
| 881 | * fail ovl_verify_inode(), so any lookup that will find some layer root |
| 882 | * will fail. |
| 883 | */ |
| 884 | struct inode *ovl_get_trap_inode(struct super_block *sb, struct dentry *dir) |
| 885 | { |
| 886 | struct inode *key = d_inode(dir); |
| 887 | struct inode *trap; |
| 888 | |
| 889 | if (!d_is_dir(dir)) |
| 890 | return ERR_PTR(-ENOTDIR); |
| 891 | |
| 892 | trap = iget5_locked(sb, (unsigned long) key, ovl_inode_test, |
| 893 | ovl_inode_set, key); |
| 894 | if (!trap) |
| 895 | return ERR_PTR(-ENOMEM); |
| 896 | |
| 897 | if (!(trap->i_state & I_NEW)) { |
| 898 | /* Conflicting layer roots? */ |
| 899 | iput(trap); |
| 900 | return ERR_PTR(-ELOOP); |
| 901 | } |
| 902 | |
| 903 | trap->i_mode = S_IFDIR; |
| 904 | trap->i_flags = S_DEAD; |
| 905 | unlock_new_inode(trap); |
| 906 | |
| 907 | return trap; |
| 908 | } |
| 909 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 910 | /* |
| 911 | * Does overlay inode need to be hashed by lower inode? |
| 912 | */ |
| 913 | static bool ovl_hash_bylower(struct super_block *sb, struct dentry *upper, |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 914 | struct dentry *lower, bool index) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 915 | { |
| 916 | struct ovl_fs *ofs = sb->s_fs_info; |
| 917 | |
| 918 | /* No, if pure upper */ |
| 919 | if (!lower) |
| 920 | return false; |
| 921 | |
| 922 | /* Yes, if already indexed */ |
| 923 | if (index) |
| 924 | return true; |
| 925 | |
| 926 | /* Yes, if won't be copied up */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 927 | if (!ovl_upper_mnt(ofs)) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 928 | return true; |
| 929 | |
| 930 | /* No, if lower hardlink is or will be broken on copy up */ |
| 931 | if ((upper || !ovl_indexdir(sb)) && |
| 932 | !d_is_dir(lower) && d_inode(lower)->i_nlink > 1) |
| 933 | return false; |
| 934 | |
| 935 | /* No, if non-indexed upper with NFS export */ |
| 936 | if (sb->s_export_op && upper) |
| 937 | return false; |
| 938 | |
| 939 | /* Otherwise, hash by lower inode for fsnotify */ |
| 940 | return true; |
| 941 | } |
| 942 | |
| 943 | static struct inode *ovl_iget5(struct super_block *sb, struct inode *newinode, |
| 944 | struct inode *key) |
| 945 | { |
| 946 | return newinode ? inode_insert5(newinode, (unsigned long) key, |
| 947 | ovl_inode_test, ovl_inode_set, key) : |
| 948 | iget5_locked(sb, (unsigned long) key, |
| 949 | ovl_inode_test, ovl_inode_set, key); |
| 950 | } |
| 951 | |
| 952 | struct inode *ovl_get_inode(struct super_block *sb, |
| 953 | struct ovl_inode_params *oip) |
| 954 | { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 955 | struct ovl_fs *ofs = OVL_FS(sb); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 956 | struct dentry *upperdentry = oip->upperdentry; |
| 957 | struct ovl_path *lowerpath = oip->lowerpath; |
| 958 | struct inode *realinode = upperdentry ? d_inode(upperdentry) : NULL; |
| 959 | struct inode *inode; |
| 960 | struct dentry *lowerdentry = lowerpath ? lowerpath->dentry : NULL; |
| 961 | bool bylower = ovl_hash_bylower(sb, upperdentry, lowerdentry, |
| 962 | oip->index); |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 963 | int fsid = bylower ? lowerpath->layer->fsid : 0; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 964 | bool is_dir; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 965 | unsigned long ino = 0; |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 966 | int err = oip->newinode ? -EEXIST : -ENOMEM; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 967 | |
| 968 | if (!realinode) |
| 969 | realinode = d_inode(lowerdentry); |
| 970 | |
| 971 | /* |
| 972 | * Copy up origin (lower) may exist for non-indexed upper, but we must |
| 973 | * not use lower as hash key if this is a broken hardlink. |
| 974 | */ |
| 975 | is_dir = S_ISDIR(realinode->i_mode); |
| 976 | if (upperdentry || bylower) { |
| 977 | struct inode *key = d_inode(bylower ? lowerdentry : |
| 978 | upperdentry); |
| 979 | unsigned int nlink = is_dir ? 1 : realinode->i_nlink; |
| 980 | |
| 981 | inode = ovl_iget5(sb, oip->newinode, key); |
| 982 | if (!inode) |
| 983 | goto out_err; |
| 984 | if (!(inode->i_state & I_NEW)) { |
| 985 | /* |
| 986 | * Verify that the underlying files stored in the inode |
| 987 | * match those in the dentry. |
| 988 | */ |
| 989 | if (!ovl_verify_inode(inode, lowerdentry, upperdentry, |
| 990 | true)) { |
| 991 | iput(inode); |
| 992 | err = -ESTALE; |
| 993 | goto out_err; |
| 994 | } |
| 995 | |
| 996 | dput(upperdentry); |
| 997 | kfree(oip->redirect); |
| 998 | goto out; |
| 999 | } |
| 1000 | |
| 1001 | /* Recalculate nlink for non-dir due to indexing */ |
| 1002 | if (!is_dir) |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1003 | nlink = ovl_get_nlink(ofs, lowerdentry, upperdentry, |
| 1004 | nlink); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1005 | set_nlink(inode, nlink); |
| 1006 | ino = key->i_ino; |
| 1007 | } else { |
| 1008 | /* Lower hardlink that will be broken on copy up */ |
| 1009 | inode = new_inode(sb); |
| 1010 | if (!inode) { |
| 1011 | err = -ENOMEM; |
| 1012 | goto out_err; |
| 1013 | } |
Olivier Deprez | 0e64123 | 2021-09-23 10:07:05 +0200 | [diff] [blame] | 1014 | ino = realinode->i_ino; |
| 1015 | fsid = lowerpath->layer->fsid; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1016 | } |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1017 | ovl_fill_inode(inode, realinode->i_mode, realinode->i_rdev); |
| 1018 | ovl_inode_init(inode, oip, ino, fsid); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1019 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1020 | if (upperdentry && ovl_is_impuredir(sb, upperdentry)) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1021 | ovl_set_flag(OVL_IMPURE, inode); |
| 1022 | |
| 1023 | if (oip->index) |
| 1024 | ovl_set_flag(OVL_INDEX, inode); |
| 1025 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1026 | OVL_I(inode)->redirect = oip->redirect; |
| 1027 | |
| 1028 | if (bylower) |
| 1029 | ovl_set_flag(OVL_CONST_INO, inode); |
| 1030 | |
| 1031 | /* Check for non-merge dir that may have whiteouts */ |
| 1032 | if (is_dir) { |
| 1033 | if (((upperdentry && lowerdentry) || oip->numlower > 1) || |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1034 | ovl_check_origin_xattr(ofs, upperdentry ?: lowerdentry)) { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1035 | ovl_set_flag(OVL_WHITEOUTS, inode); |
| 1036 | } |
| 1037 | } |
| 1038 | |
| 1039 | if (inode->i_state & I_NEW) |
| 1040 | unlock_new_inode(inode); |
| 1041 | out: |
| 1042 | return inode; |
| 1043 | |
| 1044 | out_err: |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1045 | pr_warn_ratelimited("failed to get inode (%i)\n", err); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1046 | inode = ERR_PTR(err); |
| 1047 | goto out; |
| 1048 | } |