David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame^] | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
| 2 | |
| 3 | #ifndef BTRFS_SPACE_INFO_H |
| 4 | #define BTRFS_SPACE_INFO_H |
| 5 | |
| 6 | struct btrfs_space_info { |
| 7 | spinlock_t lock; |
| 8 | |
| 9 | u64 total_bytes; /* total bytes in the space, |
| 10 | this doesn't take mirrors into account */ |
| 11 | u64 bytes_used; /* total bytes used, |
| 12 | this doesn't take mirrors into account */ |
| 13 | u64 bytes_pinned; /* total bytes pinned, will be freed when the |
| 14 | transaction finishes */ |
| 15 | u64 bytes_reserved; /* total bytes the allocator has reserved for |
| 16 | current allocations */ |
| 17 | u64 bytes_may_use; /* number of bytes that may be used for |
| 18 | delalloc/allocations */ |
| 19 | u64 bytes_readonly; /* total bytes that are read only */ |
| 20 | |
| 21 | u64 max_extent_size; /* This will hold the maximum extent size of |
| 22 | the space info if we had an ENOSPC in the |
| 23 | allocator. */ |
| 24 | |
| 25 | unsigned int full:1; /* indicates that we cannot allocate any more |
| 26 | chunks for this space */ |
| 27 | unsigned int chunk_alloc:1; /* set if we are allocating a chunk */ |
| 28 | |
| 29 | unsigned int flush:1; /* set if we are trying to make space */ |
| 30 | |
| 31 | unsigned int force_alloc; /* set if we need to force a chunk |
| 32 | alloc for this space */ |
| 33 | |
| 34 | u64 disk_used; /* total bytes used on disk */ |
| 35 | u64 disk_total; /* total bytes on disk, takes mirrors into |
| 36 | account */ |
| 37 | |
| 38 | u64 flags; |
| 39 | |
| 40 | /* |
| 41 | * bytes_pinned is kept in line with what is actually pinned, as in |
| 42 | * we've called update_block_group and dropped the bytes_used counter |
| 43 | * and increased the bytes_pinned counter. However this means that |
| 44 | * bytes_pinned does not reflect the bytes that will be pinned once the |
| 45 | * delayed refs are flushed, so this counter is inc'ed every time we |
| 46 | * call btrfs_free_extent so it is a realtime count of what will be |
| 47 | * freed once the transaction is committed. It will be zeroed every |
| 48 | * time the transaction commits. |
| 49 | */ |
| 50 | struct percpu_counter total_bytes_pinned; |
| 51 | |
| 52 | struct list_head list; |
| 53 | /* Protected by the spinlock 'lock'. */ |
| 54 | struct list_head ro_bgs; |
| 55 | struct list_head priority_tickets; |
| 56 | struct list_head tickets; |
| 57 | /* |
| 58 | * tickets_id just indicates the next ticket will be handled, so note |
| 59 | * it's not stored per ticket. |
| 60 | */ |
| 61 | u64 tickets_id; |
| 62 | |
| 63 | struct rw_semaphore groups_sem; |
| 64 | /* for block groups in our same type */ |
| 65 | struct list_head block_groups[BTRFS_NR_RAID_TYPES]; |
| 66 | wait_queue_head_t wait; |
| 67 | |
| 68 | struct kobject kobj; |
| 69 | struct kobject *block_group_kobjs[BTRFS_NR_RAID_TYPES]; |
| 70 | }; |
| 71 | |
| 72 | struct reserve_ticket { |
| 73 | u64 bytes; |
| 74 | int error; |
| 75 | struct list_head list; |
| 76 | wait_queue_head_t wait; |
| 77 | }; |
| 78 | |
| 79 | static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) |
| 80 | { |
| 81 | return ((space_info->flags & BTRFS_BLOCK_GROUP_METADATA) && |
| 82 | (space_info->flags & BTRFS_BLOCK_GROUP_DATA)); |
| 83 | } |
| 84 | |
| 85 | /* |
| 86 | * |
| 87 | * Declare a helper function to detect underflow of various space info members |
| 88 | */ |
| 89 | #define DECLARE_SPACE_INFO_UPDATE(name, trace_name) \ |
| 90 | static inline void \ |
| 91 | btrfs_space_info_update_##name(struct btrfs_fs_info *fs_info, \ |
| 92 | struct btrfs_space_info *sinfo, \ |
| 93 | s64 bytes) \ |
| 94 | { \ |
| 95 | const u64 abs_bytes = (bytes < 0) ? -bytes : bytes; \ |
| 96 | lockdep_assert_held(&sinfo->lock); \ |
| 97 | trace_update_##name(fs_info, sinfo, sinfo->name, bytes); \ |
| 98 | trace_btrfs_space_reservation(fs_info, trace_name, \ |
| 99 | sinfo->flags, abs_bytes, \ |
| 100 | bytes > 0); \ |
| 101 | if (bytes < 0 && sinfo->name < -bytes) { \ |
| 102 | WARN_ON(1); \ |
| 103 | sinfo->name = 0; \ |
| 104 | return; \ |
| 105 | } \ |
| 106 | sinfo->name += bytes; \ |
| 107 | } |
| 108 | |
| 109 | DECLARE_SPACE_INFO_UPDATE(bytes_may_use, "space_info"); |
| 110 | DECLARE_SPACE_INFO_UPDATE(bytes_pinned, "pinned"); |
| 111 | |
| 112 | int btrfs_init_space_info(struct btrfs_fs_info *fs_info); |
| 113 | void btrfs_update_space_info(struct btrfs_fs_info *info, u64 flags, |
| 114 | u64 total_bytes, u64 bytes_used, |
| 115 | u64 bytes_readonly, |
| 116 | struct btrfs_space_info **space_info); |
| 117 | struct btrfs_space_info *btrfs_find_space_info(struct btrfs_fs_info *info, |
| 118 | u64 flags); |
| 119 | u64 btrfs_space_info_used(struct btrfs_space_info *s_info, |
| 120 | bool may_use_included); |
| 121 | void btrfs_clear_space_info_full(struct btrfs_fs_info *info); |
| 122 | void btrfs_dump_space_info(struct btrfs_fs_info *fs_info, |
| 123 | struct btrfs_space_info *info, u64 bytes, |
| 124 | int dump_block_groups); |
| 125 | int btrfs_reserve_metadata_bytes(struct btrfs_root *root, |
| 126 | struct btrfs_block_rsv *block_rsv, |
| 127 | u64 orig_bytes, |
| 128 | enum btrfs_reserve_flush_enum flush); |
| 129 | void btrfs_try_granting_tickets(struct btrfs_fs_info *fs_info, |
| 130 | struct btrfs_space_info *space_info); |
| 131 | |
| 132 | static inline void btrfs_space_info_free_bytes_may_use( |
| 133 | struct btrfs_fs_info *fs_info, |
| 134 | struct btrfs_space_info *space_info, |
| 135 | u64 num_bytes) |
| 136 | { |
| 137 | spin_lock(&space_info->lock); |
| 138 | btrfs_space_info_update_bytes_may_use(fs_info, space_info, -num_bytes); |
| 139 | btrfs_try_granting_tickets(fs_info, space_info); |
| 140 | spin_unlock(&space_info->lock); |
| 141 | } |
| 142 | |
| 143 | #endif /* BTRFS_SPACE_INFO_H */ |