@@ -1774,7 +1774,7 @@ static noinline_for_stack bool lock_extent_buffer_for_io(struct extent_buffer *e
1774
1774
*/
1775
1775
spin_lock (& eb -> refs_lock );
1776
1776
if (test_and_clear_bit (EXTENT_BUFFER_DIRTY , & eb -> bflags )) {
1777
- XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> sectorsize_bits );
1777
+ XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> node_bits );
1778
1778
unsigned long flags ;
1779
1779
1780
1780
set_bit (EXTENT_BUFFER_WRITEBACK , & eb -> bflags );
@@ -1874,7 +1874,7 @@ static void set_btree_ioerr(struct extent_buffer *eb)
1874
1874
static void buffer_tree_set_mark (const struct extent_buffer * eb , xa_mark_t mark )
1875
1875
{
1876
1876
struct btrfs_fs_info * fs_info = eb -> fs_info ;
1877
- XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> sectorsize_bits );
1877
+ XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> node_bits );
1878
1878
unsigned long flags ;
1879
1879
1880
1880
xas_lock_irqsave (& xas , flags );
@@ -1886,7 +1886,7 @@ static void buffer_tree_set_mark(const struct extent_buffer *eb, xa_mark_t mark)
1886
1886
static void buffer_tree_clear_mark (const struct extent_buffer * eb , xa_mark_t mark )
1887
1887
{
1888
1888
struct btrfs_fs_info * fs_info = eb -> fs_info ;
1889
- XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> sectorsize_bits );
1889
+ XA_STATE (xas , & fs_info -> buffer_tree , eb -> start >> fs_info -> node_bits );
1890
1890
unsigned long flags ;
1891
1891
1892
1892
xas_lock_irqsave (& xas , flags );
@@ -1986,7 +1986,7 @@ static unsigned int buffer_tree_get_ebs_tag(struct btrfs_fs_info *fs_info,
1986
1986
rcu_read_lock ();
1987
1987
while ((eb = find_get_eb (& xas , end , tag )) != NULL ) {
1988
1988
if (!eb_batch_add (batch , eb )) {
1989
- * start = ((eb -> start + eb -> len ) >> fs_info -> sectorsize_bits );
1989
+ * start = ((eb -> start + eb -> len ) >> fs_info -> node_bits );
1990
1990
goto out ;
1991
1991
}
1992
1992
}
@@ -2008,7 +2008,7 @@ static struct extent_buffer *find_extent_buffer_nolock(
2008
2008
struct btrfs_fs_info * fs_info , u64 start )
2009
2009
{
2010
2010
struct extent_buffer * eb ;
2011
- unsigned long index = (start >> fs_info -> sectorsize_bits );
2011
+ unsigned long index = (start >> fs_info -> node_bits );
2012
2012
2013
2013
rcu_read_lock ();
2014
2014
eb = xa_load (& fs_info -> buffer_tree , index );
@@ -2114,8 +2114,8 @@ void btrfs_btree_wait_writeback_range(struct btrfs_fs_info *fs_info, u64 start,
2114
2114
u64 end )
2115
2115
{
2116
2116
struct eb_batch batch ;
2117
- unsigned long start_index = (start >> fs_info -> sectorsize_bits );
2118
- unsigned long end_index = (end >> fs_info -> sectorsize_bits );
2117
+ unsigned long start_index = (start >> fs_info -> node_bits );
2118
+ unsigned long end_index = (end >> fs_info -> node_bits );
2119
2119
2120
2120
eb_batch_init (& batch );
2121
2121
while (start_index <= end_index ) {
@@ -2151,7 +2151,7 @@ int btree_write_cache_pages(struct address_space *mapping,
2151
2151
2152
2152
eb_batch_init (& batch );
2153
2153
if (wbc -> range_cyclic ) {
2154
- index = ((mapping -> writeback_index << PAGE_SHIFT ) >> fs_info -> sectorsize_bits );
2154
+ index = ((mapping -> writeback_index << PAGE_SHIFT ) >> fs_info -> node_bits );
2155
2155
end = -1 ;
2156
2156
2157
2157
/*
@@ -2160,8 +2160,8 @@ int btree_write_cache_pages(struct address_space *mapping,
2160
2160
*/
2161
2161
scanned = (index == 0 );
2162
2162
} else {
2163
- index = (wbc -> range_start >> fs_info -> sectorsize_bits );
2164
- end = (wbc -> range_end >> fs_info -> sectorsize_bits );
2163
+ index = (wbc -> range_start >> fs_info -> node_bits );
2164
+ end = (wbc -> range_end >> fs_info -> node_bits );
2165
2165
2166
2166
scanned = 1 ;
2167
2167
}
@@ -3038,7 +3038,7 @@ struct extent_buffer *alloc_test_extent_buffer(struct btrfs_fs_info *fs_info,
3038
3038
eb -> fs_info = fs_info ;
3039
3039
again :
3040
3040
xa_lock_irq (& fs_info -> buffer_tree );
3041
- exists = __xa_cmpxchg (& fs_info -> buffer_tree , start >> fs_info -> sectorsize_bits ,
3041
+ exists = __xa_cmpxchg (& fs_info -> buffer_tree , start >> fs_info -> node_bits ,
3042
3042
NULL , eb , GFP_NOFS );
3043
3043
if (xa_is_err (exists )) {
3044
3044
ret = xa_err (exists );
@@ -3355,7 +3355,7 @@ struct extent_buffer *alloc_extent_buffer(struct btrfs_fs_info *fs_info,
3355
3355
again :
3356
3356
xa_lock_irq (& fs_info -> buffer_tree );
3357
3357
existing_eb = __xa_cmpxchg (& fs_info -> buffer_tree ,
3358
- start >> fs_info -> sectorsize_bits , NULL , eb ,
3358
+ start >> fs_info -> node_bits , NULL , eb ,
3359
3359
GFP_NOFS );
3360
3360
if (xa_is_err (existing_eb )) {
3361
3361
ret = xa_err (existing_eb );
@@ -3458,7 +3458,7 @@ static int release_extent_buffer(struct extent_buffer *eb)
3458
3458
* in this case.
3459
3459
*/
3460
3460
xa_cmpxchg_irq (& fs_info -> buffer_tree ,
3461
- eb -> start >> fs_info -> sectorsize_bits , eb , NULL ,
3461
+ eb -> start >> fs_info -> node_bits , eb , NULL ,
3462
3462
GFP_ATOMIC );
3463
3463
3464
3464
btrfs_leak_debug_del_eb (eb );
@@ -4300,9 +4300,9 @@ static int try_release_subpage_extent_buffer(struct folio *folio)
4300
4300
{
4301
4301
struct btrfs_fs_info * fs_info = folio_to_fs_info (folio );
4302
4302
struct extent_buffer * eb ;
4303
- unsigned long start = (folio_pos (folio ) >> fs_info -> sectorsize_bits );
4303
+ unsigned long start = (folio_pos (folio ) >> fs_info -> node_bits );
4304
4304
unsigned long index = start ;
4305
- unsigned long end = index + (PAGE_SIZE >> fs_info -> sectorsize_bits ) - 1 ;
4305
+ unsigned long end = index + (PAGE_SIZE >> fs_info -> node_bits ) - 1 ;
4306
4306
int ret ;
4307
4307
4308
4308
xa_lock_irq (& fs_info -> buffer_tree );
0 commit comments