1  /* SPDX-License-Identifier: GPL-2.0 */
2  #ifndef _BCACHEFS_BKEY_H
3  #define _BCACHEFS_BKEY_H
4  
5  #include <linux/bug.h>
6  #include "bcachefs_format.h"
7  #include "bkey_types.h"
8  #include "btree_types.h"
9  #include "util.h"
10  #include "vstructs.h"
11  
12  enum bch_validate_flags {
13  	BCH_VALIDATE_write		= BIT(0),
14  	BCH_VALIDATE_commit		= BIT(1),
15  	BCH_VALIDATE_journal		= BIT(2),
16  	BCH_VALIDATE_silent		= BIT(3),
17  };
18  
19  #if 0
20  
21  /*
22   * compiled unpack functions are disabled, pending a new interface for
23   * dynamically allocating executable memory:
24   */
25  
26  #ifdef CONFIG_X86_64
27  #define HAVE_BCACHEFS_COMPILED_UNPACK	1
28  #endif
29  #endif
30  
31  void bch2_bkey_packed_to_binary_text(struct printbuf *,
32  				     const struct bkey_format *,
33  				     const struct bkey_packed *);
34  
35  enum bkey_lr_packed {
36  	BKEY_PACKED_BOTH,
37  	BKEY_PACKED_RIGHT,
38  	BKEY_PACKED_LEFT,
39  	BKEY_PACKED_NONE,
40  };
41  
42  #define bkey_lr_packed(_l, _r)						\
43  	((_l)->format + ((_r)->format << 1))
44  
bkey_p_copy(struct bkey_packed * dst,const struct bkey_packed * src)45  static inline void bkey_p_copy(struct bkey_packed *dst, const struct bkey_packed *src)
46  {
47  	memcpy_u64s_small(dst, src, src->u64s);
48  }
49  
bkey_copy(struct bkey_i * dst,const struct bkey_i * src)50  static inline void bkey_copy(struct bkey_i *dst, const struct bkey_i *src)
51  {
52  	memcpy_u64s_small(dst, src, src->k.u64s);
53  }
54  
55  struct btree;
56  
57  __pure
58  unsigned bch2_bkey_greatest_differing_bit(const struct btree *,
59  					  const struct bkey_packed *,
60  					  const struct bkey_packed *);
61  __pure
62  unsigned bch2_bkey_ffs(const struct btree *, const struct bkey_packed *);
63  
64  __pure
65  int __bch2_bkey_cmp_packed_format_checked(const struct bkey_packed *,
66  				     const struct bkey_packed *,
67  				     const struct btree *);
68  
69  __pure
70  int __bch2_bkey_cmp_left_packed_format_checked(const struct btree *,
71  					  const struct bkey_packed *,
72  					  const struct bpos *);
73  
74  __pure
75  int bch2_bkey_cmp_packed(const struct btree *,
76  			 const struct bkey_packed *,
77  			 const struct bkey_packed *);
78  
79  __pure
80  int __bch2_bkey_cmp_left_packed(const struct btree *,
81  				const struct bkey_packed *,
82  				const struct bpos *);
83  
84  static inline __pure
bkey_cmp_left_packed(const struct btree * b,const struct bkey_packed * l,const struct bpos * r)85  int bkey_cmp_left_packed(const struct btree *b,
86  			 const struct bkey_packed *l, const struct bpos *r)
87  {
88  	return __bch2_bkey_cmp_left_packed(b, l, r);
89  }
90  
91  /*
92   * The compiler generates better code when we pass bpos by ref, but it's often
93   * enough terribly convenient to pass it by val... as much as I hate c++, const
94   * ref would be nice here:
95   */
96  __pure __flatten
bkey_cmp_left_packed_byval(const struct btree * b,const struct bkey_packed * l,struct bpos r)97  static inline int bkey_cmp_left_packed_byval(const struct btree *b,
98  					     const struct bkey_packed *l,
99  					     struct bpos r)
100  {
101  	return bkey_cmp_left_packed(b, l, &r);
102  }
103  
bpos_eq(struct bpos l,struct bpos r)104  static __always_inline bool bpos_eq(struct bpos l, struct bpos r)
105  {
106  	return  !((l.inode	^ r.inode) |
107  		  (l.offset	^ r.offset) |
108  		  (l.snapshot	^ r.snapshot));
109  }
110  
bpos_lt(struct bpos l,struct bpos r)111  static __always_inline bool bpos_lt(struct bpos l, struct bpos r)
112  {
113  	return  l.inode	!= r.inode ? l.inode < r.inode :
114  		l.offset != r.offset ? l.offset < r.offset :
115  		l.snapshot != r.snapshot ? l.snapshot < r.snapshot : false;
116  }
117  
bpos_le(struct bpos l,struct bpos r)118  static __always_inline bool bpos_le(struct bpos l, struct bpos r)
119  {
120  	return  l.inode	!= r.inode ? l.inode < r.inode :
121  		l.offset != r.offset ? l.offset < r.offset :
122  		l.snapshot != r.snapshot ? l.snapshot < r.snapshot : true;
123  }
124  
bpos_gt(struct bpos l,struct bpos r)125  static __always_inline bool bpos_gt(struct bpos l, struct bpos r)
126  {
127  	return bpos_lt(r, l);
128  }
129  
bpos_ge(struct bpos l,struct bpos r)130  static __always_inline bool bpos_ge(struct bpos l, struct bpos r)
131  {
132  	return bpos_le(r, l);
133  }
134  
bpos_cmp(struct bpos l,struct bpos r)135  static __always_inline int bpos_cmp(struct bpos l, struct bpos r)
136  {
137  	return  cmp_int(l.inode,    r.inode) ?:
138  		cmp_int(l.offset,   r.offset) ?:
139  		cmp_int(l.snapshot, r.snapshot);
140  }
141  
bpos_min(struct bpos l,struct bpos r)142  static inline struct bpos bpos_min(struct bpos l, struct bpos r)
143  {
144  	return bpos_lt(l, r) ? l : r;
145  }
146  
bpos_max(struct bpos l,struct bpos r)147  static inline struct bpos bpos_max(struct bpos l, struct bpos r)
148  {
149  	return bpos_gt(l, r) ? l : r;
150  }
151  
bkey_eq(struct bpos l,struct bpos r)152  static __always_inline bool bkey_eq(struct bpos l, struct bpos r)
153  {
154  	return  !((l.inode	^ r.inode) |
155  		  (l.offset	^ r.offset));
156  }
157  
bkey_lt(struct bpos l,struct bpos r)158  static __always_inline bool bkey_lt(struct bpos l, struct bpos r)
159  {
160  	return  l.inode	!= r.inode
161  		? l.inode < r.inode
162  		: l.offset < r.offset;
163  }
164  
bkey_le(struct bpos l,struct bpos r)165  static __always_inline bool bkey_le(struct bpos l, struct bpos r)
166  {
167  	return  l.inode	!= r.inode
168  		? l.inode < r.inode
169  		: l.offset <= r.offset;
170  }
171  
bkey_gt(struct bpos l,struct bpos r)172  static __always_inline bool bkey_gt(struct bpos l, struct bpos r)
173  {
174  	return bkey_lt(r, l);
175  }
176  
bkey_ge(struct bpos l,struct bpos r)177  static __always_inline bool bkey_ge(struct bpos l, struct bpos r)
178  {
179  	return bkey_le(r, l);
180  }
181  
bkey_cmp(struct bpos l,struct bpos r)182  static __always_inline int bkey_cmp(struct bpos l, struct bpos r)
183  {
184  	return  cmp_int(l.inode,    r.inode) ?:
185  		cmp_int(l.offset,   r.offset);
186  }
187  
bkey_min(struct bpos l,struct bpos r)188  static inline struct bpos bkey_min(struct bpos l, struct bpos r)
189  {
190  	return bkey_lt(l, r) ? l : r;
191  }
192  
bkey_max(struct bpos l,struct bpos r)193  static inline struct bpos bkey_max(struct bpos l, struct bpos r)
194  {
195  	return bkey_gt(l, r) ? l : r;
196  }
197  
bkey_and_val_eq(struct bkey_s_c l,struct bkey_s_c r)198  static inline bool bkey_and_val_eq(struct bkey_s_c l, struct bkey_s_c r)
199  {
200  	return bpos_eq(l.k->p, r.k->p) &&
201  		bkey_bytes(l.k) == bkey_bytes(r.k) &&
202  		!memcmp(l.v, r.v, bkey_val_bytes(l.k));
203  }
204  
205  void bch2_bpos_swab(struct bpos *);
206  void bch2_bkey_swab_key(const struct bkey_format *, struct bkey_packed *);
207  
bversion_cmp(struct bversion l,struct bversion r)208  static __always_inline int bversion_cmp(struct bversion l, struct bversion r)
209  {
210  	return  cmp_int(l.hi, r.hi) ?:
211  		cmp_int(l.lo, r.lo);
212  }
213  
214  #define ZERO_VERSION	((struct bversion) { .hi = 0, .lo = 0 })
215  #define MAX_VERSION	((struct bversion) { .hi = ~0, .lo = ~0ULL })
216  
bversion_zero(struct bversion v)217  static __always_inline bool bversion_zero(struct bversion v)
218  {
219  	return bversion_cmp(v, ZERO_VERSION) == 0;
220  }
221  
222  #ifdef CONFIG_BCACHEFS_DEBUG
223  /* statement expressions confusing unlikely()? */
224  #define bkey_packed(_k)							\
225  	({ EBUG_ON((_k)->format > KEY_FORMAT_CURRENT);			\
226  	 (_k)->format != KEY_FORMAT_CURRENT; })
227  #else
228  #define bkey_packed(_k)		((_k)->format != KEY_FORMAT_CURRENT)
229  #endif
230  
231  /*
232   * It's safe to treat an unpacked bkey as a packed one, but not the reverse
233   */
bkey_to_packed(struct bkey_i * k)234  static inline struct bkey_packed *bkey_to_packed(struct bkey_i *k)
235  {
236  	return (struct bkey_packed *) k;
237  }
238  
bkey_to_packed_c(const struct bkey_i * k)239  static inline const struct bkey_packed *bkey_to_packed_c(const struct bkey_i *k)
240  {
241  	return (const struct bkey_packed *) k;
242  }
243  
packed_to_bkey(struct bkey_packed * k)244  static inline struct bkey_i *packed_to_bkey(struct bkey_packed *k)
245  {
246  	return bkey_packed(k) ? NULL : (struct bkey_i *) k;
247  }
248  
packed_to_bkey_c(const struct bkey_packed * k)249  static inline const struct bkey *packed_to_bkey_c(const struct bkey_packed *k)
250  {
251  	return bkey_packed(k) ? NULL : (const struct bkey *) k;
252  }
253  
bkey_format_key_bits(const struct bkey_format * format)254  static inline unsigned bkey_format_key_bits(const struct bkey_format *format)
255  {
256  	return format->bits_per_field[BKEY_FIELD_INODE] +
257  		format->bits_per_field[BKEY_FIELD_OFFSET] +
258  		format->bits_per_field[BKEY_FIELD_SNAPSHOT];
259  }
260  
bpos_successor(struct bpos p)261  static inline struct bpos bpos_successor(struct bpos p)
262  {
263  	if (!++p.snapshot &&
264  	    !++p.offset &&
265  	    !++p.inode)
266  		BUG();
267  
268  	return p;
269  }
270  
bpos_predecessor(struct bpos p)271  static inline struct bpos bpos_predecessor(struct bpos p)
272  {
273  	if (!p.snapshot-- &&
274  	    !p.offset-- &&
275  	    !p.inode--)
276  		BUG();
277  
278  	return p;
279  }
280  
bpos_nosnap_successor(struct bpos p)281  static inline struct bpos bpos_nosnap_successor(struct bpos p)
282  {
283  	p.snapshot = 0;
284  
285  	if (!++p.offset &&
286  	    !++p.inode)
287  		BUG();
288  
289  	return p;
290  }
291  
bpos_nosnap_predecessor(struct bpos p)292  static inline struct bpos bpos_nosnap_predecessor(struct bpos p)
293  {
294  	p.snapshot = 0;
295  
296  	if (!p.offset-- &&
297  	    !p.inode--)
298  		BUG();
299  
300  	return p;
301  }
302  
bkey_start_offset(const struct bkey * k)303  static inline u64 bkey_start_offset(const struct bkey *k)
304  {
305  	return k->p.offset - k->size;
306  }
307  
bkey_start_pos(const struct bkey * k)308  static inline struct bpos bkey_start_pos(const struct bkey *k)
309  {
310  	return (struct bpos) {
311  		.inode		= k->p.inode,
312  		.offset		= bkey_start_offset(k),
313  		.snapshot	= k->p.snapshot,
314  	};
315  }
316  
317  /* Packed helpers */
318  
bkeyp_key_u64s(const struct bkey_format * format,const struct bkey_packed * k)319  static inline unsigned bkeyp_key_u64s(const struct bkey_format *format,
320  				      const struct bkey_packed *k)
321  {
322  	return bkey_packed(k) ? format->key_u64s : BKEY_U64s;
323  }
324  
bkeyp_u64s_valid(const struct bkey_format * f,const struct bkey_packed * k)325  static inline bool bkeyp_u64s_valid(const struct bkey_format *f,
326  				    const struct bkey_packed *k)
327  {
328  	return ((unsigned) k->u64s - bkeyp_key_u64s(f, k) <= U8_MAX - BKEY_U64s);
329  }
330  
bkeyp_key_bytes(const struct bkey_format * format,const struct bkey_packed * k)331  static inline unsigned bkeyp_key_bytes(const struct bkey_format *format,
332  				       const struct bkey_packed *k)
333  {
334  	return bkeyp_key_u64s(format, k) * sizeof(u64);
335  }
336  
bkeyp_val_u64s(const struct bkey_format * format,const struct bkey_packed * k)337  static inline unsigned bkeyp_val_u64s(const struct bkey_format *format,
338  				      const struct bkey_packed *k)
339  {
340  	return k->u64s - bkeyp_key_u64s(format, k);
341  }
342  
bkeyp_val_bytes(const struct bkey_format * format,const struct bkey_packed * k)343  static inline size_t bkeyp_val_bytes(const struct bkey_format *format,
344  				     const struct bkey_packed *k)
345  {
346  	return bkeyp_val_u64s(format, k) * sizeof(u64);
347  }
348  
set_bkeyp_val_u64s(const struct bkey_format * format,struct bkey_packed * k,unsigned val_u64s)349  static inline void set_bkeyp_val_u64s(const struct bkey_format *format,
350  				      struct bkey_packed *k, unsigned val_u64s)
351  {
352  	k->u64s = bkeyp_key_u64s(format, k) + val_u64s;
353  }
354  
355  #define bkeyp_val(_format, _k)						\
356  	 ((struct bch_val *) ((u64 *) (_k)->_data + bkeyp_key_u64s(_format, _k)))
357  
358  extern const struct bkey_format bch2_bkey_format_current;
359  
360  bool bch2_bkey_transform(const struct bkey_format *,
361  			 struct bkey_packed *,
362  			 const struct bkey_format *,
363  			 const struct bkey_packed *);
364  
365  struct bkey __bch2_bkey_unpack_key(const struct bkey_format *,
366  				   const struct bkey_packed *);
367  
368  #ifndef HAVE_BCACHEFS_COMPILED_UNPACK
369  struct bpos __bkey_unpack_pos(const struct bkey_format *,
370  			      const struct bkey_packed *);
371  #endif
372  
373  bool bch2_bkey_pack_key(struct bkey_packed *, const struct bkey *,
374  		   const struct bkey_format *);
375  
376  enum bkey_pack_pos_ret {
377  	BKEY_PACK_POS_EXACT,
378  	BKEY_PACK_POS_SMALLER,
379  	BKEY_PACK_POS_FAIL,
380  };
381  
382  enum bkey_pack_pos_ret bch2_bkey_pack_pos_lossy(struct bkey_packed *, struct bpos,
383  					   const struct btree *);
384  
bkey_pack_pos(struct bkey_packed * out,struct bpos in,const struct btree * b)385  static inline bool bkey_pack_pos(struct bkey_packed *out, struct bpos in,
386  				 const struct btree *b)
387  {
388  	return bch2_bkey_pack_pos_lossy(out, in, b) == BKEY_PACK_POS_EXACT;
389  }
390  
391  void bch2_bkey_unpack(const struct btree *, struct bkey_i *,
392  		 const struct bkey_packed *);
393  bool bch2_bkey_pack(struct bkey_packed *, const struct bkey_i *,
394  	       const struct bkey_format *);
395  
396  typedef void (*compiled_unpack_fn)(struct bkey *, const struct bkey_packed *);
397  
398  static inline void
__bkey_unpack_key_format_checked(const struct btree * b,struct bkey * dst,const struct bkey_packed * src)399  __bkey_unpack_key_format_checked(const struct btree *b,
400  			       struct bkey *dst,
401  			       const struct bkey_packed *src)
402  {
403  	if (IS_ENABLED(HAVE_BCACHEFS_COMPILED_UNPACK)) {
404  		compiled_unpack_fn unpack_fn = b->aux_data;
405  		unpack_fn(dst, src);
406  
407  		if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG) &&
408  		    bch2_expensive_debug_checks) {
409  			struct bkey dst2 = __bch2_bkey_unpack_key(&b->format, src);
410  
411  			BUG_ON(memcmp(dst, &dst2, sizeof(*dst)));
412  		}
413  	} else {
414  		*dst = __bch2_bkey_unpack_key(&b->format, src);
415  	}
416  }
417  
418  static inline struct bkey
bkey_unpack_key_format_checked(const struct btree * b,const struct bkey_packed * src)419  bkey_unpack_key_format_checked(const struct btree *b,
420  			       const struct bkey_packed *src)
421  {
422  	struct bkey dst;
423  
424  	__bkey_unpack_key_format_checked(b, &dst, src);
425  	return dst;
426  }
427  
__bkey_unpack_key(const struct btree * b,struct bkey * dst,const struct bkey_packed * src)428  static inline void __bkey_unpack_key(const struct btree *b,
429  				     struct bkey *dst,
430  				     const struct bkey_packed *src)
431  {
432  	if (likely(bkey_packed(src)))
433  		__bkey_unpack_key_format_checked(b, dst, src);
434  	else
435  		*dst = *packed_to_bkey_c(src);
436  }
437  
438  /**
439   * bkey_unpack_key -- unpack just the key, not the value
440   */
bkey_unpack_key(const struct btree * b,const struct bkey_packed * src)441  static inline struct bkey bkey_unpack_key(const struct btree *b,
442  					  const struct bkey_packed *src)
443  {
444  	return likely(bkey_packed(src))
445  		? bkey_unpack_key_format_checked(b, src)
446  		: *packed_to_bkey_c(src);
447  }
448  
449  static inline struct bpos
bkey_unpack_pos_format_checked(const struct btree * b,const struct bkey_packed * src)450  bkey_unpack_pos_format_checked(const struct btree *b,
451  			       const struct bkey_packed *src)
452  {
453  #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
454  	return bkey_unpack_key_format_checked(b, src).p;
455  #else
456  	return __bkey_unpack_pos(&b->format, src);
457  #endif
458  }
459  
bkey_unpack_pos(const struct btree * b,const struct bkey_packed * src)460  static inline struct bpos bkey_unpack_pos(const struct btree *b,
461  					  const struct bkey_packed *src)
462  {
463  	return likely(bkey_packed(src))
464  		? bkey_unpack_pos_format_checked(b, src)
465  		: packed_to_bkey_c(src)->p;
466  }
467  
468  /* Disassembled bkeys */
469  
bkey_disassemble(const struct btree * b,const struct bkey_packed * k,struct bkey * u)470  static inline struct bkey_s_c bkey_disassemble(const struct btree *b,
471  					       const struct bkey_packed *k,
472  					       struct bkey *u)
473  {
474  	__bkey_unpack_key(b, u, k);
475  
476  	return (struct bkey_s_c) { u, bkeyp_val(&b->format, k), };
477  }
478  
479  /* non const version: */
__bkey_disassemble(const struct btree * b,struct bkey_packed * k,struct bkey * u)480  static inline struct bkey_s __bkey_disassemble(const struct btree *b,
481  					       struct bkey_packed *k,
482  					       struct bkey *u)
483  {
484  	__bkey_unpack_key(b, u, k);
485  
486  	return (struct bkey_s) { .k = u, .v = bkeyp_val(&b->format, k), };
487  }
488  
bkey_field_max(const struct bkey_format * f,enum bch_bkey_fields nr)489  static inline u64 bkey_field_max(const struct bkey_format *f,
490  				 enum bch_bkey_fields nr)
491  {
492  	return f->bits_per_field[nr] < 64
493  		? (le64_to_cpu(f->field_offset[nr]) +
494  		   ~(~0ULL << f->bits_per_field[nr]))
495  		: U64_MAX;
496  }
497  
498  #ifdef HAVE_BCACHEFS_COMPILED_UNPACK
499  
500  int bch2_compile_bkey_format(const struct bkey_format *, void *);
501  
502  #else
503  
bch2_compile_bkey_format(const struct bkey_format * format,void * out)504  static inline int bch2_compile_bkey_format(const struct bkey_format *format,
505  					  void *out) { return 0; }
506  
507  #endif
508  
bkey_reassemble(struct bkey_i * dst,struct bkey_s_c src)509  static inline void bkey_reassemble(struct bkey_i *dst,
510  				   struct bkey_s_c src)
511  {
512  	dst->k = *src.k;
513  	memcpy_u64s_small(&dst->v, src.v, bkey_val_u64s(src.k));
514  }
515  
516  /* byte order helpers */
517  
518  #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
519  
high_word_offset(const struct bkey_format * f)520  static inline unsigned high_word_offset(const struct bkey_format *f)
521  {
522  	return f->key_u64s - 1;
523  }
524  
525  #define high_bit_offset		0
526  #define nth_word(p, n)		((p) - (n))
527  
528  #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
529  
high_word_offset(const struct bkey_format * f)530  static inline unsigned high_word_offset(const struct bkey_format *f)
531  {
532  	return 0;
533  }
534  
535  #define high_bit_offset		KEY_PACKED_BITS_START
536  #define nth_word(p, n)		((p) + (n))
537  
538  #else
539  #error edit for your odd byteorder.
540  #endif
541  
542  #define high_word(f, k)		((u64 *) (k)->_data + high_word_offset(f))
543  #define next_word(p)		nth_word(p, 1)
544  #define prev_word(p)		nth_word(p, -1)
545  
546  #ifdef CONFIG_BCACHEFS_DEBUG
547  void bch2_bkey_pack_test(void);
548  #else
bch2_bkey_pack_test(void)549  static inline void bch2_bkey_pack_test(void) {}
550  #endif
551  
552  #define bkey_fields()							\
553  	x(BKEY_FIELD_INODE,		p.inode)			\
554  	x(BKEY_FIELD_OFFSET,		p.offset)			\
555  	x(BKEY_FIELD_SNAPSHOT,		p.snapshot)			\
556  	x(BKEY_FIELD_SIZE,		size)				\
557  	x(BKEY_FIELD_VERSION_HI,	bversion.hi)			\
558  	x(BKEY_FIELD_VERSION_LO,	bversion.lo)
559  
560  struct bkey_format_state {
561  	u64 field_min[BKEY_NR_FIELDS];
562  	u64 field_max[BKEY_NR_FIELDS];
563  };
564  
565  void bch2_bkey_format_init(struct bkey_format_state *);
566  
__bkey_format_add(struct bkey_format_state * s,unsigned field,u64 v)567  static inline void __bkey_format_add(struct bkey_format_state *s, unsigned field, u64 v)
568  {
569  	s->field_min[field] = min(s->field_min[field], v);
570  	s->field_max[field] = max(s->field_max[field], v);
571  }
572  
573  /*
574   * Changes @format so that @k can be successfully packed with @format
575   */
bch2_bkey_format_add_key(struct bkey_format_state * s,const struct bkey * k)576  static inline void bch2_bkey_format_add_key(struct bkey_format_state *s, const struct bkey *k)
577  {
578  #define x(id, field) __bkey_format_add(s, id, k->field);
579  	bkey_fields()
580  #undef x
581  }
582  
583  void bch2_bkey_format_add_pos(struct bkey_format_state *, struct bpos);
584  struct bkey_format bch2_bkey_format_done(struct bkey_format_state *);
585  
bch2_bkey_format_field_overflows(struct bkey_format * f,unsigned i)586  static inline bool bch2_bkey_format_field_overflows(struct bkey_format *f, unsigned i)
587  {
588  	unsigned f_bits = f->bits_per_field[i];
589  	unsigned unpacked_bits = bch2_bkey_format_current.bits_per_field[i];
590  	u64 unpacked_mask = ~((~0ULL << 1) << (unpacked_bits - 1));
591  	u64 field_offset = le64_to_cpu(f->field_offset[i]);
592  
593  	if (f_bits > unpacked_bits)
594  		return true;
595  
596  	if ((f_bits == unpacked_bits) && field_offset)
597  		return true;
598  
599  	u64 f_mask = f_bits
600  		? ~((~0ULL << (f_bits - 1)) << 1)
601  		: 0;
602  
603  	if (((field_offset + f_mask) & unpacked_mask) < field_offset)
604  		return true;
605  	return false;
606  }
607  
608  int bch2_bkey_format_invalid(struct bch_fs *, struct bkey_format *,
609  			     enum bch_validate_flags, struct printbuf *);
610  void bch2_bkey_format_to_text(struct printbuf *, const struct bkey_format *);
611  
612  #endif /* _BCACHEFS_BKEY_H */
613