aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorCarlos Maiolino <cem@kernel.org>2024-01-23 11:21:52 +0100
committerCarlos Maiolino <cem@kernel.org>2024-01-23 11:21:52 +0100
commitdd8306b39652e07961db3779279ec1d090575c7b (patch)
treedbdf22857f28f609ce8d4b66ea76672e2c004ffe
parentfc83c7574b1fb2258c9403461e55b0cb091c670c (diff)
parent1665923a8302088744a69403ff60a1709f5d24ed (diff)
downloadxfsprogs-dev-dd8306b39652e07961db3779279ec1d090575c7b.tar.gz
Merge tag 'xfsprogs-fixes-6.6_2023-12-21' of https://git.kernel.org/pub/scm/linux/kernel/git/djwong/xfsprogs-dev into for-next
xfsprogs: various bug fixes for 6.6 [1/8] This series fixes a couple of bugs that I found in the userspace support libraries. This has been running on the djcloud for months with no problems. Enjoy! Signed-off-by: Darrick J. Wong <djwong@kernel.org> Signed-off-by: Carlos Maiolino <cem@kernel.org>
-rw-r--r--copy/xfs_copy.c24
-rw-r--r--db/block.c14
-rw-r--r--db/io.c35
-rw-r--r--db/io.h3
-rw-r--r--include/libxfs.h1
-rw-r--r--libfrog/Makefile1
-rw-r--r--libfrog/div64.h96
-rw-r--r--libxfs/defer_item.c7
-rw-r--r--libxfs/libxfs_priv.h77
9 files changed, 171 insertions, 87 deletions
diff --git a/copy/xfs_copy.c b/copy/xfs_copy.c
index 6e692e4f72..0420649de9 100644
--- a/copy/xfs_copy.c
+++ b/copy/xfs_copy.c
@@ -827,13 +827,9 @@ main(int argc, char **argv)
do_out(_("Creating file %s\n"), target[i].name);
open_flags |= O_CREAT;
- if (!buffered_output)
- open_flags |= O_DIRECT;
write_last_block = 1;
} else if (S_ISREG(statbuf.st_mode)) {
open_flags |= O_TRUNC;
- if (!buffered_output)
- open_flags |= O_DIRECT;
write_last_block = 1;
} else {
/*
@@ -850,6 +846,8 @@ main(int argc, char **argv)
exit(1);
}
}
+ if (!buffered_output)
+ open_flags |= O_DIRECT;
target[i].fd = open(target[i].name, open_flags, 0644);
if (target[i].fd < 0) {
@@ -882,20 +880,32 @@ main(int argc, char **argv)
}
}
} else {
- char *lb[XFS_MAX_SECTORSIZE] = { NULL };
+ char *lb = memalign(wbuf_align, XFS_MAX_SECTORSIZE);
off64_t off;
+ ssize_t len;
/* ensure device files are sufficiently large */
+ memset(lb, 0, XFS_MAX_SECTORSIZE);
off = mp->m_sb.sb_dblocks * source_blocksize;
- off -= sizeof(lb);
- if (pwrite(target[i].fd, lb, sizeof(lb), off) < 0) {
+ off -= XFS_MAX_SECTORSIZE;
+ len = pwrite(target[i].fd, lb, XFS_MAX_SECTORSIZE, off);
+ if (len < 0) {
do_log(_("%s: failed to write last block\n"),
progname);
do_log(_("\tIs target \"%s\" too small?\n"),
target[i].name);
die_perror();
}
+ if (len != XFS_MAX_SECTORSIZE) {
+ do_log(
+ _("%s: short write to last block: %zd bytes, %zu expected\n"),
+ progname, len, XFS_MAX_SECTORSIZE);
+ do_log(_("\tIs target \"%s\" too small?\n"),
+ target[i].name);
+ exit(1);
+ }
+ free(lb);
}
}
diff --git a/db/block.c b/db/block.c
index 788337d370..d730c77967 100644
--- a/db/block.c
+++ b/db/block.c
@@ -126,7 +126,15 @@ daddr_f(
char *p;
if (argc == 1) {
- dbprintf(_("current daddr is %lld\n"), iocur_top->off >> BBSHIFT);
+ xfs_daddr_t daddr = iocur_top->off >> BBSHIFT;
+
+ if (iocur_is_ddev(iocur_top))
+ dbprintf(_("datadev daddr is %lld\n"), daddr);
+ else if (iocur_is_extlogdev(iocur_top))
+ dbprintf(_("logdev daddr is %lld\n"), daddr);
+ else
+ dbprintf(_("current daddr is %lld\n"), daddr);
+
return 0;
}
d = (int64_t)strtoull(argv[1], &p, 0);
@@ -220,6 +228,10 @@ fsblock_f(
char *p;
if (argc == 1) {
+ if (!iocur_is_ddev(iocur_top)) {
+ dbprintf(_("cursor does not point to data device\n"));
+ return 0;
+ }
dbprintf(_("current fsblock is %lld\n"),
XFS_DADDR_TO_FSB(mp, iocur_top->off >> BBSHIFT));
return 0;
diff --git a/db/io.c b/db/io.c
index 5ccfe3b536..590dd1f82f 100644
--- a/db/io.c
+++ b/db/io.c
@@ -137,18 +137,47 @@ pop_help(void)
));
}
+bool
+iocur_is_ddev(const struct iocur *ioc)
+{
+ if (!ioc->bp)
+ return false;
+
+ return ioc->bp->b_target == ioc->bp->b_mount->m_ddev_targp;
+}
+
+bool
+iocur_is_extlogdev(const struct iocur *ioc)
+{
+ struct xfs_buf *bp = ioc->bp;
+
+ if (!bp)
+ return false;
+ if (bp->b_mount->m_logdev_targp == bp->b_mount->m_ddev_targp)
+ return false;
+
+ return bp->b_target == bp->b_mount->m_logdev_targp;
+}
+
void
print_iocur(
char *tag,
iocur_t *ioc)
{
+ const char *block_unit = "fsbno?";
int i;
+ if (iocur_is_ddev(ioc))
+ block_unit = "fsbno";
+ else if (iocur_is_extlogdev(ioc))
+ block_unit = "logbno";
+
dbprintf("%s\n", tag);
dbprintf(_("\tbyte offset %lld, length %d\n"), ioc->off, ioc->len);
- dbprintf(_("\tbuffer block %lld (fsbno %lld), %d bb%s\n"), ioc->bb,
- (xfs_fsblock_t)XFS_DADDR_TO_FSB(mp, ioc->bb), ioc->blen,
- ioc->blen == 1 ? "" : "s");
+ dbprintf(_("\tbuffer block %lld (%s %lld), %d bb%s\n"), ioc->bb,
+ block_unit,
+ (xfs_fsblock_t)XFS_DADDR_TO_FSB(mp, ioc->bb),
+ ioc->blen, ioc->blen == 1 ? "" : "s");
if (ioc->bbmap) {
dbprintf(_("\tblock map"));
for (i = 0; i < ioc->bbmap->nmaps; i++)
diff --git a/db/io.h b/db/io.h
index bd86c31f67..f48b67b47a 100644
--- a/db/io.h
+++ b/db/io.h
@@ -56,6 +56,9 @@ extern void set_iocur_type(const struct typ *type);
extern void xfs_dummy_verify(struct xfs_buf *bp);
extern void xfs_verify_recalc_crc(struct xfs_buf *bp);
+bool iocur_is_ddev(const struct iocur *ioc);
+bool iocur_is_extlogdev(const struct iocur *ioc);
+
/*
* returns -1 for unchecked, 0 for bad and 1 for good
*/
diff --git a/include/libxfs.h b/include/libxfs.h
index eb3f9ac22b..4865663916 100644
--- a/include/libxfs.h
+++ b/include/libxfs.h
@@ -18,6 +18,7 @@
#include "kmem.h"
#include "libfrog/radix-tree.h"
#include "libfrog/bitmask.h"
+#include "libfrog/div64.h"
#include "atomic.h"
#include "spinlock.h"
diff --git a/libfrog/Makefile b/libfrog/Makefile
index 8cde97d418..dcfd1fb8a9 100644
--- a/libfrog/Makefile
+++ b/libfrog/Makefile
@@ -41,6 +41,7 @@ crc32cselftest.h \
crc32defs.h \
crc32table.h \
dahashselftest.h \
+div64.h \
fsgeom.h \
logging.h \
paths.h \
diff --git a/libfrog/div64.h b/libfrog/div64.h
new file mode 100644
index 0000000000..673b01cbab
--- /dev/null
+++ b/libfrog/div64.h
@@ -0,0 +1,96 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (c) 2000-2005 Silicon Graphics, Inc.
+ * All Rights Reserved.
+ */
+#ifndef LIBFROG_DIV64_H_
+#define LIBFROG_DIV64_H_
+
+static inline int __do_div(unsigned long long *n, unsigned base)
+{
+ int __res;
+ __res = (int)(((unsigned long) *n) % (unsigned) base);
+ *n = ((unsigned long) *n) / (unsigned) base;
+ return __res;
+}
+
+#define do_div(n,base) (__do_div((unsigned long long *)&(n), (base)))
+#define do_mod(a, b) ((a) % (b))
+#define rol32(x,y) (((x) << (y)) | ((x) >> (32 - (y))))
+
+/**
+ * div_u64_rem - unsigned 64bit divide with 32bit divisor with remainder
+ * @dividend: unsigned 64bit dividend
+ * @divisor: unsigned 32bit divisor
+ * @remainder: pointer to unsigned 32bit remainder
+ *
+ * Return: sets ``*remainder``, then returns dividend / divisor
+ *
+ * This is commonly provided by 32bit archs to provide an optimized 64bit
+ * divide.
+ */
+static inline uint64_t
+div_u64_rem(uint64_t dividend, uint32_t divisor, uint32_t *remainder)
+{
+ *remainder = dividend % divisor;
+ return dividend / divisor;
+}
+
+/**
+ * div_u64 - unsigned 64bit divide with 32bit divisor
+ * @dividend: unsigned 64bit dividend
+ * @divisor: unsigned 32bit divisor
+ *
+ * This is the most common 64bit divide and should be used if possible,
+ * as many 32bit archs can optimize this variant better than a full 64bit
+ * divide.
+ */
+static inline uint64_t div_u64(uint64_t dividend, uint32_t divisor)
+{
+ uint32_t remainder;
+ return div_u64_rem(dividend, divisor, &remainder);
+}
+
+/**
+ * div64_u64_rem - unsigned 64bit divide with 64bit divisor and remainder
+ * @dividend: unsigned 64bit dividend
+ * @divisor: unsigned 64bit divisor
+ * @remainder: pointer to unsigned 64bit remainder
+ *
+ * Return: sets ``*remainder``, then returns dividend / divisor
+ */
+static inline uint64_t
+div64_u64_rem(uint64_t dividend, uint64_t divisor, uint64_t *remainder)
+{
+ *remainder = dividend % divisor;
+ return dividend / divisor;
+}
+
+static inline uint64_t rounddown_64(uint64_t x, uint32_t y)
+{
+ do_div(x, y);
+ return x * y;
+}
+
+static inline bool isaligned_64(uint64_t x, uint32_t y)
+{
+ return do_div(x, y) == 0;
+}
+
+static inline uint64_t
+roundup_64(uint64_t x, uint32_t y)
+{
+ x += y - 1;
+ do_div(x, y);
+ return x * y;
+}
+
+static inline uint64_t
+howmany_64(uint64_t x, uint32_t y)
+{
+ x += y - 1;
+ do_div(x, y);
+ return x;
+}
+
+#endif /* LIBFROG_DIV64_H_ */
diff --git a/libxfs/defer_item.c b/libxfs/defer_item.c
index 3f51925204..8731d1834b 100644
--- a/libxfs/defer_item.c
+++ b/libxfs/defer_item.c
@@ -115,6 +115,13 @@ xfs_extent_free_finish_item(
error = xfs_free_extent(tp, xefi->xefi_pag, agbno,
xefi->xefi_blockcount, &oinfo, XFS_AG_RESV_NONE);
+ /*
+ * Don't free the XEFI if we need a new transaction to complete
+ * processing of it.
+ */
+ if (error == -EAGAIN)
+ return error;
+
xfs_extent_free_put_group(xefi);
kmem_cache_free(xfs_extfree_item_cache, xefi);
return error;
diff --git a/libxfs/libxfs_priv.h b/libxfs/libxfs_priv.h
index 2729241bda..5a7decf970 100644
--- a/libxfs/libxfs_priv.h
+++ b/libxfs/libxfs_priv.h
@@ -48,6 +48,7 @@
#include "kmem.h"
#include "libfrog/radix-tree.h"
#include "libfrog/bitmask.h"
+#include "libfrog/div64.h"
#include "atomic.h"
#include "spinlock.h"
#include "linux-err.h"
@@ -215,66 +216,6 @@ static inline bool WARN_ON(bool expr) {
(inode)->i_version = (version); \
} while (0)
-static inline int __do_div(unsigned long long *n, unsigned base)
-{
- int __res;
- __res = (int)(((unsigned long) *n) % (unsigned) base);
- *n = ((unsigned long) *n) / (unsigned) base;
- return __res;
-}
-
-#define do_div(n,base) (__do_div((unsigned long long *)&(n), (base)))
-#define do_mod(a, b) ((a) % (b))
-#define rol32(x,y) (((x) << (y)) | ((x) >> (32 - (y))))
-
-/**
- * div_u64_rem - unsigned 64bit divide with 32bit divisor with remainder
- * @dividend: unsigned 64bit dividend
- * @divisor: unsigned 32bit divisor
- * @remainder: pointer to unsigned 32bit remainder
- *
- * Return: sets ``*remainder``, then returns dividend / divisor
- *
- * This is commonly provided by 32bit archs to provide an optimized 64bit
- * divide.
- */
-static inline uint64_t
-div_u64_rem(uint64_t dividend, uint32_t divisor, uint32_t *remainder)
-{
- *remainder = dividend % divisor;
- return dividend / divisor;
-}
-
-/**
- * div_u64 - unsigned 64bit divide with 32bit divisor
- * @dividend: unsigned 64bit dividend
- * @divisor: unsigned 32bit divisor
- *
- * This is the most common 64bit divide and should be used if possible,
- * as many 32bit archs can optimize this variant better than a full 64bit
- * divide.
- */
-static inline uint64_t div_u64(uint64_t dividend, uint32_t divisor)
-{
- uint32_t remainder;
- return div_u64_rem(dividend, divisor, &remainder);
-}
-
-/**
- * div64_u64_rem - unsigned 64bit divide with 64bit divisor and remainder
- * @dividend: unsigned 64bit dividend
- * @divisor: unsigned 64bit divisor
- * @remainder: pointer to unsigned 64bit remainder
- *
- * Return: sets ``*remainder``, then returns dividend / divisor
- */
-static inline uint64_t
-div64_u64_rem(uint64_t dividend, uint64_t divisor, uint64_t *remainder)
-{
- *remainder = dividend % divisor;
- return dividend / divisor;
-}
-
#define min_t(type,x,y) \
({ type __x = (x); type __y = (y); __x < __y ? __x: __y; })
#define max_t(type,x,y) \
@@ -380,22 +321,6 @@ roundup_pow_of_two(uint v)
return 0;
}
-static inline uint64_t
-roundup_64(uint64_t x, uint32_t y)
-{
- x += y - 1;
- do_div(x, y);
- return x * y;
-}
-
-static inline uint64_t
-howmany_64(uint64_t x, uint32_t y)
-{
- x += y - 1;
- do_div(x, y);
- return x;
-}
-
/* buffer management */
#define XBF_TRYLOCK 0
#define XBF_UNMAPPED 0