| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * Copyright (c) 2000-2005 Silicon Graphics, Inc. |
| * All Rights Reserved. |
| */ |
| |
| #include <sys/stat.h> |
| #include "init.h" |
| |
| #include "libxfs_priv.h" |
| #include "xfs_fs.h" |
| #include "xfs_shared.h" |
| #include "xfs_format.h" |
| #include "xfs_log_format.h" |
| #include "xfs_trans_resv.h" |
| #include "xfs_mount.h" |
| #include "xfs_defer.h" |
| #include "xfs_inode_buf.h" |
| #include "xfs_inode_fork.h" |
| #include "xfs_inode.h" |
| #include "xfs_trans.h" |
| #include "xfs_rmap_btree.h" |
| #include "xfs_refcount_btree.h" |
| #include "xfs_imeta.h" |
| #include "xfs_rtrmap_btree.h" |
| #include "libfrog/platform.h" |
| #include "xfile.h" |
| |
| #include "libxfs.h" /* for now */ |
| |
| #ifndef HAVE_LIBURCU_ATOMIC64 |
| pthread_mutex_t atomic64_lock = PTHREAD_MUTEX_INITIALIZER; |
| #endif |
| |
| char *progname = "libxfs"; /* default, changed by each tool */ |
| |
| int libxfs_bhash_size; /* #buckets in bcache */ |
| |
| int use_xfs_buf_lock; /* global flag: use xfs_buf locks for MT */ |
| |
| /* |
| * dev_map - map open devices to fd. |
| */ |
| #define MAX_DEVS 10 /* arbitary maximum */ |
| static int nextfakedev = -1; /* device number to give to next fake device */ |
| static struct dev_to_fd { |
| dev_t dev; |
| int fd; |
| } dev_map[MAX_DEVS]={{0}}; |
| |
| /* |
| * Checks whether a given device has a mounted, writable |
| * filesystem, returns 1 if it does & fatal (just warns |
| * if not fatal, but allows us to proceed). |
| * |
| * Useful to tools which will produce uncertain results |
| * if the filesystem is active - repair, check, logprint. |
| */ |
| static int |
| check_isactive(char *name, char *block, int fatal) |
| { |
| struct stat st; |
| |
| if (stat(block, &st) < 0) |
| return 0; |
| if ((st.st_mode & S_IFMT) != S_IFBLK) |
| return 0; |
| if (platform_check_ismounted(name, block, &st, 0) == 0) |
| return 0; |
| if (platform_check_iswritable(name, block, &st)) |
| return fatal ? 1 : 0; |
| return 0; |
| } |
| |
| /* libxfs_device_to_fd: |
| * lookup a device number in the device map |
| * return the associated fd |
| */ |
| int |
| libxfs_device_to_fd(dev_t device) |
| { |
| int d; |
| |
| for (d = 0; d < MAX_DEVS; d++) |
| if (dev_map[d].dev == device) |
| return dev_map[d].fd; |
| |
| fprintf(stderr, _("%s: %s: device %lld is not open\n"), |
| progname, __FUNCTION__, (long long)device); |
| exit(1); |
| /* NOTREACHED */ |
| } |
| |
| /* libxfs_device_open: |
| * open a device and return its device number |
| */ |
| dev_t |
| libxfs_device_open(char *path, int creat, int xflags, int setblksize) |
| { |
| dev_t dev; |
| int fd, d, flags; |
| int readonly, dio, excl; |
| struct stat statb; |
| |
| readonly = (xflags & LIBXFS_ISREADONLY); |
| excl = (xflags & LIBXFS_EXCLUSIVELY) && !creat; |
| dio = (xflags & LIBXFS_DIRECT) && !creat && platform_direct_blockdev(); |
| |
| retry: |
| flags = (readonly ? O_RDONLY : O_RDWR) | \ |
| (creat ? (O_CREAT|O_TRUNC) : 0) | \ |
| (dio ? O_DIRECT : 0) | \ |
| (excl ? O_EXCL : 0); |
| |
| if ((fd = open(path, flags, 0666)) < 0) { |
| if (errno == EINVAL && --dio == 0) |
| goto retry; |
| fprintf(stderr, _("%s: cannot open %s: %s\n"), |
| progname, path, strerror(errno)); |
| exit(1); |
| } |
| |
| if (fstat(fd, &statb) < 0) { |
| fprintf(stderr, _("%s: cannot stat %s: %s\n"), |
| progname, path, strerror(errno)); |
| exit(1); |
| } |
| |
| if (!readonly && setblksize && (statb.st_mode & S_IFMT) == S_IFBLK) { |
| if (setblksize == 1) |
| /* use the default blocksize */ |
| (void)platform_set_blocksize(fd, path, statb.st_rdev, XFS_MIN_SECTORSIZE, 0); |
| else { |
| /* given an explicit blocksize to use */ |
| if (platform_set_blocksize(fd, path, statb.st_rdev, setblksize, 1)) |
| exit(1); |
| } |
| } |
| |
| /* |
| * Get the device number from the stat buf - unless |
| * we're not opening a real device, in which case |
| * choose a new fake device number. |
| */ |
| dev = (statb.st_rdev) ? (statb.st_rdev) : (nextfakedev--); |
| |
| for (d = 0; d < MAX_DEVS; d++) |
| if (dev_map[d].dev == dev) { |
| fprintf(stderr, _("%s: device %lld is already open\n"), |
| progname, (long long)dev); |
| exit(1); |
| } |
| |
| for (d = 0; d < MAX_DEVS; d++) |
| if (!dev_map[d].dev) { |
| dev_map[d].dev = dev; |
| dev_map[d].fd = fd; |
| |
| return dev; |
| } |
| |
| fprintf(stderr, _("%s: %s: too many open devices\n"), |
| progname, __FUNCTION__); |
| exit(1); |
| /* NOTREACHED */ |
| } |
| |
| void |
| libxfs_device_close(dev_t dev) |
| { |
| int d; |
| |
| for (d = 0; d < MAX_DEVS; d++) |
| if (dev_map[d].dev == dev) { |
| int fd, ret; |
| |
| fd = dev_map[d].fd; |
| dev_map[d].dev = dev_map[d].fd = 0; |
| |
| ret = platform_flush_device(fd, dev); |
| if (ret) { |
| ret = -errno; |
| fprintf(stderr, |
| _("%s: flush of device %lld failed, err=%d"), |
| progname, (long long)dev, ret); |
| } |
| close(fd); |
| |
| return; |
| } |
| |
| fprintf(stderr, _("%s: %s: device %lld is not open\n"), |
| progname, __FUNCTION__, (long long)dev); |
| exit(1); |
| } |
| |
| static int |
| check_open(char *path, int flags, char **rawfile, char **blockfile) |
| { |
| int readonly = (flags & LIBXFS_ISREADONLY); |
| int inactive = (flags & LIBXFS_ISINACTIVE); |
| int dangerously = (flags & LIBXFS_DANGEROUSLY); |
| struct stat stbuf; |
| |
| if (stat(path, &stbuf) < 0) { |
| perror(path); |
| return 0; |
| } |
| if (!(*rawfile = platform_findrawpath(path))) { |
| fprintf(stderr, _("%s: " |
| "can't find a character device matching %s\n"), |
| progname, path); |
| return 0; |
| } |
| if (!(*blockfile = platform_findblockpath(path))) { |
| fprintf(stderr, _("%s: " |
| "can't find a block device matching %s\n"), |
| progname, path); |
| return 0; |
| } |
| if (!readonly && !inactive && platform_check_ismounted(path, *blockfile, NULL, 1)) |
| return 0; |
| |
| if (inactive && check_isactive(path, *blockfile, ((readonly|dangerously)?1:0))) |
| return 0; |
| |
| return 1; |
| } |
| |
| /* |
| * Initialize/destroy all of the cache allocators we use. |
| */ |
| static void |
| init_caches(void) |
| { |
| int error; |
| |
| /* initialise cache allocation */ |
| xfs_buf_cache = kmem_cache_create("xfs_buffer", |
| sizeof(struct xfs_buf), 0, 0, NULL); |
| xfs_inode_cache = kmem_cache_create("xfs_inode", |
| sizeof(struct xfs_inode), 0, 0, NULL); |
| xfs_ifork_cache = kmem_cache_create("xfs_ifork", |
| sizeof(struct xfs_ifork), 0, 0, NULL); |
| xfs_ili_cache = kmem_cache_create("xfs_inode_log_item", |
| sizeof(struct xfs_inode_log_item), 0, 0, NULL); |
| xfs_buf_item_cache = kmem_cache_create("xfs_buf_log_item", |
| sizeof(struct xfs_buf_log_item), 0, 0, NULL); |
| xfs_da_state_cache = kmem_cache_create("xfs_da_state", |
| sizeof(struct xfs_da_state), 0, 0, NULL); |
| |
| error = xfs_btree_init_cur_caches(); |
| if (error) { |
| fprintf(stderr, "Could not allocate btree cursor caches.\n"); |
| abort(); |
| } |
| |
| error = xfs_defer_init_item_caches(); |
| if (error) { |
| fprintf(stderr, "Could not allocate defer work item caches.\n"); |
| abort(); |
| } |
| |
| xfs_trans_cache = kmem_cache_create("xfs_trans", |
| sizeof(struct xfs_trans), 0, 0, NULL); |
| } |
| |
| static void |
| destroy_kmem_caches(void) |
| { |
| kmem_cache_destroy(xfs_buf_cache); |
| kmem_cache_destroy(xfs_ili_cache); |
| kmem_cache_destroy(xfs_inode_cache); |
| kmem_cache_destroy(xfs_ifork_cache); |
| kmem_cache_destroy(xfs_buf_item_cache); |
| kmem_cache_destroy(xfs_da_state_cache); |
| xfs_defer_destroy_item_caches(); |
| xfs_btree_destroy_cur_caches(); |
| kmem_cache_destroy(xfs_trans_cache); |
| } |
| |
| static void |
| libxfs_close_devices( |
| struct libxfs_xinit *li) |
| { |
| if (li->ddev) |
| libxfs_device_close(li->ddev); |
| if (li->logdev && li->logdev != li->ddev) |
| libxfs_device_close(li->logdev); |
| if (li->rtdev) |
| libxfs_device_close(li->rtdev); |
| |
| li->ddev = li->logdev = li->rtdev = 0; |
| li->dfd = li->logfd = li->rtfd = -1; |
| } |
| |
| /* |
| * libxfs initialization. |
| * Caller gets a 0 on failure (and we print a message), 1 on success. |
| */ |
| int |
| libxfs_init(libxfs_init_t *a) |
| { |
| char *blockfile; |
| char *dname; |
| char dpath[25]; |
| int fd; |
| char *logname; |
| char logpath[25]; |
| char *rawfile; |
| char *rtname; |
| char rtpath[25]; |
| int rval = 0; |
| int flags; |
| |
| dpath[0] = logpath[0] = rtpath[0] = '\0'; |
| dname = a->dname; |
| logname = a->logname; |
| rtname = a->rtname; |
| a->dfd = a->logfd = a->rtfd = -1; |
| a->ddev = a->logdev = a->rtdev = 0; |
| a->dsize = a->lbsize = a->rtbsize = 0; |
| a->dbsize = a->logBBsize = a->logBBstart = a->rtsize = 0; |
| |
| xfile_libinit(); |
| |
| fd = -1; |
| flags = (a->isreadonly | a->isdirect); |
| |
| rcu_init(); |
| rcu_register_thread(); |
| radix_tree_init(); |
| |
| if (a->volname) { |
| if(!check_open(a->volname,flags,&rawfile,&blockfile)) |
| goto done; |
| fd = open(rawfile, O_RDONLY); |
| dname = a->dname = a->volname; |
| a->volname = NULL; |
| } |
| if (dname) { |
| if (a->disfile) { |
| a->ddev= libxfs_device_open(dname, a->dcreat, flags, |
| a->setblksize); |
| a->dfd = libxfs_device_to_fd(a->ddev); |
| platform_findsizes(dname, a->dfd, &a->dsize, |
| &a->dbsize); |
| } else { |
| if (!check_open(dname, flags, &rawfile, &blockfile)) |
| goto done; |
| a->ddev = libxfs_device_open(rawfile, |
| a->dcreat, flags, a->setblksize); |
| a->dfd = libxfs_device_to_fd(a->ddev); |
| platform_findsizes(rawfile, a->dfd, |
| &a->dsize, &a->dbsize); |
| } |
| } else |
| a->dsize = 0; |
| if (logname) { |
| if (a->lisfile) { |
| a->logdev = libxfs_device_open(logname, |
| a->lcreat, flags, a->setblksize); |
| a->logfd = libxfs_device_to_fd(a->logdev); |
| platform_findsizes(dname, a->logfd, &a->logBBsize, |
| &a->lbsize); |
| } else { |
| if (!check_open(logname, flags, &rawfile, &blockfile)) |
| goto done; |
| a->logdev = libxfs_device_open(rawfile, |
| a->lcreat, flags, a->setblksize); |
| a->logfd = libxfs_device_to_fd(a->logdev); |
| platform_findsizes(rawfile, a->logfd, |
| &a->logBBsize, &a->lbsize); |
| } |
| } else |
| a->logBBsize = 0; |
| if (rtname) { |
| if (a->risfile) { |
| a->rtdev = libxfs_device_open(rtname, |
| a->rcreat, flags, a->setblksize); |
| a->rtfd = libxfs_device_to_fd(a->rtdev); |
| platform_findsizes(dname, a->rtfd, &a->rtsize, |
| &a->rtbsize); |
| } else { |
| if (!check_open(rtname, flags, &rawfile, &blockfile)) |
| goto done; |
| a->rtdev = libxfs_device_open(rawfile, |
| a->rcreat, flags, a->setblksize); |
| a->rtfd = libxfs_device_to_fd(a->rtdev); |
| platform_findsizes(rawfile, a->rtfd, |
| &a->rtsize, &a->rtbsize); |
| } |
| } else |
| a->rtsize = 0; |
| if (a->dsize < 0) { |
| fprintf(stderr, _("%s: can't get size for data subvolume\n"), |
| progname); |
| goto done; |
| } |
| if (a->logBBsize < 0) { |
| fprintf(stderr, _("%s: can't get size for log subvolume\n"), |
| progname); |
| goto done; |
| } |
| if (a->rtsize < 0) { |
| fprintf(stderr, _("%s: can't get size for realtime subvolume\n"), |
| progname); |
| goto done; |
| } |
| if (!libxfs_bhash_size) |
| libxfs_bhash_size = LIBXFS_BHASHSIZE(sbp); |
| use_xfs_buf_lock = a->usebuflock; |
| xfs_dir_startup(); |
| init_caches(); |
| rval = 1; |
| done: |
| if (dpath[0]) |
| unlink(dpath); |
| if (logpath[0]) |
| unlink(logpath); |
| if (rtpath[0]) |
| unlink(rtpath); |
| if (fd >= 0) |
| close(fd); |
| if (!rval) { |
| libxfs_close_devices(a); |
| rcu_unregister_thread(); |
| } |
| |
| return rval; |
| } |
| |
| |
| /* |
| * Initialize realtime fields in the mount structure. |
| */ |
| static int |
| rtmount_init( |
| xfs_mount_t *mp, /* file system mount structure */ |
| int flags) |
| { |
| struct xfs_buf *bp; /* buffer for last block of subvolume */ |
| xfs_daddr_t d; /* address of last block of subvolume */ |
| int error; |
| |
| if (!xfs_has_realtime(mp)) |
| return 0; |
| |
| if (mp->m_rtdev_targp->bt_bdev == 0 && !xfs_is_debugger(mp)) { |
| fprintf(stderr, _("%s: filesystem has a realtime subvolume\n"), |
| progname); |
| return -1; |
| } |
| mp->m_rsumlevels = mp->m_sb.sb_rextslog + 1; |
| mp->m_rsumsize = |
| (uint)sizeof(xfs_suminfo_t) * mp->m_rsumlevels * |
| mp->m_sb.sb_rbmblocks; |
| mp->m_rsumsize = roundup(mp->m_rsumsize, mp->m_sb.sb_blocksize); |
| mp->m_rbmip = mp->m_rsumip = mp->m_rrmapip = NULL; |
| |
| /* |
| * Allow debugger to be run without the realtime device present. |
| */ |
| if (xfs_is_debugger(mp)) |
| return 0; |
| |
| /* |
| * Check that the realtime section is an ok size. |
| */ |
| d = (xfs_daddr_t)XFS_FSB_TO_BB(mp, mp->m_sb.sb_rblocks); |
| if (XFS_BB_TO_FSB(mp, d) != mp->m_sb.sb_rblocks) { |
| fprintf(stderr, _("%s: realtime init - %llu != %llu\n"), |
| progname, (unsigned long long) XFS_BB_TO_FSB(mp, d), |
| (unsigned long long) mp->m_sb.sb_rblocks); |
| return -1; |
| } |
| error = libxfs_buf_read(mp->m_rtdev, d - XFS_FSB_TO_BB(mp, 1), |
| XFS_FSB_TO_BB(mp, 1), 0, &bp, NULL); |
| if (error) { |
| fprintf(stderr, _("%s: realtime size check failed\n"), |
| progname); |
| return -1; |
| } |
| libxfs_buf_relse(bp); |
| return 0; |
| } |
| |
| /* |
| * Set parameters for inode allocation heuristics, taking into account |
| * filesystem size and inode32/inode64 mount options; i.e. specifically |
| * whether or not XFS_MOUNT_SMALL_INUMS is set. |
| * |
| * Inode allocation patterns are altered only if inode32 is requested |
| * (XFS_MOUNT_SMALL_INUMS), and the filesystem is sufficiently large. |
| * If altered, XFS_MOUNT_32BITINODES is set as well. |
| * |
| * An agcount independent of that in the mount structure is provided |
| * because in the growfs case, mp->m_sb.sb_agcount is not yet updated |
| * to the potentially higher ag count. |
| * |
| * Returns the maximum AG index which may contain inodes. |
| */ |
| xfs_agnumber_t |
| xfs_set_inode_alloc( |
| struct xfs_mount *mp, |
| xfs_agnumber_t agcount) |
| { |
| xfs_agnumber_t index; |
| xfs_agnumber_t maxagi = 0; |
| xfs_sb_t *sbp = &mp->m_sb; |
| xfs_agnumber_t max_metadata; |
| xfs_agino_t agino; |
| xfs_ino_t ino; |
| |
| /* |
| * Calculate how much should be reserved for inodes to meet |
| * the max inode percentage. Used only for inode32. |
| */ |
| if (M_IGEO(mp)->maxicount) { |
| uint64_t icount; |
| |
| icount = sbp->sb_dblocks * sbp->sb_imax_pct; |
| do_div(icount, 100); |
| icount += sbp->sb_agblocks - 1; |
| do_div(icount, sbp->sb_agblocks); |
| max_metadata = icount; |
| } else { |
| max_metadata = agcount; |
| } |
| |
| /* Get the last possible inode in the filesystem */ |
| agino = XFS_AGB_TO_AGINO(mp, sbp->sb_agblocks - 1); |
| ino = XFS_AGINO_TO_INO(mp, agcount - 1, agino); |
| |
| /* |
| * If user asked for no more than 32-bit inodes, and the fs is |
| * sufficiently large, set XFS_MOUNT_32BITINODES if we must alter |
| * the allocator to accommodate the request. |
| */ |
| if (ino > XFS_MAXINUMBER_32) |
| xfs_set_inode32(mp); |
| else |
| xfs_clear_inode32(mp); |
| |
| for (index = 0; index < agcount; index++) { |
| struct xfs_perag *pag; |
| |
| ino = XFS_AGINO_TO_INO(mp, index, agino); |
| |
| pag = xfs_perag_get(mp, index); |
| |
| if (xfs_is_inode32(mp)) { |
| if (ino > XFS_MAXINUMBER_32) { |
| pag->pagi_inodeok = 0; |
| pag->pagf_metadata = 0; |
| } else { |
| pag->pagi_inodeok = 1; |
| maxagi++; |
| if (index < max_metadata) |
| pag->pagf_metadata = 1; |
| else |
| pag->pagf_metadata = 0; |
| } |
| } else { |
| pag->pagi_inodeok = 1; |
| pag->pagf_metadata = 0; |
| } |
| |
| xfs_perag_put(pag); |
| } |
| |
| return xfs_is_inode32(mp) ? maxagi : agcount; |
| } |
| |
| static struct xfs_buftarg * |
| libxfs_buftarg_alloc( |
| struct xfs_mount *mp, |
| dev_t dev, |
| unsigned long write_fails, |
| unsigned int buftarg_flags) |
| { |
| struct xfs_buftarg *btp; |
| unsigned int bcache_flags = 0; |
| |
| if (!write_fails) |
| buftarg_flags &= ~XFS_BUFTARG_INJECT_WRITE_FAIL; |
| |
| btp = malloc(sizeof(*btp)); |
| if (!btp) { |
| fprintf(stderr, _("%s: buftarg init failed\n"), |
| progname); |
| exit(1); |
| } |
| btp->bt_mount = mp; |
| btp->bt_bdev = dev; |
| btp->flags = buftarg_flags; |
| btp->writes_left = write_fails; |
| if (btp->flags & XFS_BUFTARG_MISCOMPARE_PURGE) |
| bcache_flags |= CACHE_MISCOMPARE_PURGE; |
| pthread_mutex_init(&btp->lock, NULL); |
| |
| btp->bcache = cache_init(bcache_flags, libxfs_bhash_size, |
| &libxfs_bcache_operations); |
| |
| return btp; |
| } |
| |
| struct xfs_buftarg * |
| libxfs_alloc_memory_buftarg( |
| struct xfs_mount *mp, |
| struct xfile *xfile) |
| { |
| struct xfs_buftarg *btp; |
| unsigned int bcache_flags = 0; |
| |
| btp = malloc(sizeof(*btp)); |
| if (!btp) { |
| fprintf(stderr, _("%s: buftarg init failed\n"), |
| progname); |
| exit(1); |
| } |
| btp->bt_mount = mp; |
| btp->bt_xfile = xfile; |
| btp->flags = XFS_BUFTARG_IN_MEMORY; |
| btp->writes_left = 0; |
| pthread_mutex_init(&btp->lock, NULL); |
| |
| /* |
| * Keep the bucket count small because the only anticipated caller is |
| * per-AG in-memory btrees, for which we don't need to scale to handle |
| * an entire filesystem. |
| */ |
| btp->bcache = cache_init(bcache_flags, 31, &libxfs_bcache_operations); |
| |
| return btp; |
| } |
| |
| enum libxfs_write_failure_nums { |
| WF_DATA = 0, |
| WF_LOG, |
| WF_RT, |
| WF_MAX_OPTS, |
| }; |
| |
| static char *wf_opts[] = { |
| [WF_DATA] = "ddev", |
| [WF_LOG] = "logdev", |
| [WF_RT] = "rtdev", |
| [WF_MAX_OPTS] = NULL, |
| }; |
| |
| void |
| libxfs_buftarg_init( |
| struct xfs_mount *mp, |
| dev_t dev, |
| dev_t logdev, |
| dev_t rtdev, |
| unsigned int btflags) |
| { |
| char *p = getenv("LIBXFS_DEBUG_WRITE_CRASH"); |
| unsigned long dfail = 0, lfail = 0, rfail = 0; |
| unsigned int dflags = 0, lflags = 0, rflags = 0; |
| |
| /* Simulate utility crash after a certain number of writes. */ |
| while (p && *p) { |
| char *val; |
| |
| switch (getsubopt(&p, wf_opts, &val)) { |
| case WF_DATA: |
| if (!val) { |
| fprintf(stderr, |
| _("ddev write fail requires a parameter\n")); |
| exit(1); |
| } |
| dfail = strtoul(val, NULL, 0); |
| if (dfail) |
| dflags |= XFS_BUFTARG_INJECT_WRITE_FAIL; |
| break; |
| case WF_LOG: |
| if (!val) { |
| fprintf(stderr, |
| _("logdev write fail requires a parameter\n")); |
| exit(1); |
| } |
| lfail = strtoul(val, NULL, 0); |
| if (lfail) |
| lflags |= XFS_BUFTARG_INJECT_WRITE_FAIL; |
| break; |
| case WF_RT: |
| if (!val) { |
| fprintf(stderr, |
| _("rtdev write fail requires a parameter\n")); |
| exit(1); |
| } |
| rfail = strtoul(val, NULL, 0); |
| if (rfail) |
| rflags |= XFS_BUFTARG_INJECT_WRITE_FAIL; |
| break; |
| default: |
| fprintf(stderr, _("unknown write fail type %s\n"), |
| val); |
| exit(1); |
| break; |
| } |
| } |
| |
| if (mp->m_ddev_targp) { |
| /* should already have all buftargs initialised */ |
| if (mp->m_ddev_targp->bt_bdev != dev || |
| mp->m_ddev_targp->bt_mount != mp) { |
| fprintf(stderr, |
| _("%s: bad buftarg reinit, ddev\n"), |
| progname); |
| exit(1); |
| } |
| if (!logdev || logdev == dev) { |
| if (mp->m_logdev_targp != mp->m_ddev_targp) { |
| fprintf(stderr, |
| _("%s: bad buftarg reinit, ldev mismatch\n"), |
| progname); |
| exit(1); |
| } |
| } else if (mp->m_logdev_targp->bt_bdev != logdev || |
| mp->m_logdev_targp->bt_mount != mp) { |
| fprintf(stderr, |
| _("%s: bad buftarg reinit, logdev\n"), |
| progname); |
| exit(1); |
| } |
| if (rtdev && (mp->m_rtdev_targp->bt_bdev != rtdev || |
| mp->m_rtdev_targp->bt_mount != mp)) { |
| fprintf(stderr, |
| _("%s: bad buftarg reinit, rtdev\n"), |
| progname); |
| exit(1); |
| } |
| return; |
| } |
| |
| mp->m_ddev_targp = libxfs_buftarg_alloc(mp, dev, dfail, |
| dflags | btflags); |
| if (!logdev || logdev == dev) |
| mp->m_logdev_targp = mp->m_ddev_targp; |
| else |
| mp->m_logdev_targp = libxfs_buftarg_alloc(mp, logdev, lfail, |
| lflags | btflags); |
| mp->m_rtdev_targp = libxfs_buftarg_alloc(mp, rtdev, rfail, |
| rflags | btflags); |
| } |
| |
| /* Compute maximum possible height for per-AG btree types for this fs. */ |
| static inline void |
| xfs_agbtree_compute_maxlevels( |
| struct xfs_mount *mp) |
| { |
| unsigned int levels; |
| |
| levels = max(mp->m_alloc_maxlevels, M_IGEO(mp)->inobt_maxlevels); |
| levels = max(levels, mp->m_rmap_maxlevels); |
| mp->m_agbtree_maxlevels = max(levels, mp->m_refc_maxlevels); |
| } |
| |
| /* Compute maximum possible height for realtime btree types for this fs. */ |
| static inline void |
| xfs_rtbtree_compute_maxlevels( |
| struct xfs_mount *mp) |
| { |
| unsigned int levels; |
| |
| levels = max(mp->m_rtrmap_maxlevels, mp->m_rtrefc_maxlevels); |
| mp->m_rtbtree_maxlevels = levels; |
| } |
| |
| STATIC void |
| libxfs_mountfs_imeta( |
| struct xfs_mount *mp) |
| { |
| int error; |
| |
| /* Ignore filesystems that are under construction. */ |
| if (mp->m_sb.sb_inprogress) |
| return; |
| |
| if (xfs_has_metadir(mp)) { |
| error = -libxfs_imeta_iget(mp, mp->m_sb.sb_metadirino, |
| XFS_DIR3_FT_DIR, &mp->m_metadirip); |
| if (error) |
| fprintf(stderr, |
| _("%s: could not open metadata directory, error %d\n"), |
| progname, error); |
| } |
| |
| error = -xfs_imeta_mount(mp); |
| if (error) { |
| if (mp->m_metadirip) |
| libxfs_imeta_irele(mp->m_metadirip); |
| mp->m_metadirip = NULL; |
| |
| fprintf(stderr, |
| _("%s: mounting metadata directory failed, error %d\n"), |
| progname, error); |
| } |
| } |
| |
| /* |
| * Mount structure initialization, provides a filled-in xfs_mount_t |
| * such that the numerous XFS_* macros can be used. If dev is zero, |
| * no IO will be performed (no size checks, read root inodes). |
| */ |
| struct xfs_mount * |
| libxfs_mount( |
| struct xfs_mount *mp, |
| struct xfs_sb *sb, |
| dev_t dev, |
| dev_t logdev, |
| dev_t rtdev, |
| unsigned int flags) |
| { |
| struct xfs_buf *bp; |
| struct xfs_sb *sbp; |
| xfs_daddr_t d; |
| unsigned int btflags = 0; |
| int error; |
| |
| |
| mp->m_features = xfs_sb_version_to_features(sb); |
| if (flags & LIBXFS_MOUNT_DEBUGGER) |
| xfs_set_debugger(mp); |
| if (flags & LIBXFS_MOUNT_REPORT_CORRUPTION) |
| xfs_set_reporting_corruption(mp); |
| if (flags & LIBXFS_MOUNT_CACHE_MISCOMPARE_PURGE) |
| btflags |= XFS_BUFTARG_MISCOMPARE_PURGE; |
| libxfs_buftarg_init(mp, dev, logdev, rtdev, btflags); |
| |
| mp->m_finobt_nores = true; |
| xfs_set_inode32(mp); |
| mp->m_sb = *sb; |
| INIT_RADIX_TREE(&mp->m_perag_tree, GFP_KERNEL); |
| sbp = &mp->m_sb; |
| spin_lock_init(&mp->m_sb_lock); |
| spin_lock_init(&mp->m_agirotor_lock); |
| |
| xfs_sb_mount_common(mp, sb); |
| |
| /* |
| * Set whether we're using stripe alignment. |
| */ |
| if (xfs_has_dalign(mp)) { |
| mp->m_dalign = sbp->sb_unit; |
| mp->m_swidth = sbp->sb_width; |
| } |
| |
| xfs_alloc_compute_maxlevels(mp); |
| xfs_bmap_compute_maxlevels(mp, XFS_DATA_FORK); |
| xfs_bmap_compute_maxlevels(mp, XFS_ATTR_FORK); |
| xfs_ialloc_setup_geometry(mp); |
| xfs_rmapbt_compute_maxlevels(mp); |
| xfs_rtrmapbt_compute_maxlevels(mp); |
| xfs_refcountbt_compute_maxlevels(mp); |
| xfs_rtrefcountbt_compute_maxlevels(mp); |
| |
| xfs_agbtree_compute_maxlevels(mp); |
| xfs_rtbtree_compute_maxlevels(mp); |
| |
| /* |
| * Check that the data (and log if separate) are an ok size. |
| */ |
| d = (xfs_daddr_t) XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks); |
| if (XFS_BB_TO_FSB(mp, d) != mp->m_sb.sb_dblocks) { |
| fprintf(stderr, _("%s: size check failed\n"), progname); |
| if (!xfs_is_debugger(mp)) |
| return NULL; |
| } |
| |
| /* |
| * We automatically convert v1 inodes to v2 inodes now, so if |
| * the NLINK bit is not set we can't operate on the filesystem. |
| */ |
| if (!(sbp->sb_versionnum & XFS_SB_VERSION_NLINKBIT)) { |
| |
| fprintf(stderr, _( |
| "%s: V1 inodes unsupported. Please try an older xfsprogs.\n"), |
| progname); |
| exit(1); |
| } |
| |
| /* Check for supported directory formats */ |
| if (!(sbp->sb_versionnum & XFS_SB_VERSION_DIRV2BIT)) { |
| |
| fprintf(stderr, _( |
| "%s: V1 directories unsupported. Please try an older xfsprogs.\n"), |
| progname); |
| exit(1); |
| } |
| |
| /* check for unsupported other features */ |
| if (!xfs_sb_good_version(sbp)) { |
| fprintf(stderr, _( |
| "%s: Unsupported features detected. Please try a newer xfsprogs.\n"), |
| progname); |
| exit(1); |
| } |
| |
| xfs_da_mount(mp); |
| |
| /* Initialize the precomputed transaction reservations values */ |
| xfs_trans_init(mp); |
| |
| if (dev == 0) /* maxtrres, we have no device so leave now */ |
| return mp; |
| |
| /* device size checks must pass unless we're a debugger. */ |
| error = libxfs_buf_read(mp->m_dev, d - XFS_FSS_TO_BB(mp, 1), |
| XFS_FSS_TO_BB(mp, 1), 0, &bp, NULL); |
| if (error) { |
| fprintf(stderr, _("%s: data size check failed\n"), progname); |
| if (!xfs_is_debugger(mp)) |
| goto out_da; |
| } else |
| libxfs_buf_relse(bp); |
| |
| if (mp->m_logdev_targp->bt_bdev && |
| mp->m_logdev_targp->bt_bdev != mp->m_ddev_targp->bt_bdev) { |
| d = (xfs_daddr_t) XFS_FSB_TO_BB(mp, mp->m_sb.sb_logblocks); |
| if (XFS_BB_TO_FSB(mp, d) != mp->m_sb.sb_logblocks || |
| libxfs_buf_read(mp->m_logdev_targp, |
| d - XFS_FSB_TO_BB(mp, 1), XFS_FSB_TO_BB(mp, 1), |
| 0, &bp, NULL)) { |
| fprintf(stderr, _("%s: log size checks failed\n"), |
| progname); |
| if (!xfs_is_debugger(mp)) |
| goto out_da; |
| } |
| if (bp) |
| libxfs_buf_relse(bp); |
| } |
| |
| /* Initialize realtime fields in the mount structure */ |
| if (rtmount_init(mp, flags)) { |
| fprintf(stderr, _("%s: realtime device init failed\n"), |
| progname); |
| goto out_da; |
| } |
| |
| /* |
| * libxfs_initialize_perag will allocate a perag structure for each ag. |
| * If agcount is corrupted and insanely high, this will OOM the box. |
| * If the agount seems (arbitrarily) high, try to read what would be |
| * the last AG, and if that fails for a relatively high agcount, just |
| * read the first one and let the user know to check the geometry. |
| */ |
| if (sbp->sb_agcount > 1000000) { |
| error = libxfs_buf_read(mp->m_dev, |
| XFS_AG_DADDR(mp, sbp->sb_agcount - 1, 0), 1, |
| 0, &bp, NULL); |
| if (error) { |
| fprintf(stderr, _("%s: read of AG %u failed\n"), |
| progname, sbp->sb_agcount); |
| if (!xfs_is_debugger(mp)) |
| goto out_da; |
| fprintf(stderr, _("%s: limiting reads to AG 0\n"), |
| progname); |
| sbp->sb_agcount = 1; |
| } else |
| libxfs_buf_relse(bp); |
| } |
| |
| error = libxfs_initialize_perag(mp, sbp->sb_agcount, &mp->m_maxagi); |
| if (error) { |
| fprintf(stderr, _("%s: perag init failed\n"), |
| progname); |
| exit(1); |
| } |
| xfs_set_perag_data_loaded(mp); |
| |
| libxfs_mountfs_imeta(mp); |
| |
| return mp; |
| out_da: |
| xfs_da_unmount(mp); |
| return NULL; |
| } |
| |
| void |
| libxfs_rtmount_destroy(xfs_mount_t *mp) |
| { |
| if (mp->m_rrefcountip) |
| libxfs_imeta_irele(mp->m_rrefcountip); |
| if (mp->m_rrmapip) |
| libxfs_imeta_irele(mp->m_rrmapip); |
| if (mp->m_rsumip) |
| libxfs_imeta_irele(mp->m_rsumip); |
| if (mp->m_rbmip) |
| libxfs_imeta_irele(mp->m_rbmip); |
| mp->m_rsumip = mp->m_rbmip = mp->m_rrmapip = mp->m_rrefcountip = NULL; |
| } |
| |
| /* Flush a device and report on writes that didn't make it to stable storage. */ |
| static inline int |
| libxfs_flush_buftarg( |
| struct xfs_buftarg *btp, |
| const char *buftarg_descr) |
| { |
| int error = 0; |
| int err2; |
| |
| /* |
| * Write verifier failures are evidence of a buggy program. Make sure |
| * that this state is always reported to the caller. |
| */ |
| if (btp->flags & XFS_BUFTARG_CORRUPT_WRITE) { |
| fprintf(stderr, |
| _("%s: Refusing to write a corrupt buffer to the %s!\n"), |
| progname, buftarg_descr); |
| error = -EFSCORRUPTED; |
| } |
| |
| if (btp->flags & XFS_BUFTARG_LOST_WRITE) { |
| fprintf(stderr, |
| _("%s: Lost a write to the %s!\n"), |
| progname, buftarg_descr); |
| if (!error) |
| error = -EIO; |
| } |
| |
| err2 = libxfs_blkdev_issue_flush(btp); |
| if (err2) { |
| fprintf(stderr, |
| _("%s: Flushing the %s failed, err=%d!\n"), |
| progname, buftarg_descr, -err2); |
| } |
| if (!error) |
| error = err2; |
| |
| return error; |
| } |
| |
| /* |
| * Flush all dirty buffers to stable storage and report on writes that didn't |
| * make it to stable storage. |
| */ |
| int |
| libxfs_flush_mount( |
| struct xfs_mount *mp) |
| { |
| int error = 0; |
| int err2; |
| |
| /* |
| * Flush the buffer cache to write all dirty buffers to disk. Buffers |
| * that fail write verification will cause the CORRUPT_WRITE flag to be |
| * set in the buftarg. Buffers that cannot be written will cause the |
| * LOST_WRITE flag to be set in the buftarg. Once that's done, |
| * instruct the disks to persist their write caches. |
| */ |
| libxfs_bcache_flush(mp); |
| |
| /* Flush all kernel and disk write caches, and report failures. */ |
| if (mp->m_ddev_targp) { |
| err2 = libxfs_flush_buftarg(mp->m_ddev_targp, _("data device")); |
| if (!error) |
| error = err2; |
| } |
| |
| if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp) { |
| err2 = libxfs_flush_buftarg(mp->m_logdev_targp, |
| _("log device")); |
| if (!error) |
| error = err2; |
| } |
| |
| if (mp->m_rtdev_targp) { |
| err2 = libxfs_flush_buftarg(mp->m_rtdev_targp, |
| _("realtime device")); |
| if (!error) |
| error = err2; |
| } |
| |
| return error; |
| } |
| |
| void |
| libxfs_buftarg_free( |
| struct xfs_buftarg *btp) |
| { |
| cache_destroy(btp->bcache); |
| kmem_free(btp); |
| } |
| |
| /* |
| * Release any resource obtained during a mount. |
| */ |
| int |
| libxfs_umount( |
| struct xfs_mount *mp) |
| { |
| int error; |
| |
| libxfs_rtmount_destroy(mp); |
| if (mp->m_metadirip) |
| libxfs_imeta_irele(mp->m_metadirip); |
| |
| /* |
| * Purge the buffer cache to write all dirty buffers to disk and free |
| * all incore buffers, then pick up the outcome when we tell the disks |
| * to persist their write caches. |
| */ |
| libxfs_bcache_purge(mp); |
| error = libxfs_flush_mount(mp); |
| |
| /* |
| * Only try to free the per-AG structures if we set them up in the |
| * first place. |
| */ |
| if (xfs_is_perag_data_loaded(mp)) |
| libxfs_free_perag(mp); |
| |
| xfs_da_unmount(mp); |
| libxfs_buftarg_free(mp->m_rtdev_targp); |
| if (mp->m_logdev_targp != mp->m_ddev_targp) |
| libxfs_buftarg_free(mp->m_logdev_targp); |
| libxfs_buftarg_free(mp->m_ddev_targp); |
| |
| return error; |
| } |
| |
| /* |
| * Release any global resources used by libxfs. |
| */ |
| void |
| libxfs_destroy( |
| struct libxfs_xinit *li) |
| { |
| kmem_start_leak_check(); |
| libxfs_close_devices(li); |
| |
| libxfs_bcache_free(); |
| destroy_kmem_caches(); |
| rcu_unregister_thread(); |
| |
| if (kmem_found_leaks()) |
| exit(1); |
| } |
| |
| int |
| libxfs_device_alignment(void) |
| { |
| return platform_align_blockdev(); |
| } |