blob: 004ed2a251e8fee5b530289abfbb0a2a5e89f004 [file] [log] [blame]
Dave Chinner0b61f8a2018-06-05 19:42:14 -07001// SPDX-License-Identifier: GPL-2.0
Linus Torvalds1da177e2005-04-16 15:20:36 -07002/*
Nathan Scott7b718762005-11-02 14:58:39 +11003 * Copyright (c) 2004-2005 Silicon Graphics, Inc.
4 * All Rights Reserved.
Linus Torvalds1da177e2005-04-16 15:20:36 -07005 */
Christoph Hellwig62e194e2009-01-19 02:03:03 +01006#include <linux/mount.h>
Darrick J. Wonge89c0412017-03-28 14:56:37 -07007#include <linux/fsmap.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -07008#include "xfs.h"
Linus Torvalds1da177e2005-04-16 15:20:36 -07009#include "xfs_fs.h"
Darrick J. Wong5467b342019-06-28 19:25:35 -070010#include "xfs_shared.h"
Dave Chinnera4fbe6a2013-10-23 10:51:50 +110011#include "xfs_format.h"
Dave Chinner239880e2013-10-23 10:50:10 +110012#include "xfs_log_format.h"
13#include "xfs_trans_resv.h"
Michal Marekfaa63e92007-07-11 11:10:19 +100014#include "xfs_mount.h"
Michal Marekfaa63e92007-07-11 11:10:19 +100015#include "xfs_inode.h"
Darrick J. Wong2810bd62019-07-02 09:39:40 -070016#include "xfs_iwalk.h"
Michal Marekfaa63e92007-07-11 11:10:19 +100017#include "xfs_itable.h"
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060018#include "xfs_fsops.h"
sandeen@sandeen.net471d5912008-11-25 21:20:10 -060019#include "xfs_rtalloc.h"
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -060020#include "xfs_attr.h"
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060021#include "xfs_ioctl.h"
David Chinnera8272ce2007-11-23 16:28:09 +110022#include "xfs_ioctl32.h"
Christoph Hellwig0b1b2132009-12-14 23:14:59 +000023#include "xfs_trace.h"
Darrick J. Wongc368ebc2018-01-08 10:51:27 -080024#include "xfs_sb.h"
Linus Torvalds1da177e2005-04-16 15:20:36 -070025
Eric Sandeen526c4202005-09-05 08:25:06 +100026#define _NATIVE_IOC(cmd, type) \
27 _IOC(_IOC_DIR(cmd), _IOC_TYPE(cmd), _IOC_NR(cmd), sizeof(type))
28
sandeen@sandeen.netffae2632008-11-25 21:20:07 -060029#ifdef BROKEN_X86_ALIGNMENT
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060030STATIC int
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060031xfs_compat_ioc_fsgeometry_v1(
32 struct xfs_mount *mp,
33 compat_xfs_fsop_geom_v1_t __user *arg32)
Michal Marek547e00c2007-07-11 11:09:57 +100034{
Dave Chinner1b6d9682019-04-12 07:41:16 -070035 struct xfs_fsop_geom fsgeo;
Michal Marek547e00c2007-07-11 11:09:57 +100036
Dave Chinner03288b12021-08-18 18:46:54 -070037 xfs_fs_geometry(mp, &fsgeo, 3);
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060038 /* The 32-bit variant simply has some padding at the end */
39 if (copy_to_user(arg32, &fsgeo, sizeof(struct compat_xfs_fsop_geom_v1)))
Eric Sandeenb474c7a2014-06-22 15:04:54 +100040 return -EFAULT;
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060041 return 0;
Michal Marek547e00c2007-07-11 11:09:57 +100042}
43
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -060044STATIC int
sandeen@sandeen.net471d5912008-11-25 21:20:10 -060045xfs_compat_growfs_data_copyin(
46 struct xfs_growfs_data *in,
47 compat_xfs_growfs_data_t __user *arg32)
48{
49 if (get_user(in->newblocks, &arg32->newblocks) ||
50 get_user(in->imaxpct, &arg32->imaxpct))
Eric Sandeenb474c7a2014-06-22 15:04:54 +100051 return -EFAULT;
sandeen@sandeen.net471d5912008-11-25 21:20:10 -060052 return 0;
53}
54
55STATIC int
56xfs_compat_growfs_rt_copyin(
57 struct xfs_growfs_rt *in,
58 compat_xfs_growfs_rt_t __user *arg32)
59{
60 if (get_user(in->newblocks, &arg32->newblocks) ||
61 get_user(in->extsize, &arg32->extsize))
Eric Sandeenb474c7a2014-06-22 15:04:54 +100062 return -EFAULT;
sandeen@sandeen.net471d5912008-11-25 21:20:10 -060063 return 0;
64}
65
66STATIC int
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -070067xfs_fsinumbers_fmt_compat(
Darrick J. Wong5f19c7f2019-07-03 20:36:27 -070068 struct xfs_ibulk *breq,
69 const struct xfs_inumbers *ig)
Michal Marekfaa63e92007-07-11 11:10:19 +100070{
Darrick J. Wong5f19c7f2019-07-03 20:36:27 -070071 struct compat_xfs_inogrp __user *p32 = breq->ubuffer;
72 struct xfs_inogrp ig1;
73 struct xfs_inogrp *igrp = &ig1;
74
75 xfs_inumbers_to_inogrp(&ig1, ig);
Michal Marekfaa63e92007-07-11 11:10:19 +100076
Darrick J. Wong677717f2019-07-02 09:39:43 -070077 if (put_user(igrp->xi_startino, &p32->xi_startino) ||
78 put_user(igrp->xi_alloccount, &p32->xi_alloccount) ||
79 put_user(igrp->xi_allocmask, &p32->xi_allocmask))
80 return -EFAULT;
81
82 return xfs_ibulk_advance(breq, sizeof(struct compat_xfs_inogrp));
Michal Marekfaa63e92007-07-11 11:10:19 +100083}
84
Linus Torvalds1da177e2005-04-16 15:20:36 -070085#else
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -070086#define xfs_fsinumbers_fmt_compat xfs_fsinumbers_fmt
sandeen@sandeen.nete5d412f2008-11-25 21:20:17 -060087#endif /* BROKEN_X86_ALIGNMENT */
Michal Marekfaa63e92007-07-11 11:10:19 +100088
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -060089STATIC int
90xfs_ioctl32_bstime_copyin(
91 xfs_bstime_t *bstime,
92 compat_xfs_bstime_t __user *bstime32)
93{
Arnd Bergmann3b62f002020-01-02 13:27:44 -080094 old_time32_t sec32; /* tv_sec differs on 64 vs. 32 */
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -060095
96 if (get_user(sec32, &bstime32->tv_sec) ||
97 get_user(bstime->tv_nsec, &bstime32->tv_nsec))
Eric Sandeenb474c7a2014-06-22 15:04:54 +100098 return -EFAULT;
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -060099 bstime->tv_sec = sec32;
100 return 0;
101}
102
Darrick J. Wong6f71fb62019-07-03 20:36:25 -0700103/*
104 * struct xfs_bstat has differing alignment on intel, & bstime_t sizes
105 * everywhere
106 */
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -0600107STATIC int
108xfs_ioctl32_bstat_copyin(
Darrick J. Wong6f71fb62019-07-03 20:36:25 -0700109 struct xfs_bstat *bstat,
110 struct compat_xfs_bstat __user *bstat32)
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -0600111{
112 if (get_user(bstat->bs_ino, &bstat32->bs_ino) ||
113 get_user(bstat->bs_mode, &bstat32->bs_mode) ||
114 get_user(bstat->bs_nlink, &bstat32->bs_nlink) ||
115 get_user(bstat->bs_uid, &bstat32->bs_uid) ||
116 get_user(bstat->bs_gid, &bstat32->bs_gid) ||
117 get_user(bstat->bs_rdev, &bstat32->bs_rdev) ||
118 get_user(bstat->bs_blksize, &bstat32->bs_blksize) ||
119 get_user(bstat->bs_size, &bstat32->bs_size) ||
120 xfs_ioctl32_bstime_copyin(&bstat->bs_atime, &bstat32->bs_atime) ||
121 xfs_ioctl32_bstime_copyin(&bstat->bs_mtime, &bstat32->bs_mtime) ||
122 xfs_ioctl32_bstime_copyin(&bstat->bs_ctime, &bstat32->bs_ctime) ||
123 get_user(bstat->bs_blocks, &bstat32->bs_size) ||
124 get_user(bstat->bs_xflags, &bstat32->bs_size) ||
125 get_user(bstat->bs_extsize, &bstat32->bs_extsize) ||
126 get_user(bstat->bs_extents, &bstat32->bs_extents) ||
127 get_user(bstat->bs_gen, &bstat32->bs_gen) ||
Arkadiusz Mi?kiewicz67430992010-09-26 06:10:18 +0000128 get_user(bstat->bs_projid_lo, &bstat32->bs_projid_lo) ||
129 get_user(bstat->bs_projid_hi, &bstat32->bs_projid_hi) ||
Dave Chinnerb1d6cc02014-10-02 09:17:58 +1000130 get_user(bstat->bs_forkoff, &bstat32->bs_forkoff) ||
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -0600131 get_user(bstat->bs_dmevmask, &bstat32->bs_dmevmask) ||
132 get_user(bstat->bs_dmstate, &bstat32->bs_dmstate) ||
133 get_user(bstat->bs_aextents, &bstat32->bs_aextents))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000134 return -EFAULT;
sandeen@sandeen.nete94fc4a2008-11-25 21:20:09 -0600135 return 0;
136}
137
Michal Marekfaa63e92007-07-11 11:10:19 +1000138/* XFS_IOC_FSBULKSTAT and friends */
139
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600140STATIC int
141xfs_bstime_store_compat(
142 compat_xfs_bstime_t __user *p32,
143 const xfs_bstime_t *p)
Michal Marekfaa63e92007-07-11 11:10:19 +1000144{
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600145 __s32 sec32;
Michal Marekfaa63e92007-07-11 11:10:19 +1000146
147 sec32 = p->tv_sec;
148 if (put_user(sec32, &p32->tv_sec) ||
149 put_user(p->tv_nsec, &p32->tv_nsec))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000150 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000151 return 0;
152}
153
sandeen@sandeen.net65fbaf22008-11-25 21:20:12 -0600154/* Return 0 on success or positive error (to xfs_bulkstat()) */
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600155STATIC int
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -0700156xfs_fsbulkstat_one_fmt_compat(
Darrick J. Wong7035f972019-07-03 20:36:26 -0700157 struct xfs_ibulk *breq,
158 const struct xfs_bulkstat *bstat)
Michal Marekfaa63e92007-07-11 11:10:19 +1000159{
Darrick J. Wong7035f972019-07-03 20:36:26 -0700160 struct compat_xfs_bstat __user *p32 = breq->ubuffer;
161 struct xfs_bstat bs1;
162 struct xfs_bstat *buffer = &bs1;
163
164 xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat);
sandeen@sandeen.net65fbaf22008-11-25 21:20:12 -0600165
166 if (put_user(buffer->bs_ino, &p32->bs_ino) ||
167 put_user(buffer->bs_mode, &p32->bs_mode) ||
168 put_user(buffer->bs_nlink, &p32->bs_nlink) ||
169 put_user(buffer->bs_uid, &p32->bs_uid) ||
170 put_user(buffer->bs_gid, &p32->bs_gid) ||
171 put_user(buffer->bs_rdev, &p32->bs_rdev) ||
172 put_user(buffer->bs_blksize, &p32->bs_blksize) ||
173 put_user(buffer->bs_size, &p32->bs_size) ||
Michal Marekfaa63e92007-07-11 11:10:19 +1000174 xfs_bstime_store_compat(&p32->bs_atime, &buffer->bs_atime) ||
175 xfs_bstime_store_compat(&p32->bs_mtime, &buffer->bs_mtime) ||
176 xfs_bstime_store_compat(&p32->bs_ctime, &buffer->bs_ctime) ||
sandeen@sandeen.net65fbaf22008-11-25 21:20:12 -0600177 put_user(buffer->bs_blocks, &p32->bs_blocks) ||
178 put_user(buffer->bs_xflags, &p32->bs_xflags) ||
179 put_user(buffer->bs_extsize, &p32->bs_extsize) ||
180 put_user(buffer->bs_extents, &p32->bs_extents) ||
181 put_user(buffer->bs_gen, &p32->bs_gen) ||
182 put_user(buffer->bs_projid, &p32->bs_projid) ||
Arkadiusz Mi?kiewicz67430992010-09-26 06:10:18 +0000183 put_user(buffer->bs_projid_hi, &p32->bs_projid_hi) ||
Dave Chinnerb1d6cc02014-10-02 09:17:58 +1000184 put_user(buffer->bs_forkoff, &p32->bs_forkoff) ||
sandeen@sandeen.net65fbaf22008-11-25 21:20:12 -0600185 put_user(buffer->bs_dmevmask, &p32->bs_dmevmask) ||
186 put_user(buffer->bs_dmstate, &p32->bs_dmstate) ||
Michal Marekfaa63e92007-07-11 11:10:19 +1000187 put_user(buffer->bs_aextents, &p32->bs_aextents))
Dave Chinner24513372014-06-25 14:58:08 +1000188 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000189
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700190 return xfs_ibulk_advance(breq, sizeof(struct compat_xfs_bstat));
sandeen@sandeen.net2ee4fa52008-11-25 21:20:11 -0600191}
192
Michal Marekfaa63e92007-07-11 11:10:19 +1000193/* copied from xfs_ioctl.c */
194STATIC int
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -0700195xfs_compat_ioc_fsbulkstat(
Christoph Hellwigf736d932021-01-21 14:19:58 +0100196 struct file *file,
sandeen@sandeen.net2ee4fa52008-11-25 21:20:11 -0600197 unsigned int cmd,
Darrick J. Wong6f71fb62019-07-03 20:36:25 -0700198 struct compat_xfs_fsop_bulkreq __user *p32)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700199{
Christoph Hellwigf736d932021-01-21 14:19:58 +0100200 struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700201 u32 addr;
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700202 struct xfs_fsop_bulkreq bulkreq;
203 struct xfs_ibulk breq = {
204 .mp = mp,
Christoph Hellwigf736d932021-01-21 14:19:58 +0100205 .mnt_userns = file_mnt_user_ns(file),
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700206 .ocount = 0,
207 };
208 xfs_ino_t lastino;
Michal Marekfaa63e92007-07-11 11:10:19 +1000209 int error;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700210
Nick Bowler7ca860e2018-12-17 09:35:27 -0800211 /*
212 * Output structure handling functions. Depending on the command,
213 * either the xfs_bstat and xfs_inogrp structures are written out
214 * to userpace memory via bulkreq.ubuffer. Normally the compat
215 * functions and structure size are the correct ones to use ...
216 */
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -0700217 inumbers_fmt_pf inumbers_func = xfs_fsinumbers_fmt_compat;
218 bulkstat_one_fmt_pf bs_one_func = xfs_fsbulkstat_one_fmt_compat;
Nick Bowler7ca860e2018-12-17 09:35:27 -0800219
220#ifdef CONFIG_X86_X32
221 if (in_x32_syscall()) {
222 /*
223 * ... but on x32 the input xfs_fsop_bulkreq has pointers
224 * which must be handled in the "compat" (32-bit) way, while
225 * the xfs_bstat and xfs_inogrp structures follow native 64-
226 * bit layout convention. So adjust accordingly, otherwise
227 * the data written out in compat layout will not match what
228 * x32 userspace expects.
229 */
Darrick J. Wong8bfe9d12019-07-03 20:36:26 -0700230 inumbers_func = xfs_fsinumbers_fmt;
231 bs_one_func = xfs_fsbulkstat_one_fmt;
Nick Bowler7ca860e2018-12-17 09:35:27 -0800232 }
233#endif
234
Michal Marekfaa63e92007-07-11 11:10:19 +1000235 /* done = 1 if there are more stats to get and if bulkstat */
236 /* should be called again (unused here, but used in dmapi) */
237
238 if (!capable(CAP_SYS_ADMIN))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000239 return -EPERM;
Michal Marekfaa63e92007-07-11 11:10:19 +1000240
Dave Chinner75c8c50f2021-08-18 18:46:53 -0700241 if (xfs_is_shutdown(mp))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000242 return -EIO;
Michal Marekfaa63e92007-07-11 11:10:19 +1000243
244 if (get_user(addr, &p32->lastip))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000245 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000246 bulkreq.lastip = compat_ptr(addr);
247 if (get_user(bulkreq.icount, &p32->icount) ||
248 get_user(addr, &p32->ubuffer))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000249 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000250 bulkreq.ubuffer = compat_ptr(addr);
251 if (get_user(addr, &p32->ocount))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000252 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000253 bulkreq.ocount = compat_ptr(addr);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700254
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700255 if (copy_from_user(&lastino, bulkreq.lastip, sizeof(__s64)))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000256 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000257
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700258 if (bulkreq.icount <= 0)
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000259 return -EINVAL;
Michal Marekfaa63e92007-07-11 11:10:19 +1000260
Lachlan McIlroycd57e592007-11-23 16:30:32 +1100261 if (bulkreq.ubuffer == NULL)
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000262 return -EINVAL;
Lachlan McIlroycd57e592007-11-23 16:30:32 +1100263
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700264 breq.ubuffer = bulkreq.ubuffer;
265 breq.icount = bulkreq.icount;
sandeen@sandeen.netaf819d22008-11-25 21:20:13 -0600266
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700267 /*
268 * FSBULKSTAT_SINGLE expects that *lastip contains the inode number
269 * that we want to stat. However, FSINUMBERS and FSBULKSTAT expect
270 * that *lastip contains either zero or the number of the last inode to
271 * be examined by the previous call and return results starting with
272 * the next inode after that. The new bulk request back end functions
273 * take the inode to start with, so we have to compute the startino
274 * parameter from lastino to maintain correct function. lastino == 0
275 * is a special case because it has traditionally meant "first inode
276 * in filesystem".
277 */
278 if (cmd == XFS_IOC_FSINUMBERS_32) {
Darrick J. Wong677717f2019-07-02 09:39:43 -0700279 breq.startino = lastino ? lastino + 1 : 0;
280 error = xfs_inumbers(&breq, inumbers_func);
281 lastino = breq.startino - 1;
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700282 } else if (cmd == XFS_IOC_FSBULKSTAT_SINGLE_32) {
283 breq.startino = lastino;
284 breq.icount = 1;
285 error = xfs_bulkstat_one(&breq, bs_one_func);
286 lastino = breq.startino;
sandeen@sandeen.netaf819d22008-11-25 21:20:13 -0600287 } else if (cmd == XFS_IOC_FSBULKSTAT_32) {
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700288 breq.startino = lastino ? lastino + 1 : 0;
289 error = xfs_bulkstat(&breq, bs_one_func);
290 lastino = breq.startino - 1;
291 } else {
Dave Chinner24513372014-06-25 14:58:08 +1000292 error = -EINVAL;
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700293 }
Michal Marekfaa63e92007-07-11 11:10:19 +1000294 if (error)
Dave Chinner24513372014-06-25 14:58:08 +1000295 return error;
Michal Marekfaa63e92007-07-11 11:10:19 +1000296
Darrick J. Wongf16fe3e2019-07-02 09:39:39 -0700297 if (bulkreq.lastip != NULL &&
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700298 copy_to_user(bulkreq.lastip, &lastino, sizeof(xfs_ino_t)))
Darrick J. Wongf16fe3e2019-07-02 09:39:39 -0700299 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000300
Darrick J. Wongf16fe3e2019-07-02 09:39:39 -0700301 if (bulkreq.ocount != NULL &&
Darrick J. Wong2810bd62019-07-02 09:39:40 -0700302 copy_to_user(bulkreq.ocount, &breq.ocount, sizeof(__s32)))
Darrick J. Wongf16fe3e2019-07-02 09:39:39 -0700303 return -EFAULT;
Michal Marekfaa63e92007-07-11 11:10:19 +1000304
305 return 0;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700306}
Michal Marekfaa63e92007-07-11 11:10:19 +1000307
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600308STATIC int
309xfs_compat_handlereq_copyin(
310 xfs_fsop_handlereq_t *hreq,
311 compat_xfs_fsop_handlereq_t __user *arg32)
Michal Marek1fa503d2007-07-11 11:10:09 +1000312{
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600313 compat_xfs_fsop_handlereq_t hreq32;
Michal Marek1fa503d2007-07-11 11:10:09 +1000314
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600315 if (copy_from_user(&hreq32, arg32, sizeof(compat_xfs_fsop_handlereq_t)))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000316 return -EFAULT;
Michal Marek1fa503d2007-07-11 11:10:09 +1000317
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600318 hreq->fd = hreq32.fd;
319 hreq->path = compat_ptr(hreq32.path);
320 hreq->oflags = hreq32.oflags;
321 hreq->ihandle = compat_ptr(hreq32.ihandle);
322 hreq->ihandlen = hreq32.ihandlen;
323 hreq->ohandle = compat_ptr(hreq32.ohandle);
324 hreq->ohandlen = compat_ptr(hreq32.ohandlen);
325
326 return 0;
Michal Marek1fa503d2007-07-11 11:10:09 +1000327}
328
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100329STATIC struct dentry *
330xfs_compat_handlereq_to_dentry(
331 struct file *parfilp,
332 compat_xfs_fsop_handlereq_t *hreq)
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600333{
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100334 return xfs_handle_to_dentry(parfilp,
335 compat_ptr(hreq->ihandle), hreq->ihandlen);
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600336}
337
338STATIC int
339xfs_compat_attrlist_by_handle(
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100340 struct file *parfilp,
Christoph Hellwig53ac39f2020-02-26 17:30:41 -0800341 compat_xfs_fsop_attrlist_handlereq_t __user *p)
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600342{
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600343 compat_xfs_fsop_attrlist_handlereq_t al_hreq;
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100344 struct dentry *dentry;
Christoph Hellwig53ac39f2020-02-26 17:30:41 -0800345 int error;
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600346
347 if (!capable(CAP_SYS_ADMIN))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000348 return -EPERM;
Christoph Hellwig53ac39f2020-02-26 17:30:41 -0800349 if (copy_from_user(&al_hreq, p, sizeof(al_hreq)))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000350 return -EFAULT;
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600351
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100352 dentry = xfs_compat_handlereq_to_dentry(parfilp, &al_hreq.hreq);
353 if (IS_ERR(dentry))
354 return PTR_ERR(dentry);
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600355
Christoph Hellwigeb241c72020-02-26 17:30:40 -0800356 error = xfs_ioc_attr_list(XFS_I(d_inode(dentry)),
357 compat_ptr(al_hreq.buffer), al_hreq.buflen,
Christoph Hellwig53ac39f2020-02-26 17:30:41 -0800358 al_hreq.flags, &p->pos);
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100359 dput(dentry);
360 return error;
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600361}
362
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600363STATIC int
364xfs_compat_attrmulti_by_handle(
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100365 struct file *parfilp,
366 void __user *arg)
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600367{
368 int error;
369 compat_xfs_attr_multiop_t *ops;
370 compat_xfs_fsop_attrmulti_handlereq_t am_hreq;
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100371 struct dentry *dentry;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600372 unsigned int i, size;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600373
374 if (!capable(CAP_SYS_ADMIN))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000375 return -EPERM;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600376 if (copy_from_user(&am_hreq, arg,
377 sizeof(compat_xfs_fsop_attrmulti_handlereq_t)))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000378 return -EFAULT;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600379
Zhitong Wangfda168c2010-03-23 09:51:22 +1100380 /* overflow check */
381 if (am_hreq.opcount >= INT_MAX / sizeof(compat_xfs_attr_multiop_t))
382 return -E2BIG;
383
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100384 dentry = xfs_compat_handlereq_to_dentry(parfilp, &am_hreq.hreq);
385 if (IS_ERR(dentry))
386 return PTR_ERR(dentry);
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600387
Dave Chinner24513372014-06-25 14:58:08 +1000388 error = -E2BIG;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600389 size = am_hreq.opcount * sizeof(compat_xfs_attr_multiop_t);
390 if (!size || size > 16 * PAGE_SIZE)
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100391 goto out_dput;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600392
Li Zefan0e639bd2009-04-08 15:08:04 +0800393 ops = memdup_user(compat_ptr(am_hreq.ops), size);
394 if (IS_ERR(ops)) {
Brian Foster4d949022015-01-22 10:04:24 +1100395 error = PTR_ERR(ops);
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100396 goto out_dput;
Li Zefan0e639bd2009-04-08 15:08:04 +0800397 }
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600398
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600399 error = 0;
400 for (i = 0; i < am_hreq.opcount; i++) {
Christoph Hellwigd0ce64392020-02-26 17:30:31 -0800401 ops[i].am_error = xfs_ioc_attrmulti_one(parfilp,
402 d_inode(dentry), ops[i].am_opcode,
403 compat_ptr(ops[i].am_attrname),
404 compat_ptr(ops[i].am_attrvalue),
405 &ops[i].am_length, ops[i].am_flags);
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600406 }
407
408 if (copy_to_user(compat_ptr(am_hreq.ops), ops, size))
Dave Chinner24513372014-06-25 14:58:08 +1000409 error = -EFAULT;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600410
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600411 kfree(ops);
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100412 out_dput:
413 dput(dentry);
Dave Chinner24513372014-06-25 14:58:08 +1000414 return error;
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600415}
416
Christoph Hellwig4d4be482008-12-09 04:47:33 -0500417long
418xfs_file_compat_ioctl(
419 struct file *filp,
420 unsigned cmd,
421 unsigned long p)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700422{
Al Viro496ad9a2013-01-23 17:07:38 -0500423 struct inode *inode = file_inode(filp);
Christoph Hellwig4d4be482008-12-09 04:47:33 -0500424 struct xfs_inode *ip = XFS_I(inode);
Christoph Hellwig4529e6d2019-08-15 23:41:06 -0700425 void __user *arg = compat_ptr(p);
Christoph Hellwig4d4be482008-12-09 04:47:33 -0500426 int error;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700427
Christoph Hellwigcca28fb2010-06-24 11:57:09 +1000428 trace_xfs_file_compat_ioctl(ip);
Christoph Hellwig4d4be482008-12-09 04:47:33 -0500429
Linus Torvalds1da177e2005-04-16 15:20:36 -0700430 switch (cmd) {
Nick Bowlera9d25bd2018-12-17 09:35:27 -0800431#if defined(BROKEN_X86_ALIGNMENT)
Michal Marek547e00c2007-07-11 11:09:57 +1000432 case XFS_IOC_FSGEOMETRY_V1_32:
Christoph Hellwigf69e8092021-02-03 18:30:10 +0100433 return xfs_compat_ioc_fsgeometry_v1(ip->i_mount, arg);
sandeen@sandeen.net471d5912008-11-25 21:20:10 -0600434 case XFS_IOC_FSGROWFSDATA_32: {
435 struct xfs_growfs_data in;
436
437 if (xfs_compat_growfs_data_copyin(&in, arg))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000438 return -EFAULT;
Jan Karad9457dc2012-06-12 16:20:39 +0200439 error = mnt_want_write_file(filp);
440 if (error)
441 return error;
Christoph Hellwigf69e8092021-02-03 18:30:10 +0100442 error = xfs_growfs_data(ip->i_mount, &in);
Jan Karad9457dc2012-06-12 16:20:39 +0200443 mnt_drop_write_file(filp);
Dave Chinner24513372014-06-25 14:58:08 +1000444 return error;
sandeen@sandeen.net471d5912008-11-25 21:20:10 -0600445 }
446 case XFS_IOC_FSGROWFSRT_32: {
447 struct xfs_growfs_rt in;
448
449 if (xfs_compat_growfs_rt_copyin(&in, arg))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000450 return -EFAULT;
Jan Karad9457dc2012-06-12 16:20:39 +0200451 error = mnt_want_write_file(filp);
452 if (error)
453 return error;
Christoph Hellwigf69e8092021-02-03 18:30:10 +0100454 error = xfs_growfs_rt(ip->i_mount, &in);
Jan Karad9457dc2012-06-12 16:20:39 +0200455 mnt_drop_write_file(filp);
Dave Chinner24513372014-06-25 14:58:08 +1000456 return error;
sandeen@sandeen.net471d5912008-11-25 21:20:10 -0600457 }
Linus Torvalds1da177e2005-04-16 15:20:36 -0700458#endif
sandeen@sandeen.nete5d412f2008-11-25 21:20:17 -0600459 /* long changes size, but xfs only copiese out 32 bits */
sandeen@sandeen.nete5d412f2008-11-25 21:20:17 -0600460 case XFS_IOC_GETVERSION_32:
461 cmd = _NATIVE_IOC(cmd, long);
Christoph Hellwig4d4be482008-12-09 04:47:33 -0500462 return xfs_file_ioctl(filp, cmd, p);
Christoph Hellwig37258672009-09-01 14:03:08 -0400463 case XFS_IOC_SWAPEXT_32: {
sandeen@sandeen.nete5d412f2008-11-25 21:20:17 -0600464 struct xfs_swapext sxp;
465 struct compat_xfs_swapext __user *sxu = arg;
466
467 /* Bulk copy in up to the sx_stat field, then copy bstat */
468 if (copy_from_user(&sxp, sxu,
469 offsetof(struct xfs_swapext, sx_stat)) ||
470 xfs_ioctl32_bstat_copyin(&sxp.sx_stat, &sxu->sx_stat))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000471 return -EFAULT;
Jan Karad9457dc2012-06-12 16:20:39 +0200472 error = mnt_want_write_file(filp);
473 if (error)
474 return error;
Dave Chinnera133d952013-08-12 20:49:48 +1000475 error = xfs_ioc_swapext(&sxp);
Jan Karad9457dc2012-06-12 16:20:39 +0200476 mnt_drop_write_file(filp);
Dave Chinner24513372014-06-25 14:58:08 +1000477 return error;
sandeen@sandeen.nete5d412f2008-11-25 21:20:17 -0600478 }
Michal Marekfaa63e92007-07-11 11:10:19 +1000479 case XFS_IOC_FSBULKSTAT_32:
480 case XFS_IOC_FSBULKSTAT_SINGLE_32:
481 case XFS_IOC_FSINUMBERS_32:
Christoph Hellwigf736d932021-01-21 14:19:58 +0100482 return xfs_compat_ioc_fsbulkstat(filp, cmd, arg);
Michal Marek1fa503d2007-07-11 11:10:09 +1000483 case XFS_IOC_FD_TO_HANDLE_32:
484 case XFS_IOC_PATH_TO_HANDLE_32:
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600485 case XFS_IOC_PATH_TO_FSHANDLE_32: {
486 struct xfs_fsop_handlereq hreq;
487
488 if (xfs_compat_handlereq_copyin(&hreq, arg))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000489 return -EFAULT;
Michal Marek1fa503d2007-07-11 11:10:09 +1000490 cmd = _NATIVE_IOC(cmd, struct xfs_fsop_handlereq);
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600491 return xfs_find_handle(cmd, &hreq);
492 }
493 case XFS_IOC_OPEN_BY_HANDLE_32: {
494 struct xfs_fsop_handlereq hreq;
495
496 if (xfs_compat_handlereq_copyin(&hreq, arg))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000497 return -EFAULT;
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100498 return xfs_open_by_handle(filp, &hreq);
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600499 }
500 case XFS_IOC_READLINK_BY_HANDLE_32: {
501 struct xfs_fsop_handlereq hreq;
502
503 if (xfs_compat_handlereq_copyin(&hreq, arg))
Eric Sandeenb474c7a2014-06-22 15:04:54 +1000504 return -EFAULT;
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100505 return xfs_readlink_by_handle(filp, &hreq);
sandeen@sandeen.netd5547f92008-11-25 21:20:08 -0600506 }
sandeen@sandeen.netebeecd22008-11-25 21:20:14 -0600507 case XFS_IOC_ATTRLIST_BY_HANDLE_32:
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100508 return xfs_compat_attrlist_by_handle(filp, arg);
sandeen@sandeen.net28750972008-11-25 21:20:15 -0600509 case XFS_IOC_ATTRMULTI_BY_HANDLE_32:
Christoph Hellwigab596ad2009-01-19 02:02:57 +0100510 return xfs_compat_attrmulti_by_handle(filp, arg);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700511 default:
Christoph Hellwig314e01a2019-08-15 23:41:06 -0700512 /* try the native version */
Christoph Hellwig4529e6d2019-08-15 23:41:06 -0700513 return xfs_file_ioctl(filp, cmd, (unsigned long)arg);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700514 }
Linus Torvalds1da177e2005-04-16 15:20:36 -0700515}