blob: e4c8f603dd58891ed1b8125a07cbfac0a05086bc [file] [log] [blame]
Christoph Probsta205d502019-05-08 21:36:25 +02001// SPDX-License-Identifier: GPL-2.0
Steve French1080ef72011-02-24 18:07:19 +00002/*
3 * SMB2 version specific operations
4 *
5 * Copyright (c) 2012, Jeff Layton <jlayton@redhat.com>
Steve French1080ef72011-02-24 18:07:19 +00006 */
7
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -07008#include <linux/pagemap.h>
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07009#include <linux/vfs.h>
Steve Frenchf29ebb42014-07-19 21:44:58 -050010#include <linux/falloc.h>
Pavel Shilovsky026e93d2016-11-03 16:47:37 -070011#include <linux/scatterlist.h>
Tobias Regnery4fa8e502017-03-30 12:34:14 +020012#include <linux/uuid.h>
Aurelien Aptel35adffe2019-09-20 06:29:39 +020013#include <linux/sort.h>
Pavel Shilovsky026e93d2016-11-03 16:47:37 -070014#include <crypto/aead.h>
Christoph Hellwig10c5db22020-05-23 09:30:11 +020015#include <linux/fiemap.h>
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +100016#include "cifsfs.h"
Steve French1080ef72011-02-24 18:07:19 +000017#include "cifsglob.h"
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +040018#include "smb2pdu.h"
19#include "smb2proto.h"
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040020#include "cifsproto.h"
21#include "cifs_debug.h"
Pavel Shilovskyb42bf882013-08-14 19:25:21 +040022#include "cifs_unicode.h"
Pavel Shilovsky2e44b282012-09-18 16:20:33 -070023#include "smb2status.h"
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -070024#include "smb2glob.h"
Steve French834170c2016-09-30 21:14:26 -050025#include "cifs_ioctl.h"
Long Li09902f82017-11-22 17:38:39 -070026#include "smbdirect.h"
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -060027#include "fs_context.h"
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040028
Pavel Shilovskyef68e832019-01-18 17:25:36 -080029/* Change credits for different ops and return the total number of credits */
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040030static int
31change_conf(struct TCP_Server_Info *server)
32{
33 server->credits += server->echo_credits + server->oplock_credits;
34 server->oplock_credits = server->echo_credits = 0;
35 switch (server->credits) {
36 case 0:
Pavel Shilovskyef68e832019-01-18 17:25:36 -080037 return 0;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040038 case 1:
39 server->echoes = false;
40 server->oplocks = false;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040041 break;
42 case 2:
43 server->echoes = true;
44 server->oplocks = false;
45 server->echo_credits = 1;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040046 break;
47 default:
48 server->echoes = true;
Steve Frenche0ddde92015-09-22 09:29:38 -050049 if (enable_oplocks) {
50 server->oplocks = true;
51 server->oplock_credits = 1;
52 } else
53 server->oplocks = false;
54
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040055 server->echo_credits = 1;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040056 }
57 server->credits -= server->echo_credits + server->oplock_credits;
Pavel Shilovskyef68e832019-01-18 17:25:36 -080058 return server->credits + server->echo_credits + server->oplock_credits;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040059}
60
61static void
Pavel Shilovsky335b7b62019-01-16 11:12:41 -080062smb2_add_credits(struct TCP_Server_Info *server,
63 const struct cifs_credits *credits, const int optype)
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040064{
Pavel Shilovskyef68e832019-01-18 17:25:36 -080065 int *val, rc = -1;
Shyam Prasad N6d82c272021-02-03 23:20:46 -080066 int scredits, in_flight;
Pavel Shilovsky335b7b62019-01-16 11:12:41 -080067 unsigned int add = credits->value;
68 unsigned int instance = credits->instance;
69 bool reconnect_detected = false;
Shyam Prasad N6d82c272021-02-03 23:20:46 -080070 bool reconnect_with_invalid_credits = false;
Pavel Shilovskyef68e832019-01-18 17:25:36 -080071
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040072 spin_lock(&server->req_lock);
73 val = server->ops->get_credits_field(server, optype);
Steve Frenchb340a4d2018-09-01 01:10:17 -050074
75 /* eg found case where write overlapping reconnect messed up credits */
76 if (((optype & CIFS_OP_MASK) == CIFS_NEG_OP) && (*val != 0))
Shyam Prasad N6d82c272021-02-03 23:20:46 -080077 reconnect_with_invalid_credits = true;
78
Pavel Shilovsky335b7b62019-01-16 11:12:41 -080079 if ((instance == 0) || (instance == server->reconnect_instance))
80 *val += add;
81 else
82 reconnect_detected = true;
Steve Frenchb340a4d2018-09-01 01:10:17 -050083
Steve French141891f2016-09-23 00:44:16 -050084 if (*val > 65000) {
85 *val = 65000; /* Don't get near 64K credits, avoid srv bugs */
Joe Perchesa0a30362020-04-14 22:42:53 -070086 pr_warn_once("server overflowed SMB3 credits\n");
Steve French141891f2016-09-23 00:44:16 -050087 }
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040088 server->in_flight--;
Shyam Prasad N0f56db82021-02-03 22:49:52 -080089 if (server->in_flight == 0 &&
90 ((optype & CIFS_OP_MASK) != CIFS_NEG_OP) &&
91 ((optype & CIFS_OP_MASK) != CIFS_SESS_OP))
Pavel Shilovsky28ea5292012-05-23 16:18:00 +040092 rc = change_conf(server);
Pavel Shilovsky983c88a2012-09-18 16:20:33 -070093 /*
94 * Sometimes server returns 0 credits on oplock break ack - we need to
95 * rebalance credits in this case.
96 */
97 else if (server->in_flight > 0 && server->oplock_credits == 0 &&
98 server->oplocks) {
99 if (server->credits > 1) {
100 server->credits--;
101 server->oplock_credits++;
102 }
103 }
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800104 scredits = *val;
105 in_flight = server->in_flight;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400106 spin_unlock(&server->req_lock);
107 wake_up(&server->request_q);
Pavel Shilovskyef68e832019-01-18 17:25:36 -0800108
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800109 if (reconnect_detected) {
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800110 trace_smb3_reconnect_detected(server->CurrentMid,
111 server->conn_id, server->hostname, scredits, add, in_flight);
112
Pavel Shilovsky335b7b62019-01-16 11:12:41 -0800113 cifs_dbg(FYI, "trying to put %d credits from the old server instance %d\n",
114 add, instance);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800115 }
Pavel Shilovsky335b7b62019-01-16 11:12:41 -0800116
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800117 if (reconnect_with_invalid_credits) {
118 trace_smb3_reconnect_with_invalid_credits(server->CurrentMid,
119 server->conn_id, server->hostname, scredits, add, in_flight);
120 cifs_dbg(FYI, "Negotiate operation when server credits is non-zero. Optype: %d, server credits: %d, credits added: %d\n",
121 optype, scredits, add);
122 }
123
Pavel Shilovsky82e04572019-01-25 10:56:41 -0800124 if (server->tcpStatus == CifsNeedReconnect
125 || server->tcpStatus == CifsExiting)
Pavel Shilovskyef68e832019-01-18 17:25:36 -0800126 return;
127
128 switch (rc) {
129 case -1:
130 /* change_conf hasn't been executed */
131 break;
132 case 0:
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000133 cifs_server_dbg(VFS, "Possible client or server bug - zero credits\n");
Pavel Shilovskyef68e832019-01-18 17:25:36 -0800134 break;
135 case 1:
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000136 cifs_server_dbg(VFS, "disabling echoes and oplocks\n");
Pavel Shilovskyef68e832019-01-18 17:25:36 -0800137 break;
138 case 2:
139 cifs_dbg(FYI, "disabling oplocks\n");
140 break;
141 default:
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800142 /* change_conf rebalanced credits for different types */
143 break;
Pavel Shilovskyef68e832019-01-18 17:25:36 -0800144 }
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800145
146 trace_smb3_add_credits(server->CurrentMid,
147 server->conn_id, server->hostname, scredits, add, in_flight);
148 cifs_dbg(FYI, "%s: added %u credits total=%d\n", __func__, add, scredits);
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400149}
150
151static void
152smb2_set_credits(struct TCP_Server_Info *server, const int val)
153{
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800154 int scredits, in_flight;
155
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400156 spin_lock(&server->req_lock);
157 server->credits = val;
Steve French9e1a37d2018-09-19 02:38:17 -0500158 if (val == 1)
159 server->reconnect_instance++;
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800160 scredits = server->credits;
161 in_flight = server->in_flight;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400162 spin_unlock(&server->req_lock);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800163
164 trace_smb3_set_credits(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800165 server->conn_id, server->hostname, scredits, val, in_flight);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800166 cifs_dbg(FYI, "%s: set %u credits\n", __func__, val);
167
Steve French6e4d3bb2018-09-22 11:25:04 -0500168 /* don't log while holding the lock */
169 if (val == 1)
170 cifs_dbg(FYI, "set credits to 1 due to smb2 reconnect\n");
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400171}
172
173static int *
174smb2_get_credits_field(struct TCP_Server_Info *server, const int optype)
175{
176 switch (optype) {
177 case CIFS_ECHO_OP:
178 return &server->echo_credits;
179 case CIFS_OBREAK_OP:
180 return &server->oplock_credits;
181 default:
182 return &server->credits;
183 }
184}
185
186static unsigned int
187smb2_get_credits(struct mid_q_entry *mid)
188{
Pavel Shilovsky86a79642019-11-21 11:35:13 -0800189 return mid->credits_received;
Pavel Shilovsky28ea5292012-05-23 16:18:00 +0400190}
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +0400191
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400192static int
193smb2_wait_mtu_credits(struct TCP_Server_Info *server, unsigned int size,
Pavel Shilovsky335b7b62019-01-16 11:12:41 -0800194 unsigned int *num, struct cifs_credits *credits)
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400195{
196 int rc = 0;
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800197 unsigned int scredits, in_flight;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400198
199 spin_lock(&server->req_lock);
200 while (1) {
201 if (server->credits <= 0) {
202 spin_unlock(&server->req_lock);
203 cifs_num_waiters_inc(server);
204 rc = wait_event_killable(server->request_q,
Ronnie Sahlbergb227d212019-03-08 12:58:20 +1000205 has_credits(server, &server->credits, 1));
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400206 cifs_num_waiters_dec(server);
207 if (rc)
208 return rc;
209 spin_lock(&server->req_lock);
210 } else {
211 if (server->tcpStatus == CifsExiting) {
212 spin_unlock(&server->req_lock);
213 return -ENOENT;
214 }
215
216 scredits = server->credits;
217 /* can deadlock with reopen */
Pavel Shilovskyacc58d02019-01-17 08:21:24 -0800218 if (scredits <= 8) {
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400219 *num = SMB2_MAX_BUFFER_SIZE;
Pavel Shilovsky335b7b62019-01-16 11:12:41 -0800220 credits->value = 0;
221 credits->instance = 0;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400222 break;
223 }
224
Pavel Shilovskyacc58d02019-01-17 08:21:24 -0800225 /* leave some credits for reopen and other ops */
226 scredits -= 8;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400227 *num = min_t(unsigned int, size,
228 scredits * SMB2_MAX_BUFFER_SIZE);
229
Pavel Shilovsky335b7b62019-01-16 11:12:41 -0800230 credits->value =
231 DIV_ROUND_UP(*num, SMB2_MAX_BUFFER_SIZE);
232 credits->instance = server->reconnect_instance;
233 server->credits -= credits->value;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400234 server->in_flight++;
Steve French1b63f182019-09-09 22:57:11 -0500235 if (server->in_flight > server->max_in_flight)
236 server->max_in_flight = server->in_flight;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400237 break;
238 }
239 }
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800240 scredits = server->credits;
241 in_flight = server->in_flight;
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400242 spin_unlock(&server->req_lock);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800243
244 trace_smb3_add_credits(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800245 server->conn_id, server->hostname, scredits, -(credits->value), in_flight);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800246 cifs_dbg(FYI, "%s: removed %u credits total=%d\n",
247 __func__, credits->value, scredits);
248
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400249 return rc;
250}
251
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800252static int
253smb2_adjust_credits(struct TCP_Server_Info *server,
254 struct cifs_credits *credits,
255 const unsigned int payload_size)
256{
257 int new_val = DIV_ROUND_UP(payload_size, SMB2_MAX_BUFFER_SIZE);
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800258 int scredits, in_flight;
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800259
260 if (!credits->value || credits->value == new_val)
261 return 0;
262
263 if (credits->value < new_val) {
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800264 trace_smb3_too_many_credits(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800265 server->conn_id, server->hostname, 0, credits->value - new_val, 0);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800266 cifs_server_dbg(VFS, "request has less credits (%d) than required (%d)",
267 credits->value, new_val);
268
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800269 return -ENOTSUPP;
270 }
271
272 spin_lock(&server->req_lock);
273
274 if (server->reconnect_instance != credits->instance) {
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800275 scredits = server->credits;
276 in_flight = server->in_flight;
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800277 spin_unlock(&server->req_lock);
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800278
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800279 trace_smb3_reconnect_detected(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800280 server->conn_id, server->hostname, scredits,
281 credits->value - new_val, in_flight);
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000282 cifs_server_dbg(VFS, "trying to return %d credits to old session\n",
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800283 credits->value - new_val);
284 return -EAGAIN;
285 }
286
287 server->credits += credits->value - new_val;
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800288 scredits = server->credits;
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800289 in_flight = server->in_flight;
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800290 spin_unlock(&server->req_lock);
291 wake_up(&server->request_q);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800292
293 trace_smb3_add_credits(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800294 server->conn_id, server->hostname, scredits,
295 credits->value - new_val, in_flight);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -0800296 cifs_dbg(FYI, "%s: adjust added %u credits total=%d\n",
297 __func__, credits->value - new_val, scredits);
298
Shyam Prasad N6d82c272021-02-03 23:20:46 -0800299 credits->value = new_val;
300
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -0800301 return 0;
302}
303
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +0400304static __u64
305smb2_get_next_mid(struct TCP_Server_Info *server)
306{
307 __u64 mid;
308 /* for SMB2 we need the current value */
309 spin_lock(&GlobalMid_Lock);
310 mid = server->CurrentMid++;
311 spin_unlock(&GlobalMid_Lock);
312 return mid;
313}
Steve French1080ef72011-02-24 18:07:19 +0000314
Pavel Shilovskyc781af72019-03-04 14:02:50 -0800315static void
316smb2_revert_current_mid(struct TCP_Server_Info *server, const unsigned int val)
317{
318 spin_lock(&GlobalMid_Lock);
319 if (server->CurrentMid >= val)
320 server->CurrentMid -= val;
321 spin_unlock(&GlobalMid_Lock);
322}
323
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400324static struct mid_q_entry *
Rohith Surabattulaac873aa2020-10-29 05:03:10 +0000325__smb2_find_mid(struct TCP_Server_Info *server, char *buf, bool dequeue)
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400326{
327 struct mid_q_entry *mid;
Ronnie Sahlberg49f466b2018-06-01 10:53:06 +1000328 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
Pavel Shilovsky31473fc2016-10-24 15:33:04 -0700329 __u64 wire_mid = le64_to_cpu(shdr->MessageId);
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400330
Pavel Shilovsky31473fc2016-10-24 15:33:04 -0700331 if (shdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000332 cifs_server_dbg(VFS, "Encrypted frame parsing not supported yet\n");
Steve French373512e2015-12-18 13:05:30 -0600333 return NULL;
334 }
335
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400336 spin_lock(&GlobalMid_Lock);
337 list_for_each_entry(mid, &server->pending_mid_q, qhead) {
Sachin Prabhu9235d092014-12-09 17:37:00 +0000338 if ((mid->mid == wire_mid) &&
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400339 (mid->mid_state == MID_REQUEST_SUBMITTED) &&
Pavel Shilovsky31473fc2016-10-24 15:33:04 -0700340 (mid->command == shdr->Command)) {
Lars Persson696e4202018-06-25 14:05:25 +0200341 kref_get(&mid->refcount);
Rohith Surabattulaac873aa2020-10-29 05:03:10 +0000342 if (dequeue) {
343 list_del_init(&mid->qhead);
344 mid->mid_flags |= MID_DELETED;
345 }
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400346 spin_unlock(&GlobalMid_Lock);
347 return mid;
348 }
349 }
350 spin_unlock(&GlobalMid_Lock);
351 return NULL;
352}
353
Rohith Surabattulaac873aa2020-10-29 05:03:10 +0000354static struct mid_q_entry *
355smb2_find_mid(struct TCP_Server_Info *server, char *buf)
356{
357 return __smb2_find_mid(server, buf, false);
358}
359
360static struct mid_q_entry *
361smb2_find_dequeue_mid(struct TCP_Server_Info *server, char *buf)
362{
363 return __smb2_find_mid(server, buf, true);
364}
365
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400366static void
Ronnie Sahlberg14547f72018-04-22 14:45:53 -0600367smb2_dump_detail(void *buf, struct TCP_Server_Info *server)
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400368{
369#ifdef CONFIG_CIFS_DEBUG2
Ronnie Sahlberg49f466b2018-06-01 10:53:06 +1000370 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400371
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000372 cifs_server_dbg(VFS, "Cmd: %d Err: 0x%x Flags: 0x%x Mid: %llu Pid: %d\n",
Pavel Shilovsky31473fc2016-10-24 15:33:04 -0700373 shdr->Command, shdr->Status, shdr->Flags, shdr->MessageId,
374 shdr->ProcessId);
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000375 cifs_server_dbg(VFS, "smb buf %p len %u\n", buf,
Steve French71992e622018-05-06 15:58:51 -0500376 server->ops->calc_smb_size(buf, server));
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +0400377#endif
378}
379
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400380static bool
381smb2_need_neg(struct TCP_Server_Info *server)
382{
383 return server->max_read == 0;
384}
385
386static int
387smb2_negotiate(const unsigned int xid, struct cifs_ses *ses)
388{
389 int rc;
Christoph Probsta205d502019-05-08 21:36:25 +0200390
Steve French0fa757b2021-06-25 13:54:32 -0500391 spin_lock(&GlobalMid_Lock);
Aurelien Aptelf6a6bf72019-09-20 06:22:14 +0200392 cifs_ses_server(ses)->CurrentMid = 0;
Steve French0fa757b2021-06-25 13:54:32 -0500393 spin_unlock(&GlobalMid_Lock);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400394 rc = SMB2_negotiate(xid, ses);
395 /* BB we probably don't need to retry with modern servers */
396 if (rc == -EAGAIN)
397 rc = -EHOSTDOWN;
398 return rc;
399}
400
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700401static unsigned int
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600402smb2_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700403{
404 struct TCP_Server_Info *server = tcon->ses->server;
405 unsigned int wsize;
406
407 /* start with specified wsize, or default */
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600408 wsize = ctx->wsize ? ctx->wsize : CIFS_DEFAULT_IOSIZE;
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700409 wsize = min_t(unsigned int, wsize, server->max_write);
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +0400410 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
411 wsize = min_t(unsigned int, wsize, SMB2_MAX_BUFFER_SIZE);
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700412
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700413 return wsize;
414}
415
416static unsigned int
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600417smb3_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
Steve French3d621232018-09-25 15:33:47 -0500418{
419 struct TCP_Server_Info *server = tcon->ses->server;
420 unsigned int wsize;
421
422 /* start with specified wsize, or default */
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600423 wsize = ctx->wsize ? ctx->wsize : SMB3_DEFAULT_IOSIZE;
Steve French3d621232018-09-25 15:33:47 -0500424 wsize = min_t(unsigned int, wsize, server->max_write);
425#ifdef CONFIG_CIFS_SMB_DIRECT
426 if (server->rdma) {
427 if (server->sign)
Long Lif7950cb2020-03-26 19:42:24 -0700428 /*
429 * Account for SMB2 data transfer packet header and
430 * possible encryption header
431 */
Steve French3d621232018-09-25 15:33:47 -0500432 wsize = min_t(unsigned int,
Long Lif7950cb2020-03-26 19:42:24 -0700433 wsize,
434 server->smbd_conn->max_fragmented_send_size -
435 SMB2_READWRITE_PDU_HEADER_SIZE -
436 sizeof(struct smb2_transform_hdr));
Steve French3d621232018-09-25 15:33:47 -0500437 else
438 wsize = min_t(unsigned int,
439 wsize, server->smbd_conn->max_readwrite_size);
440 }
441#endif
442 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
443 wsize = min_t(unsigned int, wsize, SMB2_MAX_BUFFER_SIZE);
444
445 return wsize;
446}
447
448static unsigned int
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600449smb2_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700450{
451 struct TCP_Server_Info *server = tcon->ses->server;
452 unsigned int rsize;
453
454 /* start with specified rsize, or default */
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600455 rsize = ctx->rsize ? ctx->rsize : CIFS_DEFAULT_IOSIZE;
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700456 rsize = min_t(unsigned int, rsize, server->max_read);
Pavel Shilovskybed9da02014-06-25 11:28:57 +0400457
458 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
459 rsize = min_t(unsigned int, rsize, SMB2_MAX_BUFFER_SIZE);
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700460
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -0700461 return rsize;
462}
463
Steve French3d621232018-09-25 15:33:47 -0500464static unsigned int
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600465smb3_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
Steve French3d621232018-09-25 15:33:47 -0500466{
467 struct TCP_Server_Info *server = tcon->ses->server;
468 unsigned int rsize;
469
470 /* start with specified rsize, or default */
Ronnie Sahlberg3fa1c6d2020-12-09 23:07:12 -0600471 rsize = ctx->rsize ? ctx->rsize : SMB3_DEFAULT_IOSIZE;
Steve French3d621232018-09-25 15:33:47 -0500472 rsize = min_t(unsigned int, rsize, server->max_read);
473#ifdef CONFIG_CIFS_SMB_DIRECT
474 if (server->rdma) {
475 if (server->sign)
Long Lif7950cb2020-03-26 19:42:24 -0700476 /*
477 * Account for SMB2 data transfer packet header and
478 * possible encryption header
479 */
Steve French3d621232018-09-25 15:33:47 -0500480 rsize = min_t(unsigned int,
Long Lif7950cb2020-03-26 19:42:24 -0700481 rsize,
482 server->smbd_conn->max_fragmented_recv_size -
483 SMB2_READWRITE_PDU_HEADER_SIZE -
484 sizeof(struct smb2_transform_hdr));
Steve French3d621232018-09-25 15:33:47 -0500485 else
486 rsize = min_t(unsigned int,
487 rsize, server->smbd_conn->max_readwrite_size);
488 }
489#endif
490
491 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
492 rsize = min_t(unsigned int, rsize, SMB2_MAX_BUFFER_SIZE);
493
494 return rsize;
495}
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200496
497static int
498parse_server_interfaces(struct network_interface_info_ioctl_rsp *buf,
499 size_t buf_len,
500 struct cifs_server_iface **iface_list,
501 size_t *iface_count)
502{
503 struct network_interface_info_ioctl_rsp *p;
504 struct sockaddr_in *addr4;
505 struct sockaddr_in6 *addr6;
506 struct iface_info_ipv4 *p4;
507 struct iface_info_ipv6 *p6;
508 struct cifs_server_iface *info;
509 ssize_t bytes_left;
510 size_t next = 0;
511 int nb_iface = 0;
512 int rc = 0;
513
514 *iface_list = NULL;
515 *iface_count = 0;
516
517 /*
518 * Fist pass: count and sanity check
519 */
520
521 bytes_left = buf_len;
522 p = buf;
523 while (bytes_left >= sizeof(*p)) {
524 nb_iface++;
525 next = le32_to_cpu(p->Next);
526 if (!next) {
527 bytes_left -= sizeof(*p);
528 break;
529 }
530 p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
531 bytes_left -= next;
532 }
533
534 if (!nb_iface) {
535 cifs_dbg(VFS, "%s: malformed interface info\n", __func__);
536 rc = -EINVAL;
537 goto out;
538 }
539
Steve Frenchebcd6de2020-12-08 21:13:31 -0600540 /* Azure rounds the buffer size up 8, to a 16 byte boundary */
541 if ((bytes_left > 8) || p->Next)
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200542 cifs_dbg(VFS, "%s: incomplete interface info\n", __func__);
543
544
545 /*
546 * Second pass: extract info to internal structure
547 */
548
549 *iface_list = kcalloc(nb_iface, sizeof(**iface_list), GFP_KERNEL);
550 if (!*iface_list) {
551 rc = -ENOMEM;
552 goto out;
553 }
554
555 info = *iface_list;
556 bytes_left = buf_len;
557 p = buf;
558 while (bytes_left >= sizeof(*p)) {
559 info->speed = le64_to_cpu(p->LinkSpeed);
560 info->rdma_capable = le32_to_cpu(p->Capability & RDMA_CAPABLE);
561 info->rss_capable = le32_to_cpu(p->Capability & RSS_CAPABLE);
562
563 cifs_dbg(FYI, "%s: adding iface %zu\n", __func__, *iface_count);
564 cifs_dbg(FYI, "%s: speed %zu bps\n", __func__, info->speed);
565 cifs_dbg(FYI, "%s: capabilities 0x%08x\n", __func__,
566 le32_to_cpu(p->Capability));
567
568 switch (p->Family) {
569 /*
570 * The kernel and wire socket structures have the same
571 * layout and use network byte order but make the
572 * conversion explicit in case either one changes.
573 */
574 case INTERNETWORK:
575 addr4 = (struct sockaddr_in *)&info->sockaddr;
576 p4 = (struct iface_info_ipv4 *)p->Buffer;
577 addr4->sin_family = AF_INET;
578 memcpy(&addr4->sin_addr, &p4->IPv4Address, 4);
579
580 /* [MS-SMB2] 2.2.32.5.1.1 Clients MUST ignore these */
581 addr4->sin_port = cpu_to_be16(CIFS_PORT);
582
583 cifs_dbg(FYI, "%s: ipv4 %pI4\n", __func__,
584 &addr4->sin_addr);
585 break;
586 case INTERNETWORKV6:
587 addr6 = (struct sockaddr_in6 *)&info->sockaddr;
588 p6 = (struct iface_info_ipv6 *)p->Buffer;
589 addr6->sin6_family = AF_INET6;
590 memcpy(&addr6->sin6_addr, &p6->IPv6Address, 16);
591
592 /* [MS-SMB2] 2.2.32.5.1.2 Clients MUST ignore these */
593 addr6->sin6_flowinfo = 0;
594 addr6->sin6_scope_id = 0;
595 addr6->sin6_port = cpu_to_be16(CIFS_PORT);
596
597 cifs_dbg(FYI, "%s: ipv6 %pI6\n", __func__,
598 &addr6->sin6_addr);
599 break;
600 default:
601 cifs_dbg(VFS,
602 "%s: skipping unsupported socket family\n",
603 __func__);
604 goto next_iface;
605 }
606
607 (*iface_count)++;
608 info++;
609next_iface:
610 next = le32_to_cpu(p->Next);
611 if (!next)
612 break;
613 p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
614 bytes_left -= next;
615 }
616
617 if (!*iface_count) {
618 rc = -EINVAL;
619 goto out;
620 }
621
622out:
623 if (rc) {
624 kfree(*iface_list);
625 *iface_count = 0;
626 *iface_list = NULL;
627 }
628 return rc;
629}
630
Aurelien Aptel35adffe2019-09-20 06:29:39 +0200631static int compare_iface(const void *ia, const void *ib)
632{
633 const struct cifs_server_iface *a = (struct cifs_server_iface *)ia;
634 const struct cifs_server_iface *b = (struct cifs_server_iface *)ib;
635
636 return a->speed == b->speed ? 0 : (a->speed > b->speed ? -1 : 1);
637}
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200638
Steve Frenchc481e9f2013-10-14 01:21:53 -0500639static int
640SMB3_request_interfaces(const unsigned int xid, struct cifs_tcon *tcon)
641{
642 int rc;
643 unsigned int ret_data_len = 0;
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200644 struct network_interface_info_ioctl_rsp *out_buf = NULL;
645 struct cifs_server_iface *iface_list;
646 size_t iface_count;
647 struct cifs_ses *ses = tcon->ses;
Steve Frenchc481e9f2013-10-14 01:21:53 -0500648
649 rc = SMB2_ioctl(xid, tcon, NO_FILE_ID, NO_FILE_ID,
650 FSCTL_QUERY_NETWORK_INTERFACE_INFO, true /* is_fsctl */,
651 NULL /* no data input */, 0 /* no data input */,
Steve French153322f2019-03-28 22:32:49 -0500652 CIFSMaxBufSize, (char **)&out_buf, &ret_data_len);
Steve Frenchc3ed4402018-06-28 22:53:39 -0500653 if (rc == -EOPNOTSUPP) {
654 cifs_dbg(FYI,
655 "server does not support query network interfaces\n");
656 goto out;
657 } else if (rc != 0) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +1000658 cifs_tcon_dbg(VFS, "error %d on ioctl to get interface list\n", rc);
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200659 goto out;
Steve French9ffc5412014-10-16 15:13:14 -0500660 }
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200661
662 rc = parse_server_interfaces(out_buf, ret_data_len,
663 &iface_list, &iface_count);
664 if (rc)
665 goto out;
666
Aurelien Aptel35adffe2019-09-20 06:29:39 +0200667 /* sort interfaces from fastest to slowest */
668 sort(iface_list, iface_count, sizeof(*iface_list), compare_iface, NULL);
669
Aurelien Aptelfe856be2018-06-14 17:04:51 +0200670 spin_lock(&ses->iface_lock);
671 kfree(ses->iface_list);
672 ses->iface_list = iface_list;
673 ses->iface_count = iface_count;
674 ses->iface_last_update = jiffies;
675 spin_unlock(&ses->iface_lock);
676
677out:
Steve French24df1482016-09-29 04:20:23 -0500678 kfree(out_buf);
Steve Frenchc481e9f2013-10-14 01:21:53 -0500679 return rc;
680}
Steve Frenchc481e9f2013-10-14 01:21:53 -0500681
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000682static void
683smb2_close_cached_fid(struct kref *ref)
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000684{
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000685 struct cached_fid *cfid = container_of(ref, struct cached_fid,
686 refcount);
687
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000688 if (cfid->is_valid) {
689 cifs_dbg(FYI, "clear cached root file handle\n");
690 SMB2_close(0, cfid->tcon, cfid->fid->persistent_fid,
691 cfid->fid->volatile_fid);
692 cfid->is_valid = false;
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000693 cfid->file_all_info_is_valid = false;
Pavel Shilovskyd9191312019-12-10 11:44:52 -0800694 cfid->has_lease = false;
Ronnie Sahlberg5e9c89d2021-03-09 09:07:31 +1000695 if (cfid->dentry) {
696 dput(cfid->dentry);
697 cfid->dentry = NULL;
698 }
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000699 }
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000700}
701
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000702void close_cached_dir(struct cached_fid *cfid)
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000703{
704 mutex_lock(&cfid->fid_mutex);
705 kref_put(&cfid->refcount, smb2_close_cached_fid);
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000706 mutex_unlock(&cfid->fid_mutex);
707}
708
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000709void close_cached_dir_lease_locked(struct cached_fid *cfid)
Pavel Shilovskyd9191312019-12-10 11:44:52 -0800710{
711 if (cfid->has_lease) {
712 cfid->has_lease = false;
713 kref_put(&cfid->refcount, smb2_close_cached_fid);
714 }
715}
716
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000717void close_cached_dir_lease(struct cached_fid *cfid)
Pavel Shilovskyd9191312019-12-10 11:44:52 -0800718{
719 mutex_lock(&cfid->fid_mutex);
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000720 close_cached_dir_lease_locked(cfid);
Pavel Shilovskyd9191312019-12-10 11:44:52 -0800721 mutex_unlock(&cfid->fid_mutex);
722}
723
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000724void
725smb2_cached_lease_break(struct work_struct *work)
726{
727 struct cached_fid *cfid = container_of(work,
728 struct cached_fid, lease_break);
729
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000730 close_cached_dir_lease(cfid);
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000731}
732
Steve French3d4ef9a2018-04-25 22:19:09 -0500733/*
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000734 * Open the and cache a directory handle.
735 * Only supported for the root handle.
Steve French3d4ef9a2018-04-25 22:19:09 -0500736 */
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000737int open_cached_dir(unsigned int xid, struct cifs_tcon *tcon,
Ronnie Sahlberge6eb1952021-03-09 09:07:28 +1000738 const char *path,
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000739 struct cifs_sb_info *cifs_sb,
740 struct cached_fid **cfid)
Steve French3d4ef9a2018-04-25 22:19:09 -0500741{
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000742 struct cifs_ses *ses = tcon->ses;
743 struct TCP_Server_Info *server = ses->server;
744 struct cifs_open_parms oparms;
745 struct smb2_create_rsp *o_rsp = NULL;
746 struct smb2_query_info_rsp *qi_rsp = NULL;
747 int resp_buftype[2];
748 struct smb_rqst rqst[2];
749 struct kvec rsp_iov[2];
750 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
751 struct kvec qi_iov[1];
752 int rc, flags = 0;
753 __le16 utf16_path = 0; /* Null - since an open of top of share */
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000754 u8 oplock = SMB2_OPLOCK_LEVEL_II;
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000755 struct cifs_fid *pfid;
Ronnie Sahlberg5e9c89d2021-03-09 09:07:31 +1000756 struct dentry *dentry;
Steve French3d4ef9a2018-04-25 22:19:09 -0500757
Ronnie Sahlberg4df3d972021-03-09 09:07:27 +1000758 if (tcon->nohandlecache)
759 return -ENOTSUPP;
760
Ronnie Sahlberg269f67e2021-03-09 09:07:30 +1000761 if (cifs_sb->root == NULL)
762 return -ENOENT;
763
Ronnie Sahlberge6eb1952021-03-09 09:07:28 +1000764 if (strlen(path))
Ronnie Sahlberg269f67e2021-03-09 09:07:30 +1000765 return -ENOENT;
Ronnie Sahlberge6eb1952021-03-09 09:07:28 +1000766
Ronnie Sahlberg5e9c89d2021-03-09 09:07:31 +1000767 dentry = cifs_sb->root;
768
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000769 mutex_lock(&tcon->crfid.fid_mutex);
770 if (tcon->crfid.is_valid) {
Steve French3d4ef9a2018-04-25 22:19:09 -0500771 cifs_dbg(FYI, "found a cached root file handle\n");
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000772 *cfid = &tcon->crfid;
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000773 kref_get(&tcon->crfid.refcount);
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000774 mutex_unlock(&tcon->crfid.fid_mutex);
Steve French3d4ef9a2018-04-25 22:19:09 -0500775 return 0;
776 }
777
Steve French96d9f7e2019-09-12 17:52:54 -0500778 /*
779 * We do not hold the lock for the open because in case
780 * SMB2_open needs to reconnect, it will end up calling
781 * cifs_mark_open_files_invalid() which takes the lock again
782 * thus causing a deadlock
783 */
784
785 mutex_unlock(&tcon->crfid.fid_mutex);
786
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000787 if (smb3_encryption_required(tcon))
788 flags |= CIFS_TRANSFORM_REQ;
Steve French3d4ef9a2018-04-25 22:19:09 -0500789
Paulo Alcantara0fe07812020-04-20 23:44:24 -0300790 if (!server->ops->new_lease_key)
791 return -EIO;
792
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000793 pfid = tcon->crfid.fid;
Paulo Alcantara0fe07812020-04-20 23:44:24 -0300794 server->ops->new_lease_key(pfid);
795
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000796 memset(rqst, 0, sizeof(rqst));
797 resp_buftype[0] = resp_buftype[1] = CIFS_NO_BUFFER;
798 memset(rsp_iov, 0, sizeof(rsp_iov));
799
800 /* Open */
801 memset(&open_iov, 0, sizeof(open_iov));
802 rqst[0].rq_iov = open_iov;
803 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
804
805 oparms.tcon = tcon;
Amir Goldstein0f060932020-02-03 21:46:43 +0200806 oparms.create_options = cifs_create_options(cifs_sb, 0);
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000807 oparms.desired_access = FILE_READ_ATTRIBUTES;
808 oparms.disposition = FILE_OPEN;
809 oparms.fid = pfid;
810 oparms.reconnect = false;
811
Aurelien Aptel352d96f2020-05-31 12:38:22 -0500812 rc = SMB2_open_init(tcon, server,
813 &rqst[0], &oplock, &oparms, &utf16_path);
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000814 if (rc)
Steve French96d9f7e2019-09-12 17:52:54 -0500815 goto oshr_free;
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000816 smb2_set_next_command(tcon, &rqst[0]);
817
818 memset(&qi_iov, 0, sizeof(qi_iov));
819 rqst[1].rq_iov = qi_iov;
820 rqst[1].rq_nvec = 1;
821
Aurelien Aptel352d96f2020-05-31 12:38:22 -0500822 rc = SMB2_query_info_init(tcon, server,
823 &rqst[1], COMPOUND_FID,
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000824 COMPOUND_FID, FILE_ALL_INFORMATION,
825 SMB2_O_INFO_FILE, 0,
826 sizeof(struct smb2_file_all_info) +
827 PATH_MAX * 2, 0, NULL);
828 if (rc)
Steve French96d9f7e2019-09-12 17:52:54 -0500829 goto oshr_free;
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000830
831 smb2_set_related(&rqst[1]);
832
Aurelien Aptel352d96f2020-05-31 12:38:22 -0500833 rc = compound_send_recv(xid, ses, server,
834 flags, 2, rqst,
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000835 resp_buftype, rsp_iov);
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200836 mutex_lock(&tcon->crfid.fid_mutex);
837
838 /*
839 * Now we need to check again as the cached root might have
840 * been successfully re-opened from a concurrent process
841 */
842
843 if (tcon->crfid.is_valid) {
844 /* work was already done */
845
846 /* stash fids for close() later */
847 struct cifs_fid fid = {
848 .persistent_fid = pfid->persistent_fid,
849 .volatile_fid = pfid->volatile_fid,
850 };
851
852 /*
Muhammad Usama Anjumad7567b2021-04-15 20:24:09 +0500853 * caller expects this func to set the fid in crfid to valid
854 * cached root, so increment the refcount.
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200855 */
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200856 kref_get(&tcon->crfid.refcount);
857
858 mutex_unlock(&tcon->crfid.fid_mutex);
859
860 if (rc == 0) {
861 /* close extra handle outside of crit sec */
862 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
863 }
Xiyu Yang77577de2020-06-13 20:27:09 +0800864 rc = 0;
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200865 goto oshr_free;
866 }
867
868 /* Cached root is still invalid, continue normaly */
869
Steve French7dcc82c2019-09-11 00:07:36 -0500870 if (rc) {
871 if (rc == -EREMCHG) {
872 tcon->need_reconnect = true;
Joe Perchesa0a30362020-04-14 22:42:53 -0700873 pr_warn_once("server share %s deleted\n",
874 tcon->treeName);
Steve French7dcc82c2019-09-11 00:07:36 -0500875 }
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000876 goto oshr_exit;
Steve French7dcc82c2019-09-11 00:07:36 -0500877 }
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000878
Steve Frenchd2f15422019-09-22 00:55:46 -0500879 atomic_inc(&tcon->num_remote_opens);
880
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000881 o_rsp = (struct smb2_create_rsp *)rsp_iov[0].iov_base;
882 oparms.fid->persistent_fid = o_rsp->PersistentFileId;
883 oparms.fid->volatile_fid = o_rsp->VolatileFileId;
884#ifdef CONFIG_CIFS_DEBUG2
885 oparms.fid->mid = le64_to_cpu(o_rsp->sync_hdr.MessageId);
886#endif /* CIFS_DEBUG2 */
887
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000888 tcon->crfid.tcon = tcon;
889 tcon->crfid.is_valid = true;
Ronnie Sahlberg5e9c89d2021-03-09 09:07:31 +1000890 tcon->crfid.dentry = dentry;
891 dget(dentry);
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000892 kref_init(&tcon->crfid.refcount);
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000893
Steve French89a5bfa2019-07-18 17:22:18 -0500894 /* BB TBD check to see if oplock level check can be removed below */
Ronnie Sahlberg2f94a3122019-03-28 11:20:02 +1000895 if (o_rsp->OplockLevel == SMB2_OPLOCK_LEVEL_LEASE) {
Muhammad Usama Anjumad7567b2021-04-15 20:24:09 +0500896 /*
897 * See commit 2f94a3125b87. Increment the refcount when we
898 * get a lease for root, release it if lease break occurs
899 */
Ronnie Sahlberg2f94a3122019-03-28 11:20:02 +1000900 kref_get(&tcon->crfid.refcount);
Pavel Shilovskyd9191312019-12-10 11:44:52 -0800901 tcon->crfid.has_lease = true;
Steve French89a5bfa2019-07-18 17:22:18 -0500902 smb2_parse_contexts(server, o_rsp,
903 &oparms.fid->epoch,
Aurelien Aptel69dda302020-03-02 17:53:22 +0100904 oparms.fid->lease_key, &oplock,
905 NULL, NULL);
Ronnie Sahlberg2f94a3122019-03-28 11:20:02 +1000906 } else
907 goto oshr_exit;
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000908
909 qi_rsp = (struct smb2_query_info_rsp *)rsp_iov[1].iov_base;
910 if (le32_to_cpu(qi_rsp->OutputBufferLength) < sizeof(struct smb2_file_all_info))
911 goto oshr_exit;
Ronnie Sahlberg4811e302019-04-01 09:53:44 +1000912 if (!smb2_validate_and_copy_iov(
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000913 le16_to_cpu(qi_rsp->OutputBufferOffset),
914 sizeof(struct smb2_file_all_info),
915 &rsp_iov[1], sizeof(struct smb2_file_all_info),
Ronnie Sahlberg4811e302019-04-01 09:53:44 +1000916 (char *)&tcon->crfid.file_all_info))
zhengbin720aec02019-12-25 11:30:20 +0800917 tcon->crfid.file_all_info_is_valid = true;
Ronnie Sahlberged20f542021-03-09 09:07:33 +1000918 tcon->crfid.time = jiffies;
919
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000920
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200921oshr_exit:
Ronnie Sahlberga93864d2018-06-14 06:48:35 +1000922 mutex_unlock(&tcon->crfid.fid_mutex);
Aurelien Aptel7e5a70a2019-07-17 12:46:28 +0200923oshr_free:
Ronnie Sahlbergb0f6df72019-03-12 13:58:31 +1000924 SMB2_open_free(&rqst[0]);
925 SMB2_query_info_free(&rqst[1]);
926 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
927 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000928 if (rc == 0)
929 *cfid = &tcon->crfid;
Steve French3d4ef9a2018-04-25 22:19:09 -0500930 return rc;
931}
932
Ronnie Sahlberg6ef4e9c2021-03-09 09:07:32 +1000933int open_cached_dir_by_dentry(struct cifs_tcon *tcon,
934 struct dentry *dentry,
935 struct cached_fid **cfid)
936{
937 mutex_lock(&tcon->crfid.fid_mutex);
938 if (tcon->crfid.dentry == dentry) {
939 cifs_dbg(FYI, "found a cached root file handle by dentry\n");
940 *cfid = &tcon->crfid;
941 kref_get(&tcon->crfid.refcount);
942 mutex_unlock(&tcon->crfid.fid_mutex);
943 return 0;
944 }
945 mutex_unlock(&tcon->crfid.fid_mutex);
946 return -ENOENT;
947}
948
Steve French34f62642013-10-09 02:07:00 -0500949static void
Amir Goldstein0f060932020-02-03 21:46:43 +0200950smb3_qfs_tcon(const unsigned int xid, struct cifs_tcon *tcon,
951 struct cifs_sb_info *cifs_sb)
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500952{
953 int rc;
954 __le16 srch_path = 0; /* Null - open root of share */
955 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
956 struct cifs_open_parms oparms;
957 struct cifs_fid fid;
Ronnie Sahlberg9e81e8f2020-10-05 12:37:52 +1000958 struct cached_fid *cfid = NULL;
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500959
960 oparms.tcon = tcon;
961 oparms.desired_access = FILE_READ_ATTRIBUTES;
962 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +0200963 oparms.create_options = cifs_create_options(cifs_sb, 0);
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500964 oparms.fid = &fid;
965 oparms.reconnect = false;
966
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000967 rc = open_cached_dir(xid, tcon, "", cifs_sb, &cfid);
Ronnie Sahlberg4df3d972021-03-09 09:07:27 +1000968 if (rc == 0)
969 memcpy(&fid, cfid->fid, sizeof(struct cifs_fid));
970 else
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +1000971 rc = SMB2_open(xid, &oparms, &srch_path, &oplock, NULL, NULL,
Aurelien Aptel69dda302020-03-02 17:53:22 +0100972 NULL, NULL);
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500973 if (rc)
974 return;
975
Steve Frenchc481e9f2013-10-14 01:21:53 -0500976 SMB3_request_interfaces(xid, tcon);
Steve Frenchc481e9f2013-10-14 01:21:53 -0500977
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500978 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
979 FS_ATTRIBUTE_INFORMATION);
980 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
981 FS_DEVICE_INFORMATION);
982 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
Steve French21ba3842018-06-24 23:18:52 -0500983 FS_VOLUME_INFORMATION);
984 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500985 FS_SECTOR_SIZE_INFORMATION); /* SMB3 specific */
Ronnie Sahlberg4df3d972021-03-09 09:07:27 +1000986 if (cfid == NULL)
Steve French3d4ef9a2018-04-25 22:19:09 -0500987 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
Ronnie Sahlberg9da6ec72018-07-31 08:48:22 +1000988 else
Ronnie Sahlberg45c0f1a2021-03-09 09:07:29 +1000989 close_cached_dir(cfid);
Steven Frenchaf6a12e2013-10-09 20:55:53 -0500990}
991
992static void
Amir Goldstein0f060932020-02-03 21:46:43 +0200993smb2_qfs_tcon(const unsigned int xid, struct cifs_tcon *tcon,
994 struct cifs_sb_info *cifs_sb)
Steve French34f62642013-10-09 02:07:00 -0500995{
996 int rc;
997 __le16 srch_path = 0; /* Null - open root of share */
998 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
999 struct cifs_open_parms oparms;
1000 struct cifs_fid fid;
1001
1002 oparms.tcon = tcon;
1003 oparms.desired_access = FILE_READ_ATTRIBUTES;
1004 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02001005 oparms.create_options = cifs_create_options(cifs_sb, 0);
Steve French34f62642013-10-09 02:07:00 -05001006 oparms.fid = &fid;
1007 oparms.reconnect = false;
1008
Aurelien Aptel69dda302020-03-02 17:53:22 +01001009 rc = SMB2_open(xid, &oparms, &srch_path, &oplock, NULL, NULL,
1010 NULL, NULL);
Steve French34f62642013-10-09 02:07:00 -05001011 if (rc)
1012 return;
1013
Steven French21671142013-10-09 13:36:35 -05001014 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
1015 FS_ATTRIBUTE_INFORMATION);
1016 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
1017 FS_DEVICE_INFORMATION);
Steve French34f62642013-10-09 02:07:00 -05001018 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
Steve French34f62642013-10-09 02:07:00 -05001019}
1020
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001021static int
1022smb2_is_path_accessible(const unsigned int xid, struct cifs_tcon *tcon,
1023 struct cifs_sb_info *cifs_sb, const char *full_path)
1024{
1025 int rc;
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001026 __le16 *utf16_path;
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07001027 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04001028 struct cifs_open_parms oparms;
1029 struct cifs_fid fid;
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001030
Ronnie Sahlberga93864d2018-06-14 06:48:35 +10001031 if ((*full_path == 0) && tcon->crfid.is_valid)
Steve French3d4ef9a2018-04-25 22:19:09 -05001032 return 0;
1033
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001034 utf16_path = cifs_convert_path_to_utf16(full_path, cifs_sb);
1035 if (!utf16_path)
1036 return -ENOMEM;
1037
Pavel Shilovsky064f6042013-07-09 18:20:30 +04001038 oparms.tcon = tcon;
1039 oparms.desired_access = FILE_READ_ATTRIBUTES;
1040 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02001041 oparms.create_options = cifs_create_options(cifs_sb, 0);
Pavel Shilovsky064f6042013-07-09 18:20:30 +04001042 oparms.fid = &fid;
Pavel Shilovsky9cbc0b72013-07-09 18:40:58 +04001043 oparms.reconnect = false;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04001044
Aurelien Aptel69dda302020-03-02 17:53:22 +01001045 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL, NULL,
1046 NULL);
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001047 if (rc) {
1048 kfree(utf16_path);
1049 return rc;
1050 }
1051
Pavel Shilovsky064f6042013-07-09 18:20:30 +04001052 rc = SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001053 kfree(utf16_path);
1054 return rc;
1055}
1056
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001057static int
1058smb2_get_srv_inum(const unsigned int xid, struct cifs_tcon *tcon,
1059 struct cifs_sb_info *cifs_sb, const char *full_path,
1060 u64 *uniqueid, FILE_ALL_INFO *data)
1061{
1062 *uniqueid = le64_to_cpu(data->IndexNumber);
1063 return 0;
1064}
1065
Pavel Shilovskyb7546bc2012-09-18 16:20:27 -07001066static int
1067smb2_query_file_info(const unsigned int xid, struct cifs_tcon *tcon,
1068 struct cifs_fid *fid, FILE_ALL_INFO *data)
1069{
1070 int rc;
1071 struct smb2_file_all_info *smb2_data;
1072
Pavel Shilovsky1bbe4992014-08-22 13:32:11 +04001073 smb2_data = kzalloc(sizeof(struct smb2_file_all_info) + PATH_MAX * 2,
Pavel Shilovskyb7546bc2012-09-18 16:20:27 -07001074 GFP_KERNEL);
1075 if (smb2_data == NULL)
1076 return -ENOMEM;
1077
1078 rc = SMB2_query_info(xid, tcon, fid->persistent_fid, fid->volatile_fid,
1079 smb2_data);
1080 if (!rc)
1081 move_smb2_info_to_cifs(data, smb2_data);
1082 kfree(smb2_data);
1083 return rc;
1084}
1085
Arnd Bergmann1368f152017-09-05 11:24:15 +02001086#ifdef CONFIG_CIFS_XATTR
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001087static ssize_t
1088move_smb2_ea_to_cifs(char *dst, size_t dst_size,
1089 struct smb2_file_full_ea_info *src, size_t src_size,
1090 const unsigned char *ea_name)
1091{
1092 int rc = 0;
1093 unsigned int ea_name_len = ea_name ? strlen(ea_name) : 0;
1094 char *name, *value;
Ronnie Sahlberg0c5d6cb2018-10-25 15:43:36 +10001095 size_t buf_size = dst_size;
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001096 size_t name_len, value_len, user_name_len;
1097
1098 while (src_size > 0) {
1099 name = &src->ea_data[0];
1100 name_len = (size_t)src->ea_name_length;
1101 value = &src->ea_data[src->ea_name_length + 1];
1102 value_len = (size_t)le16_to_cpu(src->ea_value_length);
1103
Christoph Probsta205d502019-05-08 21:36:25 +02001104 if (name_len == 0)
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001105 break;
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001106
1107 if (src_size < 8 + name_len + 1 + value_len) {
1108 cifs_dbg(FYI, "EA entry goes beyond length of list\n");
1109 rc = -EIO;
1110 goto out;
1111 }
1112
1113 if (ea_name) {
1114 if (ea_name_len == name_len &&
1115 memcmp(ea_name, name, name_len) == 0) {
1116 rc = value_len;
1117 if (dst_size == 0)
1118 goto out;
1119 if (dst_size < value_len) {
1120 rc = -ERANGE;
1121 goto out;
1122 }
1123 memcpy(dst, value, value_len);
1124 goto out;
1125 }
1126 } else {
1127 /* 'user.' plus a terminating null */
1128 user_name_len = 5 + 1 + name_len;
1129
Ronnie Sahlberg0c5d6cb2018-10-25 15:43:36 +10001130 if (buf_size == 0) {
1131 /* skip copy - calc size only */
1132 rc += user_name_len;
1133 } else if (dst_size >= user_name_len) {
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001134 dst_size -= user_name_len;
1135 memcpy(dst, "user.", 5);
1136 dst += 5;
1137 memcpy(dst, src->ea_data, name_len);
1138 dst += name_len;
1139 *dst = 0;
1140 ++dst;
Ronnie Sahlberg0c5d6cb2018-10-25 15:43:36 +10001141 rc += user_name_len;
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001142 } else {
1143 /* stop before overrun buffer */
1144 rc = -ERANGE;
1145 break;
1146 }
1147 }
1148
1149 if (!src->next_entry_offset)
1150 break;
1151
1152 if (src_size < le32_to_cpu(src->next_entry_offset)) {
1153 /* stop before overrun buffer */
1154 rc = -ERANGE;
1155 break;
1156 }
1157 src_size -= le32_to_cpu(src->next_entry_offset);
1158 src = (void *)((char *)src +
1159 le32_to_cpu(src->next_entry_offset));
1160 }
1161
1162 /* didn't find the named attribute */
1163 if (ea_name)
1164 rc = -ENODATA;
1165
1166out:
1167 return (ssize_t)rc;
1168}
1169
1170static ssize_t
1171smb2_query_eas(const unsigned int xid, struct cifs_tcon *tcon,
1172 const unsigned char *path, const unsigned char *ea_name,
1173 char *ea_data, size_t buf_size,
1174 struct cifs_sb_info *cifs_sb)
1175{
1176 int rc;
1177 __le16 *utf16_path;
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001178 struct kvec rsp_iov = {NULL, 0};
1179 int buftype = CIFS_NO_BUFFER;
1180 struct smb2_query_info_rsp *rsp;
1181 struct smb2_file_full_ea_info *info = NULL;
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001182
1183 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
1184 if (!utf16_path)
1185 return -ENOMEM;
1186
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001187 rc = smb2_query_info_compound(xid, tcon, utf16_path,
1188 FILE_READ_EA,
1189 FILE_FULL_EA_INFORMATION,
1190 SMB2_O_INFO_FILE,
Ronnie Sahlbergc4627e62019-01-29 12:46:17 +10001191 CIFSMaxBufSize -
1192 MAX_SMB2_CREATE_RESPONSE_SIZE -
1193 MAX_SMB2_CLOSE_RESPONSE_SIZE,
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001194 &rsp_iov, &buftype, cifs_sb);
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001195 if (rc) {
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001196 /*
1197 * If ea_name is NULL (listxattr) and there are no EAs,
1198 * return 0 as it's not an error. Otherwise, the specified
1199 * ea_name was not found.
1200 */
1201 if (!ea_name && rc == -ENODATA)
1202 rc = 0;
1203 goto qeas_exit;
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001204 }
1205
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001206 rsp = (struct smb2_query_info_rsp *)rsp_iov.iov_base;
1207 rc = smb2_validate_iov(le16_to_cpu(rsp->OutputBufferOffset),
1208 le32_to_cpu(rsp->OutputBufferLength),
1209 &rsp_iov,
1210 sizeof(struct smb2_file_full_ea_info));
1211 if (rc)
1212 goto qeas_exit;
Ronnie Sahlberg7cb3def2017-09-28 09:39:58 +10001213
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001214 info = (struct smb2_file_full_ea_info *)(
1215 le16_to_cpu(rsp->OutputBufferOffset) + (char *)rsp);
1216 rc = move_smb2_ea_to_cifs(ea_data, buf_size, info,
1217 le32_to_cpu(rsp->OutputBufferLength), ea_name);
Ronnie Sahlberg7cb3def2017-09-28 09:39:58 +10001218
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10001219 qeas_exit:
1220 kfree(utf16_path);
1221 free_rsp_buf(buftype, rsp_iov.iov_base);
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10001222 return rc;
1223}
1224
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001225
1226static int
1227smb2_set_ea(const unsigned int xid, struct cifs_tcon *tcon,
1228 const char *path, const char *ea_name, const void *ea_value,
1229 const __u16 ea_value_len, const struct nls_table *nls_codepage,
1230 struct cifs_sb_info *cifs_sb)
1231{
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001232 struct cifs_ses *ses = tcon->ses;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001233 struct TCP_Server_Info *server = cifs_pick_channel(ses);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001234 __le16 *utf16_path = NULL;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001235 int ea_name_len = strlen(ea_name);
Paulo Alcantara04ad69c2021-03-08 12:00:50 -03001236 int flags = CIFS_CP_CREATE_CLOSE_OP;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001237 int len;
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001238 struct smb_rqst rqst[3];
1239 int resp_buftype[3];
1240 struct kvec rsp_iov[3];
1241 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
1242 struct cifs_open_parms oparms;
1243 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
1244 struct cifs_fid fid;
1245 struct kvec si_iov[SMB2_SET_INFO_IOV_SIZE];
1246 unsigned int size[1];
1247 void *data[1];
1248 struct smb2_file_full_ea_info *ea = NULL;
1249 struct kvec close_iov[1];
Ronnie Sahlberg85db6b72020-02-13 12:14:47 +10001250 struct smb2_query_info_rsp *rsp;
1251 int rc, used_len = 0;
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001252
1253 if (smb3_encryption_required(tcon))
1254 flags |= CIFS_TRANSFORM_REQ;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001255
1256 if (ea_name_len > 255)
1257 return -EINVAL;
1258
1259 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
1260 if (!utf16_path)
1261 return -ENOMEM;
1262
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001263 memset(rqst, 0, sizeof(rqst));
1264 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
1265 memset(rsp_iov, 0, sizeof(rsp_iov));
1266
Ronnie Sahlberg21094642019-02-07 15:48:44 +10001267 if (ses->server->ops->query_all_EAs) {
1268 if (!ea_value) {
1269 rc = ses->server->ops->query_all_EAs(xid, tcon, path,
1270 ea_name, NULL, 0,
1271 cifs_sb);
1272 if (rc == -ENODATA)
1273 goto sea_exit;
Ronnie Sahlberg85db6b72020-02-13 12:14:47 +10001274 } else {
1275 /* If we are adding a attribute we should first check
1276 * if there will be enough space available to store
1277 * the new EA. If not we should not add it since we
1278 * would not be able to even read the EAs back.
1279 */
1280 rc = smb2_query_info_compound(xid, tcon, utf16_path,
1281 FILE_READ_EA,
1282 FILE_FULL_EA_INFORMATION,
1283 SMB2_O_INFO_FILE,
1284 CIFSMaxBufSize -
1285 MAX_SMB2_CREATE_RESPONSE_SIZE -
1286 MAX_SMB2_CLOSE_RESPONSE_SIZE,
1287 &rsp_iov[1], &resp_buftype[1], cifs_sb);
1288 if (rc == 0) {
1289 rsp = (struct smb2_query_info_rsp *)rsp_iov[1].iov_base;
1290 used_len = le32_to_cpu(rsp->OutputBufferLength);
1291 }
1292 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1293 resp_buftype[1] = CIFS_NO_BUFFER;
1294 memset(&rsp_iov[1], 0, sizeof(rsp_iov[1]));
1295 rc = 0;
1296
1297 /* Use a fudge factor of 256 bytes in case we collide
1298 * with a different set_EAs command.
1299 */
1300 if(CIFSMaxBufSize - MAX_SMB2_CREATE_RESPONSE_SIZE -
1301 MAX_SMB2_CLOSE_RESPONSE_SIZE - 256 <
1302 used_len + ea_name_len + ea_value_len + 1) {
1303 rc = -ENOSPC;
1304 goto sea_exit;
1305 }
Ronnie Sahlberg21094642019-02-07 15:48:44 +10001306 }
1307 }
1308
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001309 /* Open */
1310 memset(&open_iov, 0, sizeof(open_iov));
1311 rqst[0].rq_iov = open_iov;
1312 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
1313
1314 memset(&oparms, 0, sizeof(oparms));
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001315 oparms.tcon = tcon;
1316 oparms.desired_access = FILE_WRITE_EA;
1317 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02001318 oparms.create_options = cifs_create_options(cifs_sb, 0);
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001319 oparms.fid = &fid;
1320 oparms.reconnect = false;
1321
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001322 rc = SMB2_open_init(tcon, server,
1323 &rqst[0], &oplock, &oparms, utf16_path);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001324 if (rc)
1325 goto sea_exit;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10001326 smb2_set_next_command(tcon, &rqst[0]);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001327
1328
1329 /* Set Info */
1330 memset(&si_iov, 0, sizeof(si_iov));
1331 rqst[1].rq_iov = si_iov;
1332 rqst[1].rq_nvec = 1;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001333
Vladimir Zapolskiy64b7f672020-10-10 21:25:54 +03001334 len = sizeof(*ea) + ea_name_len + ea_value_len + 1;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001335 ea = kzalloc(len, GFP_KERNEL);
1336 if (ea == NULL) {
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001337 rc = -ENOMEM;
1338 goto sea_exit;
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001339 }
1340
1341 ea->ea_name_length = ea_name_len;
1342 ea->ea_value_length = cpu_to_le16(ea_value_len);
1343 memcpy(ea->ea_data, ea_name, ea_name_len + 1);
1344 memcpy(ea->ea_data + ea_name_len + 1, ea_value, ea_value_len);
1345
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001346 size[0] = len;
1347 data[0] = ea;
1348
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001349 rc = SMB2_set_info_init(tcon, server,
1350 &rqst[1], COMPOUND_FID,
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001351 COMPOUND_FID, current->tgid,
1352 FILE_FULL_EA_INFORMATION,
1353 SMB2_O_INFO_FILE, 0, data, size);
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10001354 smb2_set_next_command(tcon, &rqst[1]);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001355 smb2_set_related(&rqst[1]);
1356
1357
1358 /* Close */
1359 memset(&close_iov, 0, sizeof(close_iov));
1360 rqst[2].rq_iov = close_iov;
1361 rqst[2].rq_nvec = 1;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001362 rc = SMB2_close_init(tcon, server,
1363 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001364 smb2_set_related(&rqst[2]);
1365
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001366 rc = compound_send_recv(xid, ses, server,
1367 flags, 3, rqst,
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001368 resp_buftype, rsp_iov);
Steve Frenchd2f15422019-09-22 00:55:46 -05001369 /* no need to bump num_remote_opens because handle immediately closed */
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001370
1371 sea_exit:
Paulo Alcantara6aa0c112018-07-04 14:16:16 -03001372 kfree(ea);
Ronnie Sahlberg0967e542018-11-06 22:52:43 +10001373 kfree(utf16_path);
1374 SMB2_open_free(&rqst[0]);
1375 SMB2_set_info_free(&rqst[1]);
1376 SMB2_close_free(&rqst[2]);
1377 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
1378 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1379 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001380 return rc;
1381}
Arnd Bergmann1368f152017-09-05 11:24:15 +02001382#endif
Ronnie Sahlberg55175542017-08-24 11:24:56 +10001383
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04001384static bool
1385smb2_can_echo(struct TCP_Server_Info *server)
1386{
1387 return server->echoes;
1388}
1389
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001390static void
1391smb2_clear_stats(struct cifs_tcon *tcon)
1392{
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001393 int i;
Christoph Probsta205d502019-05-08 21:36:25 +02001394
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001395 for (i = 0; i < NUMBER_OF_SMB2_COMMANDS; i++) {
1396 atomic_set(&tcon->stats.smb2_stats.smb2_com_sent[i], 0);
1397 atomic_set(&tcon->stats.smb2_stats.smb2_com_failed[i], 0);
1398 }
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001399}
1400
1401static void
Steve French769ee6a2013-06-19 14:15:30 -05001402smb2_dump_share_caps(struct seq_file *m, struct cifs_tcon *tcon)
1403{
1404 seq_puts(m, "\n\tShare Capabilities:");
1405 if (tcon->capabilities & SMB2_SHARE_CAP_DFS)
1406 seq_puts(m, " DFS,");
1407 if (tcon->capabilities & SMB2_SHARE_CAP_CONTINUOUS_AVAILABILITY)
1408 seq_puts(m, " CONTINUOUS AVAILABILITY,");
1409 if (tcon->capabilities & SMB2_SHARE_CAP_SCALEOUT)
1410 seq_puts(m, " SCALEOUT,");
1411 if (tcon->capabilities & SMB2_SHARE_CAP_CLUSTER)
1412 seq_puts(m, " CLUSTER,");
1413 if (tcon->capabilities & SMB2_SHARE_CAP_ASYMMETRIC)
1414 seq_puts(m, " ASYMMETRIC,");
1415 if (tcon->capabilities == 0)
1416 seq_puts(m, " None");
Steven Frenchaf6a12e2013-10-09 20:55:53 -05001417 if (tcon->ss_flags & SSINFO_FLAGS_ALIGNED_DEVICE)
1418 seq_puts(m, " Aligned,");
1419 if (tcon->ss_flags & SSINFO_FLAGS_PARTITION_ALIGNED_ON_DEVICE)
1420 seq_puts(m, " Partition Aligned,");
1421 if (tcon->ss_flags & SSINFO_FLAGS_NO_SEEK_PENALTY)
1422 seq_puts(m, " SSD,");
1423 if (tcon->ss_flags & SSINFO_FLAGS_TRIM_ENABLED)
1424 seq_puts(m, " TRIM-support,");
1425
Steve French769ee6a2013-06-19 14:15:30 -05001426 seq_printf(m, "\tShare Flags: 0x%x", tcon->share_flags);
Steve Frenche0386e42018-05-20 01:27:03 -05001427 seq_printf(m, "\n\ttid: 0x%x", tcon->tid);
Steven Frenchaf6a12e2013-10-09 20:55:53 -05001428 if (tcon->perf_sector_size)
1429 seq_printf(m, "\tOptimal sector size: 0x%x",
1430 tcon->perf_sector_size);
Steve Frenche0386e42018-05-20 01:27:03 -05001431 seq_printf(m, "\tMaximal Access: 0x%x", tcon->maximal_access);
Steve French769ee6a2013-06-19 14:15:30 -05001432}
1433
1434static void
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001435smb2_print_stats(struct seq_file *m, struct cifs_tcon *tcon)
1436{
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001437 atomic_t *sent = tcon->stats.smb2_stats.smb2_com_sent;
1438 atomic_t *failed = tcon->stats.smb2_stats.smb2_com_failed;
Steve French1995d282018-07-27 15:14:04 -05001439
1440 /*
1441 * Can't display SMB2_NEGOTIATE, SESSION_SETUP, LOGOFF, CANCEL and ECHO
1442 * totals (requests sent) since those SMBs are per-session not per tcon
1443 */
Steve French52ce1ac2018-07-31 01:46:47 -05001444 seq_printf(m, "\nBytes read: %llu Bytes written: %llu",
1445 (long long)(tcon->bytes_read),
1446 (long long)(tcon->bytes_written));
Steve Frenchfae80442018-10-19 17:14:32 -05001447 seq_printf(m, "\nOpen files: %d total (local), %d open on server",
1448 atomic_read(&tcon->num_local_opens),
1449 atomic_read(&tcon->num_remote_opens));
Steve French1995d282018-07-27 15:14:04 -05001450 seq_printf(m, "\nTreeConnects: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001451 atomic_read(&sent[SMB2_TREE_CONNECT_HE]),
1452 atomic_read(&failed[SMB2_TREE_CONNECT_HE]));
Steve French1995d282018-07-27 15:14:04 -05001453 seq_printf(m, "\nTreeDisconnects: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001454 atomic_read(&sent[SMB2_TREE_DISCONNECT_HE]),
1455 atomic_read(&failed[SMB2_TREE_DISCONNECT_HE]));
Steve French1995d282018-07-27 15:14:04 -05001456 seq_printf(m, "\nCreates: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001457 atomic_read(&sent[SMB2_CREATE_HE]),
1458 atomic_read(&failed[SMB2_CREATE_HE]));
Steve French1995d282018-07-27 15:14:04 -05001459 seq_printf(m, "\nCloses: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001460 atomic_read(&sent[SMB2_CLOSE_HE]),
1461 atomic_read(&failed[SMB2_CLOSE_HE]));
Steve French1995d282018-07-27 15:14:04 -05001462 seq_printf(m, "\nFlushes: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001463 atomic_read(&sent[SMB2_FLUSH_HE]),
1464 atomic_read(&failed[SMB2_FLUSH_HE]));
Steve French1995d282018-07-27 15:14:04 -05001465 seq_printf(m, "\nReads: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001466 atomic_read(&sent[SMB2_READ_HE]),
1467 atomic_read(&failed[SMB2_READ_HE]));
Steve French1995d282018-07-27 15:14:04 -05001468 seq_printf(m, "\nWrites: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001469 atomic_read(&sent[SMB2_WRITE_HE]),
1470 atomic_read(&failed[SMB2_WRITE_HE]));
Steve French1995d282018-07-27 15:14:04 -05001471 seq_printf(m, "\nLocks: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001472 atomic_read(&sent[SMB2_LOCK_HE]),
1473 atomic_read(&failed[SMB2_LOCK_HE]));
Steve French1995d282018-07-27 15:14:04 -05001474 seq_printf(m, "\nIOCTLs: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001475 atomic_read(&sent[SMB2_IOCTL_HE]),
1476 atomic_read(&failed[SMB2_IOCTL_HE]));
Steve French1995d282018-07-27 15:14:04 -05001477 seq_printf(m, "\nQueryDirectories: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001478 atomic_read(&sent[SMB2_QUERY_DIRECTORY_HE]),
1479 atomic_read(&failed[SMB2_QUERY_DIRECTORY_HE]));
Steve French1995d282018-07-27 15:14:04 -05001480 seq_printf(m, "\nChangeNotifies: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001481 atomic_read(&sent[SMB2_CHANGE_NOTIFY_HE]),
1482 atomic_read(&failed[SMB2_CHANGE_NOTIFY_HE]));
Steve French1995d282018-07-27 15:14:04 -05001483 seq_printf(m, "\nQueryInfos: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001484 atomic_read(&sent[SMB2_QUERY_INFO_HE]),
1485 atomic_read(&failed[SMB2_QUERY_INFO_HE]));
Steve French1995d282018-07-27 15:14:04 -05001486 seq_printf(m, "\nSetInfos: %d total %d failed",
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001487 atomic_read(&sent[SMB2_SET_INFO_HE]),
1488 atomic_read(&failed[SMB2_SET_INFO_HE]));
1489 seq_printf(m, "\nOplockBreaks: %d sent %d failed",
1490 atomic_read(&sent[SMB2_OPLOCK_BREAK_HE]),
1491 atomic_read(&failed[SMB2_OPLOCK_BREAK_HE]));
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04001492}
1493
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001494static void
1495smb2_set_fid(struct cifsFileInfo *cfile, struct cifs_fid *fid, __u32 oplock)
1496{
David Howells2b0143b2015-03-17 22:25:59 +00001497 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04001498 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1499
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001500 cfile->fid.persistent_fid = fid->persistent_fid;
1501 cfile->fid.volatile_fid = fid->volatile_fid;
Aurelien Aptel86f740f2020-02-21 11:19:06 +01001502 cfile->fid.access = fid->access;
Steve Frenchdfe33f92018-10-30 19:50:31 -05001503#ifdef CONFIG_CIFS_DEBUG2
1504 cfile->fid.mid = fid->mid;
1505#endif /* CIFS_DEBUG2 */
Pavel Shilovsky42873b02013-09-05 21:30:16 +04001506 server->ops->set_oplock_level(cinode, oplock, fid->epoch,
1507 &fid->purge_cache);
Pavel Shilovsky18cceb62013-09-05 13:01:06 +04001508 cinode->can_cache_brlcks = CIFS_CACHE_WRITE(cinode);
Aurelien Aptel94f87372016-09-22 07:38:50 +02001509 memcpy(cfile->fid.create_guid, fid->create_guid, 16);
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001510}
1511
Pavel Shilovsky760ad0c2012-09-25 11:00:07 +04001512static void
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001513smb2_close_file(const unsigned int xid, struct cifs_tcon *tcon,
1514 struct cifs_fid *fid)
1515{
Pavel Shilovsky760ad0c2012-09-25 11:00:07 +04001516 SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001517}
1518
Steve French43f8a6a2019-12-02 21:46:54 -06001519static void
1520smb2_close_getattr(const unsigned int xid, struct cifs_tcon *tcon,
1521 struct cifsFileInfo *cfile)
1522{
1523 struct smb2_file_network_open_info file_inf;
1524 struct inode *inode;
1525 int rc;
1526
1527 rc = __SMB2_close(xid, tcon, cfile->fid.persistent_fid,
1528 cfile->fid.volatile_fid, &file_inf);
1529 if (rc)
1530 return;
1531
1532 inode = d_inode(cfile->dentry);
1533
1534 spin_lock(&inode->i_lock);
1535 CIFS_I(inode)->time = jiffies;
1536
1537 /* Creation time should not need to be updated on close */
1538 if (file_inf.LastWriteTime)
1539 inode->i_mtime = cifs_NTtimeToUnix(file_inf.LastWriteTime);
1540 if (file_inf.ChangeTime)
1541 inode->i_ctime = cifs_NTtimeToUnix(file_inf.ChangeTime);
1542 if (file_inf.LastAccessTime)
1543 inode->i_atime = cifs_NTtimeToUnix(file_inf.LastAccessTime);
1544
1545 /*
1546 * i_blocks is not related to (i_size / i_blksize),
1547 * but instead 512 byte (2**9) size is required for
1548 * calculating num blocks.
1549 */
1550 if (le64_to_cpu(file_inf.AllocationSize) > 4096)
1551 inode->i_blocks =
1552 (512 - 1 + le64_to_cpu(file_inf.AllocationSize)) >> 9;
1553
1554 /* End of file and Attributes should not have to be updated on close */
1555 spin_unlock(&inode->i_lock);
1556}
1557
Pavel Shilovsky7a5cfb12012-09-18 16:20:28 -07001558static int
Steve French41c13582013-11-14 00:05:36 -06001559SMB2_request_res_key(const unsigned int xid, struct cifs_tcon *tcon,
1560 u64 persistent_fid, u64 volatile_fid,
1561 struct copychunk_ioctl *pcchunk)
1562{
1563 int rc;
1564 unsigned int ret_data_len;
1565 struct resume_key_req *res_key;
1566
1567 rc = SMB2_ioctl(xid, tcon, persistent_fid, volatile_fid,
1568 FSCTL_SRV_REQUEST_RESUME_KEY, true /* is_fsctl */,
Steve French153322f2019-03-28 22:32:49 -05001569 NULL, 0 /* no input */, CIFSMaxBufSize,
Steve French41c13582013-11-14 00:05:36 -06001570 (char **)&res_key, &ret_data_len);
1571
Steve French423333b2021-04-19 23:22:37 -05001572 if (rc == -EOPNOTSUPP) {
1573 pr_warn_once("Server share %s does not support copy range\n", tcon->treeName);
1574 goto req_res_key_exit;
1575 } else if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10001576 cifs_tcon_dbg(VFS, "refcpy ioctl error %d getting resume key\n", rc);
Steve French41c13582013-11-14 00:05:36 -06001577 goto req_res_key_exit;
1578 }
1579 if (ret_data_len < sizeof(struct resume_key_req)) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10001580 cifs_tcon_dbg(VFS, "Invalid refcopy resume key length\n");
Steve French41c13582013-11-14 00:05:36 -06001581 rc = -EINVAL;
1582 goto req_res_key_exit;
1583 }
1584 memcpy(pcchunk->SourceKey, res_key->ResumeKey, COPY_CHUNK_RES_KEY_SIZE);
1585
1586req_res_key_exit:
1587 kfree(res_key);
1588 return rc;
1589}
1590
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001591struct iqi_vars {
1592 struct smb_rqst rqst[3];
1593 struct kvec rsp_iov[3];
1594 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
1595 struct kvec qi_iov[1];
1596 struct kvec io_iov[SMB2_IOCTL_IOV_SIZE];
1597 struct kvec si_iov[SMB2_SET_INFO_IOV_SIZE];
1598 struct kvec close_iov[1];
1599};
1600
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001601static int
1602smb2_ioctl_query_info(const unsigned int xid,
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001603 struct cifs_tcon *tcon,
Amir Goldstein0f060932020-02-03 21:46:43 +02001604 struct cifs_sb_info *cifs_sb,
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001605 __le16 *path, int is_dir,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001606 unsigned long p)
1607{
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001608 struct iqi_vars *vars;
1609 struct smb_rqst *rqst;
1610 struct kvec *rsp_iov;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001611 struct cifs_ses *ses = tcon->ses;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001612 struct TCP_Server_Info *server = cifs_pick_channel(ses);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001613 char __user *arg = (char __user *)p;
1614 struct smb_query_info qi;
1615 struct smb_query_info __user *pqi;
1616 int rc = 0;
Paulo Alcantara04ad69c2021-03-08 12:00:50 -03001617 int flags = CIFS_CP_CREATE_CLOSE_OP;
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001618 struct smb2_query_info_rsp *qi_rsp = NULL;
1619 struct smb2_ioctl_rsp *io_rsp = NULL;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001620 void *buffer = NULL;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001621 int resp_buftype[3];
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001622 struct cifs_open_parms oparms;
1623 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
1624 struct cifs_fid fid;
Ronnie Sahlberg0e906962019-07-25 13:08:43 +10001625 unsigned int size[2];
1626 void *data[2];
Amir Goldstein0f060932020-02-03 21:46:43 +02001627 int create_options = is_dir ? CREATE_NOT_FILE : CREATE_NOT_DIR;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001628
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001629 vars = kzalloc(sizeof(*vars), GFP_ATOMIC);
1630 if (vars == NULL)
1631 return -ENOMEM;
1632 rqst = &vars->rqst[0];
1633 rsp_iov = &vars->rsp_iov[0];
1634
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001635 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001636
1637 if (copy_from_user(&qi, arg, sizeof(struct smb_query_info)))
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001638 goto e_fault;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001639
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001640 if (qi.output_buffer_length > 1024) {
1641 kfree(vars);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001642 return -EINVAL;
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001643 }
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001644
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001645 if (!ses || !server) {
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001646 kfree(vars);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001647 return -EIO;
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001648 }
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001649
1650 if (smb3_encryption_required(tcon))
1651 flags |= CIFS_TRANSFORM_REQ;
1652
Markus Elfringcfaa1182019-11-05 21:30:25 +01001653 buffer = memdup_user(arg + sizeof(struct smb_query_info),
1654 qi.output_buffer_length);
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001655 if (IS_ERR(buffer)) {
1656 kfree(vars);
Markus Elfringcfaa1182019-11-05 21:30:25 +01001657 return PTR_ERR(buffer);
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001658 }
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001659
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001660 /* Open */
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001661 rqst[0].rq_iov = &vars->open_iov[0];
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001662 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001663
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001664 memset(&oparms, 0, sizeof(oparms));
1665 oparms.tcon = tcon;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001666 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02001667 oparms.create_options = cifs_create_options(cifs_sb, create_options);
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001668 oparms.fid = &fid;
1669 oparms.reconnect = false;
1670
Ronnie Sahlbergefac7792019-04-11 12:20:17 +10001671 if (qi.flags & PASSTHRU_FSCTL) {
1672 switch (qi.info_type & FSCTL_DEVICE_ACCESS_MASK) {
1673 case FSCTL_DEVICE_ACCESS_FILE_READ_WRITE_ACCESS:
1674 oparms.desired_access = FILE_READ_DATA | FILE_WRITE_DATA | FILE_READ_ATTRIBUTES | SYNCHRONIZE;
Steve French46e66612019-04-11 13:53:17 -05001675 break;
1676 case FSCTL_DEVICE_ACCESS_FILE_ANY_ACCESS:
1677 oparms.desired_access = GENERIC_ALL;
1678 break;
1679 case FSCTL_DEVICE_ACCESS_FILE_READ_ACCESS:
1680 oparms.desired_access = GENERIC_READ;
1681 break;
1682 case FSCTL_DEVICE_ACCESS_FILE_WRITE_ACCESS:
1683 oparms.desired_access = GENERIC_WRITE;
Ronnie Sahlbergefac7792019-04-11 12:20:17 +10001684 break;
1685 }
Ronnie Sahlberg0e906962019-07-25 13:08:43 +10001686 } else if (qi.flags & PASSTHRU_SET_INFO) {
1687 oparms.desired_access = GENERIC_WRITE;
1688 } else {
1689 oparms.desired_access = FILE_READ_ATTRIBUTES | READ_CONTROL;
Ronnie Sahlbergefac7792019-04-11 12:20:17 +10001690 }
1691
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001692 rc = SMB2_open_init(tcon, server,
1693 &rqst[0], &oplock, &oparms, path);
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001694 if (rc)
1695 goto iqinf_exit;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10001696 smb2_set_next_command(tcon, &rqst[0]);
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001697
1698 /* Query */
Steve French31ba4332019-03-13 02:40:07 -05001699 if (qi.flags & PASSTHRU_FSCTL) {
1700 /* Can eventually relax perm check since server enforces too */
1701 if (!capable(CAP_SYS_ADMIN))
1702 rc = -EPERM;
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001703 else {
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001704 rqst[1].rq_iov = &vars->io_iov[0];
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001705 rqst[1].rq_nvec = SMB2_IOCTL_IOV_SIZE;
1706
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001707 rc = SMB2_ioctl_init(tcon, server,
1708 &rqst[1],
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001709 COMPOUND_FID, COMPOUND_FID,
Ronnie Sahlbergefac7792019-04-11 12:20:17 +10001710 qi.info_type, true, buffer,
1711 qi.output_buffer_length,
Ronnie Sahlberg731b82b2020-01-08 13:08:07 +10001712 CIFSMaxBufSize -
1713 MAX_SMB2_CREATE_RESPONSE_SIZE -
1714 MAX_SMB2_CLOSE_RESPONSE_SIZE);
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001715 }
Ronnie Sahlberg0e906962019-07-25 13:08:43 +10001716 } else if (qi.flags == PASSTHRU_SET_INFO) {
1717 /* Can eventually relax perm check since server enforces too */
1718 if (!capable(CAP_SYS_ADMIN))
1719 rc = -EPERM;
1720 else {
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001721 rqst[1].rq_iov = &vars->si_iov[0];
Ronnie Sahlberg0e906962019-07-25 13:08:43 +10001722 rqst[1].rq_nvec = 1;
1723
1724 size[0] = 8;
1725 data[0] = buffer;
1726
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001727 rc = SMB2_set_info_init(tcon, server,
1728 &rqst[1],
Ronnie Sahlberg0e906962019-07-25 13:08:43 +10001729 COMPOUND_FID, COMPOUND_FID,
1730 current->tgid,
1731 FILE_END_OF_FILE_INFORMATION,
1732 SMB2_O_INFO_FILE, 0, data, size);
1733 }
Steve French31ba4332019-03-13 02:40:07 -05001734 } else if (qi.flags == PASSTHRU_QUERY_INFO) {
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001735 rqst[1].rq_iov = &vars->qi_iov[0];
Steve French31ba4332019-03-13 02:40:07 -05001736 rqst[1].rq_nvec = 1;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001737
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001738 rc = SMB2_query_info_init(tcon, server,
1739 &rqst[1], COMPOUND_FID,
Steve French31ba4332019-03-13 02:40:07 -05001740 COMPOUND_FID, qi.file_info_class,
1741 qi.info_type, qi.additional_information,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001742 qi.input_buffer_length,
1743 qi.output_buffer_length, buffer);
Steve French31ba4332019-03-13 02:40:07 -05001744 } else { /* unknown flags */
Joe Perchesa0a30362020-04-14 22:42:53 -07001745 cifs_tcon_dbg(VFS, "Invalid passthru query flags: 0x%x\n",
1746 qi.flags);
Steve French31ba4332019-03-13 02:40:07 -05001747 rc = -EINVAL;
1748 }
1749
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001750 if (rc)
1751 goto iqinf_exit;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10001752 smb2_set_next_command(tcon, &rqst[1]);
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001753 smb2_set_related(&rqst[1]);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001754
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001755 /* Close */
Ronnie Sahlbergb2ca6c22020-05-21 15:03:15 +10001756 rqst[2].rq_iov = &vars->close_iov[0];
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001757 rqst[2].rq_nvec = 1;
1758
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001759 rc = SMB2_close_init(tcon, server,
1760 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001761 if (rc)
1762 goto iqinf_exit;
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001763 smb2_set_related(&rqst[2]);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001764
Aurelien Aptel352d96f2020-05-31 12:38:22 -05001765 rc = compound_send_recv(xid, ses, server,
1766 flags, 3, rqst,
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001767 resp_buftype, rsp_iov);
1768 if (rc)
1769 goto iqinf_exit;
Steve Frenchd2f15422019-09-22 00:55:46 -05001770
1771 /* No need to bump num_remote_opens since handle immediately closed */
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001772 if (qi.flags & PASSTHRU_FSCTL) {
1773 pqi = (struct smb_query_info __user *)arg;
1774 io_rsp = (struct smb2_ioctl_rsp *)rsp_iov[1].iov_base;
1775 if (le32_to_cpu(io_rsp->OutputCount) < qi.input_buffer_length)
1776 qi.input_buffer_length = le32_to_cpu(io_rsp->OutputCount);
Ronnie Sahlberg5242fcb2019-04-15 12:13:52 +10001777 if (qi.input_buffer_length > 0 &&
Markus Elfring2b1116b2019-11-05 22:26:53 +01001778 le32_to_cpu(io_rsp->OutputOffset) + qi.input_buffer_length
1779 > rsp_iov[1].iov_len)
1780 goto e_fault;
1781
1782 if (copy_to_user(&pqi->input_buffer_length,
1783 &qi.input_buffer_length,
1784 sizeof(qi.input_buffer_length)))
1785 goto e_fault;
1786
Ronnie Sahlberg5242fcb2019-04-15 12:13:52 +10001787 if (copy_to_user((void __user *)pqi + sizeof(struct smb_query_info),
1788 (const void *)io_rsp + le32_to_cpu(io_rsp->OutputOffset),
Markus Elfring2b1116b2019-11-05 22:26:53 +01001789 qi.input_buffer_length))
1790 goto e_fault;
Ronnie Sahlbergf5778c32019-03-15 09:07:22 +10001791 } else {
1792 pqi = (struct smb_query_info __user *)arg;
1793 qi_rsp = (struct smb2_query_info_rsp *)rsp_iov[1].iov_base;
1794 if (le32_to_cpu(qi_rsp->OutputBufferLength) < qi.input_buffer_length)
1795 qi.input_buffer_length = le32_to_cpu(qi_rsp->OutputBufferLength);
Markus Elfring2b1116b2019-11-05 22:26:53 +01001796 if (copy_to_user(&pqi->input_buffer_length,
1797 &qi.input_buffer_length,
1798 sizeof(qi.input_buffer_length)))
1799 goto e_fault;
1800
1801 if (copy_to_user(pqi + 1, qi_rsp->Buffer,
1802 qi.input_buffer_length))
1803 goto e_fault;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001804 }
1805
1806 iqinf_exit:
Aurelien Aptelccd48ec2021-04-09 15:47:01 +02001807 cifs_small_buf_release(rqst[0].rq_iov[0].iov_base);
1808 cifs_small_buf_release(rqst[1].rq_iov[0].iov_base);
1809 cifs_small_buf_release(rqst[2].rq_iov[0].iov_base);
Ronnie Sahlberg8d8b26e2018-10-17 05:47:58 +10001810 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
1811 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1812 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
Aurelien Aptelccd48ec2021-04-09 15:47:01 +02001813 kfree(vars);
1814 kfree(buffer);
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001815 return rc;
Markus Elfring2b1116b2019-11-05 22:26:53 +01001816
1817e_fault:
1818 rc = -EFAULT;
1819 goto iqinf_exit;
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05001820}
1821
Sachin Prabhu620d8742017-02-10 16:03:51 +05301822static ssize_t
Sachin Prabhu312bbc52017-04-04 02:12:04 -05001823smb2_copychunk_range(const unsigned int xid,
Steve French41c13582013-11-14 00:05:36 -06001824 struct cifsFileInfo *srcfile,
1825 struct cifsFileInfo *trgtfile, u64 src_off,
1826 u64 len, u64 dest_off)
1827{
1828 int rc;
1829 unsigned int ret_data_len;
1830 struct copychunk_ioctl *pcchunk;
Steve French9bf0c9c2013-11-16 18:05:28 -06001831 struct copychunk_ioctl_rsp *retbuf = NULL;
1832 struct cifs_tcon *tcon;
1833 int chunks_copied = 0;
1834 bool chunk_sizes_updated = false;
Sachin Prabhu620d8742017-02-10 16:03:51 +05301835 ssize_t bytes_written, total_bytes_written = 0;
Steve French41c13582013-11-14 00:05:36 -06001836
1837 pcchunk = kmalloc(sizeof(struct copychunk_ioctl), GFP_KERNEL);
1838
1839 if (pcchunk == NULL)
1840 return -ENOMEM;
1841
Christoph Probsta205d502019-05-08 21:36:25 +02001842 cifs_dbg(FYI, "%s: about to call request res key\n", __func__);
Steve French41c13582013-11-14 00:05:36 -06001843 /* Request a key from the server to identify the source of the copy */
1844 rc = SMB2_request_res_key(xid, tlink_tcon(srcfile->tlink),
1845 srcfile->fid.persistent_fid,
1846 srcfile->fid.volatile_fid, pcchunk);
1847
1848 /* Note: request_res_key sets res_key null only if rc !=0 */
1849 if (rc)
Steve French9bf0c9c2013-11-16 18:05:28 -06001850 goto cchunk_out;
Steve French41c13582013-11-14 00:05:36 -06001851
1852 /* For now array only one chunk long, will make more flexible later */
Fabian Frederickbc09d142014-12-10 15:41:15 -08001853 pcchunk->ChunkCount = cpu_to_le32(1);
Steve French41c13582013-11-14 00:05:36 -06001854 pcchunk->Reserved = 0;
Steve French41c13582013-11-14 00:05:36 -06001855 pcchunk->Reserved2 = 0;
1856
Steve French9bf0c9c2013-11-16 18:05:28 -06001857 tcon = tlink_tcon(trgtfile->tlink);
1858
1859 while (len > 0) {
1860 pcchunk->SourceOffset = cpu_to_le64(src_off);
1861 pcchunk->TargetOffset = cpu_to_le64(dest_off);
1862 pcchunk->Length =
1863 cpu_to_le32(min_t(u32, len, tcon->max_bytes_chunk));
1864
1865 /* Request server copy to target from src identified by key */
Ronnie Sahlbergd201d762021-05-19 08:40:11 +10001866 kfree(retbuf);
1867 retbuf = NULL;
Steve French9bf0c9c2013-11-16 18:05:28 -06001868 rc = SMB2_ioctl(xid, tcon, trgtfile->fid.persistent_fid,
Steve French41c13582013-11-14 00:05:36 -06001869 trgtfile->fid.volatile_fid, FSCTL_SRV_COPYCHUNK_WRITE,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01001870 true /* is_fsctl */, (char *)pcchunk,
Steve French153322f2019-03-28 22:32:49 -05001871 sizeof(struct copychunk_ioctl), CIFSMaxBufSize,
1872 (char **)&retbuf, &ret_data_len);
Steve French9bf0c9c2013-11-16 18:05:28 -06001873 if (rc == 0) {
1874 if (ret_data_len !=
1875 sizeof(struct copychunk_ioctl_rsp)) {
Joe Perchesa0a30362020-04-14 22:42:53 -07001876 cifs_tcon_dbg(VFS, "Invalid cchunk response size\n");
Steve French9bf0c9c2013-11-16 18:05:28 -06001877 rc = -EIO;
1878 goto cchunk_out;
1879 }
1880 if (retbuf->TotalBytesWritten == 0) {
1881 cifs_dbg(FYI, "no bytes copied\n");
1882 rc = -EIO;
1883 goto cchunk_out;
1884 }
1885 /*
1886 * Check if server claimed to write more than we asked
1887 */
1888 if (le32_to_cpu(retbuf->TotalBytesWritten) >
1889 le32_to_cpu(pcchunk->Length)) {
Joe Perchesa0a30362020-04-14 22:42:53 -07001890 cifs_tcon_dbg(VFS, "Invalid copy chunk response\n");
Steve French9bf0c9c2013-11-16 18:05:28 -06001891 rc = -EIO;
1892 goto cchunk_out;
1893 }
1894 if (le32_to_cpu(retbuf->ChunksWritten) != 1) {
Joe Perchesa0a30362020-04-14 22:42:53 -07001895 cifs_tcon_dbg(VFS, "Invalid num chunks written\n");
Steve French9bf0c9c2013-11-16 18:05:28 -06001896 rc = -EIO;
1897 goto cchunk_out;
1898 }
1899 chunks_copied++;
Steve French41c13582013-11-14 00:05:36 -06001900
Sachin Prabhu620d8742017-02-10 16:03:51 +05301901 bytes_written = le32_to_cpu(retbuf->TotalBytesWritten);
1902 src_off += bytes_written;
1903 dest_off += bytes_written;
1904 len -= bytes_written;
1905 total_bytes_written += bytes_written;
Steve French41c13582013-11-14 00:05:36 -06001906
Sachin Prabhu620d8742017-02-10 16:03:51 +05301907 cifs_dbg(FYI, "Chunks %d PartialChunk %d Total %zu\n",
Steve French9bf0c9c2013-11-16 18:05:28 -06001908 le32_to_cpu(retbuf->ChunksWritten),
1909 le32_to_cpu(retbuf->ChunkBytesWritten),
Sachin Prabhu620d8742017-02-10 16:03:51 +05301910 bytes_written);
Steve French9bf0c9c2013-11-16 18:05:28 -06001911 } else if (rc == -EINVAL) {
1912 if (ret_data_len != sizeof(struct copychunk_ioctl_rsp))
1913 goto cchunk_out;
Steve French41c13582013-11-14 00:05:36 -06001914
Steve French9bf0c9c2013-11-16 18:05:28 -06001915 cifs_dbg(FYI, "MaxChunks %d BytesChunk %d MaxCopy %d\n",
1916 le32_to_cpu(retbuf->ChunksWritten),
1917 le32_to_cpu(retbuf->ChunkBytesWritten),
1918 le32_to_cpu(retbuf->TotalBytesWritten));
1919
1920 /*
1921 * Check if this is the first request using these sizes,
1922 * (ie check if copy succeed once with original sizes
1923 * and check if the server gave us different sizes after
1924 * we already updated max sizes on previous request).
1925 * if not then why is the server returning an error now
1926 */
1927 if ((chunks_copied != 0) || chunk_sizes_updated)
1928 goto cchunk_out;
1929
1930 /* Check that server is not asking us to grow size */
1931 if (le32_to_cpu(retbuf->ChunkBytesWritten) <
1932 tcon->max_bytes_chunk)
1933 tcon->max_bytes_chunk =
1934 le32_to_cpu(retbuf->ChunkBytesWritten);
1935 else
1936 goto cchunk_out; /* server gave us bogus size */
1937
1938 /* No need to change MaxChunks since already set to 1 */
1939 chunk_sizes_updated = true;
Sachin Prabhu2477bc52015-02-04 13:10:26 +00001940 } else
1941 goto cchunk_out;
Steve French9bf0c9c2013-11-16 18:05:28 -06001942 }
1943
1944cchunk_out:
Steve French41c13582013-11-14 00:05:36 -06001945 kfree(pcchunk);
Steve French24df1482016-09-29 04:20:23 -05001946 kfree(retbuf);
Sachin Prabhu620d8742017-02-10 16:03:51 +05301947 if (rc)
1948 return rc;
1949 else
1950 return total_bytes_written;
Steve French41c13582013-11-14 00:05:36 -06001951}
1952
1953static int
Pavel Shilovsky7a5cfb12012-09-18 16:20:28 -07001954smb2_flush_file(const unsigned int xid, struct cifs_tcon *tcon,
1955 struct cifs_fid *fid)
1956{
1957 return SMB2_flush(xid, tcon, fid->persistent_fid, fid->volatile_fid);
1958}
1959
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001960static unsigned int
1961smb2_read_data_offset(char *buf)
1962{
1963 struct smb2_read_rsp *rsp = (struct smb2_read_rsp *)buf;
Christoph Probsta205d502019-05-08 21:36:25 +02001964
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001965 return rsp->DataOffset;
1966}
1967
1968static unsigned int
Long Li74dcf412017-11-22 17:38:46 -07001969smb2_read_data_length(char *buf, bool in_remaining)
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001970{
1971 struct smb2_read_rsp *rsp = (struct smb2_read_rsp *)buf;
Long Li74dcf412017-11-22 17:38:46 -07001972
1973 if (in_remaining)
1974 return le32_to_cpu(rsp->DataRemaining);
1975
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001976 return le32_to_cpu(rsp->DataLength);
1977}
1978
Pavel Shilovskyd8e05032012-09-18 16:20:30 -07001979
1980static int
Steve Frenchdb8b6312014-09-22 05:13:55 -05001981smb2_sync_read(const unsigned int xid, struct cifs_fid *pfid,
Pavel Shilovskyd8e05032012-09-18 16:20:30 -07001982 struct cifs_io_parms *parms, unsigned int *bytes_read,
1983 char **buf, int *buf_type)
1984{
Steve Frenchdb8b6312014-09-22 05:13:55 -05001985 parms->persistent_fid = pfid->persistent_fid;
1986 parms->volatile_fid = pfid->volatile_fid;
Pavel Shilovskyd8e05032012-09-18 16:20:30 -07001987 return SMB2_read(xid, parms, bytes_read, buf, buf_type);
1988}
1989
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001990static int
Steve Frenchdb8b6312014-09-22 05:13:55 -05001991smb2_sync_write(const unsigned int xid, struct cifs_fid *pfid,
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001992 struct cifs_io_parms *parms, unsigned int *written,
1993 struct kvec *iov, unsigned long nr_segs)
1994{
1995
Steve Frenchdb8b6312014-09-22 05:13:55 -05001996 parms->persistent_fid = pfid->persistent_fid;
1997 parms->volatile_fid = pfid->volatile_fid;
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001998 return SMB2_write(xid, parms, written, iov, nr_segs);
1999}
2000
Steve Frenchd43cc792014-08-13 17:16:29 -05002001/* Set or clear the SPARSE_FILE attribute based on value passed in setsparse */
2002static bool smb2_set_sparse(const unsigned int xid, struct cifs_tcon *tcon,
2003 struct cifsFileInfo *cfile, struct inode *inode, __u8 setsparse)
2004{
2005 struct cifsInodeInfo *cifsi;
2006 int rc;
2007
2008 cifsi = CIFS_I(inode);
2009
2010 /* if file already sparse don't bother setting sparse again */
2011 if ((cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) && setsparse)
2012 return true; /* already sparse */
2013
2014 if (!(cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) && !setsparse)
2015 return true; /* already not sparse */
2016
2017 /*
2018 * Can't check for sparse support on share the usual way via the
2019 * FS attribute info (FILE_SUPPORTS_SPARSE_FILES) on the share
2020 * since Samba server doesn't set the flag on the share, yet
2021 * supports the set sparse FSCTL and returns sparse correctly
2022 * in the file attributes. If we fail setting sparse though we
2023 * mark that server does not support sparse files for this share
2024 * to avoid repeatedly sending the unsupported fsctl to server
2025 * if the file is repeatedly extended.
2026 */
2027 if (tcon->broken_sparse_sup)
2028 return false;
2029
2030 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
2031 cfile->fid.volatile_fid, FSCTL_SET_SPARSE,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002032 true /* is_fctl */,
Steve French153322f2019-03-28 22:32:49 -05002033 &setsparse, 1, CIFSMaxBufSize, NULL, NULL);
Steve Frenchd43cc792014-08-13 17:16:29 -05002034 if (rc) {
2035 tcon->broken_sparse_sup = true;
2036 cifs_dbg(FYI, "set sparse rc = %d\n", rc);
2037 return false;
2038 }
2039
2040 if (setsparse)
2041 cifsi->cifsAttrs |= FILE_ATTRIBUTE_SPARSE_FILE;
2042 else
2043 cifsi->cifsAttrs &= (~FILE_ATTRIBUTE_SPARSE_FILE);
2044
2045 return true;
2046}
2047
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07002048static int
2049smb2_set_file_size(const unsigned int xid, struct cifs_tcon *tcon,
2050 struct cifsFileInfo *cfile, __u64 size, bool set_alloc)
2051{
2052 __le64 eof = cpu_to_le64(size);
Steve French3d1a3742014-08-11 21:05:25 -05002053 struct inode *inode;
2054
2055 /*
2056 * If extending file more than one page make sparse. Many Linux fs
2057 * make files sparse by default when extending via ftruncate
2058 */
David Howells2b0143b2015-03-17 22:25:59 +00002059 inode = d_inode(cfile->dentry);
Steve French3d1a3742014-08-11 21:05:25 -05002060
2061 if (!set_alloc && (size > inode->i_size + 8192)) {
Steve French3d1a3742014-08-11 21:05:25 -05002062 __u8 set_sparse = 1;
Steve French3d1a3742014-08-11 21:05:25 -05002063
Steve Frenchd43cc792014-08-13 17:16:29 -05002064 /* whether set sparse succeeds or not, extend the file */
2065 smb2_set_sparse(xid, tcon, cfile, inode, set_sparse);
Steve French3d1a3742014-08-11 21:05:25 -05002066 }
2067
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07002068 return SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
Ronnie Sahlberg3764cbd2018-09-03 13:33:47 +10002069 cfile->fid.volatile_fid, cfile->pid, &eof);
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07002070}
2071
Steve French02b16662015-06-27 21:18:36 -07002072static int
2073smb2_duplicate_extents(const unsigned int xid,
2074 struct cifsFileInfo *srcfile,
2075 struct cifsFileInfo *trgtfile, u64 src_off,
2076 u64 len, u64 dest_off)
2077{
2078 int rc;
2079 unsigned int ret_data_len;
Steve Frenchcfc63fc2021-03-26 18:41:55 -05002080 struct inode *inode;
Steve French02b16662015-06-27 21:18:36 -07002081 struct duplicate_extents_to_file dup_ext_buf;
2082 struct cifs_tcon *tcon = tlink_tcon(trgtfile->tlink);
2083
2084 /* server fileays advertise duplicate extent support with this flag */
2085 if ((le32_to_cpu(tcon->fsAttrInfo.Attributes) &
2086 FILE_SUPPORTS_BLOCK_REFCOUNTING) == 0)
2087 return -EOPNOTSUPP;
2088
2089 dup_ext_buf.VolatileFileHandle = srcfile->fid.volatile_fid;
2090 dup_ext_buf.PersistentFileHandle = srcfile->fid.persistent_fid;
2091 dup_ext_buf.SourceFileOffset = cpu_to_le64(src_off);
2092 dup_ext_buf.TargetFileOffset = cpu_to_le64(dest_off);
2093 dup_ext_buf.ByteCount = cpu_to_le64(len);
Christoph Probsta205d502019-05-08 21:36:25 +02002094 cifs_dbg(FYI, "Duplicate extents: src off %lld dst off %lld len %lld\n",
Steve French02b16662015-06-27 21:18:36 -07002095 src_off, dest_off, len);
2096
Steve Frenchcfc63fc2021-03-26 18:41:55 -05002097 inode = d_inode(trgtfile->dentry);
2098 if (inode->i_size < dest_off + len) {
2099 rc = smb2_set_file_size(xid, tcon, trgtfile, dest_off + len, false);
2100 if (rc)
2101 goto duplicate_extents_out;
Steve French02b16662015-06-27 21:18:36 -07002102
Steve Frenchcfc63fc2021-03-26 18:41:55 -05002103 /*
2104 * Although also could set plausible allocation size (i_blocks)
2105 * here in addition to setting the file size, in reflink
2106 * it is likely that the target file is sparse. Its allocation
2107 * size will be queried on next revalidate, but it is important
2108 * to make sure that file's cached size is updated immediately
2109 */
2110 cifs_setsize(inode, dest_off + len);
2111 }
Steve French02b16662015-06-27 21:18:36 -07002112 rc = SMB2_ioctl(xid, tcon, trgtfile->fid.persistent_fid,
2113 trgtfile->fid.volatile_fid,
2114 FSCTL_DUPLICATE_EXTENTS_TO_FILE,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002115 true /* is_fsctl */,
Aurelien Aptel51146622017-02-28 15:08:41 +01002116 (char *)&dup_ext_buf,
Steve French02b16662015-06-27 21:18:36 -07002117 sizeof(struct duplicate_extents_to_file),
Steve French153322f2019-03-28 22:32:49 -05002118 CIFSMaxBufSize, NULL,
Steve French02b16662015-06-27 21:18:36 -07002119 &ret_data_len);
2120
2121 if (ret_data_len > 0)
Christoph Probsta205d502019-05-08 21:36:25 +02002122 cifs_dbg(FYI, "Non-zero response length in duplicate extents\n");
Steve French02b16662015-06-27 21:18:36 -07002123
2124duplicate_extents_out:
2125 return rc;
2126}
Steve French02b16662015-06-27 21:18:36 -07002127
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002128static int
Steve French64a5cfa2013-10-14 15:31:32 -05002129smb2_set_compression(const unsigned int xid, struct cifs_tcon *tcon,
2130 struct cifsFileInfo *cfile)
2131{
2132 return SMB2_set_compression(xid, tcon, cfile->fid.persistent_fid,
2133 cfile->fid.volatile_fid);
2134}
2135
2136static int
Steve Frenchb3152e22015-06-24 03:17:02 -05002137smb3_set_integrity(const unsigned int xid, struct cifs_tcon *tcon,
2138 struct cifsFileInfo *cfile)
2139{
2140 struct fsctl_set_integrity_information_req integr_info;
Steve Frenchb3152e22015-06-24 03:17:02 -05002141 unsigned int ret_data_len;
2142
2143 integr_info.ChecksumAlgorithm = cpu_to_le16(CHECKSUM_TYPE_UNCHANGED);
2144 integr_info.Flags = 0;
2145 integr_info.Reserved = 0;
2146
2147 return SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
2148 cfile->fid.volatile_fid,
2149 FSCTL_SET_INTEGRITY_INFORMATION,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002150 true /* is_fsctl */,
Aurelien Aptel51146622017-02-28 15:08:41 +01002151 (char *)&integr_info,
Steve Frenchb3152e22015-06-24 03:17:02 -05002152 sizeof(struct fsctl_set_integrity_information_req),
Steve French153322f2019-03-28 22:32:49 -05002153 CIFSMaxBufSize, NULL,
Steve Frenchb3152e22015-06-24 03:17:02 -05002154 &ret_data_len);
2155
2156}
2157
Steve Frenche02789a2018-08-09 14:33:12 -05002158/* GMT Token is @GMT-YYYY.MM.DD-HH.MM.SS Unicode which is 48 bytes + null */
2159#define GMT_TOKEN_SIZE 50
2160
Steve French153322f2019-03-28 22:32:49 -05002161#define MIN_SNAPSHOT_ARRAY_SIZE 16 /* See MS-SMB2 section 3.3.5.15.1 */
2162
Steve Frenche02789a2018-08-09 14:33:12 -05002163/*
2164 * Input buffer contains (empty) struct smb_snapshot array with size filled in
2165 * For output see struct SRV_SNAPSHOT_ARRAY in MS-SMB2 section 2.2.32.2
2166 */
Steve Frenchb3152e22015-06-24 03:17:02 -05002167static int
Steve French834170c2016-09-30 21:14:26 -05002168smb3_enum_snapshots(const unsigned int xid, struct cifs_tcon *tcon,
2169 struct cifsFileInfo *cfile, void __user *ioc_buf)
2170{
2171 char *retbuf = NULL;
2172 unsigned int ret_data_len = 0;
2173 int rc;
Steve French153322f2019-03-28 22:32:49 -05002174 u32 max_response_size;
Steve French834170c2016-09-30 21:14:26 -05002175 struct smb_snapshot_array snapshot_in;
2176
Steve French973189a2019-04-04 00:41:04 -05002177 /*
2178 * On the first query to enumerate the list of snapshots available
2179 * for this volume the buffer begins with 0 (number of snapshots
2180 * which can be returned is zero since at that point we do not know
2181 * how big the buffer needs to be). On the second query,
2182 * it (ret_data_len) is set to number of snapshots so we can
2183 * know to set the maximum response size larger (see below).
2184 */
Steve French153322f2019-03-28 22:32:49 -05002185 if (get_user(ret_data_len, (unsigned int __user *)ioc_buf))
2186 return -EFAULT;
2187
2188 /*
2189 * Note that for snapshot queries that servers like Azure expect that
2190 * the first query be minimal size (and just used to get the number/size
2191 * of previous versions) so response size must be specified as EXACTLY
2192 * sizeof(struct snapshot_array) which is 16 when rounded up to multiple
2193 * of eight bytes.
2194 */
2195 if (ret_data_len == 0)
2196 max_response_size = MIN_SNAPSHOT_ARRAY_SIZE;
2197 else
2198 max_response_size = CIFSMaxBufSize;
2199
Steve French834170c2016-09-30 21:14:26 -05002200 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
2201 cfile->fid.volatile_fid,
2202 FSCTL_SRV_ENUMERATE_SNAPSHOTS,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002203 true /* is_fsctl */,
Steve French153322f2019-03-28 22:32:49 -05002204 NULL, 0 /* no input data */, max_response_size,
Steve French834170c2016-09-30 21:14:26 -05002205 (char **)&retbuf,
2206 &ret_data_len);
2207 cifs_dbg(FYI, "enum snaphots ioctl returned %d and ret buflen is %d\n",
2208 rc, ret_data_len);
2209 if (rc)
2210 return rc;
2211
2212 if (ret_data_len && (ioc_buf != NULL) && (retbuf != NULL)) {
2213 /* Fixup buffer */
2214 if (copy_from_user(&snapshot_in, ioc_buf,
2215 sizeof(struct smb_snapshot_array))) {
2216 rc = -EFAULT;
2217 kfree(retbuf);
2218 return rc;
2219 }
Steve French834170c2016-09-30 21:14:26 -05002220
Steve Frenche02789a2018-08-09 14:33:12 -05002221 /*
2222 * Check for min size, ie not large enough to fit even one GMT
2223 * token (snapshot). On the first ioctl some users may pass in
2224 * smaller size (or zero) to simply get the size of the array
2225 * so the user space caller can allocate sufficient memory
2226 * and retry the ioctl again with larger array size sufficient
2227 * to hold all of the snapshot GMT tokens on the second try.
2228 */
2229 if (snapshot_in.snapshot_array_size < GMT_TOKEN_SIZE)
2230 ret_data_len = sizeof(struct smb_snapshot_array);
2231
2232 /*
2233 * We return struct SRV_SNAPSHOT_ARRAY, followed by
2234 * the snapshot array (of 50 byte GMT tokens) each
2235 * representing an available previous version of the data
2236 */
2237 if (ret_data_len > (snapshot_in.snapshot_array_size +
2238 sizeof(struct smb_snapshot_array)))
2239 ret_data_len = snapshot_in.snapshot_array_size +
2240 sizeof(struct smb_snapshot_array);
Steve French834170c2016-09-30 21:14:26 -05002241
2242 if (copy_to_user(ioc_buf, retbuf, ret_data_len))
2243 rc = -EFAULT;
2244 }
2245
2246 kfree(retbuf);
2247 return rc;
2248}
2249
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002250
2251
2252static int
2253smb3_notify(const unsigned int xid, struct file *pfile,
2254 void __user *ioc_buf)
2255{
2256 struct smb3_notify notify;
2257 struct dentry *dentry = pfile->f_path.dentry;
2258 struct inode *inode = file_inode(pfile);
Al Virof6a9bc32021-03-05 17:36:04 -05002259 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002260 struct cifs_open_parms oparms;
2261 struct cifs_fid fid;
2262 struct cifs_tcon *tcon;
Al Virof6a9bc32021-03-05 17:36:04 -05002263 const unsigned char *path;
2264 void *page = alloc_dentry_path();
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002265 __le16 *utf16_path = NULL;
2266 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2267 int rc = 0;
2268
Al Virof6a9bc32021-03-05 17:36:04 -05002269 path = build_path_from_dentry(dentry, page);
2270 if (IS_ERR(path)) {
2271 rc = PTR_ERR(path);
2272 goto notify_exit;
2273 }
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002274
Eugene Korenevskya637f4a2021-04-16 10:35:30 +03002275 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002276 if (utf16_path == NULL) {
2277 rc = -ENOMEM;
2278 goto notify_exit;
2279 }
2280
2281 if (copy_from_user(&notify, ioc_buf, sizeof(struct smb3_notify))) {
2282 rc = -EFAULT;
2283 goto notify_exit;
2284 }
2285
2286 tcon = cifs_sb_master_tcon(cifs_sb);
2287 oparms.tcon = tcon;
Steve French4ef9b4f2020-07-07 18:08:46 -05002288 oparms.desired_access = FILE_READ_ATTRIBUTES | FILE_READ_DATA;
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002289 oparms.disposition = FILE_OPEN;
2290 oparms.create_options = cifs_create_options(cifs_sb, 0);
2291 oparms.fid = &fid;
2292 oparms.reconnect = false;
2293
Aurelien Aptel69dda302020-03-02 17:53:22 +01002294 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL, NULL,
2295 NULL);
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002296 if (rc)
2297 goto notify_exit;
2298
2299 rc = SMB2_change_notify(xid, tcon, fid.persistent_fid, fid.volatile_fid,
2300 notify.watch_tree, notify.completion_filter);
2301
2302 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
2303
2304 cifs_dbg(FYI, "change notify for path %s rc %d\n", path, rc);
2305
2306notify_exit:
Al Virof6a9bc32021-03-05 17:36:04 -05002307 free_dentry_path(page);
Steve Frenchd26c2dd2020-02-06 06:00:14 -06002308 kfree(utf16_path);
2309 return rc;
2310}
2311
Steve French834170c2016-09-30 21:14:26 -05002312static int
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002313smb2_query_dir_first(const unsigned int xid, struct cifs_tcon *tcon,
2314 const char *path, struct cifs_sb_info *cifs_sb,
2315 struct cifs_fid *fid, __u16 search_flags,
2316 struct cifs_search_info *srch_inf)
2317{
2318 __le16 *utf16_path;
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002319 struct smb_rqst rqst[2];
2320 struct kvec rsp_iov[2];
2321 int resp_buftype[2];
2322 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
2323 struct kvec qd_iov[SMB2_QUERY_DIRECTORY_IOV_SIZE];
2324 int rc, flags = 0;
2325 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002326 struct cifs_open_parms oparms;
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002327 struct smb2_query_directory_rsp *qd_rsp = NULL;
2328 struct smb2_create_rsp *op_rsp = NULL;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002329 struct TCP_Server_Info *server = cifs_pick_channel(tcon->ses);
Thiago Rafael Becker6efa9942021-06-15 13:42:56 -03002330 int retry_count = 0;
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002331
2332 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
2333 if (!utf16_path)
2334 return -ENOMEM;
2335
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002336 if (smb3_encryption_required(tcon))
2337 flags |= CIFS_TRANSFORM_REQ;
2338
2339 memset(rqst, 0, sizeof(rqst));
2340 resp_buftype[0] = resp_buftype[1] = CIFS_NO_BUFFER;
2341 memset(rsp_iov, 0, sizeof(rsp_iov));
2342
2343 /* Open */
2344 memset(&open_iov, 0, sizeof(open_iov));
2345 rqst[0].rq_iov = open_iov;
2346 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
2347
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002348 oparms.tcon = tcon;
2349 oparms.desired_access = FILE_READ_ATTRIBUTES | FILE_READ_DATA;
2350 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02002351 oparms.create_options = cifs_create_options(cifs_sb, 0);
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002352 oparms.fid = fid;
Pavel Shilovsky9cbc0b72013-07-09 18:40:58 +04002353 oparms.reconnect = false;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002354
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002355 rc = SMB2_open_init(tcon, server,
2356 &rqst[0], &oplock, &oparms, utf16_path);
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002357 if (rc)
2358 goto qdf_free;
2359 smb2_set_next_command(tcon, &rqst[0]);
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002360
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002361 /* Query directory */
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002362 srch_inf->entries_in_buffer = 0;
Aurelien Aptel05957512018-05-17 16:35:07 +02002363 srch_inf->index_of_last_entry = 2;
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002364
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002365 memset(&qd_iov, 0, sizeof(qd_iov));
2366 rqst[1].rq_iov = qd_iov;
2367 rqst[1].rq_nvec = SMB2_QUERY_DIRECTORY_IOV_SIZE;
2368
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002369 rc = SMB2_query_directory_init(xid, tcon, server,
2370 &rqst[1],
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002371 COMPOUND_FID, COMPOUND_FID,
2372 0, srch_inf->info_level);
2373 if (rc)
2374 goto qdf_free;
2375
2376 smb2_set_related(&rqst[1]);
2377
Thiago Rafael Becker6efa9942021-06-15 13:42:56 -03002378again:
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002379 rc = compound_send_recv(xid, tcon->ses, server,
2380 flags, 2, rqst,
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002381 resp_buftype, rsp_iov);
2382
Thiago Rafael Becker6efa9942021-06-15 13:42:56 -03002383 if (rc == -EAGAIN && retry_count++ < 10)
2384 goto again;
2385
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002386 /* If the open failed there is nothing to do */
2387 op_rsp = (struct smb2_create_rsp *)rsp_iov[0].iov_base;
2388 if (op_rsp == NULL || op_rsp->sync_hdr.Status != STATUS_SUCCESS) {
2389 cifs_dbg(FYI, "query_dir_first: open failed rc=%d\n", rc);
2390 goto qdf_free;
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002391 }
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002392 fid->persistent_fid = op_rsp->PersistentFileId;
2393 fid->volatile_fid = op_rsp->VolatileFileId;
2394
2395 /* Anything else than ENODATA means a genuine error */
2396 if (rc && rc != -ENODATA) {
2397 SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
2398 cifs_dbg(FYI, "query_dir_first: query directory failed rc=%d\n", rc);
2399 trace_smb3_query_dir_err(xid, fid->persistent_fid,
2400 tcon->tid, tcon->ses->Suid, 0, 0, rc);
2401 goto qdf_free;
2402 }
2403
Shyam Prasad N1be1fa42020-03-09 01:35:09 -07002404 atomic_inc(&tcon->num_remote_opens);
2405
Ronnie Sahlberg37478602020-01-08 13:08:06 +10002406 qd_rsp = (struct smb2_query_directory_rsp *)rsp_iov[1].iov_base;
2407 if (qd_rsp->sync_hdr.Status == STATUS_NO_MORE_FILES) {
2408 trace_smb3_query_dir_done(xid, fid->persistent_fid,
2409 tcon->tid, tcon->ses->Suid, 0, 0);
2410 srch_inf->endOfSearch = true;
2411 rc = 0;
2412 goto qdf_free;
2413 }
2414
2415 rc = smb2_parse_query_directory(tcon, &rsp_iov[1], resp_buftype[1],
2416 srch_inf);
2417 if (rc) {
2418 trace_smb3_query_dir_err(xid, fid->persistent_fid, tcon->tid,
2419 tcon->ses->Suid, 0, 0, rc);
2420 goto qdf_free;
2421 }
2422 resp_buftype[1] = CIFS_NO_BUFFER;
2423
2424 trace_smb3_query_dir_done(xid, fid->persistent_fid, tcon->tid,
2425 tcon->ses->Suid, 0, srch_inf->entries_in_buffer);
2426
2427 qdf_free:
2428 kfree(utf16_path);
2429 SMB2_open_free(&rqst[0]);
2430 SMB2_query_directory_free(&rqst[1]);
2431 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
2432 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07002433 return rc;
2434}
2435
2436static int
2437smb2_query_dir_next(const unsigned int xid, struct cifs_tcon *tcon,
2438 struct cifs_fid *fid, __u16 search_flags,
2439 struct cifs_search_info *srch_inf)
2440{
2441 return SMB2_query_directory(xid, tcon, fid->persistent_fid,
2442 fid->volatile_fid, 0, srch_inf);
2443}
2444
2445static int
2446smb2_close_dir(const unsigned int xid, struct cifs_tcon *tcon,
2447 struct cifs_fid *fid)
2448{
2449 return SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
2450}
2451
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002452/*
Christoph Probsta205d502019-05-08 21:36:25 +02002453 * If we negotiate SMB2 protocol and get STATUS_PENDING - update
2454 * the number of credits and return true. Otherwise - return false.
2455 */
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002456static bool
Pavel Shilovsky66265f12019-01-23 17:11:16 -08002457smb2_is_status_pending(char *buf, struct TCP_Server_Info *server)
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002458{
Ronnie Sahlberg49f466b2018-06-01 10:53:06 +10002459 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
Shyam Prasad N6d82c272021-02-03 23:20:46 -08002460 int scredits, in_flight;
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002461
Pavel Shilovsky31473fc2016-10-24 15:33:04 -07002462 if (shdr->Status != STATUS_PENDING)
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002463 return false;
2464
Pavel Shilovsky66265f12019-01-23 17:11:16 -08002465 if (shdr->CreditRequest) {
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002466 spin_lock(&server->req_lock);
Pavel Shilovsky31473fc2016-10-24 15:33:04 -07002467 server->credits += le16_to_cpu(shdr->CreditRequest);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -08002468 scredits = server->credits;
Shyam Prasad N6d82c272021-02-03 23:20:46 -08002469 in_flight = server->in_flight;
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002470 spin_unlock(&server->req_lock);
2471 wake_up(&server->request_q);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -08002472
2473 trace_smb3_add_credits(server->CurrentMid,
Shyam Prasad N6d82c272021-02-03 23:20:46 -08002474 server->conn_id, server->hostname, scredits,
2475 le16_to_cpu(shdr->CreditRequest), in_flight);
Shyam Prasad Ncd7b6992020-11-12 08:56:49 -08002476 cifs_dbg(FYI, "%s: status pending add %u credits total=%d\n",
2477 __func__, le16_to_cpu(shdr->CreditRequest), scredits);
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07002478 }
2479
2480 return true;
2481}
2482
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07002483static bool
2484smb2_is_session_expired(char *buf)
2485{
Ronnie Sahlberg49f466b2018-06-01 10:53:06 +10002486 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07002487
Mark Symsd81243c2018-05-24 09:47:31 +01002488 if (shdr->Status != STATUS_NETWORK_SESSION_EXPIRED &&
2489 shdr->Status != STATUS_USER_SESSION_DELETED)
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07002490 return false;
2491
Steve Frenche68a9322018-07-30 14:23:58 -05002492 trace_smb3_ses_expired(shdr->TreeId, shdr->SessionId,
2493 le16_to_cpu(shdr->Command),
2494 le64_to_cpu(shdr->MessageId));
Mark Symsd81243c2018-05-24 09:47:31 +01002495 cifs_dbg(FYI, "Session expired or deleted\n");
Steve Frenche68a9322018-07-30 14:23:58 -05002496
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07002497 return true;
2498}
2499
Rohith Surabattula8e670f72020-09-18 05:37:28 +00002500static bool
2501smb2_is_status_io_timeout(char *buf)
2502{
2503 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
2504
2505 if (shdr->Status == STATUS_IO_TIMEOUT)
2506 return true;
2507 else
2508 return false;
2509}
2510
Rohith Surabattula9e550b02021-02-16 10:40:45 +00002511static void
2512smb2_is_network_name_deleted(char *buf, struct TCP_Server_Info *server)
2513{
2514 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
2515 struct list_head *tmp, *tmp1;
2516 struct cifs_ses *ses;
2517 struct cifs_tcon *tcon;
2518
Steve Frenchf1a08652021-02-20 18:52:15 -06002519 if (shdr->Status != STATUS_NETWORK_NAME_DELETED)
2520 return;
2521
2522 spin_lock(&cifs_tcp_ses_lock);
2523 list_for_each(tmp, &server->smb_ses_list) {
2524 ses = list_entry(tmp, struct cifs_ses, smb_ses_list);
2525 list_for_each(tmp1, &ses->tcon_list) {
2526 tcon = list_entry(tmp1, struct cifs_tcon, tcon_list);
2527 if (tcon->tid == shdr->TreeId) {
2528 tcon->need_reconnect = true;
2529 spin_unlock(&cifs_tcp_ses_lock);
2530 pr_warn_once("Server share %s deleted.\n",
2531 tcon->treeName);
2532 return;
Rohith Surabattula9e550b02021-02-16 10:40:45 +00002533 }
2534 }
Rohith Surabattula9e550b02021-02-16 10:40:45 +00002535 }
Steve Frenchf1a08652021-02-20 18:52:15 -06002536 spin_unlock(&cifs_tcp_ses_lock);
Rohith Surabattula9e550b02021-02-16 10:40:45 +00002537}
2538
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07002539static int
2540smb2_oplock_response(struct cifs_tcon *tcon, struct cifs_fid *fid,
2541 struct cifsInodeInfo *cinode)
2542{
Pavel Shilovsky0822f512012-09-19 06:22:45 -07002543 if (tcon->ses->server->capabilities & SMB2_GLOBAL_CAP_LEASING)
2544 return SMB2_lease_break(0, tcon, cinode->lease_key,
2545 smb2_get_lease_state(cinode));
2546
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07002547 return SMB2_oplock_break(0, tcon, fid->persistent_fid,
2548 fid->volatile_fid,
Pavel Shilovsky18cceb62013-09-05 13:01:06 +04002549 CIFS_CACHE_READ(cinode) ? 1 : 0);
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07002550}
2551
Ronnie Sahlbergc5a5f382018-09-03 13:33:41 +10002552void
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002553smb2_set_related(struct smb_rqst *rqst)
2554{
2555 struct smb2_sync_hdr *shdr;
2556
2557 shdr = (struct smb2_sync_hdr *)(rqst->rq_iov[0].iov_base);
Ronnie Sahlberg88a92c92019-07-16 10:41:46 +10002558 if (shdr == NULL) {
2559 cifs_dbg(FYI, "shdr NULL in smb2_set_related\n");
2560 return;
2561 }
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002562 shdr->Flags |= SMB2_FLAGS_RELATED_OPERATIONS;
2563}
2564
2565char smb2_padding[7] = {0, 0, 0, 0, 0, 0, 0};
2566
Ronnie Sahlbergc5a5f382018-09-03 13:33:41 +10002567void
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002568smb2_set_next_command(struct cifs_tcon *tcon, struct smb_rqst *rqst)
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002569{
2570 struct smb2_sync_hdr *shdr;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002571 struct cifs_ses *ses = tcon->ses;
2572 struct TCP_Server_Info *server = ses->server;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002573 unsigned long len = smb_rqst_len(server, rqst);
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002574 int i, num_padding;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002575
Ronnie Sahlberg88a92c92019-07-16 10:41:46 +10002576 shdr = (struct smb2_sync_hdr *)(rqst->rq_iov[0].iov_base);
2577 if (shdr == NULL) {
2578 cifs_dbg(FYI, "shdr NULL in smb2_set_next_command\n");
2579 return;
2580 }
2581
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002582 /* SMB headers in a compound are 8 byte aligned. */
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002583
2584 /* No padding needed */
2585 if (!(len & 7))
2586 goto finished;
2587
2588 num_padding = 8 - (len & 7);
2589 if (!smb3_encryption_required(tcon)) {
2590 /*
2591 * If we do not have encryption then we can just add an extra
2592 * iov for the padding.
2593 */
2594 rqst->rq_iov[rqst->rq_nvec].iov_base = smb2_padding;
2595 rqst->rq_iov[rqst->rq_nvec].iov_len = num_padding;
2596 rqst->rq_nvec++;
2597 len += num_padding;
2598 } else {
2599 /*
2600 * We can not add a small padding iov for the encryption case
2601 * because the encryption framework can not handle the padding
2602 * iovs.
2603 * We have to flatten this into a single buffer and add
2604 * the padding to it.
2605 */
2606 for (i = 1; i < rqst->rq_nvec; i++) {
2607 memcpy(rqst->rq_iov[0].iov_base +
2608 rqst->rq_iov[0].iov_len,
2609 rqst->rq_iov[i].iov_base,
2610 rqst->rq_iov[i].iov_len);
2611 rqst->rq_iov[0].iov_len += rqst->rq_iov[i].iov_len;
Ronnie Sahlberg271b9c02018-12-18 17:49:05 -06002612 }
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002613 memset(rqst->rq_iov[0].iov_base + rqst->rq_iov[0].iov_len,
2614 0, num_padding);
2615 rqst->rq_iov[0].iov_len += num_padding;
2616 len += num_padding;
2617 rqst->rq_nvec = 1;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002618 }
2619
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002620 finished:
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002621 shdr->NextCommand = cpu_to_le32(len);
2622}
2623
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002624/*
2625 * Passes the query info response back to the caller on success.
2626 * Caller need to free this with free_rsp_buf().
2627 */
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10002628int
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002629smb2_query_info_compound(const unsigned int xid, struct cifs_tcon *tcon,
2630 __le16 *utf16_path, u32 desired_access,
2631 u32 class, u32 type, u32 output_len,
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10002632 struct kvec *rsp, int *buftype,
2633 struct cifs_sb_info *cifs_sb)
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002634{
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002635 struct cifs_ses *ses = tcon->ses;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002636 struct TCP_Server_Info *server = cifs_pick_channel(ses);
Paulo Alcantara04ad69c2021-03-08 12:00:50 -03002637 int flags = CIFS_CP_CREATE_CLOSE_OP;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002638 struct smb_rqst rqst[3];
2639 int resp_buftype[3];
2640 struct kvec rsp_iov[3];
Ronnie Sahlberg4d8dfaf2018-08-21 11:49:21 +10002641 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002642 struct kvec qi_iov[1];
2643 struct kvec close_iov[1];
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002644 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002645 struct cifs_open_parms oparms;
2646 struct cifs_fid fid;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002647 int rc;
2648
2649 if (smb3_encryption_required(tcon))
2650 flags |= CIFS_TRANSFORM_REQ;
2651
2652 memset(rqst, 0, sizeof(rqst));
Ronnie Sahlbergc5a5f382018-09-03 13:33:41 +10002653 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002654 memset(rsp_iov, 0, sizeof(rsp_iov));
2655
2656 memset(&open_iov, 0, sizeof(open_iov));
2657 rqst[0].rq_iov = open_iov;
Ronnie Sahlberg4d8dfaf2018-08-21 11:49:21 +10002658 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002659
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002660 oparms.tcon = tcon;
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002661 oparms.desired_access = desired_access;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002662 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02002663 oparms.create_options = cifs_create_options(cifs_sb, 0);
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002664 oparms.fid = &fid;
Pavel Shilovsky9cbc0b72013-07-09 18:40:58 +04002665 oparms.reconnect = false;
Pavel Shilovsky064f6042013-07-09 18:20:30 +04002666
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002667 rc = SMB2_open_init(tcon, server,
2668 &rqst[0], &oplock, &oparms, utf16_path);
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002669 if (rc)
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002670 goto qic_exit;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002671 smb2_set_next_command(tcon, &rqst[0]);
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002672
2673 memset(&qi_iov, 0, sizeof(qi_iov));
2674 rqst[1].rq_iov = qi_iov;
2675 rqst[1].rq_nvec = 1;
2676
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002677 rc = SMB2_query_info_init(tcon, server,
2678 &rqst[1], COMPOUND_FID, COMPOUND_FID,
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002679 class, type, 0,
2680 output_len, 0,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05002681 NULL);
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002682 if (rc)
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002683 goto qic_exit;
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10002684 smb2_set_next_command(tcon, &rqst[1]);
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002685 smb2_set_related(&rqst[1]);
2686
2687 memset(&close_iov, 0, sizeof(close_iov));
2688 rqst[2].rq_iov = close_iov;
2689 rqst[2].rq_nvec = 1;
2690
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002691 rc = SMB2_close_init(tcon, server,
2692 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002693 if (rc)
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002694 goto qic_exit;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002695 smb2_set_related(&rqst[2]);
2696
Aurelien Aptel352d96f2020-05-31 12:38:22 -05002697 rc = compound_send_recv(xid, ses, server,
2698 flags, 3, rqst,
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002699 resp_buftype, rsp_iov);
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10002700 if (rc) {
2701 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
Steve French7dcc82c2019-09-11 00:07:36 -05002702 if (rc == -EREMCHG) {
2703 tcon->need_reconnect = true;
Joe Perchesa0a30362020-04-14 22:42:53 -07002704 pr_warn_once("server share %s deleted\n",
2705 tcon->treeName);
Steve French7dcc82c2019-09-11 00:07:36 -05002706 }
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002707 goto qic_exit;
Ronnie Sahlbergf9793b62018-11-27 09:52:04 +10002708 }
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002709 *rsp = rsp_iov[1];
2710 *buftype = resp_buftype[1];
2711
2712 qic_exit:
2713 SMB2_open_free(&rqst[0]);
2714 SMB2_query_info_free(&rqst[1]);
2715 SMB2_close_free(&rqst[2]);
2716 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
2717 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
2718 return rc;
2719}
2720
2721static int
2722smb2_queryfs(const unsigned int xid, struct cifs_tcon *tcon,
Amir Goldstein0f060932020-02-03 21:46:43 +02002723 struct cifs_sb_info *cifs_sb, struct kstatfs *buf)
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002724{
2725 struct smb2_query_info_rsp *rsp;
2726 struct smb2_fs_full_size_info *info = NULL;
2727 __le16 utf16_path = 0; /* Null - open root of share */
2728 struct kvec rsp_iov = {NULL, 0};
2729 int buftype = CIFS_NO_BUFFER;
2730 int rc;
2731
2732
2733 rc = smb2_query_info_compound(xid, tcon, &utf16_path,
2734 FILE_READ_ATTRIBUTES,
2735 FS_FULL_SIZE_INFORMATION,
2736 SMB2_O_INFO_FILESYSTEM,
2737 sizeof(struct smb2_fs_full_size_info),
Steve French87f93d82020-02-04 13:02:59 -06002738 &rsp_iov, &buftype, cifs_sb);
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002739 if (rc)
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002740 goto qfs_exit;
2741
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002742 rsp = (struct smb2_query_info_rsp *)rsp_iov.iov_base;
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002743 buf->f_type = SMB2_MAGIC_NUMBER;
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002744 info = (struct smb2_fs_full_size_info *)(
2745 le16_to_cpu(rsp->OutputBufferOffset) + (char *)rsp);
2746 rc = smb2_validate_iov(le16_to_cpu(rsp->OutputBufferOffset),
2747 le32_to_cpu(rsp->OutputBufferLength),
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002748 &rsp_iov,
Ronnie Sahlberg730928c2018-08-08 15:07:49 +10002749 sizeof(struct smb2_fs_full_size_info));
2750 if (!rc)
2751 smb2_copy_fs_info_to_kstatfs(info, buf);
2752
2753qfs_exit:
Ronnie Sahlberg07d3b2e2018-12-20 22:03:04 -06002754 free_rsp_buf(buftype, rsp_iov.iov_base);
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002755 return rc;
2756}
2757
Steve French2d304212018-06-24 23:28:12 -05002758static int
2759smb311_queryfs(const unsigned int xid, struct cifs_tcon *tcon,
Amir Goldstein0f060932020-02-03 21:46:43 +02002760 struct cifs_sb_info *cifs_sb, struct kstatfs *buf)
Steve French2d304212018-06-24 23:28:12 -05002761{
2762 int rc;
2763 __le16 srch_path = 0; /* Null - open root of share */
2764 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2765 struct cifs_open_parms oparms;
2766 struct cifs_fid fid;
2767
2768 if (!tcon->posix_extensions)
Amir Goldstein0f060932020-02-03 21:46:43 +02002769 return smb2_queryfs(xid, tcon, cifs_sb, buf);
Steve French2d304212018-06-24 23:28:12 -05002770
2771 oparms.tcon = tcon;
2772 oparms.desired_access = FILE_READ_ATTRIBUTES;
2773 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02002774 oparms.create_options = cifs_create_options(cifs_sb, 0);
Steve French2d304212018-06-24 23:28:12 -05002775 oparms.fid = &fid;
2776 oparms.reconnect = false;
2777
Aurelien Aptel69dda302020-03-02 17:53:22 +01002778 rc = SMB2_open(xid, &oparms, &srch_path, &oplock, NULL, NULL,
2779 NULL, NULL);
Steve French2d304212018-06-24 23:28:12 -05002780 if (rc)
2781 return rc;
2782
2783 rc = SMB311_posix_qfs_info(xid, tcon, fid.persistent_fid,
2784 fid.volatile_fid, buf);
2785 buf->f_type = SMB2_MAGIC_NUMBER;
2786 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
2787 return rc;
2788}
Steve French2d304212018-06-24 23:28:12 -05002789
Pavel Shilovsky027e8ee2012-09-19 06:22:43 -07002790static bool
2791smb2_compare_fids(struct cifsFileInfo *ob1, struct cifsFileInfo *ob2)
2792{
2793 return ob1->fid.persistent_fid == ob2->fid.persistent_fid &&
2794 ob1->fid.volatile_fid == ob2->fid.volatile_fid;
2795}
2796
Pavel Shilovskyf7ba7fe2012-09-19 06:22:43 -07002797static int
2798smb2_mand_lock(const unsigned int xid, struct cifsFileInfo *cfile, __u64 offset,
2799 __u64 length, __u32 type, int lock, int unlock, bool wait)
2800{
2801 if (unlock && !lock)
2802 type = SMB2_LOCKFLAG_UNLOCK;
2803 return SMB2_lock(xid, tlink_tcon(cfile->tlink),
2804 cfile->fid.persistent_fid, cfile->fid.volatile_fid,
2805 current->tgid, length, offset, type, wait);
2806}
2807
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -07002808static void
2809smb2_get_lease_key(struct inode *inode, struct cifs_fid *fid)
2810{
2811 memcpy(fid->lease_key, CIFS_I(inode)->lease_key, SMB2_LEASE_KEY_SIZE);
2812}
2813
2814static void
2815smb2_set_lease_key(struct inode *inode, struct cifs_fid *fid)
2816{
2817 memcpy(CIFS_I(inode)->lease_key, fid->lease_key, SMB2_LEASE_KEY_SIZE);
2818}
2819
2820static void
2821smb2_new_lease_key(struct cifs_fid *fid)
2822{
Steve Frenchfa70b872016-09-22 00:39:34 -05002823 generate_random_uuid(fid->lease_key);
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -07002824}
2825
Aurelien Aptel9d496402017-02-13 16:16:49 +01002826static int
2827smb2_get_dfs_refer(const unsigned int xid, struct cifs_ses *ses,
2828 const char *search_name,
2829 struct dfs_info3_param **target_nodes,
2830 unsigned int *num_of_nodes,
2831 const struct nls_table *nls_codepage, int remap)
2832{
2833 int rc;
2834 __le16 *utf16_path = NULL;
2835 int utf16_path_len = 0;
2836 struct cifs_tcon *tcon;
2837 struct fsctl_get_dfs_referral_req *dfs_req = NULL;
2838 struct get_dfs_referral_rsp *dfs_rsp = NULL;
2839 u32 dfs_req_size = 0, dfs_rsp_size = 0;
2840
Christoph Probsta205d502019-05-08 21:36:25 +02002841 cifs_dbg(FYI, "%s: path: %s\n", __func__, search_name);
Aurelien Aptel9d496402017-02-13 16:16:49 +01002842
2843 /*
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002844 * Try to use the IPC tcon, otherwise just use any
Aurelien Aptel9d496402017-02-13 16:16:49 +01002845 */
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002846 tcon = ses->tcon_ipc;
2847 if (tcon == NULL) {
2848 spin_lock(&cifs_tcp_ses_lock);
2849 tcon = list_first_entry_or_null(&ses->tcon_list,
2850 struct cifs_tcon,
2851 tcon_list);
2852 if (tcon)
2853 tcon->tc_count++;
2854 spin_unlock(&cifs_tcp_ses_lock);
2855 }
Aurelien Aptel9d496402017-02-13 16:16:49 +01002856
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002857 if (tcon == NULL) {
Aurelien Aptel9d496402017-02-13 16:16:49 +01002858 cifs_dbg(VFS, "session %p has no tcon available for a dfs referral request\n",
2859 ses);
2860 rc = -ENOTCONN;
2861 goto out;
2862 }
2863
2864 utf16_path = cifs_strndup_to_utf16(search_name, PATH_MAX,
2865 &utf16_path_len,
2866 nls_codepage, remap);
2867 if (!utf16_path) {
2868 rc = -ENOMEM;
2869 goto out;
2870 }
2871
2872 dfs_req_size = sizeof(*dfs_req) + utf16_path_len;
2873 dfs_req = kzalloc(dfs_req_size, GFP_KERNEL);
2874 if (!dfs_req) {
2875 rc = -ENOMEM;
2876 goto out;
2877 }
2878
2879 /* Highest DFS referral version understood */
2880 dfs_req->MaxReferralLevel = DFS_VERSION;
2881
2882 /* Path to resolve in an UTF-16 null-terminated string */
2883 memcpy(dfs_req->RequestFileName, utf16_path, utf16_path_len);
2884
2885 do {
Aurelien Aptel9d496402017-02-13 16:16:49 +01002886 rc = SMB2_ioctl(xid, tcon, NO_FILE_ID, NO_FILE_ID,
2887 FSCTL_DFS_GET_REFERRALS,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002888 true /* is_fsctl */,
Steve French153322f2019-03-28 22:32:49 -05002889 (char *)dfs_req, dfs_req_size, CIFSMaxBufSize,
Aurelien Aptel9d496402017-02-13 16:16:49 +01002890 (char **)&dfs_rsp, &dfs_rsp_size);
Aurelien Aptel9d496402017-02-13 16:16:49 +01002891 } while (rc == -EAGAIN);
2892
2893 if (rc) {
Steve French2564f2f2018-03-21 23:16:36 -05002894 if ((rc != -ENOENT) && (rc != -EOPNOTSUPP))
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10002895 cifs_tcon_dbg(VFS, "ioctl error in %s rc=%d\n", __func__, rc);
Aurelien Aptel9d496402017-02-13 16:16:49 +01002896 goto out;
2897 }
2898
2899 rc = parse_dfs_referrals(dfs_rsp, dfs_rsp_size,
2900 num_of_nodes, target_nodes,
2901 nls_codepage, remap, search_name,
2902 true /* is_unicode */);
2903 if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10002904 cifs_tcon_dbg(VFS, "parse error in %s rc=%d\n", __func__, rc);
Aurelien Aptel9d496402017-02-13 16:16:49 +01002905 goto out;
2906 }
2907
2908 out:
Aurelien Aptel63a83b82018-01-24 13:46:11 +01002909 if (tcon && !tcon->ipc) {
2910 /* ipc tcons are not refcounted */
Aurelien Aptel9d496402017-02-13 16:16:49 +01002911 spin_lock(&cifs_tcp_ses_lock);
2912 tcon->tc_count--;
2913 spin_unlock(&cifs_tcp_ses_lock);
2914 }
2915 kfree(utf16_path);
2916 kfree(dfs_req);
2917 kfree(dfs_rsp);
2918 return rc;
2919}
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10002920
2921static int
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002922parse_reparse_posix(struct reparse_posix_data *symlink_buf,
2923 u32 plen, char **target_path,
2924 struct cifs_sb_info *cifs_sb)
2925{
2926 unsigned int len;
2927
2928 /* See MS-FSCC 2.1.2.6 for the 'NFS' style reparse tags */
2929 len = le16_to_cpu(symlink_buf->ReparseDataLength);
2930
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002931 if (le64_to_cpu(symlink_buf->InodeType) != NFS_SPECFILE_LNK) {
2932 cifs_dbg(VFS, "%lld not a supported symlink type\n",
2933 le64_to_cpu(symlink_buf->InodeType));
2934 return -EOPNOTSUPP;
2935 }
2936
2937 *target_path = cifs_strndup_from_utf16(
2938 symlink_buf->PathBuffer,
2939 len, true, cifs_sb->local_nls);
2940 if (!(*target_path))
2941 return -ENOMEM;
2942
2943 convert_delimiter(*target_path, '/');
2944 cifs_dbg(FYI, "%s: target path: %s\n", __func__, *target_path);
2945
2946 return 0;
2947}
2948
2949static int
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10002950parse_reparse_symlink(struct reparse_symlink_data_buffer *symlink_buf,
2951 u32 plen, char **target_path,
2952 struct cifs_sb_info *cifs_sb)
2953{
2954 unsigned int sub_len;
2955 unsigned int sub_offset;
2956
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002957 /* We handle Symbolic Link reparse tag here. See: MS-FSCC 2.1.2.4 */
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10002958
2959 sub_offset = le16_to_cpu(symlink_buf->SubstituteNameOffset);
2960 sub_len = le16_to_cpu(symlink_buf->SubstituteNameLength);
2961 if (sub_offset + 20 > plen ||
2962 sub_offset + sub_len + 20 > plen) {
2963 cifs_dbg(VFS, "srv returned malformed symlink buffer\n");
2964 return -EIO;
2965 }
2966
2967 *target_path = cifs_strndup_from_utf16(
2968 symlink_buf->PathBuffer + sub_offset,
2969 sub_len, true, cifs_sb->local_nls);
2970 if (!(*target_path))
2971 return -ENOMEM;
2972
2973 convert_delimiter(*target_path, '/');
2974 cifs_dbg(FYI, "%s: target path: %s\n", __func__, *target_path);
2975
2976 return 0;
2977}
2978
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002979static int
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10002980parse_reparse_point(struct reparse_data_buffer *buf,
2981 u32 plen, char **target_path,
2982 struct cifs_sb_info *cifs_sb)
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002983{
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10002984 if (plen < sizeof(struct reparse_data_buffer)) {
Joe Perchesa0a30362020-04-14 22:42:53 -07002985 cifs_dbg(VFS, "reparse buffer is too small. Must be at least 8 bytes but was %d\n",
2986 plen);
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10002987 return -EIO;
2988 }
2989
2990 if (plen < le16_to_cpu(buf->ReparseDataLength) +
2991 sizeof(struct reparse_data_buffer)) {
Joe Perchesa0a30362020-04-14 22:42:53 -07002992 cifs_dbg(VFS, "srv returned invalid reparse buf length: %d\n",
2993 plen);
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10002994 return -EIO;
2995 }
2996
Steve Frenchd5ecebc2019-06-28 02:04:18 -05002997 /* See MS-FSCC 2.1.2 */
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10002998 switch (le32_to_cpu(buf->ReparseTag)) {
2999 case IO_REPARSE_TAG_NFS:
3000 return parse_reparse_posix(
3001 (struct reparse_posix_data *)buf,
Steve Frenchd5ecebc2019-06-28 02:04:18 -05003002 plen, target_path, cifs_sb);
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10003003 case IO_REPARSE_TAG_SYMLINK:
3004 return parse_reparse_symlink(
3005 (struct reparse_symlink_data_buffer *)buf,
3006 plen, target_path, cifs_sb);
3007 default:
Joe Perchesa0a30362020-04-14 22:42:53 -07003008 cifs_dbg(VFS, "srv returned unknown symlink buffer tag:0x%08x\n",
3009 le32_to_cpu(buf->ReparseTag));
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10003010 return -EOPNOTSUPP;
3011 }
Steve Frenchd5ecebc2019-06-28 02:04:18 -05003012}
3013
Pavel Shilovsky78932422016-07-24 10:37:38 +03003014#define SMB2_SYMLINK_STRUCT_SIZE \
3015 (sizeof(struct smb2_err_rsp) - 1 + sizeof(struct smb2_symlink_err_rsp))
3016
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003017static int
3018smb2_query_symlink(const unsigned int xid, struct cifs_tcon *tcon,
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003019 struct cifs_sb_info *cifs_sb, const char *full_path,
3020 char **target_path, bool is_reparse_point)
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003021{
3022 int rc;
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003023 __le16 *utf16_path = NULL;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003024 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3025 struct cifs_open_parms oparms;
3026 struct cifs_fid fid;
Ronnie Sahlberg91cb74f2018-04-13 09:03:19 +10003027 struct kvec err_iov = {NULL, 0};
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003028 struct smb2_err_rsp *err_buf = NULL;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003029 struct smb2_symlink_err_rsp *symlink;
Aurelien Aptel352d96f2020-05-31 12:38:22 -05003030 struct TCP_Server_Info *server = cifs_pick_channel(tcon->ses);
Pavel Shilovsky78932422016-07-24 10:37:38 +03003031 unsigned int sub_len;
3032 unsigned int sub_offset;
3033 unsigned int print_len;
3034 unsigned int print_offset;
Paulo Alcantara04ad69c2021-03-08 12:00:50 -03003035 int flags = CIFS_CP_CREATE_CLOSE_OP;
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003036 struct smb_rqst rqst[3];
3037 int resp_buftype[3];
3038 struct kvec rsp_iov[3];
3039 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
3040 struct kvec io_iov[SMB2_IOCTL_IOV_SIZE];
3041 struct kvec close_iov[1];
3042 struct smb2_create_rsp *create_rsp;
3043 struct smb2_ioctl_rsp *ioctl_rsp;
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10003044 struct reparse_data_buffer *reparse_buf;
Amir Goldstein0f060932020-02-03 21:46:43 +02003045 int create_options = is_reparse_point ? OPEN_REPARSE_POINT : 0;
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003046 u32 plen;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003047
3048 cifs_dbg(FYI, "%s: path: %s\n", __func__, full_path);
3049
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10003050 *target_path = NULL;
3051
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003052 if (smb3_encryption_required(tcon))
3053 flags |= CIFS_TRANSFORM_REQ;
3054
3055 memset(rqst, 0, sizeof(rqst));
3056 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
3057 memset(rsp_iov, 0, sizeof(rsp_iov));
3058
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003059 utf16_path = cifs_convert_path_to_utf16(full_path, cifs_sb);
3060 if (!utf16_path)
3061 return -ENOMEM;
3062
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003063 /* Open */
3064 memset(&open_iov, 0, sizeof(open_iov));
3065 rqst[0].rq_iov = open_iov;
3066 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
3067
3068 memset(&oparms, 0, sizeof(oparms));
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003069 oparms.tcon = tcon;
3070 oparms.desired_access = FILE_READ_ATTRIBUTES;
3071 oparms.disposition = FILE_OPEN;
Amir Goldstein0f060932020-02-03 21:46:43 +02003072 oparms.create_options = cifs_create_options(cifs_sb, create_options);
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003073 oparms.fid = &fid;
3074 oparms.reconnect = false;
3075
Aurelien Aptel352d96f2020-05-31 12:38:22 -05003076 rc = SMB2_open_init(tcon, server,
3077 &rqst[0], &oplock, &oparms, utf16_path);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003078 if (rc)
3079 goto querty_exit;
3080 smb2_set_next_command(tcon, &rqst[0]);
3081
3082
3083 /* IOCTL */
3084 memset(&io_iov, 0, sizeof(io_iov));
3085 rqst[1].rq_iov = io_iov;
3086 rqst[1].rq_nvec = SMB2_IOCTL_IOV_SIZE;
3087
Aurelien Aptel352d96f2020-05-31 12:38:22 -05003088 rc = SMB2_ioctl_init(tcon, server,
3089 &rqst[1], fid.persistent_fid,
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003090 fid.volatile_fid, FSCTL_GET_REPARSE_POINT,
Ronnie Sahlberg731b82b2020-01-08 13:08:07 +10003091 true /* is_fctl */, NULL, 0,
3092 CIFSMaxBufSize -
3093 MAX_SMB2_CREATE_RESPONSE_SIZE -
3094 MAX_SMB2_CLOSE_RESPONSE_SIZE);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003095 if (rc)
3096 goto querty_exit;
3097
3098 smb2_set_next_command(tcon, &rqst[1]);
3099 smb2_set_related(&rqst[1]);
3100
3101
3102 /* Close */
3103 memset(&close_iov, 0, sizeof(close_iov));
3104 rqst[2].rq_iov = close_iov;
3105 rqst[2].rq_nvec = 1;
3106
Aurelien Aptel352d96f2020-05-31 12:38:22 -05003107 rc = SMB2_close_init(tcon, server,
3108 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003109 if (rc)
3110 goto querty_exit;
3111
3112 smb2_set_related(&rqst[2]);
3113
Aurelien Aptel352d96f2020-05-31 12:38:22 -05003114 rc = compound_send_recv(xid, tcon->ses, server,
3115 flags, 3, rqst,
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003116 resp_buftype, rsp_iov);
3117
3118 create_rsp = rsp_iov[0].iov_base;
3119 if (create_rsp && create_rsp->sync_hdr.Status)
3120 err_iov = rsp_iov[0];
3121 ioctl_rsp = rsp_iov[1].iov_base;
3122
3123 /*
3124 * Open was successful and we got an ioctl response.
3125 */
3126 if ((rc == 0) && (is_reparse_point)) {
3127 /* See MS-FSCC 2.3.23 */
3128
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10003129 reparse_buf = (struct reparse_data_buffer *)
3130 ((char *)ioctl_rsp +
3131 le32_to_cpu(ioctl_rsp->OutputOffset));
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003132 plen = le32_to_cpu(ioctl_rsp->OutputCount);
3133
3134 if (plen + le32_to_cpu(ioctl_rsp->OutputOffset) >
3135 rsp_iov[1].iov_len) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10003136 cifs_tcon_dbg(VFS, "srv returned invalid ioctl len: %d\n",
Ronnie Sahlberg5de254d2019-06-27 14:57:02 +10003137 plen);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003138 rc = -EIO;
3139 goto querty_exit;
3140 }
3141
Ronnie Sahlbergf5f111c2019-07-07 07:45:42 +10003142 rc = parse_reparse_point(reparse_buf, plen, target_path,
3143 cifs_sb);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003144 goto querty_exit;
3145 }
3146
Gustavo A. R. Silva0d568cd2018-04-13 10:13:29 -05003147 if (!rc || !err_iov.iov_base) {
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003148 rc = -ENOENT;
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003149 goto querty_exit;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003150 }
Pavel Shilovsky78932422016-07-24 10:37:38 +03003151
Ronnie Sahlberg91cb74f2018-04-13 09:03:19 +10003152 err_buf = err_iov.iov_base;
Pavel Shilovsky78932422016-07-24 10:37:38 +03003153 if (le32_to_cpu(err_buf->ByteCount) < sizeof(struct smb2_symlink_err_rsp) ||
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10003154 err_iov.iov_len < SMB2_SYMLINK_STRUCT_SIZE) {
Ronnie Sahlbergdf070af2019-07-09 18:41:11 +10003155 rc = -EINVAL;
3156 goto querty_exit;
3157 }
3158
3159 symlink = (struct smb2_symlink_err_rsp *)err_buf->ErrorData;
3160 if (le32_to_cpu(symlink->SymLinkErrorTag) != SYMLINK_ERROR_TAG ||
3161 le32_to_cpu(symlink->ReparseTag) != IO_REPARSE_TAG_SYMLINK) {
3162 rc = -EINVAL;
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003163 goto querty_exit;
Pavel Shilovsky78932422016-07-24 10:37:38 +03003164 }
3165
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003166 /* open must fail on symlink - reset rc */
3167 rc = 0;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003168 sub_len = le16_to_cpu(symlink->SubstituteNameLength);
3169 sub_offset = le16_to_cpu(symlink->SubstituteNameOffset);
Pavel Shilovsky78932422016-07-24 10:37:38 +03003170 print_len = le16_to_cpu(symlink->PrintNameLength);
3171 print_offset = le16_to_cpu(symlink->PrintNameOffset);
3172
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10003173 if (err_iov.iov_len < SMB2_SYMLINK_STRUCT_SIZE + sub_offset + sub_len) {
Ronnie Sahlbergdf070af2019-07-09 18:41:11 +10003174 rc = -EINVAL;
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003175 goto querty_exit;
Pavel Shilovsky78932422016-07-24 10:37:38 +03003176 }
3177
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10003178 if (err_iov.iov_len <
3179 SMB2_SYMLINK_STRUCT_SIZE + print_offset + print_len) {
Ronnie Sahlbergdf070af2019-07-09 18:41:11 +10003180 rc = -EINVAL;
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003181 goto querty_exit;
Pavel Shilovsky78932422016-07-24 10:37:38 +03003182 }
3183
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003184 *target_path = cifs_strndup_from_utf16(
3185 (char *)symlink->PathBuffer + sub_offset,
3186 sub_len, true, cifs_sb->local_nls);
3187 if (!(*target_path)) {
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003188 rc = -ENOMEM;
3189 goto querty_exit;
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003190 }
3191 convert_delimiter(*target_path, '/');
3192 cifs_dbg(FYI, "%s: target path: %s\n", __func__, *target_path);
Ronnie Sahlberg9d874c32018-06-08 13:21:18 +10003193
3194 querty_exit:
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003195 cifs_dbg(FYI, "query symlink rc %d\n", rc);
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003196 kfree(utf16_path);
Ronnie Sahlbergebaf5462019-04-10 08:44:46 +10003197 SMB2_open_free(&rqst[0]);
3198 SMB2_ioctl_free(&rqst[1]);
3199 SMB2_close_free(&rqst[2]);
3200 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
3201 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
3202 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04003203 return rc;
3204}
3205
Steve French2e4564b2020-10-22 22:03:14 -05003206int
3207smb2_query_reparse_tag(const unsigned int xid, struct cifs_tcon *tcon,
3208 struct cifs_sb_info *cifs_sb, const char *full_path,
3209 __u32 *tag)
3210{
3211 int rc;
3212 __le16 *utf16_path = NULL;
3213 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3214 struct cifs_open_parms oparms;
3215 struct cifs_fid fid;
Steve French2e4564b2020-10-22 22:03:14 -05003216 struct TCP_Server_Info *server = cifs_pick_channel(tcon->ses);
Paulo Alcantara04ad69c2021-03-08 12:00:50 -03003217 int flags = CIFS_CP_CREATE_CLOSE_OP;
Steve French2e4564b2020-10-22 22:03:14 -05003218 struct smb_rqst rqst[3];
3219 int resp_buftype[3];
3220 struct kvec rsp_iov[3];
3221 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
3222 struct kvec io_iov[SMB2_IOCTL_IOV_SIZE];
3223 struct kvec close_iov[1];
Steve French2e4564b2020-10-22 22:03:14 -05003224 struct smb2_ioctl_rsp *ioctl_rsp;
3225 struct reparse_data_buffer *reparse_buf;
3226 u32 plen;
3227
3228 cifs_dbg(FYI, "%s: path: %s\n", __func__, full_path);
3229
3230 if (smb3_encryption_required(tcon))
3231 flags |= CIFS_TRANSFORM_REQ;
3232
3233 memset(rqst, 0, sizeof(rqst));
3234 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
3235 memset(rsp_iov, 0, sizeof(rsp_iov));
3236
3237 utf16_path = cifs_convert_path_to_utf16(full_path, cifs_sb);
3238 if (!utf16_path)
3239 return -ENOMEM;
3240
3241 /*
3242 * setup smb2open - TODO add optimization to call cifs_get_readable_path
3243 * to see if there is a handle already open that we can use
3244 */
3245 memset(&open_iov, 0, sizeof(open_iov));
3246 rqst[0].rq_iov = open_iov;
3247 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
3248
3249 memset(&oparms, 0, sizeof(oparms));
3250 oparms.tcon = tcon;
3251 oparms.desired_access = FILE_READ_ATTRIBUTES;
3252 oparms.disposition = FILE_OPEN;
3253 oparms.create_options = cifs_create_options(cifs_sb, OPEN_REPARSE_POINT);
3254 oparms.fid = &fid;
3255 oparms.reconnect = false;
3256
3257 rc = SMB2_open_init(tcon, server,
3258 &rqst[0], &oplock, &oparms, utf16_path);
3259 if (rc)
3260 goto query_rp_exit;
3261 smb2_set_next_command(tcon, &rqst[0]);
3262
3263
3264 /* IOCTL */
3265 memset(&io_iov, 0, sizeof(io_iov));
3266 rqst[1].rq_iov = io_iov;
3267 rqst[1].rq_nvec = SMB2_IOCTL_IOV_SIZE;
3268
3269 rc = SMB2_ioctl_init(tcon, server,
Namjae Jeon79631782020-12-03 12:31:36 +09003270 &rqst[1], COMPOUND_FID,
3271 COMPOUND_FID, FSCTL_GET_REPARSE_POINT,
Steve French2e4564b2020-10-22 22:03:14 -05003272 true /* is_fctl */, NULL, 0,
3273 CIFSMaxBufSize -
3274 MAX_SMB2_CREATE_RESPONSE_SIZE -
3275 MAX_SMB2_CLOSE_RESPONSE_SIZE);
3276 if (rc)
3277 goto query_rp_exit;
3278
3279 smb2_set_next_command(tcon, &rqst[1]);
3280 smb2_set_related(&rqst[1]);
3281
3282
3283 /* Close */
3284 memset(&close_iov, 0, sizeof(close_iov));
3285 rqst[2].rq_iov = close_iov;
3286 rqst[2].rq_nvec = 1;
3287
3288 rc = SMB2_close_init(tcon, server,
3289 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
3290 if (rc)
3291 goto query_rp_exit;
3292
3293 smb2_set_related(&rqst[2]);
3294
3295 rc = compound_send_recv(xid, tcon->ses, server,
3296 flags, 3, rqst,
3297 resp_buftype, rsp_iov);
3298
Steve French2e4564b2020-10-22 22:03:14 -05003299 ioctl_rsp = rsp_iov[1].iov_base;
3300
3301 /*
3302 * Open was successful and we got an ioctl response.
3303 */
3304 if (rc == 0) {
3305 /* See MS-FSCC 2.3.23 */
3306
3307 reparse_buf = (struct reparse_data_buffer *)
3308 ((char *)ioctl_rsp +
3309 le32_to_cpu(ioctl_rsp->OutputOffset));
3310 plen = le32_to_cpu(ioctl_rsp->OutputCount);
3311
3312 if (plen + le32_to_cpu(ioctl_rsp->OutputOffset) >
3313 rsp_iov[1].iov_len) {
3314 cifs_tcon_dbg(FYI, "srv returned invalid ioctl len: %d\n",
3315 plen);
3316 rc = -EIO;
3317 goto query_rp_exit;
3318 }
3319 *tag = le32_to_cpu(reparse_buf->ReparseTag);
3320 }
3321
3322 query_rp_exit:
3323 kfree(utf16_path);
3324 SMB2_open_free(&rqst[0]);
3325 SMB2_ioctl_free(&rqst[1]);
3326 SMB2_close_free(&rqst[2]);
3327 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
3328 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
3329 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
3330 return rc;
3331}
3332
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003333static struct cifs_ntsd *
3334get_smb2_acl_by_fid(struct cifs_sb_info *cifs_sb,
Boris Protopopov3970acf2020-12-18 11:30:12 -06003335 const struct cifs_fid *cifsfid, u32 *pacllen, u32 info)
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003336{
3337 struct cifs_ntsd *pntsd = NULL;
3338 unsigned int xid;
3339 int rc = -EOPNOTSUPP;
3340 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3341
3342 if (IS_ERR(tlink))
3343 return ERR_CAST(tlink);
3344
3345 xid = get_xid();
3346 cifs_dbg(FYI, "trying to get acl\n");
3347
3348 rc = SMB2_query_acl(xid, tlink_tcon(tlink), cifsfid->persistent_fid,
Boris Protopopov3970acf2020-12-18 11:30:12 -06003349 cifsfid->volatile_fid, (void **)&pntsd, pacllen,
3350 info);
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003351 free_xid(xid);
3352
3353 cifs_put_tlink(tlink);
3354
3355 cifs_dbg(FYI, "%s: rc = %d ACL len %d\n", __func__, rc, *pacllen);
3356 if (rc)
3357 return ERR_PTR(rc);
3358 return pntsd;
3359
3360}
3361
3362static struct cifs_ntsd *
3363get_smb2_acl_by_path(struct cifs_sb_info *cifs_sb,
Boris Protopopov3970acf2020-12-18 11:30:12 -06003364 const char *path, u32 *pacllen, u32 info)
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003365{
3366 struct cifs_ntsd *pntsd = NULL;
3367 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3368 unsigned int xid;
3369 int rc;
3370 struct cifs_tcon *tcon;
3371 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3372 struct cifs_fid fid;
3373 struct cifs_open_parms oparms;
3374 __le16 *utf16_path;
3375
3376 cifs_dbg(FYI, "get smb3 acl for path %s\n", path);
3377 if (IS_ERR(tlink))
3378 return ERR_CAST(tlink);
3379
3380 tcon = tlink_tcon(tlink);
3381 xid = get_xid();
3382
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003383 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
Steve Frenchcfe89092018-05-19 02:04:55 -05003384 if (!utf16_path) {
3385 rc = -ENOMEM;
3386 free_xid(xid);
3387 return ERR_PTR(rc);
3388 }
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003389
3390 oparms.tcon = tcon;
3391 oparms.desired_access = READ_CONTROL;
3392 oparms.disposition = FILE_OPEN;
Steve French3c3317d2020-10-21 13:12:08 -05003393 /*
3394 * When querying an ACL, even if the file is a symlink we want to open
3395 * the source not the target, and so the protocol requires that the
3396 * client specify this flag when opening a reparse point
3397 */
3398 oparms.create_options = cifs_create_options(cifs_sb, 0) | OPEN_REPARSE_POINT;
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003399 oparms.fid = &fid;
3400 oparms.reconnect = false;
3401
Boris Protopopov3970acf2020-12-18 11:30:12 -06003402 if (info & SACL_SECINFO)
3403 oparms.desired_access |= SYSTEM_SECURITY;
3404
Aurelien Aptel69dda302020-03-02 17:53:22 +01003405 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL, NULL,
3406 NULL);
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003407 kfree(utf16_path);
3408 if (!rc) {
3409 rc = SMB2_query_acl(xid, tlink_tcon(tlink), fid.persistent_fid,
Boris Protopopov3970acf2020-12-18 11:30:12 -06003410 fid.volatile_fid, (void **)&pntsd, pacllen,
3411 info);
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003412 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
3413 }
3414
3415 cifs_put_tlink(tlink);
3416 free_xid(xid);
3417
3418 cifs_dbg(FYI, "%s: rc = %d ACL len %d\n", __func__, rc, *pacllen);
3419 if (rc)
3420 return ERR_PTR(rc);
3421 return pntsd;
3422}
3423
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003424static int
3425set_smb2_acl(struct cifs_ntsd *pnntsd, __u32 acllen,
3426 struct inode *inode, const char *path, int aclflag)
3427{
3428 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3429 unsigned int xid;
3430 int rc, access_flags = 0;
3431 struct cifs_tcon *tcon;
3432 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3433 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3434 struct cifs_fid fid;
3435 struct cifs_open_parms oparms;
3436 __le16 *utf16_path;
3437
3438 cifs_dbg(FYI, "set smb3 acl for path %s\n", path);
3439 if (IS_ERR(tlink))
3440 return PTR_ERR(tlink);
3441
3442 tcon = tlink_tcon(tlink);
3443 xid = get_xid();
3444
Boris Protopopov3970acf2020-12-18 11:30:12 -06003445 if (aclflag & CIFS_ACL_OWNER || aclflag & CIFS_ACL_GROUP)
3446 access_flags |= WRITE_OWNER;
3447 if (aclflag & CIFS_ACL_SACL)
3448 access_flags |= SYSTEM_SECURITY;
3449 if (aclflag & CIFS_ACL_DACL)
3450 access_flags |= WRITE_DAC;
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003451
3452 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
Steve Frenchcfe89092018-05-19 02:04:55 -05003453 if (!utf16_path) {
3454 rc = -ENOMEM;
3455 free_xid(xid);
3456 return rc;
3457 }
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003458
3459 oparms.tcon = tcon;
3460 oparms.desired_access = access_flags;
Amir Goldstein0f060932020-02-03 21:46:43 +02003461 oparms.create_options = cifs_create_options(cifs_sb, 0);
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003462 oparms.disposition = FILE_OPEN;
3463 oparms.path = path;
3464 oparms.fid = &fid;
3465 oparms.reconnect = false;
3466
Aurelien Aptel69dda302020-03-02 17:53:22 +01003467 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL,
3468 NULL, NULL);
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003469 kfree(utf16_path);
3470 if (!rc) {
3471 rc = SMB2_set_acl(xid, tlink_tcon(tlink), fid.persistent_fid,
3472 fid.volatile_fid, pnntsd, acllen, aclflag);
3473 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
3474 }
3475
3476 cifs_put_tlink(tlink);
3477 free_xid(xid);
3478 return rc;
3479}
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05003480
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003481/* Retrieve an ACL from the server */
3482static struct cifs_ntsd *
3483get_smb2_acl(struct cifs_sb_info *cifs_sb,
Boris Protopopov3970acf2020-12-18 11:30:12 -06003484 struct inode *inode, const char *path,
3485 u32 *pacllen, u32 info)
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003486{
3487 struct cifs_ntsd *pntsd = NULL;
3488 struct cifsFileInfo *open_file = NULL;
3489
Boris Protopopov9541b812020-12-17 20:58:08 +00003490 if (inode && !(info & SACL_SECINFO))
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003491 open_file = find_readable_file(CIFS_I(inode), true);
Boris Protopopov9541b812020-12-17 20:58:08 +00003492 if (!open_file || (info & SACL_SECINFO))
Boris Protopopov3970acf2020-12-18 11:30:12 -06003493 return get_smb2_acl_by_path(cifs_sb, path, pacllen, info);
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003494
Boris Protopopov3970acf2020-12-18 11:30:12 -06003495 pntsd = get_smb2_acl_by_fid(cifs_sb, &open_file->fid, pacllen, info);
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003496 cifsFileInfo_put(open_file);
3497 return pntsd;
3498}
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05003499
Steve French30175622014-08-17 18:16:40 -05003500static long smb3_zero_range(struct file *file, struct cifs_tcon *tcon,
3501 loff_t offset, loff_t len, bool keep_size)
3502{
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003503 struct cifs_ses *ses = tcon->ses;
Steve French30175622014-08-17 18:16:40 -05003504 struct inode *inode;
3505 struct cifsInodeInfo *cifsi;
3506 struct cifsFileInfo *cfile = file->private_data;
3507 struct file_zero_data_information fsctl_buf;
3508 long rc;
3509 unsigned int xid;
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003510 __le64 eof;
Steve French30175622014-08-17 18:16:40 -05003511
3512 xid = get_xid();
3513
David Howells2b0143b2015-03-17 22:25:59 +00003514 inode = d_inode(cfile->dentry);
Steve French30175622014-08-17 18:16:40 -05003515 cifsi = CIFS_I(inode);
3516
Christoph Probsta205d502019-05-08 21:36:25 +02003517 trace_smb3_zero_enter(xid, cfile->fid.persistent_fid, tcon->tid,
Steve French779ede02019-03-13 01:41:49 -05003518 ses->Suid, offset, len);
3519
Zhang Xiaoxu6b690402020-06-23 07:31:54 -04003520 /*
3521 * We zero the range through ioctl, so we need remove the page caches
3522 * first, otherwise the data may be inconsistent with the server.
3523 */
3524 truncate_pagecache_range(inode, offset, offset + len - 1);
Steve French779ede02019-03-13 01:41:49 -05003525
Steve French30175622014-08-17 18:16:40 -05003526 /* if file not oplocked can't be sure whether asking to extend size */
3527 if (!CIFS_CACHE_READ(cifsi))
Steve Frenchcfe89092018-05-19 02:04:55 -05003528 if (keep_size == false) {
3529 rc = -EOPNOTSUPP;
Steve French779ede02019-03-13 01:41:49 -05003530 trace_smb3_zero_err(xid, cfile->fid.persistent_fid,
3531 tcon->tid, ses->Suid, offset, len, rc);
Steve Frenchcfe89092018-05-19 02:04:55 -05003532 free_xid(xid);
3533 return rc;
3534 }
Steve French30175622014-08-17 18:16:40 -05003535
Steve Frenchd1c35af2019-05-09 00:09:37 -05003536 cifs_dbg(FYI, "Offset %lld len %lld\n", offset, len);
Steve French30175622014-08-17 18:16:40 -05003537
3538 fsctl_buf.FileOffset = cpu_to_le64(offset);
3539 fsctl_buf.BeyondFinalZero = cpu_to_le64(offset + len);
3540
Ronnie Sahlbergc4250142019-05-02 15:52:57 +10003541 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3542 cfile->fid.volatile_fid, FSCTL_SET_ZERO_DATA, true,
3543 (char *)&fsctl_buf,
3544 sizeof(struct file_zero_data_information),
3545 0, NULL, NULL);
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003546 if (rc)
3547 goto zero_range_exit;
3548
3549 /*
3550 * do we also need to change the size of the file?
3551 */
3552 if (keep_size == false && i_size_read(inode) < offset + len) {
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003553 eof = cpu_to_le64(offset + len);
Ronnie Sahlbergc4250142019-05-02 15:52:57 +10003554 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3555 cfile->fid.volatile_fid, cfile->pid, &eof);
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003556 }
3557
Ronnie Sahlberg72c419d2019-03-13 14:37:49 +10003558 zero_range_exit:
Steve French30175622014-08-17 18:16:40 -05003559 free_xid(xid);
Steve French779ede02019-03-13 01:41:49 -05003560 if (rc)
3561 trace_smb3_zero_err(xid, cfile->fid.persistent_fid, tcon->tid,
3562 ses->Suid, offset, len, rc);
3563 else
3564 trace_smb3_zero_done(xid, cfile->fid.persistent_fid, tcon->tid,
3565 ses->Suid, offset, len);
Steve French30175622014-08-17 18:16:40 -05003566 return rc;
3567}
3568
Steve French31742c52014-08-17 08:38:47 -05003569static long smb3_punch_hole(struct file *file, struct cifs_tcon *tcon,
3570 loff_t offset, loff_t len)
3571{
3572 struct inode *inode;
Steve French31742c52014-08-17 08:38:47 -05003573 struct cifsFileInfo *cfile = file->private_data;
3574 struct file_zero_data_information fsctl_buf;
3575 long rc;
3576 unsigned int xid;
3577 __u8 set_sparse = 1;
3578
3579 xid = get_xid();
3580
David Howells2b0143b2015-03-17 22:25:59 +00003581 inode = d_inode(cfile->dentry);
Steve French31742c52014-08-17 08:38:47 -05003582
3583 /* Need to make file sparse, if not already, before freeing range. */
3584 /* Consider adding equivalent for compressed since it could also work */
Steve Frenchcfe89092018-05-19 02:04:55 -05003585 if (!smb2_set_sparse(xid, tcon, cfile, inode, set_sparse)) {
3586 rc = -EOPNOTSUPP;
3587 free_xid(xid);
3588 return rc;
3589 }
Steve French31742c52014-08-17 08:38:47 -05003590
Zhang Xiaoxuacc91c22020-06-23 07:31:53 -04003591 /*
3592 * We implement the punch hole through ioctl, so we need remove the page
3593 * caches first, otherwise the data may be inconsistent with the server.
3594 */
3595 truncate_pagecache_range(inode, offset, offset + len - 1);
3596
Christoph Probsta205d502019-05-08 21:36:25 +02003597 cifs_dbg(FYI, "Offset %lld len %lld\n", offset, len);
Steve French31742c52014-08-17 08:38:47 -05003598
3599 fsctl_buf.FileOffset = cpu_to_le64(offset);
3600 fsctl_buf.BeyondFinalZero = cpu_to_le64(offset + len);
3601
3602 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3603 cfile->fid.volatile_fid, FSCTL_SET_ZERO_DATA,
Aurelien Aptel63a83b82018-01-24 13:46:11 +01003604 true /* is_fctl */, (char *)&fsctl_buf,
Steve French153322f2019-03-28 22:32:49 -05003605 sizeof(struct file_zero_data_information),
3606 CIFSMaxBufSize, NULL, NULL);
Steve French31742c52014-08-17 08:38:47 -05003607 free_xid(xid);
3608 return rc;
3609}
3610
Ronnie Sahlberg966a3cb2021-06-03 15:31:01 +10003611static int smb3_simple_fallocate_write_range(unsigned int xid,
3612 struct cifs_tcon *tcon,
3613 struct cifsFileInfo *cfile,
3614 loff_t off, loff_t len,
3615 char *buf)
3616{
3617 struct cifs_io_parms io_parms = {0};
3618 int nbytes;
3619 struct kvec iov[2];
3620
3621 io_parms.netfid = cfile->fid.netfid;
3622 io_parms.pid = current->tgid;
3623 io_parms.tcon = tcon;
3624 io_parms.persistent_fid = cfile->fid.persistent_fid;
3625 io_parms.volatile_fid = cfile->fid.volatile_fid;
3626 io_parms.offset = off;
3627 io_parms.length = len;
3628
3629 /* iov[0] is reserved for smb header */
3630 iov[1].iov_base = buf;
3631 iov[1].iov_len = io_parms.length;
3632 return SMB2_write(xid, &io_parms, &nbytes, iov, 1);
3633}
3634
3635static int smb3_simple_fallocate_range(unsigned int xid,
3636 struct cifs_tcon *tcon,
3637 struct cifsFileInfo *cfile,
3638 loff_t off, loff_t len)
3639{
3640 struct file_allocated_range_buffer in_data, *out_data = NULL, *tmp_data;
3641 u32 out_data_len;
3642 char *buf = NULL;
3643 loff_t l;
3644 int rc;
3645
3646 in_data.file_offset = cpu_to_le64(off);
3647 in_data.length = cpu_to_le64(len);
3648 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3649 cfile->fid.volatile_fid,
3650 FSCTL_QUERY_ALLOCATED_RANGES, true,
3651 (char *)&in_data, sizeof(in_data),
3652 1024 * sizeof(struct file_allocated_range_buffer),
3653 (char **)&out_data, &out_data_len);
3654 if (rc)
3655 goto out;
3656 /*
3657 * It is already all allocated
3658 */
3659 if (out_data_len == 0)
3660 goto out;
3661
3662 buf = kzalloc(1024 * 1024, GFP_KERNEL);
3663 if (buf == NULL) {
3664 rc = -ENOMEM;
3665 goto out;
3666 }
3667
3668 tmp_data = out_data;
3669 while (len) {
3670 /*
3671 * The rest of the region is unmapped so write it all.
3672 */
3673 if (out_data_len == 0) {
3674 rc = smb3_simple_fallocate_write_range(xid, tcon,
3675 cfile, off, len, buf);
3676 goto out;
3677 }
3678
3679 if (out_data_len < sizeof(struct file_allocated_range_buffer)) {
3680 rc = -EINVAL;
3681 goto out;
3682 }
3683
3684 if (off < le64_to_cpu(tmp_data->file_offset)) {
3685 /*
3686 * We are at a hole. Write until the end of the region
3687 * or until the next allocated data,
3688 * whichever comes next.
3689 */
3690 l = le64_to_cpu(tmp_data->file_offset) - off;
3691 if (len < l)
3692 l = len;
3693 rc = smb3_simple_fallocate_write_range(xid, tcon,
3694 cfile, off, l, buf);
3695 if (rc)
3696 goto out;
3697 off = off + l;
3698 len = len - l;
3699 if (len == 0)
3700 goto out;
3701 }
3702 /*
3703 * We are at a section of allocated data, just skip forward
3704 * until the end of the data or the end of the region
3705 * we are supposed to fallocate, whichever comes first.
3706 */
3707 l = le64_to_cpu(tmp_data->length);
3708 if (len < l)
3709 l = len;
3710 off += l;
3711 len -= l;
3712
3713 tmp_data = &tmp_data[1];
3714 out_data_len -= sizeof(struct file_allocated_range_buffer);
3715 }
3716
3717 out:
3718 kfree(out_data);
3719 kfree(buf);
3720 return rc;
3721}
3722
3723
Steve French9ccf3212014-10-18 17:01:15 -05003724static long smb3_simple_falloc(struct file *file, struct cifs_tcon *tcon,
3725 loff_t off, loff_t len, bool keep_size)
3726{
3727 struct inode *inode;
3728 struct cifsInodeInfo *cifsi;
3729 struct cifsFileInfo *cfile = file->private_data;
3730 long rc = -EOPNOTSUPP;
3731 unsigned int xid;
Ronnie Sahlbergf1699472019-03-15 00:08:48 +10003732 __le64 eof;
Steve French9ccf3212014-10-18 17:01:15 -05003733
3734 xid = get_xid();
3735
David Howells2b0143b2015-03-17 22:25:59 +00003736 inode = d_inode(cfile->dentry);
Steve French9ccf3212014-10-18 17:01:15 -05003737 cifsi = CIFS_I(inode);
3738
Steve French779ede02019-03-13 01:41:49 -05003739 trace_smb3_falloc_enter(xid, cfile->fid.persistent_fid, tcon->tid,
3740 tcon->ses->Suid, off, len);
Steve French9ccf3212014-10-18 17:01:15 -05003741 /* if file not oplocked can't be sure whether asking to extend size */
3742 if (!CIFS_CACHE_READ(cifsi))
Steve Frenchcfe89092018-05-19 02:04:55 -05003743 if (keep_size == false) {
Steve French779ede02019-03-13 01:41:49 -05003744 trace_smb3_falloc_err(xid, cfile->fid.persistent_fid,
3745 tcon->tid, tcon->ses->Suid, off, len, rc);
Steve Frenchcfe89092018-05-19 02:04:55 -05003746 free_xid(xid);
3747 return rc;
3748 }
Steve French9ccf3212014-10-18 17:01:15 -05003749
3750 /*
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003751 * Extending the file
3752 */
3753 if ((keep_size == false) && i_size_read(inode) < off + len) {
Murphy Zhouef4a6322020-03-18 20:43:38 +08003754 rc = inode_newsize_ok(inode, off + len);
3755 if (rc)
3756 goto out;
3757
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003758 if ((cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) == 0)
3759 smb2_set_sparse(xid, tcon, cfile, inode, false);
3760
3761 eof = cpu_to_le64(off + len);
3762 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3763 cfile->fid.volatile_fid, cfile->pid, &eof);
3764 if (rc == 0) {
3765 cifsi->server_eof = off + len;
3766 cifs_setsize(inode, off + len);
3767 cifs_truncate_page(inode->i_mapping, inode->i_size);
3768 truncate_setsize(inode, off + len);
3769 }
3770 goto out;
3771 }
3772
3773 /*
Steve French9ccf3212014-10-18 17:01:15 -05003774 * Files are non-sparse by default so falloc may be a no-op
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003775 * Must check if file sparse. If not sparse, and since we are not
3776 * extending then no need to do anything since file already allocated
Steve French9ccf3212014-10-18 17:01:15 -05003777 */
3778 if ((cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) == 0) {
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003779 rc = 0;
3780 goto out;
Steve French9ccf3212014-10-18 17:01:15 -05003781 }
3782
3783 if ((keep_size == true) || (i_size_read(inode) >= off + len)) {
3784 /*
Ronnie Sahlberg966a3cb2021-06-03 15:31:01 +10003785 * At this point, we are trying to fallocate an internal
3786 * regions of a sparse file. Since smb2 does not have a
3787 * fallocate command we have two otions on how to emulate this.
3788 * We can either turn the entire file to become non-sparse
3789 * which we only do if the fallocate is for virtually
3790 * the whole file, or we can overwrite the region with zeroes
3791 * using SMB2_write, which could be prohibitevly expensive
3792 * if len is large.
3793 */
3794 /*
3795 * We are only trying to fallocate a small region so
3796 * just write it with zero.
3797 */
3798 if (len <= 1024 * 1024) {
3799 rc = smb3_simple_fallocate_range(xid, tcon, cfile,
3800 off, len);
3801 goto out;
3802 }
3803
3804 /*
Steve French9ccf3212014-10-18 17:01:15 -05003805 * Check if falloc starts within first few pages of file
3806 * and ends within a few pages of the end of file to
3807 * ensure that most of file is being forced to be
3808 * fallocated now. If so then setting whole file sparse
3809 * ie potentially making a few extra pages at the beginning
3810 * or end of the file non-sparse via set_sparse is harmless.
3811 */
Steve Frenchcfe89092018-05-19 02:04:55 -05003812 if ((off > 8192) || (off + len + 8192 < i_size_read(inode))) {
3813 rc = -EOPNOTSUPP;
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003814 goto out;
Ronnie Sahlbergf1699472019-03-15 00:08:48 +10003815 }
Steve French9ccf3212014-10-18 17:01:15 -05003816 }
Steve French9ccf3212014-10-18 17:01:15 -05003817
Ronnie Sahlberg8bd0d702020-01-17 11:45:02 +10003818 smb2_set_sparse(xid, tcon, cfile, inode, false);
3819 rc = 0;
3820
3821out:
Steve French779ede02019-03-13 01:41:49 -05003822 if (rc)
3823 trace_smb3_falloc_err(xid, cfile->fid.persistent_fid, tcon->tid,
3824 tcon->ses->Suid, off, len, rc);
3825 else
3826 trace_smb3_falloc_done(xid, cfile->fid.persistent_fid, tcon->tid,
3827 tcon->ses->Suid, off, len);
Steve French9ccf3212014-10-18 17:01:15 -05003828
3829 free_xid(xid);
3830 return rc;
3831}
3832
Ronnie Sahlberg5476b5d2021-03-27 05:52:29 +10003833static long smb3_collapse_range(struct file *file, struct cifs_tcon *tcon,
3834 loff_t off, loff_t len)
3835{
3836 int rc;
3837 unsigned int xid;
3838 struct cifsFileInfo *cfile = file->private_data;
3839 __le64 eof;
3840
3841 xid = get_xid();
3842
3843 if (off >= i_size_read(file->f_inode) ||
3844 off + len >= i_size_read(file->f_inode)) {
3845 rc = -EINVAL;
3846 goto out;
3847 }
3848
3849 rc = smb2_copychunk_range(xid, cfile, cfile, off + len,
3850 i_size_read(file->f_inode) - off - len, off);
3851 if (rc < 0)
3852 goto out;
3853
3854 eof = cpu_to_le64(i_size_read(file->f_inode) - len);
3855 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3856 cfile->fid.volatile_fid, cfile->pid, &eof);
3857 if (rc < 0)
3858 goto out;
3859
3860 rc = 0;
3861 out:
3862 free_xid(xid);
3863 return rc;
3864}
3865
Ronnie Sahlberg7fe6fe92021-03-27 06:31:30 +10003866static long smb3_insert_range(struct file *file, struct cifs_tcon *tcon,
3867 loff_t off, loff_t len)
3868{
3869 int rc;
3870 unsigned int xid;
3871 struct cifsFileInfo *cfile = file->private_data;
3872 __le64 eof;
3873 __u64 count;
3874
3875 xid = get_xid();
3876
3877 if (off >= i_size_read(file->f_inode)) {
3878 rc = -EINVAL;
3879 goto out;
3880 }
3881
3882 count = i_size_read(file->f_inode) - off;
3883 eof = cpu_to_le64(i_size_read(file->f_inode) + len);
3884
3885 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3886 cfile->fid.volatile_fid, cfile->pid, &eof);
3887 if (rc < 0)
3888 goto out;
3889
3890 rc = smb2_copychunk_range(xid, cfile, cfile, off, count, off + len);
3891 if (rc < 0)
3892 goto out;
3893
3894 rc = smb3_zero_range(file, tcon, off, len, 1);
3895 if (rc < 0)
3896 goto out;
3897
3898 rc = 0;
3899 out:
3900 free_xid(xid);
3901 return rc;
3902}
3903
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10003904static loff_t smb3_llseek(struct file *file, struct cifs_tcon *tcon, loff_t offset, int whence)
3905{
3906 struct cifsFileInfo *wrcfile, *cfile = file->private_data;
3907 struct cifsInodeInfo *cifsi;
3908 struct inode *inode;
3909 int rc = 0;
3910 struct file_allocated_range_buffer in_data, *out_data = NULL;
3911 u32 out_data_len;
3912 unsigned int xid;
3913
3914 if (whence != SEEK_HOLE && whence != SEEK_DATA)
3915 return generic_file_llseek(file, offset, whence);
3916
3917 inode = d_inode(cfile->dentry);
3918 cifsi = CIFS_I(inode);
3919
3920 if (offset < 0 || offset >= i_size_read(inode))
3921 return -ENXIO;
3922
3923 xid = get_xid();
3924 /*
3925 * We need to be sure that all dirty pages are written as they
3926 * might fill holes on the server.
3927 * Note that we also MUST flush any written pages since at least
3928 * some servers (Windows2016) will not reflect recent writes in
3929 * QUERY_ALLOCATED_RANGES until SMB2_flush is called.
3930 */
Aurelien Aptel86f740f2020-02-21 11:19:06 +01003931 wrcfile = find_writable_file(cifsi, FIND_WR_ANY);
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10003932 if (wrcfile) {
3933 filemap_write_and_wait(inode->i_mapping);
3934 smb2_flush_file(xid, tcon, &wrcfile->fid);
3935 cifsFileInfo_put(wrcfile);
3936 }
3937
3938 if (!(cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE)) {
3939 if (whence == SEEK_HOLE)
3940 offset = i_size_read(inode);
3941 goto lseek_exit;
3942 }
3943
3944 in_data.file_offset = cpu_to_le64(offset);
3945 in_data.length = cpu_to_le64(i_size_read(inode));
3946
3947 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3948 cfile->fid.volatile_fid,
3949 FSCTL_QUERY_ALLOCATED_RANGES, true,
3950 (char *)&in_data, sizeof(in_data),
3951 sizeof(struct file_allocated_range_buffer),
3952 (char **)&out_data, &out_data_len);
3953 if (rc == -E2BIG)
3954 rc = 0;
3955 if (rc)
3956 goto lseek_exit;
3957
3958 if (whence == SEEK_HOLE && out_data_len == 0)
3959 goto lseek_exit;
3960
3961 if (whence == SEEK_DATA && out_data_len == 0) {
3962 rc = -ENXIO;
3963 goto lseek_exit;
3964 }
3965
3966 if (out_data_len < sizeof(struct file_allocated_range_buffer)) {
3967 rc = -EINVAL;
3968 goto lseek_exit;
3969 }
3970 if (whence == SEEK_DATA) {
3971 offset = le64_to_cpu(out_data->file_offset);
3972 goto lseek_exit;
3973 }
3974 if (offset < le64_to_cpu(out_data->file_offset))
3975 goto lseek_exit;
3976
3977 offset = le64_to_cpu(out_data->file_offset) + le64_to_cpu(out_data->length);
3978
3979 lseek_exit:
3980 free_xid(xid);
3981 kfree(out_data);
3982 if (!rc)
3983 return vfs_setpos(file, offset, inode->i_sb->s_maxbytes);
3984 else
3985 return rc;
3986}
3987
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10003988static int smb3_fiemap(struct cifs_tcon *tcon,
3989 struct cifsFileInfo *cfile,
3990 struct fiemap_extent_info *fei, u64 start, u64 len)
3991{
3992 unsigned int xid;
3993 struct file_allocated_range_buffer in_data, *out_data;
3994 u32 out_data_len;
3995 int i, num, rc, flags, last_blob;
3996 u64 next;
3997
Christoph Hellwig45dd0522020-05-23 09:30:14 +02003998 rc = fiemap_prep(d_inode(cfile->dentry), fei, start, &len, 0);
Christoph Hellwigcddf8a22020-05-23 09:30:13 +02003999 if (rc)
4000 return rc;
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10004001
4002 xid = get_xid();
4003 again:
4004 in_data.file_offset = cpu_to_le64(start);
4005 in_data.length = cpu_to_le64(len);
4006
4007 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
4008 cfile->fid.volatile_fid,
4009 FSCTL_QUERY_ALLOCATED_RANGES, true,
4010 (char *)&in_data, sizeof(in_data),
4011 1024 * sizeof(struct file_allocated_range_buffer),
4012 (char **)&out_data, &out_data_len);
4013 if (rc == -E2BIG) {
4014 last_blob = 0;
4015 rc = 0;
4016 } else
4017 last_blob = 1;
4018 if (rc)
4019 goto out;
4020
Murphy Zhou979a2662020-03-14 11:38:31 +08004021 if (out_data_len && out_data_len < sizeof(struct file_allocated_range_buffer)) {
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10004022 rc = -EINVAL;
4023 goto out;
4024 }
4025 if (out_data_len % sizeof(struct file_allocated_range_buffer)) {
4026 rc = -EINVAL;
4027 goto out;
4028 }
4029
4030 num = out_data_len / sizeof(struct file_allocated_range_buffer);
4031 for (i = 0; i < num; i++) {
4032 flags = 0;
4033 if (i == num - 1 && last_blob)
4034 flags |= FIEMAP_EXTENT_LAST;
4035
4036 rc = fiemap_fill_next_extent(fei,
4037 le64_to_cpu(out_data[i].file_offset),
4038 le64_to_cpu(out_data[i].file_offset),
4039 le64_to_cpu(out_data[i].length),
4040 flags);
4041 if (rc < 0)
4042 goto out;
4043 if (rc == 1) {
4044 rc = 0;
4045 goto out;
4046 }
4047 }
4048
4049 if (!last_blob) {
4050 next = le64_to_cpu(out_data[num - 1].file_offset) +
4051 le64_to_cpu(out_data[num - 1].length);
4052 len = len - (next - start);
4053 start = next;
4054 goto again;
4055 }
4056
4057 out:
4058 free_xid(xid);
4059 kfree(out_data);
4060 return rc;
4061}
Steve French9ccf3212014-10-18 17:01:15 -05004062
Steve French31742c52014-08-17 08:38:47 -05004063static long smb3_fallocate(struct file *file, struct cifs_tcon *tcon, int mode,
4064 loff_t off, loff_t len)
4065{
4066 /* KEEP_SIZE already checked for by do_fallocate */
4067 if (mode & FALLOC_FL_PUNCH_HOLE)
4068 return smb3_punch_hole(file, tcon, off, len);
Steve French30175622014-08-17 18:16:40 -05004069 else if (mode & FALLOC_FL_ZERO_RANGE) {
4070 if (mode & FALLOC_FL_KEEP_SIZE)
4071 return smb3_zero_range(file, tcon, off, len, true);
4072 return smb3_zero_range(file, tcon, off, len, false);
Steve French9ccf3212014-10-18 17:01:15 -05004073 } else if (mode == FALLOC_FL_KEEP_SIZE)
4074 return smb3_simple_falloc(file, tcon, off, len, true);
Ronnie Sahlberg5476b5d2021-03-27 05:52:29 +10004075 else if (mode == FALLOC_FL_COLLAPSE_RANGE)
4076 return smb3_collapse_range(file, tcon, off, len);
Ronnie Sahlberg7fe6fe92021-03-27 06:31:30 +10004077 else if (mode == FALLOC_FL_INSERT_RANGE)
4078 return smb3_insert_range(file, tcon, off, len);
Steve French9ccf3212014-10-18 17:01:15 -05004079 else if (mode == 0)
4080 return smb3_simple_falloc(file, tcon, off, len, false);
Steve French31742c52014-08-17 08:38:47 -05004081
4082 return -EOPNOTSUPP;
4083}
4084
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004085static void
Sachin Prabhuc11f1df2014-03-11 16:11:47 +00004086smb2_downgrade_oplock(struct TCP_Server_Info *server,
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07004087 struct cifsInodeInfo *cinode, __u32 oplock,
4088 unsigned int epoch, bool *purge_cache)
Sachin Prabhuc11f1df2014-03-11 16:11:47 +00004089{
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07004090 server->ops->set_oplock_level(cinode, oplock, 0, NULL);
Sachin Prabhuc11f1df2014-03-11 16:11:47 +00004091}
4092
4093static void
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07004094smb21_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
4095 unsigned int epoch, bool *purge_cache);
4096
4097static void
4098smb3_downgrade_oplock(struct TCP_Server_Info *server,
4099 struct cifsInodeInfo *cinode, __u32 oplock,
4100 unsigned int epoch, bool *purge_cache)
Pavel Shilovsky7b9b9ed2019-02-13 15:43:08 -08004101{
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07004102 unsigned int old_state = cinode->oplock;
4103 unsigned int old_epoch = cinode->epoch;
4104 unsigned int new_state;
4105
4106 if (epoch > old_epoch) {
4107 smb21_set_oplock_level(cinode, oplock, 0, NULL);
4108 cinode->epoch = epoch;
4109 }
4110
4111 new_state = cinode->oplock;
4112 *purge_cache = false;
4113
4114 if ((old_state & CIFS_CACHE_READ_FLG) != 0 &&
4115 (new_state & CIFS_CACHE_READ_FLG) == 0)
4116 *purge_cache = true;
4117 else if (old_state == new_state && (epoch - old_epoch > 1))
4118 *purge_cache = true;
Pavel Shilovsky7b9b9ed2019-02-13 15:43:08 -08004119}
4120
4121static void
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004122smb2_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
4123 unsigned int epoch, bool *purge_cache)
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004124{
4125 oplock &= 0xFF;
Rohith Surabattula0ab95c22021-05-17 11:28:34 +00004126 cinode->lease_granted = false;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004127 if (oplock == SMB2_OPLOCK_LEVEL_NOCHANGE)
4128 return;
4129 if (oplock == SMB2_OPLOCK_LEVEL_BATCH) {
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004130 cinode->oplock = CIFS_CACHE_RHW_FLG;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004131 cifs_dbg(FYI, "Batch Oplock granted on inode %p\n",
4132 &cinode->vfs_inode);
4133 } else if (oplock == SMB2_OPLOCK_LEVEL_EXCLUSIVE) {
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004134 cinode->oplock = CIFS_CACHE_RW_FLG;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004135 cifs_dbg(FYI, "Exclusive Oplock granted on inode %p\n",
4136 &cinode->vfs_inode);
4137 } else if (oplock == SMB2_OPLOCK_LEVEL_II) {
4138 cinode->oplock = CIFS_CACHE_READ_FLG;
4139 cifs_dbg(FYI, "Level II Oplock granted on inode %p\n",
4140 &cinode->vfs_inode);
4141 } else
4142 cinode->oplock = 0;
4143}
4144
4145static void
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004146smb21_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
4147 unsigned int epoch, bool *purge_cache)
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004148{
4149 char message[5] = {0};
Christoph Probst6a54b2e2019-05-07 17:16:40 +02004150 unsigned int new_oplock = 0;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004151
4152 oplock &= 0xFF;
Rohith Surabattula0ab95c22021-05-17 11:28:34 +00004153 cinode->lease_granted = true;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004154 if (oplock == SMB2_OPLOCK_LEVEL_NOCHANGE)
4155 return;
4156
Pavel Shilovskya016e272019-09-26 12:31:20 -07004157 /* Check if the server granted an oplock rather than a lease */
4158 if (oplock & SMB2_OPLOCK_LEVEL_EXCLUSIVE)
4159 return smb2_set_oplock_level(cinode, oplock, epoch,
4160 purge_cache);
4161
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004162 if (oplock & SMB2_LEASE_READ_CACHING_HE) {
Christoph Probst6a54b2e2019-05-07 17:16:40 +02004163 new_oplock |= CIFS_CACHE_READ_FLG;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004164 strcat(message, "R");
4165 }
4166 if (oplock & SMB2_LEASE_HANDLE_CACHING_HE) {
Christoph Probst6a54b2e2019-05-07 17:16:40 +02004167 new_oplock |= CIFS_CACHE_HANDLE_FLG;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004168 strcat(message, "H");
4169 }
4170 if (oplock & SMB2_LEASE_WRITE_CACHING_HE) {
Christoph Probst6a54b2e2019-05-07 17:16:40 +02004171 new_oplock |= CIFS_CACHE_WRITE_FLG;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004172 strcat(message, "W");
4173 }
Christoph Probst6a54b2e2019-05-07 17:16:40 +02004174 if (!new_oplock)
4175 strncpy(message, "None", sizeof(message));
4176
4177 cinode->oplock = new_oplock;
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004178 cifs_dbg(FYI, "%s Lease granted on inode %p\n", message,
4179 &cinode->vfs_inode);
4180}
4181
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004182static void
4183smb3_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
4184 unsigned int epoch, bool *purge_cache)
4185{
4186 unsigned int old_oplock = cinode->oplock;
4187
4188 smb21_set_oplock_level(cinode, oplock, epoch, purge_cache);
4189
4190 if (purge_cache) {
4191 *purge_cache = false;
4192 if (old_oplock == CIFS_CACHE_READ_FLG) {
4193 if (cinode->oplock == CIFS_CACHE_READ_FLG &&
4194 (epoch - cinode->epoch > 0))
4195 *purge_cache = true;
4196 else if (cinode->oplock == CIFS_CACHE_RH_FLG &&
4197 (epoch - cinode->epoch > 1))
4198 *purge_cache = true;
4199 else if (cinode->oplock == CIFS_CACHE_RHW_FLG &&
4200 (epoch - cinode->epoch > 1))
4201 *purge_cache = true;
4202 else if (cinode->oplock == 0 &&
4203 (epoch - cinode->epoch > 0))
4204 *purge_cache = true;
4205 } else if (old_oplock == CIFS_CACHE_RH_FLG) {
4206 if (cinode->oplock == CIFS_CACHE_RH_FLG &&
4207 (epoch - cinode->epoch > 0))
4208 *purge_cache = true;
4209 else if (cinode->oplock == CIFS_CACHE_RHW_FLG &&
4210 (epoch - cinode->epoch > 1))
4211 *purge_cache = true;
4212 }
4213 cinode->epoch = epoch;
4214 }
4215}
4216
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04004217static bool
4218smb2_is_read_op(__u32 oplock)
4219{
4220 return oplock == SMB2_OPLOCK_LEVEL_II;
4221}
4222
4223static bool
4224smb21_is_read_op(__u32 oplock)
4225{
4226 return (oplock & SMB2_LEASE_READ_CACHING_HE) &&
4227 !(oplock & SMB2_LEASE_WRITE_CACHING_HE);
4228}
4229
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004230static __le32
4231map_oplock_to_lease(u8 oplock)
4232{
4233 if (oplock == SMB2_OPLOCK_LEVEL_EXCLUSIVE)
4234 return SMB2_LEASE_WRITE_CACHING | SMB2_LEASE_READ_CACHING;
4235 else if (oplock == SMB2_OPLOCK_LEVEL_II)
4236 return SMB2_LEASE_READ_CACHING;
4237 else if (oplock == SMB2_OPLOCK_LEVEL_BATCH)
4238 return SMB2_LEASE_HANDLE_CACHING | SMB2_LEASE_READ_CACHING |
4239 SMB2_LEASE_WRITE_CACHING;
4240 return 0;
4241}
4242
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04004243static char *
4244smb2_create_lease_buf(u8 *lease_key, u8 oplock)
4245{
4246 struct create_lease *buf;
4247
4248 buf = kzalloc(sizeof(struct create_lease), GFP_KERNEL);
4249 if (!buf)
4250 return NULL;
4251
Stefano Brivio729c0c92018-07-05 15:10:02 +02004252 memcpy(&buf->lcontext.LeaseKey, lease_key, SMB2_LEASE_KEY_SIZE);
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004253 buf->lcontext.LeaseState = map_oplock_to_lease(oplock);
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04004254
4255 buf->ccontext.DataOffset = cpu_to_le16(offsetof
4256 (struct create_lease, lcontext));
4257 buf->ccontext.DataLength = cpu_to_le32(sizeof(struct lease_context));
4258 buf->ccontext.NameOffset = cpu_to_le16(offsetof
4259 (struct create_lease, Name));
4260 buf->ccontext.NameLength = cpu_to_le16(4);
Steve French12197a72014-05-14 05:29:40 -07004261 /* SMB2_CREATE_REQUEST_LEASE is "RqLs" */
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04004262 buf->Name[0] = 'R';
4263 buf->Name[1] = 'q';
4264 buf->Name[2] = 'L';
4265 buf->Name[3] = 's';
4266 return (char *)buf;
4267}
4268
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004269static char *
4270smb3_create_lease_buf(u8 *lease_key, u8 oplock)
4271{
4272 struct create_lease_v2 *buf;
4273
4274 buf = kzalloc(sizeof(struct create_lease_v2), GFP_KERNEL);
4275 if (!buf)
4276 return NULL;
4277
Stefano Brivio729c0c92018-07-05 15:10:02 +02004278 memcpy(&buf->lcontext.LeaseKey, lease_key, SMB2_LEASE_KEY_SIZE);
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004279 buf->lcontext.LeaseState = map_oplock_to_lease(oplock);
4280
4281 buf->ccontext.DataOffset = cpu_to_le16(offsetof
4282 (struct create_lease_v2, lcontext));
4283 buf->ccontext.DataLength = cpu_to_le32(sizeof(struct lease_context_v2));
4284 buf->ccontext.NameOffset = cpu_to_le16(offsetof
4285 (struct create_lease_v2, Name));
4286 buf->ccontext.NameLength = cpu_to_le16(4);
Steve French12197a72014-05-14 05:29:40 -07004287 /* SMB2_CREATE_REQUEST_LEASE is "RqLs" */
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004288 buf->Name[0] = 'R';
4289 buf->Name[1] = 'q';
4290 buf->Name[2] = 'L';
4291 buf->Name[3] = 's';
4292 return (char *)buf;
4293}
4294
Pavel Shilovskyb5c7cde2013-09-05 20:16:45 +04004295static __u8
Ronnie Sahlberg96164ab2018-04-26 08:10:18 -06004296smb2_parse_lease_buf(void *buf, unsigned int *epoch, char *lease_key)
Pavel Shilovskyb5c7cde2013-09-05 20:16:45 +04004297{
4298 struct create_lease *lc = (struct create_lease *)buf;
4299
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004300 *epoch = 0; /* not used */
Pavel Shilovskyb5c7cde2013-09-05 20:16:45 +04004301 if (lc->lcontext.LeaseFlags & SMB2_LEASE_FLAG_BREAK_IN_PROGRESS)
4302 return SMB2_OPLOCK_LEVEL_NOCHANGE;
4303 return le32_to_cpu(lc->lcontext.LeaseState);
4304}
4305
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004306static __u8
Ronnie Sahlberg96164ab2018-04-26 08:10:18 -06004307smb3_parse_lease_buf(void *buf, unsigned int *epoch, char *lease_key)
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004308{
4309 struct create_lease_v2 *lc = (struct create_lease_v2 *)buf;
4310
Pavel Shilovsky42873b02013-09-05 21:30:16 +04004311 *epoch = le16_to_cpu(lc->lcontext.Epoch);
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004312 if (lc->lcontext.LeaseFlags & SMB2_LEASE_FLAG_BREAK_IN_PROGRESS)
4313 return SMB2_OPLOCK_LEVEL_NOCHANGE;
Ronnie Sahlberg96164ab2018-04-26 08:10:18 -06004314 if (lease_key)
Stefano Brivio729c0c92018-07-05 15:10:02 +02004315 memcpy(lease_key, &lc->lcontext.LeaseKey, SMB2_LEASE_KEY_SIZE);
Pavel Shilovskyf0473902013-09-04 13:44:05 +04004316 return le32_to_cpu(lc->lcontext.LeaseState);
4317}
4318
Pavel Shilovsky7f6c5002014-06-22 11:03:22 +04004319static unsigned int
4320smb2_wp_retry_size(struct inode *inode)
4321{
Ronnie Sahlberg522aa3b2020-12-14 16:40:17 +10004322 return min_t(unsigned int, CIFS_SB(inode->i_sb)->ctx->wsize,
Pavel Shilovsky7f6c5002014-06-22 11:03:22 +04004323 SMB2_MAX_BUFFER_SIZE);
4324}
4325
Pavel Shilovsky52755802014-08-18 20:49:57 +04004326static bool
4327smb2_dir_needs_close(struct cifsFileInfo *cfile)
4328{
4329 return !cfile->invalidHandle;
4330}
4331
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004332static void
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10004333fill_transform_hdr(struct smb2_transform_hdr *tr_hdr, unsigned int orig_len,
Steve French2b2f7542019-06-07 15:16:10 -05004334 struct smb_rqst *old_rq, __le16 cipher_type)
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004335{
4336 struct smb2_sync_hdr *shdr =
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004337 (struct smb2_sync_hdr *)old_rq->rq_iov[0].iov_base;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004338
4339 memset(tr_hdr, 0, sizeof(struct smb2_transform_hdr));
4340 tr_hdr->ProtocolId = SMB2_TRANSFORM_PROTO_NUM;
4341 tr_hdr->OriginalMessageSize = cpu_to_le32(orig_len);
4342 tr_hdr->Flags = cpu_to_le16(0x01);
Steve French63ca5652020-10-15 23:41:40 -05004343 if ((cipher_type == SMB2_ENCRYPTION_AES128_GCM) ||
4344 (cipher_type == SMB2_ENCRYPTION_AES256_GCM))
Steve Frenchfd08f2d2020-10-15 00:25:02 -05004345 get_random_bytes(&tr_hdr->Nonce, SMB3_AES_GCM_NONCE);
Steve French2b2f7542019-06-07 15:16:10 -05004346 else
Steve Frenchfd08f2d2020-10-15 00:25:02 -05004347 get_random_bytes(&tr_hdr->Nonce, SMB3_AES_CCM_NONCE);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004348 memcpy(&tr_hdr->SessionId, &shdr->SessionId, 8);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004349}
4350
Ronnie Sahlberg262916b2018-02-20 12:45:21 +11004351/* We can not use the normal sg_set_buf() as we will sometimes pass a
4352 * stack object as buf.
4353 */
4354static inline void smb2_sg_set_buf(struct scatterlist *sg, const void *buf,
4355 unsigned int buflen)
4356{
Sebastien Tisserantee9d6612019-08-01 12:06:08 -05004357 void *addr;
4358 /*
4359 * VMAP_STACK (at least) puts stack into the vmalloc address space
4360 */
4361 if (is_vmalloc_addr(buf))
4362 addr = vmalloc_to_page(buf);
4363 else
4364 addr = virt_to_page(buf);
4365 sg_set_page(sg, addr, buflen, offset_in_page(buf));
Ronnie Sahlberg262916b2018-02-20 12:45:21 +11004366}
4367
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004368/* Assumes the first rqst has a transform header as the first iov.
4369 * I.e.
4370 * rqst[0].rq_iov[0] is transform header
4371 * rqst[0].rq_iov[1+] data to be encrypted/decrypted
4372 * rqst[1+].rq_iov[0+] data to be encrypted/decrypted
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10004373 */
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004374static struct scatterlist *
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004375init_sg(int num_rqst, struct smb_rqst *rqst, u8 *sign)
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004376{
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004377 unsigned int sg_len;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004378 struct scatterlist *sg;
4379 unsigned int i;
4380 unsigned int j;
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004381 unsigned int idx = 0;
4382 int skip;
4383
4384 sg_len = 1;
4385 for (i = 0; i < num_rqst; i++)
4386 sg_len += rqst[i].rq_nvec + rqst[i].rq_npages;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004387
4388 sg = kmalloc_array(sg_len, sizeof(struct scatterlist), GFP_KERNEL);
4389 if (!sg)
4390 return NULL;
4391
4392 sg_init_table(sg, sg_len);
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004393 for (i = 0; i < num_rqst; i++) {
4394 for (j = 0; j < rqst[i].rq_nvec; j++) {
4395 /*
4396 * The first rqst has a transform header where the
4397 * first 20 bytes are not part of the encrypted blob
4398 */
4399 skip = (i == 0) && (j == 0) ? 20 : 0;
4400 smb2_sg_set_buf(&sg[idx++],
4401 rqst[i].rq_iov[j].iov_base + skip,
4402 rqst[i].rq_iov[j].iov_len - skip);
Ronnie Sahlberge77fe732018-12-31 13:43:40 +10004403 }
Steve Frenchd5f07fb2018-06-05 17:46:24 -05004404
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004405 for (j = 0; j < rqst[i].rq_npages; j++) {
4406 unsigned int len, offset;
4407
4408 rqst_page_get_length(&rqst[i], j, &len, &offset);
4409 sg_set_page(&sg[idx++], rqst[i].rq_pages[j], len, offset);
4410 }
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004411 }
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004412 smb2_sg_set_buf(&sg[idx], sign, SMB2_SIGNATURE_SIZE);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004413 return sg;
4414}
4415
Pavel Shilovsky61cfac6f2017-02-28 16:05:19 -08004416static int
4417smb2_get_enc_key(struct TCP_Server_Info *server, __u64 ses_id, int enc, u8 *key)
4418{
4419 struct cifs_ses *ses;
4420 u8 *ses_enc_key;
4421
4422 spin_lock(&cifs_tcp_ses_lock);
Aurelien Apteld70e9fa2019-09-20 06:31:10 +02004423 list_for_each_entry(server, &cifs_tcp_ses_list, tcp_ses_list) {
4424 list_for_each_entry(ses, &server->smb_ses_list, smb_ses_list) {
4425 if (ses->Suid == ses_id) {
4426 ses_enc_key = enc ? ses->smb3encryptionkey :
4427 ses->smb3decryptionkey;
Shyam Prasad N45a45462021-03-25 12:34:54 +00004428 memcpy(key, ses_enc_key, SMB3_ENC_DEC_KEY_SIZE);
Aurelien Apteld70e9fa2019-09-20 06:31:10 +02004429 spin_unlock(&cifs_tcp_ses_lock);
4430 return 0;
4431 }
4432 }
Pavel Shilovsky61cfac6f2017-02-28 16:05:19 -08004433 }
4434 spin_unlock(&cifs_tcp_ses_lock);
4435
Paul Aurich83728cb2021-04-13 14:25:27 -07004436 return -EAGAIN;
Pavel Shilovsky61cfac6f2017-02-28 16:05:19 -08004437}
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004438/*
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004439 * Encrypt or decrypt @rqst message. @rqst[0] has the following format:
4440 * iov[0] - transform header (associate data),
4441 * iov[1-N] - SMB2 header and pages - data to encrypt.
4442 * On success return encrypted data in iov[1-N] and pages, leave iov[0]
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004443 * untouched.
4444 */
4445static int
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004446crypt_message(struct TCP_Server_Info *server, int num_rqst,
4447 struct smb_rqst *rqst, int enc)
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004448{
4449 struct smb2_transform_hdr *tr_hdr =
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004450 (struct smb2_transform_hdr *)rqst[0].rq_iov[0].iov_base;
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10004451 unsigned int assoc_data_len = sizeof(struct smb2_transform_hdr) - 20;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004452 int rc = 0;
4453 struct scatterlist *sg;
4454 u8 sign[SMB2_SIGNATURE_SIZE] = {};
Shyam Prasad N45a45462021-03-25 12:34:54 +00004455 u8 key[SMB3_ENC_DEC_KEY_SIZE];
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004456 struct aead_request *req;
4457 char *iv;
4458 unsigned int iv_len;
Gilad Ben-Yossefa5186b82017-10-18 08:00:46 +01004459 DECLARE_CRYPTO_WAIT(wait);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004460 struct crypto_aead *tfm;
4461 unsigned int crypt_len = le32_to_cpu(tr_hdr->OriginalMessageSize);
4462
Pavel Shilovsky61cfac6f2017-02-28 16:05:19 -08004463 rc = smb2_get_enc_key(server, tr_hdr->SessionId, enc, key);
4464 if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004465 cifs_server_dbg(VFS, "%s: Could not get %scryption key\n", __func__,
Pavel Shilovsky61cfac6f2017-02-28 16:05:19 -08004466 enc ? "en" : "de");
Shyam Prasad N0bd294b2020-10-15 10:41:31 -07004467 return rc;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004468 }
4469
4470 rc = smb3_crypto_aead_allocate(server);
4471 if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004472 cifs_server_dbg(VFS, "%s: crypto alloc failed\n", __func__);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004473 return rc;
4474 }
4475
4476 tfm = enc ? server->secmech.ccmaesencrypt :
4477 server->secmech.ccmaesdecrypt;
Steve French63ca5652020-10-15 23:41:40 -05004478
Shyam Prasad N45a45462021-03-25 12:34:54 +00004479 if ((server->cipher_type == SMB2_ENCRYPTION_AES256_CCM) ||
4480 (server->cipher_type == SMB2_ENCRYPTION_AES256_GCM))
Steve French63ca5652020-10-15 23:41:40 -05004481 rc = crypto_aead_setkey(tfm, key, SMB3_GCM256_CRYPTKEY_SIZE);
4482 else
Shyam Prasad N45a45462021-03-25 12:34:54 +00004483 rc = crypto_aead_setkey(tfm, key, SMB3_GCM128_CRYPTKEY_SIZE);
Steve French63ca5652020-10-15 23:41:40 -05004484
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004485 if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004486 cifs_server_dbg(VFS, "%s: Failed to set aead key %d\n", __func__, rc);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004487 return rc;
4488 }
4489
4490 rc = crypto_aead_setauthsize(tfm, SMB2_SIGNATURE_SIZE);
4491 if (rc) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004492 cifs_server_dbg(VFS, "%s: Failed to set authsize %d\n", __func__, rc);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004493 return rc;
4494 }
4495
4496 req = aead_request_alloc(tfm, GFP_KERNEL);
4497 if (!req) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004498 cifs_server_dbg(VFS, "%s: Failed to alloc aead request\n", __func__);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004499 return -ENOMEM;
4500 }
4501
4502 if (!enc) {
4503 memcpy(sign, &tr_hdr->Signature, SMB2_SIGNATURE_SIZE);
4504 crypt_len += SMB2_SIGNATURE_SIZE;
4505 }
4506
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004507 sg = init_sg(num_rqst, rqst, sign);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004508 if (!sg) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004509 cifs_server_dbg(VFS, "%s: Failed to init sg\n", __func__);
Christophe Jaillet517a6e42017-06-11 09:12:47 +02004510 rc = -ENOMEM;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004511 goto free_req;
4512 }
4513
4514 iv_len = crypto_aead_ivsize(tfm);
4515 iv = kzalloc(iv_len, GFP_KERNEL);
4516 if (!iv) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004517 cifs_server_dbg(VFS, "%s: Failed to alloc iv\n", __func__);
Christophe Jaillet517a6e42017-06-11 09:12:47 +02004518 rc = -ENOMEM;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004519 goto free_sg;
4520 }
Steve French2b2f7542019-06-07 15:16:10 -05004521
Steve French63ca5652020-10-15 23:41:40 -05004522 if ((server->cipher_type == SMB2_ENCRYPTION_AES128_GCM) ||
4523 (server->cipher_type == SMB2_ENCRYPTION_AES256_GCM))
Steve Frenchfd08f2d2020-10-15 00:25:02 -05004524 memcpy(iv, (char *)tr_hdr->Nonce, SMB3_AES_GCM_NONCE);
Steve French2b2f7542019-06-07 15:16:10 -05004525 else {
4526 iv[0] = 3;
Steve Frenchfd08f2d2020-10-15 00:25:02 -05004527 memcpy(iv + 1, (char *)tr_hdr->Nonce, SMB3_AES_CCM_NONCE);
Steve French2b2f7542019-06-07 15:16:10 -05004528 }
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004529
4530 aead_request_set_crypt(req, sg, sg, crypt_len, iv);
4531 aead_request_set_ad(req, assoc_data_len);
4532
4533 aead_request_set_callback(req, CRYPTO_TFM_REQ_MAY_BACKLOG,
Gilad Ben-Yossefa5186b82017-10-18 08:00:46 +01004534 crypto_req_done, &wait);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004535
Gilad Ben-Yossefa5186b82017-10-18 08:00:46 +01004536 rc = crypto_wait_req(enc ? crypto_aead_encrypt(req)
4537 : crypto_aead_decrypt(req), &wait);
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004538
4539 if (!rc && enc)
4540 memcpy(&tr_hdr->Signature, sign, SMB2_SIGNATURE_SIZE);
4541
4542 kfree(iv);
4543free_sg:
4544 kfree(sg);
4545free_req:
4546 kfree(req);
4547 return rc;
4548}
4549
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004550void
4551smb3_free_compound_rqst(int num_rqst, struct smb_rqst *rqst)
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004552{
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004553 int i, j;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004554
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004555 for (i = 0; i < num_rqst; i++) {
4556 if (rqst[i].rq_pages) {
4557 for (j = rqst[i].rq_npages - 1; j >= 0; j--)
4558 put_page(rqst[i].rq_pages[j]);
4559 kfree(rqst[i].rq_pages);
4560 }
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004561 }
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004562}
4563
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004564/*
4565 * This function will initialize new_rq and encrypt the content.
4566 * The first entry, new_rq[0], only contains a single iov which contains
4567 * a smb2_transform_hdr and is pre-allocated by the caller.
4568 * This function then populates new_rq[1+] with the content from olq_rq[0+].
4569 *
4570 * The end result is an array of smb_rqst structures where the first structure
4571 * only contains a single iov for the transform header which we then can pass
4572 * to crypt_message().
4573 *
4574 * new_rq[0].rq_iov[0] : smb2_transform_hdr pre-allocated by the caller
4575 * new_rq[1+].rq_iov[*] == old_rq[0+].rq_iov[*] : SMB2/3 requests
4576 */
4577static int
4578smb3_init_transform_rq(struct TCP_Server_Info *server, int num_rqst,
4579 struct smb_rqst *new_rq, struct smb_rqst *old_rq)
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004580{
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004581 struct page **pages;
4582 struct smb2_transform_hdr *tr_hdr = new_rq[0].rq_iov[0].iov_base;
4583 unsigned int npages;
4584 unsigned int orig_len = 0;
4585 int i, j;
4586 int rc = -ENOMEM;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004587
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004588 for (i = 1; i < num_rqst; i++) {
4589 npages = old_rq[i - 1].rq_npages;
4590 pages = kmalloc_array(npages, sizeof(struct page *),
4591 GFP_KERNEL);
4592 if (!pages)
4593 goto err_free;
4594
4595 new_rq[i].rq_pages = pages;
4596 new_rq[i].rq_npages = npages;
4597 new_rq[i].rq_offset = old_rq[i - 1].rq_offset;
4598 new_rq[i].rq_pagesz = old_rq[i - 1].rq_pagesz;
4599 new_rq[i].rq_tailsz = old_rq[i - 1].rq_tailsz;
4600 new_rq[i].rq_iov = old_rq[i - 1].rq_iov;
4601 new_rq[i].rq_nvec = old_rq[i - 1].rq_nvec;
4602
4603 orig_len += smb_rqst_len(server, &old_rq[i - 1]);
4604
4605 for (j = 0; j < npages; j++) {
4606 pages[j] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
4607 if (!pages[j])
4608 goto err_free;
4609 }
4610
4611 /* copy pages form the old */
4612 for (j = 0; j < npages; j++) {
4613 char *dst, *src;
4614 unsigned int offset, len;
4615
4616 rqst_page_get_length(&new_rq[i], j, &len, &offset);
4617
4618 dst = (char *) kmap(new_rq[i].rq_pages[j]) + offset;
4619 src = (char *) kmap(old_rq[i - 1].rq_pages[j]) + offset;
4620
4621 memcpy(dst, src, len);
4622 kunmap(new_rq[i].rq_pages[j]);
4623 kunmap(old_rq[i - 1].rq_pages[j]);
4624 }
4625 }
4626
4627 /* fill the 1st iov with a transform header */
Steve French2b2f7542019-06-07 15:16:10 -05004628 fill_transform_hdr(tr_hdr, orig_len, old_rq, server->cipher_type);
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004629
4630 rc = crypt_message(server, num_rqst, new_rq, 1);
Christoph Probsta205d502019-05-08 21:36:25 +02004631 cifs_dbg(FYI, "Encrypt message returned %d\n", rc);
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004632 if (rc)
4633 goto err_free;
4634
4635 return rc;
4636
4637err_free:
4638 smb3_free_compound_rqst(num_rqst - 1, &new_rq[1]);
4639 return rc;
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07004640}
4641
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004642static int
4643smb3_is_transform_hdr(void *buf)
4644{
4645 struct smb2_transform_hdr *trhdr = buf;
4646
4647 return trhdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM;
4648}
4649
4650static int
4651decrypt_raw_data(struct TCP_Server_Info *server, char *buf,
4652 unsigned int buf_data_size, struct page **pages,
Rohith Surabattula62593012020-10-08 09:58:41 +00004653 unsigned int npages, unsigned int page_data_size,
4654 bool is_offloaded)
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004655{
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004656 struct kvec iov[2];
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004657 struct smb_rqst rqst = {NULL};
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004658 int rc;
4659
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004660 iov[0].iov_base = buf;
4661 iov[0].iov_len = sizeof(struct smb2_transform_hdr);
4662 iov[1].iov_base = buf + sizeof(struct smb2_transform_hdr);
4663 iov[1].iov_len = buf_data_size;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004664
4665 rqst.rq_iov = iov;
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004666 rqst.rq_nvec = 2;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004667 rqst.rq_pages = pages;
4668 rqst.rq_npages = npages;
4669 rqst.rq_pagesz = PAGE_SIZE;
4670 rqst.rq_tailsz = (page_data_size % PAGE_SIZE) ? : PAGE_SIZE;
4671
Ronnie Sahlbergb2c96de2018-08-01 09:26:11 +10004672 rc = crypt_message(server, 1, &rqst, 0);
Christoph Probsta205d502019-05-08 21:36:25 +02004673 cifs_dbg(FYI, "Decrypt message returned %d\n", rc);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004674
4675 if (rc)
4676 return rc;
4677
Ronnie Sahlbergc713c872018-06-12 08:00:58 +10004678 memmove(buf, iov[1].iov_base, buf_data_size);
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10004679
Rohith Surabattula62593012020-10-08 09:58:41 +00004680 if (!is_offloaded)
4681 server->total_read = buf_data_size + page_data_size;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004682
4683 return rc;
4684}
4685
4686static int
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004687read_data_into_pages(struct TCP_Server_Info *server, struct page **pages,
4688 unsigned int npages, unsigned int len)
4689{
4690 int i;
4691 int length;
4692
4693 for (i = 0; i < npages; i++) {
4694 struct page *page = pages[i];
4695 size_t n;
4696
4697 n = len;
4698 if (len >= PAGE_SIZE) {
4699 /* enough data to fill the page */
4700 n = PAGE_SIZE;
4701 len -= n;
4702 } else {
4703 zero_user(page, len, PAGE_SIZE - len);
4704 len = 0;
4705 }
Long Li1dbe3462018-05-30 12:47:55 -07004706 length = cifs_read_page_from_socket(server, page, 0, n);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004707 if (length < 0)
4708 return length;
4709 server->total_read += length;
4710 }
4711
4712 return 0;
4713}
4714
4715static int
4716init_read_bvec(struct page **pages, unsigned int npages, unsigned int data_size,
4717 unsigned int cur_off, struct bio_vec **page_vec)
4718{
4719 struct bio_vec *bvec;
4720 int i;
4721
4722 bvec = kcalloc(npages, sizeof(struct bio_vec), GFP_KERNEL);
4723 if (!bvec)
4724 return -ENOMEM;
4725
4726 for (i = 0; i < npages; i++) {
4727 bvec[i].bv_page = pages[i];
4728 bvec[i].bv_offset = (i == 0) ? cur_off : 0;
4729 bvec[i].bv_len = min_t(unsigned int, PAGE_SIZE, data_size);
4730 data_size -= bvec[i].bv_len;
4731 }
4732
4733 if (data_size != 0) {
4734 cifs_dbg(VFS, "%s: something went wrong\n", __func__);
4735 kfree(bvec);
4736 return -EIO;
4737 }
4738
4739 *page_vec = bvec;
4740 return 0;
4741}
4742
4743static int
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004744handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
4745 char *buf, unsigned int buf_len, struct page **pages,
Rohith Surabattulade9ac0a2020-10-28 13:42:21 +00004746 unsigned int npages, unsigned int page_data_size,
4747 bool is_offloaded)
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004748{
4749 unsigned int data_offset;
4750 unsigned int data_len;
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004751 unsigned int cur_off;
4752 unsigned int cur_page_idx;
4753 unsigned int pad_len;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004754 struct cifs_readdata *rdata = mid->callback_data;
Ronnie Sahlberg49f466b2018-06-01 10:53:06 +10004755 struct smb2_sync_hdr *shdr = (struct smb2_sync_hdr *)buf;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004756 struct bio_vec *bvec = NULL;
4757 struct iov_iter iter;
4758 struct kvec iov;
4759 int length;
Long Li74dcf412017-11-22 17:38:46 -07004760 bool use_rdma_mr = false;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004761
4762 if (shdr->Command != SMB2_READ) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10004763 cifs_server_dbg(VFS, "only big read responses are supported\n");
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004764 return -ENOTSUPP;
4765 }
4766
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07004767 if (server->ops->is_session_expired &&
4768 server->ops->is_session_expired(buf)) {
Rohith Surabattulade9ac0a2020-10-28 13:42:21 +00004769 if (!is_offloaded)
4770 cifs_reconnect(server);
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07004771 return -1;
4772 }
4773
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004774 if (server->ops->is_status_pending &&
Pavel Shilovsky66265f12019-01-23 17:11:16 -08004775 server->ops->is_status_pending(buf, server))
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004776 return -1;
4777
Pavel Shilovskyec678ea2019-01-18 15:38:11 -08004778 /* set up first two iov to get credits */
4779 rdata->iov[0].iov_base = buf;
Pavel Shilovskybb1bccb2019-01-17 16:18:38 -08004780 rdata->iov[0].iov_len = 0;
4781 rdata->iov[1].iov_base = buf;
Pavel Shilovskyec678ea2019-01-18 15:38:11 -08004782 rdata->iov[1].iov_len =
Pavel Shilovskybb1bccb2019-01-17 16:18:38 -08004783 min_t(unsigned int, buf_len, server->vals->read_rsp_size);
Pavel Shilovskyec678ea2019-01-18 15:38:11 -08004784 cifs_dbg(FYI, "0: iov_base=%p iov_len=%zu\n",
4785 rdata->iov[0].iov_base, rdata->iov[0].iov_len);
4786 cifs_dbg(FYI, "1: iov_base=%p iov_len=%zu\n",
4787 rdata->iov[1].iov_base, rdata->iov[1].iov_len);
4788
4789 rdata->result = server->ops->map_error(buf, true);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004790 if (rdata->result != 0) {
4791 cifs_dbg(FYI, "%s: server returned error %d\n",
4792 __func__, rdata->result);
Pavel Shilovskyec678ea2019-01-18 15:38:11 -08004793 /* normal error on read response */
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004794 if (is_offloaded)
4795 mid->mid_state = MID_RESPONSE_RECEIVED;
4796 else
4797 dequeue_mid(mid, false);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004798 return 0;
4799 }
4800
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10004801 data_offset = server->ops->read_data_offset(buf);
Long Li74dcf412017-11-22 17:38:46 -07004802#ifdef CONFIG_CIFS_SMB_DIRECT
4803 use_rdma_mr = rdata->mr;
4804#endif
4805 data_len = server->ops->read_data_length(buf, use_rdma_mr);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004806
4807 if (data_offset < server->vals->read_rsp_size) {
4808 /*
4809 * win2k8 sometimes sends an offset of 0 when the read
4810 * is beyond the EOF. Treat it as if the data starts just after
4811 * the header.
4812 */
4813 cifs_dbg(FYI, "%s: data offset (%u) inside read response header\n",
4814 __func__, data_offset);
4815 data_offset = server->vals->read_rsp_size;
4816 } else if (data_offset > MAX_CIFS_SMALL_BUFFER_SIZE) {
4817 /* data_offset is beyond the end of smallbuf */
4818 cifs_dbg(FYI, "%s: data offset (%u) beyond end of smallbuf\n",
4819 __func__, data_offset);
4820 rdata->result = -EIO;
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004821 if (is_offloaded)
4822 mid->mid_state = MID_RESPONSE_MALFORMED;
4823 else
4824 dequeue_mid(mid, rdata->result);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004825 return 0;
4826 }
4827
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004828 pad_len = data_offset - server->vals->read_rsp_size;
4829
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004830 if (buf_len <= data_offset) {
4831 /* read response payload is in pages */
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004832 cur_page_idx = pad_len / PAGE_SIZE;
4833 cur_off = pad_len % PAGE_SIZE;
4834
4835 if (cur_page_idx != 0) {
4836 /* data offset is beyond the 1st page of response */
4837 cifs_dbg(FYI, "%s: data offset (%u) beyond 1st page of response\n",
4838 __func__, data_offset);
4839 rdata->result = -EIO;
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004840 if (is_offloaded)
4841 mid->mid_state = MID_RESPONSE_MALFORMED;
4842 else
4843 dequeue_mid(mid, rdata->result);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004844 return 0;
4845 }
4846
4847 if (data_len > page_data_size - pad_len) {
4848 /* data_len is corrupt -- discard frame */
4849 rdata->result = -EIO;
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004850 if (is_offloaded)
4851 mid->mid_state = MID_RESPONSE_MALFORMED;
4852 else
4853 dequeue_mid(mid, rdata->result);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004854 return 0;
4855 }
4856
4857 rdata->result = init_read_bvec(pages, npages, page_data_size,
4858 cur_off, &bvec);
4859 if (rdata->result != 0) {
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004860 if (is_offloaded)
4861 mid->mid_state = MID_RESPONSE_MALFORMED;
4862 else
4863 dequeue_mid(mid, rdata->result);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004864 return 0;
4865 }
4866
David Howellsaa563d72018-10-20 00:57:56 +01004867 iov_iter_bvec(&iter, WRITE, bvec, npages, data_len);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004868 } else if (buf_len >= data_offset + data_len) {
4869 /* read response payload is in buf */
4870 WARN_ONCE(npages > 0, "read data can be either in buf or in pages");
4871 iov.iov_base = buf + data_offset;
4872 iov.iov_len = data_len;
David Howellsaa563d72018-10-20 00:57:56 +01004873 iov_iter_kvec(&iter, WRITE, &iov, 1, data_len);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004874 } else {
4875 /* read response payload cannot be in both buf and pages */
4876 WARN_ONCE(1, "buf can not contain only a part of read data");
4877 rdata->result = -EIO;
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004878 if (is_offloaded)
4879 mid->mid_state = MID_RESPONSE_MALFORMED;
4880 else
4881 dequeue_mid(mid, rdata->result);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004882 return 0;
4883 }
4884
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004885 length = rdata->copy_into_pages(server, rdata, &iter);
4886
4887 kfree(bvec);
4888
4889 if (length < 0)
4890 return length;
4891
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004892 if (is_offloaded)
4893 mid->mid_state = MID_RESPONSE_RECEIVED;
4894 else
4895 dequeue_mid(mid, false);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004896 return length;
4897}
4898
Steve French35cf94a2019-09-07 01:09:49 -05004899struct smb2_decrypt_work {
4900 struct work_struct decrypt;
4901 struct TCP_Server_Info *server;
4902 struct page **ppages;
4903 char *buf;
4904 unsigned int npages;
4905 unsigned int len;
4906};
4907
4908
4909static void smb2_decrypt_offload(struct work_struct *work)
4910{
4911 struct smb2_decrypt_work *dw = container_of(work,
4912 struct smb2_decrypt_work, decrypt);
4913 int i, rc;
4914 struct mid_q_entry *mid;
4915
4916 rc = decrypt_raw_data(dw->server, dw->buf, dw->server->vals->read_rsp_size,
Rohith Surabattula62593012020-10-08 09:58:41 +00004917 dw->ppages, dw->npages, dw->len, true);
Steve French35cf94a2019-09-07 01:09:49 -05004918 if (rc) {
4919 cifs_dbg(VFS, "error decrypting rc=%d\n", rc);
4920 goto free_pages;
4921 }
4922
Steve French22553972019-09-13 16:47:31 -05004923 dw->server->lstrp = jiffies;
Rohith Surabattulaac873aa2020-10-29 05:03:10 +00004924 mid = smb2_find_dequeue_mid(dw->server, dw->buf);
Steve French35cf94a2019-09-07 01:09:49 -05004925 if (mid == NULL)
4926 cifs_dbg(FYI, "mid not found\n");
4927 else {
4928 mid->decrypted = true;
4929 rc = handle_read_data(dw->server, mid, dw->buf,
4930 dw->server->vals->read_rsp_size,
Rohith Surabattulade9ac0a2020-10-28 13:42:21 +00004931 dw->ppages, dw->npages, dw->len,
4932 true);
Rohith Surabattula12541002020-10-29 06:07:56 +00004933 if (rc >= 0) {
4934#ifdef CONFIG_CIFS_STATS2
4935 mid->when_received = jiffies;
4936#endif
Rohith Surabattula9e550b02021-02-16 10:40:45 +00004937 if (dw->server->ops->is_network_name_deleted)
4938 dw->server->ops->is_network_name_deleted(dw->buf,
4939 dw->server);
4940
Rohith Surabattula12541002020-10-29 06:07:56 +00004941 mid->callback(mid);
4942 } else {
4943 spin_lock(&GlobalMid_Lock);
4944 if (dw->server->tcpStatus == CifsNeedReconnect) {
4945 mid->mid_state = MID_RETRY_NEEDED;
4946 spin_unlock(&GlobalMid_Lock);
4947 mid->callback(mid);
4948 } else {
4949 mid->mid_state = MID_REQUEST_SUBMITTED;
4950 mid->mid_flags &= ~(MID_DELETED);
4951 list_add_tail(&mid->qhead,
4952 &dw->server->pending_mid_q);
4953 spin_unlock(&GlobalMid_Lock);
4954 }
4955 }
Steve French22553972019-09-13 16:47:31 -05004956 cifs_mid_q_entry_release(mid);
Steve French35cf94a2019-09-07 01:09:49 -05004957 }
4958
Steve French35cf94a2019-09-07 01:09:49 -05004959free_pages:
4960 for (i = dw->npages-1; i >= 0; i--)
4961 put_page(dw->ppages[i]);
4962
4963 kfree(dw->ppages);
4964 cifs_small_buf_release(dw->buf);
Steve Frencha08d8972019-10-26 16:00:44 -05004965 kfree(dw);
Steve French35cf94a2019-09-07 01:09:49 -05004966}
4967
4968
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08004969static int
Steve French35cf94a2019-09-07 01:09:49 -05004970receive_encrypted_read(struct TCP_Server_Info *server, struct mid_q_entry **mid,
4971 int *num_mids)
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004972{
4973 char *buf = server->smallbuf;
4974 struct smb2_transform_hdr *tr_hdr = (struct smb2_transform_hdr *)buf;
4975 unsigned int npages;
4976 struct page **pages;
4977 unsigned int len;
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10004978 unsigned int buflen = server->pdu_size;
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004979 int rc;
4980 int i = 0;
Steve French35cf94a2019-09-07 01:09:49 -05004981 struct smb2_decrypt_work *dw;
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004982
Steve French35cf94a2019-09-07 01:09:49 -05004983 *num_mids = 1;
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10004984 len = min_t(unsigned int, buflen, server->vals->read_rsp_size +
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004985 sizeof(struct smb2_transform_hdr)) - HEADER_SIZE(server) + 1;
4986
4987 rc = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1, len);
4988 if (rc < 0)
4989 return rc;
4990 server->total_read += rc;
4991
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10004992 len = le32_to_cpu(tr_hdr->OriginalMessageSize) -
Ronnie Sahlberg93012bf2018-03-31 11:45:31 +11004993 server->vals->read_rsp_size;
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08004994 npages = DIV_ROUND_UP(len, PAGE_SIZE);
4995
4996 pages = kmalloc_array(npages, sizeof(struct page *), GFP_KERNEL);
4997 if (!pages) {
4998 rc = -ENOMEM;
4999 goto discard_data;
5000 }
5001
5002 for (; i < npages; i++) {
5003 pages[i] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
5004 if (!pages[i]) {
5005 rc = -ENOMEM;
5006 goto discard_data;
5007 }
5008 }
5009
5010 /* read read data into pages */
5011 rc = read_data_into_pages(server, pages, npages, len);
5012 if (rc)
5013 goto free_pages;
5014
Pavel Shilovsky350be252017-04-10 10:31:33 -07005015 rc = cifs_discard_remaining_data(server);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08005016 if (rc)
5017 goto free_pages;
5018
Steve French35cf94a2019-09-07 01:09:49 -05005019 /*
5020 * For large reads, offload to different thread for better performance,
5021 * use more cores decrypting which can be expensive
5022 */
5023
Steve French10328c42019-09-09 13:30:15 -05005024 if ((server->min_offload) && (server->in_flight > 1) &&
Steve French563317e2019-09-08 23:22:02 -05005025 (server->pdu_size >= server->min_offload)) {
Steve French35cf94a2019-09-07 01:09:49 -05005026 dw = kmalloc(sizeof(struct smb2_decrypt_work), GFP_KERNEL);
5027 if (dw == NULL)
5028 goto non_offloaded_decrypt;
5029
5030 dw->buf = server->smallbuf;
5031 server->smallbuf = (char *)cifs_small_buf_get();
5032
5033 INIT_WORK(&dw->decrypt, smb2_decrypt_offload);
5034
5035 dw->npages = npages;
5036 dw->server = server;
5037 dw->ppages = pages;
5038 dw->len = len;
Steve Frencha08d8972019-10-26 16:00:44 -05005039 queue_work(decrypt_wq, &dw->decrypt);
Steve French35cf94a2019-09-07 01:09:49 -05005040 *num_mids = 0; /* worker thread takes care of finding mid */
5041 return -1;
5042 }
5043
5044non_offloaded_decrypt:
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005045 rc = decrypt_raw_data(server, buf, server->vals->read_rsp_size,
Rohith Surabattula62593012020-10-08 09:58:41 +00005046 pages, npages, len, false);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08005047 if (rc)
5048 goto free_pages;
5049
5050 *mid = smb2_find_mid(server, buf);
5051 if (*mid == NULL)
5052 cifs_dbg(FYI, "mid not found\n");
5053 else {
5054 cifs_dbg(FYI, "mid found\n");
5055 (*mid)->decrypted = true;
5056 rc = handle_read_data(server, *mid, buf,
5057 server->vals->read_rsp_size,
Rohith Surabattulade9ac0a2020-10-28 13:42:21 +00005058 pages, npages, len, false);
Rohith Surabattula9e550b02021-02-16 10:40:45 +00005059 if (rc >= 0) {
5060 if (server->ops->is_network_name_deleted) {
5061 server->ops->is_network_name_deleted(buf,
5062 server);
5063 }
5064 }
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08005065 }
5066
5067free_pages:
5068 for (i = i - 1; i >= 0; i--)
5069 put_page(pages[i]);
5070 kfree(pages);
5071 return rc;
5072discard_data:
Pavel Shilovsky350be252017-04-10 10:31:33 -07005073 cifs_discard_remaining_data(server);
Pavel Shilovskyc42a6ab2016-11-17 16:20:23 -08005074 goto free_pages;
5075}
5076
5077static int
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005078receive_encrypted_standard(struct TCP_Server_Info *server,
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005079 struct mid_q_entry **mids, char **bufs,
5080 int *num_mids)
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005081{
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005082 int ret, length;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005083 char *buf = server->smallbuf;
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005084 struct smb2_sync_hdr *shdr;
Ronnie Sahlberg2e964672018-04-09 18:06:26 +10005085 unsigned int pdu_length = server->pdu_size;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005086 unsigned int buf_size;
5087 struct mid_q_entry *mid_entry;
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005088 int next_is_large;
5089 char *next_buffer = NULL;
5090
5091 *num_mids = 0;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005092
5093 /* switch to large buffer if too big for a small one */
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005094 if (pdu_length > MAX_CIFS_SMALL_BUFFER_SIZE) {
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005095 server->large_buf = true;
5096 memcpy(server->bigbuf, buf, server->total_read);
5097 buf = server->bigbuf;
5098 }
5099
5100 /* now read the rest */
5101 length = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1,
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005102 pdu_length - HEADER_SIZE(server) + 1);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005103 if (length < 0)
5104 return length;
5105 server->total_read += length;
5106
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005107 buf_size = pdu_length - sizeof(struct smb2_transform_hdr);
Rohith Surabattula62593012020-10-08 09:58:41 +00005108 length = decrypt_raw_data(server, buf, buf_size, NULL, 0, 0, false);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005109 if (length)
5110 return length;
5111
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005112 next_is_large = server->large_buf;
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005113one_more:
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005114 shdr = (struct smb2_sync_hdr *)buf;
5115 if (shdr->NextCommand) {
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005116 if (next_is_large)
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005117 next_buffer = (char *)cifs_buf_get();
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005118 else
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005119 next_buffer = (char *)cifs_small_buf_get();
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005120 memcpy(next_buffer,
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005121 buf + le32_to_cpu(shdr->NextCommand),
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005122 pdu_length - le32_to_cpu(shdr->NextCommand));
5123 }
5124
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005125 mid_entry = smb2_find_mid(server, buf);
5126 if (mid_entry == NULL)
5127 cifs_dbg(FYI, "mid not found\n");
5128 else {
5129 cifs_dbg(FYI, "mid found\n");
5130 mid_entry->decrypted = true;
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005131 mid_entry->resp_buf_size = server->pdu_size;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005132 }
5133
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005134 if (*num_mids >= MAX_COMPOUND) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10005135 cifs_server_dbg(VFS, "too many PDUs in compound\n");
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005136 return -1;
5137 }
5138 bufs[*num_mids] = buf;
5139 mids[(*num_mids)++] = mid_entry;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005140
5141 if (mid_entry && mid_entry->handle)
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005142 ret = mid_entry->handle(server, mid_entry);
5143 else
5144 ret = cifs_handle_standard(server, mid_entry);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005145
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005146 if (ret == 0 && shdr->NextCommand) {
5147 pdu_length -= le32_to_cpu(shdr->NextCommand);
5148 server->large_buf = next_is_large;
5149 if (next_is_large)
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005150 server->bigbuf = buf = next_buffer;
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005151 else
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005152 server->smallbuf = buf = next_buffer;
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005153 goto one_more;
Pavel Shilovsky3edeb4a2019-07-22 11:38:22 -07005154 } else if (ret != 0) {
5155 /*
5156 * ret != 0 here means that we didn't get to handle_mid() thus
5157 * server->smallbuf and server->bigbuf are still valid. We need
5158 * to free next_buffer because it is not going to be used
5159 * anywhere.
5160 */
5161 if (next_is_large)
5162 free_rsp_buf(CIFS_LARGE_BUFFER, next_buffer);
5163 else
5164 free_rsp_buf(CIFS_SMALL_BUFFER, next_buffer);
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005165 }
5166
5167 return ret;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005168}
5169
5170static int
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005171smb3_receive_transform(struct TCP_Server_Info *server,
5172 struct mid_q_entry **mids, char **bufs, int *num_mids)
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005173{
5174 char *buf = server->smallbuf;
Ronnie Sahlberg2e964672018-04-09 18:06:26 +10005175 unsigned int pdu_length = server->pdu_size;
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005176 struct smb2_transform_hdr *tr_hdr = (struct smb2_transform_hdr *)buf;
5177 unsigned int orig_len = le32_to_cpu(tr_hdr->OriginalMessageSize);
5178
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005179 if (pdu_length < sizeof(struct smb2_transform_hdr) +
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005180 sizeof(struct smb2_sync_hdr)) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10005181 cifs_server_dbg(VFS, "Transform message is too small (%u)\n",
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005182 pdu_length);
5183 cifs_reconnect(server);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005184 return -ECONNABORTED;
5185 }
5186
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005187 if (pdu_length < orig_len + sizeof(struct smb2_transform_hdr)) {
Ronnie Sahlberg3175eb92019-09-04 12:32:41 +10005188 cifs_server_dbg(VFS, "Transform message is broken\n");
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005189 cifs_reconnect(server);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005190 return -ECONNABORTED;
5191 }
5192
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005193 /* TODO: add support for compounds containing READ. */
Paul Aurich6d2f84e2018-12-31 14:13:34 -08005194 if (pdu_length > CIFSMaxBufSize + MAX_HEADER_SIZE(server)) {
Steve French35cf94a2019-09-07 01:09:49 -05005195 return receive_encrypted_read(server, &mids[0], num_mids);
Paul Aurich6d2f84e2018-12-31 14:13:34 -08005196 }
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005197
Ronnie Sahlbergb24df3e2018-08-08 15:07:45 +10005198 return receive_encrypted_standard(server, mids, bufs, num_mids);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005199}
5200
5201int
5202smb3_handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid)
5203{
5204 char *buf = server->large_buf ? server->bigbuf : server->smallbuf;
5205
Ronnie Sahlberg1fc6ad22018-06-01 10:53:07 +10005206 return handle_read_data(server, mid, buf, server->pdu_size,
Rohith Surabattulade9ac0a2020-10-28 13:42:21 +00005207 NULL, 0, 0, false);
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005208}
5209
Ronnie Sahlberg8ce79ec2018-06-01 10:53:08 +10005210static int
5211smb2_next_header(char *buf)
5212{
5213 struct smb2_sync_hdr *hdr = (struct smb2_sync_hdr *)buf;
5214 struct smb2_transform_hdr *t_hdr = (struct smb2_transform_hdr *)buf;
5215
5216 if (hdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM)
5217 return sizeof(struct smb2_transform_hdr) +
5218 le32_to_cpu(t_hdr->OriginalMessageSize);
5219
5220 return le32_to_cpu(hdr->NextCommand);
5221}
5222
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005223static int
5224smb2_make_node(unsigned int xid, struct inode *inode,
5225 struct dentry *dentry, struct cifs_tcon *tcon,
Al Viro55869132021-03-18 01:38:53 -04005226 const char *full_path, umode_t mode, dev_t dev)
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005227{
5228 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
5229 int rc = -EPERM;
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005230 FILE_ALL_INFO *buf = NULL;
Aurelien Aptel7c065142020-06-04 17:23:55 +02005231 struct cifs_io_parms io_parms = {0};
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005232 __u32 oplock = 0;
5233 struct cifs_fid fid;
5234 struct cifs_open_parms oparms;
5235 unsigned int bytes_written;
5236 struct win_dev *pdev;
5237 struct kvec iov[2];
5238
5239 /*
5240 * Check if mounted with mount parm 'sfu' mount parm.
5241 * SFU emulation should work with all servers, but only
5242 * supports block and char device (no socket & fifo),
5243 * and was used by default in earlier versions of Windows
5244 */
5245 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_UNX_EMUL))
5246 goto out;
5247
5248 /*
5249 * TODO: Add ability to create instead via reparse point. Windows (e.g.
5250 * their current NFS server) uses this approach to expose special files
5251 * over SMB2/SMB3 and Samba will do this with SMB3.1.1 POSIX Extensions
5252 */
5253
5254 if (!S_ISCHR(mode) && !S_ISBLK(mode))
5255 goto out;
5256
5257 cifs_dbg(FYI, "sfu compat create special file\n");
5258
5259 buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL);
5260 if (buf == NULL) {
5261 rc = -ENOMEM;
5262 goto out;
5263 }
5264
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005265 oparms.tcon = tcon;
5266 oparms.cifs_sb = cifs_sb;
5267 oparms.desired_access = GENERIC_WRITE;
Amir Goldstein0f060932020-02-03 21:46:43 +02005268 oparms.create_options = cifs_create_options(cifs_sb, CREATE_NOT_DIR |
5269 CREATE_OPTION_SPECIAL);
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005270 oparms.disposition = FILE_CREATE;
5271 oparms.path = full_path;
5272 oparms.fid = &fid;
5273 oparms.reconnect = false;
5274
5275 if (tcon->ses->server->oplocks)
5276 oplock = REQ_OPLOCK;
5277 else
5278 oplock = 0;
5279 rc = tcon->ses->server->ops->open(xid, &oparms, &oplock, buf);
5280 if (rc)
5281 goto out;
5282
5283 /*
5284 * BB Do not bother to decode buf since no local inode yet to put
5285 * timestamps in, but we can reuse it safely.
5286 */
5287
5288 pdev = (struct win_dev *)buf;
5289 io_parms.pid = current->tgid;
5290 io_parms.tcon = tcon;
5291 io_parms.offset = 0;
5292 io_parms.length = sizeof(struct win_dev);
5293 iov[1].iov_base = buf;
5294 iov[1].iov_len = sizeof(struct win_dev);
5295 if (S_ISCHR(mode)) {
5296 memcpy(pdev->type, "IntxCHR", 8);
5297 pdev->major = cpu_to_le64(MAJOR(dev));
5298 pdev->minor = cpu_to_le64(MINOR(dev));
5299 rc = tcon->ses->server->ops->sync_write(xid, &fid, &io_parms,
5300 &bytes_written, iov, 1);
5301 } else if (S_ISBLK(mode)) {
5302 memcpy(pdev->type, "IntxBLK", 8);
5303 pdev->major = cpu_to_le64(MAJOR(dev));
5304 pdev->minor = cpu_to_le64(MINOR(dev));
5305 rc = tcon->ses->server->ops->sync_write(xid, &fid, &io_parms,
5306 &bytes_written, iov, 1);
5307 }
5308 tcon->ses->server->ops->close(xid, tcon, &fid);
5309 d_drop(dentry);
5310
5311 /* FIXME: add code here to set EAs */
5312out:
5313 kfree(buf);
5314 return rc;
5315}
5316
5317
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005318struct smb_version_operations smb20_operations = {
5319 .compare_fids = smb2_compare_fids,
5320 .setup_request = smb2_setup_request,
5321 .setup_async_request = smb2_setup_async_request,
5322 .check_receive = smb2_check_receive,
5323 .add_credits = smb2_add_credits,
5324 .set_credits = smb2_set_credits,
5325 .get_credits_field = smb2_get_credits_field,
5326 .get_credits = smb2_get_credits,
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +04005327 .wait_mtu_credits = cifs_wait_mtu_credits,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005328 .get_next_mid = smb2_get_next_mid,
Pavel Shilovskyc781af72019-03-04 14:02:50 -08005329 .revert_current_mid = smb2_revert_current_mid,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005330 .read_data_offset = smb2_read_data_offset,
5331 .read_data_length = smb2_read_data_length,
5332 .map_error = map_smb2_to_linux_error,
5333 .find_mid = smb2_find_mid,
5334 .check_message = smb2_check_message,
5335 .dump_detail = smb2_dump_detail,
5336 .clear_stats = smb2_clear_stats,
5337 .print_stats = smb2_print_stats,
5338 .is_oplock_break = smb2_is_valid_oplock_break,
Sachin Prabhu38bd4902017-03-03 15:41:38 -08005339 .handle_cancelled_mid = smb2_handle_cancelled_mid,
Sachin Prabhuc11f1df2014-03-11 16:11:47 +00005340 .downgrade_oplock = smb2_downgrade_oplock,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005341 .need_neg = smb2_need_neg,
5342 .negotiate = smb2_negotiate,
5343 .negotiate_wsize = smb2_negotiate_wsize,
5344 .negotiate_rsize = smb2_negotiate_rsize,
5345 .sess_setup = SMB2_sess_setup,
5346 .logoff = SMB2_logoff,
5347 .tree_connect = SMB2_tcon,
5348 .tree_disconnect = SMB2_tdis,
Steve French34f62642013-10-09 02:07:00 -05005349 .qfs_tcon = smb2_qfs_tcon,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005350 .is_path_accessible = smb2_is_path_accessible,
5351 .can_echo = smb2_can_echo,
5352 .echo = SMB2_echo,
5353 .query_path_info = smb2_query_path_info,
5354 .get_srv_inum = smb2_get_srv_inum,
5355 .query_file_info = smb2_query_file_info,
5356 .set_path_size = smb2_set_path_size,
5357 .set_file_size = smb2_set_file_size,
5358 .set_file_info = smb2_set_file_info,
Steve French64a5cfa2013-10-14 15:31:32 -05005359 .set_compression = smb2_set_compression,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005360 .mkdir = smb2_mkdir,
5361 .mkdir_setinfo = smb2_mkdir_setinfo,
5362 .rmdir = smb2_rmdir,
5363 .unlink = smb2_unlink,
5364 .rename = smb2_rename_path,
5365 .create_hardlink = smb2_create_hardlink,
5366 .query_symlink = smb2_query_symlink,
Sachin Prabhu5b23c972016-07-11 16:53:20 +01005367 .query_mf_symlink = smb3_query_mf_symlink,
5368 .create_mf_symlink = smb3_create_mf_symlink,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005369 .open = smb2_open_file,
5370 .set_fid = smb2_set_fid,
5371 .close = smb2_close_file,
5372 .flush = smb2_flush_file,
5373 .async_readv = smb2_async_readv,
5374 .async_writev = smb2_async_writev,
5375 .sync_read = smb2_sync_read,
5376 .sync_write = smb2_sync_write,
5377 .query_dir_first = smb2_query_dir_first,
5378 .query_dir_next = smb2_query_dir_next,
5379 .close_dir = smb2_close_dir,
5380 .calc_smb_size = smb2_calc_size,
5381 .is_status_pending = smb2_is_status_pending,
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07005382 .is_session_expired = smb2_is_session_expired,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005383 .oplock_response = smb2_oplock_response,
5384 .queryfs = smb2_queryfs,
5385 .mand_lock = smb2_mand_lock,
5386 .mand_unlock_range = smb2_unlock_range,
5387 .push_mand_locks = smb2_push_mandatory_locks,
5388 .get_lease_key = smb2_get_lease_key,
5389 .set_lease_key = smb2_set_lease_key,
5390 .new_lease_key = smb2_new_lease_key,
5391 .calc_signature = smb2_calc_signature,
5392 .is_read_op = smb2_is_read_op,
5393 .set_oplock_level = smb2_set_oplock_level,
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04005394 .create_lease_buf = smb2_create_lease_buf,
Pavel Shilovskyb5c7cde2013-09-05 20:16:45 +04005395 .parse_lease_buf = smb2_parse_lease_buf,
Sachin Prabhu312bbc52017-04-04 02:12:04 -05005396 .copychunk_range = smb2_copychunk_range,
Pavel Shilovsky7f6c5002014-06-22 11:03:22 +04005397 .wp_retry_size = smb2_wp_retry_size,
Pavel Shilovsky52755802014-08-18 20:49:57 +04005398 .dir_needs_close = smb2_dir_needs_close,
Aurelien Aptel9d496402017-02-13 16:16:49 +01005399 .get_dfs_refer = smb2_get_dfs_refer,
Sachin Prabhuef65aae2017-01-18 15:35:57 +05305400 .select_sectype = smb2_select_sectype,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005401#ifdef CONFIG_CIFS_XATTR
5402 .query_all_EAs = smb2_query_eas,
Ronnie Sahlberg55175542017-08-24 11:24:56 +10005403 .set_EA = smb2_set_ea,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005404#endif /* CIFS_XATTR */
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05005405 .get_acl = get_smb2_acl,
5406 .get_acl_by_fid = get_smb2_acl_by_fid,
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05005407 .set_acl = set_smb2_acl,
Ronnie Sahlberg8ce79ec2018-06-01 10:53:08 +10005408 .next_header = smb2_next_header,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05005409 .ioctl_query_info = smb2_ioctl_query_info,
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005410 .make_node = smb2_make_node,
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10005411 .fiemap = smb3_fiemap,
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10005412 .llseek = smb3_llseek,
Rohith Surabattula8e670f72020-09-18 05:37:28 +00005413 .is_status_io_timeout = smb2_is_status_io_timeout,
Rohith Surabattula9e550b02021-02-16 10:40:45 +00005414 .is_network_name_deleted = smb2_is_network_name_deleted,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005415};
5416
Steve French1080ef72011-02-24 18:07:19 +00005417struct smb_version_operations smb21_operations = {
Pavel Shilovsky027e8ee2012-09-19 06:22:43 -07005418 .compare_fids = smb2_compare_fids,
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +04005419 .setup_request = smb2_setup_request,
Pavel Shilovskyc95b8ee2012-07-11 14:45:28 +04005420 .setup_async_request = smb2_setup_async_request,
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +04005421 .check_receive = smb2_check_receive,
Pavel Shilovsky28ea5292012-05-23 16:18:00 +04005422 .add_credits = smb2_add_credits,
5423 .set_credits = smb2_set_credits,
5424 .get_credits_field = smb2_get_credits_field,
5425 .get_credits = smb2_get_credits,
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +04005426 .wait_mtu_credits = smb2_wait_mtu_credits,
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -08005427 .adjust_credits = smb2_adjust_credits,
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +04005428 .get_next_mid = smb2_get_next_mid,
Pavel Shilovskyc781af72019-03-04 14:02:50 -08005429 .revert_current_mid = smb2_revert_current_mid,
Pavel Shilovsky09a47072012-09-18 16:20:29 -07005430 .read_data_offset = smb2_read_data_offset,
5431 .read_data_length = smb2_read_data_length,
5432 .map_error = map_smb2_to_linux_error,
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +04005433 .find_mid = smb2_find_mid,
5434 .check_message = smb2_check_message,
5435 .dump_detail = smb2_dump_detail,
Pavel Shilovskyd60622e2012-05-28 15:19:39 +04005436 .clear_stats = smb2_clear_stats,
5437 .print_stats = smb2_print_stats,
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07005438 .is_oplock_break = smb2_is_valid_oplock_break,
Sachin Prabhu38bd4902017-03-03 15:41:38 -08005439 .handle_cancelled_mid = smb2_handle_cancelled_mid,
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07005440 .downgrade_oplock = smb2_downgrade_oplock,
Pavel Shilovskyec2e4522011-12-27 16:12:43 +04005441 .need_neg = smb2_need_neg,
5442 .negotiate = smb2_negotiate,
Pavel Shilovsky3a3bab52012-09-18 16:20:28 -07005443 .negotiate_wsize = smb2_negotiate_wsize,
5444 .negotiate_rsize = smb2_negotiate_rsize,
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +04005445 .sess_setup = SMB2_sess_setup,
5446 .logoff = SMB2_logoff,
Pavel Shilovskyfaaf9462011-12-27 16:04:00 +04005447 .tree_connect = SMB2_tcon,
5448 .tree_disconnect = SMB2_tdis,
Steve French34f62642013-10-09 02:07:00 -05005449 .qfs_tcon = smb2_qfs_tcon,
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04005450 .is_path_accessible = smb2_is_path_accessible,
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04005451 .can_echo = smb2_can_echo,
5452 .echo = SMB2_echo,
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04005453 .query_path_info = smb2_query_path_info,
5454 .get_srv_inum = smb2_get_srv_inum,
Pavel Shilovskyb7546bc2012-09-18 16:20:27 -07005455 .query_file_info = smb2_query_file_info,
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07005456 .set_path_size = smb2_set_path_size,
5457 .set_file_size = smb2_set_file_size,
Pavel Shilovsky1feeaac2012-09-18 16:20:32 -07005458 .set_file_info = smb2_set_file_info,
Steve French64a5cfa2013-10-14 15:31:32 -05005459 .set_compression = smb2_set_compression,
Pavel Shilovskya0e73182011-07-19 12:56:37 +04005460 .mkdir = smb2_mkdir,
5461 .mkdir_setinfo = smb2_mkdir_setinfo,
Pavel Shilovsky1a500f02012-07-10 16:14:38 +04005462 .rmdir = smb2_rmdir,
Pavel Shilovskycbe6f432012-09-18 16:20:25 -07005463 .unlink = smb2_unlink,
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07005464 .rename = smb2_rename_path,
Pavel Shilovsky568798c2012-09-18 16:20:31 -07005465 .create_hardlink = smb2_create_hardlink,
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04005466 .query_symlink = smb2_query_symlink,
Steve Frenchc22870e2014-09-16 07:18:19 -05005467 .query_mf_symlink = smb3_query_mf_symlink,
Steve French5ab97572014-09-15 04:49:28 -05005468 .create_mf_symlink = smb3_create_mf_symlink,
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07005469 .open = smb2_open_file,
5470 .set_fid = smb2_set_fid,
5471 .close = smb2_close_file,
Pavel Shilovsky7a5cfb12012-09-18 16:20:28 -07005472 .flush = smb2_flush_file,
Pavel Shilovsky09a47072012-09-18 16:20:29 -07005473 .async_readv = smb2_async_readv,
Pavel Shilovsky33319142012-09-18 16:20:29 -07005474 .async_writev = smb2_async_writev,
Pavel Shilovskyd8e05032012-09-18 16:20:30 -07005475 .sync_read = smb2_sync_read,
Pavel Shilovsky009d3442012-09-18 16:20:30 -07005476 .sync_write = smb2_sync_write,
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07005477 .query_dir_first = smb2_query_dir_first,
5478 .query_dir_next = smb2_query_dir_next,
5479 .close_dir = smb2_close_dir,
5480 .calc_smb_size = smb2_calc_size,
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07005481 .is_status_pending = smb2_is_status_pending,
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07005482 .is_session_expired = smb2_is_session_expired,
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07005483 .oplock_response = smb2_oplock_response,
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07005484 .queryfs = smb2_queryfs,
Pavel Shilovskyf7ba7fe2012-09-19 06:22:43 -07005485 .mand_lock = smb2_mand_lock,
5486 .mand_unlock_range = smb2_unlock_range,
Pavel Shilovskyb1407992012-09-19 06:22:44 -07005487 .push_mand_locks = smb2_push_mandatory_locks,
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -07005488 .get_lease_key = smb2_get_lease_key,
5489 .set_lease_key = smb2_set_lease_key,
5490 .new_lease_key = smb2_new_lease_key,
Steve French38107d42012-12-08 22:08:06 -06005491 .calc_signature = smb2_calc_signature,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005492 .is_read_op = smb21_is_read_op,
5493 .set_oplock_level = smb21_set_oplock_level,
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04005494 .create_lease_buf = smb2_create_lease_buf,
Pavel Shilovskyb5c7cde2013-09-05 20:16:45 +04005495 .parse_lease_buf = smb2_parse_lease_buf,
Sachin Prabhu312bbc52017-04-04 02:12:04 -05005496 .copychunk_range = smb2_copychunk_range,
Pavel Shilovsky7f6c5002014-06-22 11:03:22 +04005497 .wp_retry_size = smb2_wp_retry_size,
Pavel Shilovsky52755802014-08-18 20:49:57 +04005498 .dir_needs_close = smb2_dir_needs_close,
Steve French834170c2016-09-30 21:14:26 -05005499 .enum_snapshots = smb3_enum_snapshots,
Steve French2c6251a2020-02-12 22:37:08 -06005500 .notify = smb3_notify,
Aurelien Aptel9d496402017-02-13 16:16:49 +01005501 .get_dfs_refer = smb2_get_dfs_refer,
Sachin Prabhuef65aae2017-01-18 15:35:57 +05305502 .select_sectype = smb2_select_sectype,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005503#ifdef CONFIG_CIFS_XATTR
5504 .query_all_EAs = smb2_query_eas,
Ronnie Sahlberg55175542017-08-24 11:24:56 +10005505 .set_EA = smb2_set_ea,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005506#endif /* CIFS_XATTR */
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05005507 .get_acl = get_smb2_acl,
5508 .get_acl_by_fid = get_smb2_acl_by_fid,
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05005509 .set_acl = set_smb2_acl,
Ronnie Sahlberg8ce79ec2018-06-01 10:53:08 +10005510 .next_header = smb2_next_header,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05005511 .ioctl_query_info = smb2_ioctl_query_info,
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005512 .make_node = smb2_make_node,
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10005513 .fiemap = smb3_fiemap,
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10005514 .llseek = smb3_llseek,
Rohith Surabattula8e670f72020-09-18 05:37:28 +00005515 .is_status_io_timeout = smb2_is_status_io_timeout,
Rohith Surabattula9e550b02021-02-16 10:40:45 +00005516 .is_network_name_deleted = smb2_is_network_name_deleted,
Steve French38107d42012-12-08 22:08:06 -06005517};
5518
Steve French38107d42012-12-08 22:08:06 -06005519struct smb_version_operations smb30_operations = {
5520 .compare_fids = smb2_compare_fids,
5521 .setup_request = smb2_setup_request,
5522 .setup_async_request = smb2_setup_async_request,
5523 .check_receive = smb2_check_receive,
5524 .add_credits = smb2_add_credits,
5525 .set_credits = smb2_set_credits,
5526 .get_credits_field = smb2_get_credits_field,
5527 .get_credits = smb2_get_credits,
Pavel Shilovskycb7e9ea2014-06-05 19:03:27 +04005528 .wait_mtu_credits = smb2_wait_mtu_credits,
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -08005529 .adjust_credits = smb2_adjust_credits,
Steve French38107d42012-12-08 22:08:06 -06005530 .get_next_mid = smb2_get_next_mid,
Pavel Shilovskyc781af72019-03-04 14:02:50 -08005531 .revert_current_mid = smb2_revert_current_mid,
Steve French38107d42012-12-08 22:08:06 -06005532 .read_data_offset = smb2_read_data_offset,
5533 .read_data_length = smb2_read_data_length,
5534 .map_error = map_smb2_to_linux_error,
5535 .find_mid = smb2_find_mid,
5536 .check_message = smb2_check_message,
5537 .dump_detail = smb2_dump_detail,
5538 .clear_stats = smb2_clear_stats,
5539 .print_stats = smb2_print_stats,
Steve French769ee6a2013-06-19 14:15:30 -05005540 .dump_share_caps = smb2_dump_share_caps,
Steve French38107d42012-12-08 22:08:06 -06005541 .is_oplock_break = smb2_is_valid_oplock_break,
Sachin Prabhu38bd4902017-03-03 15:41:38 -08005542 .handle_cancelled_mid = smb2_handle_cancelled_mid,
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07005543 .downgrade_oplock = smb3_downgrade_oplock,
Steve French38107d42012-12-08 22:08:06 -06005544 .need_neg = smb2_need_neg,
5545 .negotiate = smb2_negotiate,
Steve French3d621232018-09-25 15:33:47 -05005546 .negotiate_wsize = smb3_negotiate_wsize,
5547 .negotiate_rsize = smb3_negotiate_rsize,
Steve French38107d42012-12-08 22:08:06 -06005548 .sess_setup = SMB2_sess_setup,
5549 .logoff = SMB2_logoff,
5550 .tree_connect = SMB2_tcon,
5551 .tree_disconnect = SMB2_tdis,
Steven Frenchaf6a12e2013-10-09 20:55:53 -05005552 .qfs_tcon = smb3_qfs_tcon,
Steve French38107d42012-12-08 22:08:06 -06005553 .is_path_accessible = smb2_is_path_accessible,
5554 .can_echo = smb2_can_echo,
5555 .echo = SMB2_echo,
5556 .query_path_info = smb2_query_path_info,
Steve French2e4564b2020-10-22 22:03:14 -05005557 /* WSL tags introduced long after smb2.1, enable for SMB3, 3.11 only */
5558 .query_reparse_tag = smb2_query_reparse_tag,
Steve French38107d42012-12-08 22:08:06 -06005559 .get_srv_inum = smb2_get_srv_inum,
5560 .query_file_info = smb2_query_file_info,
5561 .set_path_size = smb2_set_path_size,
5562 .set_file_size = smb2_set_file_size,
5563 .set_file_info = smb2_set_file_info,
Steve French64a5cfa2013-10-14 15:31:32 -05005564 .set_compression = smb2_set_compression,
Steve French38107d42012-12-08 22:08:06 -06005565 .mkdir = smb2_mkdir,
5566 .mkdir_setinfo = smb2_mkdir_setinfo,
5567 .rmdir = smb2_rmdir,
5568 .unlink = smb2_unlink,
5569 .rename = smb2_rename_path,
5570 .create_hardlink = smb2_create_hardlink,
Pavel Shilovskyb42bf882013-08-14 19:25:21 +04005571 .query_symlink = smb2_query_symlink,
Steve Frenchc22870e2014-09-16 07:18:19 -05005572 .query_mf_symlink = smb3_query_mf_symlink,
Steve French5ab97572014-09-15 04:49:28 -05005573 .create_mf_symlink = smb3_create_mf_symlink,
Steve French38107d42012-12-08 22:08:06 -06005574 .open = smb2_open_file,
5575 .set_fid = smb2_set_fid,
5576 .close = smb2_close_file,
Steve French43f8a6a2019-12-02 21:46:54 -06005577 .close_getattr = smb2_close_getattr,
Steve French38107d42012-12-08 22:08:06 -06005578 .flush = smb2_flush_file,
5579 .async_readv = smb2_async_readv,
5580 .async_writev = smb2_async_writev,
5581 .sync_read = smb2_sync_read,
5582 .sync_write = smb2_sync_write,
5583 .query_dir_first = smb2_query_dir_first,
5584 .query_dir_next = smb2_query_dir_next,
5585 .close_dir = smb2_close_dir,
5586 .calc_smb_size = smb2_calc_size,
5587 .is_status_pending = smb2_is_status_pending,
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07005588 .is_session_expired = smb2_is_session_expired,
Steve French38107d42012-12-08 22:08:06 -06005589 .oplock_response = smb2_oplock_response,
5590 .queryfs = smb2_queryfs,
5591 .mand_lock = smb2_mand_lock,
5592 .mand_unlock_range = smb2_unlock_range,
5593 .push_mand_locks = smb2_push_mandatory_locks,
5594 .get_lease_key = smb2_get_lease_key,
5595 .set_lease_key = smb2_set_lease_key,
5596 .new_lease_key = smb2_new_lease_key,
Steve French373512e2015-12-18 13:05:30 -06005597 .generate_signingkey = generate_smb30signingkey,
Steve French38107d42012-12-08 22:08:06 -06005598 .calc_signature = smb3_calc_signature,
Steve Frenchb3152e22015-06-24 03:17:02 -05005599 .set_integrity = smb3_set_integrity,
Pavel Shilovsky53ef1012013-09-05 16:11:28 +04005600 .is_read_op = smb21_is_read_op,
Pavel Shilovsky42873b02013-09-05 21:30:16 +04005601 .set_oplock_level = smb3_set_oplock_level,
Pavel Shilovskyf0473902013-09-04 13:44:05 +04005602 .create_lease_buf = smb3_create_lease_buf,
5603 .parse_lease_buf = smb3_parse_lease_buf,
Sachin Prabhu312bbc52017-04-04 02:12:04 -05005604 .copychunk_range = smb2_copychunk_range,
Steve Frenchca9e7a12015-10-01 21:40:10 -05005605 .duplicate_extents = smb2_duplicate_extents,
Steve Frenchff1c0382013-11-19 23:44:46 -06005606 .validate_negotiate = smb3_validate_negotiate,
Pavel Shilovsky7f6c5002014-06-22 11:03:22 +04005607 .wp_retry_size = smb2_wp_retry_size,
Pavel Shilovsky52755802014-08-18 20:49:57 +04005608 .dir_needs_close = smb2_dir_needs_close,
Steve French31742c52014-08-17 08:38:47 -05005609 .fallocate = smb3_fallocate,
Steve French834170c2016-09-30 21:14:26 -05005610 .enum_snapshots = smb3_enum_snapshots,
Steve Frenchd26c2dd2020-02-06 06:00:14 -06005611 .notify = smb3_notify,
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07005612 .init_transform_rq = smb3_init_transform_rq,
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005613 .is_transform_hdr = smb3_is_transform_hdr,
5614 .receive_transform = smb3_receive_transform,
Aurelien Aptel9d496402017-02-13 16:16:49 +01005615 .get_dfs_refer = smb2_get_dfs_refer,
Sachin Prabhuef65aae2017-01-18 15:35:57 +05305616 .select_sectype = smb2_select_sectype,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005617#ifdef CONFIG_CIFS_XATTR
5618 .query_all_EAs = smb2_query_eas,
Ronnie Sahlberg55175542017-08-24 11:24:56 +10005619 .set_EA = smb2_set_ea,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005620#endif /* CIFS_XATTR */
Shirish Pargaonkar2f1afe22017-06-22 22:52:05 -05005621 .get_acl = get_smb2_acl,
5622 .get_acl_by_fid = get_smb2_acl_by_fid,
Shirish Pargaonkar366ed842017-06-28 22:37:32 -05005623 .set_acl = set_smb2_acl,
Ronnie Sahlberg8ce79ec2018-06-01 10:53:08 +10005624 .next_header = smb2_next_header,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05005625 .ioctl_query_info = smb2_ioctl_query_info,
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005626 .make_node = smb2_make_node,
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10005627 .fiemap = smb3_fiemap,
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10005628 .llseek = smb3_llseek,
Rohith Surabattula8e670f72020-09-18 05:37:28 +00005629 .is_status_io_timeout = smb2_is_status_io_timeout,
Rohith Surabattula9e550b02021-02-16 10:40:45 +00005630 .is_network_name_deleted = smb2_is_network_name_deleted,
Steve French1080ef72011-02-24 18:07:19 +00005631};
5632
Steve Frenchaab18932015-06-23 23:37:11 -05005633struct smb_version_operations smb311_operations = {
5634 .compare_fids = smb2_compare_fids,
5635 .setup_request = smb2_setup_request,
5636 .setup_async_request = smb2_setup_async_request,
5637 .check_receive = smb2_check_receive,
5638 .add_credits = smb2_add_credits,
5639 .set_credits = smb2_set_credits,
5640 .get_credits_field = smb2_get_credits_field,
5641 .get_credits = smb2_get_credits,
5642 .wait_mtu_credits = smb2_wait_mtu_credits,
Pavel Shilovsky9a1c67e2019-01-23 18:15:52 -08005643 .adjust_credits = smb2_adjust_credits,
Steve Frenchaab18932015-06-23 23:37:11 -05005644 .get_next_mid = smb2_get_next_mid,
Pavel Shilovskyc781af72019-03-04 14:02:50 -08005645 .revert_current_mid = smb2_revert_current_mid,
Steve Frenchaab18932015-06-23 23:37:11 -05005646 .read_data_offset = smb2_read_data_offset,
5647 .read_data_length = smb2_read_data_length,
5648 .map_error = map_smb2_to_linux_error,
5649 .find_mid = smb2_find_mid,
5650 .check_message = smb2_check_message,
5651 .dump_detail = smb2_dump_detail,
5652 .clear_stats = smb2_clear_stats,
5653 .print_stats = smb2_print_stats,
5654 .dump_share_caps = smb2_dump_share_caps,
5655 .is_oplock_break = smb2_is_valid_oplock_break,
Sachin Prabhu38bd4902017-03-03 15:41:38 -08005656 .handle_cancelled_mid = smb2_handle_cancelled_mid,
Pavel Shilovsky9bd45402019-10-29 16:51:19 -07005657 .downgrade_oplock = smb3_downgrade_oplock,
Steve Frenchaab18932015-06-23 23:37:11 -05005658 .need_neg = smb2_need_neg,
5659 .negotiate = smb2_negotiate,
Steve French3d621232018-09-25 15:33:47 -05005660 .negotiate_wsize = smb3_negotiate_wsize,
5661 .negotiate_rsize = smb3_negotiate_rsize,
Steve Frenchaab18932015-06-23 23:37:11 -05005662 .sess_setup = SMB2_sess_setup,
5663 .logoff = SMB2_logoff,
5664 .tree_connect = SMB2_tcon,
5665 .tree_disconnect = SMB2_tdis,
5666 .qfs_tcon = smb3_qfs_tcon,
5667 .is_path_accessible = smb2_is_path_accessible,
5668 .can_echo = smb2_can_echo,
5669 .echo = SMB2_echo,
5670 .query_path_info = smb2_query_path_info,
Steve French2e4564b2020-10-22 22:03:14 -05005671 .query_reparse_tag = smb2_query_reparse_tag,
Steve Frenchaab18932015-06-23 23:37:11 -05005672 .get_srv_inum = smb2_get_srv_inum,
5673 .query_file_info = smb2_query_file_info,
5674 .set_path_size = smb2_set_path_size,
5675 .set_file_size = smb2_set_file_size,
5676 .set_file_info = smb2_set_file_info,
5677 .set_compression = smb2_set_compression,
5678 .mkdir = smb2_mkdir,
5679 .mkdir_setinfo = smb2_mkdir_setinfo,
Steve Frenchbea851b2018-06-14 21:56:32 -05005680 .posix_mkdir = smb311_posix_mkdir,
Steve Frenchaab18932015-06-23 23:37:11 -05005681 .rmdir = smb2_rmdir,
5682 .unlink = smb2_unlink,
5683 .rename = smb2_rename_path,
5684 .create_hardlink = smb2_create_hardlink,
5685 .query_symlink = smb2_query_symlink,
5686 .query_mf_symlink = smb3_query_mf_symlink,
5687 .create_mf_symlink = smb3_create_mf_symlink,
5688 .open = smb2_open_file,
5689 .set_fid = smb2_set_fid,
5690 .close = smb2_close_file,
Steve French43f8a6a2019-12-02 21:46:54 -06005691 .close_getattr = smb2_close_getattr,
Steve Frenchaab18932015-06-23 23:37:11 -05005692 .flush = smb2_flush_file,
5693 .async_readv = smb2_async_readv,
5694 .async_writev = smb2_async_writev,
5695 .sync_read = smb2_sync_read,
5696 .sync_write = smb2_sync_write,
5697 .query_dir_first = smb2_query_dir_first,
5698 .query_dir_next = smb2_query_dir_next,
5699 .close_dir = smb2_close_dir,
5700 .calc_smb_size = smb2_calc_size,
5701 .is_status_pending = smb2_is_status_pending,
Pavel Shilovsky511c54a2017-07-08 14:32:00 -07005702 .is_session_expired = smb2_is_session_expired,
Steve Frenchaab18932015-06-23 23:37:11 -05005703 .oplock_response = smb2_oplock_response,
Steve French2d304212018-06-24 23:28:12 -05005704 .queryfs = smb311_queryfs,
Steve Frenchaab18932015-06-23 23:37:11 -05005705 .mand_lock = smb2_mand_lock,
5706 .mand_unlock_range = smb2_unlock_range,
5707 .push_mand_locks = smb2_push_mandatory_locks,
5708 .get_lease_key = smb2_get_lease_key,
5709 .set_lease_key = smb2_set_lease_key,
5710 .new_lease_key = smb2_new_lease_key,
Steve French373512e2015-12-18 13:05:30 -06005711 .generate_signingkey = generate_smb311signingkey,
Steve Frenchaab18932015-06-23 23:37:11 -05005712 .calc_signature = smb3_calc_signature,
Steve Frenchb3152e22015-06-24 03:17:02 -05005713 .set_integrity = smb3_set_integrity,
Steve Frenchaab18932015-06-23 23:37:11 -05005714 .is_read_op = smb21_is_read_op,
5715 .set_oplock_level = smb3_set_oplock_level,
5716 .create_lease_buf = smb3_create_lease_buf,
5717 .parse_lease_buf = smb3_parse_lease_buf,
Sachin Prabhu312bbc52017-04-04 02:12:04 -05005718 .copychunk_range = smb2_copychunk_range,
Steve French02b16662015-06-27 21:18:36 -07005719 .duplicate_extents = smb2_duplicate_extents,
Steve Frenchaab18932015-06-23 23:37:11 -05005720/* .validate_negotiate = smb3_validate_negotiate, */ /* not used in 3.11 */
5721 .wp_retry_size = smb2_wp_retry_size,
5722 .dir_needs_close = smb2_dir_needs_close,
5723 .fallocate = smb3_fallocate,
Steve French834170c2016-09-30 21:14:26 -05005724 .enum_snapshots = smb3_enum_snapshots,
Steve Frenchd26c2dd2020-02-06 06:00:14 -06005725 .notify = smb3_notify,
Pavel Shilovsky026e93d2016-11-03 16:47:37 -07005726 .init_transform_rq = smb3_init_transform_rq,
Pavel Shilovsky4326ed22016-11-17 15:24:46 -08005727 .is_transform_hdr = smb3_is_transform_hdr,
5728 .receive_transform = smb3_receive_transform,
Aurelien Aptel9d496402017-02-13 16:16:49 +01005729 .get_dfs_refer = smb2_get_dfs_refer,
Sachin Prabhuef65aae2017-01-18 15:35:57 +05305730 .select_sectype = smb2_select_sectype,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005731#ifdef CONFIG_CIFS_XATTR
5732 .query_all_EAs = smb2_query_eas,
Ronnie Sahlberg55175542017-08-24 11:24:56 +10005733 .set_EA = smb2_set_ea,
Ronnie Sahlberg95907fe2017-08-24 11:24:55 +10005734#endif /* CIFS_XATTR */
Ronnie Sahlbergc1777df2018-08-10 11:03:55 +10005735 .get_acl = get_smb2_acl,
5736 .get_acl_by_fid = get_smb2_acl_by_fid,
5737 .set_acl = set_smb2_acl,
Ronnie Sahlberg8ce79ec2018-06-01 10:53:08 +10005738 .next_header = smb2_next_header,
Ronnie Sahlbergf5b05d62018-10-07 19:19:58 -05005739 .ioctl_query_info = smb2_ioctl_query_info,
Aurelien Aptelc847dcc2019-03-14 00:29:17 -05005740 .make_node = smb2_make_node,
Ronnie Sahlberg2f3ebab2019-04-25 16:45:29 +10005741 .fiemap = smb3_fiemap,
Ronnie Sahlbergdece44e2019-05-15 07:17:02 +10005742 .llseek = smb3_llseek,
Rohith Surabattula8e670f72020-09-18 05:37:28 +00005743 .is_status_io_timeout = smb2_is_status_io_timeout,
Rohith Surabattula9e550b02021-02-16 10:40:45 +00005744 .is_network_name_deleted = smb2_is_network_name_deleted,
Steve Frenchaab18932015-06-23 23:37:11 -05005745};
Steve Frenchaab18932015-06-23 23:37:11 -05005746
Steve Frenchdd446b12012-11-28 23:21:06 -06005747struct smb_version_values smb20_values = {
5748 .version_string = SMB20_VERSION_STRING,
5749 .protocol_id = SMB20_PROT_ID,
5750 .req_capabilities = 0, /* MBZ */
5751 .large_lock_type = 0,
5752 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5753 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5754 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005755 .header_size = sizeof(struct smb2_sync_hdr),
5756 .header_preamble_size = 0,
Steve Frenchdd446b12012-11-28 23:21:06 -06005757 .max_header_size = MAX_SMB2_HDR_SIZE,
5758 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5759 .lock_cmd = SMB2_LOCK,
5760 .cap_unix = 0,
5761 .cap_nt_find = SMB2_NT_FIND,
5762 .cap_large_files = SMB2_LARGE_FILES,
Jeff Layton502858822013-06-27 12:45:00 -04005763 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5764 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04005765 .create_lease_size = sizeof(struct create_lease),
Steve Frenchdd446b12012-11-28 23:21:06 -06005766};
5767
Steve French1080ef72011-02-24 18:07:19 +00005768struct smb_version_values smb21_values = {
5769 .version_string = SMB21_VERSION_STRING,
Steve Frenche4aa25e2012-10-01 12:26:22 -05005770 .protocol_id = SMB21_PROT_ID,
5771 .req_capabilities = 0, /* MBZ on negotiate req until SMB3 dialect */
5772 .large_lock_type = 0,
5773 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5774 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5775 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005776 .header_size = sizeof(struct smb2_sync_hdr),
5777 .header_preamble_size = 0,
Steve Frenche4aa25e2012-10-01 12:26:22 -05005778 .max_header_size = MAX_SMB2_HDR_SIZE,
5779 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5780 .lock_cmd = SMB2_LOCK,
5781 .cap_unix = 0,
5782 .cap_nt_find = SMB2_NT_FIND,
5783 .cap_large_files = SMB2_LARGE_FILES,
Jeff Layton502858822013-06-27 12:45:00 -04005784 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5785 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
Pavel Shilovskya41a28b2013-09-04 13:07:41 +04005786 .create_lease_size = sizeof(struct create_lease),
Steve Frenche4aa25e2012-10-01 12:26:22 -05005787};
5788
Steve French9764c022017-09-17 10:41:35 -05005789struct smb_version_values smb3any_values = {
5790 .version_string = SMB3ANY_VERSION_STRING,
5791 .protocol_id = SMB302_PROT_ID, /* doesn't matter, send protocol array */
Steve Frenchf8015682018-08-31 15:12:10 -05005792 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
Steve French9764c022017-09-17 10:41:35 -05005793 .large_lock_type = 0,
5794 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5795 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5796 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005797 .header_size = sizeof(struct smb2_sync_hdr),
5798 .header_preamble_size = 0,
Steve French9764c022017-09-17 10:41:35 -05005799 .max_header_size = MAX_SMB2_HDR_SIZE,
5800 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5801 .lock_cmd = SMB2_LOCK,
5802 .cap_unix = 0,
5803 .cap_nt_find = SMB2_NT_FIND,
5804 .cap_large_files = SMB2_LARGE_FILES,
5805 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5806 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5807 .create_lease_size = sizeof(struct create_lease_v2),
5808};
5809
5810struct smb_version_values smbdefault_values = {
5811 .version_string = SMBDEFAULT_VERSION_STRING,
5812 .protocol_id = SMB302_PROT_ID, /* doesn't matter, send protocol array */
Steve Frenchf8015682018-08-31 15:12:10 -05005813 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
Steve French9764c022017-09-17 10:41:35 -05005814 .large_lock_type = 0,
5815 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5816 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5817 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005818 .header_size = sizeof(struct smb2_sync_hdr),
5819 .header_preamble_size = 0,
Steve French9764c022017-09-17 10:41:35 -05005820 .max_header_size = MAX_SMB2_HDR_SIZE,
5821 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5822 .lock_cmd = SMB2_LOCK,
5823 .cap_unix = 0,
5824 .cap_nt_find = SMB2_NT_FIND,
5825 .cap_large_files = SMB2_LARGE_FILES,
5826 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5827 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5828 .create_lease_size = sizeof(struct create_lease_v2),
5829};
5830
Steve Frenche4aa25e2012-10-01 12:26:22 -05005831struct smb_version_values smb30_values = {
5832 .version_string = SMB30_VERSION_STRING,
5833 .protocol_id = SMB30_PROT_ID,
Steve Frenchf8015682018-08-31 15:12:10 -05005834 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
Pavel Shilovsky027e8ee2012-09-19 06:22:43 -07005835 .large_lock_type = 0,
5836 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5837 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5838 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005839 .header_size = sizeof(struct smb2_sync_hdr),
5840 .header_preamble_size = 0,
Pavel Shilovsky093b2bd2011-06-08 15:51:07 +04005841 .max_header_size = MAX_SMB2_HDR_SIZE,
Pavel Shilovsky09a47072012-09-18 16:20:29 -07005842 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
Pavel Shilovsky2dc7e1c2011-12-26 22:53:34 +04005843 .lock_cmd = SMB2_LOCK,
Pavel Shilovsky29e20f92012-07-13 13:58:14 +04005844 .cap_unix = 0,
5845 .cap_nt_find = SMB2_NT_FIND,
5846 .cap_large_files = SMB2_LARGE_FILES,
Jeff Layton502858822013-06-27 12:45:00 -04005847 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5848 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
Pavel Shilovskyf0473902013-09-04 13:44:05 +04005849 .create_lease_size = sizeof(struct create_lease_v2),
Steve French1080ef72011-02-24 18:07:19 +00005850};
Steve French20b6d8b2013-06-12 22:48:41 -05005851
5852struct smb_version_values smb302_values = {
5853 .version_string = SMB302_VERSION_STRING,
5854 .protocol_id = SMB302_PROT_ID,
Steve Frenchf8015682018-08-31 15:12:10 -05005855 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
Steve French20b6d8b2013-06-12 22:48:41 -05005856 .large_lock_type = 0,
5857 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5858 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5859 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005860 .header_size = sizeof(struct smb2_sync_hdr),
5861 .header_preamble_size = 0,
Steve French20b6d8b2013-06-12 22:48:41 -05005862 .max_header_size = MAX_SMB2_HDR_SIZE,
5863 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5864 .lock_cmd = SMB2_LOCK,
5865 .cap_unix = 0,
5866 .cap_nt_find = SMB2_NT_FIND,
5867 .cap_large_files = SMB2_LARGE_FILES,
Jeff Layton502858822013-06-27 12:45:00 -04005868 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5869 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
Pavel Shilovskyf0473902013-09-04 13:44:05 +04005870 .create_lease_size = sizeof(struct create_lease_v2),
Steve French20b6d8b2013-06-12 22:48:41 -05005871};
Steve French5f7fbf72014-12-17 22:52:58 -06005872
Steve French5f7fbf72014-12-17 22:52:58 -06005873struct smb_version_values smb311_values = {
5874 .version_string = SMB311_VERSION_STRING,
5875 .protocol_id = SMB311_PROT_ID,
Steve Frenchf8015682018-08-31 15:12:10 -05005876 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
Steve French5f7fbf72014-12-17 22:52:58 -06005877 .large_lock_type = 0,
5878 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE_LOCK,
5879 .shared_lock_type = SMB2_LOCKFLAG_SHARED_LOCK,
5880 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
Ronnie Sahlberg977b6172018-06-01 10:53:02 +10005881 .header_size = sizeof(struct smb2_sync_hdr),
5882 .header_preamble_size = 0,
Steve French5f7fbf72014-12-17 22:52:58 -06005883 .max_header_size = MAX_SMB2_HDR_SIZE,
5884 .read_rsp_size = sizeof(struct smb2_read_rsp) - 1,
5885 .lock_cmd = SMB2_LOCK,
5886 .cap_unix = 0,
5887 .cap_nt_find = SMB2_NT_FIND,
5888 .cap_large_files = SMB2_LARGE_FILES,
5889 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5890 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5891 .create_lease_size = sizeof(struct create_lease_v2),
5892};