blob: fa7b0fa72bb360d7133b13b05788e1e591faf547 [file] [log] [blame]
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001/*
2 * fs/cifs/file.c
3 *
4 * vfs operations that deal with files
5 *
6 * Copyright (C) International Business Machines Corp., 2002,2010
7 * Author(s): Steve French (sfrench@us.ibm.com)
8 * Jeremy Allison (jra@samba.org)
9 *
10 * This library is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU Lesser General Public License as published
12 * by the Free Software Foundation; either version 2.1 of the License, or
13 * (at your option) any later version.
14 *
15 * This library is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
18 * the GNU Lesser General Public License for more details.
19 *
20 * You should have received a copy of the GNU Lesser General Public License
21 * along with this library; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 */
24#include <linux/fs.h>
25#include <linux/backing-dev.h>
26#include <linux/stat.h>
27#include <linux/fcntl.h>
28#include <linux/pagemap.h>
29#include <linux/pagevec.h>
30#include <linux/writeback.h>
31#include <linux/task_io_accounting_ops.h>
32#include <linux/delay.h>
33#include <linux/mount.h>
34#include <linux/slab.h>
35#include <linux/swap.h>
David Brazdil0f672f62019-12-10 10:32:29 +000036#include <linux/mm.h>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000037#include <asm/div64.h>
38#include "cifsfs.h"
39#include "cifspdu.h"
40#include "cifsglob.h"
41#include "cifsproto.h"
42#include "cifs_unicode.h"
43#include "cifs_debug.h"
44#include "cifs_fs_sb.h"
45#include "fscache.h"
46#include "smbdirect.h"
47
48static inline int cifs_convert_flags(unsigned int flags)
49{
50 if ((flags & O_ACCMODE) == O_RDONLY)
51 return GENERIC_READ;
52 else if ((flags & O_ACCMODE) == O_WRONLY)
53 return GENERIC_WRITE;
54 else if ((flags & O_ACCMODE) == O_RDWR) {
55 /* GENERIC_ALL is too much permission to request
56 can cause unnecessary access denied on create */
57 /* return GENERIC_ALL; */
58 return (GENERIC_READ | GENERIC_WRITE);
59 }
60
61 return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES |
62 FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA |
63 FILE_READ_DATA);
64}
65
66static u32 cifs_posix_convert_flags(unsigned int flags)
67{
68 u32 posix_flags = 0;
69
70 if ((flags & O_ACCMODE) == O_RDONLY)
71 posix_flags = SMB_O_RDONLY;
72 else if ((flags & O_ACCMODE) == O_WRONLY)
73 posix_flags = SMB_O_WRONLY;
74 else if ((flags & O_ACCMODE) == O_RDWR)
75 posix_flags = SMB_O_RDWR;
76
77 if (flags & O_CREAT) {
78 posix_flags |= SMB_O_CREAT;
79 if (flags & O_EXCL)
80 posix_flags |= SMB_O_EXCL;
81 } else if (flags & O_EXCL)
82 cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n",
83 current->comm, current->tgid);
84
85 if (flags & O_TRUNC)
86 posix_flags |= SMB_O_TRUNC;
87 /* be safe and imply O_SYNC for O_DSYNC */
88 if (flags & O_DSYNC)
89 posix_flags |= SMB_O_SYNC;
90 if (flags & O_DIRECTORY)
91 posix_flags |= SMB_O_DIRECTORY;
92 if (flags & O_NOFOLLOW)
93 posix_flags |= SMB_O_NOFOLLOW;
94 if (flags & O_DIRECT)
95 posix_flags |= SMB_O_DIRECT;
96
97 return posix_flags;
98}
99
100static inline int cifs_get_disposition(unsigned int flags)
101{
102 if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL))
103 return FILE_CREATE;
104 else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC))
105 return FILE_OVERWRITE_IF;
106 else if ((flags & O_CREAT) == O_CREAT)
107 return FILE_OPEN_IF;
108 else if ((flags & O_TRUNC) == O_TRUNC)
109 return FILE_OVERWRITE;
110 else
111 return FILE_OPEN;
112}
113
114int cifs_posix_open(char *full_path, struct inode **pinode,
115 struct super_block *sb, int mode, unsigned int f_flags,
116 __u32 *poplock, __u16 *pnetfid, unsigned int xid)
117{
118 int rc;
119 FILE_UNIX_BASIC_INFO *presp_data;
120 __u32 posix_flags = 0;
121 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
122 struct cifs_fattr fattr;
123 struct tcon_link *tlink;
124 struct cifs_tcon *tcon;
125
126 cifs_dbg(FYI, "posix open %s\n", full_path);
127
128 presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL);
129 if (presp_data == NULL)
130 return -ENOMEM;
131
132 tlink = cifs_sb_tlink(cifs_sb);
133 if (IS_ERR(tlink)) {
134 rc = PTR_ERR(tlink);
135 goto posix_open_ret;
136 }
137
138 tcon = tlink_tcon(tlink);
139 mode &= ~current_umask();
140
141 posix_flags = cifs_posix_convert_flags(f_flags);
142 rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data,
143 poplock, full_path, cifs_sb->local_nls,
144 cifs_remap(cifs_sb));
145 cifs_put_tlink(tlink);
146
147 if (rc)
148 goto posix_open_ret;
149
150 if (presp_data->Type == cpu_to_le32(-1))
151 goto posix_open_ret; /* open ok, caller does qpathinfo */
152
153 if (!pinode)
154 goto posix_open_ret; /* caller does not need info */
155
156 cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb);
157
158 /* get new inode and set it up */
159 if (*pinode == NULL) {
160 cifs_fill_uniqueid(sb, &fattr);
161 *pinode = cifs_iget(sb, &fattr);
162 if (!*pinode) {
163 rc = -ENOMEM;
164 goto posix_open_ret;
165 }
166 } else {
167 cifs_fattr_to_inode(*pinode, &fattr);
168 }
169
170posix_open_ret:
171 kfree(presp_data);
172 return rc;
173}
174
175static int
176cifs_nt_open(char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb,
177 struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock,
178 struct cifs_fid *fid, unsigned int xid)
179{
180 int rc;
181 int desired_access;
182 int disposition;
183 int create_options = CREATE_NOT_DIR;
184 FILE_ALL_INFO *buf;
185 struct TCP_Server_Info *server = tcon->ses->server;
186 struct cifs_open_parms oparms;
187
188 if (!server->ops->open)
189 return -ENOSYS;
190
191 desired_access = cifs_convert_flags(f_flags);
192
193/*********************************************************************
194 * open flag mapping table:
195 *
196 * POSIX Flag CIFS Disposition
197 * ---------- ----------------
198 * O_CREAT FILE_OPEN_IF
199 * O_CREAT | O_EXCL FILE_CREATE
200 * O_CREAT | O_TRUNC FILE_OVERWRITE_IF
201 * O_TRUNC FILE_OVERWRITE
202 * none of the above FILE_OPEN
203 *
204 * Note that there is not a direct match between disposition
205 * FILE_SUPERSEDE (ie create whether or not file exists although
206 * O_CREAT | O_TRUNC is similar but truncates the existing
207 * file rather than creating a new file as FILE_SUPERSEDE does
208 * (which uses the attributes / metadata passed in on open call)
209 *?
210 *? O_SYNC is a reasonable match to CIFS writethrough flag
211 *? and the read write flags match reasonably. O_LARGEFILE
212 *? is irrelevant because largefile support is always used
213 *? by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY,
214 * O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation
215 *********************************************************************/
216
217 disposition = cifs_get_disposition(f_flags);
218
219 /* BB pass O_SYNC flag through on file attributes .. BB */
220
221 buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL);
222 if (!buf)
223 return -ENOMEM;
224
225 if (backup_cred(cifs_sb))
226 create_options |= CREATE_OPEN_BACKUP_INTENT;
227
228 /* O_SYNC also has bit for O_DSYNC so following check picks up either */
229 if (f_flags & O_SYNC)
230 create_options |= CREATE_WRITE_THROUGH;
231
232 if (f_flags & O_DIRECT)
233 create_options |= CREATE_NO_BUFFER;
234
235 oparms.tcon = tcon;
236 oparms.cifs_sb = cifs_sb;
237 oparms.desired_access = desired_access;
238 oparms.create_options = create_options;
239 oparms.disposition = disposition;
240 oparms.path = full_path;
241 oparms.fid = fid;
242 oparms.reconnect = false;
243
244 rc = server->ops->open(xid, &oparms, oplock, buf);
245
246 if (rc)
247 goto out;
248
249 if (tcon->unix_ext)
250 rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb,
251 xid);
252 else
253 rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb,
254 xid, fid);
255
David Brazdil0f672f62019-12-10 10:32:29 +0000256 if (rc) {
257 server->ops->close(xid, tcon, fid);
258 if (rc == -ESTALE)
259 rc = -EOPENSTALE;
260 }
261
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000262out:
263 kfree(buf);
264 return rc;
265}
266
267static bool
268cifs_has_mand_locks(struct cifsInodeInfo *cinode)
269{
270 struct cifs_fid_locks *cur;
271 bool has_locks = false;
272
273 down_read(&cinode->lock_sem);
274 list_for_each_entry(cur, &cinode->llist, llist) {
275 if (!list_empty(&cur->locks)) {
276 has_locks = true;
277 break;
278 }
279 }
280 up_read(&cinode->lock_sem);
281 return has_locks;
282}
283
David Brazdil0f672f62019-12-10 10:32:29 +0000284void
285cifs_down_write(struct rw_semaphore *sem)
286{
287 while (!down_write_trylock(sem))
288 msleep(10);
289}
290
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000291struct cifsFileInfo *
292cifs_new_fileinfo(struct cifs_fid *fid, struct file *file,
293 struct tcon_link *tlink, __u32 oplock)
294{
295 struct dentry *dentry = file_dentry(file);
296 struct inode *inode = d_inode(dentry);
297 struct cifsInodeInfo *cinode = CIFS_I(inode);
298 struct cifsFileInfo *cfile;
299 struct cifs_fid_locks *fdlocks;
300 struct cifs_tcon *tcon = tlink_tcon(tlink);
301 struct TCP_Server_Info *server = tcon->ses->server;
302
303 cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL);
304 if (cfile == NULL)
305 return cfile;
306
307 fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL);
308 if (!fdlocks) {
309 kfree(cfile);
310 return NULL;
311 }
312
313 INIT_LIST_HEAD(&fdlocks->locks);
314 fdlocks->cfile = cfile;
315 cfile->llist = fdlocks;
David Brazdil0f672f62019-12-10 10:32:29 +0000316 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000317 list_add(&fdlocks->llist, &cinode->llist);
318 up_write(&cinode->lock_sem);
319
320 cfile->count = 1;
321 cfile->pid = current->tgid;
322 cfile->uid = current_fsuid();
323 cfile->dentry = dget(dentry);
324 cfile->f_flags = file->f_flags;
325 cfile->invalidHandle = false;
326 cfile->tlink = cifs_get_tlink(tlink);
327 INIT_WORK(&cfile->oplock_break, cifs_oplock_break);
328 mutex_init(&cfile->fh_mutex);
329 spin_lock_init(&cfile->file_info_lock);
330
331 cifs_sb_active(inode->i_sb);
332
333 /*
334 * If the server returned a read oplock and we have mandatory brlocks,
335 * set oplock level to None.
336 */
337 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
338 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
339 oplock = 0;
340 }
341
342 spin_lock(&tcon->open_file_lock);
343 if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock)
344 oplock = fid->pending_open->oplock;
345 list_del(&fid->pending_open->olist);
346
347 fid->purge_cache = false;
348 server->ops->set_fid(cfile, fid, oplock);
349
350 list_add(&cfile->tlist, &tcon->openFileList);
David Brazdil0f672f62019-12-10 10:32:29 +0000351 atomic_inc(&tcon->num_local_opens);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000352
353 /* if readable file instance put first in list*/
David Brazdil0f672f62019-12-10 10:32:29 +0000354 spin_lock(&cinode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000355 if (file->f_mode & FMODE_READ)
356 list_add(&cfile->flist, &cinode->openFileList);
357 else
358 list_add_tail(&cfile->flist, &cinode->openFileList);
David Brazdil0f672f62019-12-10 10:32:29 +0000359 spin_unlock(&cinode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000360 spin_unlock(&tcon->open_file_lock);
361
362 if (fid->purge_cache)
363 cifs_zap_mapping(inode);
364
365 file->private_data = cfile;
366 return cfile;
367}
368
369struct cifsFileInfo *
370cifsFileInfo_get(struct cifsFileInfo *cifs_file)
371{
372 spin_lock(&cifs_file->file_info_lock);
373 cifsFileInfo_get_locked(cifs_file);
374 spin_unlock(&cifs_file->file_info_lock);
375 return cifs_file;
376}
377
David Brazdil0f672f62019-12-10 10:32:29 +0000378/**
379 * cifsFileInfo_put - release a reference of file priv data
380 *
381 * Always potentially wait for oplock handler. See _cifsFileInfo_put().
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000382 */
383void cifsFileInfo_put(struct cifsFileInfo *cifs_file)
384{
David Brazdil0f672f62019-12-10 10:32:29 +0000385 _cifsFileInfo_put(cifs_file, true);
386}
387
388/**
389 * _cifsFileInfo_put - release a reference of file priv data
390 *
391 * This may involve closing the filehandle @cifs_file out on the
392 * server. Must be called without holding tcon->open_file_lock and
393 * cifs_file->file_info_lock.
394 *
395 * If @wait_for_oplock_handler is true and we are releasing the last
396 * reference, wait for any running oplock break handler of the file
397 * and cancel any pending one. If calling this function from the
398 * oplock break handler, you need to pass false.
399 *
400 */
401void _cifsFileInfo_put(struct cifsFileInfo *cifs_file, bool wait_oplock_handler)
402{
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000403 struct inode *inode = d_inode(cifs_file->dentry);
404 struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
405 struct TCP_Server_Info *server = tcon->ses->server;
406 struct cifsInodeInfo *cifsi = CIFS_I(inode);
407 struct super_block *sb = inode->i_sb;
408 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
409 struct cifsLockInfo *li, *tmp;
410 struct cifs_fid fid;
411 struct cifs_pending_open open;
412 bool oplock_break_cancelled;
413
414 spin_lock(&tcon->open_file_lock);
David Brazdil0f672f62019-12-10 10:32:29 +0000415 spin_lock(&cifsi->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000416 spin_lock(&cifs_file->file_info_lock);
417 if (--cifs_file->count > 0) {
418 spin_unlock(&cifs_file->file_info_lock);
David Brazdil0f672f62019-12-10 10:32:29 +0000419 spin_unlock(&cifsi->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000420 spin_unlock(&tcon->open_file_lock);
421 return;
422 }
423 spin_unlock(&cifs_file->file_info_lock);
424
425 if (server->ops->get_lease_key)
426 server->ops->get_lease_key(inode, &fid);
427
428 /* store open in pending opens to make sure we don't miss lease break */
429 cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open);
430
431 /* remove it from the lists */
432 list_del(&cifs_file->flist);
433 list_del(&cifs_file->tlist);
David Brazdil0f672f62019-12-10 10:32:29 +0000434 atomic_dec(&tcon->num_local_opens);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000435
436 if (list_empty(&cifsi->openFileList)) {
437 cifs_dbg(FYI, "closing last open instance for inode %p\n",
438 d_inode(cifs_file->dentry));
439 /*
440 * In strict cache mode we need invalidate mapping on the last
441 * close because it may cause a error when we open this file
442 * again and get at least level II oplock.
443 */
444 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO)
445 set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags);
446 cifs_set_oplock_level(cifsi, 0);
447 }
448
David Brazdil0f672f62019-12-10 10:32:29 +0000449 spin_unlock(&cifsi->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000450 spin_unlock(&tcon->open_file_lock);
451
David Brazdil0f672f62019-12-10 10:32:29 +0000452 oplock_break_cancelled = wait_oplock_handler ?
453 cancel_work_sync(&cifs_file->oplock_break) : false;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000454
455 if (!tcon->need_reconnect && !cifs_file->invalidHandle) {
456 struct TCP_Server_Info *server = tcon->ses->server;
457 unsigned int xid;
458
459 xid = get_xid();
460 if (server->ops->close)
461 server->ops->close(xid, tcon, &cifs_file->fid);
462 _free_xid(xid);
463 }
464
465 if (oplock_break_cancelled)
466 cifs_done_oplock_break(cifsi);
467
468 cifs_del_pending_open(&open);
469
470 /*
471 * Delete any outstanding lock records. We'll lose them when the file
472 * is closed anyway.
473 */
David Brazdil0f672f62019-12-10 10:32:29 +0000474 cifs_down_write(&cifsi->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000475 list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) {
476 list_del(&li->llist);
477 cifs_del_lock_waiters(li);
478 kfree(li);
479 }
480 list_del(&cifs_file->llist->llist);
481 kfree(cifs_file->llist);
482 up_write(&cifsi->lock_sem);
483
484 cifs_put_tlink(cifs_file->tlink);
485 dput(cifs_file->dentry);
486 cifs_sb_deactive(sb);
487 kfree(cifs_file);
488}
489
490int cifs_open(struct inode *inode, struct file *file)
491
492{
493 int rc = -EACCES;
494 unsigned int xid;
495 __u32 oplock;
496 struct cifs_sb_info *cifs_sb;
497 struct TCP_Server_Info *server;
498 struct cifs_tcon *tcon;
499 struct tcon_link *tlink;
500 struct cifsFileInfo *cfile = NULL;
501 char *full_path = NULL;
502 bool posix_open_ok = false;
503 struct cifs_fid fid;
504 struct cifs_pending_open open;
505
506 xid = get_xid();
507
508 cifs_sb = CIFS_SB(inode->i_sb);
509 tlink = cifs_sb_tlink(cifs_sb);
510 if (IS_ERR(tlink)) {
511 free_xid(xid);
512 return PTR_ERR(tlink);
513 }
514 tcon = tlink_tcon(tlink);
515 server = tcon->ses->server;
516
517 full_path = build_path_from_dentry(file_dentry(file));
518 if (full_path == NULL) {
519 rc = -ENOMEM;
520 goto out;
521 }
522
523 cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n",
524 inode, file->f_flags, full_path);
525
526 if (file->f_flags & O_DIRECT &&
527 cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) {
528 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL)
529 file->f_op = &cifs_file_direct_nobrl_ops;
530 else
531 file->f_op = &cifs_file_direct_ops;
532 }
533
534 if (server->oplocks)
535 oplock = REQ_OPLOCK;
536 else
537 oplock = 0;
538
539 if (!tcon->broken_posix_open && tcon->unix_ext &&
540 cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP &
541 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
542 /* can not refresh inode info since size could be stale */
543 rc = cifs_posix_open(full_path, &inode, inode->i_sb,
544 cifs_sb->mnt_file_mode /* ignored */,
545 file->f_flags, &oplock, &fid.netfid, xid);
546 if (rc == 0) {
547 cifs_dbg(FYI, "posix open succeeded\n");
548 posix_open_ok = true;
549 } else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) {
550 if (tcon->ses->serverNOS)
551 cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n",
552 tcon->ses->serverName,
553 tcon->ses->serverNOS);
554 tcon->broken_posix_open = true;
555 } else if ((rc != -EIO) && (rc != -EREMOTE) &&
556 (rc != -EOPNOTSUPP)) /* path not found or net err */
557 goto out;
558 /*
559 * Else fallthrough to retry open the old way on network i/o
560 * or DFS errors.
561 */
562 }
563
564 if (server->ops->get_lease_key)
565 server->ops->get_lease_key(inode, &fid);
566
567 cifs_add_pending_open(&fid, tlink, &open);
568
569 if (!posix_open_ok) {
570 if (server->ops->get_lease_key)
571 server->ops->get_lease_key(inode, &fid);
572
573 rc = cifs_nt_open(full_path, inode, cifs_sb, tcon,
574 file->f_flags, &oplock, &fid, xid);
575 if (rc) {
576 cifs_del_pending_open(&open);
577 goto out;
578 }
579 }
580
581 cfile = cifs_new_fileinfo(&fid, file, tlink, oplock);
582 if (cfile == NULL) {
583 if (server->ops->close)
584 server->ops->close(xid, tcon, &fid);
585 cifs_del_pending_open(&open);
586 rc = -ENOMEM;
587 goto out;
588 }
589
590 cifs_fscache_set_inode_cookie(inode, file);
591
592 if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) {
593 /*
594 * Time to set mode which we can not set earlier due to
595 * problems creating new read-only files.
596 */
597 struct cifs_unix_set_info_args args = {
598 .mode = inode->i_mode,
599 .uid = INVALID_UID, /* no change */
600 .gid = INVALID_GID, /* no change */
601 .ctime = NO_CHANGE_64,
602 .atime = NO_CHANGE_64,
603 .mtime = NO_CHANGE_64,
604 .device = 0,
605 };
606 CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid,
607 cfile->pid);
608 }
609
610out:
611 kfree(full_path);
612 free_xid(xid);
613 cifs_put_tlink(tlink);
614 return rc;
615}
616
617static int cifs_push_posix_locks(struct cifsFileInfo *cfile);
618
619/*
620 * Try to reacquire byte range locks that were released when session
621 * to server was lost.
622 */
623static int
624cifs_relock_file(struct cifsFileInfo *cfile)
625{
626 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
627 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
628 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
629 int rc = 0;
630
631 down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING);
632 if (cinode->can_cache_brlcks) {
633 /* can cache locks - no need to relock */
634 up_read(&cinode->lock_sem);
635 return rc;
636 }
637
638 if (cap_unix(tcon->ses) &&
639 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
640 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
641 rc = cifs_push_posix_locks(cfile);
642 else
643 rc = tcon->ses->server->ops->push_mand_locks(cfile);
644
645 up_read(&cinode->lock_sem);
646 return rc;
647}
648
649static int
650cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush)
651{
652 int rc = -EACCES;
653 unsigned int xid;
654 __u32 oplock;
655 struct cifs_sb_info *cifs_sb;
656 struct cifs_tcon *tcon;
657 struct TCP_Server_Info *server;
658 struct cifsInodeInfo *cinode;
659 struct inode *inode;
660 char *full_path = NULL;
661 int desired_access;
662 int disposition = FILE_OPEN;
663 int create_options = CREATE_NOT_DIR;
664 struct cifs_open_parms oparms;
665
666 xid = get_xid();
667 mutex_lock(&cfile->fh_mutex);
668 if (!cfile->invalidHandle) {
669 mutex_unlock(&cfile->fh_mutex);
670 rc = 0;
671 free_xid(xid);
672 return rc;
673 }
674
675 inode = d_inode(cfile->dentry);
676 cifs_sb = CIFS_SB(inode->i_sb);
677 tcon = tlink_tcon(cfile->tlink);
678 server = tcon->ses->server;
679
680 /*
681 * Can not grab rename sem here because various ops, including those
682 * that already have the rename sem can end up causing writepage to get
683 * called and if the server was down that means we end up here, and we
684 * can never tell if the caller already has the rename_sem.
685 */
686 full_path = build_path_from_dentry(cfile->dentry);
687 if (full_path == NULL) {
688 rc = -ENOMEM;
689 mutex_unlock(&cfile->fh_mutex);
690 free_xid(xid);
691 return rc;
692 }
693
694 cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n",
695 inode, cfile->f_flags, full_path);
696
697 if (tcon->ses->server->oplocks)
698 oplock = REQ_OPLOCK;
699 else
700 oplock = 0;
701
702 if (tcon->unix_ext && cap_unix(tcon->ses) &&
703 (CIFS_UNIX_POSIX_PATH_OPS_CAP &
704 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
705 /*
706 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the
707 * original open. Must mask them off for a reopen.
708 */
709 unsigned int oflags = cfile->f_flags &
710 ~(O_CREAT | O_EXCL | O_TRUNC);
711
712 rc = cifs_posix_open(full_path, NULL, inode->i_sb,
713 cifs_sb->mnt_file_mode /* ignored */,
714 oflags, &oplock, &cfile->fid.netfid, xid);
715 if (rc == 0) {
716 cifs_dbg(FYI, "posix reopen succeeded\n");
717 oparms.reconnect = true;
718 goto reopen_success;
719 }
720 /*
721 * fallthrough to retry open the old way on errors, especially
722 * in the reconnect path it is important to retry hard
723 */
724 }
725
726 desired_access = cifs_convert_flags(cfile->f_flags);
727
728 if (backup_cred(cifs_sb))
729 create_options |= CREATE_OPEN_BACKUP_INTENT;
730
731 if (server->ops->get_lease_key)
732 server->ops->get_lease_key(inode, &cfile->fid);
733
734 oparms.tcon = tcon;
735 oparms.cifs_sb = cifs_sb;
736 oparms.desired_access = desired_access;
737 oparms.create_options = create_options;
738 oparms.disposition = disposition;
739 oparms.path = full_path;
740 oparms.fid = &cfile->fid;
741 oparms.reconnect = true;
742
743 /*
744 * Can not refresh inode by passing in file_info buf to be returned by
745 * ops->open and then calling get_inode_info with returned buf since
746 * file might have write behind data that needs to be flushed and server
747 * version of file size can be stale. If we knew for sure that inode was
748 * not dirty locally we could do this.
749 */
750 rc = server->ops->open(xid, &oparms, &oplock, NULL);
751 if (rc == -ENOENT && oparms.reconnect == false) {
752 /* durable handle timeout is expired - open the file again */
753 rc = server->ops->open(xid, &oparms, &oplock, NULL);
754 /* indicate that we need to relock the file */
755 oparms.reconnect = true;
756 }
757
758 if (rc) {
759 mutex_unlock(&cfile->fh_mutex);
760 cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc);
761 cifs_dbg(FYI, "oplock: %d\n", oplock);
762 goto reopen_error_exit;
763 }
764
765reopen_success:
766 cfile->invalidHandle = false;
767 mutex_unlock(&cfile->fh_mutex);
768 cinode = CIFS_I(inode);
769
770 if (can_flush) {
771 rc = filemap_write_and_wait(inode->i_mapping);
David Brazdil0f672f62019-12-10 10:32:29 +0000772 if (!is_interrupt_error(rc))
773 mapping_set_error(inode->i_mapping, rc);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000774
775 if (tcon->unix_ext)
776 rc = cifs_get_inode_info_unix(&inode, full_path,
777 inode->i_sb, xid);
778 else
779 rc = cifs_get_inode_info(&inode, full_path, NULL,
780 inode->i_sb, xid, NULL);
781 }
782 /*
783 * Else we are writing out data to server already and could deadlock if
784 * we tried to flush data, and since we do not know if we have data that
785 * would invalidate the current end of file on the server we can not go
786 * to the server to get the new inode info.
787 */
788
789 /*
790 * If the server returned a read oplock and we have mandatory brlocks,
791 * set oplock level to None.
792 */
793 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
794 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
795 oplock = 0;
796 }
797
798 server->ops->set_fid(cfile, &cfile->fid, oplock);
799 if (oparms.reconnect)
800 cifs_relock_file(cfile);
801
802reopen_error_exit:
803 kfree(full_path);
804 free_xid(xid);
805 return rc;
806}
807
808int cifs_close(struct inode *inode, struct file *file)
809{
810 if (file->private_data != NULL) {
811 cifsFileInfo_put(file->private_data);
812 file->private_data = NULL;
813 }
814
815 /* return code from the ->release op is always ignored */
816 return 0;
817}
818
819void
820cifs_reopen_persistent_handles(struct cifs_tcon *tcon)
821{
822 struct cifsFileInfo *open_file;
823 struct list_head *tmp;
824 struct list_head *tmp1;
825 struct list_head tmp_list;
826
827 if (!tcon->use_persistent || !tcon->need_reopen_files)
828 return;
829
830 tcon->need_reopen_files = false;
831
832 cifs_dbg(FYI, "Reopen persistent handles");
833 INIT_LIST_HEAD(&tmp_list);
834
835 /* list all files open on tree connection, reopen resilient handles */
836 spin_lock(&tcon->open_file_lock);
837 list_for_each(tmp, &tcon->openFileList) {
838 open_file = list_entry(tmp, struct cifsFileInfo, tlist);
839 if (!open_file->invalidHandle)
840 continue;
841 cifsFileInfo_get(open_file);
842 list_add_tail(&open_file->rlist, &tmp_list);
843 }
844 spin_unlock(&tcon->open_file_lock);
845
846 list_for_each_safe(tmp, tmp1, &tmp_list) {
847 open_file = list_entry(tmp, struct cifsFileInfo, rlist);
848 if (cifs_reopen_file(open_file, false /* do not flush */))
849 tcon->need_reopen_files = true;
850 list_del_init(&open_file->rlist);
851 cifsFileInfo_put(open_file);
852 }
853}
854
855int cifs_closedir(struct inode *inode, struct file *file)
856{
857 int rc = 0;
858 unsigned int xid;
859 struct cifsFileInfo *cfile = file->private_data;
860 struct cifs_tcon *tcon;
861 struct TCP_Server_Info *server;
862 char *buf;
863
864 cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode);
865
866 if (cfile == NULL)
867 return rc;
868
869 xid = get_xid();
870 tcon = tlink_tcon(cfile->tlink);
871 server = tcon->ses->server;
872
873 cifs_dbg(FYI, "Freeing private data in close dir\n");
874 spin_lock(&cfile->file_info_lock);
875 if (server->ops->dir_needs_close(cfile)) {
876 cfile->invalidHandle = true;
877 spin_unlock(&cfile->file_info_lock);
878 if (server->ops->close_dir)
879 rc = server->ops->close_dir(xid, tcon, &cfile->fid);
880 else
881 rc = -ENOSYS;
882 cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc);
883 /* not much we can do if it fails anyway, ignore rc */
884 rc = 0;
885 } else
886 spin_unlock(&cfile->file_info_lock);
887
888 buf = cfile->srch_inf.ntwrk_buf_start;
889 if (buf) {
890 cifs_dbg(FYI, "closedir free smb buf in srch struct\n");
891 cfile->srch_inf.ntwrk_buf_start = NULL;
892 if (cfile->srch_inf.smallBuf)
893 cifs_small_buf_release(buf);
894 else
895 cifs_buf_release(buf);
896 }
897
898 cifs_put_tlink(cfile->tlink);
899 kfree(file->private_data);
900 file->private_data = NULL;
901 /* BB can we lock the filestruct while this is going on? */
902 free_xid(xid);
903 return rc;
904}
905
906static struct cifsLockInfo *
David Brazdil0f672f62019-12-10 10:32:29 +0000907cifs_lock_init(__u64 offset, __u64 length, __u8 type, __u16 flags)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000908{
909 struct cifsLockInfo *lock =
910 kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL);
911 if (!lock)
912 return lock;
913 lock->offset = offset;
914 lock->length = length;
915 lock->type = type;
916 lock->pid = current->tgid;
David Brazdil0f672f62019-12-10 10:32:29 +0000917 lock->flags = flags;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000918 INIT_LIST_HEAD(&lock->blist);
919 init_waitqueue_head(&lock->block_q);
920 return lock;
921}
922
923void
924cifs_del_lock_waiters(struct cifsLockInfo *lock)
925{
926 struct cifsLockInfo *li, *tmp;
927 list_for_each_entry_safe(li, tmp, &lock->blist, blist) {
928 list_del_init(&li->blist);
929 wake_up(&li->block_q);
930 }
931}
932
933#define CIFS_LOCK_OP 0
934#define CIFS_READ_OP 1
935#define CIFS_WRITE_OP 2
936
937/* @rw_check : 0 - no op, 1 - read, 2 - write */
938static bool
939cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset,
David Brazdil0f672f62019-12-10 10:32:29 +0000940 __u64 length, __u8 type, __u16 flags,
941 struct cifsFileInfo *cfile,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000942 struct cifsLockInfo **conf_lock, int rw_check)
943{
944 struct cifsLockInfo *li;
945 struct cifsFileInfo *cur_cfile = fdlocks->cfile;
946 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
947
948 list_for_each_entry(li, &fdlocks->locks, llist) {
949 if (offset + length <= li->offset ||
950 offset >= li->offset + li->length)
951 continue;
952 if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
953 server->ops->compare_fids(cfile, cur_cfile)) {
954 /* shared lock prevents write op through the same fid */
955 if (!(li->type & server->vals->shared_lock_type) ||
956 rw_check != CIFS_WRITE_OP)
957 continue;
958 }
959 if ((type & server->vals->shared_lock_type) &&
960 ((server->ops->compare_fids(cfile, cur_cfile) &&
961 current->tgid == li->pid) || type == li->type))
962 continue;
David Brazdil0f672f62019-12-10 10:32:29 +0000963 if (rw_check == CIFS_LOCK_OP &&
964 (flags & FL_OFDLCK) && (li->flags & FL_OFDLCK) &&
965 server->ops->compare_fids(cfile, cur_cfile))
966 continue;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000967 if (conf_lock)
968 *conf_lock = li;
969 return true;
970 }
971 return false;
972}
973
974bool
975cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
David Brazdil0f672f62019-12-10 10:32:29 +0000976 __u8 type, __u16 flags,
977 struct cifsLockInfo **conf_lock, int rw_check)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000978{
979 bool rc = false;
980 struct cifs_fid_locks *cur;
981 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
982
983 list_for_each_entry(cur, &cinode->llist, llist) {
984 rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
David Brazdil0f672f62019-12-10 10:32:29 +0000985 flags, cfile, conf_lock,
986 rw_check);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000987 if (rc)
988 break;
989 }
990
991 return rc;
992}
993
994/*
995 * Check if there is another lock that prevents us to set the lock (mandatory
996 * style). If such a lock exists, update the flock structure with its
997 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
998 * or leave it the same if we can't. Returns 0 if we don't need to request to
999 * the server or 1 otherwise.
1000 */
1001static int
1002cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1003 __u8 type, struct file_lock *flock)
1004{
1005 int rc = 0;
1006 struct cifsLockInfo *conf_lock;
1007 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1008 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1009 bool exist;
1010
1011 down_read(&cinode->lock_sem);
1012
1013 exist = cifs_find_lock_conflict(cfile, offset, length, type,
David Brazdil0f672f62019-12-10 10:32:29 +00001014 flock->fl_flags, &conf_lock,
1015 CIFS_LOCK_OP);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001016 if (exist) {
1017 flock->fl_start = conf_lock->offset;
1018 flock->fl_end = conf_lock->offset + conf_lock->length - 1;
1019 flock->fl_pid = conf_lock->pid;
1020 if (conf_lock->type & server->vals->shared_lock_type)
1021 flock->fl_type = F_RDLCK;
1022 else
1023 flock->fl_type = F_WRLCK;
1024 } else if (!cinode->can_cache_brlcks)
1025 rc = 1;
1026 else
1027 flock->fl_type = F_UNLCK;
1028
1029 up_read(&cinode->lock_sem);
1030 return rc;
1031}
1032
1033static void
1034cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock)
1035{
1036 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
David Brazdil0f672f62019-12-10 10:32:29 +00001037 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001038 list_add_tail(&lock->llist, &cfile->llist->locks);
1039 up_write(&cinode->lock_sem);
1040}
1041
1042/*
1043 * Set the byte-range lock (mandatory style). Returns:
1044 * 1) 0, if we set the lock and don't need to request to the server;
1045 * 2) 1, if no locks prevent us but we need to request to the server;
David Brazdil0f672f62019-12-10 10:32:29 +00001046 * 3) -EACCES, if there is a lock that prevents us and wait is false.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001047 */
1048static int
1049cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock,
1050 bool wait)
1051{
1052 struct cifsLockInfo *conf_lock;
1053 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1054 bool exist;
1055 int rc = 0;
1056
1057try_again:
1058 exist = false;
David Brazdil0f672f62019-12-10 10:32:29 +00001059 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001060
1061 exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length,
David Brazdil0f672f62019-12-10 10:32:29 +00001062 lock->type, lock->flags, &conf_lock,
1063 CIFS_LOCK_OP);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001064 if (!exist && cinode->can_cache_brlcks) {
1065 list_add_tail(&lock->llist, &cfile->llist->locks);
1066 up_write(&cinode->lock_sem);
1067 return rc;
1068 }
1069
1070 if (!exist)
1071 rc = 1;
1072 else if (!wait)
1073 rc = -EACCES;
1074 else {
1075 list_add_tail(&lock->blist, &conf_lock->blist);
1076 up_write(&cinode->lock_sem);
1077 rc = wait_event_interruptible(lock->block_q,
1078 (lock->blist.prev == &lock->blist) &&
1079 (lock->blist.next == &lock->blist));
1080 if (!rc)
1081 goto try_again;
David Brazdil0f672f62019-12-10 10:32:29 +00001082 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001083 list_del_init(&lock->blist);
1084 }
1085
1086 up_write(&cinode->lock_sem);
1087 return rc;
1088}
1089
1090/*
1091 * Check if there is another lock that prevents us to set the lock (posix
1092 * style). If such a lock exists, update the flock structure with its
1093 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1094 * or leave it the same if we can't. Returns 0 if we don't need to request to
1095 * the server or 1 otherwise.
1096 */
1097static int
1098cifs_posix_lock_test(struct file *file, struct file_lock *flock)
1099{
1100 int rc = 0;
1101 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1102 unsigned char saved_type = flock->fl_type;
1103
1104 if ((flock->fl_flags & FL_POSIX) == 0)
1105 return 1;
1106
1107 down_read(&cinode->lock_sem);
1108 posix_test_lock(file, flock);
1109
1110 if (flock->fl_type == F_UNLCK && !cinode->can_cache_brlcks) {
1111 flock->fl_type = saved_type;
1112 rc = 1;
1113 }
1114
1115 up_read(&cinode->lock_sem);
1116 return rc;
1117}
1118
1119/*
1120 * Set the byte-range lock (posix style). Returns:
1121 * 1) 0, if we set the lock and don't need to request to the server;
1122 * 2) 1, if we need to request to the server;
1123 * 3) <0, if the error occurs while setting the lock.
1124 */
1125static int
1126cifs_posix_lock_set(struct file *file, struct file_lock *flock)
1127{
1128 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1129 int rc = 1;
1130
1131 if ((flock->fl_flags & FL_POSIX) == 0)
1132 return rc;
1133
1134try_again:
David Brazdil0f672f62019-12-10 10:32:29 +00001135 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001136 if (!cinode->can_cache_brlcks) {
1137 up_write(&cinode->lock_sem);
1138 return rc;
1139 }
1140
1141 rc = posix_lock_file(file, flock, NULL);
1142 up_write(&cinode->lock_sem);
1143 if (rc == FILE_LOCK_DEFERRED) {
David Brazdil0f672f62019-12-10 10:32:29 +00001144 rc = wait_event_interruptible(flock->fl_wait, !flock->fl_blocker);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001145 if (!rc)
1146 goto try_again;
David Brazdil0f672f62019-12-10 10:32:29 +00001147 locks_delete_block(flock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001148 }
1149 return rc;
1150}
1151
1152int
1153cifs_push_mandatory_locks(struct cifsFileInfo *cfile)
1154{
1155 unsigned int xid;
1156 int rc = 0, stored_rc;
1157 struct cifsLockInfo *li, *tmp;
1158 struct cifs_tcon *tcon;
1159 unsigned int num, max_num, max_buf;
1160 LOCKING_ANDX_RANGE *buf, *cur;
1161 static const int types[] = {
1162 LOCKING_ANDX_LARGE_FILES,
1163 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1164 };
1165 int i;
1166
1167 xid = get_xid();
1168 tcon = tlink_tcon(cfile->tlink);
1169
1170 /*
1171 * Accessing maxBuf is racy with cifs_reconnect - need to store value
David Brazdil0f672f62019-12-10 10:32:29 +00001172 * and check it before using.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001173 */
1174 max_buf = tcon->ses->server->maxBuf;
David Brazdil0f672f62019-12-10 10:32:29 +00001175 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001176 free_xid(xid);
1177 return -EINVAL;
1178 }
1179
David Brazdil0f672f62019-12-10 10:32:29 +00001180 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1181 PAGE_SIZE);
1182 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1183 PAGE_SIZE);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001184 max_num = (max_buf - sizeof(struct smb_hdr)) /
1185 sizeof(LOCKING_ANDX_RANGE);
1186 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
1187 if (!buf) {
1188 free_xid(xid);
1189 return -ENOMEM;
1190 }
1191
1192 for (i = 0; i < 2; i++) {
1193 cur = buf;
1194 num = 0;
1195 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
1196 if (li->type != types[i])
1197 continue;
1198 cur->Pid = cpu_to_le16(li->pid);
1199 cur->LengthLow = cpu_to_le32((u32)li->length);
1200 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1201 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1202 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1203 if (++num == max_num) {
1204 stored_rc = cifs_lockv(xid, tcon,
1205 cfile->fid.netfid,
1206 (__u8)li->type, 0, num,
1207 buf);
1208 if (stored_rc)
1209 rc = stored_rc;
1210 cur = buf;
1211 num = 0;
1212 } else
1213 cur++;
1214 }
1215
1216 if (num) {
1217 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
1218 (__u8)types[i], 0, num, buf);
1219 if (stored_rc)
1220 rc = stored_rc;
1221 }
1222 }
1223
1224 kfree(buf);
1225 free_xid(xid);
1226 return rc;
1227}
1228
1229static __u32
1230hash_lockowner(fl_owner_t owner)
1231{
1232 return cifs_lock_secret ^ hash32_ptr((const void *)owner);
1233}
1234
1235struct lock_to_push {
1236 struct list_head llist;
1237 __u64 offset;
1238 __u64 length;
1239 __u32 pid;
1240 __u16 netfid;
1241 __u8 type;
1242};
1243
1244static int
1245cifs_push_posix_locks(struct cifsFileInfo *cfile)
1246{
1247 struct inode *inode = d_inode(cfile->dentry);
1248 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1249 struct file_lock *flock;
1250 struct file_lock_context *flctx = inode->i_flctx;
1251 unsigned int count = 0, i;
1252 int rc = 0, xid, type;
1253 struct list_head locks_to_send, *el;
1254 struct lock_to_push *lck, *tmp;
1255 __u64 length;
1256
1257 xid = get_xid();
1258
1259 if (!flctx)
1260 goto out;
1261
1262 spin_lock(&flctx->flc_lock);
1263 list_for_each(el, &flctx->flc_posix) {
1264 count++;
1265 }
1266 spin_unlock(&flctx->flc_lock);
1267
1268 INIT_LIST_HEAD(&locks_to_send);
1269
1270 /*
1271 * Allocating count locks is enough because no FL_POSIX locks can be
1272 * added to the list while we are holding cinode->lock_sem that
1273 * protects locking operations of this inode.
1274 */
1275 for (i = 0; i < count; i++) {
1276 lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL);
1277 if (!lck) {
1278 rc = -ENOMEM;
1279 goto err_out;
1280 }
1281 list_add_tail(&lck->llist, &locks_to_send);
1282 }
1283
1284 el = locks_to_send.next;
1285 spin_lock(&flctx->flc_lock);
1286 list_for_each_entry(flock, &flctx->flc_posix, fl_list) {
1287 if (el == &locks_to_send) {
1288 /*
1289 * The list ended. We don't have enough allocated
1290 * structures - something is really wrong.
1291 */
1292 cifs_dbg(VFS, "Can't push all brlocks!\n");
1293 break;
1294 }
1295 length = 1 + flock->fl_end - flock->fl_start;
1296 if (flock->fl_type == F_RDLCK || flock->fl_type == F_SHLCK)
1297 type = CIFS_RDLCK;
1298 else
1299 type = CIFS_WRLCK;
1300 lck = list_entry(el, struct lock_to_push, llist);
1301 lck->pid = hash_lockowner(flock->fl_owner);
1302 lck->netfid = cfile->fid.netfid;
1303 lck->length = length;
1304 lck->type = type;
1305 lck->offset = flock->fl_start;
1306 }
1307 spin_unlock(&flctx->flc_lock);
1308
1309 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1310 int stored_rc;
1311
1312 stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid,
1313 lck->offset, lck->length, NULL,
1314 lck->type, 0);
1315 if (stored_rc)
1316 rc = stored_rc;
1317 list_del(&lck->llist);
1318 kfree(lck);
1319 }
1320
1321out:
1322 free_xid(xid);
1323 return rc;
1324err_out:
1325 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1326 list_del(&lck->llist);
1327 kfree(lck);
1328 }
1329 goto out;
1330}
1331
1332static int
1333cifs_push_locks(struct cifsFileInfo *cfile)
1334{
1335 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
1336 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1337 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1338 int rc = 0;
1339
1340 /* we are going to update can_cache_brlcks here - need a write access */
David Brazdil0f672f62019-12-10 10:32:29 +00001341 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001342 if (!cinode->can_cache_brlcks) {
1343 up_write(&cinode->lock_sem);
1344 return rc;
1345 }
1346
1347 if (cap_unix(tcon->ses) &&
1348 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1349 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1350 rc = cifs_push_posix_locks(cfile);
1351 else
1352 rc = tcon->ses->server->ops->push_mand_locks(cfile);
1353
1354 cinode->can_cache_brlcks = false;
1355 up_write(&cinode->lock_sem);
1356 return rc;
1357}
1358
1359static void
1360cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock,
1361 bool *wait_flag, struct TCP_Server_Info *server)
1362{
1363 if (flock->fl_flags & FL_POSIX)
1364 cifs_dbg(FYI, "Posix\n");
1365 if (flock->fl_flags & FL_FLOCK)
1366 cifs_dbg(FYI, "Flock\n");
1367 if (flock->fl_flags & FL_SLEEP) {
1368 cifs_dbg(FYI, "Blocking lock\n");
1369 *wait_flag = true;
1370 }
1371 if (flock->fl_flags & FL_ACCESS)
1372 cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n");
1373 if (flock->fl_flags & FL_LEASE)
1374 cifs_dbg(FYI, "Lease on file - not implemented yet\n");
1375 if (flock->fl_flags &
1376 (~(FL_POSIX | FL_FLOCK | FL_SLEEP |
David Brazdil0f672f62019-12-10 10:32:29 +00001377 FL_ACCESS | FL_LEASE | FL_CLOSE | FL_OFDLCK)))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001378 cifs_dbg(FYI, "Unknown lock flags 0x%x\n", flock->fl_flags);
1379
1380 *type = server->vals->large_lock_type;
1381 if (flock->fl_type == F_WRLCK) {
1382 cifs_dbg(FYI, "F_WRLCK\n");
1383 *type |= server->vals->exclusive_lock_type;
1384 *lock = 1;
1385 } else if (flock->fl_type == F_UNLCK) {
1386 cifs_dbg(FYI, "F_UNLCK\n");
1387 *type |= server->vals->unlock_lock_type;
1388 *unlock = 1;
1389 /* Check if unlock includes more than one lock range */
1390 } else if (flock->fl_type == F_RDLCK) {
1391 cifs_dbg(FYI, "F_RDLCK\n");
1392 *type |= server->vals->shared_lock_type;
1393 *lock = 1;
1394 } else if (flock->fl_type == F_EXLCK) {
1395 cifs_dbg(FYI, "F_EXLCK\n");
1396 *type |= server->vals->exclusive_lock_type;
1397 *lock = 1;
1398 } else if (flock->fl_type == F_SHLCK) {
1399 cifs_dbg(FYI, "F_SHLCK\n");
1400 *type |= server->vals->shared_lock_type;
1401 *lock = 1;
1402 } else
1403 cifs_dbg(FYI, "Unknown type of lock\n");
1404}
1405
1406static int
1407cifs_getlk(struct file *file, struct file_lock *flock, __u32 type,
1408 bool wait_flag, bool posix_lck, unsigned int xid)
1409{
1410 int rc = 0;
1411 __u64 length = 1 + flock->fl_end - flock->fl_start;
1412 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1413 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1414 struct TCP_Server_Info *server = tcon->ses->server;
1415 __u16 netfid = cfile->fid.netfid;
1416
1417 if (posix_lck) {
1418 int posix_lock_type;
1419
1420 rc = cifs_posix_lock_test(file, flock);
1421 if (!rc)
1422 return rc;
1423
1424 if (type & server->vals->shared_lock_type)
1425 posix_lock_type = CIFS_RDLCK;
1426 else
1427 posix_lock_type = CIFS_WRLCK;
1428 rc = CIFSSMBPosixLock(xid, tcon, netfid,
1429 hash_lockowner(flock->fl_owner),
1430 flock->fl_start, length, flock,
1431 posix_lock_type, wait_flag);
1432 return rc;
1433 }
1434
1435 rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock);
1436 if (!rc)
1437 return rc;
1438
1439 /* BB we could chain these into one lock request BB */
1440 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type,
1441 1, 0, false);
1442 if (rc == 0) {
1443 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1444 type, 0, 1, false);
1445 flock->fl_type = F_UNLCK;
1446 if (rc != 0)
1447 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1448 rc);
1449 return 0;
1450 }
1451
1452 if (type & server->vals->shared_lock_type) {
1453 flock->fl_type = F_WRLCK;
1454 return 0;
1455 }
1456
1457 type &= ~server->vals->exclusive_lock_type;
1458
1459 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1460 type | server->vals->shared_lock_type,
1461 1, 0, false);
1462 if (rc == 0) {
1463 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1464 type | server->vals->shared_lock_type, 0, 1, false);
1465 flock->fl_type = F_RDLCK;
1466 if (rc != 0)
1467 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1468 rc);
1469 } else
1470 flock->fl_type = F_WRLCK;
1471
1472 return 0;
1473}
1474
1475void
1476cifs_move_llist(struct list_head *source, struct list_head *dest)
1477{
1478 struct list_head *li, *tmp;
1479 list_for_each_safe(li, tmp, source)
1480 list_move(li, dest);
1481}
1482
1483void
1484cifs_free_llist(struct list_head *llist)
1485{
1486 struct cifsLockInfo *li, *tmp;
1487 list_for_each_entry_safe(li, tmp, llist, llist) {
1488 cifs_del_lock_waiters(li);
1489 list_del(&li->llist);
1490 kfree(li);
1491 }
1492}
1493
1494int
1495cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock,
1496 unsigned int xid)
1497{
1498 int rc = 0, stored_rc;
1499 static const int types[] = {
1500 LOCKING_ANDX_LARGE_FILES,
1501 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1502 };
1503 unsigned int i;
1504 unsigned int max_num, num, max_buf;
1505 LOCKING_ANDX_RANGE *buf, *cur;
1506 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1507 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1508 struct cifsLockInfo *li, *tmp;
1509 __u64 length = 1 + flock->fl_end - flock->fl_start;
1510 struct list_head tmp_llist;
1511
1512 INIT_LIST_HEAD(&tmp_llist);
1513
1514 /*
1515 * Accessing maxBuf is racy with cifs_reconnect - need to store value
David Brazdil0f672f62019-12-10 10:32:29 +00001516 * and check it before using.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001517 */
1518 max_buf = tcon->ses->server->maxBuf;
David Brazdil0f672f62019-12-10 10:32:29 +00001519 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE)))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001520 return -EINVAL;
1521
David Brazdil0f672f62019-12-10 10:32:29 +00001522 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1523 PAGE_SIZE);
1524 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1525 PAGE_SIZE);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001526 max_num = (max_buf - sizeof(struct smb_hdr)) /
1527 sizeof(LOCKING_ANDX_RANGE);
1528 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
1529 if (!buf)
1530 return -ENOMEM;
1531
David Brazdil0f672f62019-12-10 10:32:29 +00001532 cifs_down_write(&cinode->lock_sem);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001533 for (i = 0; i < 2; i++) {
1534 cur = buf;
1535 num = 0;
1536 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
1537 if (flock->fl_start > li->offset ||
1538 (flock->fl_start + length) <
1539 (li->offset + li->length))
1540 continue;
1541 if (current->tgid != li->pid)
1542 continue;
1543 if (types[i] != li->type)
1544 continue;
1545 if (cinode->can_cache_brlcks) {
1546 /*
1547 * We can cache brlock requests - simply remove
1548 * a lock from the file's list.
1549 */
1550 list_del(&li->llist);
1551 cifs_del_lock_waiters(li);
1552 kfree(li);
1553 continue;
1554 }
1555 cur->Pid = cpu_to_le16(li->pid);
1556 cur->LengthLow = cpu_to_le32((u32)li->length);
1557 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1558 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1559 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1560 /*
1561 * We need to save a lock here to let us add it again to
1562 * the file's list if the unlock range request fails on
1563 * the server.
1564 */
1565 list_move(&li->llist, &tmp_llist);
1566 if (++num == max_num) {
1567 stored_rc = cifs_lockv(xid, tcon,
1568 cfile->fid.netfid,
1569 li->type, num, 0, buf);
1570 if (stored_rc) {
1571 /*
1572 * We failed on the unlock range
1573 * request - add all locks from the tmp
1574 * list to the head of the file's list.
1575 */
1576 cifs_move_llist(&tmp_llist,
1577 &cfile->llist->locks);
1578 rc = stored_rc;
1579 } else
1580 /*
1581 * The unlock range request succeed -
1582 * free the tmp list.
1583 */
1584 cifs_free_llist(&tmp_llist);
1585 cur = buf;
1586 num = 0;
1587 } else
1588 cur++;
1589 }
1590 if (num) {
1591 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
1592 types[i], num, 0, buf);
1593 if (stored_rc) {
1594 cifs_move_llist(&tmp_llist,
1595 &cfile->llist->locks);
1596 rc = stored_rc;
1597 } else
1598 cifs_free_llist(&tmp_llist);
1599 }
1600 }
1601
1602 up_write(&cinode->lock_sem);
1603 kfree(buf);
1604 return rc;
1605}
1606
1607static int
1608cifs_setlk(struct file *file, struct file_lock *flock, __u32 type,
1609 bool wait_flag, bool posix_lck, int lock, int unlock,
1610 unsigned int xid)
1611{
1612 int rc = 0;
1613 __u64 length = 1 + flock->fl_end - flock->fl_start;
1614 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1615 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1616 struct TCP_Server_Info *server = tcon->ses->server;
1617 struct inode *inode = d_inode(cfile->dentry);
1618
1619 if (posix_lck) {
1620 int posix_lock_type;
1621
1622 rc = cifs_posix_lock_set(file, flock);
1623 if (!rc || rc < 0)
1624 return rc;
1625
1626 if (type & server->vals->shared_lock_type)
1627 posix_lock_type = CIFS_RDLCK;
1628 else
1629 posix_lock_type = CIFS_WRLCK;
1630
1631 if (unlock == 1)
1632 posix_lock_type = CIFS_UNLCK;
1633
1634 rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid,
1635 hash_lockowner(flock->fl_owner),
1636 flock->fl_start, length,
1637 NULL, posix_lock_type, wait_flag);
1638 goto out;
1639 }
1640
1641 if (lock) {
1642 struct cifsLockInfo *lock;
1643
David Brazdil0f672f62019-12-10 10:32:29 +00001644 lock = cifs_lock_init(flock->fl_start, length, type,
1645 flock->fl_flags);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001646 if (!lock)
1647 return -ENOMEM;
1648
1649 rc = cifs_lock_add_if(cfile, lock, wait_flag);
1650 if (rc < 0) {
1651 kfree(lock);
1652 return rc;
1653 }
1654 if (!rc)
1655 goto out;
1656
1657 /*
1658 * Windows 7 server can delay breaking lease from read to None
1659 * if we set a byte-range lock on a file - break it explicitly
1660 * before sending the lock to the server to be sure the next
1661 * read won't conflict with non-overlapted locks due to
1662 * pagereading.
1663 */
1664 if (!CIFS_CACHE_WRITE(CIFS_I(inode)) &&
1665 CIFS_CACHE_READ(CIFS_I(inode))) {
1666 cifs_zap_mapping(inode);
1667 cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n",
1668 inode);
1669 CIFS_I(inode)->oplock = 0;
1670 }
1671
1672 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1673 type, 1, 0, wait_flag);
1674 if (rc) {
1675 kfree(lock);
1676 return rc;
1677 }
1678
1679 cifs_lock_add(cfile, lock);
1680 } else if (unlock)
1681 rc = server->ops->mand_unlock_range(cfile, flock, xid);
1682
1683out:
David Brazdil0f672f62019-12-10 10:32:29 +00001684 if (flock->fl_flags & FL_POSIX) {
1685 /*
1686 * If this is a request to remove all locks because we
1687 * are closing the file, it doesn't matter if the
1688 * unlocking failed as both cifs.ko and the SMB server
1689 * remove the lock on file close
1690 */
1691 if (rc) {
1692 cifs_dbg(VFS, "%s failed rc=%d\n", __func__, rc);
1693 if (!(flock->fl_flags & FL_CLOSE))
1694 return rc;
1695 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001696 rc = locks_lock_file_wait(file, flock);
David Brazdil0f672f62019-12-10 10:32:29 +00001697 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001698 return rc;
1699}
1700
1701int cifs_lock(struct file *file, int cmd, struct file_lock *flock)
1702{
1703 int rc, xid;
1704 int lock = 0, unlock = 0;
1705 bool wait_flag = false;
1706 bool posix_lck = false;
1707 struct cifs_sb_info *cifs_sb;
1708 struct cifs_tcon *tcon;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001709 struct cifsFileInfo *cfile;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001710 __u32 type;
1711
1712 rc = -EACCES;
1713 xid = get_xid();
1714
1715 cifs_dbg(FYI, "Lock parm: 0x%x flockflags: 0x%x flocktype: 0x%x start: %lld end: %lld\n",
1716 cmd, flock->fl_flags, flock->fl_type,
1717 flock->fl_start, flock->fl_end);
1718
1719 cfile = (struct cifsFileInfo *)file->private_data;
1720 tcon = tlink_tcon(cfile->tlink);
1721
1722 cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag,
1723 tcon->ses->server);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001724 cifs_sb = CIFS_FILE_SB(file);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001725
1726 if (cap_unix(tcon->ses) &&
1727 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1728 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1729 posix_lck = true;
1730 /*
1731 * BB add code here to normalize offset and length to account for
1732 * negative length which we can not accept over the wire.
1733 */
1734 if (IS_GETLK(cmd)) {
1735 rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid);
1736 free_xid(xid);
1737 return rc;
1738 }
1739
1740 if (!lock && !unlock) {
1741 /*
1742 * if no lock or unlock then nothing to do since we do not
1743 * know what it is
1744 */
1745 free_xid(xid);
1746 return -EOPNOTSUPP;
1747 }
1748
1749 rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock,
1750 xid);
1751 free_xid(xid);
1752 return rc;
1753}
1754
1755/*
1756 * update the file size (if needed) after a write. Should be called with
1757 * the inode->i_lock held
1758 */
1759void
1760cifs_update_eof(struct cifsInodeInfo *cifsi, loff_t offset,
1761 unsigned int bytes_written)
1762{
1763 loff_t end_of_write = offset + bytes_written;
1764
1765 if (end_of_write > cifsi->server_eof)
1766 cifsi->server_eof = end_of_write;
1767}
1768
1769static ssize_t
1770cifs_write(struct cifsFileInfo *open_file, __u32 pid, const char *write_data,
1771 size_t write_size, loff_t *offset)
1772{
1773 int rc = 0;
1774 unsigned int bytes_written = 0;
1775 unsigned int total_written;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001776 struct cifs_tcon *tcon;
1777 struct TCP_Server_Info *server;
1778 unsigned int xid;
1779 struct dentry *dentry = open_file->dentry;
1780 struct cifsInodeInfo *cifsi = CIFS_I(d_inode(dentry));
1781 struct cifs_io_parms io_parms;
1782
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001783 cifs_dbg(FYI, "write %zd bytes to offset %lld of %pd\n",
1784 write_size, *offset, dentry);
1785
1786 tcon = tlink_tcon(open_file->tlink);
1787 server = tcon->ses->server;
1788
1789 if (!server->ops->sync_write)
1790 return -ENOSYS;
1791
1792 xid = get_xid();
1793
1794 for (total_written = 0; write_size > total_written;
1795 total_written += bytes_written) {
1796 rc = -EAGAIN;
1797 while (rc == -EAGAIN) {
1798 struct kvec iov[2];
1799 unsigned int len;
1800
1801 if (open_file->invalidHandle) {
1802 /* we could deadlock if we called
1803 filemap_fdatawait from here so tell
1804 reopen_file not to flush data to
1805 server now */
1806 rc = cifs_reopen_file(open_file, false);
1807 if (rc != 0)
1808 break;
1809 }
1810
1811 len = min(server->ops->wp_retry_size(d_inode(dentry)),
1812 (unsigned int)write_size - total_written);
1813 /* iov[0] is reserved for smb header */
1814 iov[1].iov_base = (char *)write_data + total_written;
1815 iov[1].iov_len = len;
1816 io_parms.pid = pid;
1817 io_parms.tcon = tcon;
1818 io_parms.offset = *offset;
1819 io_parms.length = len;
1820 rc = server->ops->sync_write(xid, &open_file->fid,
1821 &io_parms, &bytes_written, iov, 1);
1822 }
1823 if (rc || (bytes_written == 0)) {
1824 if (total_written)
1825 break;
1826 else {
1827 free_xid(xid);
1828 return rc;
1829 }
1830 } else {
1831 spin_lock(&d_inode(dentry)->i_lock);
1832 cifs_update_eof(cifsi, *offset, bytes_written);
1833 spin_unlock(&d_inode(dentry)->i_lock);
1834 *offset += bytes_written;
1835 }
1836 }
1837
1838 cifs_stats_bytes_written(tcon, total_written);
1839
1840 if (total_written > 0) {
1841 spin_lock(&d_inode(dentry)->i_lock);
1842 if (*offset > d_inode(dentry)->i_size)
1843 i_size_write(d_inode(dentry), *offset);
1844 spin_unlock(&d_inode(dentry)->i_lock);
1845 }
1846 mark_inode_dirty_sync(d_inode(dentry));
1847 free_xid(xid);
1848 return total_written;
1849}
1850
1851struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode,
1852 bool fsuid_only)
1853{
1854 struct cifsFileInfo *open_file = NULL;
1855 struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001856
1857 /* only filter by fsuid on multiuser mounts */
1858 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1859 fsuid_only = false;
1860
David Brazdil0f672f62019-12-10 10:32:29 +00001861 spin_lock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001862 /* we could simply get the first_list_entry since write-only entries
1863 are always at the end of the list but since the first entry might
1864 have a close pending, we go through the whole list */
1865 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
1866 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
1867 continue;
1868 if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) {
1869 if (!open_file->invalidHandle) {
1870 /* found a good file */
1871 /* lock it so it will not be closed on us */
1872 cifsFileInfo_get(open_file);
David Brazdil0f672f62019-12-10 10:32:29 +00001873 spin_unlock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001874 return open_file;
1875 } /* else might as well continue, and look for
1876 another, or simply have the caller reopen it
1877 again rather than trying to fix this handle */
1878 } else /* write only file */
1879 break; /* write only files are last so must be done */
1880 }
David Brazdil0f672f62019-12-10 10:32:29 +00001881 spin_unlock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001882 return NULL;
1883}
1884
David Brazdil0f672f62019-12-10 10:32:29 +00001885/* Return -EBADF if no handle is found and general rc otherwise */
1886int
1887cifs_get_writable_file(struct cifsInodeInfo *cifs_inode, bool fsuid_only,
1888 struct cifsFileInfo **ret_file)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001889{
1890 struct cifsFileInfo *open_file, *inv_file = NULL;
1891 struct cifs_sb_info *cifs_sb;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001892 bool any_available = false;
David Brazdil0f672f62019-12-10 10:32:29 +00001893 int rc = -EBADF;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001894 unsigned int refind = 0;
1895
David Brazdil0f672f62019-12-10 10:32:29 +00001896 *ret_file = NULL;
1897
1898 /*
1899 * Having a null inode here (because mapping->host was set to zero by
1900 * the VFS or MM) should not happen but we had reports of on oops (due
1901 * to it being zero) during stress testcases so we need to check for it
1902 */
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001903
1904 if (cifs_inode == NULL) {
1905 cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n");
1906 dump_stack();
David Brazdil0f672f62019-12-10 10:32:29 +00001907 return rc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001908 }
1909
1910 cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001911
1912 /* only filter by fsuid on multiuser mounts */
1913 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1914 fsuid_only = false;
1915
David Brazdil0f672f62019-12-10 10:32:29 +00001916 spin_lock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001917refind_writable:
1918 if (refind > MAX_REOPEN_ATT) {
David Brazdil0f672f62019-12-10 10:32:29 +00001919 spin_unlock(&cifs_inode->open_file_lock);
1920 return rc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001921 }
1922 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
1923 if (!any_available && open_file->pid != current->tgid)
1924 continue;
1925 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
1926 continue;
1927 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
1928 if (!open_file->invalidHandle) {
1929 /* found a good writable file */
1930 cifsFileInfo_get(open_file);
David Brazdil0f672f62019-12-10 10:32:29 +00001931 spin_unlock(&cifs_inode->open_file_lock);
1932 *ret_file = open_file;
1933 return 0;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001934 } else {
1935 if (!inv_file)
1936 inv_file = open_file;
1937 }
1938 }
1939 }
1940 /* couldn't find useable FH with same pid, try any available */
1941 if (!any_available) {
1942 any_available = true;
1943 goto refind_writable;
1944 }
1945
1946 if (inv_file) {
1947 any_available = false;
1948 cifsFileInfo_get(inv_file);
1949 }
1950
David Brazdil0f672f62019-12-10 10:32:29 +00001951 spin_unlock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001952
1953 if (inv_file) {
1954 rc = cifs_reopen_file(inv_file, false);
David Brazdil0f672f62019-12-10 10:32:29 +00001955 if (!rc) {
1956 *ret_file = inv_file;
1957 return 0;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001958 }
David Brazdil0f672f62019-12-10 10:32:29 +00001959
1960 spin_lock(&cifs_inode->open_file_lock);
1961 list_move_tail(&inv_file->flist, &cifs_inode->openFileList);
1962 spin_unlock(&cifs_inode->open_file_lock);
1963 cifsFileInfo_put(inv_file);
1964 ++refind;
1965 inv_file = NULL;
1966 spin_lock(&cifs_inode->open_file_lock);
1967 goto refind_writable;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001968 }
1969
David Brazdil0f672f62019-12-10 10:32:29 +00001970 return rc;
1971}
1972
1973struct cifsFileInfo *
1974find_writable_file(struct cifsInodeInfo *cifs_inode, bool fsuid_only)
1975{
1976 struct cifsFileInfo *cfile;
1977 int rc;
1978
1979 rc = cifs_get_writable_file(cifs_inode, fsuid_only, &cfile);
1980 if (rc)
1981 cifs_dbg(FYI, "couldn't find writable handle rc=%d", rc);
1982
1983 return cfile;
1984}
1985
1986int
1987cifs_get_writable_path(struct cifs_tcon *tcon, const char *name,
1988 struct cifsFileInfo **ret_file)
1989{
1990 struct list_head *tmp;
1991 struct cifsFileInfo *cfile;
1992 struct cifsInodeInfo *cinode;
1993 char *full_path;
1994
1995 *ret_file = NULL;
1996
1997 spin_lock(&tcon->open_file_lock);
1998 list_for_each(tmp, &tcon->openFileList) {
1999 cfile = list_entry(tmp, struct cifsFileInfo,
2000 tlist);
2001 full_path = build_path_from_dentry(cfile->dentry);
2002 if (full_path == NULL) {
2003 spin_unlock(&tcon->open_file_lock);
2004 return -ENOMEM;
2005 }
2006 if (strcmp(full_path, name)) {
2007 kfree(full_path);
2008 continue;
2009 }
2010
2011 kfree(full_path);
2012 cinode = CIFS_I(d_inode(cfile->dentry));
2013 spin_unlock(&tcon->open_file_lock);
2014 return cifs_get_writable_file(cinode, 0, ret_file);
2015 }
2016
2017 spin_unlock(&tcon->open_file_lock);
2018 return -ENOENT;
2019}
2020
2021int
2022cifs_get_readable_path(struct cifs_tcon *tcon, const char *name,
2023 struct cifsFileInfo **ret_file)
2024{
2025 struct list_head *tmp;
2026 struct cifsFileInfo *cfile;
2027 struct cifsInodeInfo *cinode;
2028 char *full_path;
2029
2030 *ret_file = NULL;
2031
2032 spin_lock(&tcon->open_file_lock);
2033 list_for_each(tmp, &tcon->openFileList) {
2034 cfile = list_entry(tmp, struct cifsFileInfo,
2035 tlist);
2036 full_path = build_path_from_dentry(cfile->dentry);
2037 if (full_path == NULL) {
2038 spin_unlock(&tcon->open_file_lock);
2039 return -ENOMEM;
2040 }
2041 if (strcmp(full_path, name)) {
2042 kfree(full_path);
2043 continue;
2044 }
2045
2046 kfree(full_path);
2047 cinode = CIFS_I(d_inode(cfile->dentry));
2048 spin_unlock(&tcon->open_file_lock);
2049 *ret_file = find_readable_file(cinode, 0);
2050 return *ret_file ? 0 : -ENOENT;
2051 }
2052
2053 spin_unlock(&tcon->open_file_lock);
2054 return -ENOENT;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002055}
2056
2057static int cifs_partialpagewrite(struct page *page, unsigned from, unsigned to)
2058{
2059 struct address_space *mapping = page->mapping;
2060 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
2061 char *write_data;
2062 int rc = -EFAULT;
2063 int bytes_written = 0;
2064 struct inode *inode;
2065 struct cifsFileInfo *open_file;
2066
2067 if (!mapping || !mapping->host)
2068 return -EFAULT;
2069
2070 inode = page->mapping->host;
2071
2072 offset += (loff_t)from;
2073 write_data = kmap(page);
2074 write_data += from;
2075
2076 if ((to > PAGE_SIZE) || (from > to)) {
2077 kunmap(page);
2078 return -EIO;
2079 }
2080
2081 /* racing with truncate? */
2082 if (offset > mapping->host->i_size) {
2083 kunmap(page);
2084 return 0; /* don't care */
2085 }
2086
2087 /* check to make sure that we are not extending the file */
2088 if (mapping->host->i_size - offset < (loff_t)to)
2089 to = (unsigned)(mapping->host->i_size - offset);
2090
David Brazdil0f672f62019-12-10 10:32:29 +00002091 rc = cifs_get_writable_file(CIFS_I(mapping->host), false, &open_file);
2092 if (!rc) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002093 bytes_written = cifs_write(open_file, open_file->pid,
2094 write_data, to - from, &offset);
2095 cifsFileInfo_put(open_file);
2096 /* Does mm or vfs already set times? */
2097 inode->i_atime = inode->i_mtime = current_time(inode);
2098 if ((bytes_written > 0) && (offset))
2099 rc = 0;
2100 else if (bytes_written < 0)
2101 rc = bytes_written;
David Brazdil0f672f62019-12-10 10:32:29 +00002102 else
2103 rc = -EFAULT;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002104 } else {
David Brazdil0f672f62019-12-10 10:32:29 +00002105 cifs_dbg(FYI, "No writable handle for write page rc=%d\n", rc);
2106 if (!is_retryable_error(rc))
2107 rc = -EIO;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002108 }
2109
2110 kunmap(page);
2111 return rc;
2112}
2113
2114static struct cifs_writedata *
2115wdata_alloc_and_fillpages(pgoff_t tofind, struct address_space *mapping,
2116 pgoff_t end, pgoff_t *index,
2117 unsigned int *found_pages)
2118{
2119 struct cifs_writedata *wdata;
2120
2121 wdata = cifs_writedata_alloc((unsigned int)tofind,
2122 cifs_writev_complete);
2123 if (!wdata)
2124 return NULL;
2125
2126 *found_pages = find_get_pages_range_tag(mapping, index, end,
2127 PAGECACHE_TAG_DIRTY, tofind, wdata->pages);
2128 return wdata;
2129}
2130
2131static unsigned int
2132wdata_prepare_pages(struct cifs_writedata *wdata, unsigned int found_pages,
2133 struct address_space *mapping,
2134 struct writeback_control *wbc,
2135 pgoff_t end, pgoff_t *index, pgoff_t *next, bool *done)
2136{
2137 unsigned int nr_pages = 0, i;
2138 struct page *page;
2139
2140 for (i = 0; i < found_pages; i++) {
2141 page = wdata->pages[i];
2142 /*
2143 * At this point we hold neither the i_pages lock nor the
2144 * page lock: the page may be truncated or invalidated
2145 * (changing page->mapping to NULL), or even swizzled
2146 * back from swapper_space to tmpfs file mapping
2147 */
2148
2149 if (nr_pages == 0)
2150 lock_page(page);
2151 else if (!trylock_page(page))
2152 break;
2153
2154 if (unlikely(page->mapping != mapping)) {
2155 unlock_page(page);
2156 break;
2157 }
2158
2159 if (!wbc->range_cyclic && page->index > end) {
2160 *done = true;
2161 unlock_page(page);
2162 break;
2163 }
2164
2165 if (*next && (page->index != *next)) {
2166 /* Not next consecutive page */
2167 unlock_page(page);
2168 break;
2169 }
2170
2171 if (wbc->sync_mode != WB_SYNC_NONE)
2172 wait_on_page_writeback(page);
2173
2174 if (PageWriteback(page) ||
2175 !clear_page_dirty_for_io(page)) {
2176 unlock_page(page);
2177 break;
2178 }
2179
2180 /*
2181 * This actually clears the dirty bit in the radix tree.
2182 * See cifs_writepage() for more commentary.
2183 */
2184 set_page_writeback(page);
2185 if (page_offset(page) >= i_size_read(mapping->host)) {
2186 *done = true;
2187 unlock_page(page);
2188 end_page_writeback(page);
2189 break;
2190 }
2191
2192 wdata->pages[i] = page;
2193 *next = page->index + 1;
2194 ++nr_pages;
2195 }
2196
2197 /* reset index to refind any pages skipped */
2198 if (nr_pages == 0)
2199 *index = wdata->pages[0]->index + 1;
2200
2201 /* put any pages we aren't going to use */
2202 for (i = nr_pages; i < found_pages; i++) {
2203 put_page(wdata->pages[i]);
2204 wdata->pages[i] = NULL;
2205 }
2206
2207 return nr_pages;
2208}
2209
2210static int
2211wdata_send_pages(struct cifs_writedata *wdata, unsigned int nr_pages,
2212 struct address_space *mapping, struct writeback_control *wbc)
2213{
David Brazdil0f672f62019-12-10 10:32:29 +00002214 int rc;
2215 struct TCP_Server_Info *server =
2216 tlink_tcon(wdata->cfile->tlink)->ses->server;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002217
2218 wdata->sync_mode = wbc->sync_mode;
2219 wdata->nr_pages = nr_pages;
2220 wdata->offset = page_offset(wdata->pages[0]);
2221 wdata->pagesz = PAGE_SIZE;
2222 wdata->tailsz = min(i_size_read(mapping->host) -
2223 page_offset(wdata->pages[nr_pages - 1]),
2224 (loff_t)PAGE_SIZE);
2225 wdata->bytes = ((nr_pages - 1) * PAGE_SIZE) + wdata->tailsz;
David Brazdil0f672f62019-12-10 10:32:29 +00002226 wdata->pid = wdata->cfile->pid;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002227
David Brazdil0f672f62019-12-10 10:32:29 +00002228 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2229 if (rc)
2230 return rc;
2231
2232 if (wdata->cfile->invalidHandle)
2233 rc = -EAGAIN;
2234 else
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002235 rc = server->ops->async_writev(wdata, cifs_writedata_release);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002236
2237 return rc;
2238}
2239
2240static int cifs_writepages(struct address_space *mapping,
2241 struct writeback_control *wbc)
2242{
David Brazdil0f672f62019-12-10 10:32:29 +00002243 struct inode *inode = mapping->host;
2244 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002245 struct TCP_Server_Info *server;
2246 bool done = false, scanned = false, range_whole = false;
2247 pgoff_t end, index;
2248 struct cifs_writedata *wdata;
David Brazdil0f672f62019-12-10 10:32:29 +00002249 struct cifsFileInfo *cfile = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002250 int rc = 0;
David Brazdil0f672f62019-12-10 10:32:29 +00002251 int saved_rc = 0;
2252 unsigned int xid;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002253
2254 /*
2255 * If wsize is smaller than the page cache size, default to writing
2256 * one page at a time via cifs_writepage
2257 */
2258 if (cifs_sb->wsize < PAGE_SIZE)
2259 return generic_writepages(mapping, wbc);
2260
David Brazdil0f672f62019-12-10 10:32:29 +00002261 xid = get_xid();
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002262 if (wbc->range_cyclic) {
2263 index = mapping->writeback_index; /* Start from prev offset */
2264 end = -1;
2265 } else {
2266 index = wbc->range_start >> PAGE_SHIFT;
2267 end = wbc->range_end >> PAGE_SHIFT;
2268 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX)
2269 range_whole = true;
2270 scanned = true;
2271 }
2272 server = cifs_sb_master_tcon(cifs_sb)->ses->server;
2273retry:
2274 while (!done && index <= end) {
David Brazdil0f672f62019-12-10 10:32:29 +00002275 unsigned int i, nr_pages, found_pages, wsize;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002276 pgoff_t next = 0, tofind, saved_index = index;
David Brazdil0f672f62019-12-10 10:32:29 +00002277 struct cifs_credits credits_on_stack;
2278 struct cifs_credits *credits = &credits_on_stack;
2279 int get_file_rc = 0;
2280
2281 if (cfile)
2282 cifsFileInfo_put(cfile);
2283
2284 rc = cifs_get_writable_file(CIFS_I(inode), false, &cfile);
2285
2286 /* in case of an error store it to return later */
2287 if (rc)
2288 get_file_rc = rc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002289
2290 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
David Brazdil0f672f62019-12-10 10:32:29 +00002291 &wsize, credits);
2292 if (rc != 0) {
2293 done = true;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002294 break;
David Brazdil0f672f62019-12-10 10:32:29 +00002295 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002296
2297 tofind = min((wsize / PAGE_SIZE) - 1, end - index) + 1;
2298
2299 wdata = wdata_alloc_and_fillpages(tofind, mapping, end, &index,
2300 &found_pages);
2301 if (!wdata) {
2302 rc = -ENOMEM;
David Brazdil0f672f62019-12-10 10:32:29 +00002303 done = true;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002304 add_credits_and_wake_if(server, credits, 0);
2305 break;
2306 }
2307
2308 if (found_pages == 0) {
2309 kref_put(&wdata->refcount, cifs_writedata_release);
2310 add_credits_and_wake_if(server, credits, 0);
2311 break;
2312 }
2313
2314 nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc,
2315 end, &index, &next, &done);
2316
2317 /* nothing to write? */
2318 if (nr_pages == 0) {
2319 kref_put(&wdata->refcount, cifs_writedata_release);
2320 add_credits_and_wake_if(server, credits, 0);
2321 continue;
2322 }
2323
David Brazdil0f672f62019-12-10 10:32:29 +00002324 wdata->credits = credits_on_stack;
2325 wdata->cfile = cfile;
2326 cfile = NULL;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002327
David Brazdil0f672f62019-12-10 10:32:29 +00002328 if (!wdata->cfile) {
2329 cifs_dbg(VFS, "No writable handle in writepages rc=%d\n",
2330 get_file_rc);
2331 if (is_retryable_error(get_file_rc))
2332 rc = get_file_rc;
2333 else
2334 rc = -EBADF;
2335 } else
2336 rc = wdata_send_pages(wdata, nr_pages, mapping, wbc);
2337
2338 for (i = 0; i < nr_pages; ++i)
2339 unlock_page(wdata->pages[i]);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002340
2341 /* send failure -- clean up the mess */
2342 if (rc != 0) {
David Brazdil0f672f62019-12-10 10:32:29 +00002343 add_credits_and_wake_if(server, &wdata->credits, 0);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002344 for (i = 0; i < nr_pages; ++i) {
David Brazdil0f672f62019-12-10 10:32:29 +00002345 if (is_retryable_error(rc))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002346 redirty_page_for_writepage(wbc,
2347 wdata->pages[i]);
2348 else
2349 SetPageError(wdata->pages[i]);
2350 end_page_writeback(wdata->pages[i]);
2351 put_page(wdata->pages[i]);
2352 }
David Brazdil0f672f62019-12-10 10:32:29 +00002353 if (!is_retryable_error(rc))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002354 mapping_set_error(mapping, rc);
2355 }
2356 kref_put(&wdata->refcount, cifs_writedata_release);
2357
2358 if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN) {
2359 index = saved_index;
2360 continue;
2361 }
2362
David Brazdil0f672f62019-12-10 10:32:29 +00002363 /* Return immediately if we received a signal during writing */
2364 if (is_interrupt_error(rc)) {
2365 done = true;
2366 break;
2367 }
2368
2369 if (rc != 0 && saved_rc == 0)
2370 saved_rc = rc;
2371
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002372 wbc->nr_to_write -= nr_pages;
2373 if (wbc->nr_to_write <= 0)
2374 done = true;
2375
2376 index = next;
2377 }
2378
2379 if (!scanned && !done) {
2380 /*
2381 * We hit the last page and there is more work to be done: wrap
2382 * back to the start of the file
2383 */
2384 scanned = true;
2385 index = 0;
2386 goto retry;
2387 }
2388
David Brazdil0f672f62019-12-10 10:32:29 +00002389 if (saved_rc != 0)
2390 rc = saved_rc;
2391
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002392 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0))
2393 mapping->writeback_index = index;
2394
David Brazdil0f672f62019-12-10 10:32:29 +00002395 if (cfile)
2396 cifsFileInfo_put(cfile);
2397 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002398 return rc;
2399}
2400
2401static int
2402cifs_writepage_locked(struct page *page, struct writeback_control *wbc)
2403{
2404 int rc;
2405 unsigned int xid;
2406
2407 xid = get_xid();
2408/* BB add check for wbc flags */
2409 get_page(page);
2410 if (!PageUptodate(page))
2411 cifs_dbg(FYI, "ppw - page not up to date\n");
2412
2413 /*
2414 * Set the "writeback" flag, and clear "dirty" in the radix tree.
2415 *
2416 * A writepage() implementation always needs to do either this,
2417 * or re-dirty the page with "redirty_page_for_writepage()" in
2418 * the case of a failure.
2419 *
2420 * Just unlocking the page will cause the radix tree tag-bits
2421 * to fail to update with the state of the page correctly.
2422 */
2423 set_page_writeback(page);
2424retry_write:
2425 rc = cifs_partialpagewrite(page, 0, PAGE_SIZE);
David Brazdil0f672f62019-12-10 10:32:29 +00002426 if (is_retryable_error(rc)) {
2427 if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002428 goto retry_write;
2429 redirty_page_for_writepage(wbc, page);
2430 } else if (rc != 0) {
2431 SetPageError(page);
2432 mapping_set_error(page->mapping, rc);
2433 } else {
2434 SetPageUptodate(page);
2435 }
2436 end_page_writeback(page);
2437 put_page(page);
2438 free_xid(xid);
2439 return rc;
2440}
2441
2442static int cifs_writepage(struct page *page, struct writeback_control *wbc)
2443{
2444 int rc = cifs_writepage_locked(page, wbc);
2445 unlock_page(page);
2446 return rc;
2447}
2448
2449static int cifs_write_end(struct file *file, struct address_space *mapping,
2450 loff_t pos, unsigned len, unsigned copied,
2451 struct page *page, void *fsdata)
2452{
2453 int rc;
2454 struct inode *inode = mapping->host;
2455 struct cifsFileInfo *cfile = file->private_data;
2456 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2457 __u32 pid;
2458
2459 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2460 pid = cfile->pid;
2461 else
2462 pid = current->tgid;
2463
2464 cifs_dbg(FYI, "write_end for page %p from pos %lld with %d bytes\n",
2465 page, pos, copied);
2466
2467 if (PageChecked(page)) {
2468 if (copied == len)
2469 SetPageUptodate(page);
2470 ClearPageChecked(page);
2471 } else if (!PageUptodate(page) && copied == PAGE_SIZE)
2472 SetPageUptodate(page);
2473
2474 if (!PageUptodate(page)) {
2475 char *page_data;
2476 unsigned offset = pos & (PAGE_SIZE - 1);
2477 unsigned int xid;
2478
2479 xid = get_xid();
2480 /* this is probably better than directly calling
2481 partialpage_write since in this function the file handle is
2482 known which we might as well leverage */
2483 /* BB check if anything else missing out of ppw
2484 such as updating last write time */
2485 page_data = kmap(page);
2486 rc = cifs_write(cfile, pid, page_data + offset, copied, &pos);
2487 /* if (rc < 0) should we set writebehind rc? */
2488 kunmap(page);
2489
2490 free_xid(xid);
2491 } else {
2492 rc = copied;
2493 pos += copied;
2494 set_page_dirty(page);
2495 }
2496
2497 if (rc > 0) {
2498 spin_lock(&inode->i_lock);
2499 if (pos > inode->i_size)
2500 i_size_write(inode, pos);
2501 spin_unlock(&inode->i_lock);
2502 }
2503
2504 unlock_page(page);
2505 put_page(page);
2506
2507 return rc;
2508}
2509
2510int cifs_strict_fsync(struct file *file, loff_t start, loff_t end,
2511 int datasync)
2512{
2513 unsigned int xid;
2514 int rc = 0;
2515 struct cifs_tcon *tcon;
2516 struct TCP_Server_Info *server;
2517 struct cifsFileInfo *smbfile = file->private_data;
2518 struct inode *inode = file_inode(file);
2519 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2520
2521 rc = file_write_and_wait_range(file, start, end);
2522 if (rc)
2523 return rc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002524
2525 xid = get_xid();
2526
2527 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2528 file, datasync);
2529
2530 if (!CIFS_CACHE_READ(CIFS_I(inode))) {
2531 rc = cifs_zap_mapping(inode);
2532 if (rc) {
2533 cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc);
2534 rc = 0; /* don't care about it in fsync */
2535 }
2536 }
2537
2538 tcon = tlink_tcon(smbfile->tlink);
2539 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2540 server = tcon->ses->server;
2541 if (server->ops->flush)
2542 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2543 else
2544 rc = -ENOSYS;
2545 }
2546
2547 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002548 return rc;
2549}
2550
2551int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
2552{
2553 unsigned int xid;
2554 int rc = 0;
2555 struct cifs_tcon *tcon;
2556 struct TCP_Server_Info *server;
2557 struct cifsFileInfo *smbfile = file->private_data;
2558 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002559
2560 rc = file_write_and_wait_range(file, start, end);
2561 if (rc)
2562 return rc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002563
2564 xid = get_xid();
2565
2566 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2567 file, datasync);
2568
2569 tcon = tlink_tcon(smbfile->tlink);
2570 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2571 server = tcon->ses->server;
2572 if (server->ops->flush)
2573 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2574 else
2575 rc = -ENOSYS;
2576 }
2577
2578 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002579 return rc;
2580}
2581
2582/*
2583 * As file closes, flush all cached write data for this inode checking
2584 * for write behind errors.
2585 */
2586int cifs_flush(struct file *file, fl_owner_t id)
2587{
2588 struct inode *inode = file_inode(file);
2589 int rc = 0;
2590
2591 if (file->f_mode & FMODE_WRITE)
2592 rc = filemap_write_and_wait(inode->i_mapping);
2593
2594 cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc);
2595
2596 return rc;
2597}
2598
2599static int
2600cifs_write_allocate_pages(struct page **pages, unsigned long num_pages)
2601{
2602 int rc = 0;
2603 unsigned long i;
2604
2605 for (i = 0; i < num_pages; i++) {
2606 pages[i] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
2607 if (!pages[i]) {
2608 /*
2609 * save number of pages we have already allocated and
2610 * return with ENOMEM error
2611 */
2612 num_pages = i;
2613 rc = -ENOMEM;
2614 break;
2615 }
2616 }
2617
2618 if (rc) {
2619 for (i = 0; i < num_pages; i++)
2620 put_page(pages[i]);
2621 }
2622 return rc;
2623}
2624
2625static inline
2626size_t get_numpages(const size_t wsize, const size_t len, size_t *cur_len)
2627{
2628 size_t num_pages;
2629 size_t clen;
2630
2631 clen = min_t(const size_t, len, wsize);
2632 num_pages = DIV_ROUND_UP(clen, PAGE_SIZE);
2633
2634 if (cur_len)
2635 *cur_len = clen;
2636
2637 return num_pages;
2638}
2639
2640static void
2641cifs_uncached_writedata_release(struct kref *refcount)
2642{
2643 int i;
2644 struct cifs_writedata *wdata = container_of(refcount,
2645 struct cifs_writedata, refcount);
2646
2647 kref_put(&wdata->ctx->refcount, cifs_aio_ctx_release);
2648 for (i = 0; i < wdata->nr_pages; i++)
2649 put_page(wdata->pages[i]);
2650 cifs_writedata_release(refcount);
2651}
2652
2653static void collect_uncached_write_data(struct cifs_aio_ctx *ctx);
2654
2655static void
2656cifs_uncached_writev_complete(struct work_struct *work)
2657{
2658 struct cifs_writedata *wdata = container_of(work,
2659 struct cifs_writedata, work);
2660 struct inode *inode = d_inode(wdata->cfile->dentry);
2661 struct cifsInodeInfo *cifsi = CIFS_I(inode);
2662
2663 spin_lock(&inode->i_lock);
2664 cifs_update_eof(cifsi, wdata->offset, wdata->bytes);
2665 if (cifsi->server_eof > inode->i_size)
2666 i_size_write(inode, cifsi->server_eof);
2667 spin_unlock(&inode->i_lock);
2668
2669 complete(&wdata->done);
2670 collect_uncached_write_data(wdata->ctx);
2671 /* the below call can possibly free the last ref to aio ctx */
2672 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
2673}
2674
2675static int
2676wdata_fill_from_iovec(struct cifs_writedata *wdata, struct iov_iter *from,
2677 size_t *len, unsigned long *num_pages)
2678{
2679 size_t save_len, copied, bytes, cur_len = *len;
2680 unsigned long i, nr_pages = *num_pages;
2681
2682 save_len = cur_len;
2683 for (i = 0; i < nr_pages; i++) {
2684 bytes = min_t(const size_t, cur_len, PAGE_SIZE);
2685 copied = copy_page_from_iter(wdata->pages[i], 0, bytes, from);
2686 cur_len -= copied;
2687 /*
2688 * If we didn't copy as much as we expected, then that
2689 * may mean we trod into an unmapped area. Stop copying
2690 * at that point. On the next pass through the big
2691 * loop, we'll likely end up getting a zero-length
2692 * write and bailing out of it.
2693 */
2694 if (copied < bytes)
2695 break;
2696 }
2697 cur_len = save_len - cur_len;
2698 *len = cur_len;
2699
2700 /*
2701 * If we have no data to send, then that probably means that
2702 * the copy above failed altogether. That's most likely because
2703 * the address in the iovec was bogus. Return -EFAULT and let
2704 * the caller free anything we allocated and bail out.
2705 */
2706 if (!cur_len)
2707 return -EFAULT;
2708
2709 /*
2710 * i + 1 now represents the number of pages we actually used in
2711 * the copy phase above.
2712 */
2713 *num_pages = i + 1;
2714 return 0;
2715}
2716
2717static int
David Brazdil0f672f62019-12-10 10:32:29 +00002718cifs_resend_wdata(struct cifs_writedata *wdata, struct list_head *wdata_list,
2719 struct cifs_aio_ctx *ctx)
2720{
2721 unsigned int wsize;
2722 struct cifs_credits credits;
2723 int rc;
2724 struct TCP_Server_Info *server =
2725 tlink_tcon(wdata->cfile->tlink)->ses->server;
2726
2727 do {
2728 if (wdata->cfile->invalidHandle) {
2729 rc = cifs_reopen_file(wdata->cfile, false);
2730 if (rc == -EAGAIN)
2731 continue;
2732 else if (rc)
2733 break;
2734 }
2735
2736
2737 /*
2738 * Wait for credits to resend this wdata.
2739 * Note: we are attempting to resend the whole wdata not in
2740 * segments
2741 */
2742 do {
2743 rc = server->ops->wait_mtu_credits(server, wdata->bytes,
2744 &wsize, &credits);
2745 if (rc)
2746 goto fail;
2747
2748 if (wsize < wdata->bytes) {
2749 add_credits_and_wake_if(server, &credits, 0);
2750 msleep(1000);
2751 }
2752 } while (wsize < wdata->bytes);
2753 wdata->credits = credits;
2754
2755 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2756
2757 if (!rc) {
2758 if (wdata->cfile->invalidHandle)
2759 rc = -EAGAIN;
2760 else
2761 rc = server->ops->async_writev(wdata,
2762 cifs_uncached_writedata_release);
2763 }
2764
2765 /* If the write was successfully sent, we are done */
2766 if (!rc) {
2767 list_add_tail(&wdata->list, wdata_list);
2768 return 0;
2769 }
2770
2771 /* Roll back credits and retry if needed */
2772 add_credits_and_wake_if(server, &wdata->credits, 0);
2773 } while (rc == -EAGAIN);
2774
2775fail:
2776 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
2777 return rc;
2778}
2779
2780static int
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002781cifs_write_from_iter(loff_t offset, size_t len, struct iov_iter *from,
2782 struct cifsFileInfo *open_file,
2783 struct cifs_sb_info *cifs_sb, struct list_head *wdata_list,
2784 struct cifs_aio_ctx *ctx)
2785{
2786 int rc = 0;
2787 size_t cur_len;
2788 unsigned long nr_pages, num_pages, i;
2789 struct cifs_writedata *wdata;
2790 struct iov_iter saved_from = *from;
2791 loff_t saved_offset = offset;
2792 pid_t pid;
2793 struct TCP_Server_Info *server;
David Brazdil0f672f62019-12-10 10:32:29 +00002794 struct page **pagevec;
2795 size_t start;
2796 unsigned int xid;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002797
2798 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2799 pid = open_file->pid;
2800 else
2801 pid = current->tgid;
2802
2803 server = tlink_tcon(open_file->tlink)->ses->server;
David Brazdil0f672f62019-12-10 10:32:29 +00002804 xid = get_xid();
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002805
2806 do {
David Brazdil0f672f62019-12-10 10:32:29 +00002807 unsigned int wsize;
2808 struct cifs_credits credits_on_stack;
2809 struct cifs_credits *credits = &credits_on_stack;
2810
2811 if (open_file->invalidHandle) {
2812 rc = cifs_reopen_file(open_file, false);
2813 if (rc == -EAGAIN)
2814 continue;
2815 else if (rc)
2816 break;
2817 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002818
2819 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
David Brazdil0f672f62019-12-10 10:32:29 +00002820 &wsize, credits);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002821 if (rc)
2822 break;
2823
David Brazdil0f672f62019-12-10 10:32:29 +00002824 cur_len = min_t(const size_t, len, wsize);
2825
2826 if (ctx->direct_io) {
2827 ssize_t result;
2828
2829 result = iov_iter_get_pages_alloc(
2830 from, &pagevec, cur_len, &start);
2831 if (result < 0) {
2832 cifs_dbg(VFS,
2833 "direct_writev couldn't get user pages "
2834 "(rc=%zd) iter type %d iov_offset %zd "
2835 "count %zd\n",
2836 result, from->type,
2837 from->iov_offset, from->count);
2838 dump_stack();
2839
2840 rc = result;
2841 add_credits_and_wake_if(server, credits, 0);
2842 break;
2843 }
2844 cur_len = (size_t)result;
2845 iov_iter_advance(from, cur_len);
2846
2847 nr_pages =
2848 (cur_len + start + PAGE_SIZE - 1) / PAGE_SIZE;
2849
2850 wdata = cifs_writedata_direct_alloc(pagevec,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002851 cifs_uncached_writev_complete);
David Brazdil0f672f62019-12-10 10:32:29 +00002852 if (!wdata) {
2853 rc = -ENOMEM;
2854 add_credits_and_wake_if(server, credits, 0);
2855 break;
2856 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002857
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002858
David Brazdil0f672f62019-12-10 10:32:29 +00002859 wdata->page_offset = start;
2860 wdata->tailsz =
2861 nr_pages > 1 ?
2862 cur_len - (PAGE_SIZE - start) -
2863 (nr_pages - 2) * PAGE_SIZE :
2864 cur_len;
2865 } else {
2866 nr_pages = get_numpages(wsize, len, &cur_len);
2867 wdata = cifs_writedata_alloc(nr_pages,
2868 cifs_uncached_writev_complete);
2869 if (!wdata) {
2870 rc = -ENOMEM;
2871 add_credits_and_wake_if(server, credits, 0);
2872 break;
2873 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002874
David Brazdil0f672f62019-12-10 10:32:29 +00002875 rc = cifs_write_allocate_pages(wdata->pages, nr_pages);
2876 if (rc) {
2877 kvfree(wdata->pages);
2878 kfree(wdata);
2879 add_credits_and_wake_if(server, credits, 0);
2880 break;
2881 }
2882
2883 num_pages = nr_pages;
2884 rc = wdata_fill_from_iovec(
2885 wdata, from, &cur_len, &num_pages);
2886 if (rc) {
2887 for (i = 0; i < nr_pages; i++)
2888 put_page(wdata->pages[i]);
2889 kvfree(wdata->pages);
2890 kfree(wdata);
2891 add_credits_and_wake_if(server, credits, 0);
2892 break;
2893 }
2894
2895 /*
2896 * Bring nr_pages down to the number of pages we
2897 * actually used, and free any pages that we didn't use.
2898 */
2899 for ( ; nr_pages > num_pages; nr_pages--)
2900 put_page(wdata->pages[nr_pages - 1]);
2901
2902 wdata->tailsz = cur_len - ((nr_pages - 1) * PAGE_SIZE);
2903 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002904
2905 wdata->sync_mode = WB_SYNC_ALL;
2906 wdata->nr_pages = nr_pages;
2907 wdata->offset = (__u64)offset;
2908 wdata->cfile = cifsFileInfo_get(open_file);
2909 wdata->pid = pid;
2910 wdata->bytes = cur_len;
2911 wdata->pagesz = PAGE_SIZE;
David Brazdil0f672f62019-12-10 10:32:29 +00002912 wdata->credits = credits_on_stack;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002913 wdata->ctx = ctx;
2914 kref_get(&ctx->refcount);
2915
David Brazdil0f672f62019-12-10 10:32:29 +00002916 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2917
2918 if (!rc) {
2919 if (wdata->cfile->invalidHandle)
2920 rc = -EAGAIN;
2921 else
2922 rc = server->ops->async_writev(wdata,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002923 cifs_uncached_writedata_release);
David Brazdil0f672f62019-12-10 10:32:29 +00002924 }
2925
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002926 if (rc) {
David Brazdil0f672f62019-12-10 10:32:29 +00002927 add_credits_and_wake_if(server, &wdata->credits, 0);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002928 kref_put(&wdata->refcount,
2929 cifs_uncached_writedata_release);
2930 if (rc == -EAGAIN) {
2931 *from = saved_from;
2932 iov_iter_advance(from, offset - saved_offset);
2933 continue;
2934 }
2935 break;
2936 }
2937
2938 list_add_tail(&wdata->list, wdata_list);
2939 offset += cur_len;
2940 len -= cur_len;
2941 } while (len > 0);
2942
David Brazdil0f672f62019-12-10 10:32:29 +00002943 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002944 return rc;
2945}
2946
2947static void collect_uncached_write_data(struct cifs_aio_ctx *ctx)
2948{
2949 struct cifs_writedata *wdata, *tmp;
2950 struct cifs_tcon *tcon;
2951 struct cifs_sb_info *cifs_sb;
2952 struct dentry *dentry = ctx->cfile->dentry;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002953 int rc;
2954
2955 tcon = tlink_tcon(ctx->cfile->tlink);
2956 cifs_sb = CIFS_SB(dentry->d_sb);
2957
2958 mutex_lock(&ctx->aio_mutex);
2959
2960 if (list_empty(&ctx->list)) {
2961 mutex_unlock(&ctx->aio_mutex);
2962 return;
2963 }
2964
2965 rc = ctx->rc;
2966 /*
2967 * Wait for and collect replies for any successful sends in order of
2968 * increasing offset. Once an error is hit, then return without waiting
2969 * for any more replies.
2970 */
2971restart_loop:
2972 list_for_each_entry_safe(wdata, tmp, &ctx->list, list) {
2973 if (!rc) {
2974 if (!try_wait_for_completion(&wdata->done)) {
2975 mutex_unlock(&ctx->aio_mutex);
2976 return;
2977 }
2978
2979 if (wdata->result)
2980 rc = wdata->result;
2981 else
2982 ctx->total_len += wdata->bytes;
2983
2984 /* resend call if it's a retryable error */
2985 if (rc == -EAGAIN) {
2986 struct list_head tmp_list;
2987 struct iov_iter tmp_from = ctx->iter;
2988
2989 INIT_LIST_HEAD(&tmp_list);
2990 list_del_init(&wdata->list);
2991
David Brazdil0f672f62019-12-10 10:32:29 +00002992 if (ctx->direct_io)
2993 rc = cifs_resend_wdata(
2994 wdata, &tmp_list, ctx);
2995 else {
2996 iov_iter_advance(&tmp_from,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002997 wdata->offset - ctx->pos);
2998
David Brazdil0f672f62019-12-10 10:32:29 +00002999 rc = cifs_write_from_iter(wdata->offset,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003000 wdata->bytes, &tmp_from,
3001 ctx->cfile, cifs_sb, &tmp_list,
3002 ctx);
3003
David Brazdil0f672f62019-12-10 10:32:29 +00003004 kref_put(&wdata->refcount,
3005 cifs_uncached_writedata_release);
3006 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003007
David Brazdil0f672f62019-12-10 10:32:29 +00003008 list_splice(&tmp_list, &ctx->list);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003009 goto restart_loop;
3010 }
3011 }
3012 list_del_init(&wdata->list);
3013 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
3014 }
3015
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003016 cifs_stats_bytes_written(tcon, ctx->total_len);
3017 set_bit(CIFS_INO_INVALID_MAPPING, &CIFS_I(dentry->d_inode)->flags);
3018
3019 ctx->rc = (rc == 0) ? ctx->total_len : rc;
3020
3021 mutex_unlock(&ctx->aio_mutex);
3022
3023 if (ctx->iocb && ctx->iocb->ki_complete)
3024 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
3025 else
3026 complete(&ctx->done);
3027}
3028
David Brazdil0f672f62019-12-10 10:32:29 +00003029static ssize_t __cifs_writev(
3030 struct kiocb *iocb, struct iov_iter *from, bool direct)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003031{
3032 struct file *file = iocb->ki_filp;
3033 ssize_t total_written = 0;
3034 struct cifsFileInfo *cfile;
3035 struct cifs_tcon *tcon;
3036 struct cifs_sb_info *cifs_sb;
3037 struct cifs_aio_ctx *ctx;
3038 struct iov_iter saved_from = *from;
David Brazdil0f672f62019-12-10 10:32:29 +00003039 size_t len = iov_iter_count(from);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003040 int rc;
3041
3042 /*
David Brazdil0f672f62019-12-10 10:32:29 +00003043 * iov_iter_get_pages_alloc doesn't work with ITER_KVEC.
3044 * In this case, fall back to non-direct write function.
3045 * this could be improved by getting pages directly in ITER_KVEC
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003046 */
David Brazdil0f672f62019-12-10 10:32:29 +00003047 if (direct && from->type & ITER_KVEC) {
3048 cifs_dbg(FYI, "use non-direct cifs_writev for kvec I/O\n");
3049 direct = false;
3050 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003051
3052 rc = generic_write_checks(iocb, from);
3053 if (rc <= 0)
3054 return rc;
3055
3056 cifs_sb = CIFS_FILE_SB(file);
3057 cfile = file->private_data;
3058 tcon = tlink_tcon(cfile->tlink);
3059
3060 if (!tcon->ses->server->ops->async_writev)
3061 return -ENOSYS;
3062
3063 ctx = cifs_aio_ctx_alloc();
3064 if (!ctx)
3065 return -ENOMEM;
3066
3067 ctx->cfile = cifsFileInfo_get(cfile);
3068
3069 if (!is_sync_kiocb(iocb))
3070 ctx->iocb = iocb;
3071
3072 ctx->pos = iocb->ki_pos;
3073
David Brazdil0f672f62019-12-10 10:32:29 +00003074 if (direct) {
3075 ctx->direct_io = true;
3076 ctx->iter = *from;
3077 ctx->len = len;
3078 } else {
3079 rc = setup_aio_ctx_iter(ctx, from, WRITE);
3080 if (rc) {
3081 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3082 return rc;
3083 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003084 }
3085
3086 /* grab a lock here due to read response handlers can access ctx */
3087 mutex_lock(&ctx->aio_mutex);
3088
3089 rc = cifs_write_from_iter(iocb->ki_pos, ctx->len, &saved_from,
3090 cfile, cifs_sb, &ctx->list, ctx);
3091
3092 /*
3093 * If at least one write was successfully sent, then discard any rc
3094 * value from the later writes. If the other write succeeds, then
3095 * we'll end up returning whatever was written. If it fails, then
3096 * we'll get a new rc value from that.
3097 */
3098 if (!list_empty(&ctx->list))
3099 rc = 0;
3100
3101 mutex_unlock(&ctx->aio_mutex);
3102
3103 if (rc) {
3104 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3105 return rc;
3106 }
3107
3108 if (!is_sync_kiocb(iocb)) {
3109 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3110 return -EIOCBQUEUED;
3111 }
3112
3113 rc = wait_for_completion_killable(&ctx->done);
3114 if (rc) {
3115 mutex_lock(&ctx->aio_mutex);
3116 ctx->rc = rc = -EINTR;
3117 total_written = ctx->total_len;
3118 mutex_unlock(&ctx->aio_mutex);
3119 } else {
3120 rc = ctx->rc;
3121 total_written = ctx->total_len;
3122 }
3123
3124 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3125
3126 if (unlikely(!total_written))
3127 return rc;
3128
3129 iocb->ki_pos += total_written;
3130 return total_written;
3131}
3132
David Brazdil0f672f62019-12-10 10:32:29 +00003133ssize_t cifs_direct_writev(struct kiocb *iocb, struct iov_iter *from)
3134{
3135 return __cifs_writev(iocb, from, true);
3136}
3137
3138ssize_t cifs_user_writev(struct kiocb *iocb, struct iov_iter *from)
3139{
3140 return __cifs_writev(iocb, from, false);
3141}
3142
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003143static ssize_t
3144cifs_writev(struct kiocb *iocb, struct iov_iter *from)
3145{
3146 struct file *file = iocb->ki_filp;
3147 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
3148 struct inode *inode = file->f_mapping->host;
3149 struct cifsInodeInfo *cinode = CIFS_I(inode);
3150 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
3151 ssize_t rc;
3152
3153 inode_lock(inode);
3154 /*
3155 * We need to hold the sem to be sure nobody modifies lock list
3156 * with a brlock that prevents writing.
3157 */
3158 down_read(&cinode->lock_sem);
3159
3160 rc = generic_write_checks(iocb, from);
3161 if (rc <= 0)
3162 goto out;
3163
3164 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from),
David Brazdil0f672f62019-12-10 10:32:29 +00003165 server->vals->exclusive_lock_type, 0,
3166 NULL, CIFS_WRITE_OP))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003167 rc = __generic_file_write_iter(iocb, from);
3168 else
3169 rc = -EACCES;
3170out:
3171 up_read(&cinode->lock_sem);
3172 inode_unlock(inode);
3173
3174 if (rc > 0)
3175 rc = generic_write_sync(iocb, rc);
3176 return rc;
3177}
3178
3179ssize_t
3180cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from)
3181{
3182 struct inode *inode = file_inode(iocb->ki_filp);
3183 struct cifsInodeInfo *cinode = CIFS_I(inode);
3184 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3185 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3186 iocb->ki_filp->private_data;
3187 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
3188 ssize_t written;
3189
3190 written = cifs_get_writer(cinode);
3191 if (written)
3192 return written;
3193
3194 if (CIFS_CACHE_WRITE(cinode)) {
3195 if (cap_unix(tcon->ses) &&
3196 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability))
3197 && ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
3198 written = generic_file_write_iter(iocb, from);
3199 goto out;
3200 }
3201 written = cifs_writev(iocb, from);
3202 goto out;
3203 }
3204 /*
3205 * For non-oplocked files in strict cache mode we need to write the data
3206 * to the server exactly from the pos to pos+len-1 rather than flush all
3207 * affected pages because it may cause a error with mandatory locks on
3208 * these pages but not on the region from pos to ppos+len-1.
3209 */
3210 written = cifs_user_writev(iocb, from);
David Brazdil0f672f62019-12-10 10:32:29 +00003211 if (CIFS_CACHE_READ(cinode)) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003212 /*
David Brazdil0f672f62019-12-10 10:32:29 +00003213 * We have read level caching and we have just sent a write
3214 * request to the server thus making data in the cache stale.
3215 * Zap the cache and set oplock/lease level to NONE to avoid
3216 * reading stale data from the cache. All subsequent read
3217 * operations will read new data from the server.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003218 */
3219 cifs_zap_mapping(inode);
David Brazdil0f672f62019-12-10 10:32:29 +00003220 cifs_dbg(FYI, "Set Oplock/Lease to NONE for inode=%p after write\n",
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003221 inode);
3222 cinode->oplock = 0;
3223 }
3224out:
3225 cifs_put_writer(cinode);
3226 return written;
3227}
3228
3229static struct cifs_readdata *
3230cifs_readdata_direct_alloc(struct page **pages, work_func_t complete)
3231{
3232 struct cifs_readdata *rdata;
3233
3234 rdata = kzalloc(sizeof(*rdata), GFP_KERNEL);
3235 if (rdata != NULL) {
3236 rdata->pages = pages;
3237 kref_init(&rdata->refcount);
3238 INIT_LIST_HEAD(&rdata->list);
3239 init_completion(&rdata->done);
3240 INIT_WORK(&rdata->work, complete);
3241 }
3242
3243 return rdata;
3244}
3245
3246static struct cifs_readdata *
3247cifs_readdata_alloc(unsigned int nr_pages, work_func_t complete)
3248{
3249 struct page **pages =
3250 kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
3251 struct cifs_readdata *ret = NULL;
3252
3253 if (pages) {
3254 ret = cifs_readdata_direct_alloc(pages, complete);
3255 if (!ret)
3256 kfree(pages);
3257 }
3258
3259 return ret;
3260}
3261
3262void
3263cifs_readdata_release(struct kref *refcount)
3264{
3265 struct cifs_readdata *rdata = container_of(refcount,
3266 struct cifs_readdata, refcount);
3267#ifdef CONFIG_CIFS_SMB_DIRECT
3268 if (rdata->mr) {
3269 smbd_deregister_mr(rdata->mr);
3270 rdata->mr = NULL;
3271 }
3272#endif
3273 if (rdata->cfile)
3274 cifsFileInfo_put(rdata->cfile);
3275
3276 kvfree(rdata->pages);
3277 kfree(rdata);
3278}
3279
3280static int
3281cifs_read_allocate_pages(struct cifs_readdata *rdata, unsigned int nr_pages)
3282{
3283 int rc = 0;
3284 struct page *page;
3285 unsigned int i;
3286
3287 for (i = 0; i < nr_pages; i++) {
3288 page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
3289 if (!page) {
3290 rc = -ENOMEM;
3291 break;
3292 }
3293 rdata->pages[i] = page;
3294 }
3295
3296 if (rc) {
David Brazdil0f672f62019-12-10 10:32:29 +00003297 unsigned int nr_page_failed = i;
3298
3299 for (i = 0; i < nr_page_failed; i++) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003300 put_page(rdata->pages[i]);
3301 rdata->pages[i] = NULL;
3302 }
3303 }
3304 return rc;
3305}
3306
3307static void
3308cifs_uncached_readdata_release(struct kref *refcount)
3309{
3310 struct cifs_readdata *rdata = container_of(refcount,
3311 struct cifs_readdata, refcount);
3312 unsigned int i;
3313
3314 kref_put(&rdata->ctx->refcount, cifs_aio_ctx_release);
3315 for (i = 0; i < rdata->nr_pages; i++) {
3316 put_page(rdata->pages[i]);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003317 }
3318 cifs_readdata_release(refcount);
3319}
3320
3321/**
3322 * cifs_readdata_to_iov - copy data from pages in response to an iovec
3323 * @rdata: the readdata response with list of pages holding data
3324 * @iter: destination for our data
3325 *
3326 * This function copies data from a list of pages in a readdata response into
3327 * an array of iovecs. It will first calculate where the data should go
3328 * based on the info in the readdata and then copy the data into that spot.
3329 */
3330static int
3331cifs_readdata_to_iov(struct cifs_readdata *rdata, struct iov_iter *iter)
3332{
3333 size_t remaining = rdata->got_bytes;
3334 unsigned int i;
3335
3336 for (i = 0; i < rdata->nr_pages; i++) {
3337 struct page *page = rdata->pages[i];
3338 size_t copy = min_t(size_t, remaining, PAGE_SIZE);
3339 size_t written;
3340
David Brazdil0f672f62019-12-10 10:32:29 +00003341 if (unlikely(iov_iter_is_pipe(iter))) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003342 void *addr = kmap_atomic(page);
3343
3344 written = copy_to_iter(addr, copy, iter);
3345 kunmap_atomic(addr);
3346 } else
3347 written = copy_page_to_iter(page, 0, copy, iter);
3348 remaining -= written;
3349 if (written < copy && iov_iter_count(iter) > 0)
3350 break;
3351 }
3352 return remaining ? -EFAULT : 0;
3353}
3354
3355static void collect_uncached_read_data(struct cifs_aio_ctx *ctx);
3356
3357static void
3358cifs_uncached_readv_complete(struct work_struct *work)
3359{
3360 struct cifs_readdata *rdata = container_of(work,
3361 struct cifs_readdata, work);
3362
3363 complete(&rdata->done);
3364 collect_uncached_read_data(rdata->ctx);
3365 /* the below call can possibly free the last ref to aio ctx */
3366 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
3367}
3368
3369static int
3370uncached_fill_pages(struct TCP_Server_Info *server,
3371 struct cifs_readdata *rdata, struct iov_iter *iter,
3372 unsigned int len)
3373{
3374 int result = 0;
3375 unsigned int i;
3376 unsigned int nr_pages = rdata->nr_pages;
3377 unsigned int page_offset = rdata->page_offset;
3378
3379 rdata->got_bytes = 0;
3380 rdata->tailsz = PAGE_SIZE;
3381 for (i = 0; i < nr_pages; i++) {
3382 struct page *page = rdata->pages[i];
3383 size_t n;
3384 unsigned int segment_size = rdata->pagesz;
3385
3386 if (i == 0)
3387 segment_size -= page_offset;
3388 else
3389 page_offset = 0;
3390
3391
3392 if (len <= 0) {
3393 /* no need to hold page hostage */
3394 rdata->pages[i] = NULL;
3395 rdata->nr_pages--;
3396 put_page(page);
3397 continue;
3398 }
3399
3400 n = len;
3401 if (len >= segment_size)
3402 /* enough data to fill the page */
3403 n = segment_size;
3404 else
3405 rdata->tailsz = len;
3406 len -= n;
3407
3408 if (iter)
3409 result = copy_page_from_iter(
3410 page, page_offset, n, iter);
3411#ifdef CONFIG_CIFS_SMB_DIRECT
3412 else if (rdata->mr)
3413 result = n;
3414#endif
3415 else
3416 result = cifs_read_page_from_socket(
3417 server, page, page_offset, n);
3418 if (result < 0)
3419 break;
3420
3421 rdata->got_bytes += result;
3422 }
3423
3424 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
3425 rdata->got_bytes : result;
3426}
3427
3428static int
3429cifs_uncached_read_into_pages(struct TCP_Server_Info *server,
3430 struct cifs_readdata *rdata, unsigned int len)
3431{
3432 return uncached_fill_pages(server, rdata, NULL, len);
3433}
3434
3435static int
3436cifs_uncached_copy_into_pages(struct TCP_Server_Info *server,
3437 struct cifs_readdata *rdata,
3438 struct iov_iter *iter)
3439{
3440 return uncached_fill_pages(server, rdata, iter, iter->count);
3441}
3442
David Brazdil0f672f62019-12-10 10:32:29 +00003443static int cifs_resend_rdata(struct cifs_readdata *rdata,
3444 struct list_head *rdata_list,
3445 struct cifs_aio_ctx *ctx)
3446{
3447 unsigned int rsize;
3448 struct cifs_credits credits;
3449 int rc;
3450 struct TCP_Server_Info *server =
3451 tlink_tcon(rdata->cfile->tlink)->ses->server;
3452
3453 do {
3454 if (rdata->cfile->invalidHandle) {
3455 rc = cifs_reopen_file(rdata->cfile, true);
3456 if (rc == -EAGAIN)
3457 continue;
3458 else if (rc)
3459 break;
3460 }
3461
3462 /*
3463 * Wait for credits to resend this rdata.
3464 * Note: we are attempting to resend the whole rdata not in
3465 * segments
3466 */
3467 do {
3468 rc = server->ops->wait_mtu_credits(server, rdata->bytes,
3469 &rsize, &credits);
3470
3471 if (rc)
3472 goto fail;
3473
3474 if (rsize < rdata->bytes) {
3475 add_credits_and_wake_if(server, &credits, 0);
3476 msleep(1000);
3477 }
3478 } while (rsize < rdata->bytes);
3479 rdata->credits = credits;
3480
3481 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
3482 if (!rc) {
3483 if (rdata->cfile->invalidHandle)
3484 rc = -EAGAIN;
3485 else
3486 rc = server->ops->async_readv(rdata);
3487 }
3488
3489 /* If the read was successfully sent, we are done */
3490 if (!rc) {
3491 /* Add to aio pending list */
3492 list_add_tail(&rdata->list, rdata_list);
3493 return 0;
3494 }
3495
3496 /* Roll back credits and retry if needed */
3497 add_credits_and_wake_if(server, &rdata->credits, 0);
3498 } while (rc == -EAGAIN);
3499
3500fail:
3501 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
3502 return rc;
3503}
3504
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003505static int
3506cifs_send_async_read(loff_t offset, size_t len, struct cifsFileInfo *open_file,
3507 struct cifs_sb_info *cifs_sb, struct list_head *rdata_list,
3508 struct cifs_aio_ctx *ctx)
3509{
3510 struct cifs_readdata *rdata;
David Brazdil0f672f62019-12-10 10:32:29 +00003511 unsigned int npages, rsize;
3512 struct cifs_credits credits_on_stack;
3513 struct cifs_credits *credits = &credits_on_stack;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003514 size_t cur_len;
3515 int rc;
3516 pid_t pid;
3517 struct TCP_Server_Info *server;
David Brazdil0f672f62019-12-10 10:32:29 +00003518 struct page **pagevec;
3519 size_t start;
3520 struct iov_iter direct_iov = ctx->iter;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003521
3522 server = tlink_tcon(open_file->tlink)->ses->server;
3523
3524 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3525 pid = open_file->pid;
3526 else
3527 pid = current->tgid;
3528
David Brazdil0f672f62019-12-10 10:32:29 +00003529 if (ctx->direct_io)
3530 iov_iter_advance(&direct_iov, offset - ctx->pos);
3531
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003532 do {
David Brazdil0f672f62019-12-10 10:32:29 +00003533 if (open_file->invalidHandle) {
3534 rc = cifs_reopen_file(open_file, true);
3535 if (rc == -EAGAIN)
3536 continue;
3537 else if (rc)
3538 break;
3539 }
3540
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003541 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
David Brazdil0f672f62019-12-10 10:32:29 +00003542 &rsize, credits);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003543 if (rc)
3544 break;
3545
3546 cur_len = min_t(const size_t, len, rsize);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003547
David Brazdil0f672f62019-12-10 10:32:29 +00003548 if (ctx->direct_io) {
3549 ssize_t result;
3550
3551 result = iov_iter_get_pages_alloc(
3552 &direct_iov, &pagevec,
3553 cur_len, &start);
3554 if (result < 0) {
3555 cifs_dbg(VFS,
3556 "couldn't get user pages (rc=%zd)"
3557 " iter type %d"
3558 " iov_offset %zd count %zd\n",
3559 result, direct_iov.type,
3560 direct_iov.iov_offset,
3561 direct_iov.count);
3562 dump_stack();
3563
3564 rc = result;
3565 add_credits_and_wake_if(server, credits, 0);
3566 break;
3567 }
3568 cur_len = (size_t)result;
3569 iov_iter_advance(&direct_iov, cur_len);
3570
3571 rdata = cifs_readdata_direct_alloc(
3572 pagevec, cifs_uncached_readv_complete);
3573 if (!rdata) {
3574 add_credits_and_wake_if(server, credits, 0);
3575 rc = -ENOMEM;
3576 break;
3577 }
3578
3579 npages = (cur_len + start + PAGE_SIZE-1) / PAGE_SIZE;
3580 rdata->page_offset = start;
3581 rdata->tailsz = npages > 1 ?
3582 cur_len-(PAGE_SIZE-start)-(npages-2)*PAGE_SIZE :
3583 cur_len;
3584
3585 } else {
3586
3587 npages = DIV_ROUND_UP(cur_len, PAGE_SIZE);
3588 /* allocate a readdata struct */
3589 rdata = cifs_readdata_alloc(npages,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003590 cifs_uncached_readv_complete);
David Brazdil0f672f62019-12-10 10:32:29 +00003591 if (!rdata) {
3592 add_credits_and_wake_if(server, credits, 0);
3593 rc = -ENOMEM;
3594 break;
3595 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003596
David Brazdil0f672f62019-12-10 10:32:29 +00003597 rc = cifs_read_allocate_pages(rdata, npages);
3598 if (rc) {
3599 kvfree(rdata->pages);
3600 kfree(rdata);
3601 add_credits_and_wake_if(server, credits, 0);
3602 break;
3603 }
3604
3605 rdata->tailsz = PAGE_SIZE;
3606 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003607
3608 rdata->cfile = cifsFileInfo_get(open_file);
3609 rdata->nr_pages = npages;
3610 rdata->offset = offset;
3611 rdata->bytes = cur_len;
3612 rdata->pid = pid;
3613 rdata->pagesz = PAGE_SIZE;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003614 rdata->read_into_pages = cifs_uncached_read_into_pages;
3615 rdata->copy_into_pages = cifs_uncached_copy_into_pages;
David Brazdil0f672f62019-12-10 10:32:29 +00003616 rdata->credits = credits_on_stack;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003617 rdata->ctx = ctx;
3618 kref_get(&ctx->refcount);
3619
David Brazdil0f672f62019-12-10 10:32:29 +00003620 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
3621
3622 if (!rc) {
3623 if (rdata->cfile->invalidHandle)
3624 rc = -EAGAIN;
3625 else
3626 rc = server->ops->async_readv(rdata);
3627 }
3628
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003629 if (rc) {
David Brazdil0f672f62019-12-10 10:32:29 +00003630 add_credits_and_wake_if(server, &rdata->credits, 0);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003631 kref_put(&rdata->refcount,
David Brazdil0f672f62019-12-10 10:32:29 +00003632 cifs_uncached_readdata_release);
3633 if (rc == -EAGAIN) {
3634 iov_iter_revert(&direct_iov, cur_len);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003635 continue;
David Brazdil0f672f62019-12-10 10:32:29 +00003636 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003637 break;
3638 }
3639
3640 list_add_tail(&rdata->list, rdata_list);
3641 offset += cur_len;
3642 len -= cur_len;
3643 } while (len > 0);
3644
3645 return rc;
3646}
3647
3648static void
3649collect_uncached_read_data(struct cifs_aio_ctx *ctx)
3650{
3651 struct cifs_readdata *rdata, *tmp;
3652 struct iov_iter *to = &ctx->iter;
3653 struct cifs_sb_info *cifs_sb;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003654 int rc;
3655
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003656 cifs_sb = CIFS_SB(ctx->cfile->dentry->d_sb);
3657
3658 mutex_lock(&ctx->aio_mutex);
3659
3660 if (list_empty(&ctx->list)) {
3661 mutex_unlock(&ctx->aio_mutex);
3662 return;
3663 }
3664
3665 rc = ctx->rc;
3666 /* the loop below should proceed in the order of increasing offsets */
3667again:
3668 list_for_each_entry_safe(rdata, tmp, &ctx->list, list) {
3669 if (!rc) {
3670 if (!try_wait_for_completion(&rdata->done)) {
3671 mutex_unlock(&ctx->aio_mutex);
3672 return;
3673 }
3674
3675 if (rdata->result == -EAGAIN) {
3676 /* resend call if it's a retryable error */
3677 struct list_head tmp_list;
3678 unsigned int got_bytes = rdata->got_bytes;
3679
3680 list_del_init(&rdata->list);
3681 INIT_LIST_HEAD(&tmp_list);
3682
3683 /*
3684 * Got a part of data and then reconnect has
3685 * happened -- fill the buffer and continue
3686 * reading.
3687 */
3688 if (got_bytes && got_bytes < rdata->bytes) {
David Brazdil0f672f62019-12-10 10:32:29 +00003689 rc = 0;
3690 if (!ctx->direct_io)
3691 rc = cifs_readdata_to_iov(rdata, to);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003692 if (rc) {
3693 kref_put(&rdata->refcount,
David Brazdil0f672f62019-12-10 10:32:29 +00003694 cifs_uncached_readdata_release);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003695 continue;
3696 }
3697 }
3698
David Brazdil0f672f62019-12-10 10:32:29 +00003699 if (ctx->direct_io) {
3700 /*
3701 * Re-use rdata as this is a
3702 * direct I/O
3703 */
3704 rc = cifs_resend_rdata(
3705 rdata,
3706 &tmp_list, ctx);
3707 } else {
3708 rc = cifs_send_async_read(
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003709 rdata->offset + got_bytes,
3710 rdata->bytes - got_bytes,
3711 rdata->cfile, cifs_sb,
3712 &tmp_list, ctx);
3713
David Brazdil0f672f62019-12-10 10:32:29 +00003714 kref_put(&rdata->refcount,
3715 cifs_uncached_readdata_release);
3716 }
3717
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003718 list_splice(&tmp_list, &ctx->list);
3719
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003720 goto again;
3721 } else if (rdata->result)
3722 rc = rdata->result;
David Brazdil0f672f62019-12-10 10:32:29 +00003723 else if (!ctx->direct_io)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003724 rc = cifs_readdata_to_iov(rdata, to);
3725
3726 /* if there was a short read -- discard anything left */
3727 if (rdata->got_bytes && rdata->got_bytes < rdata->bytes)
3728 rc = -ENODATA;
David Brazdil0f672f62019-12-10 10:32:29 +00003729
3730 ctx->total_len += rdata->got_bytes;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003731 }
3732 list_del_init(&rdata->list);
3733 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
3734 }
3735
David Brazdil0f672f62019-12-10 10:32:29 +00003736 if (!ctx->direct_io)
3737 ctx->total_len = ctx->len - iov_iter_count(to);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003738
3739 /* mask nodata case */
3740 if (rc == -ENODATA)
3741 rc = 0;
3742
3743 ctx->rc = (rc == 0) ? ctx->total_len : rc;
3744
3745 mutex_unlock(&ctx->aio_mutex);
3746
3747 if (ctx->iocb && ctx->iocb->ki_complete)
3748 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
3749 else
3750 complete(&ctx->done);
3751}
3752
David Brazdil0f672f62019-12-10 10:32:29 +00003753static ssize_t __cifs_readv(
3754 struct kiocb *iocb, struct iov_iter *to, bool direct)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003755{
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003756 size_t len;
David Brazdil0f672f62019-12-10 10:32:29 +00003757 struct file *file = iocb->ki_filp;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003758 struct cifs_sb_info *cifs_sb;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003759 struct cifsFileInfo *cfile;
David Brazdil0f672f62019-12-10 10:32:29 +00003760 struct cifs_tcon *tcon;
3761 ssize_t rc, total_read = 0;
3762 loff_t offset = iocb->ki_pos;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003763 struct cifs_aio_ctx *ctx;
3764
David Brazdil0f672f62019-12-10 10:32:29 +00003765 /*
3766 * iov_iter_get_pages_alloc() doesn't work with ITER_KVEC,
3767 * fall back to data copy read path
3768 * this could be improved by getting pages directly in ITER_KVEC
3769 */
3770 if (direct && to->type & ITER_KVEC) {
3771 cifs_dbg(FYI, "use non-direct cifs_user_readv for kvec I/O\n");
3772 direct = false;
3773 }
3774
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003775 len = iov_iter_count(to);
3776 if (!len)
3777 return 0;
3778
3779 cifs_sb = CIFS_FILE_SB(file);
3780 cfile = file->private_data;
3781 tcon = tlink_tcon(cfile->tlink);
3782
3783 if (!tcon->ses->server->ops->async_readv)
3784 return -ENOSYS;
3785
3786 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
3787 cifs_dbg(FYI, "attempting read on write only file instance\n");
3788
3789 ctx = cifs_aio_ctx_alloc();
3790 if (!ctx)
3791 return -ENOMEM;
3792
3793 ctx->cfile = cifsFileInfo_get(cfile);
3794
3795 if (!is_sync_kiocb(iocb))
3796 ctx->iocb = iocb;
3797
David Brazdil0f672f62019-12-10 10:32:29 +00003798 if (iter_is_iovec(to))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003799 ctx->should_dirty = true;
3800
David Brazdil0f672f62019-12-10 10:32:29 +00003801 if (direct) {
3802 ctx->pos = offset;
3803 ctx->direct_io = true;
3804 ctx->iter = *to;
3805 ctx->len = len;
3806 } else {
3807 rc = setup_aio_ctx_iter(ctx, to, READ);
3808 if (rc) {
3809 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3810 return rc;
3811 }
3812 len = ctx->len;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003813 }
3814
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003815 /* grab a lock here due to read response handlers can access ctx */
3816 mutex_lock(&ctx->aio_mutex);
3817
3818 rc = cifs_send_async_read(offset, len, cfile, cifs_sb, &ctx->list, ctx);
3819
3820 /* if at least one read request send succeeded, then reset rc */
3821 if (!list_empty(&ctx->list))
3822 rc = 0;
3823
3824 mutex_unlock(&ctx->aio_mutex);
3825
3826 if (rc) {
3827 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3828 return rc;
3829 }
3830
3831 if (!is_sync_kiocb(iocb)) {
3832 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3833 return -EIOCBQUEUED;
3834 }
3835
3836 rc = wait_for_completion_killable(&ctx->done);
3837 if (rc) {
3838 mutex_lock(&ctx->aio_mutex);
3839 ctx->rc = rc = -EINTR;
3840 total_read = ctx->total_len;
3841 mutex_unlock(&ctx->aio_mutex);
3842 } else {
3843 rc = ctx->rc;
3844 total_read = ctx->total_len;
3845 }
3846
3847 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3848
3849 if (total_read) {
3850 iocb->ki_pos += total_read;
3851 return total_read;
3852 }
3853 return rc;
3854}
3855
David Brazdil0f672f62019-12-10 10:32:29 +00003856ssize_t cifs_direct_readv(struct kiocb *iocb, struct iov_iter *to)
3857{
3858 return __cifs_readv(iocb, to, true);
3859}
3860
3861ssize_t cifs_user_readv(struct kiocb *iocb, struct iov_iter *to)
3862{
3863 return __cifs_readv(iocb, to, false);
3864}
3865
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003866ssize_t
3867cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to)
3868{
3869 struct inode *inode = file_inode(iocb->ki_filp);
3870 struct cifsInodeInfo *cinode = CIFS_I(inode);
3871 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3872 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3873 iocb->ki_filp->private_data;
3874 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
3875 int rc = -EACCES;
3876
3877 /*
3878 * In strict cache mode we need to read from the server all the time
3879 * if we don't have level II oplock because the server can delay mtime
3880 * change - so we can't make a decision about inode invalidating.
3881 * And we can also fail with pagereading if there are mandatory locks
3882 * on pages affected by this read but not on the region from pos to
3883 * pos+len-1.
3884 */
3885 if (!CIFS_CACHE_READ(cinode))
3886 return cifs_user_readv(iocb, to);
3887
3888 if (cap_unix(tcon->ses) &&
3889 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
3890 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
3891 return generic_file_read_iter(iocb, to);
3892
3893 /*
3894 * We need to hold the sem to be sure nobody modifies lock list
3895 * with a brlock that prevents reading.
3896 */
3897 down_read(&cinode->lock_sem);
3898 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(to),
3899 tcon->ses->server->vals->shared_lock_type,
David Brazdil0f672f62019-12-10 10:32:29 +00003900 0, NULL, CIFS_READ_OP))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003901 rc = generic_file_read_iter(iocb, to);
3902 up_read(&cinode->lock_sem);
3903 return rc;
3904}
3905
3906static ssize_t
3907cifs_read(struct file *file, char *read_data, size_t read_size, loff_t *offset)
3908{
3909 int rc = -EACCES;
3910 unsigned int bytes_read = 0;
3911 unsigned int total_read;
3912 unsigned int current_read_size;
3913 unsigned int rsize;
3914 struct cifs_sb_info *cifs_sb;
3915 struct cifs_tcon *tcon;
3916 struct TCP_Server_Info *server;
3917 unsigned int xid;
3918 char *cur_offset;
3919 struct cifsFileInfo *open_file;
3920 struct cifs_io_parms io_parms;
3921 int buf_type = CIFS_NO_BUFFER;
3922 __u32 pid;
3923
3924 xid = get_xid();
3925 cifs_sb = CIFS_FILE_SB(file);
3926
3927 /* FIXME: set up handlers for larger reads and/or convert to async */
3928 rsize = min_t(unsigned int, cifs_sb->rsize, CIFSMaxBufSize);
3929
3930 if (file->private_data == NULL) {
3931 rc = -EBADF;
3932 free_xid(xid);
3933 return rc;
3934 }
3935 open_file = file->private_data;
3936 tcon = tlink_tcon(open_file->tlink);
3937 server = tcon->ses->server;
3938
3939 if (!server->ops->sync_read) {
3940 free_xid(xid);
3941 return -ENOSYS;
3942 }
3943
3944 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3945 pid = open_file->pid;
3946 else
3947 pid = current->tgid;
3948
3949 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
3950 cifs_dbg(FYI, "attempting read on write only file instance\n");
3951
3952 for (total_read = 0, cur_offset = read_data; read_size > total_read;
3953 total_read += bytes_read, cur_offset += bytes_read) {
3954 do {
3955 current_read_size = min_t(uint, read_size - total_read,
3956 rsize);
3957 /*
3958 * For windows me and 9x we do not want to request more
3959 * than it negotiated since it will refuse the read
3960 * then.
3961 */
3962 if ((tcon->ses) && !(tcon->ses->capabilities &
3963 tcon->ses->server->vals->cap_large_files)) {
3964 current_read_size = min_t(uint,
3965 current_read_size, CIFSMaxBufSize);
3966 }
3967 if (open_file->invalidHandle) {
3968 rc = cifs_reopen_file(open_file, true);
3969 if (rc != 0)
3970 break;
3971 }
3972 io_parms.pid = pid;
3973 io_parms.tcon = tcon;
3974 io_parms.offset = *offset;
3975 io_parms.length = current_read_size;
3976 rc = server->ops->sync_read(xid, &open_file->fid, &io_parms,
3977 &bytes_read, &cur_offset,
3978 &buf_type);
3979 } while (rc == -EAGAIN);
3980
3981 if (rc || (bytes_read == 0)) {
3982 if (total_read) {
3983 break;
3984 } else {
3985 free_xid(xid);
3986 return rc;
3987 }
3988 } else {
3989 cifs_stats_bytes_read(tcon, total_read);
3990 *offset += bytes_read;
3991 }
3992 }
3993 free_xid(xid);
3994 return total_read;
3995}
3996
3997/*
3998 * If the page is mmap'ed into a process' page tables, then we need to make
3999 * sure that it doesn't change while being written back.
4000 */
4001static vm_fault_t
4002cifs_page_mkwrite(struct vm_fault *vmf)
4003{
4004 struct page *page = vmf->page;
4005
4006 lock_page(page);
4007 return VM_FAULT_LOCKED;
4008}
4009
4010static const struct vm_operations_struct cifs_file_vm_ops = {
4011 .fault = filemap_fault,
4012 .map_pages = filemap_map_pages,
4013 .page_mkwrite = cifs_page_mkwrite,
4014};
4015
4016int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma)
4017{
4018 int xid, rc = 0;
4019 struct inode *inode = file_inode(file);
4020
4021 xid = get_xid();
4022
4023 if (!CIFS_CACHE_READ(CIFS_I(inode)))
4024 rc = cifs_zap_mapping(inode);
4025 if (!rc)
4026 rc = generic_file_mmap(file, vma);
4027 if (!rc)
4028 vma->vm_ops = &cifs_file_vm_ops;
4029
4030 free_xid(xid);
4031 return rc;
4032}
4033
4034int cifs_file_mmap(struct file *file, struct vm_area_struct *vma)
4035{
4036 int rc, xid;
4037
4038 xid = get_xid();
4039
4040 rc = cifs_revalidate_file(file);
4041 if (rc)
4042 cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n",
4043 rc);
4044 if (!rc)
4045 rc = generic_file_mmap(file, vma);
4046 if (!rc)
4047 vma->vm_ops = &cifs_file_vm_ops;
4048
4049 free_xid(xid);
4050 return rc;
4051}
4052
4053static void
4054cifs_readv_complete(struct work_struct *work)
4055{
4056 unsigned int i, got_bytes;
4057 struct cifs_readdata *rdata = container_of(work,
4058 struct cifs_readdata, work);
4059
4060 got_bytes = rdata->got_bytes;
4061 for (i = 0; i < rdata->nr_pages; i++) {
4062 struct page *page = rdata->pages[i];
4063
4064 lru_cache_add_file(page);
4065
4066 if (rdata->result == 0 ||
4067 (rdata->result == -EAGAIN && got_bytes)) {
4068 flush_dcache_page(page);
4069 SetPageUptodate(page);
4070 }
4071
4072 unlock_page(page);
4073
4074 if (rdata->result == 0 ||
4075 (rdata->result == -EAGAIN && got_bytes))
4076 cifs_readpage_to_fscache(rdata->mapping->host, page);
4077
4078 got_bytes -= min_t(unsigned int, PAGE_SIZE, got_bytes);
4079
4080 put_page(page);
4081 rdata->pages[i] = NULL;
4082 }
4083 kref_put(&rdata->refcount, cifs_readdata_release);
4084}
4085
4086static int
4087readpages_fill_pages(struct TCP_Server_Info *server,
4088 struct cifs_readdata *rdata, struct iov_iter *iter,
4089 unsigned int len)
4090{
4091 int result = 0;
4092 unsigned int i;
4093 u64 eof;
4094 pgoff_t eof_index;
4095 unsigned int nr_pages = rdata->nr_pages;
4096 unsigned int page_offset = rdata->page_offset;
4097
4098 /* determine the eof that the server (probably) has */
4099 eof = CIFS_I(rdata->mapping->host)->server_eof;
4100 eof_index = eof ? (eof - 1) >> PAGE_SHIFT : 0;
4101 cifs_dbg(FYI, "eof=%llu eof_index=%lu\n", eof, eof_index);
4102
4103 rdata->got_bytes = 0;
4104 rdata->tailsz = PAGE_SIZE;
4105 for (i = 0; i < nr_pages; i++) {
4106 struct page *page = rdata->pages[i];
4107 unsigned int to_read = rdata->pagesz;
4108 size_t n;
4109
4110 if (i == 0)
4111 to_read -= page_offset;
4112 else
4113 page_offset = 0;
4114
4115 n = to_read;
4116
4117 if (len >= to_read) {
4118 len -= to_read;
4119 } else if (len > 0) {
4120 /* enough for partial page, fill and zero the rest */
4121 zero_user(page, len + page_offset, to_read - len);
4122 n = rdata->tailsz = len;
4123 len = 0;
4124 } else if (page->index > eof_index) {
4125 /*
4126 * The VFS will not try to do readahead past the
4127 * i_size, but it's possible that we have outstanding
4128 * writes with gaps in the middle and the i_size hasn't
4129 * caught up yet. Populate those with zeroed out pages
4130 * to prevent the VFS from repeatedly attempting to
4131 * fill them until the writes are flushed.
4132 */
4133 zero_user(page, 0, PAGE_SIZE);
4134 lru_cache_add_file(page);
4135 flush_dcache_page(page);
4136 SetPageUptodate(page);
4137 unlock_page(page);
4138 put_page(page);
4139 rdata->pages[i] = NULL;
4140 rdata->nr_pages--;
4141 continue;
4142 } else {
4143 /* no need to hold page hostage */
4144 lru_cache_add_file(page);
4145 unlock_page(page);
4146 put_page(page);
4147 rdata->pages[i] = NULL;
4148 rdata->nr_pages--;
4149 continue;
4150 }
4151
4152 if (iter)
4153 result = copy_page_from_iter(
4154 page, page_offset, n, iter);
4155#ifdef CONFIG_CIFS_SMB_DIRECT
4156 else if (rdata->mr)
4157 result = n;
4158#endif
4159 else
4160 result = cifs_read_page_from_socket(
4161 server, page, page_offset, n);
4162 if (result < 0)
4163 break;
4164
4165 rdata->got_bytes += result;
4166 }
4167
4168 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
4169 rdata->got_bytes : result;
4170}
4171
4172static int
4173cifs_readpages_read_into_pages(struct TCP_Server_Info *server,
4174 struct cifs_readdata *rdata, unsigned int len)
4175{
4176 return readpages_fill_pages(server, rdata, NULL, len);
4177}
4178
4179static int
4180cifs_readpages_copy_into_pages(struct TCP_Server_Info *server,
4181 struct cifs_readdata *rdata,
4182 struct iov_iter *iter)
4183{
4184 return readpages_fill_pages(server, rdata, iter, iter->count);
4185}
4186
4187static int
4188readpages_get_pages(struct address_space *mapping, struct list_head *page_list,
4189 unsigned int rsize, struct list_head *tmplist,
4190 unsigned int *nr_pages, loff_t *offset, unsigned int *bytes)
4191{
4192 struct page *page, *tpage;
4193 unsigned int expected_index;
4194 int rc;
4195 gfp_t gfp = readahead_gfp_mask(mapping);
4196
4197 INIT_LIST_HEAD(tmplist);
4198
David Brazdil0f672f62019-12-10 10:32:29 +00004199 page = lru_to_page(page_list);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004200
4201 /*
4202 * Lock the page and put it in the cache. Since no one else
4203 * should have access to this page, we're safe to simply set
4204 * PG_locked without checking it first.
4205 */
4206 __SetPageLocked(page);
4207 rc = add_to_page_cache_locked(page, mapping,
4208 page->index, gfp);
4209
4210 /* give up if we can't stick it in the cache */
4211 if (rc) {
4212 __ClearPageLocked(page);
4213 return rc;
4214 }
4215
4216 /* move first page to the tmplist */
4217 *offset = (loff_t)page->index << PAGE_SHIFT;
4218 *bytes = PAGE_SIZE;
4219 *nr_pages = 1;
4220 list_move_tail(&page->lru, tmplist);
4221
4222 /* now try and add more pages onto the request */
4223 expected_index = page->index + 1;
4224 list_for_each_entry_safe_reverse(page, tpage, page_list, lru) {
4225 /* discontinuity ? */
4226 if (page->index != expected_index)
4227 break;
4228
4229 /* would this page push the read over the rsize? */
4230 if (*bytes + PAGE_SIZE > rsize)
4231 break;
4232
4233 __SetPageLocked(page);
4234 if (add_to_page_cache_locked(page, mapping, page->index, gfp)) {
4235 __ClearPageLocked(page);
4236 break;
4237 }
4238 list_move_tail(&page->lru, tmplist);
4239 (*bytes) += PAGE_SIZE;
4240 expected_index++;
4241 (*nr_pages)++;
4242 }
4243 return rc;
4244}
4245
4246static int cifs_readpages(struct file *file, struct address_space *mapping,
4247 struct list_head *page_list, unsigned num_pages)
4248{
4249 int rc;
4250 struct list_head tmplist;
4251 struct cifsFileInfo *open_file = file->private_data;
4252 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
4253 struct TCP_Server_Info *server;
4254 pid_t pid;
David Brazdil0f672f62019-12-10 10:32:29 +00004255 unsigned int xid;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004256
David Brazdil0f672f62019-12-10 10:32:29 +00004257 xid = get_xid();
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004258 /*
4259 * Reads as many pages as possible from fscache. Returns -ENOBUFS
4260 * immediately if the cookie is negative
4261 *
4262 * After this point, every page in the list might have PG_fscache set,
4263 * so we will need to clean that up off of every page we don't use.
4264 */
4265 rc = cifs_readpages_from_fscache(mapping->host, mapping, page_list,
4266 &num_pages);
David Brazdil0f672f62019-12-10 10:32:29 +00004267 if (rc == 0) {
4268 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004269 return rc;
David Brazdil0f672f62019-12-10 10:32:29 +00004270 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004271
4272 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
4273 pid = open_file->pid;
4274 else
4275 pid = current->tgid;
4276
4277 rc = 0;
4278 server = tlink_tcon(open_file->tlink)->ses->server;
4279
4280 cifs_dbg(FYI, "%s: file=%p mapping=%p num_pages=%u\n",
4281 __func__, file, mapping, num_pages);
4282
4283 /*
4284 * Start with the page at end of list and move it to private
4285 * list. Do the same with any following pages until we hit
4286 * the rsize limit, hit an index discontinuity, or run out of
4287 * pages. Issue the async read and then start the loop again
4288 * until the list is empty.
4289 *
4290 * Note that list order is important. The page_list is in
4291 * the order of declining indexes. When we put the pages in
4292 * the rdata->pages, then we want them in increasing order.
4293 */
4294 while (!list_empty(page_list)) {
4295 unsigned int i, nr_pages, bytes, rsize;
4296 loff_t offset;
4297 struct page *page, *tpage;
4298 struct cifs_readdata *rdata;
David Brazdil0f672f62019-12-10 10:32:29 +00004299 struct cifs_credits credits_on_stack;
4300 struct cifs_credits *credits = &credits_on_stack;
4301
4302 if (open_file->invalidHandle) {
4303 rc = cifs_reopen_file(open_file, true);
4304 if (rc == -EAGAIN)
4305 continue;
4306 else if (rc)
4307 break;
4308 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004309
4310 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
David Brazdil0f672f62019-12-10 10:32:29 +00004311 &rsize, credits);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004312 if (rc)
4313 break;
4314
4315 /*
4316 * Give up immediately if rsize is too small to read an entire
4317 * page. The VFS will fall back to readpage. We should never
4318 * reach this point however since we set ra_pages to 0 when the
4319 * rsize is smaller than a cache page.
4320 */
4321 if (unlikely(rsize < PAGE_SIZE)) {
4322 add_credits_and_wake_if(server, credits, 0);
David Brazdil0f672f62019-12-10 10:32:29 +00004323 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004324 return 0;
4325 }
4326
4327 rc = readpages_get_pages(mapping, page_list, rsize, &tmplist,
4328 &nr_pages, &offset, &bytes);
4329 if (rc) {
4330 add_credits_and_wake_if(server, credits, 0);
4331 break;
4332 }
4333
4334 rdata = cifs_readdata_alloc(nr_pages, cifs_readv_complete);
4335 if (!rdata) {
4336 /* best to give up if we're out of mem */
4337 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
4338 list_del(&page->lru);
4339 lru_cache_add_file(page);
4340 unlock_page(page);
4341 put_page(page);
4342 }
4343 rc = -ENOMEM;
4344 add_credits_and_wake_if(server, credits, 0);
4345 break;
4346 }
4347
4348 rdata->cfile = cifsFileInfo_get(open_file);
4349 rdata->mapping = mapping;
4350 rdata->offset = offset;
4351 rdata->bytes = bytes;
4352 rdata->pid = pid;
4353 rdata->pagesz = PAGE_SIZE;
4354 rdata->tailsz = PAGE_SIZE;
4355 rdata->read_into_pages = cifs_readpages_read_into_pages;
4356 rdata->copy_into_pages = cifs_readpages_copy_into_pages;
David Brazdil0f672f62019-12-10 10:32:29 +00004357 rdata->credits = credits_on_stack;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004358
4359 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
4360 list_del(&page->lru);
4361 rdata->pages[rdata->nr_pages++] = page;
4362 }
4363
David Brazdil0f672f62019-12-10 10:32:29 +00004364 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
4365
4366 if (!rc) {
4367 if (rdata->cfile->invalidHandle)
4368 rc = -EAGAIN;
4369 else
4370 rc = server->ops->async_readv(rdata);
4371 }
4372
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004373 if (rc) {
David Brazdil0f672f62019-12-10 10:32:29 +00004374 add_credits_and_wake_if(server, &rdata->credits, 0);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004375 for (i = 0; i < rdata->nr_pages; i++) {
4376 page = rdata->pages[i];
4377 lru_cache_add_file(page);
4378 unlock_page(page);
4379 put_page(page);
4380 }
4381 /* Fallback to the readpage in error/reconnect cases */
4382 kref_put(&rdata->refcount, cifs_readdata_release);
4383 break;
4384 }
4385
4386 kref_put(&rdata->refcount, cifs_readdata_release);
4387 }
4388
4389 /* Any pages that have been shown to fscache but didn't get added to
4390 * the pagecache must be uncached before they get returned to the
4391 * allocator.
4392 */
4393 cifs_fscache_readpages_cancel(mapping->host, page_list);
David Brazdil0f672f62019-12-10 10:32:29 +00004394 free_xid(xid);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004395 return rc;
4396}
4397
4398/*
4399 * cifs_readpage_worker must be called with the page pinned
4400 */
4401static int cifs_readpage_worker(struct file *file, struct page *page,
4402 loff_t *poffset)
4403{
4404 char *read_data;
4405 int rc;
4406
4407 /* Is the page cached? */
4408 rc = cifs_readpage_from_fscache(file_inode(file), page);
4409 if (rc == 0)
4410 goto read_complete;
4411
4412 read_data = kmap(page);
4413 /* for reads over a certain size could initiate async read ahead */
4414
4415 rc = cifs_read(file, read_data, PAGE_SIZE, poffset);
4416
4417 if (rc < 0)
4418 goto io_error;
4419 else
4420 cifs_dbg(FYI, "Bytes read %d\n", rc);
4421
David Brazdil0f672f62019-12-10 10:32:29 +00004422 /* we do not want atime to be less than mtime, it broke some apps */
4423 file_inode(file)->i_atime = current_time(file_inode(file));
4424 if (timespec64_compare(&(file_inode(file)->i_atime), &(file_inode(file)->i_mtime)))
4425 file_inode(file)->i_atime = file_inode(file)->i_mtime;
4426 else
4427 file_inode(file)->i_atime = current_time(file_inode(file));
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004428
4429 if (PAGE_SIZE > rc)
4430 memset(read_data + rc, 0, PAGE_SIZE - rc);
4431
4432 flush_dcache_page(page);
4433 SetPageUptodate(page);
4434
4435 /* send this page to the cache */
4436 cifs_readpage_to_fscache(file_inode(file), page);
4437
4438 rc = 0;
4439
4440io_error:
4441 kunmap(page);
4442 unlock_page(page);
4443
4444read_complete:
4445 return rc;
4446}
4447
4448static int cifs_readpage(struct file *file, struct page *page)
4449{
4450 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
4451 int rc = -EACCES;
4452 unsigned int xid;
4453
4454 xid = get_xid();
4455
4456 if (file->private_data == NULL) {
4457 rc = -EBADF;
4458 free_xid(xid);
4459 return rc;
4460 }
4461
4462 cifs_dbg(FYI, "readpage %p at offset %d 0x%x\n",
4463 page, (int)offset, (int)offset);
4464
4465 rc = cifs_readpage_worker(file, page, &offset);
4466
4467 free_xid(xid);
4468 return rc;
4469}
4470
4471static int is_inode_writable(struct cifsInodeInfo *cifs_inode)
4472{
4473 struct cifsFileInfo *open_file;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004474
David Brazdil0f672f62019-12-10 10:32:29 +00004475 spin_lock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004476 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
4477 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
David Brazdil0f672f62019-12-10 10:32:29 +00004478 spin_unlock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004479 return 1;
4480 }
4481 }
David Brazdil0f672f62019-12-10 10:32:29 +00004482 spin_unlock(&cifs_inode->open_file_lock);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004483 return 0;
4484}
4485
4486/* We do not want to update the file size from server for inodes
4487 open for write - to avoid races with writepage extending
4488 the file - in the future we could consider allowing
4489 refreshing the inode only on increases in the file size
4490 but this is tricky to do without racing with writebehind
4491 page caching in the current Linux kernel design */
4492bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file)
4493{
4494 if (!cifsInode)
4495 return true;
4496
4497 if (is_inode_writable(cifsInode)) {
4498 /* This inode is open for write at least once */
4499 struct cifs_sb_info *cifs_sb;
4500
4501 cifs_sb = CIFS_SB(cifsInode->vfs_inode.i_sb);
4502 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) {
4503 /* since no page cache to corrupt on directio
4504 we can change size safely */
4505 return true;
4506 }
4507
4508 if (i_size_read(&cifsInode->vfs_inode) < end_of_file)
4509 return true;
4510
4511 return false;
4512 } else
4513 return true;
4514}
4515
4516static int cifs_write_begin(struct file *file, struct address_space *mapping,
4517 loff_t pos, unsigned len, unsigned flags,
4518 struct page **pagep, void **fsdata)
4519{
4520 int oncethru = 0;
4521 pgoff_t index = pos >> PAGE_SHIFT;
4522 loff_t offset = pos & (PAGE_SIZE - 1);
4523 loff_t page_start = pos & PAGE_MASK;
4524 loff_t i_size;
4525 struct page *page;
4526 int rc = 0;
4527
4528 cifs_dbg(FYI, "write_begin from %lld len %d\n", (long long)pos, len);
4529
4530start:
4531 page = grab_cache_page_write_begin(mapping, index, flags);
4532 if (!page) {
4533 rc = -ENOMEM;
4534 goto out;
4535 }
4536
4537 if (PageUptodate(page))
4538 goto out;
4539
4540 /*
4541 * If we write a full page it will be up to date, no need to read from
4542 * the server. If the write is short, we'll end up doing a sync write
4543 * instead.
4544 */
4545 if (len == PAGE_SIZE)
4546 goto out;
4547
4548 /*
4549 * optimize away the read when we have an oplock, and we're not
4550 * expecting to use any of the data we'd be reading in. That
4551 * is, when the page lies beyond the EOF, or straddles the EOF
4552 * and the write will cover all of the existing data.
4553 */
4554 if (CIFS_CACHE_READ(CIFS_I(mapping->host))) {
4555 i_size = i_size_read(mapping->host);
4556 if (page_start >= i_size ||
4557 (offset == 0 && (pos + len) >= i_size)) {
4558 zero_user_segments(page, 0, offset,
4559 offset + len,
4560 PAGE_SIZE);
4561 /*
4562 * PageChecked means that the parts of the page
4563 * to which we're not writing are considered up
4564 * to date. Once the data is copied to the
4565 * page, it can be set uptodate.
4566 */
4567 SetPageChecked(page);
4568 goto out;
4569 }
4570 }
4571
4572 if ((file->f_flags & O_ACCMODE) != O_WRONLY && !oncethru) {
4573 /*
4574 * might as well read a page, it is fast enough. If we get
4575 * an error, we don't need to return it. cifs_write_end will
4576 * do a sync write instead since PG_uptodate isn't set.
4577 */
4578 cifs_readpage_worker(file, page, &page_start);
4579 put_page(page);
4580 oncethru = 1;
4581 goto start;
4582 } else {
4583 /* we could try using another file handle if there is one -
4584 but how would we lock it to prevent close of that handle
4585 racing with this read? In any case
4586 this will be written out by write_end so is fine */
4587 }
4588out:
4589 *pagep = page;
4590 return rc;
4591}
4592
4593static int cifs_release_page(struct page *page, gfp_t gfp)
4594{
4595 if (PagePrivate(page))
4596 return 0;
4597
4598 return cifs_fscache_release_page(page, gfp);
4599}
4600
4601static void cifs_invalidate_page(struct page *page, unsigned int offset,
4602 unsigned int length)
4603{
4604 struct cifsInodeInfo *cifsi = CIFS_I(page->mapping->host);
4605
4606 if (offset == 0 && length == PAGE_SIZE)
4607 cifs_fscache_invalidate_page(page, &cifsi->vfs_inode);
4608}
4609
4610static int cifs_launder_page(struct page *page)
4611{
4612 int rc = 0;
4613 loff_t range_start = page_offset(page);
4614 loff_t range_end = range_start + (loff_t)(PAGE_SIZE - 1);
4615 struct writeback_control wbc = {
4616 .sync_mode = WB_SYNC_ALL,
4617 .nr_to_write = 0,
4618 .range_start = range_start,
4619 .range_end = range_end,
4620 };
4621
4622 cifs_dbg(FYI, "Launder page: %p\n", page);
4623
4624 if (clear_page_dirty_for_io(page))
4625 rc = cifs_writepage_locked(page, &wbc);
4626
4627 cifs_fscache_invalidate_page(page, page->mapping->host);
4628 return rc;
4629}
4630
4631void cifs_oplock_break(struct work_struct *work)
4632{
4633 struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo,
4634 oplock_break);
4635 struct inode *inode = d_inode(cfile->dentry);
4636 struct cifsInodeInfo *cinode = CIFS_I(inode);
4637 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
4638 struct TCP_Server_Info *server = tcon->ses->server;
4639 int rc = 0;
4640
4641 wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS,
4642 TASK_UNINTERRUPTIBLE);
4643
4644 server->ops->downgrade_oplock(server, cinode,
4645 test_bit(CIFS_INODE_DOWNGRADE_OPLOCK_TO_L2, &cinode->flags));
4646
4647 if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) &&
4648 cifs_has_mand_locks(cinode)) {
4649 cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n",
4650 inode);
4651 cinode->oplock = 0;
4652 }
4653
4654 if (inode && S_ISREG(inode->i_mode)) {
4655 if (CIFS_CACHE_READ(cinode))
4656 break_lease(inode, O_RDONLY);
4657 else
4658 break_lease(inode, O_WRONLY);
4659 rc = filemap_fdatawrite(inode->i_mapping);
4660 if (!CIFS_CACHE_READ(cinode)) {
4661 rc = filemap_fdatawait(inode->i_mapping);
4662 mapping_set_error(inode->i_mapping, rc);
4663 cifs_zap_mapping(inode);
4664 }
4665 cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc);
4666 }
4667
4668 rc = cifs_push_locks(cfile);
4669 if (rc)
4670 cifs_dbg(VFS, "Push locks rc = %d\n", rc);
4671
4672 /*
4673 * releasing stale oplock after recent reconnect of smb session using
4674 * a now incorrect file handle is not a data integrity issue but do
4675 * not bother sending an oplock release if session to server still is
4676 * disconnected since oplock already released by the server
4677 */
4678 if (!cfile->oplock_break_cancelled) {
4679 rc = tcon->ses->server->ops->oplock_response(tcon, &cfile->fid,
4680 cinode);
4681 cifs_dbg(FYI, "Oplock release rc = %d\n", rc);
4682 }
David Brazdil0f672f62019-12-10 10:32:29 +00004683 _cifsFileInfo_put(cfile, false /* do not wait for ourself */);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00004684 cifs_done_oplock_break(cinode);
4685}
4686
4687/*
4688 * The presence of cifs_direct_io() in the address space ops vector
4689 * allowes open() O_DIRECT flags which would have failed otherwise.
4690 *
4691 * In the non-cached mode (mount with cache=none), we shunt off direct read and write requests
4692 * so this method should never be called.
4693 *
4694 * Direct IO is not yet supported in the cached mode.
4695 */
4696static ssize_t
4697cifs_direct_io(struct kiocb *iocb, struct iov_iter *iter)
4698{
4699 /*
4700 * FIXME
4701 * Eventually need to support direct IO for non forcedirectio mounts
4702 */
4703 return -EINVAL;
4704}
4705
4706
4707const struct address_space_operations cifs_addr_ops = {
4708 .readpage = cifs_readpage,
4709 .readpages = cifs_readpages,
4710 .writepage = cifs_writepage,
4711 .writepages = cifs_writepages,
4712 .write_begin = cifs_write_begin,
4713 .write_end = cifs_write_end,
4714 .set_page_dirty = __set_page_dirty_nobuffers,
4715 .releasepage = cifs_release_page,
4716 .direct_IO = cifs_direct_io,
4717 .invalidatepage = cifs_invalidate_page,
4718 .launder_page = cifs_launder_page,
4719};
4720
4721/*
4722 * cifs_readpages requires the server to support a buffer large enough to
4723 * contain the header plus one complete page of data. Otherwise, we need
4724 * to leave cifs_readpages out of the address space operations.
4725 */
4726const struct address_space_operations cifs_addr_ops_smallbuf = {
4727 .readpage = cifs_readpage,
4728 .writepage = cifs_writepage,
4729 .writepages = cifs_writepages,
4730 .write_begin = cifs_write_begin,
4731 .write_end = cifs_write_end,
4732 .set_page_dirty = __set_page_dirty_nobuffers,
4733 .releasepage = cifs_release_page,
4734 .invalidatepage = cifs_invalidate_page,
4735 .launder_page = cifs_launder_page,
4736};