]> git.ipfire.org Git - people/ms/linux.git/blame - fs/cifs/file.c
cifs: remove unused variable 'sid_user'
[people/ms/linux.git] / fs / cifs / file.c
CommitLineData
1da177e4
LT
1/*
2 * fs/cifs/file.c
3 *
4 * vfs operations that deal with files
fb8c4b14 5 *
f19159dc 6 * Copyright (C) International Business Machines Corp., 2002,2010
1da177e4 7 * Author(s): Steve French (sfrench@us.ibm.com)
7ee1af76 8 * Jeremy Allison (jra@samba.org)
1da177e4
LT
9 *
10 * This library is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU Lesser General Public License as published
12 * by the Free Software Foundation; either version 2.1 of the License, or
13 * (at your option) any later version.
14 *
15 * This library is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
18 * the GNU Lesser General Public License for more details.
19 *
20 * You should have received a copy of the GNU Lesser General Public License
21 * along with this library; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 */
24#include <linux/fs.h>
37c0eb46 25#include <linux/backing-dev.h>
1da177e4
LT
26#include <linux/stat.h>
27#include <linux/fcntl.h>
28#include <linux/pagemap.h>
29#include <linux/pagevec.h>
37c0eb46 30#include <linux/writeback.h>
6f88cc2e 31#include <linux/task_io_accounting_ops.h>
23e7dd7d 32#include <linux/delay.h>
3bc303c2 33#include <linux/mount.h>
5a0e3ad6 34#include <linux/slab.h>
690c5e31 35#include <linux/swap.h>
f86196ea 36#include <linux/mm.h>
1da177e4
LT
37#include <asm/div64.h>
38#include "cifsfs.h"
39#include "cifspdu.h"
40#include "cifsglob.h"
41#include "cifsproto.h"
42#include "cifs_unicode.h"
43#include "cifs_debug.h"
44#include "cifs_fs_sb.h"
9451a9a5 45#include "fscache.h"
bd3dcc6a 46#include "smbdirect.h"
07b92d0d 47
1da177e4
LT
48static inline int cifs_convert_flags(unsigned int flags)
49{
50 if ((flags & O_ACCMODE) == O_RDONLY)
51 return GENERIC_READ;
52 else if ((flags & O_ACCMODE) == O_WRONLY)
53 return GENERIC_WRITE;
54 else if ((flags & O_ACCMODE) == O_RDWR) {
55 /* GENERIC_ALL is too much permission to request
56 can cause unnecessary access denied on create */
57 /* return GENERIC_ALL; */
58 return (GENERIC_READ | GENERIC_WRITE);
59 }
60
e10f7b55
JL
61 return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES |
62 FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA |
63 FILE_READ_DATA);
7fc8f4e9 64}
e10f7b55 65
608712fe 66static u32 cifs_posix_convert_flags(unsigned int flags)
7fc8f4e9 67{
608712fe 68 u32 posix_flags = 0;
e10f7b55 69
7fc8f4e9 70 if ((flags & O_ACCMODE) == O_RDONLY)
608712fe 71 posix_flags = SMB_O_RDONLY;
7fc8f4e9 72 else if ((flags & O_ACCMODE) == O_WRONLY)
608712fe
JL
73 posix_flags = SMB_O_WRONLY;
74 else if ((flags & O_ACCMODE) == O_RDWR)
75 posix_flags = SMB_O_RDWR;
76
07b92d0d 77 if (flags & O_CREAT) {
608712fe 78 posix_flags |= SMB_O_CREAT;
07b92d0d
SF
79 if (flags & O_EXCL)
80 posix_flags |= SMB_O_EXCL;
81 } else if (flags & O_EXCL)
f96637be
JP
82 cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n",
83 current->comm, current->tgid);
07b92d0d 84
608712fe
JL
85 if (flags & O_TRUNC)
86 posix_flags |= SMB_O_TRUNC;
87 /* be safe and imply O_SYNC for O_DSYNC */
6b2f3d1f 88 if (flags & O_DSYNC)
608712fe 89 posix_flags |= SMB_O_SYNC;
7fc8f4e9 90 if (flags & O_DIRECTORY)
608712fe 91 posix_flags |= SMB_O_DIRECTORY;
7fc8f4e9 92 if (flags & O_NOFOLLOW)
608712fe 93 posix_flags |= SMB_O_NOFOLLOW;
7fc8f4e9 94 if (flags & O_DIRECT)
608712fe 95 posix_flags |= SMB_O_DIRECT;
7fc8f4e9
SF
96
97 return posix_flags;
1da177e4
LT
98}
99
100static inline int cifs_get_disposition(unsigned int flags)
101{
102 if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL))
103 return FILE_CREATE;
104 else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC))
105 return FILE_OVERWRITE_IF;
106 else if ((flags & O_CREAT) == O_CREAT)
107 return FILE_OPEN_IF;
55aa2e09
SF
108 else if ((flags & O_TRUNC) == O_TRUNC)
109 return FILE_OVERWRITE;
1da177e4
LT
110 else
111 return FILE_OPEN;
112}
113
608712fe
JL
114int cifs_posix_open(char *full_path, struct inode **pinode,
115 struct super_block *sb, int mode, unsigned int f_flags,
6d5786a3 116 __u32 *poplock, __u16 *pnetfid, unsigned int xid)
608712fe
JL
117{
118 int rc;
119 FILE_UNIX_BASIC_INFO *presp_data;
120 __u32 posix_flags = 0;
121 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
122 struct cifs_fattr fattr;
123 struct tcon_link *tlink;
96daf2b0 124 struct cifs_tcon *tcon;
608712fe 125
f96637be 126 cifs_dbg(FYI, "posix open %s\n", full_path);
608712fe
JL
127
128 presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL);
129 if (presp_data == NULL)
130 return -ENOMEM;
131
132 tlink = cifs_sb_tlink(cifs_sb);
133 if (IS_ERR(tlink)) {
134 rc = PTR_ERR(tlink);
135 goto posix_open_ret;
136 }
137
138 tcon = tlink_tcon(tlink);
139 mode &= ~current_umask();
140
141 posix_flags = cifs_posix_convert_flags(f_flags);
142 rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data,
143 poplock, full_path, cifs_sb->local_nls,
bc8ebdc4 144 cifs_remap(cifs_sb));
608712fe
JL
145 cifs_put_tlink(tlink);
146
147 if (rc)
148 goto posix_open_ret;
149
150 if (presp_data->Type == cpu_to_le32(-1))
151 goto posix_open_ret; /* open ok, caller does qpathinfo */
152
153 if (!pinode)
154 goto posix_open_ret; /* caller does not need info */
155
156 cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb);
157
158 /* get new inode and set it up */
159 if (*pinode == NULL) {
160 cifs_fill_uniqueid(sb, &fattr);
161 *pinode = cifs_iget(sb, &fattr);
162 if (!*pinode) {
163 rc = -ENOMEM;
164 goto posix_open_ret;
165 }
166 } else {
167 cifs_fattr_to_inode(*pinode, &fattr);
168 }
169
170posix_open_ret:
171 kfree(presp_data);
172 return rc;
173}
174
eeb910a6
PS
175static int
176cifs_nt_open(char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb,
fb1214e4
PS
177 struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock,
178 struct cifs_fid *fid, unsigned int xid)
eeb910a6
PS
179{
180 int rc;
fb1214e4 181 int desired_access;
eeb910a6 182 int disposition;
3d3ea8e6 183 int create_options = CREATE_NOT_DIR;
eeb910a6 184 FILE_ALL_INFO *buf;
b8c32dbb 185 struct TCP_Server_Info *server = tcon->ses->server;
226730b4 186 struct cifs_open_parms oparms;
eeb910a6 187
b8c32dbb 188 if (!server->ops->open)
fb1214e4
PS
189 return -ENOSYS;
190
191 desired_access = cifs_convert_flags(f_flags);
eeb910a6
PS
192
193/*********************************************************************
194 * open flag mapping table:
195 *
196 * POSIX Flag CIFS Disposition
197 * ---------- ----------------
198 * O_CREAT FILE_OPEN_IF
199 * O_CREAT | O_EXCL FILE_CREATE
200 * O_CREAT | O_TRUNC FILE_OVERWRITE_IF
201 * O_TRUNC FILE_OVERWRITE
202 * none of the above FILE_OPEN
203 *
204 * Note that there is not a direct match between disposition
205 * FILE_SUPERSEDE (ie create whether or not file exists although
206 * O_CREAT | O_TRUNC is similar but truncates the existing
207 * file rather than creating a new file as FILE_SUPERSEDE does
208 * (which uses the attributes / metadata passed in on open call)
209 *?
210 *? O_SYNC is a reasonable match to CIFS writethrough flag
211 *? and the read write flags match reasonably. O_LARGEFILE
212 *? is irrelevant because largefile support is always used
213 *? by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY,
214 * O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation
215 *********************************************************************/
216
217 disposition = cifs_get_disposition(f_flags);
218
219 /* BB pass O_SYNC flag through on file attributes .. BB */
220
221 buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL);
222 if (!buf)
223 return -ENOMEM;
224
3d3ea8e6
SP
225 if (backup_cred(cifs_sb))
226 create_options |= CREATE_OPEN_BACKUP_INTENT;
227
1013e760
SF
228 /* O_SYNC also has bit for O_DSYNC so following check picks up either */
229 if (f_flags & O_SYNC)
230 create_options |= CREATE_WRITE_THROUGH;
231
232 if (f_flags & O_DIRECT)
233 create_options |= CREATE_NO_BUFFER;
234
226730b4
PS
235 oparms.tcon = tcon;
236 oparms.cifs_sb = cifs_sb;
237 oparms.desired_access = desired_access;
238 oparms.create_options = create_options;
239 oparms.disposition = disposition;
240 oparms.path = full_path;
241 oparms.fid = fid;
9cbc0b73 242 oparms.reconnect = false;
226730b4
PS
243
244 rc = server->ops->open(xid, &oparms, oplock, buf);
eeb910a6
PS
245
246 if (rc)
247 goto out;
248
249 if (tcon->unix_ext)
250 rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb,
251 xid);
252 else
253 rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb,
42eacf9e 254 xid, fid);
eeb910a6 255
30573a82
PS
256 if (rc) {
257 server->ops->close(xid, tcon, fid);
258 if (rc == -ESTALE)
259 rc = -EOPENSTALE;
260 }
261
eeb910a6
PS
262out:
263 kfree(buf);
264 return rc;
265}
266
63b7d3a4
PS
267static bool
268cifs_has_mand_locks(struct cifsInodeInfo *cinode)
269{
270 struct cifs_fid_locks *cur;
271 bool has_locks = false;
272
273 down_read(&cinode->lock_sem);
274 list_for_each_entry(cur, &cinode->llist, llist) {
275 if (!list_empty(&cur->locks)) {
276 has_locks = true;
277 break;
278 }
279 }
280 up_read(&cinode->lock_sem);
281 return has_locks;
282}
283
d46b0da7
DW
284void
285cifs_down_write(struct rw_semaphore *sem)
286{
287 while (!down_write_trylock(sem))
288 msleep(10);
289}
290
15ecb436 291struct cifsFileInfo *
fb1214e4 292cifs_new_fileinfo(struct cifs_fid *fid, struct file *file,
15ecb436
JL
293 struct tcon_link *tlink, __u32 oplock)
294{
1f1735cb 295 struct dentry *dentry = file_dentry(file);
2b0143b5 296 struct inode *inode = d_inode(dentry);
4b4de76e
PS
297 struct cifsInodeInfo *cinode = CIFS_I(inode);
298 struct cifsFileInfo *cfile;
f45d3416 299 struct cifs_fid_locks *fdlocks;
233839b1 300 struct cifs_tcon *tcon = tlink_tcon(tlink);
63b7d3a4 301 struct TCP_Server_Info *server = tcon->ses->server;
4b4de76e
PS
302
303 cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL);
304 if (cfile == NULL)
305 return cfile;
306
f45d3416
PS
307 fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL);
308 if (!fdlocks) {
309 kfree(cfile);
310 return NULL;
311 }
312
313 INIT_LIST_HEAD(&fdlocks->locks);
314 fdlocks->cfile = cfile;
315 cfile->llist = fdlocks;
d46b0da7 316 cifs_down_write(&cinode->lock_sem);
f45d3416 317 list_add(&fdlocks->llist, &cinode->llist);
1b4b55a1 318 up_write(&cinode->lock_sem);
f45d3416 319
4b4de76e 320 cfile->count = 1;
4b4de76e
PS
321 cfile->pid = current->tgid;
322 cfile->uid = current_fsuid();
323 cfile->dentry = dget(dentry);
324 cfile->f_flags = file->f_flags;
325 cfile->invalidHandle = false;
326 cfile->tlink = cifs_get_tlink(tlink);
4b4de76e 327 INIT_WORK(&cfile->oplock_break, cifs_oplock_break);
f45d3416 328 mutex_init(&cfile->fh_mutex);
3afca265 329 spin_lock_init(&cfile->file_info_lock);
15ecb436 330
24261fc2
MG
331 cifs_sb_active(inode->i_sb);
332
63b7d3a4
PS
333 /*
334 * If the server returned a read oplock and we have mandatory brlocks,
335 * set oplock level to None.
336 */
53ef1016 337 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
f96637be 338 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
63b7d3a4
PS
339 oplock = 0;
340 }
341
3afca265 342 spin_lock(&tcon->open_file_lock);
63b7d3a4 343 if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock)
233839b1
PS
344 oplock = fid->pending_open->oplock;
345 list_del(&fid->pending_open->olist);
346
42873b0a 347 fid->purge_cache = false;
63b7d3a4 348 server->ops->set_fid(cfile, fid, oplock);
233839b1
PS
349
350 list_add(&cfile->tlist, &tcon->openFileList);
fae8044c 351 atomic_inc(&tcon->num_local_opens);
3afca265 352
15ecb436 353 /* if readable file instance put first in list*/
487317c9 354 spin_lock(&cinode->open_file_lock);
15ecb436 355 if (file->f_mode & FMODE_READ)
4b4de76e 356 list_add(&cfile->flist, &cinode->openFileList);
15ecb436 357 else
4b4de76e 358 list_add_tail(&cfile->flist, &cinode->openFileList);
487317c9 359 spin_unlock(&cinode->open_file_lock);
3afca265 360 spin_unlock(&tcon->open_file_lock);
15ecb436 361
42873b0a 362 if (fid->purge_cache)
4f73c7d3 363 cifs_zap_mapping(inode);
42873b0a 364
4b4de76e
PS
365 file->private_data = cfile;
366 return cfile;
15ecb436
JL
367}
368
764a1b1a
JL
369struct cifsFileInfo *
370cifsFileInfo_get(struct cifsFileInfo *cifs_file)
371{
3afca265 372 spin_lock(&cifs_file->file_info_lock);
764a1b1a 373 cifsFileInfo_get_locked(cifs_file);
3afca265 374 spin_unlock(&cifs_file->file_info_lock);
764a1b1a
JL
375 return cifs_file;
376}
377
b98749ca
AA
378/**
379 * cifsFileInfo_put - release a reference of file priv data
380 *
381 * Always potentially wait for oplock handler. See _cifsFileInfo_put().
cdff08e7 382 */
b33879aa 383void cifsFileInfo_put(struct cifsFileInfo *cifs_file)
b98749ca
AA
384{
385 _cifsFileInfo_put(cifs_file, true);
386}
387
388/**
389 * _cifsFileInfo_put - release a reference of file priv data
390 *
391 * This may involve closing the filehandle @cifs_file out on the
392 * server. Must be called without holding tcon->open_file_lock and
393 * cifs_file->file_info_lock.
394 *
395 * If @wait_for_oplock_handler is true and we are releasing the last
396 * reference, wait for any running oplock break handler of the file
397 * and cancel any pending one. If calling this function from the
398 * oplock break handler, you need to pass false.
399 *
400 */
401void _cifsFileInfo_put(struct cifsFileInfo *cifs_file, bool wait_oplock_handler)
b33879aa 402{
2b0143b5 403 struct inode *inode = d_inode(cifs_file->dentry);
96daf2b0 404 struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
233839b1 405 struct TCP_Server_Info *server = tcon->ses->server;
e66673e3 406 struct cifsInodeInfo *cifsi = CIFS_I(inode);
24261fc2
MG
407 struct super_block *sb = inode->i_sb;
408 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
cdff08e7 409 struct cifsLockInfo *li, *tmp;
233839b1
PS
410 struct cifs_fid fid;
411 struct cifs_pending_open open;
ca7df8e0 412 bool oplock_break_cancelled;
cdff08e7 413
3afca265 414 spin_lock(&tcon->open_file_lock);
1a67c415 415 spin_lock(&cifsi->open_file_lock);
3afca265 416 spin_lock(&cifs_file->file_info_lock);
5f6dbc9e 417 if (--cifs_file->count > 0) {
3afca265 418 spin_unlock(&cifs_file->file_info_lock);
1a67c415 419 spin_unlock(&cifsi->open_file_lock);
3afca265 420 spin_unlock(&tcon->open_file_lock);
cdff08e7
SF
421 return;
422 }
3afca265 423 spin_unlock(&cifs_file->file_info_lock);
cdff08e7 424
233839b1
PS
425 if (server->ops->get_lease_key)
426 server->ops->get_lease_key(inode, &fid);
427
428 /* store open in pending opens to make sure we don't miss lease break */
429 cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open);
430
cdff08e7
SF
431 /* remove it from the lists */
432 list_del(&cifs_file->flist);
433 list_del(&cifs_file->tlist);
fae8044c 434 atomic_dec(&tcon->num_local_opens);
cdff08e7
SF
435
436 if (list_empty(&cifsi->openFileList)) {
f96637be 437 cifs_dbg(FYI, "closing last open instance for inode %p\n",
2b0143b5 438 d_inode(cifs_file->dentry));
25364138
PS
439 /*
440 * In strict cache mode we need invalidate mapping on the last
441 * close because it may cause a error when we open this file
442 * again and get at least level II oplock.
443 */
4f8ba8a0 444 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO)
aff8d5ca 445 set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags);
c6723628 446 cifs_set_oplock_level(cifsi, 0);
cdff08e7 447 }
3afca265 448
1a67c415 449 spin_unlock(&cifsi->open_file_lock);
3afca265 450 spin_unlock(&tcon->open_file_lock);
cdff08e7 451
b98749ca
AA
452 oplock_break_cancelled = wait_oplock_handler ?
453 cancel_work_sync(&cifs_file->oplock_break) : false;
ad635942 454
cdff08e7 455 if (!tcon->need_reconnect && !cifs_file->invalidHandle) {
0ff78a22 456 struct TCP_Server_Info *server = tcon->ses->server;
6d5786a3 457 unsigned int xid;
0ff78a22 458
6d5786a3 459 xid = get_xid();
0ff78a22 460 if (server->ops->close)
760ad0ca
PS
461 server->ops->close(xid, tcon, &cifs_file->fid);
462 _free_xid(xid);
cdff08e7
SF
463 }
464
ca7df8e0
SP
465 if (oplock_break_cancelled)
466 cifs_done_oplock_break(cifsi);
467
233839b1
PS
468 cifs_del_pending_open(&open);
469
f45d3416
PS
470 /*
471 * Delete any outstanding lock records. We'll lose them when the file
cdff08e7
SF
472 * is closed anyway.
473 */
d46b0da7 474 cifs_down_write(&cifsi->lock_sem);
f45d3416 475 list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) {
cdff08e7 476 list_del(&li->llist);
85160e03 477 cifs_del_lock_waiters(li);
cdff08e7 478 kfree(li);
b33879aa 479 }
f45d3416
PS
480 list_del(&cifs_file->llist->llist);
481 kfree(cifs_file->llist);
1b4b55a1 482 up_write(&cifsi->lock_sem);
cdff08e7
SF
483
484 cifs_put_tlink(cifs_file->tlink);
485 dput(cifs_file->dentry);
24261fc2 486 cifs_sb_deactive(sb);
cdff08e7 487 kfree(cifs_file);
b33879aa
JL
488}
489
1da177e4 490int cifs_open(struct inode *inode, struct file *file)
233839b1 491
1da177e4
LT
492{
493 int rc = -EACCES;
6d5786a3 494 unsigned int xid;
590a3fe0 495 __u32 oplock;
1da177e4 496 struct cifs_sb_info *cifs_sb;
b8c32dbb 497 struct TCP_Server_Info *server;
96daf2b0 498 struct cifs_tcon *tcon;
7ffec372 499 struct tcon_link *tlink;
fb1214e4 500 struct cifsFileInfo *cfile = NULL;
1da177e4 501 char *full_path = NULL;
7e12eddb 502 bool posix_open_ok = false;
fb1214e4 503 struct cifs_fid fid;
233839b1 504 struct cifs_pending_open open;
1da177e4 505
6d5786a3 506 xid = get_xid();
1da177e4
LT
507
508 cifs_sb = CIFS_SB(inode->i_sb);
7ffec372
JL
509 tlink = cifs_sb_tlink(cifs_sb);
510 if (IS_ERR(tlink)) {
6d5786a3 511 free_xid(xid);
7ffec372
JL
512 return PTR_ERR(tlink);
513 }
514 tcon = tlink_tcon(tlink);
b8c32dbb 515 server = tcon->ses->server;
1da177e4 516
1f1735cb 517 full_path = build_path_from_dentry(file_dentry(file));
1da177e4 518 if (full_path == NULL) {
0f3bc09e 519 rc = -ENOMEM;
232341ba 520 goto out;
1da177e4
LT
521 }
522
f96637be 523 cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n",
b6b38f70 524 inode, file->f_flags, full_path);
276a74a4 525
787aded6
NJ
526 if (file->f_flags & O_DIRECT &&
527 cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) {
528 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL)
529 file->f_op = &cifs_file_direct_nobrl_ops;
530 else
531 file->f_op = &cifs_file_direct_ops;
532 }
533
233839b1 534 if (server->oplocks)
276a74a4
SF
535 oplock = REQ_OPLOCK;
536 else
537 oplock = 0;
538
64cc2c63 539 if (!tcon->broken_posix_open && tcon->unix_ext &&
29e20f9c
PS
540 cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP &
541 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
276a74a4 542 /* can not refresh inode info since size could be stale */
2422f676 543 rc = cifs_posix_open(full_path, &inode, inode->i_sb,
fa588e0c 544 cifs_sb->mnt_file_mode /* ignored */,
fb1214e4 545 file->f_flags, &oplock, &fid.netfid, xid);
276a74a4 546 if (rc == 0) {
f96637be 547 cifs_dbg(FYI, "posix open succeeded\n");
7e12eddb 548 posix_open_ok = true;
64cc2c63
SF
549 } else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) {
550 if (tcon->ses->serverNOS)
f96637be
JP
551 cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n",
552 tcon->ses->serverName,
553 tcon->ses->serverNOS);
64cc2c63 554 tcon->broken_posix_open = true;
276a74a4
SF
555 } else if ((rc != -EIO) && (rc != -EREMOTE) &&
556 (rc != -EOPNOTSUPP)) /* path not found or net err */
557 goto out;
fb1214e4
PS
558 /*
559 * Else fallthrough to retry open the old way on network i/o
560 * or DFS errors.
561 */
276a74a4
SF
562 }
563
233839b1
PS
564 if (server->ops->get_lease_key)
565 server->ops->get_lease_key(inode, &fid);
566
567 cifs_add_pending_open(&fid, tlink, &open);
568
7e12eddb 569 if (!posix_open_ok) {
b8c32dbb
PS
570 if (server->ops->get_lease_key)
571 server->ops->get_lease_key(inode, &fid);
572
7e12eddb 573 rc = cifs_nt_open(full_path, inode, cifs_sb, tcon,
fb1214e4 574 file->f_flags, &oplock, &fid, xid);
233839b1
PS
575 if (rc) {
576 cifs_del_pending_open(&open);
7e12eddb 577 goto out;
233839b1 578 }
7e12eddb 579 }
47c78b7f 580
fb1214e4
PS
581 cfile = cifs_new_fileinfo(&fid, file, tlink, oplock);
582 if (cfile == NULL) {
b8c32dbb
PS
583 if (server->ops->close)
584 server->ops->close(xid, tcon, &fid);
233839b1 585 cifs_del_pending_open(&open);
1da177e4
LT
586 rc = -ENOMEM;
587 goto out;
588 }
1da177e4 589
9451a9a5
SJ
590 cifs_fscache_set_inode_cookie(inode, file);
591
7e12eddb 592 if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) {
fb1214e4
PS
593 /*
594 * Time to set mode which we can not set earlier due to
595 * problems creating new read-only files.
596 */
7e12eddb
PS
597 struct cifs_unix_set_info_args args = {
598 .mode = inode->i_mode,
49418b2c
EB
599 .uid = INVALID_UID, /* no change */
600 .gid = INVALID_GID, /* no change */
7e12eddb
PS
601 .ctime = NO_CHANGE_64,
602 .atime = NO_CHANGE_64,
603 .mtime = NO_CHANGE_64,
604 .device = 0,
605 };
fb1214e4
PS
606 CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid,
607 cfile->pid);
1da177e4
LT
608 }
609
610out:
1da177e4 611 kfree(full_path);
6d5786a3 612 free_xid(xid);
7ffec372 613 cifs_put_tlink(tlink);
1da177e4
LT
614 return rc;
615}
616
f152fd5f
PS
617static int cifs_push_posix_locks(struct cifsFileInfo *cfile);
618
2ae78ba8
PS
619/*
620 * Try to reacquire byte range locks that were released when session
f152fd5f 621 * to server was lost.
2ae78ba8 622 */
f152fd5f
PS
623static int
624cifs_relock_file(struct cifsFileInfo *cfile)
1da177e4 625{
f152fd5f 626 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2b0143b5 627 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
f152fd5f 628 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1da177e4
LT
629 int rc = 0;
630
560d3889 631 down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING);
f152fd5f 632 if (cinode->can_cache_brlcks) {
689c3db4
PS
633 /* can cache locks - no need to relock */
634 up_read(&cinode->lock_sem);
f152fd5f
PS
635 return rc;
636 }
637
638 if (cap_unix(tcon->ses) &&
639 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
640 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
641 rc = cifs_push_posix_locks(cfile);
642 else
643 rc = tcon->ses->server->ops->push_mand_locks(cfile);
1da177e4 644
689c3db4 645 up_read(&cinode->lock_sem);
1da177e4
LT
646 return rc;
647}
648
2ae78ba8
PS
649static int
650cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush)
1da177e4
LT
651{
652 int rc = -EACCES;
6d5786a3 653 unsigned int xid;
590a3fe0 654 __u32 oplock;
1da177e4 655 struct cifs_sb_info *cifs_sb;
96daf2b0 656 struct cifs_tcon *tcon;
2ae78ba8
PS
657 struct TCP_Server_Info *server;
658 struct cifsInodeInfo *cinode;
fb8c4b14 659 struct inode *inode;
1da177e4 660 char *full_path = NULL;
2ae78ba8 661 int desired_access;
1da177e4 662 int disposition = FILE_OPEN;
3d3ea8e6 663 int create_options = CREATE_NOT_DIR;
226730b4 664 struct cifs_open_parms oparms;
1da177e4 665
6d5786a3 666 xid = get_xid();
2ae78ba8
PS
667 mutex_lock(&cfile->fh_mutex);
668 if (!cfile->invalidHandle) {
669 mutex_unlock(&cfile->fh_mutex);
0f3bc09e 670 rc = 0;
6d5786a3 671 free_xid(xid);
0f3bc09e 672 return rc;
1da177e4
LT
673 }
674
2b0143b5 675 inode = d_inode(cfile->dentry);
1da177e4 676 cifs_sb = CIFS_SB(inode->i_sb);
2ae78ba8
PS
677 tcon = tlink_tcon(cfile->tlink);
678 server = tcon->ses->server;
679
680 /*
681 * Can not grab rename sem here because various ops, including those
682 * that already have the rename sem can end up causing writepage to get
683 * called and if the server was down that means we end up here, and we
684 * can never tell if the caller already has the rename_sem.
685 */
686 full_path = build_path_from_dentry(cfile->dentry);
1da177e4 687 if (full_path == NULL) {
3a9f462f 688 rc = -ENOMEM;
2ae78ba8 689 mutex_unlock(&cfile->fh_mutex);
6d5786a3 690 free_xid(xid);
3a9f462f 691 return rc;
1da177e4
LT
692 }
693
f96637be
JP
694 cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n",
695 inode, cfile->f_flags, full_path);
1da177e4 696
10b9b98e 697 if (tcon->ses->server->oplocks)
1da177e4
LT
698 oplock = REQ_OPLOCK;
699 else
4b18f2a9 700 oplock = 0;
1da177e4 701
29e20f9c 702 if (tcon->unix_ext && cap_unix(tcon->ses) &&
7fc8f4e9 703 (CIFS_UNIX_POSIX_PATH_OPS_CAP &
29e20f9c 704 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
608712fe
JL
705 /*
706 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the
707 * original open. Must mask them off for a reopen.
708 */
2ae78ba8 709 unsigned int oflags = cfile->f_flags &
15886177 710 ~(O_CREAT | O_EXCL | O_TRUNC);
608712fe 711
2422f676 712 rc = cifs_posix_open(full_path, NULL, inode->i_sb,
2ae78ba8 713 cifs_sb->mnt_file_mode /* ignored */,
9cbc0b73 714 oflags, &oplock, &cfile->fid.netfid, xid);
7fc8f4e9 715 if (rc == 0) {
f96637be 716 cifs_dbg(FYI, "posix reopen succeeded\n");
fe090e4e 717 oparms.reconnect = true;
7fc8f4e9
SF
718 goto reopen_success;
719 }
2ae78ba8
PS
720 /*
721 * fallthrough to retry open the old way on errors, especially
722 * in the reconnect path it is important to retry hard
723 */
7fc8f4e9
SF
724 }
725
2ae78ba8 726 desired_access = cifs_convert_flags(cfile->f_flags);
7fc8f4e9 727
3d3ea8e6
SP
728 if (backup_cred(cifs_sb))
729 create_options |= CREATE_OPEN_BACKUP_INTENT;
730
b8c32dbb 731 if (server->ops->get_lease_key)
9cbc0b73 732 server->ops->get_lease_key(inode, &cfile->fid);
b8c32dbb 733
226730b4
PS
734 oparms.tcon = tcon;
735 oparms.cifs_sb = cifs_sb;
736 oparms.desired_access = desired_access;
737 oparms.create_options = create_options;
738 oparms.disposition = disposition;
739 oparms.path = full_path;
9cbc0b73
PS
740 oparms.fid = &cfile->fid;
741 oparms.reconnect = true;
226730b4 742
2ae78ba8
PS
743 /*
744 * Can not refresh inode by passing in file_info buf to be returned by
d81b8a40 745 * ops->open and then calling get_inode_info with returned buf since
2ae78ba8
PS
746 * file might have write behind data that needs to be flushed and server
747 * version of file size can be stale. If we knew for sure that inode was
748 * not dirty locally we could do this.
749 */
226730b4 750 rc = server->ops->open(xid, &oparms, &oplock, NULL);
b33fcf1c
PS
751 if (rc == -ENOENT && oparms.reconnect == false) {
752 /* durable handle timeout is expired - open the file again */
753 rc = server->ops->open(xid, &oparms, &oplock, NULL);
754 /* indicate that we need to relock the file */
755 oparms.reconnect = true;
756 }
757
1da177e4 758 if (rc) {
2ae78ba8 759 mutex_unlock(&cfile->fh_mutex);
f96637be
JP
760 cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc);
761 cifs_dbg(FYI, "oplock: %d\n", oplock);
15886177
JL
762 goto reopen_error_exit;
763 }
764
7fc8f4e9 765reopen_success:
2ae78ba8
PS
766 cfile->invalidHandle = false;
767 mutex_unlock(&cfile->fh_mutex);
768 cinode = CIFS_I(inode);
15886177
JL
769
770 if (can_flush) {
771 rc = filemap_write_and_wait(inode->i_mapping);
9a66396f
PS
772 if (!is_interrupt_error(rc))
773 mapping_set_error(inode->i_mapping, rc);
15886177 774
15886177 775 if (tcon->unix_ext)
2ae78ba8
PS
776 rc = cifs_get_inode_info_unix(&inode, full_path,
777 inode->i_sb, xid);
15886177 778 else
2ae78ba8
PS
779 rc = cifs_get_inode_info(&inode, full_path, NULL,
780 inode->i_sb, xid, NULL);
781 }
782 /*
783 * Else we are writing out data to server already and could deadlock if
784 * we tried to flush data, and since we do not know if we have data that
785 * would invalidate the current end of file on the server we can not go
786 * to the server to get the new inode info.
787 */
788
de740250
PS
789 /*
790 * If the server returned a read oplock and we have mandatory brlocks,
791 * set oplock level to None.
792 */
793 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
794 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
795 oplock = 0;
796 }
797
9cbc0b73
PS
798 server->ops->set_fid(cfile, &cfile->fid, oplock);
799 if (oparms.reconnect)
800 cifs_relock_file(cfile);
15886177
JL
801
802reopen_error_exit:
1da177e4 803 kfree(full_path);
6d5786a3 804 free_xid(xid);
1da177e4
LT
805 return rc;
806}
807
808int cifs_close(struct inode *inode, struct file *file)
809{
77970693
JL
810 if (file->private_data != NULL) {
811 cifsFileInfo_put(file->private_data);
812 file->private_data = NULL;
813 }
7ee1af76 814
cdff08e7
SF
815 /* return code from the ->release op is always ignored */
816 return 0;
1da177e4
LT
817}
818
52ace1ef
SF
819void
820cifs_reopen_persistent_handles(struct cifs_tcon *tcon)
821{
f2cca6a7 822 struct cifsFileInfo *open_file;
52ace1ef
SF
823 struct list_head *tmp;
824 struct list_head *tmp1;
f2cca6a7
PS
825 struct list_head tmp_list;
826
96a988ff
PS
827 if (!tcon->use_persistent || !tcon->need_reopen_files)
828 return;
829
830 tcon->need_reopen_files = false;
831
f2cca6a7
PS
832 cifs_dbg(FYI, "Reopen persistent handles");
833 INIT_LIST_HEAD(&tmp_list);
52ace1ef
SF
834
835 /* list all files open on tree connection, reopen resilient handles */
836 spin_lock(&tcon->open_file_lock);
f2cca6a7 837 list_for_each(tmp, &tcon->openFileList) {
52ace1ef 838 open_file = list_entry(tmp, struct cifsFileInfo, tlist);
f2cca6a7
PS
839 if (!open_file->invalidHandle)
840 continue;
841 cifsFileInfo_get(open_file);
842 list_add_tail(&open_file->rlist, &tmp_list);
52ace1ef
SF
843 }
844 spin_unlock(&tcon->open_file_lock);
f2cca6a7
PS
845
846 list_for_each_safe(tmp, tmp1, &tmp_list) {
847 open_file = list_entry(tmp, struct cifsFileInfo, rlist);
96a988ff
PS
848 if (cifs_reopen_file(open_file, false /* do not flush */))
849 tcon->need_reopen_files = true;
f2cca6a7
PS
850 list_del_init(&open_file->rlist);
851 cifsFileInfo_put(open_file);
852 }
52ace1ef
SF
853}
854
1da177e4
LT
855int cifs_closedir(struct inode *inode, struct file *file)
856{
857 int rc = 0;
6d5786a3 858 unsigned int xid;
4b4de76e 859 struct cifsFileInfo *cfile = file->private_data;
92fc65a7
PS
860 struct cifs_tcon *tcon;
861 struct TCP_Server_Info *server;
862 char *buf;
1da177e4 863
f96637be 864 cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode);
1da177e4 865
92fc65a7
PS
866 if (cfile == NULL)
867 return rc;
868
6d5786a3 869 xid = get_xid();
92fc65a7
PS
870 tcon = tlink_tcon(cfile->tlink);
871 server = tcon->ses->server;
1da177e4 872
f96637be 873 cifs_dbg(FYI, "Freeing private data in close dir\n");
3afca265 874 spin_lock(&cfile->file_info_lock);
52755808 875 if (server->ops->dir_needs_close(cfile)) {
92fc65a7 876 cfile->invalidHandle = true;
3afca265 877 spin_unlock(&cfile->file_info_lock);
92fc65a7
PS
878 if (server->ops->close_dir)
879 rc = server->ops->close_dir(xid, tcon, &cfile->fid);
880 else
881 rc = -ENOSYS;
f96637be 882 cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc);
92fc65a7
PS
883 /* not much we can do if it fails anyway, ignore rc */
884 rc = 0;
885 } else
3afca265 886 spin_unlock(&cfile->file_info_lock);
92fc65a7
PS
887
888 buf = cfile->srch_inf.ntwrk_buf_start;
889 if (buf) {
f96637be 890 cifs_dbg(FYI, "closedir free smb buf in srch struct\n");
92fc65a7
PS
891 cfile->srch_inf.ntwrk_buf_start = NULL;
892 if (cfile->srch_inf.smallBuf)
893 cifs_small_buf_release(buf);
894 else
895 cifs_buf_release(buf);
1da177e4 896 }
92fc65a7
PS
897
898 cifs_put_tlink(cfile->tlink);
899 kfree(file->private_data);
900 file->private_data = NULL;
1da177e4 901 /* BB can we lock the filestruct while this is going on? */
6d5786a3 902 free_xid(xid);
1da177e4
LT
903 return rc;
904}
905
85160e03 906static struct cifsLockInfo *
9645759c 907cifs_lock_init(__u64 offset, __u64 length, __u8 type, __u16 flags)
7ee1af76 908{
a88b4707 909 struct cifsLockInfo *lock =
fb8c4b14 910 kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL);
a88b4707
PS
911 if (!lock)
912 return lock;
913 lock->offset = offset;
914 lock->length = length;
915 lock->type = type;
a88b4707 916 lock->pid = current->tgid;
9645759c 917 lock->flags = flags;
a88b4707
PS
918 INIT_LIST_HEAD(&lock->blist);
919 init_waitqueue_head(&lock->block_q);
920 return lock;
85160e03
PS
921}
922
f7ba7fe6 923void
85160e03
PS
924cifs_del_lock_waiters(struct cifsLockInfo *lock)
925{
926 struct cifsLockInfo *li, *tmp;
927 list_for_each_entry_safe(li, tmp, &lock->blist, blist) {
928 list_del_init(&li->blist);
929 wake_up(&li->block_q);
930 }
931}
932
081c0414
PS
933#define CIFS_LOCK_OP 0
934#define CIFS_READ_OP 1
935#define CIFS_WRITE_OP 2
936
937/* @rw_check : 0 - no op, 1 - read, 2 - write */
85160e03 938static bool
f45d3416 939cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset,
9645759c
RS
940 __u64 length, __u8 type, __u16 flags,
941 struct cifsFileInfo *cfile,
081c0414 942 struct cifsLockInfo **conf_lock, int rw_check)
85160e03 943{
fbd35aca 944 struct cifsLockInfo *li;
f45d3416 945 struct cifsFileInfo *cur_cfile = fdlocks->cfile;
106dc538 946 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
85160e03 947
f45d3416 948 list_for_each_entry(li, &fdlocks->locks, llist) {
85160e03
PS
949 if (offset + length <= li->offset ||
950 offset >= li->offset + li->length)
951 continue;
081c0414
PS
952 if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
953 server->ops->compare_fids(cfile, cur_cfile)) {
954 /* shared lock prevents write op through the same fid */
955 if (!(li->type & server->vals->shared_lock_type) ||
956 rw_check != CIFS_WRITE_OP)
957 continue;
958 }
f45d3416
PS
959 if ((type & server->vals->shared_lock_type) &&
960 ((server->ops->compare_fids(cfile, cur_cfile) &&
961 current->tgid == li->pid) || type == li->type))
85160e03 962 continue;
9645759c
RS
963 if (rw_check == CIFS_LOCK_OP &&
964 (flags & FL_OFDLCK) && (li->flags & FL_OFDLCK) &&
965 server->ops->compare_fids(cfile, cur_cfile))
966 continue;
579f9053
PS
967 if (conf_lock)
968 *conf_lock = li;
f45d3416 969 return true;
85160e03
PS
970 }
971 return false;
972}
973
579f9053 974bool
55157dfb 975cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
9645759c
RS
976 __u8 type, __u16 flags,
977 struct cifsLockInfo **conf_lock, int rw_check)
161ebf9f 978{
fbd35aca 979 bool rc = false;
f45d3416 980 struct cifs_fid_locks *cur;
2b0143b5 981 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
fbd35aca 982
f45d3416
PS
983 list_for_each_entry(cur, &cinode->llist, llist) {
984 rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
9645759c
RS
985 flags, cfile, conf_lock,
986 rw_check);
fbd35aca
PS
987 if (rc)
988 break;
989 }
fbd35aca
PS
990
991 return rc;
161ebf9f
PS
992}
993
9a5101c8
PS
994/*
995 * Check if there is another lock that prevents us to set the lock (mandatory
996 * style). If such a lock exists, update the flock structure with its
997 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
998 * or leave it the same if we can't. Returns 0 if we don't need to request to
999 * the server or 1 otherwise.
1000 */
85160e03 1001static int
fbd35aca
PS
1002cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1003 __u8 type, struct file_lock *flock)
85160e03
PS
1004{
1005 int rc = 0;
1006 struct cifsLockInfo *conf_lock;
2b0143b5 1007 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
106dc538 1008 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
85160e03
PS
1009 bool exist;
1010
1b4b55a1 1011 down_read(&cinode->lock_sem);
85160e03 1012
55157dfb 1013 exist = cifs_find_lock_conflict(cfile, offset, length, type,
9645759c
RS
1014 flock->fl_flags, &conf_lock,
1015 CIFS_LOCK_OP);
85160e03
PS
1016 if (exist) {
1017 flock->fl_start = conf_lock->offset;
1018 flock->fl_end = conf_lock->offset + conf_lock->length - 1;
1019 flock->fl_pid = conf_lock->pid;
106dc538 1020 if (conf_lock->type & server->vals->shared_lock_type)
85160e03
PS
1021 flock->fl_type = F_RDLCK;
1022 else
1023 flock->fl_type = F_WRLCK;
1024 } else if (!cinode->can_cache_brlcks)
1025 rc = 1;
1026 else
1027 flock->fl_type = F_UNLCK;
1028
1b4b55a1 1029 up_read(&cinode->lock_sem);
85160e03
PS
1030 return rc;
1031}
1032
161ebf9f 1033static void
fbd35aca 1034cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock)
85160e03 1035{
2b0143b5 1036 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
d46b0da7 1037 cifs_down_write(&cinode->lock_sem);
f45d3416 1038 list_add_tail(&lock->llist, &cfile->llist->locks);
1b4b55a1 1039 up_write(&cinode->lock_sem);
7ee1af76
JA
1040}
1041
9a5101c8
PS
1042/*
1043 * Set the byte-range lock (mandatory style). Returns:
1044 * 1) 0, if we set the lock and don't need to request to the server;
1045 * 2) 1, if no locks prevent us but we need to request to the server;
413d6100 1046 * 3) -EACCES, if there is a lock that prevents us and wait is false.
9a5101c8 1047 */
85160e03 1048static int
fbd35aca 1049cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock,
161ebf9f 1050 bool wait)
85160e03 1051{
161ebf9f 1052 struct cifsLockInfo *conf_lock;
2b0143b5 1053 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
85160e03
PS
1054 bool exist;
1055 int rc = 0;
1056
85160e03
PS
1057try_again:
1058 exist = false;
d46b0da7 1059 cifs_down_write(&cinode->lock_sem);
85160e03 1060
55157dfb 1061 exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length,
9645759c
RS
1062 lock->type, lock->flags, &conf_lock,
1063 CIFS_LOCK_OP);
85160e03 1064 if (!exist && cinode->can_cache_brlcks) {
f45d3416 1065 list_add_tail(&lock->llist, &cfile->llist->locks);
1b4b55a1 1066 up_write(&cinode->lock_sem);
85160e03
PS
1067 return rc;
1068 }
1069
1070 if (!exist)
1071 rc = 1;
1072 else if (!wait)
1073 rc = -EACCES;
1074 else {
1075 list_add_tail(&lock->blist, &conf_lock->blist);
1b4b55a1 1076 up_write(&cinode->lock_sem);
85160e03
PS
1077 rc = wait_event_interruptible(lock->block_q,
1078 (lock->blist.prev == &lock->blist) &&
1079 (lock->blist.next == &lock->blist));
1080 if (!rc)
1081 goto try_again;
d46b0da7 1082 cifs_down_write(&cinode->lock_sem);
a88b4707 1083 list_del_init(&lock->blist);
85160e03
PS
1084 }
1085
1b4b55a1 1086 up_write(&cinode->lock_sem);
85160e03
PS
1087 return rc;
1088}
1089
9a5101c8
PS
1090/*
1091 * Check if there is another lock that prevents us to set the lock (posix
1092 * style). If such a lock exists, update the flock structure with its
1093 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1094 * or leave it the same if we can't. Returns 0 if we don't need to request to
1095 * the server or 1 otherwise.
1096 */
85160e03 1097static int
4f6bcec9
PS
1098cifs_posix_lock_test(struct file *file, struct file_lock *flock)
1099{
1100 int rc = 0;
496ad9aa 1101 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
4f6bcec9
PS
1102 unsigned char saved_type = flock->fl_type;
1103
50792760
PS
1104 if ((flock->fl_flags & FL_POSIX) == 0)
1105 return 1;
1106
1b4b55a1 1107 down_read(&cinode->lock_sem);
4f6bcec9
PS
1108 posix_test_lock(file, flock);
1109
1110 if (flock->fl_type == F_UNLCK && !cinode->can_cache_brlcks) {
1111 flock->fl_type = saved_type;
1112 rc = 1;
1113 }
1114
1b4b55a1 1115 up_read(&cinode->lock_sem);
4f6bcec9
PS
1116 return rc;
1117}
1118
9a5101c8
PS
1119/*
1120 * Set the byte-range lock (posix style). Returns:
1121 * 1) 0, if we set the lock and don't need to request to the server;
1122 * 2) 1, if we need to request to the server;
1123 * 3) <0, if the error occurs while setting the lock.
1124 */
4f6bcec9
PS
1125static int
1126cifs_posix_lock_set(struct file *file, struct file_lock *flock)
1127{
496ad9aa 1128 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
50792760
PS
1129 int rc = 1;
1130
1131 if ((flock->fl_flags & FL_POSIX) == 0)
1132 return rc;
4f6bcec9 1133
66189be7 1134try_again:
d46b0da7 1135 cifs_down_write(&cinode->lock_sem);
4f6bcec9 1136 if (!cinode->can_cache_brlcks) {
1b4b55a1 1137 up_write(&cinode->lock_sem);
50792760 1138 return rc;
4f6bcec9 1139 }
66189be7
PS
1140
1141 rc = posix_lock_file(file, flock, NULL);
1b4b55a1 1142 up_write(&cinode->lock_sem);
66189be7 1143 if (rc == FILE_LOCK_DEFERRED) {
ada5c1da 1144 rc = wait_event_interruptible(flock->fl_wait, !flock->fl_blocker);
66189be7
PS
1145 if (!rc)
1146 goto try_again;
cb03f94f 1147 locks_delete_block(flock);
66189be7 1148 }
9ebb389d 1149 return rc;
4f6bcec9
PS
1150}
1151
d39a4f71 1152int
4f6bcec9 1153cifs_push_mandatory_locks(struct cifsFileInfo *cfile)
85160e03 1154{
6d5786a3
PS
1155 unsigned int xid;
1156 int rc = 0, stored_rc;
85160e03
PS
1157 struct cifsLockInfo *li, *tmp;
1158 struct cifs_tcon *tcon;
0013fb4c 1159 unsigned int num, max_num, max_buf;
32b9aaf1 1160 LOCKING_ANDX_RANGE *buf, *cur;
4d61eda8
CIK
1161 static const int types[] = {
1162 LOCKING_ANDX_LARGE_FILES,
1163 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1164 };
32b9aaf1 1165 int i;
85160e03 1166
6d5786a3 1167 xid = get_xid();
85160e03
PS
1168 tcon = tlink_tcon(cfile->tlink);
1169
0013fb4c
PS
1170 /*
1171 * Accessing maxBuf is racy with cifs_reconnect - need to store value
b9a74cde 1172 * and check it before using.
0013fb4c
PS
1173 */
1174 max_buf = tcon->ses->server->maxBuf;
b9a74cde 1175 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) {
6d5786a3 1176 free_xid(xid);
0013fb4c
PS
1177 return -EINVAL;
1178 }
1179
92a8109e
RL
1180 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1181 PAGE_SIZE);
1182 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1183 PAGE_SIZE);
0013fb4c
PS
1184 max_num = (max_buf - sizeof(struct smb_hdr)) /
1185 sizeof(LOCKING_ANDX_RANGE);
4b99d39b 1186 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
32b9aaf1 1187 if (!buf) {
6d5786a3 1188 free_xid(xid);
e2f2886a 1189 return -ENOMEM;
32b9aaf1
PS
1190 }
1191
1192 for (i = 0; i < 2; i++) {
1193 cur = buf;
1194 num = 0;
f45d3416 1195 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
32b9aaf1
PS
1196 if (li->type != types[i])
1197 continue;
1198 cur->Pid = cpu_to_le16(li->pid);
1199 cur->LengthLow = cpu_to_le32((u32)li->length);
1200 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1201 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1202 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1203 if (++num == max_num) {
4b4de76e
PS
1204 stored_rc = cifs_lockv(xid, tcon,
1205 cfile->fid.netfid,
04a6aa8a
PS
1206 (__u8)li->type, 0, num,
1207 buf);
32b9aaf1
PS
1208 if (stored_rc)
1209 rc = stored_rc;
1210 cur = buf;
1211 num = 0;
1212 } else
1213 cur++;
1214 }
1215
1216 if (num) {
4b4de76e 1217 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
04a6aa8a 1218 (__u8)types[i], 0, num, buf);
32b9aaf1
PS
1219 if (stored_rc)
1220 rc = stored_rc;
1221 }
85160e03
PS
1222 }
1223
32b9aaf1 1224 kfree(buf);
6d5786a3 1225 free_xid(xid);
85160e03
PS
1226 return rc;
1227}
1228
3d22462a
JL
1229static __u32
1230hash_lockowner(fl_owner_t owner)
1231{
1232 return cifs_lock_secret ^ hash32_ptr((const void *)owner);
1233}
1234
d5751469
PS
1235struct lock_to_push {
1236 struct list_head llist;
1237 __u64 offset;
1238 __u64 length;
1239 __u32 pid;
1240 __u16 netfid;
1241 __u8 type;
1242};
1243
4f6bcec9 1244static int
b8db928b 1245cifs_push_posix_locks(struct cifsFileInfo *cfile)
4f6bcec9 1246{
2b0143b5 1247 struct inode *inode = d_inode(cfile->dentry);
4f6bcec9 1248 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
bd61e0a9
JL
1249 struct file_lock *flock;
1250 struct file_lock_context *flctx = inode->i_flctx;
e084c1bd 1251 unsigned int count = 0, i;
4f6bcec9 1252 int rc = 0, xid, type;
d5751469
PS
1253 struct list_head locks_to_send, *el;
1254 struct lock_to_push *lck, *tmp;
4f6bcec9 1255 __u64 length;
4f6bcec9 1256
6d5786a3 1257 xid = get_xid();
4f6bcec9 1258
bd61e0a9
JL
1259 if (!flctx)
1260 goto out;
d5751469 1261
e084c1bd
JL
1262 spin_lock(&flctx->flc_lock);
1263 list_for_each(el, &flctx->flc_posix) {
1264 count++;
1265 }
1266 spin_unlock(&flctx->flc_lock);
1267
4f6bcec9
PS
1268 INIT_LIST_HEAD(&locks_to_send);
1269
d5751469 1270 /*
e084c1bd
JL
1271 * Allocating count locks is enough because no FL_POSIX locks can be
1272 * added to the list while we are holding cinode->lock_sem that
ce85852b 1273 * protects locking operations of this inode.
d5751469 1274 */
e084c1bd 1275 for (i = 0; i < count; i++) {
d5751469
PS
1276 lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL);
1277 if (!lck) {
1278 rc = -ENOMEM;
1279 goto err_out;
1280 }
1281 list_add_tail(&lck->llist, &locks_to_send);
1282 }
1283
d5751469 1284 el = locks_to_send.next;
6109c850 1285 spin_lock(&flctx->flc_lock);
bd61e0a9 1286 list_for_each_entry(flock, &flctx->flc_posix, fl_list) {
d5751469 1287 if (el == &locks_to_send) {
ce85852b
PS
1288 /*
1289 * The list ended. We don't have enough allocated
1290 * structures - something is really wrong.
1291 */
f96637be 1292 cifs_dbg(VFS, "Can't push all brlocks!\n");
d5751469
PS
1293 break;
1294 }
4f6bcec9
PS
1295 length = 1 + flock->fl_end - flock->fl_start;
1296 if (flock->fl_type == F_RDLCK || flock->fl_type == F_SHLCK)
1297 type = CIFS_RDLCK;
1298 else
1299 type = CIFS_WRLCK;
d5751469 1300 lck = list_entry(el, struct lock_to_push, llist);
3d22462a 1301 lck->pid = hash_lockowner(flock->fl_owner);
4b4de76e 1302 lck->netfid = cfile->fid.netfid;
d5751469
PS
1303 lck->length = length;
1304 lck->type = type;
1305 lck->offset = flock->fl_start;
4f6bcec9 1306 }
6109c850 1307 spin_unlock(&flctx->flc_lock);
4f6bcec9
PS
1308
1309 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
4f6bcec9
PS
1310 int stored_rc;
1311
4f6bcec9 1312 stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid,
c5fd363d 1313 lck->offset, lck->length, NULL,
4f6bcec9
PS
1314 lck->type, 0);
1315 if (stored_rc)
1316 rc = stored_rc;
1317 list_del(&lck->llist);
1318 kfree(lck);
1319 }
1320
d5751469 1321out:
6d5786a3 1322 free_xid(xid);
4f6bcec9 1323 return rc;
d5751469
PS
1324err_out:
1325 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1326 list_del(&lck->llist);
1327 kfree(lck);
1328 }
1329 goto out;
4f6bcec9
PS
1330}
1331
9ec3c882 1332static int
b8db928b 1333cifs_push_locks(struct cifsFileInfo *cfile)
9ec3c882 1334{
b8db928b 1335 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2b0143b5 1336 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
b8db928b 1337 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
9ec3c882
PS
1338 int rc = 0;
1339
1340 /* we are going to update can_cache_brlcks here - need a write access */
d46b0da7 1341 cifs_down_write(&cinode->lock_sem);
9ec3c882
PS
1342 if (!cinode->can_cache_brlcks) {
1343 up_write(&cinode->lock_sem);
1344 return rc;
1345 }
4f6bcec9 1346
29e20f9c 1347 if (cap_unix(tcon->ses) &&
4f6bcec9
PS
1348 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1349 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
b8db928b
PS
1350 rc = cifs_push_posix_locks(cfile);
1351 else
1352 rc = tcon->ses->server->ops->push_mand_locks(cfile);
4f6bcec9 1353
b8db928b
PS
1354 cinode->can_cache_brlcks = false;
1355 up_write(&cinode->lock_sem);
1356 return rc;
4f6bcec9
PS
1357}
1358
03776f45 1359static void
04a6aa8a 1360cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock,
106dc538 1361 bool *wait_flag, struct TCP_Server_Info *server)
1da177e4 1362{
03776f45 1363 if (flock->fl_flags & FL_POSIX)
f96637be 1364 cifs_dbg(FYI, "Posix\n");
03776f45 1365 if (flock->fl_flags & FL_FLOCK)
f96637be 1366 cifs_dbg(FYI, "Flock\n");
03776f45 1367 if (flock->fl_flags & FL_SLEEP) {
f96637be 1368 cifs_dbg(FYI, "Blocking lock\n");
03776f45 1369 *wait_flag = true;
1da177e4 1370 }
03776f45 1371 if (flock->fl_flags & FL_ACCESS)
f96637be 1372 cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n");
03776f45 1373 if (flock->fl_flags & FL_LEASE)
f96637be 1374 cifs_dbg(FYI, "Lease on file - not implemented yet\n");
03776f45 1375 if (flock->fl_flags &
3d6d854a 1376 (~(FL_POSIX | FL_FLOCK | FL_SLEEP |
9645759c 1377 FL_ACCESS | FL_LEASE | FL_CLOSE | FL_OFDLCK)))
f96637be 1378 cifs_dbg(FYI, "Unknown lock flags 0x%x\n", flock->fl_flags);
1da177e4 1379
106dc538 1380 *type = server->vals->large_lock_type;
03776f45 1381 if (flock->fl_type == F_WRLCK) {
f96637be 1382 cifs_dbg(FYI, "F_WRLCK\n");
106dc538 1383 *type |= server->vals->exclusive_lock_type;
03776f45
PS
1384 *lock = 1;
1385 } else if (flock->fl_type == F_UNLCK) {
f96637be 1386 cifs_dbg(FYI, "F_UNLCK\n");
106dc538 1387 *type |= server->vals->unlock_lock_type;
03776f45
PS
1388 *unlock = 1;
1389 /* Check if unlock includes more than one lock range */
1390 } else if (flock->fl_type == F_RDLCK) {
f96637be 1391 cifs_dbg(FYI, "F_RDLCK\n");
106dc538 1392 *type |= server->vals->shared_lock_type;
03776f45
PS
1393 *lock = 1;
1394 } else if (flock->fl_type == F_EXLCK) {
f96637be 1395 cifs_dbg(FYI, "F_EXLCK\n");
106dc538 1396 *type |= server->vals->exclusive_lock_type;
03776f45
PS
1397 *lock = 1;
1398 } else if (flock->fl_type == F_SHLCK) {
f96637be 1399 cifs_dbg(FYI, "F_SHLCK\n");
106dc538 1400 *type |= server->vals->shared_lock_type;
03776f45 1401 *lock = 1;
1da177e4 1402 } else
f96637be 1403 cifs_dbg(FYI, "Unknown type of lock\n");
03776f45 1404}
1da177e4 1405
03776f45 1406static int
04a6aa8a 1407cifs_getlk(struct file *file, struct file_lock *flock, __u32 type,
6d5786a3 1408 bool wait_flag, bool posix_lck, unsigned int xid)
03776f45
PS
1409{
1410 int rc = 0;
1411 __u64 length = 1 + flock->fl_end - flock->fl_start;
4f6bcec9
PS
1412 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1413 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
106dc538 1414 struct TCP_Server_Info *server = tcon->ses->server;
4b4de76e 1415 __u16 netfid = cfile->fid.netfid;
f05337c6 1416
03776f45
PS
1417 if (posix_lck) {
1418 int posix_lock_type;
4f6bcec9
PS
1419
1420 rc = cifs_posix_lock_test(file, flock);
1421 if (!rc)
1422 return rc;
1423
106dc538 1424 if (type & server->vals->shared_lock_type)
03776f45
PS
1425 posix_lock_type = CIFS_RDLCK;
1426 else
1427 posix_lock_type = CIFS_WRLCK;
3d22462a
JL
1428 rc = CIFSSMBPosixLock(xid, tcon, netfid,
1429 hash_lockowner(flock->fl_owner),
c5fd363d 1430 flock->fl_start, length, flock,
4f6bcec9 1431 posix_lock_type, wait_flag);
03776f45
PS
1432 return rc;
1433 }
1da177e4 1434
fbd35aca 1435 rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock);
85160e03
PS
1436 if (!rc)
1437 return rc;
1438
03776f45 1439 /* BB we could chain these into one lock request BB */
d39a4f71
PS
1440 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type,
1441 1, 0, false);
03776f45 1442 if (rc == 0) {
d39a4f71
PS
1443 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1444 type, 0, 1, false);
03776f45
PS
1445 flock->fl_type = F_UNLCK;
1446 if (rc != 0)
f96637be
JP
1447 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1448 rc);
a88b4707 1449 return 0;
1da177e4 1450 }
7ee1af76 1451
106dc538 1452 if (type & server->vals->shared_lock_type) {
03776f45 1453 flock->fl_type = F_WRLCK;
a88b4707 1454 return 0;
7ee1af76
JA
1455 }
1456
d39a4f71
PS
1457 type &= ~server->vals->exclusive_lock_type;
1458
1459 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1460 type | server->vals->shared_lock_type,
1461 1, 0, false);
03776f45 1462 if (rc == 0) {
d39a4f71
PS
1463 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1464 type | server->vals->shared_lock_type, 0, 1, false);
03776f45
PS
1465 flock->fl_type = F_RDLCK;
1466 if (rc != 0)
f96637be
JP
1467 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1468 rc);
03776f45
PS
1469 } else
1470 flock->fl_type = F_WRLCK;
1471
a88b4707 1472 return 0;
03776f45
PS
1473}
1474
f7ba7fe6 1475void
9ee305b7
PS
1476cifs_move_llist(struct list_head *source, struct list_head *dest)
1477{
1478 struct list_head *li, *tmp;
1479 list_for_each_safe(li, tmp, source)
1480 list_move(li, dest);
1481}
1482
f7ba7fe6 1483void
9ee305b7
PS
1484cifs_free_llist(struct list_head *llist)
1485{
1486 struct cifsLockInfo *li, *tmp;
1487 list_for_each_entry_safe(li, tmp, llist, llist) {
1488 cifs_del_lock_waiters(li);
1489 list_del(&li->llist);
1490 kfree(li);
1491 }
1492}
1493
d39a4f71 1494int
6d5786a3
PS
1495cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock,
1496 unsigned int xid)
9ee305b7
PS
1497{
1498 int rc = 0, stored_rc;
4d61eda8
CIK
1499 static const int types[] = {
1500 LOCKING_ANDX_LARGE_FILES,
1501 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1502 };
9ee305b7 1503 unsigned int i;
0013fb4c 1504 unsigned int max_num, num, max_buf;
9ee305b7
PS
1505 LOCKING_ANDX_RANGE *buf, *cur;
1506 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2b0143b5 1507 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
9ee305b7
PS
1508 struct cifsLockInfo *li, *tmp;
1509 __u64 length = 1 + flock->fl_end - flock->fl_start;
1510 struct list_head tmp_llist;
1511
1512 INIT_LIST_HEAD(&tmp_llist);
1513
0013fb4c
PS
1514 /*
1515 * Accessing maxBuf is racy with cifs_reconnect - need to store value
b9a74cde 1516 * and check it before using.
0013fb4c
PS
1517 */
1518 max_buf = tcon->ses->server->maxBuf;
b9a74cde 1519 if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE)))
0013fb4c
PS
1520 return -EINVAL;
1521
92a8109e
RL
1522 BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1523 PAGE_SIZE);
1524 max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1525 PAGE_SIZE);
0013fb4c
PS
1526 max_num = (max_buf - sizeof(struct smb_hdr)) /
1527 sizeof(LOCKING_ANDX_RANGE);
4b99d39b 1528 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
9ee305b7
PS
1529 if (!buf)
1530 return -ENOMEM;
1531
d46b0da7 1532 cifs_down_write(&cinode->lock_sem);
9ee305b7
PS
1533 for (i = 0; i < 2; i++) {
1534 cur = buf;
1535 num = 0;
f45d3416 1536 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
9ee305b7
PS
1537 if (flock->fl_start > li->offset ||
1538 (flock->fl_start + length) <
1539 (li->offset + li->length))
1540 continue;
1541 if (current->tgid != li->pid)
1542 continue;
9ee305b7
PS
1543 if (types[i] != li->type)
1544 continue;
ea319d57 1545 if (cinode->can_cache_brlcks) {
9ee305b7
PS
1546 /*
1547 * We can cache brlock requests - simply remove
fbd35aca 1548 * a lock from the file's list.
9ee305b7
PS
1549 */
1550 list_del(&li->llist);
1551 cifs_del_lock_waiters(li);
1552 kfree(li);
ea319d57 1553 continue;
9ee305b7 1554 }
ea319d57
PS
1555 cur->Pid = cpu_to_le16(li->pid);
1556 cur->LengthLow = cpu_to_le32((u32)li->length);
1557 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1558 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1559 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1560 /*
1561 * We need to save a lock here to let us add it again to
1562 * the file's list if the unlock range request fails on
1563 * the server.
1564 */
1565 list_move(&li->llist, &tmp_llist);
1566 if (++num == max_num) {
4b4de76e
PS
1567 stored_rc = cifs_lockv(xid, tcon,
1568 cfile->fid.netfid,
ea319d57
PS
1569 li->type, num, 0, buf);
1570 if (stored_rc) {
1571 /*
1572 * We failed on the unlock range
1573 * request - add all locks from the tmp
1574 * list to the head of the file's list.
1575 */
1576 cifs_move_llist(&tmp_llist,
f45d3416 1577 &cfile->llist->locks);
ea319d57
PS
1578 rc = stored_rc;
1579 } else
1580 /*
1581 * The unlock range request succeed -
1582 * free the tmp list.
1583 */
1584 cifs_free_llist(&tmp_llist);
1585 cur = buf;
1586 num = 0;
1587 } else
1588 cur++;
9ee305b7
PS
1589 }
1590 if (num) {
4b4de76e 1591 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
9ee305b7
PS
1592 types[i], num, 0, buf);
1593 if (stored_rc) {
f45d3416
PS
1594 cifs_move_llist(&tmp_llist,
1595 &cfile->llist->locks);
9ee305b7
PS
1596 rc = stored_rc;
1597 } else
1598 cifs_free_llist(&tmp_llist);
1599 }
1600 }
1601
1b4b55a1 1602 up_write(&cinode->lock_sem);
9ee305b7
PS
1603 kfree(buf);
1604 return rc;
1605}
1606
03776f45 1607static int
f45d3416 1608cifs_setlk(struct file *file, struct file_lock *flock, __u32 type,
6d5786a3
PS
1609 bool wait_flag, bool posix_lck, int lock, int unlock,
1610 unsigned int xid)
03776f45
PS
1611{
1612 int rc = 0;
1613 __u64 length = 1 + flock->fl_end - flock->fl_start;
1614 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1615 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
106dc538 1616 struct TCP_Server_Info *server = tcon->ses->server;
2b0143b5 1617 struct inode *inode = d_inode(cfile->dentry);
03776f45
PS
1618
1619 if (posix_lck) {
08547b03 1620 int posix_lock_type;
4f6bcec9
PS
1621
1622 rc = cifs_posix_lock_set(file, flock);
1623 if (!rc || rc < 0)
1624 return rc;
1625
106dc538 1626 if (type & server->vals->shared_lock_type)
08547b03
SF
1627 posix_lock_type = CIFS_RDLCK;
1628 else
1629 posix_lock_type = CIFS_WRLCK;
50c2f753 1630
03776f45 1631 if (unlock == 1)
beb84dc8 1632 posix_lock_type = CIFS_UNLCK;
7ee1af76 1633
f45d3416 1634 rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid,
3d22462a
JL
1635 hash_lockowner(flock->fl_owner),
1636 flock->fl_start, length,
f45d3416 1637 NULL, posix_lock_type, wait_flag);
03776f45
PS
1638 goto out;
1639 }
7ee1af76 1640
03776f45 1641 if (lock) {
161ebf9f
PS
1642 struct cifsLockInfo *lock;
1643
9645759c
RS
1644 lock = cifs_lock_init(flock->fl_start, length, type,
1645 flock->fl_flags);
161ebf9f
PS
1646 if (!lock)
1647 return -ENOMEM;
1648
fbd35aca 1649 rc = cifs_lock_add_if(cfile, lock, wait_flag);
21cb2d90 1650 if (rc < 0) {
161ebf9f 1651 kfree(lock);
21cb2d90
PS
1652 return rc;
1653 }
1654 if (!rc)
85160e03
PS
1655 goto out;
1656
63b7d3a4
PS
1657 /*
1658 * Windows 7 server can delay breaking lease from read to None
1659 * if we set a byte-range lock on a file - break it explicitly
1660 * before sending the lock to the server to be sure the next
1661 * read won't conflict with non-overlapted locks due to
1662 * pagereading.
1663 */
18cceb6a
PS
1664 if (!CIFS_CACHE_WRITE(CIFS_I(inode)) &&
1665 CIFS_CACHE_READ(CIFS_I(inode))) {
4f73c7d3 1666 cifs_zap_mapping(inode);
f96637be
JP
1667 cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n",
1668 inode);
18cceb6a 1669 CIFS_I(inode)->oplock = 0;
63b7d3a4
PS
1670 }
1671
d39a4f71
PS
1672 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1673 type, 1, 0, wait_flag);
161ebf9f
PS
1674 if (rc) {
1675 kfree(lock);
21cb2d90 1676 return rc;
03776f45 1677 }
161ebf9f 1678
fbd35aca 1679 cifs_lock_add(cfile, lock);
9ee305b7 1680 } else if (unlock)
d39a4f71 1681 rc = server->ops->mand_unlock_range(cfile, flock, xid);
03776f45 1682
03776f45 1683out:
bc31d0cd
AA
1684 if (flock->fl_flags & FL_POSIX) {
1685 /*
1686 * If this is a request to remove all locks because we
1687 * are closing the file, it doesn't matter if the
1688 * unlocking failed as both cifs.ko and the SMB server
1689 * remove the lock on file close
1690 */
1691 if (rc) {
1692 cifs_dbg(VFS, "%s failed rc=%d\n", __func__, rc);
1693 if (!(flock->fl_flags & FL_CLOSE))
1694 return rc;
1695 }
4f656367 1696 rc = locks_lock_file_wait(file, flock);
bc31d0cd 1697 }
03776f45
PS
1698 return rc;
1699}
1700
1701int cifs_lock(struct file *file, int cmd, struct file_lock *flock)
1702{
1703 int rc, xid;
1704 int lock = 0, unlock = 0;
1705 bool wait_flag = false;
1706 bool posix_lck = false;
1707 struct cifs_sb_info *cifs_sb;
1708 struct cifs_tcon *tcon;
03776f45 1709 struct cifsFileInfo *cfile;
04a6aa8a 1710 __u32 type;
03776f45
PS
1711
1712 rc = -EACCES;
6d5786a3 1713 xid = get_xid();
03776f45 1714
f96637be
JP
1715 cifs_dbg(FYI, "Lock parm: 0x%x flockflags: 0x%x flocktype: 0x%x start: %lld end: %lld\n",
1716 cmd, flock->fl_flags, flock->fl_type,
1717 flock->fl_start, flock->fl_end);
03776f45 1718
03776f45
PS
1719 cfile = (struct cifsFileInfo *)file->private_data;
1720 tcon = tlink_tcon(cfile->tlink);
106dc538
PS
1721
1722 cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag,
1723 tcon->ses->server);
7119e220 1724 cifs_sb = CIFS_FILE_SB(file);
03776f45 1725
29e20f9c 1726 if (cap_unix(tcon->ses) &&
03776f45
PS
1727 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1728 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1729 posix_lck = true;
1730 /*
1731 * BB add code here to normalize offset and length to account for
1732 * negative length which we can not accept over the wire.
1733 */
1734 if (IS_GETLK(cmd)) {
4f6bcec9 1735 rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid);
6d5786a3 1736 free_xid(xid);
03776f45
PS
1737 return rc;
1738 }
1739
1740 if (!lock && !unlock) {
1741 /*
1742 * if no lock or unlock then nothing to do since we do not
1743 * know what it is
1744 */
6d5786a3 1745 free_xid(xid);
03776f45 1746 return -EOPNOTSUPP;
7ee1af76
JA
1747 }
1748
03776f45
PS
1749 rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock,
1750 xid);
6d5786a3 1751 free_xid(xid);
1da177e4
LT
1752 return rc;
1753}
1754
597b027f
JL
1755/*
1756 * update the file size (if needed) after a write. Should be called with
1757 * the inode->i_lock held
1758 */
72432ffc 1759void
fbec9ab9
JL
1760cifs_update_eof(struct cifsInodeInfo *cifsi, loff_t offset,
1761 unsigned int bytes_written)
1762{
1763 loff_t end_of_write = offset + bytes_written;
1764
1765 if (end_of_write > cifsi->server_eof)
1766 cifsi->server_eof = end_of_write;
1767}
1768
ba9ad725
PS
1769static ssize_t
1770cifs_write(struct cifsFileInfo *open_file, __u32 pid, const char *write_data,
1771 size_t write_size, loff_t *offset)
1da177e4
LT
1772{
1773 int rc = 0;
1774 unsigned int bytes_written = 0;
1775 unsigned int total_written;
ba9ad725
PS
1776 struct cifs_tcon *tcon;
1777 struct TCP_Server_Info *server;
6d5786a3 1778 unsigned int xid;
7da4b49a 1779 struct dentry *dentry = open_file->dentry;
2b0143b5 1780 struct cifsInodeInfo *cifsi = CIFS_I(d_inode(dentry));
fa2989f4 1781 struct cifs_io_parms io_parms;
1da177e4 1782
35c265e0
AV
1783 cifs_dbg(FYI, "write %zd bytes to offset %lld of %pd\n",
1784 write_size, *offset, dentry);
1da177e4 1785
ba9ad725
PS
1786 tcon = tlink_tcon(open_file->tlink);
1787 server = tcon->ses->server;
1788
1789 if (!server->ops->sync_write)
1790 return -ENOSYS;
50c2f753 1791
6d5786a3 1792 xid = get_xid();
1da177e4 1793
1da177e4
LT
1794 for (total_written = 0; write_size > total_written;
1795 total_written += bytes_written) {
1796 rc = -EAGAIN;
1797 while (rc == -EAGAIN) {
ca83ce3d
JL
1798 struct kvec iov[2];
1799 unsigned int len;
1800
1da177e4 1801 if (open_file->invalidHandle) {
1da177e4
LT
1802 /* we could deadlock if we called
1803 filemap_fdatawait from here so tell
fb8c4b14 1804 reopen_file not to flush data to
1da177e4 1805 server now */
15886177 1806 rc = cifs_reopen_file(open_file, false);
1da177e4
LT
1807 if (rc != 0)
1808 break;
1809 }
ca83ce3d 1810
2b0143b5 1811 len = min(server->ops->wp_retry_size(d_inode(dentry)),
cb7e9eab 1812 (unsigned int)write_size - total_written);
ca83ce3d
JL
1813 /* iov[0] is reserved for smb header */
1814 iov[1].iov_base = (char *)write_data + total_written;
1815 iov[1].iov_len = len;
fa2989f4 1816 io_parms.pid = pid;
ba9ad725
PS
1817 io_parms.tcon = tcon;
1818 io_parms.offset = *offset;
fa2989f4 1819 io_parms.length = len;
db8b631d
SF
1820 rc = server->ops->sync_write(xid, &open_file->fid,
1821 &io_parms, &bytes_written, iov, 1);
1da177e4
LT
1822 }
1823 if (rc || (bytes_written == 0)) {
1824 if (total_written)
1825 break;
1826 else {
6d5786a3 1827 free_xid(xid);
1da177e4
LT
1828 return rc;
1829 }
fbec9ab9 1830 } else {
2b0143b5 1831 spin_lock(&d_inode(dentry)->i_lock);
ba9ad725 1832 cifs_update_eof(cifsi, *offset, bytes_written);
2b0143b5 1833 spin_unlock(&d_inode(dentry)->i_lock);
ba9ad725 1834 *offset += bytes_written;
fbec9ab9 1835 }
1da177e4
LT
1836 }
1837
ba9ad725 1838 cifs_stats_bytes_written(tcon, total_written);
1da177e4 1839
7da4b49a 1840 if (total_written > 0) {
2b0143b5
DH
1841 spin_lock(&d_inode(dentry)->i_lock);
1842 if (*offset > d_inode(dentry)->i_size)
1843 i_size_write(d_inode(dentry), *offset);
1844 spin_unlock(&d_inode(dentry)->i_lock);
1da177e4 1845 }
2b0143b5 1846 mark_inode_dirty_sync(d_inode(dentry));
6d5786a3 1847 free_xid(xid);
1da177e4
LT
1848 return total_written;
1849}
1850
6508d904
JL
1851struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode,
1852 bool fsuid_only)
630f3f0c
SF
1853{
1854 struct cifsFileInfo *open_file = NULL;
6508d904
JL
1855 struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
1856
1857 /* only filter by fsuid on multiuser mounts */
1858 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1859 fsuid_only = false;
630f3f0c 1860
cb248819 1861 spin_lock(&cifs_inode->open_file_lock);
630f3f0c
SF
1862 /* we could simply get the first_list_entry since write-only entries
1863 are always at the end of the list but since the first entry might
1864 have a close pending, we go through the whole list */
1865 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
fef59fd7 1866 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
6508d904 1867 continue;
2e396b83 1868 if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) {
630f3f0c
SF
1869 if (!open_file->invalidHandle) {
1870 /* found a good file */
1871 /* lock it so it will not be closed on us */
3afca265 1872 cifsFileInfo_get(open_file);
cb248819 1873 spin_unlock(&cifs_inode->open_file_lock);
630f3f0c
SF
1874 return open_file;
1875 } /* else might as well continue, and look for
1876 another, or simply have the caller reopen it
1877 again rather than trying to fix this handle */
1878 } else /* write only file */
1879 break; /* write only files are last so must be done */
1880 }
cb248819 1881 spin_unlock(&cifs_inode->open_file_lock);
630f3f0c
SF
1882 return NULL;
1883}
630f3f0c 1884
fe768d51
PS
1885/* Return -EBADF if no handle is found and general rc otherwise */
1886int
1887cifs_get_writable_file(struct cifsInodeInfo *cifs_inode, bool fsuid_only,
1888 struct cifsFileInfo **ret_file)
6148a742 1889{
2c0c2a08 1890 struct cifsFileInfo *open_file, *inv_file = NULL;
d3892294 1891 struct cifs_sb_info *cifs_sb;
2846d386 1892 bool any_available = false;
fe768d51 1893 int rc = -EBADF;
2c0c2a08 1894 unsigned int refind = 0;
6148a742 1895
fe768d51
PS
1896 *ret_file = NULL;
1897
1898 /*
1899 * Having a null inode here (because mapping->host was set to zero by
1900 * the VFS or MM) should not happen but we had reports of on oops (due
1901 * to it being zero) during stress testcases so we need to check for it
1902 */
60808233 1903
fb8c4b14 1904 if (cifs_inode == NULL) {
f96637be 1905 cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n");
60808233 1906 dump_stack();
fe768d51 1907 return rc;
60808233
SF
1908 }
1909
d3892294
JL
1910 cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
1911
6508d904
JL
1912 /* only filter by fsuid on multiuser mounts */
1913 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1914 fsuid_only = false;
1915
cb248819 1916 spin_lock(&cifs_inode->open_file_lock);
9b22b0b7 1917refind_writable:
2c0c2a08 1918 if (refind > MAX_REOPEN_ATT) {
cb248819 1919 spin_unlock(&cifs_inode->open_file_lock);
fe768d51 1920 return rc;
2c0c2a08 1921 }
6148a742 1922 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
6508d904
JL
1923 if (!any_available && open_file->pid != current->tgid)
1924 continue;
fef59fd7 1925 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
6148a742 1926 continue;
2e396b83 1927 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
9b22b0b7
SF
1928 if (!open_file->invalidHandle) {
1929 /* found a good writable file */
3afca265 1930 cifsFileInfo_get(open_file);
cb248819 1931 spin_unlock(&cifs_inode->open_file_lock);
fe768d51
PS
1932 *ret_file = open_file;
1933 return 0;
2c0c2a08
SP
1934 } else {
1935 if (!inv_file)
1936 inv_file = open_file;
9b22b0b7 1937 }
6148a742
SF
1938 }
1939 }
2846d386
JL
1940 /* couldn't find useable FH with same pid, try any available */
1941 if (!any_available) {
1942 any_available = true;
1943 goto refind_writable;
1944 }
2c0c2a08
SP
1945
1946 if (inv_file) {
1947 any_available = false;
3afca265 1948 cifsFileInfo_get(inv_file);
2c0c2a08
SP
1949 }
1950
cb248819 1951 spin_unlock(&cifs_inode->open_file_lock);
2c0c2a08
SP
1952
1953 if (inv_file) {
1954 rc = cifs_reopen_file(inv_file, false);
fe768d51
PS
1955 if (!rc) {
1956 *ret_file = inv_file;
1957 return 0;
2c0c2a08 1958 }
fe768d51 1959
487317c9 1960 spin_lock(&cifs_inode->open_file_lock);
fe768d51 1961 list_move_tail(&inv_file->flist, &cifs_inode->openFileList);
487317c9 1962 spin_unlock(&cifs_inode->open_file_lock);
fe768d51
PS
1963 cifsFileInfo_put(inv_file);
1964 ++refind;
1965 inv_file = NULL;
cb248819 1966 spin_lock(&cifs_inode->open_file_lock);
fe768d51 1967 goto refind_writable;
2c0c2a08
SP
1968 }
1969
fe768d51
PS
1970 return rc;
1971}
1972
1973struct cifsFileInfo *
1974find_writable_file(struct cifsInodeInfo *cifs_inode, bool fsuid_only)
1975{
1976 struct cifsFileInfo *cfile;
1977 int rc;
1978
1979 rc = cifs_get_writable_file(cifs_inode, fsuid_only, &cfile);
1980 if (rc)
1981 cifs_dbg(FYI, "couldn't find writable handle rc=%d", rc);
1982
1983 return cfile;
6148a742
SF
1984}
1985
8de9e86c
RS
1986int
1987cifs_get_writable_path(struct cifs_tcon *tcon, const char *name,
1988 struct cifsFileInfo **ret_file)
1989{
1990 struct list_head *tmp;
1991 struct cifsFileInfo *cfile;
1992 struct cifsInodeInfo *cinode;
1993 char *full_path;
1994
1995 *ret_file = NULL;
1996
1997 spin_lock(&tcon->open_file_lock);
1998 list_for_each(tmp, &tcon->openFileList) {
1999 cfile = list_entry(tmp, struct cifsFileInfo,
2000 tlist);
2001 full_path = build_path_from_dentry(cfile->dentry);
2002 if (full_path == NULL) {
2003 spin_unlock(&tcon->open_file_lock);
2004 return -ENOMEM;
2005 }
2006 if (strcmp(full_path, name)) {
2007 kfree(full_path);
2008 continue;
2009 }
2010
2011 kfree(full_path);
2012 cinode = CIFS_I(d_inode(cfile->dentry));
2013 spin_unlock(&tcon->open_file_lock);
2014 return cifs_get_writable_file(cinode, 0, ret_file);
2015 }
2016
2017 spin_unlock(&tcon->open_file_lock);
2018 return -ENOENT;
2019}
2020
496902dc
RS
2021int
2022cifs_get_readable_path(struct cifs_tcon *tcon, const char *name,
2023 struct cifsFileInfo **ret_file)
2024{
2025 struct list_head *tmp;
2026 struct cifsFileInfo *cfile;
2027 struct cifsInodeInfo *cinode;
2028 char *full_path;
2029
2030 *ret_file = NULL;
2031
2032 spin_lock(&tcon->open_file_lock);
2033 list_for_each(tmp, &tcon->openFileList) {
2034 cfile = list_entry(tmp, struct cifsFileInfo,
2035 tlist);
2036 full_path = build_path_from_dentry(cfile->dentry);
2037 if (full_path == NULL) {
2038 spin_unlock(&tcon->open_file_lock);
2039 return -ENOMEM;
2040 }
2041 if (strcmp(full_path, name)) {
2042 kfree(full_path);
2043 continue;
2044 }
2045
2046 kfree(full_path);
2047 cinode = CIFS_I(d_inode(cfile->dentry));
2048 spin_unlock(&tcon->open_file_lock);
2049 *ret_file = find_readable_file(cinode, 0);
2050 return *ret_file ? 0 : -ENOENT;
2051 }
2052
2053 spin_unlock(&tcon->open_file_lock);
2054 return -ENOENT;
2055}
2056
1da177e4
LT
2057static int cifs_partialpagewrite(struct page *page, unsigned from, unsigned to)
2058{
2059 struct address_space *mapping = page->mapping;
09cbfeaf 2060 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
1da177e4
LT
2061 char *write_data;
2062 int rc = -EFAULT;
2063 int bytes_written = 0;
1da177e4 2064 struct inode *inode;
6148a742 2065 struct cifsFileInfo *open_file;
1da177e4
LT
2066
2067 if (!mapping || !mapping->host)
2068 return -EFAULT;
2069
2070 inode = page->mapping->host;
1da177e4
LT
2071
2072 offset += (loff_t)from;
2073 write_data = kmap(page);
2074 write_data += from;
2075
09cbfeaf 2076 if ((to > PAGE_SIZE) || (from > to)) {
1da177e4
LT
2077 kunmap(page);
2078 return -EIO;
2079 }
2080
2081 /* racing with truncate? */
2082 if (offset > mapping->host->i_size) {
2083 kunmap(page);
2084 return 0; /* don't care */
2085 }
2086
2087 /* check to make sure that we are not extending the file */
2088 if (mapping->host->i_size - offset < (loff_t)to)
fb8c4b14 2089 to = (unsigned)(mapping->host->i_size - offset);
1da177e4 2090
fe768d51
PS
2091 rc = cifs_get_writable_file(CIFS_I(mapping->host), false, &open_file);
2092 if (!rc) {
fa2989f4
PS
2093 bytes_written = cifs_write(open_file, open_file->pid,
2094 write_data, to - from, &offset);
6ab409b5 2095 cifsFileInfo_put(open_file);
1da177e4 2096 /* Does mm or vfs already set times? */
c2050a45 2097 inode->i_atime = inode->i_mtime = current_time(inode);
bb5a9a04 2098 if ((bytes_written > 0) && (offset))
6148a742 2099 rc = 0;
bb5a9a04
SF
2100 else if (bytes_written < 0)
2101 rc = bytes_written;
fe768d51
PS
2102 else
2103 rc = -EFAULT;
6148a742 2104 } else {
fe768d51
PS
2105 cifs_dbg(FYI, "No writable handle for write page rc=%d\n", rc);
2106 if (!is_retryable_error(rc))
2107 rc = -EIO;
1da177e4
LT
2108 }
2109
2110 kunmap(page);
2111 return rc;
2112}
2113
90ac1387
PS
2114static struct cifs_writedata *
2115wdata_alloc_and_fillpages(pgoff_t tofind, struct address_space *mapping,
2116 pgoff_t end, pgoff_t *index,
2117 unsigned int *found_pages)
2118{
90ac1387
PS
2119 struct cifs_writedata *wdata;
2120
2121 wdata = cifs_writedata_alloc((unsigned int)tofind,
2122 cifs_writev_complete);
2123 if (!wdata)
2124 return NULL;
2125
9c19a9cb
JK
2126 *found_pages = find_get_pages_range_tag(mapping, index, end,
2127 PAGECACHE_TAG_DIRTY, tofind, wdata->pages);
90ac1387
PS
2128 return wdata;
2129}
2130
7e48ff82
PS
2131static unsigned int
2132wdata_prepare_pages(struct cifs_writedata *wdata, unsigned int found_pages,
2133 struct address_space *mapping,
2134 struct writeback_control *wbc,
2135 pgoff_t end, pgoff_t *index, pgoff_t *next, bool *done)
2136{
2137 unsigned int nr_pages = 0, i;
2138 struct page *page;
2139
2140 for (i = 0; i < found_pages; i++) {
2141 page = wdata->pages[i];
2142 /*
b93b0163
MW
2143 * At this point we hold neither the i_pages lock nor the
2144 * page lock: the page may be truncated or invalidated
2145 * (changing page->mapping to NULL), or even swizzled
2146 * back from swapper_space to tmpfs file mapping
7e48ff82
PS
2147 */
2148
2149 if (nr_pages == 0)
2150 lock_page(page);
2151 else if (!trylock_page(page))
2152 break;
2153
2154 if (unlikely(page->mapping != mapping)) {
2155 unlock_page(page);
2156 break;
2157 }
2158
2159 if (!wbc->range_cyclic && page->index > end) {
2160 *done = true;
2161 unlock_page(page);
2162 break;
2163 }
2164
2165 if (*next && (page->index != *next)) {
2166 /* Not next consecutive page */
2167 unlock_page(page);
2168 break;
2169 }
2170
2171 if (wbc->sync_mode != WB_SYNC_NONE)
2172 wait_on_page_writeback(page);
2173
2174 if (PageWriteback(page) ||
2175 !clear_page_dirty_for_io(page)) {
2176 unlock_page(page);
2177 break;
2178 }
2179
2180 /*
2181 * This actually clears the dirty bit in the radix tree.
2182 * See cifs_writepage() for more commentary.
2183 */
2184 set_page_writeback(page);
2185 if (page_offset(page) >= i_size_read(mapping->host)) {
2186 *done = true;
2187 unlock_page(page);
2188 end_page_writeback(page);
2189 break;
2190 }
2191
2192 wdata->pages[i] = page;
2193 *next = page->index + 1;
2194 ++nr_pages;
2195 }
2196
2197 /* reset index to refind any pages skipped */
2198 if (nr_pages == 0)
2199 *index = wdata->pages[0]->index + 1;
2200
2201 /* put any pages we aren't going to use */
2202 for (i = nr_pages; i < found_pages; i++) {
09cbfeaf 2203 put_page(wdata->pages[i]);
7e48ff82
PS
2204 wdata->pages[i] = NULL;
2205 }
2206
2207 return nr_pages;
2208}
2209
619aa48e 2210static int
c4b8f657
PS
2211wdata_send_pages(struct cifs_writedata *wdata, unsigned int nr_pages,
2212 struct address_space *mapping, struct writeback_control *wbc)
619aa48e 2213{
258f0603 2214 int rc;
c4b8f657
PS
2215 struct TCP_Server_Info *server =
2216 tlink_tcon(wdata->cfile->tlink)->ses->server;
619aa48e
PS
2217
2218 wdata->sync_mode = wbc->sync_mode;
2219 wdata->nr_pages = nr_pages;
2220 wdata->offset = page_offset(wdata->pages[0]);
09cbfeaf 2221 wdata->pagesz = PAGE_SIZE;
619aa48e
PS
2222 wdata->tailsz = min(i_size_read(mapping->host) -
2223 page_offset(wdata->pages[nr_pages - 1]),
09cbfeaf
KS
2224 (loff_t)PAGE_SIZE);
2225 wdata->bytes = ((nr_pages - 1) * PAGE_SIZE) + wdata->tailsz;
c4b8f657 2226 wdata->pid = wdata->cfile->pid;
619aa48e 2227
9a1c67e8
PS
2228 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2229 if (rc)
258f0603 2230 return rc;
9a1c67e8 2231
c4b8f657
PS
2232 if (wdata->cfile->invalidHandle)
2233 rc = -EAGAIN;
2234 else
2235 rc = server->ops->async_writev(wdata, cifs_writedata_release);
619aa48e 2236
619aa48e
PS
2237 return rc;
2238}
2239
1da177e4 2240static int cifs_writepages(struct address_space *mapping,
37c0eb46 2241 struct writeback_control *wbc)
1da177e4 2242{
c7d38dbe
PS
2243 struct inode *inode = mapping->host;
2244 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
cb7e9eab 2245 struct TCP_Server_Info *server;
c3d17b63
JL
2246 bool done = false, scanned = false, range_whole = false;
2247 pgoff_t end, index;
2248 struct cifs_writedata *wdata;
c7d38dbe 2249 struct cifsFileInfo *cfile = NULL;
37c0eb46 2250 int rc = 0;
9a66396f 2251 int saved_rc = 0;
0cb012d1 2252 unsigned int xid;
50c2f753 2253
37c0eb46 2254 /*
c3d17b63 2255 * If wsize is smaller than the page cache size, default to writing
37c0eb46
SF
2256 * one page at a time via cifs_writepage
2257 */
09cbfeaf 2258 if (cifs_sb->wsize < PAGE_SIZE)
37c0eb46
SF
2259 return generic_writepages(mapping, wbc);
2260
0cb012d1 2261 xid = get_xid();
111ebb6e 2262 if (wbc->range_cyclic) {
37c0eb46 2263 index = mapping->writeback_index; /* Start from prev offset */
111ebb6e
OH
2264 end = -1;
2265 } else {
09cbfeaf
KS
2266 index = wbc->range_start >> PAGE_SHIFT;
2267 end = wbc->range_end >> PAGE_SHIFT;
111ebb6e 2268 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX)
c3d17b63
JL
2269 range_whole = true;
2270 scanned = true;
37c0eb46 2271 }
cb7e9eab 2272 server = cifs_sb_master_tcon(cifs_sb)->ses->server;
37c0eb46 2273retry:
c3d17b63 2274 while (!done && index <= end) {
335b7b62 2275 unsigned int i, nr_pages, found_pages, wsize;
66231a47 2276 pgoff_t next = 0, tofind, saved_index = index;
335b7b62
PS
2277 struct cifs_credits credits_on_stack;
2278 struct cifs_credits *credits = &credits_on_stack;
fe768d51 2279 int get_file_rc = 0;
c3d17b63 2280
c7d38dbe
PS
2281 if (cfile)
2282 cifsFileInfo_put(cfile);
2283
fe768d51
PS
2284 rc = cifs_get_writable_file(CIFS_I(inode), false, &cfile);
2285
2286 /* in case of an error store it to return later */
2287 if (rc)
2288 get_file_rc = rc;
c7d38dbe 2289
cb7e9eab 2290 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
335b7b62 2291 &wsize, credits);
9a66396f
PS
2292 if (rc != 0) {
2293 done = true;
cb7e9eab 2294 break;
9a66396f 2295 }
c3d17b63 2296
09cbfeaf 2297 tofind = min((wsize / PAGE_SIZE) - 1, end - index) + 1;
c3d17b63 2298
90ac1387
PS
2299 wdata = wdata_alloc_and_fillpages(tofind, mapping, end, &index,
2300 &found_pages);
c3d17b63
JL
2301 if (!wdata) {
2302 rc = -ENOMEM;
9a66396f 2303 done = true;
cb7e9eab 2304 add_credits_and_wake_if(server, credits, 0);
c3d17b63
JL
2305 break;
2306 }
2307
c3d17b63
JL
2308 if (found_pages == 0) {
2309 kref_put(&wdata->refcount, cifs_writedata_release);
cb7e9eab 2310 add_credits_and_wake_if(server, credits, 0);
c3d17b63
JL
2311 break;
2312 }
2313
7e48ff82
PS
2314 nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc,
2315 end, &index, &next, &done);
37c0eb46 2316
c3d17b63
JL
2317 /* nothing to write? */
2318 if (nr_pages == 0) {
2319 kref_put(&wdata->refcount, cifs_writedata_release);
cb7e9eab 2320 add_credits_and_wake_if(server, credits, 0);
c3d17b63 2321 continue;
37c0eb46 2322 }
fbec9ab9 2323
335b7b62 2324 wdata->credits = credits_on_stack;
c7d38dbe
PS
2325 wdata->cfile = cfile;
2326 cfile = NULL;
941b853d 2327
c4b8f657 2328 if (!wdata->cfile) {
fe768d51
PS
2329 cifs_dbg(VFS, "No writable handle in writepages rc=%d\n",
2330 get_file_rc);
2331 if (is_retryable_error(get_file_rc))
2332 rc = get_file_rc;
2333 else
2334 rc = -EBADF;
c4b8f657
PS
2335 } else
2336 rc = wdata_send_pages(wdata, nr_pages, mapping, wbc);
f3983c21 2337
258f0603
PS
2338 for (i = 0; i < nr_pages; ++i)
2339 unlock_page(wdata->pages[i]);
2340
c3d17b63
JL
2341 /* send failure -- clean up the mess */
2342 if (rc != 0) {
335b7b62 2343 add_credits_and_wake_if(server, &wdata->credits, 0);
c3d17b63 2344 for (i = 0; i < nr_pages; ++i) {
9a66396f 2345 if (is_retryable_error(rc))
c3d17b63
JL
2346 redirty_page_for_writepage(wbc,
2347 wdata->pages[i]);
2348 else
2349 SetPageError(wdata->pages[i]);
2350 end_page_writeback(wdata->pages[i]);
09cbfeaf 2351 put_page(wdata->pages[i]);
37c0eb46 2352 }
9a66396f 2353 if (!is_retryable_error(rc))
941b853d 2354 mapping_set_error(mapping, rc);
c3d17b63
JL
2355 }
2356 kref_put(&wdata->refcount, cifs_writedata_release);
941b853d 2357
66231a47
PS
2358 if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN) {
2359 index = saved_index;
2360 continue;
2361 }
2362
9a66396f
PS
2363 /* Return immediately if we received a signal during writing */
2364 if (is_interrupt_error(rc)) {
2365 done = true;
2366 break;
2367 }
2368
2369 if (rc != 0 && saved_rc == 0)
2370 saved_rc = rc;
2371
c3d17b63
JL
2372 wbc->nr_to_write -= nr_pages;
2373 if (wbc->nr_to_write <= 0)
2374 done = true;
b066a48c 2375
c3d17b63 2376 index = next;
37c0eb46 2377 }
c3d17b63 2378
37c0eb46
SF
2379 if (!scanned && !done) {
2380 /*
2381 * We hit the last page and there is more work to be done: wrap
2382 * back to the start of the file
2383 */
c3d17b63 2384 scanned = true;
37c0eb46
SF
2385 index = 0;
2386 goto retry;
2387 }
c3d17b63 2388
9a66396f
PS
2389 if (saved_rc != 0)
2390 rc = saved_rc;
2391
111ebb6e 2392 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0))
37c0eb46
SF
2393 mapping->writeback_index = index;
2394
c7d38dbe
PS
2395 if (cfile)
2396 cifsFileInfo_put(cfile);
0cb012d1 2397 free_xid(xid);
1da177e4
LT
2398 return rc;
2399}
1da177e4 2400
9ad1506b
PS
2401static int
2402cifs_writepage_locked(struct page *page, struct writeback_control *wbc)
1da177e4 2403{
9ad1506b 2404 int rc;
6d5786a3 2405 unsigned int xid;
1da177e4 2406
6d5786a3 2407 xid = get_xid();
1da177e4 2408/* BB add check for wbc flags */
09cbfeaf 2409 get_page(page);
ad7a2926 2410 if (!PageUptodate(page))
f96637be 2411 cifs_dbg(FYI, "ppw - page not up to date\n");
cb876f45
LT
2412
2413 /*
2414 * Set the "writeback" flag, and clear "dirty" in the radix tree.
2415 *
2416 * A writepage() implementation always needs to do either this,
2417 * or re-dirty the page with "redirty_page_for_writepage()" in
2418 * the case of a failure.
2419 *
2420 * Just unlocking the page will cause the radix tree tag-bits
2421 * to fail to update with the state of the page correctly.
2422 */
fb8c4b14 2423 set_page_writeback(page);
9ad1506b 2424retry_write:
09cbfeaf 2425 rc = cifs_partialpagewrite(page, 0, PAGE_SIZE);
9a66396f
PS
2426 if (is_retryable_error(rc)) {
2427 if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN)
97b37f24 2428 goto retry_write;
9ad1506b 2429 redirty_page_for_writepage(wbc, page);
97b37f24 2430 } else if (rc != 0) {
9ad1506b 2431 SetPageError(page);
97b37f24
JL
2432 mapping_set_error(page->mapping, rc);
2433 } else {
9ad1506b 2434 SetPageUptodate(page);
97b37f24 2435 }
cb876f45 2436 end_page_writeback(page);
09cbfeaf 2437 put_page(page);
6d5786a3 2438 free_xid(xid);
1da177e4
LT
2439 return rc;
2440}
2441
9ad1506b
PS
2442static int cifs_writepage(struct page *page, struct writeback_control *wbc)
2443{
2444 int rc = cifs_writepage_locked(page, wbc);
2445 unlock_page(page);
2446 return rc;
2447}
2448
d9414774
NP
2449static int cifs_write_end(struct file *file, struct address_space *mapping,
2450 loff_t pos, unsigned len, unsigned copied,
2451 struct page *page, void *fsdata)
1da177e4 2452{
d9414774
NP
2453 int rc;
2454 struct inode *inode = mapping->host;
d4ffff1f
PS
2455 struct cifsFileInfo *cfile = file->private_data;
2456 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2457 __u32 pid;
2458
2459 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2460 pid = cfile->pid;
2461 else
2462 pid = current->tgid;
1da177e4 2463
f96637be 2464 cifs_dbg(FYI, "write_end for page %p from pos %lld with %d bytes\n",
b6b38f70 2465 page, pos, copied);
d9414774 2466
a98ee8c1
JL
2467 if (PageChecked(page)) {
2468 if (copied == len)
2469 SetPageUptodate(page);
2470 ClearPageChecked(page);
09cbfeaf 2471 } else if (!PageUptodate(page) && copied == PAGE_SIZE)
d9414774 2472 SetPageUptodate(page);
ad7a2926 2473
1da177e4 2474 if (!PageUptodate(page)) {
d9414774 2475 char *page_data;
09cbfeaf 2476 unsigned offset = pos & (PAGE_SIZE - 1);
6d5786a3 2477 unsigned int xid;
d9414774 2478
6d5786a3 2479 xid = get_xid();
1da177e4
LT
2480 /* this is probably better than directly calling
2481 partialpage_write since in this function the file handle is
2482 known which we might as well leverage */
2483 /* BB check if anything else missing out of ppw
2484 such as updating last write time */
2485 page_data = kmap(page);
d4ffff1f 2486 rc = cifs_write(cfile, pid, page_data + offset, copied, &pos);
d9414774 2487 /* if (rc < 0) should we set writebehind rc? */
1da177e4 2488 kunmap(page);
d9414774 2489
6d5786a3 2490 free_xid(xid);
fb8c4b14 2491 } else {
d9414774
NP
2492 rc = copied;
2493 pos += copied;
ca8aa29c 2494 set_page_dirty(page);
1da177e4
LT
2495 }
2496
d9414774
NP
2497 if (rc > 0) {
2498 spin_lock(&inode->i_lock);
2499 if (pos > inode->i_size)
2500 i_size_write(inode, pos);
2501 spin_unlock(&inode->i_lock);
2502 }
2503
2504 unlock_page(page);
09cbfeaf 2505 put_page(page);
d9414774 2506
1da177e4
LT
2507 return rc;
2508}
2509
02c24a82
JB
2510int cifs_strict_fsync(struct file *file, loff_t start, loff_t end,
2511 int datasync)
1da177e4 2512{
6d5786a3 2513 unsigned int xid;
1da177e4 2514 int rc = 0;
96daf2b0 2515 struct cifs_tcon *tcon;
1d8c4c00 2516 struct TCP_Server_Info *server;
c21dfb69 2517 struct cifsFileInfo *smbfile = file->private_data;
496ad9aa 2518 struct inode *inode = file_inode(file);
8be7e6ba 2519 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1da177e4 2520
3b49c9a1 2521 rc = file_write_and_wait_range(file, start, end);
02c24a82
JB
2522 if (rc)
2523 return rc;
02c24a82 2524
6d5786a3 2525 xid = get_xid();
1da177e4 2526
35c265e0
AV
2527 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2528 file, datasync);
50c2f753 2529
18cceb6a 2530 if (!CIFS_CACHE_READ(CIFS_I(inode))) {
4f73c7d3 2531 rc = cifs_zap_mapping(inode);
6feb9891 2532 if (rc) {
f96637be 2533 cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc);
6feb9891
PS
2534 rc = 0; /* don't care about it in fsync */
2535 }
2536 }
eb4b756b 2537
8be7e6ba 2538 tcon = tlink_tcon(smbfile->tlink);
1d8c4c00
PS
2539 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2540 server = tcon->ses->server;
2541 if (server->ops->flush)
2542 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2543 else
2544 rc = -ENOSYS;
2545 }
8be7e6ba 2546
6d5786a3 2547 free_xid(xid);
8be7e6ba
PS
2548 return rc;
2549}
2550
02c24a82 2551int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
8be7e6ba 2552{
6d5786a3 2553 unsigned int xid;
8be7e6ba 2554 int rc = 0;
96daf2b0 2555 struct cifs_tcon *tcon;
1d8c4c00 2556 struct TCP_Server_Info *server;
8be7e6ba 2557 struct cifsFileInfo *smbfile = file->private_data;
7119e220 2558 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
02c24a82 2559
3b49c9a1 2560 rc = file_write_and_wait_range(file, start, end);
02c24a82
JB
2561 if (rc)
2562 return rc;
8be7e6ba 2563
6d5786a3 2564 xid = get_xid();
8be7e6ba 2565
35c265e0
AV
2566 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2567 file, datasync);
8be7e6ba
PS
2568
2569 tcon = tlink_tcon(smbfile->tlink);
1d8c4c00
PS
2570 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2571 server = tcon->ses->server;
2572 if (server->ops->flush)
2573 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2574 else
2575 rc = -ENOSYS;
2576 }
b298f223 2577
6d5786a3 2578 free_xid(xid);
1da177e4
LT
2579 return rc;
2580}
2581
1da177e4
LT
2582/*
2583 * As file closes, flush all cached write data for this inode checking
2584 * for write behind errors.
2585 */
75e1fcc0 2586int cifs_flush(struct file *file, fl_owner_t id)
1da177e4 2587{
496ad9aa 2588 struct inode *inode = file_inode(file);
1da177e4
LT
2589 int rc = 0;
2590
eb4b756b 2591 if (file->f_mode & FMODE_WRITE)
d3f1322a 2592 rc = filemap_write_and_wait(inode->i_mapping);
50c2f753 2593
f96637be 2594 cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc);
1da177e4
LT
2595
2596 return rc;
2597}
2598
72432ffc
PS
2599static int
2600cifs_write_allocate_pages(struct page **pages, unsigned long num_pages)
2601{
2602 int rc = 0;
2603 unsigned long i;
2604
2605 for (i = 0; i < num_pages; i++) {
e94f7ba1 2606 pages[i] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
72432ffc
PS
2607 if (!pages[i]) {
2608 /*
2609 * save number of pages we have already allocated and
2610 * return with ENOMEM error
2611 */
2612 num_pages = i;
2613 rc = -ENOMEM;
e94f7ba1 2614 break;
72432ffc
PS
2615 }
2616 }
2617
e94f7ba1
JL
2618 if (rc) {
2619 for (i = 0; i < num_pages; i++)
2620 put_page(pages[i]);
2621 }
72432ffc
PS
2622 return rc;
2623}
2624
2625static inline
2626size_t get_numpages(const size_t wsize, const size_t len, size_t *cur_len)
2627{
2628 size_t num_pages;
2629 size_t clen;
2630
2631 clen = min_t(const size_t, len, wsize);
a7103b99 2632 num_pages = DIV_ROUND_UP(clen, PAGE_SIZE);
72432ffc
PS
2633
2634 if (cur_len)
2635 *cur_len = clen;
2636
2637 return num_pages;
2638}
2639
da82f7e7 2640static void
4a5c80d7 2641cifs_uncached_writedata_release(struct kref *refcount)
da82f7e7
JL
2642{
2643 int i;
4a5c80d7
SF
2644 struct cifs_writedata *wdata = container_of(refcount,
2645 struct cifs_writedata, refcount);
2646
c610c4b6 2647 kref_put(&wdata->ctx->refcount, cifs_aio_ctx_release);
4a5c80d7
SF
2648 for (i = 0; i < wdata->nr_pages; i++)
2649 put_page(wdata->pages[i]);
2650 cifs_writedata_release(refcount);
2651}
2652
c610c4b6
PS
2653static void collect_uncached_write_data(struct cifs_aio_ctx *ctx);
2654
4a5c80d7
SF
2655static void
2656cifs_uncached_writev_complete(struct work_struct *work)
2657{
da82f7e7
JL
2658 struct cifs_writedata *wdata = container_of(work,
2659 struct cifs_writedata, work);
2b0143b5 2660 struct inode *inode = d_inode(wdata->cfile->dentry);
da82f7e7
JL
2661 struct cifsInodeInfo *cifsi = CIFS_I(inode);
2662
2663 spin_lock(&inode->i_lock);
2664 cifs_update_eof(cifsi, wdata->offset, wdata->bytes);
2665 if (cifsi->server_eof > inode->i_size)
2666 i_size_write(inode, cifsi->server_eof);
2667 spin_unlock(&inode->i_lock);
2668
2669 complete(&wdata->done);
c610c4b6
PS
2670 collect_uncached_write_data(wdata->ctx);
2671 /* the below call can possibly free the last ref to aio ctx */
4a5c80d7 2672 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
da82f7e7
JL
2673}
2674
da82f7e7 2675static int
66386c08
PS
2676wdata_fill_from_iovec(struct cifs_writedata *wdata, struct iov_iter *from,
2677 size_t *len, unsigned long *num_pages)
da82f7e7 2678{
66386c08
PS
2679 size_t save_len, copied, bytes, cur_len = *len;
2680 unsigned long i, nr_pages = *num_pages;
c9de5c80 2681
66386c08
PS
2682 save_len = cur_len;
2683 for (i = 0; i < nr_pages; i++) {
2684 bytes = min_t(const size_t, cur_len, PAGE_SIZE);
2685 copied = copy_page_from_iter(wdata->pages[i], 0, bytes, from);
2686 cur_len -= copied;
2687 /*
2688 * If we didn't copy as much as we expected, then that
2689 * may mean we trod into an unmapped area. Stop copying
2690 * at that point. On the next pass through the big
2691 * loop, we'll likely end up getting a zero-length
2692 * write and bailing out of it.
2693 */
2694 if (copied < bytes)
2695 break;
2696 }
2697 cur_len = save_len - cur_len;
2698 *len = cur_len;
da82f7e7 2699
66386c08
PS
2700 /*
2701 * If we have no data to send, then that probably means that
2702 * the copy above failed altogether. That's most likely because
2703 * the address in the iovec was bogus. Return -EFAULT and let
2704 * the caller free anything we allocated and bail out.
2705 */
2706 if (!cur_len)
2707 return -EFAULT;
da82f7e7 2708
66386c08
PS
2709 /*
2710 * i + 1 now represents the number of pages we actually used in
2711 * the copy phase above.
2712 */
2713 *num_pages = i + 1;
2714 return 0;
da82f7e7
JL
2715}
2716
8c5f9c1a
LL
2717static int
2718cifs_resend_wdata(struct cifs_writedata *wdata, struct list_head *wdata_list,
2719 struct cifs_aio_ctx *ctx)
2720{
335b7b62
PS
2721 unsigned int wsize;
2722 struct cifs_credits credits;
8c5f9c1a
LL
2723 int rc;
2724 struct TCP_Server_Info *server =
2725 tlink_tcon(wdata->cfile->tlink)->ses->server;
2726
8c5f9c1a 2727 do {
d53e292f
LL
2728 if (wdata->cfile->invalidHandle) {
2729 rc = cifs_reopen_file(wdata->cfile, false);
2730 if (rc == -EAGAIN)
2731 continue;
2732 else if (rc)
2733 break;
2734 }
8c5f9c1a 2735
8c5f9c1a 2736
d53e292f
LL
2737 /*
2738 * Wait for credits to resend this wdata.
2739 * Note: we are attempting to resend the whole wdata not in
2740 * segments
2741 */
2742 do {
2743 rc = server->ops->wait_mtu_credits(server, wdata->bytes,
2744 &wsize, &credits);
2745 if (rc)
2746 goto fail;
2747
2748 if (wsize < wdata->bytes) {
2749 add_credits_and_wake_if(server, &credits, 0);
2750 msleep(1000);
2751 }
2752 } while (wsize < wdata->bytes);
2753 wdata->credits = credits;
8c5f9c1a 2754
d53e292f
LL
2755 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2756
2757 if (!rc) {
2758 if (wdata->cfile->invalidHandle)
2759 rc = -EAGAIN;
2760 else
2761 rc = server->ops->async_writev(wdata,
8c5f9c1a 2762 cifs_uncached_writedata_release);
d53e292f 2763 }
8c5f9c1a 2764
d53e292f
LL
2765 /* If the write was successfully sent, we are done */
2766 if (!rc) {
2767 list_add_tail(&wdata->list, wdata_list);
2768 return 0;
2769 }
8c5f9c1a 2770
d53e292f
LL
2771 /* Roll back credits and retry if needed */
2772 add_credits_and_wake_if(server, &wdata->credits, 0);
2773 } while (rc == -EAGAIN);
8c5f9c1a 2774
d53e292f
LL
2775fail:
2776 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
8c5f9c1a
LL
2777 return rc;
2778}
2779
43de94ea
PS
2780static int
2781cifs_write_from_iter(loff_t offset, size_t len, struct iov_iter *from,
2782 struct cifsFileInfo *open_file,
c610c4b6
PS
2783 struct cifs_sb_info *cifs_sb, struct list_head *wdata_list,
2784 struct cifs_aio_ctx *ctx)
72432ffc 2785{
43de94ea
PS
2786 int rc = 0;
2787 size_t cur_len;
66386c08 2788 unsigned long nr_pages, num_pages, i;
43de94ea 2789 struct cifs_writedata *wdata;
fc56b983 2790 struct iov_iter saved_from = *from;
6ec0b01b 2791 loff_t saved_offset = offset;
da82f7e7 2792 pid_t pid;
6ec0b01b 2793 struct TCP_Server_Info *server;
8c5f9c1a
LL
2794 struct page **pagevec;
2795 size_t start;
335b7b62 2796 unsigned int xid;
d4ffff1f
PS
2797
2798 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2799 pid = open_file->pid;
2800 else
2801 pid = current->tgid;
2802
6ec0b01b 2803 server = tlink_tcon(open_file->tlink)->ses->server;
335b7b62 2804 xid = get_xid();
6ec0b01b 2805
72432ffc 2806 do {
335b7b62
PS
2807 unsigned int wsize;
2808 struct cifs_credits credits_on_stack;
2809 struct cifs_credits *credits = &credits_on_stack;
cb7e9eab 2810
3e952994
PS
2811 if (open_file->invalidHandle) {
2812 rc = cifs_reopen_file(open_file, false);
2813 if (rc == -EAGAIN)
2814 continue;
2815 else if (rc)
2816 break;
2817 }
2818
cb7e9eab 2819 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
335b7b62 2820 &wsize, credits);
cb7e9eab
PS
2821 if (rc)
2822 break;
da82f7e7 2823
b6bc8a7b
LL
2824 cur_len = min_t(const size_t, len, wsize);
2825
8c5f9c1a 2826 if (ctx->direct_io) {
b98e26df
SF
2827 ssize_t result;
2828
2829 result = iov_iter_get_pages_alloc(
b6bc8a7b 2830 from, &pagevec, cur_len, &start);
b98e26df 2831 if (result < 0) {
8c5f9c1a
LL
2832 cifs_dbg(VFS,
2833 "direct_writev couldn't get user pages "
2834 "(rc=%zd) iter type %d iov_offset %zd "
2835 "count %zd\n",
b98e26df 2836 result, from->type,
8c5f9c1a
LL
2837 from->iov_offset, from->count);
2838 dump_stack();
54e94ff9
LL
2839
2840 rc = result;
2841 add_credits_and_wake_if(server, credits, 0);
8c5f9c1a
LL
2842 break;
2843 }
b98e26df 2844 cur_len = (size_t)result;
8c5f9c1a
LL
2845 iov_iter_advance(from, cur_len);
2846
2847 nr_pages =
2848 (cur_len + start + PAGE_SIZE - 1) / PAGE_SIZE;
2849
2850 wdata = cifs_writedata_direct_alloc(pagevec,
da82f7e7 2851 cifs_uncached_writev_complete);
8c5f9c1a
LL
2852 if (!wdata) {
2853 rc = -ENOMEM;
2854 add_credits_and_wake_if(server, credits, 0);
2855 break;
2856 }
da82f7e7 2857
da82f7e7 2858
8c5f9c1a
LL
2859 wdata->page_offset = start;
2860 wdata->tailsz =
2861 nr_pages > 1 ?
2862 cur_len - (PAGE_SIZE - start) -
2863 (nr_pages - 2) * PAGE_SIZE :
2864 cur_len;
2865 } else {
2866 nr_pages = get_numpages(wsize, len, &cur_len);
2867 wdata = cifs_writedata_alloc(nr_pages,
2868 cifs_uncached_writev_complete);
2869 if (!wdata) {
2870 rc = -ENOMEM;
2871 add_credits_and_wake_if(server, credits, 0);
2872 break;
2873 }
5d81de8e 2874
8c5f9c1a
LL
2875 rc = cifs_write_allocate_pages(wdata->pages, nr_pages);
2876 if (rc) {
9bda8723 2877 kvfree(wdata->pages);
8c5f9c1a
LL
2878 kfree(wdata);
2879 add_credits_and_wake_if(server, credits, 0);
2880 break;
2881 }
2882
2883 num_pages = nr_pages;
2884 rc = wdata_fill_from_iovec(
2885 wdata, from, &cur_len, &num_pages);
2886 if (rc) {
2887 for (i = 0; i < nr_pages; i++)
2888 put_page(wdata->pages[i]);
9bda8723 2889 kvfree(wdata->pages);
8c5f9c1a
LL
2890 kfree(wdata);
2891 add_credits_and_wake_if(server, credits, 0);
2892 break;
2893 }
2894
2895 /*
2896 * Bring nr_pages down to the number of pages we
2897 * actually used, and free any pages that we didn't use.
2898 */
2899 for ( ; nr_pages > num_pages; nr_pages--)
2900 put_page(wdata->pages[nr_pages - 1]);
2901
2902 wdata->tailsz = cur_len - ((nr_pages - 1) * PAGE_SIZE);
2903 }
5d81de8e 2904
da82f7e7
JL
2905 wdata->sync_mode = WB_SYNC_ALL;
2906 wdata->nr_pages = nr_pages;
2907 wdata->offset = (__u64)offset;
2908 wdata->cfile = cifsFileInfo_get(open_file);
2909 wdata->pid = pid;
2910 wdata->bytes = cur_len;
eddb079d 2911 wdata->pagesz = PAGE_SIZE;
335b7b62 2912 wdata->credits = credits_on_stack;
c610c4b6
PS
2913 wdata->ctx = ctx;
2914 kref_get(&ctx->refcount);
6ec0b01b 2915
9a1c67e8
PS
2916 rc = adjust_credits(server, &wdata->credits, wdata->bytes);
2917
2918 if (!rc) {
2919 if (wdata->cfile->invalidHandle)
3e952994
PS
2920 rc = -EAGAIN;
2921 else
9a1c67e8 2922 rc = server->ops->async_writev(wdata,
6ec0b01b 2923 cifs_uncached_writedata_release);
9a1c67e8
PS
2924 }
2925
da82f7e7 2926 if (rc) {
335b7b62 2927 add_credits_and_wake_if(server, &wdata->credits, 0);
4a5c80d7
SF
2928 kref_put(&wdata->refcount,
2929 cifs_uncached_writedata_release);
6ec0b01b 2930 if (rc == -EAGAIN) {
fc56b983 2931 *from = saved_from;
6ec0b01b
PS
2932 iov_iter_advance(from, offset - saved_offset);
2933 continue;
2934 }
72432ffc
PS
2935 break;
2936 }
2937
43de94ea 2938 list_add_tail(&wdata->list, wdata_list);
da82f7e7
JL
2939 offset += cur_len;
2940 len -= cur_len;
72432ffc
PS
2941 } while (len > 0);
2942
335b7b62 2943 free_xid(xid);
43de94ea
PS
2944 return rc;
2945}
2946
c610c4b6 2947static void collect_uncached_write_data(struct cifs_aio_ctx *ctx)
43de94ea 2948{
c610c4b6 2949 struct cifs_writedata *wdata, *tmp;
43de94ea
PS
2950 struct cifs_tcon *tcon;
2951 struct cifs_sb_info *cifs_sb;
c610c4b6 2952 struct dentry *dentry = ctx->cfile->dentry;
43de94ea
PS
2953 int rc;
2954
c610c4b6
PS
2955 tcon = tlink_tcon(ctx->cfile->tlink);
2956 cifs_sb = CIFS_SB(dentry->d_sb);
43de94ea 2957
c610c4b6 2958 mutex_lock(&ctx->aio_mutex);
43de94ea 2959
c610c4b6
PS
2960 if (list_empty(&ctx->list)) {
2961 mutex_unlock(&ctx->aio_mutex);
2962 return;
2963 }
da82f7e7 2964
c610c4b6 2965 rc = ctx->rc;
da82f7e7
JL
2966 /*
2967 * Wait for and collect replies for any successful sends in order of
c610c4b6
PS
2968 * increasing offset. Once an error is hit, then return without waiting
2969 * for any more replies.
da82f7e7
JL
2970 */
2971restart_loop:
c610c4b6 2972 list_for_each_entry_safe(wdata, tmp, &ctx->list, list) {
da82f7e7 2973 if (!rc) {
c610c4b6
PS
2974 if (!try_wait_for_completion(&wdata->done)) {
2975 mutex_unlock(&ctx->aio_mutex);
2976 return;
2977 }
2978
2979 if (wdata->result)
da82f7e7
JL
2980 rc = wdata->result;
2981 else
c610c4b6 2982 ctx->total_len += wdata->bytes;
da82f7e7
JL
2983
2984 /* resend call if it's a retryable error */
2985 if (rc == -EAGAIN) {
6ec0b01b 2986 struct list_head tmp_list;
c610c4b6 2987 struct iov_iter tmp_from = ctx->iter;
6ec0b01b
PS
2988
2989 INIT_LIST_HEAD(&tmp_list);
2990 list_del_init(&wdata->list);
2991
8c5f9c1a
LL
2992 if (ctx->direct_io)
2993 rc = cifs_resend_wdata(
2994 wdata, &tmp_list, ctx);
2995 else {
2996 iov_iter_advance(&tmp_from,
c610c4b6 2997 wdata->offset - ctx->pos);
6ec0b01b 2998
8c5f9c1a 2999 rc = cifs_write_from_iter(wdata->offset,
6ec0b01b 3000 wdata->bytes, &tmp_from,
c610c4b6
PS
3001 ctx->cfile, cifs_sb, &tmp_list,
3002 ctx);
d53e292f
LL
3003
3004 kref_put(&wdata->refcount,
3005 cifs_uncached_writedata_release);
8c5f9c1a 3006 }
6ec0b01b 3007
c610c4b6 3008 list_splice(&tmp_list, &ctx->list);
da82f7e7
JL
3009 goto restart_loop;
3010 }
3011 }
3012 list_del_init(&wdata->list);
4a5c80d7 3013 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
72432ffc
PS
3014 }
3015
c610c4b6
PS
3016 cifs_stats_bytes_written(tcon, ctx->total_len);
3017 set_bit(CIFS_INO_INVALID_MAPPING, &CIFS_I(dentry->d_inode)->flags);
3018
3019 ctx->rc = (rc == 0) ? ctx->total_len : rc;
3020
3021 mutex_unlock(&ctx->aio_mutex);
3022
3023 if (ctx->iocb && ctx->iocb->ki_complete)
3024 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
3025 else
3026 complete(&ctx->done);
3027}
3028
8c5f9c1a
LL
3029static ssize_t __cifs_writev(
3030 struct kiocb *iocb, struct iov_iter *from, bool direct)
c610c4b6
PS
3031{
3032 struct file *file = iocb->ki_filp;
3033 ssize_t total_written = 0;
3034 struct cifsFileInfo *cfile;
3035 struct cifs_tcon *tcon;
3036 struct cifs_sb_info *cifs_sb;
3037 struct cifs_aio_ctx *ctx;
3038 struct iov_iter saved_from = *from;
8c5f9c1a 3039 size_t len = iov_iter_count(from);
c610c4b6
PS
3040 int rc;
3041
3042 /*
8c5f9c1a
LL
3043 * iov_iter_get_pages_alloc doesn't work with ITER_KVEC.
3044 * In this case, fall back to non-direct write function.
3045 * this could be improved by getting pages directly in ITER_KVEC
c610c4b6 3046 */
8c5f9c1a
LL
3047 if (direct && from->type & ITER_KVEC) {
3048 cifs_dbg(FYI, "use non-direct cifs_writev for kvec I/O\n");
3049 direct = false;
3050 }
c610c4b6
PS
3051
3052 rc = generic_write_checks(iocb, from);
3053 if (rc <= 0)
3054 return rc;
3055
3056 cifs_sb = CIFS_FILE_SB(file);
3057 cfile = file->private_data;
3058 tcon = tlink_tcon(cfile->tlink);
3059
3060 if (!tcon->ses->server->ops->async_writev)
3061 return -ENOSYS;
3062
3063 ctx = cifs_aio_ctx_alloc();
3064 if (!ctx)
3065 return -ENOMEM;
3066
3067 ctx->cfile = cifsFileInfo_get(cfile);
3068
3069 if (!is_sync_kiocb(iocb))
3070 ctx->iocb = iocb;
3071
3072 ctx->pos = iocb->ki_pos;
3073
8c5f9c1a
LL
3074 if (direct) {
3075 ctx->direct_io = true;
3076 ctx->iter = *from;
3077 ctx->len = len;
3078 } else {
3079 rc = setup_aio_ctx_iter(ctx, from, WRITE);
3080 if (rc) {
3081 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3082 return rc;
3083 }
c610c4b6
PS
3084 }
3085
3086 /* grab a lock here due to read response handlers can access ctx */
3087 mutex_lock(&ctx->aio_mutex);
3088
3089 rc = cifs_write_from_iter(iocb->ki_pos, ctx->len, &saved_from,
3090 cfile, cifs_sb, &ctx->list, ctx);
3091
3092 /*
3093 * If at least one write was successfully sent, then discard any rc
3094 * value from the later writes. If the other write succeeds, then
3095 * we'll end up returning whatever was written. If it fails, then
3096 * we'll get a new rc value from that.
3097 */
3098 if (!list_empty(&ctx->list))
3099 rc = 0;
3100
3101 mutex_unlock(&ctx->aio_mutex);
3102
3103 if (rc) {
3104 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3105 return rc;
3106 }
3107
3108 if (!is_sync_kiocb(iocb)) {
3109 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3110 return -EIOCBQUEUED;
3111 }
3112
3113 rc = wait_for_completion_killable(&ctx->done);
3114 if (rc) {
3115 mutex_lock(&ctx->aio_mutex);
3116 ctx->rc = rc = -EINTR;
3117 total_written = ctx->total_len;
3118 mutex_unlock(&ctx->aio_mutex);
3119 } else {
3120 rc = ctx->rc;
3121 total_written = ctx->total_len;
3122 }
3123
3124 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3125
e9d1593d
AV
3126 if (unlikely(!total_written))
3127 return rc;
72432ffc 3128
e9d1593d 3129 iocb->ki_pos += total_written;
e9d1593d 3130 return total_written;
72432ffc
PS
3131}
3132
8c5f9c1a
LL
3133ssize_t cifs_direct_writev(struct kiocb *iocb, struct iov_iter *from)
3134{
3135 return __cifs_writev(iocb, from, true);
3136}
3137
3138ssize_t cifs_user_writev(struct kiocb *iocb, struct iov_iter *from)
3139{
3140 return __cifs_writev(iocb, from, false);
3141}
3142
579f9053 3143static ssize_t
3dae8750 3144cifs_writev(struct kiocb *iocb, struct iov_iter *from)
72432ffc 3145{
579f9053
PS
3146 struct file *file = iocb->ki_filp;
3147 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
3148 struct inode *inode = file->f_mapping->host;
3149 struct cifsInodeInfo *cinode = CIFS_I(inode);
3150 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
5f380c7f 3151 ssize_t rc;
72432ffc 3152
966681c9 3153 inode_lock(inode);
579f9053
PS
3154 /*
3155 * We need to hold the sem to be sure nobody modifies lock list
3156 * with a brlock that prevents writing.
3157 */
3158 down_read(&cinode->lock_sem);
5f380c7f 3159
3309dd04
AV
3160 rc = generic_write_checks(iocb, from);
3161 if (rc <= 0)
5f380c7f
AV
3162 goto out;
3163
5f380c7f 3164 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from),
9645759c
RS
3165 server->vals->exclusive_lock_type, 0,
3166 NULL, CIFS_WRITE_OP))
3dae8750 3167 rc = __generic_file_write_iter(iocb, from);
5f380c7f
AV
3168 else
3169 rc = -EACCES;
3170out:
966681c9 3171 up_read(&cinode->lock_sem);
5955102c 3172 inode_unlock(inode);
19dfc1f5 3173
e2592217
CH
3174 if (rc > 0)
3175 rc = generic_write_sync(iocb, rc);
579f9053
PS
3176 return rc;
3177}
3178
3179ssize_t
3dae8750 3180cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from)
579f9053 3181{
496ad9aa 3182 struct inode *inode = file_inode(iocb->ki_filp);
579f9053
PS
3183 struct cifsInodeInfo *cinode = CIFS_I(inode);
3184 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3185 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3186 iocb->ki_filp->private_data;
3187 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
88cf75aa 3188 ssize_t written;
ca8aa29c 3189
c11f1df5
SP
3190 written = cifs_get_writer(cinode);
3191 if (written)
3192 return written;
3193
18cceb6a 3194 if (CIFS_CACHE_WRITE(cinode)) {
88cf75aa
PS
3195 if (cap_unix(tcon->ses) &&
3196 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability))
c11f1df5 3197 && ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
3dae8750 3198 written = generic_file_write_iter(iocb, from);
c11f1df5
SP
3199 goto out;
3200 }
3dae8750 3201 written = cifs_writev(iocb, from);
c11f1df5 3202 goto out;
25078105 3203 }
25078105 3204 /*
ca8aa29c
PS
3205 * For non-oplocked files in strict cache mode we need to write the data
3206 * to the server exactly from the pos to pos+len-1 rather than flush all
3207 * affected pages because it may cause a error with mandatory locks on
3208 * these pages but not on the region from pos to ppos+len-1.
72432ffc 3209 */
3dae8750 3210 written = cifs_user_writev(iocb, from);
6dfbd846 3211 if (CIFS_CACHE_READ(cinode)) {
88cf75aa 3212 /*
6dfbd846
PS
3213 * We have read level caching and we have just sent a write
3214 * request to the server thus making data in the cache stale.
3215 * Zap the cache and set oplock/lease level to NONE to avoid
3216 * reading stale data from the cache. All subsequent read
3217 * operations will read new data from the server.
88cf75aa 3218 */
4f73c7d3 3219 cifs_zap_mapping(inode);
6dfbd846 3220 cifs_dbg(FYI, "Set Oplock/Lease to NONE for inode=%p after write\n",
f96637be 3221 inode);
18cceb6a 3222 cinode->oplock = 0;
88cf75aa 3223 }
c11f1df5
SP
3224out:
3225 cifs_put_writer(cinode);
88cf75aa 3226 return written;
72432ffc
PS
3227}
3228
0471ca3f 3229static struct cifs_readdata *
f9f5aca1 3230cifs_readdata_direct_alloc(struct page **pages, work_func_t complete)
0471ca3f
JL
3231{
3232 struct cifs_readdata *rdata;
f4e49cd2 3233
f9f5aca1 3234 rdata = kzalloc(sizeof(*rdata), GFP_KERNEL);
0471ca3f 3235 if (rdata != NULL) {
f9f5aca1 3236 rdata->pages = pages;
6993f74a 3237 kref_init(&rdata->refcount);
1c892549
JL
3238 INIT_LIST_HEAD(&rdata->list);
3239 init_completion(&rdata->done);
0471ca3f 3240 INIT_WORK(&rdata->work, complete);
0471ca3f 3241 }
f4e49cd2 3242
0471ca3f
JL
3243 return rdata;
3244}
3245
f9f5aca1
LL
3246static struct cifs_readdata *
3247cifs_readdata_alloc(unsigned int nr_pages, work_func_t complete)
3248{
3249 struct page **pages =
6396bb22 3250 kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
f9f5aca1
LL
3251 struct cifs_readdata *ret = NULL;
3252
3253 if (pages) {
3254 ret = cifs_readdata_direct_alloc(pages, complete);
3255 if (!ret)
3256 kfree(pages);
3257 }
3258
3259 return ret;
3260}
3261
6993f74a
JL
3262void
3263cifs_readdata_release(struct kref *refcount)
0471ca3f 3264{
6993f74a
JL
3265 struct cifs_readdata *rdata = container_of(refcount,
3266 struct cifs_readdata, refcount);
bd3dcc6a
LL
3267#ifdef CONFIG_CIFS_SMB_DIRECT
3268 if (rdata->mr) {
3269 smbd_deregister_mr(rdata->mr);
3270 rdata->mr = NULL;
3271 }
3272#endif
6993f74a
JL
3273 if (rdata->cfile)
3274 cifsFileInfo_put(rdata->cfile);
3275
f9f5aca1 3276 kvfree(rdata->pages);
0471ca3f
JL
3277 kfree(rdata);
3278}
3279
1c892549 3280static int
c5fab6f4 3281cifs_read_allocate_pages(struct cifs_readdata *rdata, unsigned int nr_pages)
1c892549
JL
3282{
3283 int rc = 0;
c5fab6f4 3284 struct page *page;
1c892549
JL
3285 unsigned int i;
3286
c5fab6f4 3287 for (i = 0; i < nr_pages; i++) {
1c892549
JL
3288 page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
3289 if (!page) {
3290 rc = -ENOMEM;
3291 break;
3292 }
c5fab6f4 3293 rdata->pages[i] = page;
1c892549
JL
3294 }
3295
3296 if (rc) {
31fad7d4
RBC
3297 unsigned int nr_page_failed = i;
3298
3299 for (i = 0; i < nr_page_failed; i++) {
c5fab6f4
JL
3300 put_page(rdata->pages[i]);
3301 rdata->pages[i] = NULL;
1c892549
JL
3302 }
3303 }
3304 return rc;
3305}
3306
3307static void
3308cifs_uncached_readdata_release(struct kref *refcount)
3309{
1c892549
JL
3310 struct cifs_readdata *rdata = container_of(refcount,
3311 struct cifs_readdata, refcount);
c5fab6f4 3312 unsigned int i;
1c892549 3313
6685c5e2 3314 kref_put(&rdata->ctx->refcount, cifs_aio_ctx_release);
c5fab6f4
JL
3315 for (i = 0; i < rdata->nr_pages; i++) {
3316 put_page(rdata->pages[i]);
1c892549
JL
3317 }
3318 cifs_readdata_release(refcount);
3319}
3320
1c892549
JL
3321/**
3322 * cifs_readdata_to_iov - copy data from pages in response to an iovec
3323 * @rdata: the readdata response with list of pages holding data
7f25bba8 3324 * @iter: destination for our data
1c892549
JL
3325 *
3326 * This function copies data from a list of pages in a readdata response into
3327 * an array of iovecs. It will first calculate where the data should go
3328 * based on the info in the readdata and then copy the data into that spot.
3329 */
7f25bba8
AV
3330static int
3331cifs_readdata_to_iov(struct cifs_readdata *rdata, struct iov_iter *iter)
1c892549 3332{
34a54d61 3333 size_t remaining = rdata->got_bytes;
c5fab6f4 3334 unsigned int i;
1c892549 3335
c5fab6f4 3336 for (i = 0; i < rdata->nr_pages; i++) {
c5fab6f4 3337 struct page *page = rdata->pages[i];
e686bd8d 3338 size_t copy = min_t(size_t, remaining, PAGE_SIZE);
9c25702c
PS
3339 size_t written;
3340
00e23707 3341 if (unlikely(iov_iter_is_pipe(iter))) {
9c25702c
PS
3342 void *addr = kmap_atomic(page);
3343
3344 written = copy_to_iter(addr, copy, iter);
3345 kunmap_atomic(addr);
3346 } else
3347 written = copy_page_to_iter(page, 0, copy, iter);
7f25bba8
AV
3348 remaining -= written;
3349 if (written < copy && iov_iter_count(iter) > 0)
3350 break;
1c892549 3351 }
7f25bba8 3352 return remaining ? -EFAULT : 0;
1c892549
JL
3353}
3354
6685c5e2
PS
3355static void collect_uncached_read_data(struct cifs_aio_ctx *ctx);
3356
1c892549
JL
3357static void
3358cifs_uncached_readv_complete(struct work_struct *work)
3359{
3360 struct cifs_readdata *rdata = container_of(work,
3361 struct cifs_readdata, work);
1c892549
JL
3362
3363 complete(&rdata->done);
6685c5e2
PS
3364 collect_uncached_read_data(rdata->ctx);
3365 /* the below call can possibly free the last ref to aio ctx */
1c892549
JL
3366 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
3367}
3368
3369static int
d70b9104
PS
3370uncached_fill_pages(struct TCP_Server_Info *server,
3371 struct cifs_readdata *rdata, struct iov_iter *iter,
3372 unsigned int len)
1c892549 3373{
b3160aeb 3374 int result = 0;
c5fab6f4
JL
3375 unsigned int i;
3376 unsigned int nr_pages = rdata->nr_pages;
1dbe3466 3377 unsigned int page_offset = rdata->page_offset;
1c892549 3378
b3160aeb 3379 rdata->got_bytes = 0;
8321fec4 3380 rdata->tailsz = PAGE_SIZE;
c5fab6f4
JL
3381 for (i = 0; i < nr_pages; i++) {
3382 struct page *page = rdata->pages[i];
71335664 3383 size_t n;
1dbe3466
LL
3384 unsigned int segment_size = rdata->pagesz;
3385
3386 if (i == 0)
3387 segment_size -= page_offset;
3388 else
3389 page_offset = 0;
3390
c5fab6f4 3391
71335664 3392 if (len <= 0) {
1c892549 3393 /* no need to hold page hostage */
c5fab6f4
JL
3394 rdata->pages[i] = NULL;
3395 rdata->nr_pages--;
1c892549 3396 put_page(page);
8321fec4 3397 continue;
1c892549 3398 }
1dbe3466 3399
71335664 3400 n = len;
1dbe3466 3401 if (len >= segment_size)
71335664 3402 /* enough data to fill the page */
1dbe3466
LL
3403 n = segment_size;
3404 else
71335664 3405 rdata->tailsz = len;
1dbe3466
LL
3406 len -= n;
3407
d70b9104 3408 if (iter)
1dbe3466
LL
3409 result = copy_page_from_iter(
3410 page, page_offset, n, iter);
bd3dcc6a
LL
3411#ifdef CONFIG_CIFS_SMB_DIRECT
3412 else if (rdata->mr)
3413 result = n;
3414#endif
d70b9104 3415 else
1dbe3466
LL
3416 result = cifs_read_page_from_socket(
3417 server, page, page_offset, n);
8321fec4
JL
3418 if (result < 0)
3419 break;
3420
b3160aeb 3421 rdata->got_bytes += result;
1c892549
JL
3422 }
3423
b3160aeb
PS
3424 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
3425 rdata->got_bytes : result;
1c892549
JL
3426}
3427
d70b9104
PS
3428static int
3429cifs_uncached_read_into_pages(struct TCP_Server_Info *server,
3430 struct cifs_readdata *rdata, unsigned int len)
3431{
3432 return uncached_fill_pages(server, rdata, NULL, len);
3433}
3434
3435static int
3436cifs_uncached_copy_into_pages(struct TCP_Server_Info *server,
3437 struct cifs_readdata *rdata,
3438 struct iov_iter *iter)
3439{
3440 return uncached_fill_pages(server, rdata, iter, iter->count);
3441}
3442
6e6e2b86
LL
3443static int cifs_resend_rdata(struct cifs_readdata *rdata,
3444 struct list_head *rdata_list,
3445 struct cifs_aio_ctx *ctx)
3446{
335b7b62
PS
3447 unsigned int rsize;
3448 struct cifs_credits credits;
6e6e2b86
LL
3449 int rc;
3450 struct TCP_Server_Info *server =
3451 tlink_tcon(rdata->cfile->tlink)->ses->server;
3452
6e6e2b86 3453 do {
0b0dfd59
LL
3454 if (rdata->cfile->invalidHandle) {
3455 rc = cifs_reopen_file(rdata->cfile, true);
3456 if (rc == -EAGAIN)
3457 continue;
3458 else if (rc)
3459 break;
3460 }
3461
3462 /*
3463 * Wait for credits to resend this rdata.
3464 * Note: we are attempting to resend the whole rdata not in
3465 * segments
3466 */
3467 do {
3468 rc = server->ops->wait_mtu_credits(server, rdata->bytes,
6e6e2b86
LL
3469 &rsize, &credits);
3470
0b0dfd59
LL
3471 if (rc)
3472 goto fail;
6e6e2b86 3473
0b0dfd59
LL
3474 if (rsize < rdata->bytes) {
3475 add_credits_and_wake_if(server, &credits, 0);
3476 msleep(1000);
3477 }
3478 } while (rsize < rdata->bytes);
3479 rdata->credits = credits;
6e6e2b86 3480
0b0dfd59
LL
3481 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
3482 if (!rc) {
3483 if (rdata->cfile->invalidHandle)
3484 rc = -EAGAIN;
3485 else
3486 rc = server->ops->async_readv(rdata);
3487 }
6e6e2b86 3488
0b0dfd59
LL
3489 /* If the read was successfully sent, we are done */
3490 if (!rc) {
3491 /* Add to aio pending list */
3492 list_add_tail(&rdata->list, rdata_list);
3493 return 0;
3494 }
6e6e2b86 3495
0b0dfd59
LL
3496 /* Roll back credits and retry if needed */
3497 add_credits_and_wake_if(server, &rdata->credits, 0);
3498 } while (rc == -EAGAIN);
6e6e2b86 3499
0b0dfd59
LL
3500fail:
3501 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
6e6e2b86
LL
3502 return rc;
3503}
3504
0ada36b2
PS
3505static int
3506cifs_send_async_read(loff_t offset, size_t len, struct cifsFileInfo *open_file,
6685c5e2
PS
3507 struct cifs_sb_info *cifs_sb, struct list_head *rdata_list,
3508 struct cifs_aio_ctx *ctx)
1da177e4 3509{
0ada36b2 3510 struct cifs_readdata *rdata;
335b7b62
PS
3511 unsigned int npages, rsize;
3512 struct cifs_credits credits_on_stack;
3513 struct cifs_credits *credits = &credits_on_stack;
0ada36b2
PS
3514 size_t cur_len;
3515 int rc;
1c892549 3516 pid_t pid;
25f40259 3517 struct TCP_Server_Info *server;
6e6e2b86
LL
3518 struct page **pagevec;
3519 size_t start;
3520 struct iov_iter direct_iov = ctx->iter;
a70307ee 3521
25f40259 3522 server = tlink_tcon(open_file->tlink)->ses->server;
fc9c5966 3523
d4ffff1f
PS
3524 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3525 pid = open_file->pid;
3526 else
3527 pid = current->tgid;
3528
6e6e2b86
LL
3529 if (ctx->direct_io)
3530 iov_iter_advance(&direct_iov, offset - ctx->pos);
3531
1c892549 3532 do {
3e952994
PS
3533 if (open_file->invalidHandle) {
3534 rc = cifs_reopen_file(open_file, true);
3535 if (rc == -EAGAIN)
3536 continue;
3537 else if (rc)
3538 break;
3539 }
3540
bed9da02 3541 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
335b7b62 3542 &rsize, credits);
bed9da02
PS
3543 if (rc)
3544 break;
3545
3546 cur_len = min_t(const size_t, len, rsize);
a70307ee 3547
6e6e2b86 3548 if (ctx->direct_io) {
b98e26df 3549 ssize_t result;
6e6e2b86 3550
b98e26df 3551 result = iov_iter_get_pages_alloc(
6e6e2b86
LL
3552 &direct_iov, &pagevec,
3553 cur_len, &start);
b98e26df 3554 if (result < 0) {
6e6e2b86 3555 cifs_dbg(VFS,
54e94ff9 3556 "couldn't get user pages (rc=%zd)"
6e6e2b86
LL
3557 " iter type %d"
3558 " iov_offset %zd count %zd\n",
b98e26df 3559 result, direct_iov.type,
6e6e2b86
LL
3560 direct_iov.iov_offset,
3561 direct_iov.count);
3562 dump_stack();
54e94ff9
LL
3563
3564 rc = result;
3565 add_credits_and_wake_if(server, credits, 0);
6e6e2b86
LL
3566 break;
3567 }
b98e26df 3568 cur_len = (size_t)result;
6e6e2b86
LL
3569 iov_iter_advance(&direct_iov, cur_len);
3570
3571 rdata = cifs_readdata_direct_alloc(
3572 pagevec, cifs_uncached_readv_complete);
3573 if (!rdata) {
3574 add_credits_and_wake_if(server, credits, 0);
3575 rc = -ENOMEM;
3576 break;
3577 }
3578
3579 npages = (cur_len + start + PAGE_SIZE-1) / PAGE_SIZE;
3580 rdata->page_offset = start;
3581 rdata->tailsz = npages > 1 ?
3582 cur_len-(PAGE_SIZE-start)-(npages-2)*PAGE_SIZE :
3583 cur_len;
3584
3585 } else {
3586
3587 npages = DIV_ROUND_UP(cur_len, PAGE_SIZE);
3588 /* allocate a readdata struct */
3589 rdata = cifs_readdata_alloc(npages,
1c892549 3590 cifs_uncached_readv_complete);
6e6e2b86
LL
3591 if (!rdata) {
3592 add_credits_and_wake_if(server, credits, 0);
3593 rc = -ENOMEM;
3594 break;
3595 }
a70307ee 3596
6e6e2b86 3597 rc = cifs_read_allocate_pages(rdata, npages);
9bda8723
PS
3598 if (rc) {
3599 kvfree(rdata->pages);
3600 kfree(rdata);
3601 add_credits_and_wake_if(server, credits, 0);
3602 break;
3603 }
6e6e2b86
LL
3604
3605 rdata->tailsz = PAGE_SIZE;
3606 }
1c892549
JL
3607
3608 rdata->cfile = cifsFileInfo_get(open_file);
c5fab6f4 3609 rdata->nr_pages = npages;
1c892549
JL
3610 rdata->offset = offset;
3611 rdata->bytes = cur_len;
3612 rdata->pid = pid;
8321fec4
JL
3613 rdata->pagesz = PAGE_SIZE;
3614 rdata->read_into_pages = cifs_uncached_read_into_pages;
d70b9104 3615 rdata->copy_into_pages = cifs_uncached_copy_into_pages;
335b7b62 3616 rdata->credits = credits_on_stack;
6685c5e2
PS
3617 rdata->ctx = ctx;
3618 kref_get(&ctx->refcount);
1c892549 3619
9a1c67e8
PS
3620 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
3621
3622 if (!rc) {
3623 if (rdata->cfile->invalidHandle)
3e952994
PS
3624 rc = -EAGAIN;
3625 else
9a1c67e8
PS
3626 rc = server->ops->async_readv(rdata);
3627 }
3628
1c892549 3629 if (rc) {
335b7b62 3630 add_credits_and_wake_if(server, &rdata->credits, 0);
1c892549 3631 kref_put(&rdata->refcount,
6e6e2b86
LL
3632 cifs_uncached_readdata_release);
3633 if (rc == -EAGAIN) {
3634 iov_iter_revert(&direct_iov, cur_len);
25f40259 3635 continue;
6e6e2b86 3636 }
1c892549
JL
3637 break;
3638 }
3639
0ada36b2 3640 list_add_tail(&rdata->list, rdata_list);
1c892549
JL
3641 offset += cur_len;
3642 len -= cur_len;
3643 } while (len > 0);
3644
0ada36b2
PS
3645 return rc;
3646}
3647
6685c5e2
PS
3648static void
3649collect_uncached_read_data(struct cifs_aio_ctx *ctx)
0ada36b2 3650{
6685c5e2
PS
3651 struct cifs_readdata *rdata, *tmp;
3652 struct iov_iter *to = &ctx->iter;
0ada36b2 3653 struct cifs_sb_info *cifs_sb;
6685c5e2 3654 int rc;
0ada36b2 3655
6685c5e2 3656 cifs_sb = CIFS_SB(ctx->cfile->dentry->d_sb);
0ada36b2 3657
6685c5e2 3658 mutex_lock(&ctx->aio_mutex);
0ada36b2 3659
6685c5e2
PS
3660 if (list_empty(&ctx->list)) {
3661 mutex_unlock(&ctx->aio_mutex);
3662 return;
3663 }
1c892549 3664
6685c5e2 3665 rc = ctx->rc;
1c892549 3666 /* the loop below should proceed in the order of increasing offsets */
25f40259 3667again:
6685c5e2 3668 list_for_each_entry_safe(rdata, tmp, &ctx->list, list) {
1c892549 3669 if (!rc) {
6685c5e2
PS
3670 if (!try_wait_for_completion(&rdata->done)) {
3671 mutex_unlock(&ctx->aio_mutex);
3672 return;
3673 }
3674
3675 if (rdata->result == -EAGAIN) {
74027f4a 3676 /* resend call if it's a retryable error */
fb8a3e52 3677 struct list_head tmp_list;
d913ed17 3678 unsigned int got_bytes = rdata->got_bytes;
25f40259 3679
fb8a3e52
PS
3680 list_del_init(&rdata->list);
3681 INIT_LIST_HEAD(&tmp_list);
25f40259 3682
d913ed17
PS
3683 /*
3684 * Got a part of data and then reconnect has
3685 * happened -- fill the buffer and continue
3686 * reading.
3687 */
3688 if (got_bytes && got_bytes < rdata->bytes) {
6e6e2b86
LL
3689 rc = 0;
3690 if (!ctx->direct_io)
3691 rc = cifs_readdata_to_iov(rdata, to);
d913ed17
PS
3692 if (rc) {
3693 kref_put(&rdata->refcount,
6e6e2b86 3694 cifs_uncached_readdata_release);
d913ed17
PS
3695 continue;
3696 }
74027f4a 3697 }
d913ed17 3698
6e6e2b86
LL
3699 if (ctx->direct_io) {
3700 /*
3701 * Re-use rdata as this is a
3702 * direct I/O
3703 */
3704 rc = cifs_resend_rdata(
3705 rdata,
3706 &tmp_list, ctx);
3707 } else {
3708 rc = cifs_send_async_read(
d913ed17
PS
3709 rdata->offset + got_bytes,
3710 rdata->bytes - got_bytes,
3711 rdata->cfile, cifs_sb,
6685c5e2 3712 &tmp_list, ctx);
25f40259 3713
6e6e2b86
LL
3714 kref_put(&rdata->refcount,
3715 cifs_uncached_readdata_release);
3716 }
3717
6685c5e2 3718 list_splice(&tmp_list, &ctx->list);
25f40259 3719
fb8a3e52
PS
3720 goto again;
3721 } else if (rdata->result)
3722 rc = rdata->result;
6e6e2b86 3723 else if (!ctx->direct_io)
e6a7bcb4 3724 rc = cifs_readdata_to_iov(rdata, to);
1c892549 3725
2e8a05d8
PS
3726 /* if there was a short read -- discard anything left */
3727 if (rdata->got_bytes && rdata->got_bytes < rdata->bytes)
3728 rc = -ENODATA;
6e6e2b86
LL
3729
3730 ctx->total_len += rdata->got_bytes;
1da177e4 3731 }
1c892549
JL
3732 list_del_init(&rdata->list);
3733 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
1da177e4 3734 }
a70307ee 3735
13f5938d 3736 if (!ctx->direct_io)
6e6e2b86 3737 ctx->total_len = ctx->len - iov_iter_count(to);
6685c5e2 3738
09a4707e
PS
3739 /* mask nodata case */
3740 if (rc == -ENODATA)
3741 rc = 0;
3742
6685c5e2
PS
3743 ctx->rc = (rc == 0) ? ctx->total_len : rc;
3744
3745 mutex_unlock(&ctx->aio_mutex);
3746
3747 if (ctx->iocb && ctx->iocb->ki_complete)
3748 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
3749 else
3750 complete(&ctx->done);
3751}
3752
6e6e2b86
LL
3753static ssize_t __cifs_readv(
3754 struct kiocb *iocb, struct iov_iter *to, bool direct)
6685c5e2 3755{
6685c5e2 3756 size_t len;
6e6e2b86 3757 struct file *file = iocb->ki_filp;
6685c5e2 3758 struct cifs_sb_info *cifs_sb;
6685c5e2 3759 struct cifsFileInfo *cfile;
6e6e2b86
LL
3760 struct cifs_tcon *tcon;
3761 ssize_t rc, total_read = 0;
3762 loff_t offset = iocb->ki_pos;
6685c5e2
PS
3763 struct cifs_aio_ctx *ctx;
3764
6e6e2b86
LL
3765 /*
3766 * iov_iter_get_pages_alloc() doesn't work with ITER_KVEC,
3767 * fall back to data copy read path
3768 * this could be improved by getting pages directly in ITER_KVEC
3769 */
3770 if (direct && to->type & ITER_KVEC) {
3771 cifs_dbg(FYI, "use non-direct cifs_user_readv for kvec I/O\n");
3772 direct = false;
3773 }
3774
6685c5e2
PS
3775 len = iov_iter_count(to);
3776 if (!len)
3777 return 0;
3778
3779 cifs_sb = CIFS_FILE_SB(file);
3780 cfile = file->private_data;
3781 tcon = tlink_tcon(cfile->tlink);
3782
3783 if (!tcon->ses->server->ops->async_readv)
3784 return -ENOSYS;
3785
3786 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
3787 cifs_dbg(FYI, "attempting read on write only file instance\n");
3788
3789 ctx = cifs_aio_ctx_alloc();
3790 if (!ctx)
3791 return -ENOMEM;
3792
3793 ctx->cfile = cifsFileInfo_get(cfile);
3794
3795 if (!is_sync_kiocb(iocb))
3796 ctx->iocb = iocb;
3797
00e23707 3798 if (iter_is_iovec(to))
6685c5e2
PS
3799 ctx->should_dirty = true;
3800
6e6e2b86
LL
3801 if (direct) {
3802 ctx->pos = offset;
3803 ctx->direct_io = true;
3804 ctx->iter = *to;
3805 ctx->len = len;
3806 } else {
3807 rc = setup_aio_ctx_iter(ctx, to, READ);
3808 if (rc) {
3809 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3810 return rc;
3811 }
3812 len = ctx->len;
6685c5e2
PS
3813 }
3814
6685c5e2
PS
3815 /* grab a lock here due to read response handlers can access ctx */
3816 mutex_lock(&ctx->aio_mutex);
3817
3818 rc = cifs_send_async_read(offset, len, cfile, cifs_sb, &ctx->list, ctx);
3819
3820 /* if at least one read request send succeeded, then reset rc */
3821 if (!list_empty(&ctx->list))
3822 rc = 0;
3823
3824 mutex_unlock(&ctx->aio_mutex);
3825
3826 if (rc) {
3827 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3828 return rc;
3829 }
3830
3831 if (!is_sync_kiocb(iocb)) {
3832 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3833 return -EIOCBQUEUED;
3834 }
3835
3836 rc = wait_for_completion_killable(&ctx->done);
3837 if (rc) {
3838 mutex_lock(&ctx->aio_mutex);
3839 ctx->rc = rc = -EINTR;
3840 total_read = ctx->total_len;
3841 mutex_unlock(&ctx->aio_mutex);
3842 } else {
3843 rc = ctx->rc;
3844 total_read = ctx->total_len;
3845 }
3846
3847 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3848
0165e810 3849 if (total_read) {
e6a7bcb4 3850 iocb->ki_pos += total_read;
0165e810
AV
3851 return total_read;
3852 }
3853 return rc;
a70307ee
PS
3854}
3855
6e6e2b86
LL
3856ssize_t cifs_direct_readv(struct kiocb *iocb, struct iov_iter *to)
3857{
3858 return __cifs_readv(iocb, to, true);
3859}
3860
3861ssize_t cifs_user_readv(struct kiocb *iocb, struct iov_iter *to)
3862{
3863 return __cifs_readv(iocb, to, false);
3864}
3865
579f9053 3866ssize_t
e6a7bcb4 3867cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to)
a70307ee 3868{
496ad9aa 3869 struct inode *inode = file_inode(iocb->ki_filp);
579f9053
PS
3870 struct cifsInodeInfo *cinode = CIFS_I(inode);
3871 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3872 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3873 iocb->ki_filp->private_data;
3874 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
3875 int rc = -EACCES;
a70307ee
PS
3876
3877 /*
3878 * In strict cache mode we need to read from the server all the time
3879 * if we don't have level II oplock because the server can delay mtime
3880 * change - so we can't make a decision about inode invalidating.
3881 * And we can also fail with pagereading if there are mandatory locks
3882 * on pages affected by this read but not on the region from pos to
3883 * pos+len-1.
3884 */
18cceb6a 3885 if (!CIFS_CACHE_READ(cinode))
e6a7bcb4 3886 return cifs_user_readv(iocb, to);
a70307ee 3887
579f9053
PS
3888 if (cap_unix(tcon->ses) &&
3889 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
3890 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
e6a7bcb4 3891 return generic_file_read_iter(iocb, to);
579f9053
PS
3892
3893 /*
3894 * We need to hold the sem to be sure nobody modifies lock list
3895 * with a brlock that prevents reading.
3896 */
3897 down_read(&cinode->lock_sem);
e6a7bcb4 3898 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(to),
579f9053 3899 tcon->ses->server->vals->shared_lock_type,
9645759c 3900 0, NULL, CIFS_READ_OP))
e6a7bcb4 3901 rc = generic_file_read_iter(iocb, to);
579f9053
PS
3902 up_read(&cinode->lock_sem);
3903 return rc;
a70307ee 3904}
1da177e4 3905
f9c6e234
PS
3906static ssize_t
3907cifs_read(struct file *file, char *read_data, size_t read_size, loff_t *offset)
1da177e4
LT
3908{
3909 int rc = -EACCES;
3910 unsigned int bytes_read = 0;
3911 unsigned int total_read;
3912 unsigned int current_read_size;
5eba8ab3 3913 unsigned int rsize;
1da177e4 3914 struct cifs_sb_info *cifs_sb;
29e20f9c 3915 struct cifs_tcon *tcon;
f9c6e234 3916 struct TCP_Server_Info *server;
6d5786a3 3917 unsigned int xid;
f9c6e234 3918 char *cur_offset;
1da177e4 3919 struct cifsFileInfo *open_file;
d4ffff1f 3920 struct cifs_io_parms io_parms;
ec637e3f 3921 int buf_type = CIFS_NO_BUFFER;
d4ffff1f 3922 __u32 pid;
1da177e4 3923
6d5786a3 3924 xid = get_xid();
7119e220 3925 cifs_sb = CIFS_FILE_SB(file);
1da177e4 3926
5eba8ab3
JL
3927 /* FIXME: set up handlers for larger reads and/or convert to async */
3928 rsize = min_t(unsigned int, cifs_sb->rsize, CIFSMaxBufSize);
3929
1da177e4 3930 if (file->private_data == NULL) {
0f3bc09e 3931 rc = -EBADF;
6d5786a3 3932 free_xid(xid);
0f3bc09e 3933 return rc;
1da177e4 3934 }
c21dfb69 3935 open_file = file->private_data;
29e20f9c 3936 tcon = tlink_tcon(open_file->tlink);
f9c6e234
PS
3937 server = tcon->ses->server;
3938
3939 if (!server->ops->sync_read) {
3940 free_xid(xid);
3941 return -ENOSYS;
3942 }
1da177e4 3943
d4ffff1f
PS
3944 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3945 pid = open_file->pid;
3946 else
3947 pid = current->tgid;
3948
1da177e4 3949 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
f96637be 3950 cifs_dbg(FYI, "attempting read on write only file instance\n");
1da177e4 3951
f9c6e234
PS
3952 for (total_read = 0, cur_offset = read_data; read_size > total_read;
3953 total_read += bytes_read, cur_offset += bytes_read) {
e374d90f
PS
3954 do {
3955 current_read_size = min_t(uint, read_size - total_read,
3956 rsize);
3957 /*
3958 * For windows me and 9x we do not want to request more
3959 * than it negotiated since it will refuse the read
3960 * then.
3961 */
3962 if ((tcon->ses) && !(tcon->ses->capabilities &
29e20f9c 3963 tcon->ses->server->vals->cap_large_files)) {
e374d90f
PS
3964 current_read_size = min_t(uint,
3965 current_read_size, CIFSMaxBufSize);
3966 }
cdff08e7 3967 if (open_file->invalidHandle) {
15886177 3968 rc = cifs_reopen_file(open_file, true);
1da177e4
LT
3969 if (rc != 0)
3970 break;
3971 }
d4ffff1f 3972 io_parms.pid = pid;
29e20f9c 3973 io_parms.tcon = tcon;
f9c6e234 3974 io_parms.offset = *offset;
d4ffff1f 3975 io_parms.length = current_read_size;
db8b631d 3976 rc = server->ops->sync_read(xid, &open_file->fid, &io_parms,
f9c6e234
PS
3977 &bytes_read, &cur_offset,
3978 &buf_type);
e374d90f
PS
3979 } while (rc == -EAGAIN);
3980
1da177e4
LT
3981 if (rc || (bytes_read == 0)) {
3982 if (total_read) {
3983 break;
3984 } else {
6d5786a3 3985 free_xid(xid);
1da177e4
LT
3986 return rc;
3987 }
3988 } else {
29e20f9c 3989 cifs_stats_bytes_read(tcon, total_read);
f9c6e234 3990 *offset += bytes_read;
1da177e4
LT
3991 }
3992 }
6d5786a3 3993 free_xid(xid);
1da177e4
LT
3994 return total_read;
3995}
3996
ca83ce3d
JL
3997/*
3998 * If the page is mmap'ed into a process' page tables, then we need to make
3999 * sure that it doesn't change while being written back.
4000 */
a5240cbd 4001static vm_fault_t
11bac800 4002cifs_page_mkwrite(struct vm_fault *vmf)
ca83ce3d
JL
4003{
4004 struct page *page = vmf->page;
4005
4006 lock_page(page);
4007 return VM_FAULT_LOCKED;
4008}
4009
7cbea8dc 4010static const struct vm_operations_struct cifs_file_vm_ops = {
ca83ce3d 4011 .fault = filemap_fault,
f1820361 4012 .map_pages = filemap_map_pages,
ca83ce3d
JL
4013 .page_mkwrite = cifs_page_mkwrite,
4014};
4015
7a6a19b1
PS
4016int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma)
4017{
f04a703c 4018 int xid, rc = 0;
496ad9aa 4019 struct inode *inode = file_inode(file);
7a6a19b1 4020
6d5786a3 4021 xid = get_xid();
7a6a19b1 4022
f04a703c 4023 if (!CIFS_CACHE_READ(CIFS_I(inode)))
4f73c7d3 4024 rc = cifs_zap_mapping(inode);
f04a703c
MW
4025 if (!rc)
4026 rc = generic_file_mmap(file, vma);
4027 if (!rc)
ca83ce3d 4028 vma->vm_ops = &cifs_file_vm_ops;
f04a703c 4029
6d5786a3 4030 free_xid(xid);
7a6a19b1
PS
4031 return rc;
4032}
4033
1da177e4
LT
4034int cifs_file_mmap(struct file *file, struct vm_area_struct *vma)
4035{
1da177e4
LT
4036 int rc, xid;
4037
6d5786a3 4038 xid = get_xid();
f04a703c 4039
abab095d 4040 rc = cifs_revalidate_file(file);
f04a703c 4041 if (rc)
f96637be
JP
4042 cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n",
4043 rc);
f04a703c
MW
4044 if (!rc)
4045 rc = generic_file_mmap(file, vma);
4046 if (!rc)
ca83ce3d 4047 vma->vm_ops = &cifs_file_vm_ops;
f04a703c 4048
6d5786a3 4049 free_xid(xid);
1da177e4
LT
4050 return rc;
4051}
4052
0471ca3f
JL
4053static void
4054cifs_readv_complete(struct work_struct *work)
4055{
b770ddfa 4056 unsigned int i, got_bytes;
0471ca3f
JL
4057 struct cifs_readdata *rdata = container_of(work,
4058 struct cifs_readdata, work);
0471ca3f 4059
b770ddfa 4060 got_bytes = rdata->got_bytes;
c5fab6f4
JL
4061 for (i = 0; i < rdata->nr_pages; i++) {
4062 struct page *page = rdata->pages[i];
4063
0471ca3f
JL
4064 lru_cache_add_file(page);
4065
b770ddfa
PS
4066 if (rdata->result == 0 ||
4067 (rdata->result == -EAGAIN && got_bytes)) {
0471ca3f
JL
4068 flush_dcache_page(page);
4069 SetPageUptodate(page);
4070 }
4071
4072 unlock_page(page);
4073
b770ddfa
PS
4074 if (rdata->result == 0 ||
4075 (rdata->result == -EAGAIN && got_bytes))
0471ca3f
JL
4076 cifs_readpage_to_fscache(rdata->mapping->host, page);
4077
09cbfeaf 4078 got_bytes -= min_t(unsigned int, PAGE_SIZE, got_bytes);
b770ddfa 4079
09cbfeaf 4080 put_page(page);
c5fab6f4 4081 rdata->pages[i] = NULL;
0471ca3f 4082 }
6993f74a 4083 kref_put(&rdata->refcount, cifs_readdata_release);
0471ca3f
JL
4084}
4085
8d5ce4d2 4086static int
d70b9104
PS
4087readpages_fill_pages(struct TCP_Server_Info *server,
4088 struct cifs_readdata *rdata, struct iov_iter *iter,
4089 unsigned int len)
8d5ce4d2 4090{
b3160aeb 4091 int result = 0;
c5fab6f4 4092 unsigned int i;
8d5ce4d2
JL
4093 u64 eof;
4094 pgoff_t eof_index;
c5fab6f4 4095 unsigned int nr_pages = rdata->nr_pages;
1dbe3466 4096 unsigned int page_offset = rdata->page_offset;
8d5ce4d2
JL
4097
4098 /* determine the eof that the server (probably) has */
4099 eof = CIFS_I(rdata->mapping->host)->server_eof;
09cbfeaf 4100 eof_index = eof ? (eof - 1) >> PAGE_SHIFT : 0;
f96637be 4101 cifs_dbg(FYI, "eof=%llu eof_index=%lu\n", eof, eof_index);
8d5ce4d2 4102
b3160aeb 4103 rdata->got_bytes = 0;
09cbfeaf 4104 rdata->tailsz = PAGE_SIZE;
c5fab6f4
JL
4105 for (i = 0; i < nr_pages; i++) {
4106 struct page *page = rdata->pages[i];
1dbe3466
LL
4107 unsigned int to_read = rdata->pagesz;
4108 size_t n;
4109
4110 if (i == 0)
4111 to_read -= page_offset;
4112 else
4113 page_offset = 0;
4114
4115 n = to_read;
c5fab6f4 4116
1dbe3466
LL
4117 if (len >= to_read) {
4118 len -= to_read;
8321fec4 4119 } else if (len > 0) {
8d5ce4d2 4120 /* enough for partial page, fill and zero the rest */
1dbe3466 4121 zero_user(page, len + page_offset, to_read - len);
71335664 4122 n = rdata->tailsz = len;
8321fec4 4123 len = 0;
8d5ce4d2
JL
4124 } else if (page->index > eof_index) {
4125 /*
4126 * The VFS will not try to do readahead past the
4127 * i_size, but it's possible that we have outstanding
4128 * writes with gaps in the middle and the i_size hasn't
4129 * caught up yet. Populate those with zeroed out pages
4130 * to prevent the VFS from repeatedly attempting to
4131 * fill them until the writes are flushed.
4132 */
09cbfeaf 4133 zero_user(page, 0, PAGE_SIZE);
8d5ce4d2
JL
4134 lru_cache_add_file(page);
4135 flush_dcache_page(page);
4136 SetPageUptodate(page);
4137 unlock_page(page);
09cbfeaf 4138 put_page(page);
c5fab6f4
JL
4139 rdata->pages[i] = NULL;
4140 rdata->nr_pages--;
8321fec4 4141 continue;
8d5ce4d2
JL
4142 } else {
4143 /* no need to hold page hostage */
8d5ce4d2
JL
4144 lru_cache_add_file(page);
4145 unlock_page(page);
09cbfeaf 4146 put_page(page);
c5fab6f4
JL
4147 rdata->pages[i] = NULL;
4148 rdata->nr_pages--;
8321fec4 4149 continue;
8d5ce4d2 4150 }
8321fec4 4151
d70b9104 4152 if (iter)
1dbe3466
LL
4153 result = copy_page_from_iter(
4154 page, page_offset, n, iter);
bd3dcc6a
LL
4155#ifdef CONFIG_CIFS_SMB_DIRECT
4156 else if (rdata->mr)
4157 result = n;
4158#endif
d70b9104 4159 else
1dbe3466
LL
4160 result = cifs_read_page_from_socket(
4161 server, page, page_offset, n);
8321fec4
JL
4162 if (result < 0)
4163 break;
4164
b3160aeb 4165 rdata->got_bytes += result;
8d5ce4d2
JL
4166 }
4167
b3160aeb
PS
4168 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
4169 rdata->got_bytes : result;
8d5ce4d2
JL
4170}
4171
d70b9104
PS
4172static int
4173cifs_readpages_read_into_pages(struct TCP_Server_Info *server,
4174 struct cifs_readdata *rdata, unsigned int len)
4175{
4176 return readpages_fill_pages(server, rdata, NULL, len);
4177}
4178
4179static int
4180cifs_readpages_copy_into_pages(struct TCP_Server_Info *server,
4181 struct cifs_readdata *rdata,
4182 struct iov_iter *iter)
4183{
4184 return readpages_fill_pages(server, rdata, iter, iter->count);
4185}
4186
387eb92a
PS
4187static int
4188readpages_get_pages(struct address_space *mapping, struct list_head *page_list,
4189 unsigned int rsize, struct list_head *tmplist,
4190 unsigned int *nr_pages, loff_t *offset, unsigned int *bytes)
4191{
4192 struct page *page, *tpage;
4193 unsigned int expected_index;
4194 int rc;
8a5c743e 4195 gfp_t gfp = readahead_gfp_mask(mapping);
387eb92a 4196
69cebd75
PS
4197 INIT_LIST_HEAD(tmplist);
4198
f86196ea 4199 page = lru_to_page(page_list);
387eb92a
PS
4200
4201 /*
4202 * Lock the page and put it in the cache. Since no one else
4203 * should have access to this page, we're safe to simply set
4204 * PG_locked without checking it first.
4205 */
48c935ad 4206 __SetPageLocked(page);
387eb92a 4207 rc = add_to_page_cache_locked(page, mapping,
063d99b4 4208 page->index, gfp);
387eb92a
PS
4209
4210 /* give up if we can't stick it in the cache */
4211 if (rc) {
48c935ad 4212 __ClearPageLocked(page);
387eb92a
PS
4213 return rc;
4214 }
4215
4216 /* move first page to the tmplist */
09cbfeaf
KS
4217 *offset = (loff_t)page->index << PAGE_SHIFT;
4218 *bytes = PAGE_SIZE;
387eb92a
PS
4219 *nr_pages = 1;
4220 list_move_tail(&page->lru, tmplist);
4221
4222 /* now try and add more pages onto the request */
4223 expected_index = page->index + 1;
4224 list_for_each_entry_safe_reverse(page, tpage, page_list, lru) {
4225 /* discontinuity ? */
4226 if (page->index != expected_index)
4227 break;
4228
4229 /* would this page push the read over the rsize? */
09cbfeaf 4230 if (*bytes + PAGE_SIZE > rsize)
387eb92a
PS
4231 break;
4232
48c935ad 4233 __SetPageLocked(page);
063d99b4 4234 if (add_to_page_cache_locked(page, mapping, page->index, gfp)) {
48c935ad 4235 __ClearPageLocked(page);
387eb92a
PS
4236 break;
4237 }
4238 list_move_tail(&page->lru, tmplist);
09cbfeaf 4239 (*bytes) += PAGE_SIZE;
387eb92a
PS
4240 expected_index++;
4241 (*nr_pages)++;
4242 }
4243 return rc;
8d5ce4d2
JL
4244}
4245
1da177e4
LT
4246static int cifs_readpages(struct file *file, struct address_space *mapping,
4247 struct list_head *page_list, unsigned num_pages)
4248{
690c5e31
JL
4249 int rc;
4250 struct list_head tmplist;
4251 struct cifsFileInfo *open_file = file->private_data;
7119e220 4252 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
69cebd75 4253 struct TCP_Server_Info *server;
690c5e31 4254 pid_t pid;
0cb012d1 4255 unsigned int xid;
1da177e4 4256
0cb012d1 4257 xid = get_xid();
56698236
SJ
4258 /*
4259 * Reads as many pages as possible from fscache. Returns -ENOBUFS
4260 * immediately if the cookie is negative
54afa990
DH
4261 *
4262 * After this point, every page in the list might have PG_fscache set,
4263 * so we will need to clean that up off of every page we don't use.
56698236
SJ
4264 */
4265 rc = cifs_readpages_from_fscache(mapping->host, mapping, page_list,
4266 &num_pages);
0cb012d1
SF
4267 if (rc == 0) {
4268 free_xid(xid);
690c5e31 4269 return rc;
0cb012d1 4270 }
56698236 4271
d4ffff1f
PS
4272 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
4273 pid = open_file->pid;
4274 else
4275 pid = current->tgid;
4276
690c5e31 4277 rc = 0;
69cebd75 4278 server = tlink_tcon(open_file->tlink)->ses->server;
1da177e4 4279
f96637be
JP
4280 cifs_dbg(FYI, "%s: file=%p mapping=%p num_pages=%u\n",
4281 __func__, file, mapping, num_pages);
690c5e31
JL
4282
4283 /*
4284 * Start with the page at end of list and move it to private
4285 * list. Do the same with any following pages until we hit
4286 * the rsize limit, hit an index discontinuity, or run out of
4287 * pages. Issue the async read and then start the loop again
4288 * until the list is empty.
4289 *
4290 * Note that list order is important. The page_list is in
4291 * the order of declining indexes. When we put the pages in
4292 * the rdata->pages, then we want them in increasing order.
4293 */
4294 while (!list_empty(page_list)) {
bed9da02 4295 unsigned int i, nr_pages, bytes, rsize;
690c5e31
JL
4296 loff_t offset;
4297 struct page *page, *tpage;
4298 struct cifs_readdata *rdata;
335b7b62
PS
4299 struct cifs_credits credits_on_stack;
4300 struct cifs_credits *credits = &credits_on_stack;
1da177e4 4301
3e952994
PS
4302 if (open_file->invalidHandle) {
4303 rc = cifs_reopen_file(open_file, true);
4304 if (rc == -EAGAIN)
4305 continue;
4306 else if (rc)
4307 break;
4308 }
4309
bed9da02 4310 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
335b7b62 4311 &rsize, credits);
bed9da02
PS
4312 if (rc)
4313 break;
690c5e31
JL
4314
4315 /*
69cebd75
PS
4316 * Give up immediately if rsize is too small to read an entire
4317 * page. The VFS will fall back to readpage. We should never
4318 * reach this point however since we set ra_pages to 0 when the
4319 * rsize is smaller than a cache page.
690c5e31 4320 */
09cbfeaf 4321 if (unlikely(rsize < PAGE_SIZE)) {
bed9da02 4322 add_credits_and_wake_if(server, credits, 0);
0cb012d1 4323 free_xid(xid);
69cebd75 4324 return 0;
bed9da02 4325 }
690c5e31 4326
bed9da02
PS
4327 rc = readpages_get_pages(mapping, page_list, rsize, &tmplist,
4328 &nr_pages, &offset, &bytes);
690c5e31 4329 if (rc) {
bed9da02 4330 add_credits_and_wake_if(server, credits, 0);
690c5e31
JL
4331 break;
4332 }
4333
0471ca3f 4334 rdata = cifs_readdata_alloc(nr_pages, cifs_readv_complete);
690c5e31
JL
4335 if (!rdata) {
4336 /* best to give up if we're out of mem */
4337 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
4338 list_del(&page->lru);
4339 lru_cache_add_file(page);
4340 unlock_page(page);
09cbfeaf 4341 put_page(page);
690c5e31
JL
4342 }
4343 rc = -ENOMEM;
bed9da02 4344 add_credits_and_wake_if(server, credits, 0);
690c5e31
JL
4345 break;
4346 }
4347
6993f74a 4348 rdata->cfile = cifsFileInfo_get(open_file);
690c5e31
JL
4349 rdata->mapping = mapping;
4350 rdata->offset = offset;
4351 rdata->bytes = bytes;
4352 rdata->pid = pid;
09cbfeaf 4353 rdata->pagesz = PAGE_SIZE;
1dbe3466 4354 rdata->tailsz = PAGE_SIZE;
8321fec4 4355 rdata->read_into_pages = cifs_readpages_read_into_pages;
d70b9104 4356 rdata->copy_into_pages = cifs_readpages_copy_into_pages;
335b7b62 4357 rdata->credits = credits_on_stack;
c5fab6f4
JL
4358
4359 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
4360 list_del(&page->lru);
4361 rdata->pages[rdata->nr_pages++] = page;
4362 }
690c5e31 4363
9a1c67e8
PS
4364 rc = adjust_credits(server, &rdata->credits, rdata->bytes);
4365
4366 if (!rc) {
4367 if (rdata->cfile->invalidHandle)
3e952994
PS
4368 rc = -EAGAIN;
4369 else
9a1c67e8
PS
4370 rc = server->ops->async_readv(rdata);
4371 }
4372
69cebd75 4373 if (rc) {
335b7b62 4374 add_credits_and_wake_if(server, &rdata->credits, 0);
c5fab6f4
JL
4375 for (i = 0; i < rdata->nr_pages; i++) {
4376 page = rdata->pages[i];
690c5e31
JL
4377 lru_cache_add_file(page);
4378 unlock_page(page);
09cbfeaf 4379 put_page(page);
1da177e4 4380 }
1209bbdf 4381 /* Fallback to the readpage in error/reconnect cases */
6993f74a 4382 kref_put(&rdata->refcount, cifs_readdata_release);
1da177e4
LT
4383 break;
4384 }
6993f74a
JL
4385
4386 kref_put(&rdata->refcount, cifs_readdata_release);
1da177e4
LT
4387 }
4388
54afa990
DH
4389 /* Any pages that have been shown to fscache but didn't get added to
4390 * the pagecache must be uncached before they get returned to the
4391 * allocator.
4392 */
4393 cifs_fscache_readpages_cancel(mapping->host, page_list);
0cb012d1 4394 free_xid(xid);
1da177e4
LT
4395 return rc;
4396}
4397
a9e9b7bc
SP
4398/*
4399 * cifs_readpage_worker must be called with the page pinned
4400 */
1da177e4
LT
4401static int cifs_readpage_worker(struct file *file, struct page *page,
4402 loff_t *poffset)
4403{
4404 char *read_data;
4405 int rc;
4406
56698236 4407 /* Is the page cached? */
496ad9aa 4408 rc = cifs_readpage_from_fscache(file_inode(file), page);
56698236
SJ
4409 if (rc == 0)
4410 goto read_complete;
4411
1da177e4
LT
4412 read_data = kmap(page);
4413 /* for reads over a certain size could initiate async read ahead */
fb8c4b14 4414
09cbfeaf 4415 rc = cifs_read(file, read_data, PAGE_SIZE, poffset);
fb8c4b14 4416
1da177e4
LT
4417 if (rc < 0)
4418 goto io_error;
4419 else
f96637be 4420 cifs_dbg(FYI, "Bytes read %d\n", rc);
fb8c4b14 4421
9b9c5bea
SF
4422 /* we do not want atime to be less than mtime, it broke some apps */
4423 file_inode(file)->i_atime = current_time(file_inode(file));
4424 if (timespec64_compare(&(file_inode(file)->i_atime), &(file_inode(file)->i_mtime)))
4425 file_inode(file)->i_atime = file_inode(file)->i_mtime;
4426 else
4427 file_inode(file)->i_atime = current_time(file_inode(file));
fb8c4b14 4428
09cbfeaf
KS
4429 if (PAGE_SIZE > rc)
4430 memset(read_data + rc, 0, PAGE_SIZE - rc);
1da177e4
LT
4431
4432 flush_dcache_page(page);
4433 SetPageUptodate(page);
9dc06558
SJ
4434
4435 /* send this page to the cache */
496ad9aa 4436 cifs_readpage_to_fscache(file_inode(file), page);
9dc06558 4437
1da177e4 4438 rc = 0;
fb8c4b14 4439
1da177e4 4440io_error:
fb8c4b14 4441 kunmap(page);
466bd31b 4442 unlock_page(page);
56698236
SJ
4443
4444read_complete:
1da177e4
LT
4445 return rc;
4446}
4447
4448static int cifs_readpage(struct file *file, struct page *page)
4449{
09cbfeaf 4450 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
1da177e4 4451 int rc = -EACCES;
6d5786a3 4452 unsigned int xid;
1da177e4 4453
6d5786a3 4454 xid = get_xid();
1da177e4
LT
4455
4456 if (file->private_data == NULL) {
0f3bc09e 4457 rc = -EBADF;
6d5786a3 4458 free_xid(xid);
0f3bc09e 4459 return rc;
1da177e4
LT
4460 }
4461
f96637be 4462 cifs_dbg(FYI, "readpage %p at offset %d 0x%x\n",
b6b38f70 4463 page, (int)offset, (int)offset);
1da177e4
LT
4464
4465 rc = cifs_readpage_worker(file, page, &offset);
4466
6d5786a3 4467 free_xid(xid);
1da177e4
LT
4468 return rc;
4469}
4470
a403a0a3
SF
4471static int is_inode_writable(struct cifsInodeInfo *cifs_inode)
4472{
4473 struct cifsFileInfo *open_file;
4474
cb248819 4475 spin_lock(&cifs_inode->open_file_lock);
a403a0a3 4476 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2e396b83 4477 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
cb248819 4478 spin_unlock(&cifs_inode->open_file_lock);
a403a0a3
SF
4479 return 1;
4480 }
4481 }
cb248819 4482 spin_unlock(&cifs_inode->open_file_lock);
a403a0a3
SF
4483 return 0;
4484}
4485
1da177e4
LT
4486/* We do not want to update the file size from server for inodes
4487 open for write - to avoid races with writepage extending
4488 the file - in the future we could consider allowing
fb8c4b14 4489 refreshing the inode only on increases in the file size
1da177e4
LT
4490 but this is tricky to do without racing with writebehind
4491 page caching in the current Linux kernel design */
4b18f2a9 4492bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file)
1da177e4 4493{
a403a0a3 4494 if (!cifsInode)
4b18f2a9 4495 return true;
50c2f753 4496
a403a0a3
SF
4497 if (is_inode_writable(cifsInode)) {
4498 /* This inode is open for write at least once */
c32a0b68
SF
4499 struct cifs_sb_info *cifs_sb;
4500
c32a0b68 4501 cifs_sb = CIFS_SB(cifsInode->vfs_inode.i_sb);
ad7a2926 4502 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) {
fb8c4b14 4503 /* since no page cache to corrupt on directio
c32a0b68 4504 we can change size safely */
4b18f2a9 4505 return true;
c32a0b68
SF
4506 }
4507
fb8c4b14 4508 if (i_size_read(&cifsInode->vfs_inode) < end_of_file)
4b18f2a9 4509 return true;
7ba52631 4510
4b18f2a9 4511 return false;
23e7dd7d 4512 } else
4b18f2a9 4513 return true;
1da177e4
LT
4514}
4515
d9414774
NP
4516static int cifs_write_begin(struct file *file, struct address_space *mapping,
4517 loff_t pos, unsigned len, unsigned flags,
4518 struct page **pagep, void **fsdata)
1da177e4 4519{
466bd31b 4520 int oncethru = 0;
09cbfeaf
KS
4521 pgoff_t index = pos >> PAGE_SHIFT;
4522 loff_t offset = pos & (PAGE_SIZE - 1);
a98ee8c1
JL
4523 loff_t page_start = pos & PAGE_MASK;
4524 loff_t i_size;
4525 struct page *page;
4526 int rc = 0;
d9414774 4527
f96637be 4528 cifs_dbg(FYI, "write_begin from %lld len %d\n", (long long)pos, len);
d9414774 4529
466bd31b 4530start:
54566b2c 4531 page = grab_cache_page_write_begin(mapping, index, flags);
a98ee8c1
JL
4532 if (!page) {
4533 rc = -ENOMEM;
4534 goto out;
4535 }
8a236264 4536
a98ee8c1
JL
4537 if (PageUptodate(page))
4538 goto out;
8a236264 4539
a98ee8c1
JL
4540 /*
4541 * If we write a full page it will be up to date, no need to read from
4542 * the server. If the write is short, we'll end up doing a sync write
4543 * instead.
4544 */
09cbfeaf 4545 if (len == PAGE_SIZE)
a98ee8c1 4546 goto out;
8a236264 4547
a98ee8c1
JL
4548 /*
4549 * optimize away the read when we have an oplock, and we're not
4550 * expecting to use any of the data we'd be reading in. That
4551 * is, when the page lies beyond the EOF, or straddles the EOF
4552 * and the write will cover all of the existing data.
4553 */
18cceb6a 4554 if (CIFS_CACHE_READ(CIFS_I(mapping->host))) {
a98ee8c1
JL
4555 i_size = i_size_read(mapping->host);
4556 if (page_start >= i_size ||
4557 (offset == 0 && (pos + len) >= i_size)) {
4558 zero_user_segments(page, 0, offset,
4559 offset + len,
09cbfeaf 4560 PAGE_SIZE);
a98ee8c1
JL
4561 /*
4562 * PageChecked means that the parts of the page
4563 * to which we're not writing are considered up
4564 * to date. Once the data is copied to the
4565 * page, it can be set uptodate.
4566 */
4567 SetPageChecked(page);
4568 goto out;
4569 }
4570 }
d9414774 4571
466bd31b 4572 if ((file->f_flags & O_ACCMODE) != O_WRONLY && !oncethru) {
a98ee8c1
JL
4573 /*
4574 * might as well read a page, it is fast enough. If we get
4575 * an error, we don't need to return it. cifs_write_end will
4576 * do a sync write instead since PG_uptodate isn't set.
4577 */
4578 cifs_readpage_worker(file, page, &page_start);
09cbfeaf 4579 put_page(page);
466bd31b
SP
4580 oncethru = 1;
4581 goto start;
8a236264
SF
4582 } else {
4583 /* we could try using another file handle if there is one -
4584 but how would we lock it to prevent close of that handle
4585 racing with this read? In any case
d9414774 4586 this will be written out by write_end so is fine */
1da177e4 4587 }
a98ee8c1
JL
4588out:
4589 *pagep = page;
4590 return rc;
1da177e4
LT
4591}
4592
85f2d6b4
SJ
4593static int cifs_release_page(struct page *page, gfp_t gfp)
4594{
4595 if (PagePrivate(page))
4596 return 0;
4597
4598 return cifs_fscache_release_page(page, gfp);
4599}
4600
d47992f8
LC
4601static void cifs_invalidate_page(struct page *page, unsigned int offset,
4602 unsigned int length)
85f2d6b4
SJ
4603{
4604 struct cifsInodeInfo *cifsi = CIFS_I(page->mapping->host);
4605
09cbfeaf 4606 if (offset == 0 && length == PAGE_SIZE)
85f2d6b4
SJ
4607 cifs_fscache_invalidate_page(page, &cifsi->vfs_inode);
4608}
4609
9ad1506b
PS
4610static int cifs_launder_page(struct page *page)
4611{
4612 int rc = 0;
4613 loff_t range_start = page_offset(page);
09cbfeaf 4614 loff_t range_end = range_start + (loff_t)(PAGE_SIZE - 1);
9ad1506b
PS
4615 struct writeback_control wbc = {
4616 .sync_mode = WB_SYNC_ALL,
4617 .nr_to_write = 0,
4618 .range_start = range_start,
4619 .range_end = range_end,
4620 };
4621
f96637be 4622 cifs_dbg(FYI, "Launder page: %p\n", page);
9ad1506b
PS
4623
4624 if (clear_page_dirty_for_io(page))
4625 rc = cifs_writepage_locked(page, &wbc);
4626
4627 cifs_fscache_invalidate_page(page, page->mapping->host);
4628 return rc;
4629}
4630
9b646972 4631void cifs_oplock_break(struct work_struct *work)
3bc303c2
JL
4632{
4633 struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo,
4634 oplock_break);
2b0143b5 4635 struct inode *inode = d_inode(cfile->dentry);
3bc303c2 4636 struct cifsInodeInfo *cinode = CIFS_I(inode);
95a3f2f3 4637 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
c11f1df5 4638 struct TCP_Server_Info *server = tcon->ses->server;
eb4b756b 4639 int rc = 0;
3bc303c2 4640
c11f1df5 4641 wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS,
74316201 4642 TASK_UNINTERRUPTIBLE);
c11f1df5
SP
4643
4644 server->ops->downgrade_oplock(server, cinode,
4645 test_bit(CIFS_INODE_DOWNGRADE_OPLOCK_TO_L2, &cinode->flags));
4646
18cceb6a 4647 if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) &&
63b7d3a4 4648 cifs_has_mand_locks(cinode)) {
f96637be
JP
4649 cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n",
4650 inode);
18cceb6a 4651 cinode->oplock = 0;
63b7d3a4
PS
4652 }
4653
3bc303c2 4654 if (inode && S_ISREG(inode->i_mode)) {
18cceb6a 4655 if (CIFS_CACHE_READ(cinode))
8737c930 4656 break_lease(inode, O_RDONLY);
d54ff732 4657 else
8737c930 4658 break_lease(inode, O_WRONLY);
3bc303c2 4659 rc = filemap_fdatawrite(inode->i_mapping);
18cceb6a 4660 if (!CIFS_CACHE_READ(cinode)) {
eb4b756b
JL
4661 rc = filemap_fdatawait(inode->i_mapping);
4662 mapping_set_error(inode->i_mapping, rc);
4f73c7d3 4663 cifs_zap_mapping(inode);
3bc303c2 4664 }
f96637be 4665 cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc);
3bc303c2
JL
4666 }
4667
85160e03
PS
4668 rc = cifs_push_locks(cfile);
4669 if (rc)
f96637be 4670 cifs_dbg(VFS, "Push locks rc = %d\n", rc);
85160e03 4671
3bc303c2
JL
4672 /*
4673 * releasing stale oplock after recent reconnect of smb session using
4674 * a now incorrect file handle is not a data integrity issue but do
4675 * not bother sending an oplock release if session to server still is
4676 * disconnected since oplock already released by the server
4677 */
cdff08e7 4678 if (!cfile->oplock_break_cancelled) {
95a3f2f3
PS
4679 rc = tcon->ses->server->ops->oplock_response(tcon, &cfile->fid,
4680 cinode);
f96637be 4681 cifs_dbg(FYI, "Oplock release rc = %d\n", rc);
3bc303c2 4682 }
b98749ca 4683 _cifsFileInfo_put(cfile, false /* do not wait for ourself */);
c11f1df5 4684 cifs_done_oplock_break(cinode);
3bc303c2
JL
4685}
4686
dca69288
SF
4687/*
4688 * The presence of cifs_direct_io() in the address space ops vector
4689 * allowes open() O_DIRECT flags which would have failed otherwise.
4690 *
4691 * In the non-cached mode (mount with cache=none), we shunt off direct read and write requests
4692 * so this method should never be called.
4693 *
4694 * Direct IO is not yet supported in the cached mode.
4695 */
4696static ssize_t
c8b8e32d 4697cifs_direct_io(struct kiocb *iocb, struct iov_iter *iter)
dca69288
SF
4698{
4699 /*
4700 * FIXME
4701 * Eventually need to support direct IO for non forcedirectio mounts
4702 */
4703 return -EINVAL;
4704}
4705
4706
f5e54d6e 4707const struct address_space_operations cifs_addr_ops = {
1da177e4
LT
4708 .readpage = cifs_readpage,
4709 .readpages = cifs_readpages,
4710 .writepage = cifs_writepage,
37c0eb46 4711 .writepages = cifs_writepages,
d9414774
NP
4712 .write_begin = cifs_write_begin,
4713 .write_end = cifs_write_end,
1da177e4 4714 .set_page_dirty = __set_page_dirty_nobuffers,
85f2d6b4 4715 .releasepage = cifs_release_page,
dca69288 4716 .direct_IO = cifs_direct_io,
85f2d6b4 4717 .invalidatepage = cifs_invalidate_page,
9ad1506b 4718 .launder_page = cifs_launder_page,
1da177e4 4719};
273d81d6
DK
4720
4721/*
4722 * cifs_readpages requires the server to support a buffer large enough to
4723 * contain the header plus one complete page of data. Otherwise, we need
4724 * to leave cifs_readpages out of the address space operations.
4725 */
f5e54d6e 4726const struct address_space_operations cifs_addr_ops_smallbuf = {
273d81d6
DK
4727 .readpage = cifs_readpage,
4728 .writepage = cifs_writepage,
4729 .writepages = cifs_writepages,
d9414774
NP
4730 .write_begin = cifs_write_begin,
4731 .write_end = cifs_write_end,
273d81d6 4732 .set_page_dirty = __set_page_dirty_nobuffers,
85f2d6b4
SJ
4733 .releasepage = cifs_release_page,
4734 .invalidatepage = cifs_invalidate_page,
9ad1506b 4735 .launder_page = cifs_launder_page,
273d81d6 4736};