treewide: kzalloc() -> kcalloc()
[linux-2.6-block.git] / fs / cifs / file.c
CommitLineData
1da177e4
LT
1/*
2 * fs/cifs/file.c
3 *
4 * vfs operations that deal with files
fb8c4b14 5 *
f19159dc 6 * Copyright (C) International Business Machines Corp., 2002,2010
1da177e4 7 * Author(s): Steve French (sfrench@us.ibm.com)
7ee1af76 8 * Jeremy Allison (jra@samba.org)
1da177e4
LT
9 *
10 * This library is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU Lesser General Public License as published
12 * by the Free Software Foundation; either version 2.1 of the License, or
13 * (at your option) any later version.
14 *
15 * This library is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
18 * the GNU Lesser General Public License for more details.
19 *
20 * You should have received a copy of the GNU Lesser General Public License
21 * along with this library; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 */
24#include <linux/fs.h>
37c0eb46 25#include <linux/backing-dev.h>
1da177e4
LT
26#include <linux/stat.h>
27#include <linux/fcntl.h>
28#include <linux/pagemap.h>
29#include <linux/pagevec.h>
37c0eb46 30#include <linux/writeback.h>
6f88cc2e 31#include <linux/task_io_accounting_ops.h>
23e7dd7d 32#include <linux/delay.h>
3bc303c2 33#include <linux/mount.h>
5a0e3ad6 34#include <linux/slab.h>
690c5e31 35#include <linux/swap.h>
1da177e4
LT
36#include <asm/div64.h>
37#include "cifsfs.h"
38#include "cifspdu.h"
39#include "cifsglob.h"
40#include "cifsproto.h"
41#include "cifs_unicode.h"
42#include "cifs_debug.h"
43#include "cifs_fs_sb.h"
9451a9a5 44#include "fscache.h"
bd3dcc6a 45#include "smbdirect.h"
07b92d0d 46
1da177e4
LT
47static inline int cifs_convert_flags(unsigned int flags)
48{
49 if ((flags & O_ACCMODE) == O_RDONLY)
50 return GENERIC_READ;
51 else if ((flags & O_ACCMODE) == O_WRONLY)
52 return GENERIC_WRITE;
53 else if ((flags & O_ACCMODE) == O_RDWR) {
54 /* GENERIC_ALL is too much permission to request
55 can cause unnecessary access denied on create */
56 /* return GENERIC_ALL; */
57 return (GENERIC_READ | GENERIC_WRITE);
58 }
59
e10f7b55
JL
60 return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES |
61 FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA |
62 FILE_READ_DATA);
7fc8f4e9 63}
e10f7b55 64
608712fe 65static u32 cifs_posix_convert_flags(unsigned int flags)
7fc8f4e9 66{
608712fe 67 u32 posix_flags = 0;
e10f7b55 68
7fc8f4e9 69 if ((flags & O_ACCMODE) == O_RDONLY)
608712fe 70 posix_flags = SMB_O_RDONLY;
7fc8f4e9 71 else if ((flags & O_ACCMODE) == O_WRONLY)
608712fe
JL
72 posix_flags = SMB_O_WRONLY;
73 else if ((flags & O_ACCMODE) == O_RDWR)
74 posix_flags = SMB_O_RDWR;
75
07b92d0d 76 if (flags & O_CREAT) {
608712fe 77 posix_flags |= SMB_O_CREAT;
07b92d0d
SF
78 if (flags & O_EXCL)
79 posix_flags |= SMB_O_EXCL;
80 } else if (flags & O_EXCL)
f96637be
JP
81 cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n",
82 current->comm, current->tgid);
07b92d0d 83
608712fe
JL
84 if (flags & O_TRUNC)
85 posix_flags |= SMB_O_TRUNC;
86 /* be safe and imply O_SYNC for O_DSYNC */
6b2f3d1f 87 if (flags & O_DSYNC)
608712fe 88 posix_flags |= SMB_O_SYNC;
7fc8f4e9 89 if (flags & O_DIRECTORY)
608712fe 90 posix_flags |= SMB_O_DIRECTORY;
7fc8f4e9 91 if (flags & O_NOFOLLOW)
608712fe 92 posix_flags |= SMB_O_NOFOLLOW;
7fc8f4e9 93 if (flags & O_DIRECT)
608712fe 94 posix_flags |= SMB_O_DIRECT;
7fc8f4e9
SF
95
96 return posix_flags;
1da177e4
LT
97}
98
99static inline int cifs_get_disposition(unsigned int flags)
100{
101 if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL))
102 return FILE_CREATE;
103 else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC))
104 return FILE_OVERWRITE_IF;
105 else if ((flags & O_CREAT) == O_CREAT)
106 return FILE_OPEN_IF;
55aa2e09
SF
107 else if ((flags & O_TRUNC) == O_TRUNC)
108 return FILE_OVERWRITE;
1da177e4
LT
109 else
110 return FILE_OPEN;
111}
112
608712fe
JL
113int cifs_posix_open(char *full_path, struct inode **pinode,
114 struct super_block *sb, int mode, unsigned int f_flags,
6d5786a3 115 __u32 *poplock, __u16 *pnetfid, unsigned int xid)
608712fe
JL
116{
117 int rc;
118 FILE_UNIX_BASIC_INFO *presp_data;
119 __u32 posix_flags = 0;
120 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
121 struct cifs_fattr fattr;
122 struct tcon_link *tlink;
96daf2b0 123 struct cifs_tcon *tcon;
608712fe 124
f96637be 125 cifs_dbg(FYI, "posix open %s\n", full_path);
608712fe
JL
126
127 presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL);
128 if (presp_data == NULL)
129 return -ENOMEM;
130
131 tlink = cifs_sb_tlink(cifs_sb);
132 if (IS_ERR(tlink)) {
133 rc = PTR_ERR(tlink);
134 goto posix_open_ret;
135 }
136
137 tcon = tlink_tcon(tlink);
138 mode &= ~current_umask();
139
140 posix_flags = cifs_posix_convert_flags(f_flags);
141 rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data,
142 poplock, full_path, cifs_sb->local_nls,
bc8ebdc4 143 cifs_remap(cifs_sb));
608712fe
JL
144 cifs_put_tlink(tlink);
145
146 if (rc)
147 goto posix_open_ret;
148
149 if (presp_data->Type == cpu_to_le32(-1))
150 goto posix_open_ret; /* open ok, caller does qpathinfo */
151
152 if (!pinode)
153 goto posix_open_ret; /* caller does not need info */
154
155 cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb);
156
157 /* get new inode and set it up */
158 if (*pinode == NULL) {
159 cifs_fill_uniqueid(sb, &fattr);
160 *pinode = cifs_iget(sb, &fattr);
161 if (!*pinode) {
162 rc = -ENOMEM;
163 goto posix_open_ret;
164 }
165 } else {
166 cifs_fattr_to_inode(*pinode, &fattr);
167 }
168
169posix_open_ret:
170 kfree(presp_data);
171 return rc;
172}
173
eeb910a6
PS
174static int
175cifs_nt_open(char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb,
fb1214e4
PS
176 struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock,
177 struct cifs_fid *fid, unsigned int xid)
eeb910a6
PS
178{
179 int rc;
fb1214e4 180 int desired_access;
eeb910a6 181 int disposition;
3d3ea8e6 182 int create_options = CREATE_NOT_DIR;
eeb910a6 183 FILE_ALL_INFO *buf;
b8c32dbb 184 struct TCP_Server_Info *server = tcon->ses->server;
226730b4 185 struct cifs_open_parms oparms;
eeb910a6 186
b8c32dbb 187 if (!server->ops->open)
fb1214e4
PS
188 return -ENOSYS;
189
190 desired_access = cifs_convert_flags(f_flags);
eeb910a6
PS
191
192/*********************************************************************
193 * open flag mapping table:
194 *
195 * POSIX Flag CIFS Disposition
196 * ---------- ----------------
197 * O_CREAT FILE_OPEN_IF
198 * O_CREAT | O_EXCL FILE_CREATE
199 * O_CREAT | O_TRUNC FILE_OVERWRITE_IF
200 * O_TRUNC FILE_OVERWRITE
201 * none of the above FILE_OPEN
202 *
203 * Note that there is not a direct match between disposition
204 * FILE_SUPERSEDE (ie create whether or not file exists although
205 * O_CREAT | O_TRUNC is similar but truncates the existing
206 * file rather than creating a new file as FILE_SUPERSEDE does
207 * (which uses the attributes / metadata passed in on open call)
208 *?
209 *? O_SYNC is a reasonable match to CIFS writethrough flag
210 *? and the read write flags match reasonably. O_LARGEFILE
211 *? is irrelevant because largefile support is always used
212 *? by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY,
213 * O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation
214 *********************************************************************/
215
216 disposition = cifs_get_disposition(f_flags);
217
218 /* BB pass O_SYNC flag through on file attributes .. BB */
219
220 buf = kmalloc(sizeof(FILE_ALL_INFO), GFP_KERNEL);
221 if (!buf)
222 return -ENOMEM;
223
3d3ea8e6
SP
224 if (backup_cred(cifs_sb))
225 create_options |= CREATE_OPEN_BACKUP_INTENT;
226
1013e760
SF
227 /* O_SYNC also has bit for O_DSYNC so following check picks up either */
228 if (f_flags & O_SYNC)
229 create_options |= CREATE_WRITE_THROUGH;
230
231 if (f_flags & O_DIRECT)
232 create_options |= CREATE_NO_BUFFER;
233
226730b4
PS
234 oparms.tcon = tcon;
235 oparms.cifs_sb = cifs_sb;
236 oparms.desired_access = desired_access;
237 oparms.create_options = create_options;
238 oparms.disposition = disposition;
239 oparms.path = full_path;
240 oparms.fid = fid;
9cbc0b73 241 oparms.reconnect = false;
226730b4
PS
242
243 rc = server->ops->open(xid, &oparms, oplock, buf);
eeb910a6
PS
244
245 if (rc)
246 goto out;
247
248 if (tcon->unix_ext)
249 rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb,
250 xid);
251 else
252 rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb,
42eacf9e 253 xid, fid);
eeb910a6
PS
254
255out:
256 kfree(buf);
257 return rc;
258}
259
63b7d3a4
PS
260static bool
261cifs_has_mand_locks(struct cifsInodeInfo *cinode)
262{
263 struct cifs_fid_locks *cur;
264 bool has_locks = false;
265
266 down_read(&cinode->lock_sem);
267 list_for_each_entry(cur, &cinode->llist, llist) {
268 if (!list_empty(&cur->locks)) {
269 has_locks = true;
270 break;
271 }
272 }
273 up_read(&cinode->lock_sem);
274 return has_locks;
275}
276
15ecb436 277struct cifsFileInfo *
fb1214e4 278cifs_new_fileinfo(struct cifs_fid *fid, struct file *file,
15ecb436
JL
279 struct tcon_link *tlink, __u32 oplock)
280{
1f1735cb 281 struct dentry *dentry = file_dentry(file);
2b0143b5 282 struct inode *inode = d_inode(dentry);
4b4de76e
PS
283 struct cifsInodeInfo *cinode = CIFS_I(inode);
284 struct cifsFileInfo *cfile;
f45d3416 285 struct cifs_fid_locks *fdlocks;
233839b1 286 struct cifs_tcon *tcon = tlink_tcon(tlink);
63b7d3a4 287 struct TCP_Server_Info *server = tcon->ses->server;
4b4de76e
PS
288
289 cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL);
290 if (cfile == NULL)
291 return cfile;
292
f45d3416
PS
293 fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL);
294 if (!fdlocks) {
295 kfree(cfile);
296 return NULL;
297 }
298
299 INIT_LIST_HEAD(&fdlocks->locks);
300 fdlocks->cfile = cfile;
301 cfile->llist = fdlocks;
1b4b55a1 302 down_write(&cinode->lock_sem);
f45d3416 303 list_add(&fdlocks->llist, &cinode->llist);
1b4b55a1 304 up_write(&cinode->lock_sem);
f45d3416 305
4b4de76e 306 cfile->count = 1;
4b4de76e
PS
307 cfile->pid = current->tgid;
308 cfile->uid = current_fsuid();
309 cfile->dentry = dget(dentry);
310 cfile->f_flags = file->f_flags;
311 cfile->invalidHandle = false;
312 cfile->tlink = cifs_get_tlink(tlink);
4b4de76e 313 INIT_WORK(&cfile->oplock_break, cifs_oplock_break);
f45d3416 314 mutex_init(&cfile->fh_mutex);
3afca265 315 spin_lock_init(&cfile->file_info_lock);
15ecb436 316
24261fc2
MG
317 cifs_sb_active(inode->i_sb);
318
63b7d3a4
PS
319 /*
320 * If the server returned a read oplock and we have mandatory brlocks,
321 * set oplock level to None.
322 */
53ef1016 323 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
f96637be 324 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
63b7d3a4
PS
325 oplock = 0;
326 }
327
3afca265 328 spin_lock(&tcon->open_file_lock);
63b7d3a4 329 if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock)
233839b1
PS
330 oplock = fid->pending_open->oplock;
331 list_del(&fid->pending_open->olist);
332
42873b0a 333 fid->purge_cache = false;
63b7d3a4 334 server->ops->set_fid(cfile, fid, oplock);
233839b1
PS
335
336 list_add(&cfile->tlist, &tcon->openFileList);
3afca265 337
15ecb436
JL
338 /* if readable file instance put first in list*/
339 if (file->f_mode & FMODE_READ)
4b4de76e 340 list_add(&cfile->flist, &cinode->openFileList);
15ecb436 341 else
4b4de76e 342 list_add_tail(&cfile->flist, &cinode->openFileList);
3afca265 343 spin_unlock(&tcon->open_file_lock);
15ecb436 344
42873b0a 345 if (fid->purge_cache)
4f73c7d3 346 cifs_zap_mapping(inode);
42873b0a 347
4b4de76e
PS
348 file->private_data = cfile;
349 return cfile;
15ecb436
JL
350}
351
764a1b1a
JL
352struct cifsFileInfo *
353cifsFileInfo_get(struct cifsFileInfo *cifs_file)
354{
3afca265 355 spin_lock(&cifs_file->file_info_lock);
764a1b1a 356 cifsFileInfo_get_locked(cifs_file);
3afca265 357 spin_unlock(&cifs_file->file_info_lock);
764a1b1a
JL
358 return cifs_file;
359}
360
cdff08e7
SF
361/*
362 * Release a reference on the file private data. This may involve closing
5f6dbc9e 363 * the filehandle out on the server. Must be called without holding
3afca265 364 * tcon->open_file_lock and cifs_file->file_info_lock.
cdff08e7 365 */
b33879aa
JL
366void cifsFileInfo_put(struct cifsFileInfo *cifs_file)
367{
2b0143b5 368 struct inode *inode = d_inode(cifs_file->dentry);
96daf2b0 369 struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
233839b1 370 struct TCP_Server_Info *server = tcon->ses->server;
e66673e3 371 struct cifsInodeInfo *cifsi = CIFS_I(inode);
24261fc2
MG
372 struct super_block *sb = inode->i_sb;
373 struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
cdff08e7 374 struct cifsLockInfo *li, *tmp;
233839b1
PS
375 struct cifs_fid fid;
376 struct cifs_pending_open open;
ca7df8e0 377 bool oplock_break_cancelled;
cdff08e7 378
3afca265
SF
379 spin_lock(&tcon->open_file_lock);
380
381 spin_lock(&cifs_file->file_info_lock);
5f6dbc9e 382 if (--cifs_file->count > 0) {
3afca265
SF
383 spin_unlock(&cifs_file->file_info_lock);
384 spin_unlock(&tcon->open_file_lock);
cdff08e7
SF
385 return;
386 }
3afca265 387 spin_unlock(&cifs_file->file_info_lock);
cdff08e7 388
233839b1
PS
389 if (server->ops->get_lease_key)
390 server->ops->get_lease_key(inode, &fid);
391
392 /* store open in pending opens to make sure we don't miss lease break */
393 cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open);
394
cdff08e7
SF
395 /* remove it from the lists */
396 list_del(&cifs_file->flist);
397 list_del(&cifs_file->tlist);
398
399 if (list_empty(&cifsi->openFileList)) {
f96637be 400 cifs_dbg(FYI, "closing last open instance for inode %p\n",
2b0143b5 401 d_inode(cifs_file->dentry));
25364138
PS
402 /*
403 * In strict cache mode we need invalidate mapping on the last
404 * close because it may cause a error when we open this file
405 * again and get at least level II oplock.
406 */
4f8ba8a0 407 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO)
aff8d5ca 408 set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags);
c6723628 409 cifs_set_oplock_level(cifsi, 0);
cdff08e7 410 }
3afca265
SF
411
412 spin_unlock(&tcon->open_file_lock);
cdff08e7 413
ca7df8e0 414 oplock_break_cancelled = cancel_work_sync(&cifs_file->oplock_break);
ad635942 415
cdff08e7 416 if (!tcon->need_reconnect && !cifs_file->invalidHandle) {
0ff78a22 417 struct TCP_Server_Info *server = tcon->ses->server;
6d5786a3 418 unsigned int xid;
0ff78a22 419
6d5786a3 420 xid = get_xid();
0ff78a22 421 if (server->ops->close)
760ad0ca
PS
422 server->ops->close(xid, tcon, &cifs_file->fid);
423 _free_xid(xid);
cdff08e7
SF
424 }
425
ca7df8e0
SP
426 if (oplock_break_cancelled)
427 cifs_done_oplock_break(cifsi);
428
233839b1
PS
429 cifs_del_pending_open(&open);
430
f45d3416
PS
431 /*
432 * Delete any outstanding lock records. We'll lose them when the file
cdff08e7
SF
433 * is closed anyway.
434 */
1b4b55a1 435 down_write(&cifsi->lock_sem);
f45d3416 436 list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) {
cdff08e7 437 list_del(&li->llist);
85160e03 438 cifs_del_lock_waiters(li);
cdff08e7 439 kfree(li);
b33879aa 440 }
f45d3416
PS
441 list_del(&cifs_file->llist->llist);
442 kfree(cifs_file->llist);
1b4b55a1 443 up_write(&cifsi->lock_sem);
cdff08e7
SF
444
445 cifs_put_tlink(cifs_file->tlink);
446 dput(cifs_file->dentry);
24261fc2 447 cifs_sb_deactive(sb);
cdff08e7 448 kfree(cifs_file);
b33879aa
JL
449}
450
1da177e4 451int cifs_open(struct inode *inode, struct file *file)
233839b1 452
1da177e4
LT
453{
454 int rc = -EACCES;
6d5786a3 455 unsigned int xid;
590a3fe0 456 __u32 oplock;
1da177e4 457 struct cifs_sb_info *cifs_sb;
b8c32dbb 458 struct TCP_Server_Info *server;
96daf2b0 459 struct cifs_tcon *tcon;
7ffec372 460 struct tcon_link *tlink;
fb1214e4 461 struct cifsFileInfo *cfile = NULL;
1da177e4 462 char *full_path = NULL;
7e12eddb 463 bool posix_open_ok = false;
fb1214e4 464 struct cifs_fid fid;
233839b1 465 struct cifs_pending_open open;
1da177e4 466
6d5786a3 467 xid = get_xid();
1da177e4
LT
468
469 cifs_sb = CIFS_SB(inode->i_sb);
7ffec372
JL
470 tlink = cifs_sb_tlink(cifs_sb);
471 if (IS_ERR(tlink)) {
6d5786a3 472 free_xid(xid);
7ffec372
JL
473 return PTR_ERR(tlink);
474 }
475 tcon = tlink_tcon(tlink);
b8c32dbb 476 server = tcon->ses->server;
1da177e4 477
1f1735cb 478 full_path = build_path_from_dentry(file_dentry(file));
1da177e4 479 if (full_path == NULL) {
0f3bc09e 480 rc = -ENOMEM;
232341ba 481 goto out;
1da177e4
LT
482 }
483
f96637be 484 cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n",
b6b38f70 485 inode, file->f_flags, full_path);
276a74a4 486
787aded6
NJ
487 if (file->f_flags & O_DIRECT &&
488 cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) {
489 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL)
490 file->f_op = &cifs_file_direct_nobrl_ops;
491 else
492 file->f_op = &cifs_file_direct_ops;
493 }
494
233839b1 495 if (server->oplocks)
276a74a4
SF
496 oplock = REQ_OPLOCK;
497 else
498 oplock = 0;
499
64cc2c63 500 if (!tcon->broken_posix_open && tcon->unix_ext &&
29e20f9c
PS
501 cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP &
502 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
276a74a4 503 /* can not refresh inode info since size could be stale */
2422f676 504 rc = cifs_posix_open(full_path, &inode, inode->i_sb,
fa588e0c 505 cifs_sb->mnt_file_mode /* ignored */,
fb1214e4 506 file->f_flags, &oplock, &fid.netfid, xid);
276a74a4 507 if (rc == 0) {
f96637be 508 cifs_dbg(FYI, "posix open succeeded\n");
7e12eddb 509 posix_open_ok = true;
64cc2c63
SF
510 } else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) {
511 if (tcon->ses->serverNOS)
f96637be
JP
512 cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n",
513 tcon->ses->serverName,
514 tcon->ses->serverNOS);
64cc2c63 515 tcon->broken_posix_open = true;
276a74a4
SF
516 } else if ((rc != -EIO) && (rc != -EREMOTE) &&
517 (rc != -EOPNOTSUPP)) /* path not found or net err */
518 goto out;
fb1214e4
PS
519 /*
520 * Else fallthrough to retry open the old way on network i/o
521 * or DFS errors.
522 */
276a74a4
SF
523 }
524
233839b1
PS
525 if (server->ops->get_lease_key)
526 server->ops->get_lease_key(inode, &fid);
527
528 cifs_add_pending_open(&fid, tlink, &open);
529
7e12eddb 530 if (!posix_open_ok) {
b8c32dbb
PS
531 if (server->ops->get_lease_key)
532 server->ops->get_lease_key(inode, &fid);
533
7e12eddb 534 rc = cifs_nt_open(full_path, inode, cifs_sb, tcon,
fb1214e4 535 file->f_flags, &oplock, &fid, xid);
233839b1
PS
536 if (rc) {
537 cifs_del_pending_open(&open);
7e12eddb 538 goto out;
233839b1 539 }
7e12eddb 540 }
47c78b7f 541
fb1214e4
PS
542 cfile = cifs_new_fileinfo(&fid, file, tlink, oplock);
543 if (cfile == NULL) {
b8c32dbb
PS
544 if (server->ops->close)
545 server->ops->close(xid, tcon, &fid);
233839b1 546 cifs_del_pending_open(&open);
1da177e4
LT
547 rc = -ENOMEM;
548 goto out;
549 }
1da177e4 550
9451a9a5
SJ
551 cifs_fscache_set_inode_cookie(inode, file);
552
7e12eddb 553 if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) {
fb1214e4
PS
554 /*
555 * Time to set mode which we can not set earlier due to
556 * problems creating new read-only files.
557 */
7e12eddb
PS
558 struct cifs_unix_set_info_args args = {
559 .mode = inode->i_mode,
49418b2c
EB
560 .uid = INVALID_UID, /* no change */
561 .gid = INVALID_GID, /* no change */
7e12eddb
PS
562 .ctime = NO_CHANGE_64,
563 .atime = NO_CHANGE_64,
564 .mtime = NO_CHANGE_64,
565 .device = 0,
566 };
fb1214e4
PS
567 CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid,
568 cfile->pid);
1da177e4
LT
569 }
570
571out:
1da177e4 572 kfree(full_path);
6d5786a3 573 free_xid(xid);
7ffec372 574 cifs_put_tlink(tlink);
1da177e4
LT
575 return rc;
576}
577
f152fd5f
PS
578static int cifs_push_posix_locks(struct cifsFileInfo *cfile);
579
2ae78ba8
PS
580/*
581 * Try to reacquire byte range locks that were released when session
f152fd5f 582 * to server was lost.
2ae78ba8 583 */
f152fd5f
PS
584static int
585cifs_relock_file(struct cifsFileInfo *cfile)
1da177e4 586{
f152fd5f 587 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2b0143b5 588 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
f152fd5f 589 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1da177e4
LT
590 int rc = 0;
591
560d3889 592 down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING);
f152fd5f 593 if (cinode->can_cache_brlcks) {
689c3db4
PS
594 /* can cache locks - no need to relock */
595 up_read(&cinode->lock_sem);
f152fd5f
PS
596 return rc;
597 }
598
599 if (cap_unix(tcon->ses) &&
600 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
601 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
602 rc = cifs_push_posix_locks(cfile);
603 else
604 rc = tcon->ses->server->ops->push_mand_locks(cfile);
1da177e4 605
689c3db4 606 up_read(&cinode->lock_sem);
1da177e4
LT
607 return rc;
608}
609
2ae78ba8
PS
610static int
611cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush)
1da177e4
LT
612{
613 int rc = -EACCES;
6d5786a3 614 unsigned int xid;
590a3fe0 615 __u32 oplock;
1da177e4 616 struct cifs_sb_info *cifs_sb;
96daf2b0 617 struct cifs_tcon *tcon;
2ae78ba8
PS
618 struct TCP_Server_Info *server;
619 struct cifsInodeInfo *cinode;
fb8c4b14 620 struct inode *inode;
1da177e4 621 char *full_path = NULL;
2ae78ba8 622 int desired_access;
1da177e4 623 int disposition = FILE_OPEN;
3d3ea8e6 624 int create_options = CREATE_NOT_DIR;
226730b4 625 struct cifs_open_parms oparms;
1da177e4 626
6d5786a3 627 xid = get_xid();
2ae78ba8
PS
628 mutex_lock(&cfile->fh_mutex);
629 if (!cfile->invalidHandle) {
630 mutex_unlock(&cfile->fh_mutex);
0f3bc09e 631 rc = 0;
6d5786a3 632 free_xid(xid);
0f3bc09e 633 return rc;
1da177e4
LT
634 }
635
2b0143b5 636 inode = d_inode(cfile->dentry);
1da177e4 637 cifs_sb = CIFS_SB(inode->i_sb);
2ae78ba8
PS
638 tcon = tlink_tcon(cfile->tlink);
639 server = tcon->ses->server;
640
641 /*
642 * Can not grab rename sem here because various ops, including those
643 * that already have the rename sem can end up causing writepage to get
644 * called and if the server was down that means we end up here, and we
645 * can never tell if the caller already has the rename_sem.
646 */
647 full_path = build_path_from_dentry(cfile->dentry);
1da177e4 648 if (full_path == NULL) {
3a9f462f 649 rc = -ENOMEM;
2ae78ba8 650 mutex_unlock(&cfile->fh_mutex);
6d5786a3 651 free_xid(xid);
3a9f462f 652 return rc;
1da177e4
LT
653 }
654
f96637be
JP
655 cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n",
656 inode, cfile->f_flags, full_path);
1da177e4 657
10b9b98e 658 if (tcon->ses->server->oplocks)
1da177e4
LT
659 oplock = REQ_OPLOCK;
660 else
4b18f2a9 661 oplock = 0;
1da177e4 662
29e20f9c 663 if (tcon->unix_ext && cap_unix(tcon->ses) &&
7fc8f4e9 664 (CIFS_UNIX_POSIX_PATH_OPS_CAP &
29e20f9c 665 le64_to_cpu(tcon->fsUnixInfo.Capability))) {
608712fe
JL
666 /*
667 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the
668 * original open. Must mask them off for a reopen.
669 */
2ae78ba8 670 unsigned int oflags = cfile->f_flags &
15886177 671 ~(O_CREAT | O_EXCL | O_TRUNC);
608712fe 672
2422f676 673 rc = cifs_posix_open(full_path, NULL, inode->i_sb,
2ae78ba8 674 cifs_sb->mnt_file_mode /* ignored */,
9cbc0b73 675 oflags, &oplock, &cfile->fid.netfid, xid);
7fc8f4e9 676 if (rc == 0) {
f96637be 677 cifs_dbg(FYI, "posix reopen succeeded\n");
fe090e4e 678 oparms.reconnect = true;
7fc8f4e9
SF
679 goto reopen_success;
680 }
2ae78ba8
PS
681 /*
682 * fallthrough to retry open the old way on errors, especially
683 * in the reconnect path it is important to retry hard
684 */
7fc8f4e9
SF
685 }
686
2ae78ba8 687 desired_access = cifs_convert_flags(cfile->f_flags);
7fc8f4e9 688
3d3ea8e6
SP
689 if (backup_cred(cifs_sb))
690 create_options |= CREATE_OPEN_BACKUP_INTENT;
691
b8c32dbb 692 if (server->ops->get_lease_key)
9cbc0b73 693 server->ops->get_lease_key(inode, &cfile->fid);
b8c32dbb 694
226730b4
PS
695 oparms.tcon = tcon;
696 oparms.cifs_sb = cifs_sb;
697 oparms.desired_access = desired_access;
698 oparms.create_options = create_options;
699 oparms.disposition = disposition;
700 oparms.path = full_path;
9cbc0b73
PS
701 oparms.fid = &cfile->fid;
702 oparms.reconnect = true;
226730b4 703
2ae78ba8
PS
704 /*
705 * Can not refresh inode by passing in file_info buf to be returned by
d81b8a40 706 * ops->open and then calling get_inode_info with returned buf since
2ae78ba8
PS
707 * file might have write behind data that needs to be flushed and server
708 * version of file size can be stale. If we knew for sure that inode was
709 * not dirty locally we could do this.
710 */
226730b4 711 rc = server->ops->open(xid, &oparms, &oplock, NULL);
b33fcf1c
PS
712 if (rc == -ENOENT && oparms.reconnect == false) {
713 /* durable handle timeout is expired - open the file again */
714 rc = server->ops->open(xid, &oparms, &oplock, NULL);
715 /* indicate that we need to relock the file */
716 oparms.reconnect = true;
717 }
718
1da177e4 719 if (rc) {
2ae78ba8 720 mutex_unlock(&cfile->fh_mutex);
f96637be
JP
721 cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc);
722 cifs_dbg(FYI, "oplock: %d\n", oplock);
15886177
JL
723 goto reopen_error_exit;
724 }
725
7fc8f4e9 726reopen_success:
2ae78ba8
PS
727 cfile->invalidHandle = false;
728 mutex_unlock(&cfile->fh_mutex);
729 cinode = CIFS_I(inode);
15886177
JL
730
731 if (can_flush) {
732 rc = filemap_write_and_wait(inode->i_mapping);
eb4b756b 733 mapping_set_error(inode->i_mapping, rc);
15886177 734
15886177 735 if (tcon->unix_ext)
2ae78ba8
PS
736 rc = cifs_get_inode_info_unix(&inode, full_path,
737 inode->i_sb, xid);
15886177 738 else
2ae78ba8
PS
739 rc = cifs_get_inode_info(&inode, full_path, NULL,
740 inode->i_sb, xid, NULL);
741 }
742 /*
743 * Else we are writing out data to server already and could deadlock if
744 * we tried to flush data, and since we do not know if we have data that
745 * would invalidate the current end of file on the server we can not go
746 * to the server to get the new inode info.
747 */
748
de740250
PS
749 /*
750 * If the server returned a read oplock and we have mandatory brlocks,
751 * set oplock level to None.
752 */
753 if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
754 cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
755 oplock = 0;
756 }
757
9cbc0b73
PS
758 server->ops->set_fid(cfile, &cfile->fid, oplock);
759 if (oparms.reconnect)
760 cifs_relock_file(cfile);
15886177
JL
761
762reopen_error_exit:
1da177e4 763 kfree(full_path);
6d5786a3 764 free_xid(xid);
1da177e4
LT
765 return rc;
766}
767
768int cifs_close(struct inode *inode, struct file *file)
769{
77970693
JL
770 if (file->private_data != NULL) {
771 cifsFileInfo_put(file->private_data);
772 file->private_data = NULL;
773 }
7ee1af76 774
cdff08e7
SF
775 /* return code from the ->release op is always ignored */
776 return 0;
1da177e4
LT
777}
778
52ace1ef
SF
779void
780cifs_reopen_persistent_handles(struct cifs_tcon *tcon)
781{
f2cca6a7 782 struct cifsFileInfo *open_file;
52ace1ef
SF
783 struct list_head *tmp;
784 struct list_head *tmp1;
f2cca6a7
PS
785 struct list_head tmp_list;
786
96a988ff
PS
787 if (!tcon->use_persistent || !tcon->need_reopen_files)
788 return;
789
790 tcon->need_reopen_files = false;
791
f2cca6a7
PS
792 cifs_dbg(FYI, "Reopen persistent handles");
793 INIT_LIST_HEAD(&tmp_list);
52ace1ef
SF
794
795 /* list all files open on tree connection, reopen resilient handles */
796 spin_lock(&tcon->open_file_lock);
f2cca6a7 797 list_for_each(tmp, &tcon->openFileList) {
52ace1ef 798 open_file = list_entry(tmp, struct cifsFileInfo, tlist);
f2cca6a7
PS
799 if (!open_file->invalidHandle)
800 continue;
801 cifsFileInfo_get(open_file);
802 list_add_tail(&open_file->rlist, &tmp_list);
52ace1ef
SF
803 }
804 spin_unlock(&tcon->open_file_lock);
f2cca6a7
PS
805
806 list_for_each_safe(tmp, tmp1, &tmp_list) {
807 open_file = list_entry(tmp, struct cifsFileInfo, rlist);
96a988ff
PS
808 if (cifs_reopen_file(open_file, false /* do not flush */))
809 tcon->need_reopen_files = true;
f2cca6a7
PS
810 list_del_init(&open_file->rlist);
811 cifsFileInfo_put(open_file);
812 }
52ace1ef
SF
813}
814
1da177e4
LT
815int cifs_closedir(struct inode *inode, struct file *file)
816{
817 int rc = 0;
6d5786a3 818 unsigned int xid;
4b4de76e 819 struct cifsFileInfo *cfile = file->private_data;
92fc65a7
PS
820 struct cifs_tcon *tcon;
821 struct TCP_Server_Info *server;
822 char *buf;
1da177e4 823
f96637be 824 cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode);
1da177e4 825
92fc65a7
PS
826 if (cfile == NULL)
827 return rc;
828
6d5786a3 829 xid = get_xid();
92fc65a7
PS
830 tcon = tlink_tcon(cfile->tlink);
831 server = tcon->ses->server;
1da177e4 832
f96637be 833 cifs_dbg(FYI, "Freeing private data in close dir\n");
3afca265 834 spin_lock(&cfile->file_info_lock);
52755808 835 if (server->ops->dir_needs_close(cfile)) {
92fc65a7 836 cfile->invalidHandle = true;
3afca265 837 spin_unlock(&cfile->file_info_lock);
92fc65a7
PS
838 if (server->ops->close_dir)
839 rc = server->ops->close_dir(xid, tcon, &cfile->fid);
840 else
841 rc = -ENOSYS;
f96637be 842 cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc);
92fc65a7
PS
843 /* not much we can do if it fails anyway, ignore rc */
844 rc = 0;
845 } else
3afca265 846 spin_unlock(&cfile->file_info_lock);
92fc65a7
PS
847
848 buf = cfile->srch_inf.ntwrk_buf_start;
849 if (buf) {
f96637be 850 cifs_dbg(FYI, "closedir free smb buf in srch struct\n");
92fc65a7
PS
851 cfile->srch_inf.ntwrk_buf_start = NULL;
852 if (cfile->srch_inf.smallBuf)
853 cifs_small_buf_release(buf);
854 else
855 cifs_buf_release(buf);
1da177e4 856 }
92fc65a7
PS
857
858 cifs_put_tlink(cfile->tlink);
859 kfree(file->private_data);
860 file->private_data = NULL;
1da177e4 861 /* BB can we lock the filestruct while this is going on? */
6d5786a3 862 free_xid(xid);
1da177e4
LT
863 return rc;
864}
865
85160e03 866static struct cifsLockInfo *
fbd35aca 867cifs_lock_init(__u64 offset, __u64 length, __u8 type)
7ee1af76 868{
a88b4707 869 struct cifsLockInfo *lock =
fb8c4b14 870 kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL);
a88b4707
PS
871 if (!lock)
872 return lock;
873 lock->offset = offset;
874 lock->length = length;
875 lock->type = type;
a88b4707
PS
876 lock->pid = current->tgid;
877 INIT_LIST_HEAD(&lock->blist);
878 init_waitqueue_head(&lock->block_q);
879 return lock;
85160e03
PS
880}
881
f7ba7fe6 882void
85160e03
PS
883cifs_del_lock_waiters(struct cifsLockInfo *lock)
884{
885 struct cifsLockInfo *li, *tmp;
886 list_for_each_entry_safe(li, tmp, &lock->blist, blist) {
887 list_del_init(&li->blist);
888 wake_up(&li->block_q);
889 }
890}
891
081c0414
PS
892#define CIFS_LOCK_OP 0
893#define CIFS_READ_OP 1
894#define CIFS_WRITE_OP 2
895
896/* @rw_check : 0 - no op, 1 - read, 2 - write */
85160e03 897static bool
f45d3416
PS
898cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset,
899 __u64 length, __u8 type, struct cifsFileInfo *cfile,
081c0414 900 struct cifsLockInfo **conf_lock, int rw_check)
85160e03 901{
fbd35aca 902 struct cifsLockInfo *li;
f45d3416 903 struct cifsFileInfo *cur_cfile = fdlocks->cfile;
106dc538 904 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
85160e03 905
f45d3416 906 list_for_each_entry(li, &fdlocks->locks, llist) {
85160e03
PS
907 if (offset + length <= li->offset ||
908 offset >= li->offset + li->length)
909 continue;
081c0414
PS
910 if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
911 server->ops->compare_fids(cfile, cur_cfile)) {
912 /* shared lock prevents write op through the same fid */
913 if (!(li->type & server->vals->shared_lock_type) ||
914 rw_check != CIFS_WRITE_OP)
915 continue;
916 }
f45d3416
PS
917 if ((type & server->vals->shared_lock_type) &&
918 ((server->ops->compare_fids(cfile, cur_cfile) &&
919 current->tgid == li->pid) || type == li->type))
85160e03 920 continue;
579f9053
PS
921 if (conf_lock)
922 *conf_lock = li;
f45d3416 923 return true;
85160e03
PS
924 }
925 return false;
926}
927
579f9053 928bool
55157dfb 929cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
579f9053 930 __u8 type, struct cifsLockInfo **conf_lock,
081c0414 931 int rw_check)
161ebf9f 932{
fbd35aca 933 bool rc = false;
f45d3416 934 struct cifs_fid_locks *cur;
2b0143b5 935 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
fbd35aca 936
f45d3416
PS
937 list_for_each_entry(cur, &cinode->llist, llist) {
938 rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
579f9053 939 cfile, conf_lock, rw_check);
fbd35aca
PS
940 if (rc)
941 break;
942 }
fbd35aca
PS
943
944 return rc;
161ebf9f
PS
945}
946
9a5101c8
PS
947/*
948 * Check if there is another lock that prevents us to set the lock (mandatory
949 * style). If such a lock exists, update the flock structure with its
950 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
951 * or leave it the same if we can't. Returns 0 if we don't need to request to
952 * the server or 1 otherwise.
953 */
85160e03 954static int
fbd35aca
PS
955cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
956 __u8 type, struct file_lock *flock)
85160e03
PS
957{
958 int rc = 0;
959 struct cifsLockInfo *conf_lock;
2b0143b5 960 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
106dc538 961 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
85160e03
PS
962 bool exist;
963
1b4b55a1 964 down_read(&cinode->lock_sem);
85160e03 965
55157dfb 966 exist = cifs_find_lock_conflict(cfile, offset, length, type,
081c0414 967 &conf_lock, CIFS_LOCK_OP);
85160e03
PS
968 if (exist) {
969 flock->fl_start = conf_lock->offset;
970 flock->fl_end = conf_lock->offset + conf_lock->length - 1;
971 flock->fl_pid = conf_lock->pid;
106dc538 972 if (conf_lock->type & server->vals->shared_lock_type)
85160e03
PS
973 flock->fl_type = F_RDLCK;
974 else
975 flock->fl_type = F_WRLCK;
976 } else if (!cinode->can_cache_brlcks)
977 rc = 1;
978 else
979 flock->fl_type = F_UNLCK;
980
1b4b55a1 981 up_read(&cinode->lock_sem);
85160e03
PS
982 return rc;
983}
984
161ebf9f 985static void
fbd35aca 986cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock)
85160e03 987{
2b0143b5 988 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1b4b55a1 989 down_write(&cinode->lock_sem);
f45d3416 990 list_add_tail(&lock->llist, &cfile->llist->locks);
1b4b55a1 991 up_write(&cinode->lock_sem);
7ee1af76
JA
992}
993
9a5101c8
PS
994/*
995 * Set the byte-range lock (mandatory style). Returns:
996 * 1) 0, if we set the lock and don't need to request to the server;
997 * 2) 1, if no locks prevent us but we need to request to the server;
998 * 3) -EACCESS, if there is a lock that prevents us and wait is false.
999 */
85160e03 1000static int
fbd35aca 1001cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock,
161ebf9f 1002 bool wait)
85160e03 1003{
161ebf9f 1004 struct cifsLockInfo *conf_lock;
2b0143b5 1005 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
85160e03
PS
1006 bool exist;
1007 int rc = 0;
1008
85160e03
PS
1009try_again:
1010 exist = false;
1b4b55a1 1011 down_write(&cinode->lock_sem);
85160e03 1012
55157dfb 1013 exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length,
081c0414 1014 lock->type, &conf_lock, CIFS_LOCK_OP);
85160e03 1015 if (!exist && cinode->can_cache_brlcks) {
f45d3416 1016 list_add_tail(&lock->llist, &cfile->llist->locks);
1b4b55a1 1017 up_write(&cinode->lock_sem);
85160e03
PS
1018 return rc;
1019 }
1020
1021 if (!exist)
1022 rc = 1;
1023 else if (!wait)
1024 rc = -EACCES;
1025 else {
1026 list_add_tail(&lock->blist, &conf_lock->blist);
1b4b55a1 1027 up_write(&cinode->lock_sem);
85160e03
PS
1028 rc = wait_event_interruptible(lock->block_q,
1029 (lock->blist.prev == &lock->blist) &&
1030 (lock->blist.next == &lock->blist));
1031 if (!rc)
1032 goto try_again;
1b4b55a1 1033 down_write(&cinode->lock_sem);
a88b4707 1034 list_del_init(&lock->blist);
85160e03
PS
1035 }
1036
1b4b55a1 1037 up_write(&cinode->lock_sem);
85160e03
PS
1038 return rc;
1039}
1040
9a5101c8
PS
1041/*
1042 * Check if there is another lock that prevents us to set the lock (posix
1043 * style). If such a lock exists, update the flock structure with its
1044 * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1045 * or leave it the same if we can't. Returns 0 if we don't need to request to
1046 * the server or 1 otherwise.
1047 */
85160e03 1048static int
4f6bcec9
PS
1049cifs_posix_lock_test(struct file *file, struct file_lock *flock)
1050{
1051 int rc = 0;
496ad9aa 1052 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
4f6bcec9
PS
1053 unsigned char saved_type = flock->fl_type;
1054
50792760
PS
1055 if ((flock->fl_flags & FL_POSIX) == 0)
1056 return 1;
1057
1b4b55a1 1058 down_read(&cinode->lock_sem);
4f6bcec9
PS
1059 posix_test_lock(file, flock);
1060
1061 if (flock->fl_type == F_UNLCK && !cinode->can_cache_brlcks) {
1062 flock->fl_type = saved_type;
1063 rc = 1;
1064 }
1065
1b4b55a1 1066 up_read(&cinode->lock_sem);
4f6bcec9
PS
1067 return rc;
1068}
1069
9a5101c8
PS
1070/*
1071 * Set the byte-range lock (posix style). Returns:
1072 * 1) 0, if we set the lock and don't need to request to the server;
1073 * 2) 1, if we need to request to the server;
1074 * 3) <0, if the error occurs while setting the lock.
1075 */
4f6bcec9
PS
1076static int
1077cifs_posix_lock_set(struct file *file, struct file_lock *flock)
1078{
496ad9aa 1079 struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
50792760
PS
1080 int rc = 1;
1081
1082 if ((flock->fl_flags & FL_POSIX) == 0)
1083 return rc;
4f6bcec9 1084
66189be7 1085try_again:
1b4b55a1 1086 down_write(&cinode->lock_sem);
4f6bcec9 1087 if (!cinode->can_cache_brlcks) {
1b4b55a1 1088 up_write(&cinode->lock_sem);
50792760 1089 return rc;
4f6bcec9 1090 }
66189be7
PS
1091
1092 rc = posix_lock_file(file, flock, NULL);
1b4b55a1 1093 up_write(&cinode->lock_sem);
66189be7
PS
1094 if (rc == FILE_LOCK_DEFERRED) {
1095 rc = wait_event_interruptible(flock->fl_wait, !flock->fl_next);
1096 if (!rc)
1097 goto try_again;
1a9e64a7 1098 posix_unblock_lock(flock);
66189be7 1099 }
9ebb389d 1100 return rc;
4f6bcec9
PS
1101}
1102
d39a4f71 1103int
4f6bcec9 1104cifs_push_mandatory_locks(struct cifsFileInfo *cfile)
85160e03 1105{
6d5786a3
PS
1106 unsigned int xid;
1107 int rc = 0, stored_rc;
85160e03
PS
1108 struct cifsLockInfo *li, *tmp;
1109 struct cifs_tcon *tcon;
0013fb4c 1110 unsigned int num, max_num, max_buf;
32b9aaf1 1111 LOCKING_ANDX_RANGE *buf, *cur;
4d61eda8
CIK
1112 static const int types[] = {
1113 LOCKING_ANDX_LARGE_FILES,
1114 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1115 };
32b9aaf1 1116 int i;
85160e03 1117
6d5786a3 1118 xid = get_xid();
85160e03
PS
1119 tcon = tlink_tcon(cfile->tlink);
1120
0013fb4c
PS
1121 /*
1122 * Accessing maxBuf is racy with cifs_reconnect - need to store value
1123 * and check it for zero before using.
1124 */
1125 max_buf = tcon->ses->server->maxBuf;
1126 if (!max_buf) {
6d5786a3 1127 free_xid(xid);
0013fb4c
PS
1128 return -EINVAL;
1129 }
1130
1131 max_num = (max_buf - sizeof(struct smb_hdr)) /
1132 sizeof(LOCKING_ANDX_RANGE);
4b99d39b 1133 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
32b9aaf1 1134 if (!buf) {
6d5786a3 1135 free_xid(xid);
e2f2886a 1136 return -ENOMEM;
32b9aaf1
PS
1137 }
1138
1139 for (i = 0; i < 2; i++) {
1140 cur = buf;
1141 num = 0;
f45d3416 1142 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
32b9aaf1
PS
1143 if (li->type != types[i])
1144 continue;
1145 cur->Pid = cpu_to_le16(li->pid);
1146 cur->LengthLow = cpu_to_le32((u32)li->length);
1147 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1148 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1149 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1150 if (++num == max_num) {
4b4de76e
PS
1151 stored_rc = cifs_lockv(xid, tcon,
1152 cfile->fid.netfid,
04a6aa8a
PS
1153 (__u8)li->type, 0, num,
1154 buf);
32b9aaf1
PS
1155 if (stored_rc)
1156 rc = stored_rc;
1157 cur = buf;
1158 num = 0;
1159 } else
1160 cur++;
1161 }
1162
1163 if (num) {
4b4de76e 1164 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
04a6aa8a 1165 (__u8)types[i], 0, num, buf);
32b9aaf1
PS
1166 if (stored_rc)
1167 rc = stored_rc;
1168 }
85160e03
PS
1169 }
1170
32b9aaf1 1171 kfree(buf);
6d5786a3 1172 free_xid(xid);
85160e03
PS
1173 return rc;
1174}
1175
3d22462a
JL
1176static __u32
1177hash_lockowner(fl_owner_t owner)
1178{
1179 return cifs_lock_secret ^ hash32_ptr((const void *)owner);
1180}
1181
d5751469
PS
1182struct lock_to_push {
1183 struct list_head llist;
1184 __u64 offset;
1185 __u64 length;
1186 __u32 pid;
1187 __u16 netfid;
1188 __u8 type;
1189};
1190
4f6bcec9 1191static int
b8db928b 1192cifs_push_posix_locks(struct cifsFileInfo *cfile)
4f6bcec9 1193{
2b0143b5 1194 struct inode *inode = d_inode(cfile->dentry);
4f6bcec9 1195 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
bd61e0a9
JL
1196 struct file_lock *flock;
1197 struct file_lock_context *flctx = inode->i_flctx;
e084c1bd 1198 unsigned int count = 0, i;
4f6bcec9 1199 int rc = 0, xid, type;
d5751469
PS
1200 struct list_head locks_to_send, *el;
1201 struct lock_to_push *lck, *tmp;
4f6bcec9 1202 __u64 length;
4f6bcec9 1203
6d5786a3 1204 xid = get_xid();
4f6bcec9 1205
bd61e0a9
JL
1206 if (!flctx)
1207 goto out;
d5751469 1208
e084c1bd
JL
1209 spin_lock(&flctx->flc_lock);
1210 list_for_each(el, &flctx->flc_posix) {
1211 count++;
1212 }
1213 spin_unlock(&flctx->flc_lock);
1214
4f6bcec9
PS
1215 INIT_LIST_HEAD(&locks_to_send);
1216
d5751469 1217 /*
e084c1bd
JL
1218 * Allocating count locks is enough because no FL_POSIX locks can be
1219 * added to the list while we are holding cinode->lock_sem that
ce85852b 1220 * protects locking operations of this inode.
d5751469 1221 */
e084c1bd 1222 for (i = 0; i < count; i++) {
d5751469
PS
1223 lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL);
1224 if (!lck) {
1225 rc = -ENOMEM;
1226 goto err_out;
1227 }
1228 list_add_tail(&lck->llist, &locks_to_send);
1229 }
1230
d5751469 1231 el = locks_to_send.next;
6109c850 1232 spin_lock(&flctx->flc_lock);
bd61e0a9 1233 list_for_each_entry(flock, &flctx->flc_posix, fl_list) {
d5751469 1234 if (el == &locks_to_send) {
ce85852b
PS
1235 /*
1236 * The list ended. We don't have enough allocated
1237 * structures - something is really wrong.
1238 */
f96637be 1239 cifs_dbg(VFS, "Can't push all brlocks!\n");
d5751469
PS
1240 break;
1241 }
4f6bcec9
PS
1242 length = 1 + flock->fl_end - flock->fl_start;
1243 if (flock->fl_type == F_RDLCK || flock->fl_type == F_SHLCK)
1244 type = CIFS_RDLCK;
1245 else
1246 type = CIFS_WRLCK;
d5751469 1247 lck = list_entry(el, struct lock_to_push, llist);
3d22462a 1248 lck->pid = hash_lockowner(flock->fl_owner);
4b4de76e 1249 lck->netfid = cfile->fid.netfid;
d5751469
PS
1250 lck->length = length;
1251 lck->type = type;
1252 lck->offset = flock->fl_start;
4f6bcec9 1253 }
6109c850 1254 spin_unlock(&flctx->flc_lock);
4f6bcec9
PS
1255
1256 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
4f6bcec9
PS
1257 int stored_rc;
1258
4f6bcec9 1259 stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid,
c5fd363d 1260 lck->offset, lck->length, NULL,
4f6bcec9
PS
1261 lck->type, 0);
1262 if (stored_rc)
1263 rc = stored_rc;
1264 list_del(&lck->llist);
1265 kfree(lck);
1266 }
1267
d5751469 1268out:
6d5786a3 1269 free_xid(xid);
4f6bcec9 1270 return rc;
d5751469
PS
1271err_out:
1272 list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1273 list_del(&lck->llist);
1274 kfree(lck);
1275 }
1276 goto out;
4f6bcec9
PS
1277}
1278
9ec3c882 1279static int
b8db928b 1280cifs_push_locks(struct cifsFileInfo *cfile)
9ec3c882 1281{
b8db928b 1282 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2b0143b5 1283 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
b8db928b 1284 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
9ec3c882
PS
1285 int rc = 0;
1286
1287 /* we are going to update can_cache_brlcks here - need a write access */
1288 down_write(&cinode->lock_sem);
1289 if (!cinode->can_cache_brlcks) {
1290 up_write(&cinode->lock_sem);
1291 return rc;
1292 }
4f6bcec9 1293
29e20f9c 1294 if (cap_unix(tcon->ses) &&
4f6bcec9
PS
1295 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1296 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
b8db928b
PS
1297 rc = cifs_push_posix_locks(cfile);
1298 else
1299 rc = tcon->ses->server->ops->push_mand_locks(cfile);
4f6bcec9 1300
b8db928b
PS
1301 cinode->can_cache_brlcks = false;
1302 up_write(&cinode->lock_sem);
1303 return rc;
4f6bcec9
PS
1304}
1305
03776f45 1306static void
04a6aa8a 1307cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock,
106dc538 1308 bool *wait_flag, struct TCP_Server_Info *server)
1da177e4 1309{
03776f45 1310 if (flock->fl_flags & FL_POSIX)
f96637be 1311 cifs_dbg(FYI, "Posix\n");
03776f45 1312 if (flock->fl_flags & FL_FLOCK)
f96637be 1313 cifs_dbg(FYI, "Flock\n");
03776f45 1314 if (flock->fl_flags & FL_SLEEP) {
f96637be 1315 cifs_dbg(FYI, "Blocking lock\n");
03776f45 1316 *wait_flag = true;
1da177e4 1317 }
03776f45 1318 if (flock->fl_flags & FL_ACCESS)
f96637be 1319 cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n");
03776f45 1320 if (flock->fl_flags & FL_LEASE)
f96637be 1321 cifs_dbg(FYI, "Lease on file - not implemented yet\n");
03776f45 1322 if (flock->fl_flags &
3d6d854a
JL
1323 (~(FL_POSIX | FL_FLOCK | FL_SLEEP |
1324 FL_ACCESS | FL_LEASE | FL_CLOSE)))
f96637be 1325 cifs_dbg(FYI, "Unknown lock flags 0x%x\n", flock->fl_flags);
1da177e4 1326
106dc538 1327 *type = server->vals->large_lock_type;
03776f45 1328 if (flock->fl_type == F_WRLCK) {
f96637be 1329 cifs_dbg(FYI, "F_WRLCK\n");
106dc538 1330 *type |= server->vals->exclusive_lock_type;
03776f45
PS
1331 *lock = 1;
1332 } else if (flock->fl_type == F_UNLCK) {
f96637be 1333 cifs_dbg(FYI, "F_UNLCK\n");
106dc538 1334 *type |= server->vals->unlock_lock_type;
03776f45
PS
1335 *unlock = 1;
1336 /* Check if unlock includes more than one lock range */
1337 } else if (flock->fl_type == F_RDLCK) {
f96637be 1338 cifs_dbg(FYI, "F_RDLCK\n");
106dc538 1339 *type |= server->vals->shared_lock_type;
03776f45
PS
1340 *lock = 1;
1341 } else if (flock->fl_type == F_EXLCK) {
f96637be 1342 cifs_dbg(FYI, "F_EXLCK\n");
106dc538 1343 *type |= server->vals->exclusive_lock_type;
03776f45
PS
1344 *lock = 1;
1345 } else if (flock->fl_type == F_SHLCK) {
f96637be 1346 cifs_dbg(FYI, "F_SHLCK\n");
106dc538 1347 *type |= server->vals->shared_lock_type;
03776f45 1348 *lock = 1;
1da177e4 1349 } else
f96637be 1350 cifs_dbg(FYI, "Unknown type of lock\n");
03776f45 1351}
1da177e4 1352
03776f45 1353static int
04a6aa8a 1354cifs_getlk(struct file *file, struct file_lock *flock, __u32 type,
6d5786a3 1355 bool wait_flag, bool posix_lck, unsigned int xid)
03776f45
PS
1356{
1357 int rc = 0;
1358 __u64 length = 1 + flock->fl_end - flock->fl_start;
4f6bcec9
PS
1359 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1360 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
106dc538 1361 struct TCP_Server_Info *server = tcon->ses->server;
4b4de76e 1362 __u16 netfid = cfile->fid.netfid;
f05337c6 1363
03776f45
PS
1364 if (posix_lck) {
1365 int posix_lock_type;
4f6bcec9
PS
1366
1367 rc = cifs_posix_lock_test(file, flock);
1368 if (!rc)
1369 return rc;
1370
106dc538 1371 if (type & server->vals->shared_lock_type)
03776f45
PS
1372 posix_lock_type = CIFS_RDLCK;
1373 else
1374 posix_lock_type = CIFS_WRLCK;
3d22462a
JL
1375 rc = CIFSSMBPosixLock(xid, tcon, netfid,
1376 hash_lockowner(flock->fl_owner),
c5fd363d 1377 flock->fl_start, length, flock,
4f6bcec9 1378 posix_lock_type, wait_flag);
03776f45
PS
1379 return rc;
1380 }
1da177e4 1381
fbd35aca 1382 rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock);
85160e03
PS
1383 if (!rc)
1384 return rc;
1385
03776f45 1386 /* BB we could chain these into one lock request BB */
d39a4f71
PS
1387 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type,
1388 1, 0, false);
03776f45 1389 if (rc == 0) {
d39a4f71
PS
1390 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1391 type, 0, 1, false);
03776f45
PS
1392 flock->fl_type = F_UNLCK;
1393 if (rc != 0)
f96637be
JP
1394 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1395 rc);
a88b4707 1396 return 0;
1da177e4 1397 }
7ee1af76 1398
106dc538 1399 if (type & server->vals->shared_lock_type) {
03776f45 1400 flock->fl_type = F_WRLCK;
a88b4707 1401 return 0;
7ee1af76
JA
1402 }
1403
d39a4f71
PS
1404 type &= ~server->vals->exclusive_lock_type;
1405
1406 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1407 type | server->vals->shared_lock_type,
1408 1, 0, false);
03776f45 1409 if (rc == 0) {
d39a4f71
PS
1410 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1411 type | server->vals->shared_lock_type, 0, 1, false);
03776f45
PS
1412 flock->fl_type = F_RDLCK;
1413 if (rc != 0)
f96637be
JP
1414 cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1415 rc);
03776f45
PS
1416 } else
1417 flock->fl_type = F_WRLCK;
1418
a88b4707 1419 return 0;
03776f45
PS
1420}
1421
f7ba7fe6 1422void
9ee305b7
PS
1423cifs_move_llist(struct list_head *source, struct list_head *dest)
1424{
1425 struct list_head *li, *tmp;
1426 list_for_each_safe(li, tmp, source)
1427 list_move(li, dest);
1428}
1429
f7ba7fe6 1430void
9ee305b7
PS
1431cifs_free_llist(struct list_head *llist)
1432{
1433 struct cifsLockInfo *li, *tmp;
1434 list_for_each_entry_safe(li, tmp, llist, llist) {
1435 cifs_del_lock_waiters(li);
1436 list_del(&li->llist);
1437 kfree(li);
1438 }
1439}
1440
d39a4f71 1441int
6d5786a3
PS
1442cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock,
1443 unsigned int xid)
9ee305b7
PS
1444{
1445 int rc = 0, stored_rc;
4d61eda8
CIK
1446 static const int types[] = {
1447 LOCKING_ANDX_LARGE_FILES,
1448 LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1449 };
9ee305b7 1450 unsigned int i;
0013fb4c 1451 unsigned int max_num, num, max_buf;
9ee305b7
PS
1452 LOCKING_ANDX_RANGE *buf, *cur;
1453 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2b0143b5 1454 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
9ee305b7
PS
1455 struct cifsLockInfo *li, *tmp;
1456 __u64 length = 1 + flock->fl_end - flock->fl_start;
1457 struct list_head tmp_llist;
1458
1459 INIT_LIST_HEAD(&tmp_llist);
1460
0013fb4c
PS
1461 /*
1462 * Accessing maxBuf is racy with cifs_reconnect - need to store value
1463 * and check it for zero before using.
1464 */
1465 max_buf = tcon->ses->server->maxBuf;
1466 if (!max_buf)
1467 return -EINVAL;
1468
1469 max_num = (max_buf - sizeof(struct smb_hdr)) /
1470 sizeof(LOCKING_ANDX_RANGE);
4b99d39b 1471 buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
9ee305b7
PS
1472 if (!buf)
1473 return -ENOMEM;
1474
1b4b55a1 1475 down_write(&cinode->lock_sem);
9ee305b7
PS
1476 for (i = 0; i < 2; i++) {
1477 cur = buf;
1478 num = 0;
f45d3416 1479 list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
9ee305b7
PS
1480 if (flock->fl_start > li->offset ||
1481 (flock->fl_start + length) <
1482 (li->offset + li->length))
1483 continue;
1484 if (current->tgid != li->pid)
1485 continue;
9ee305b7
PS
1486 if (types[i] != li->type)
1487 continue;
ea319d57 1488 if (cinode->can_cache_brlcks) {
9ee305b7
PS
1489 /*
1490 * We can cache brlock requests - simply remove
fbd35aca 1491 * a lock from the file's list.
9ee305b7
PS
1492 */
1493 list_del(&li->llist);
1494 cifs_del_lock_waiters(li);
1495 kfree(li);
ea319d57 1496 continue;
9ee305b7 1497 }
ea319d57
PS
1498 cur->Pid = cpu_to_le16(li->pid);
1499 cur->LengthLow = cpu_to_le32((u32)li->length);
1500 cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1501 cur->OffsetLow = cpu_to_le32((u32)li->offset);
1502 cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1503 /*
1504 * We need to save a lock here to let us add it again to
1505 * the file's list if the unlock range request fails on
1506 * the server.
1507 */
1508 list_move(&li->llist, &tmp_llist);
1509 if (++num == max_num) {
4b4de76e
PS
1510 stored_rc = cifs_lockv(xid, tcon,
1511 cfile->fid.netfid,
ea319d57
PS
1512 li->type, num, 0, buf);
1513 if (stored_rc) {
1514 /*
1515 * We failed on the unlock range
1516 * request - add all locks from the tmp
1517 * list to the head of the file's list.
1518 */
1519 cifs_move_llist(&tmp_llist,
f45d3416 1520 &cfile->llist->locks);
ea319d57
PS
1521 rc = stored_rc;
1522 } else
1523 /*
1524 * The unlock range request succeed -
1525 * free the tmp list.
1526 */
1527 cifs_free_llist(&tmp_llist);
1528 cur = buf;
1529 num = 0;
1530 } else
1531 cur++;
9ee305b7
PS
1532 }
1533 if (num) {
4b4de76e 1534 stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
9ee305b7
PS
1535 types[i], num, 0, buf);
1536 if (stored_rc) {
f45d3416
PS
1537 cifs_move_llist(&tmp_llist,
1538 &cfile->llist->locks);
9ee305b7
PS
1539 rc = stored_rc;
1540 } else
1541 cifs_free_llist(&tmp_llist);
1542 }
1543 }
1544
1b4b55a1 1545 up_write(&cinode->lock_sem);
9ee305b7
PS
1546 kfree(buf);
1547 return rc;
1548}
1549
03776f45 1550static int
f45d3416 1551cifs_setlk(struct file *file, struct file_lock *flock, __u32 type,
6d5786a3
PS
1552 bool wait_flag, bool posix_lck, int lock, int unlock,
1553 unsigned int xid)
03776f45
PS
1554{
1555 int rc = 0;
1556 __u64 length = 1 + flock->fl_end - flock->fl_start;
1557 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1558 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
106dc538 1559 struct TCP_Server_Info *server = tcon->ses->server;
2b0143b5 1560 struct inode *inode = d_inode(cfile->dentry);
03776f45
PS
1561
1562 if (posix_lck) {
08547b03 1563 int posix_lock_type;
4f6bcec9
PS
1564
1565 rc = cifs_posix_lock_set(file, flock);
1566 if (!rc || rc < 0)
1567 return rc;
1568
106dc538 1569 if (type & server->vals->shared_lock_type)
08547b03
SF
1570 posix_lock_type = CIFS_RDLCK;
1571 else
1572 posix_lock_type = CIFS_WRLCK;
50c2f753 1573
03776f45 1574 if (unlock == 1)
beb84dc8 1575 posix_lock_type = CIFS_UNLCK;
7ee1af76 1576
f45d3416 1577 rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid,
3d22462a
JL
1578 hash_lockowner(flock->fl_owner),
1579 flock->fl_start, length,
f45d3416 1580 NULL, posix_lock_type, wait_flag);
03776f45
PS
1581 goto out;
1582 }
7ee1af76 1583
03776f45 1584 if (lock) {
161ebf9f
PS
1585 struct cifsLockInfo *lock;
1586
fbd35aca 1587 lock = cifs_lock_init(flock->fl_start, length, type);
161ebf9f
PS
1588 if (!lock)
1589 return -ENOMEM;
1590
fbd35aca 1591 rc = cifs_lock_add_if(cfile, lock, wait_flag);
21cb2d90 1592 if (rc < 0) {
161ebf9f 1593 kfree(lock);
21cb2d90
PS
1594 return rc;
1595 }
1596 if (!rc)
85160e03
PS
1597 goto out;
1598
63b7d3a4
PS
1599 /*
1600 * Windows 7 server can delay breaking lease from read to None
1601 * if we set a byte-range lock on a file - break it explicitly
1602 * before sending the lock to the server to be sure the next
1603 * read won't conflict with non-overlapted locks due to
1604 * pagereading.
1605 */
18cceb6a
PS
1606 if (!CIFS_CACHE_WRITE(CIFS_I(inode)) &&
1607 CIFS_CACHE_READ(CIFS_I(inode))) {
4f73c7d3 1608 cifs_zap_mapping(inode);
f96637be
JP
1609 cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n",
1610 inode);
18cceb6a 1611 CIFS_I(inode)->oplock = 0;
63b7d3a4
PS
1612 }
1613
d39a4f71
PS
1614 rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1615 type, 1, 0, wait_flag);
161ebf9f
PS
1616 if (rc) {
1617 kfree(lock);
21cb2d90 1618 return rc;
03776f45 1619 }
161ebf9f 1620
fbd35aca 1621 cifs_lock_add(cfile, lock);
9ee305b7 1622 } else if (unlock)
d39a4f71 1623 rc = server->ops->mand_unlock_range(cfile, flock, xid);
03776f45 1624
03776f45 1625out:
00b8c95b 1626 if (flock->fl_flags & FL_POSIX && !rc)
4f656367 1627 rc = locks_lock_file_wait(file, flock);
03776f45
PS
1628 return rc;
1629}
1630
1631int cifs_lock(struct file *file, int cmd, struct file_lock *flock)
1632{
1633 int rc, xid;
1634 int lock = 0, unlock = 0;
1635 bool wait_flag = false;
1636 bool posix_lck = false;
1637 struct cifs_sb_info *cifs_sb;
1638 struct cifs_tcon *tcon;
1639 struct cifsInodeInfo *cinode;
1640 struct cifsFileInfo *cfile;
1641 __u16 netfid;
04a6aa8a 1642 __u32 type;
03776f45
PS
1643
1644 rc = -EACCES;
6d5786a3 1645 xid = get_xid();
03776f45 1646
f96637be
JP
1647 cifs_dbg(FYI, "Lock parm: 0x%x flockflags: 0x%x flocktype: 0x%x start: %lld end: %lld\n",
1648 cmd, flock->fl_flags, flock->fl_type,
1649 flock->fl_start, flock->fl_end);
03776f45 1650
03776f45
PS
1651 cfile = (struct cifsFileInfo *)file->private_data;
1652 tcon = tlink_tcon(cfile->tlink);
106dc538
PS
1653
1654 cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag,
1655 tcon->ses->server);
1656
7119e220 1657 cifs_sb = CIFS_FILE_SB(file);
4b4de76e 1658 netfid = cfile->fid.netfid;
496ad9aa 1659 cinode = CIFS_I(file_inode(file));
03776f45 1660
29e20f9c 1661 if (cap_unix(tcon->ses) &&
03776f45
PS
1662 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1663 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1664 posix_lck = true;
1665 /*
1666 * BB add code here to normalize offset and length to account for
1667 * negative length which we can not accept over the wire.
1668 */
1669 if (IS_GETLK(cmd)) {
4f6bcec9 1670 rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid);
6d5786a3 1671 free_xid(xid);
03776f45
PS
1672 return rc;
1673 }
1674
1675 if (!lock && !unlock) {
1676 /*
1677 * if no lock or unlock then nothing to do since we do not
1678 * know what it is
1679 */
6d5786a3 1680 free_xid(xid);
03776f45 1681 return -EOPNOTSUPP;
7ee1af76
JA
1682 }
1683
03776f45
PS
1684 rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock,
1685 xid);
6d5786a3 1686 free_xid(xid);
1da177e4
LT
1687 return rc;
1688}
1689
597b027f
JL
1690/*
1691 * update the file size (if needed) after a write. Should be called with
1692 * the inode->i_lock held
1693 */
72432ffc 1694void
fbec9ab9
JL
1695cifs_update_eof(struct cifsInodeInfo *cifsi, loff_t offset,
1696 unsigned int bytes_written)
1697{
1698 loff_t end_of_write = offset + bytes_written;
1699
1700 if (end_of_write > cifsi->server_eof)
1701 cifsi->server_eof = end_of_write;
1702}
1703
ba9ad725
PS
1704static ssize_t
1705cifs_write(struct cifsFileInfo *open_file, __u32 pid, const char *write_data,
1706 size_t write_size, loff_t *offset)
1da177e4
LT
1707{
1708 int rc = 0;
1709 unsigned int bytes_written = 0;
1710 unsigned int total_written;
1711 struct cifs_sb_info *cifs_sb;
ba9ad725
PS
1712 struct cifs_tcon *tcon;
1713 struct TCP_Server_Info *server;
6d5786a3 1714 unsigned int xid;
7da4b49a 1715 struct dentry *dentry = open_file->dentry;
2b0143b5 1716 struct cifsInodeInfo *cifsi = CIFS_I(d_inode(dentry));
fa2989f4 1717 struct cifs_io_parms io_parms;
1da177e4 1718
7da4b49a 1719 cifs_sb = CIFS_SB(dentry->d_sb);
1da177e4 1720
35c265e0
AV
1721 cifs_dbg(FYI, "write %zd bytes to offset %lld of %pd\n",
1722 write_size, *offset, dentry);
1da177e4 1723
ba9ad725
PS
1724 tcon = tlink_tcon(open_file->tlink);
1725 server = tcon->ses->server;
1726
1727 if (!server->ops->sync_write)
1728 return -ENOSYS;
50c2f753 1729
6d5786a3 1730 xid = get_xid();
1da177e4 1731
1da177e4
LT
1732 for (total_written = 0; write_size > total_written;
1733 total_written += bytes_written) {
1734 rc = -EAGAIN;
1735 while (rc == -EAGAIN) {
ca83ce3d
JL
1736 struct kvec iov[2];
1737 unsigned int len;
1738
1da177e4 1739 if (open_file->invalidHandle) {
1da177e4
LT
1740 /* we could deadlock if we called
1741 filemap_fdatawait from here so tell
fb8c4b14 1742 reopen_file not to flush data to
1da177e4 1743 server now */
15886177 1744 rc = cifs_reopen_file(open_file, false);
1da177e4
LT
1745 if (rc != 0)
1746 break;
1747 }
ca83ce3d 1748
2b0143b5 1749 len = min(server->ops->wp_retry_size(d_inode(dentry)),
cb7e9eab 1750 (unsigned int)write_size - total_written);
ca83ce3d
JL
1751 /* iov[0] is reserved for smb header */
1752 iov[1].iov_base = (char *)write_data + total_written;
1753 iov[1].iov_len = len;
fa2989f4 1754 io_parms.pid = pid;
ba9ad725
PS
1755 io_parms.tcon = tcon;
1756 io_parms.offset = *offset;
fa2989f4 1757 io_parms.length = len;
db8b631d
SF
1758 rc = server->ops->sync_write(xid, &open_file->fid,
1759 &io_parms, &bytes_written, iov, 1);
1da177e4
LT
1760 }
1761 if (rc || (bytes_written == 0)) {
1762 if (total_written)
1763 break;
1764 else {
6d5786a3 1765 free_xid(xid);
1da177e4
LT
1766 return rc;
1767 }
fbec9ab9 1768 } else {
2b0143b5 1769 spin_lock(&d_inode(dentry)->i_lock);
ba9ad725 1770 cifs_update_eof(cifsi, *offset, bytes_written);
2b0143b5 1771 spin_unlock(&d_inode(dentry)->i_lock);
ba9ad725 1772 *offset += bytes_written;
fbec9ab9 1773 }
1da177e4
LT
1774 }
1775
ba9ad725 1776 cifs_stats_bytes_written(tcon, total_written);
1da177e4 1777
7da4b49a 1778 if (total_written > 0) {
2b0143b5
DH
1779 spin_lock(&d_inode(dentry)->i_lock);
1780 if (*offset > d_inode(dentry)->i_size)
1781 i_size_write(d_inode(dentry), *offset);
1782 spin_unlock(&d_inode(dentry)->i_lock);
1da177e4 1783 }
2b0143b5 1784 mark_inode_dirty_sync(d_inode(dentry));
6d5786a3 1785 free_xid(xid);
1da177e4
LT
1786 return total_written;
1787}
1788
6508d904
JL
1789struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode,
1790 bool fsuid_only)
630f3f0c
SF
1791{
1792 struct cifsFileInfo *open_file = NULL;
6508d904 1793 struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
3afca265 1794 struct cifs_tcon *tcon = cifs_sb_master_tcon(cifs_sb);
6508d904
JL
1795
1796 /* only filter by fsuid on multiuser mounts */
1797 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1798 fsuid_only = false;
630f3f0c 1799
3afca265 1800 spin_lock(&tcon->open_file_lock);
630f3f0c
SF
1801 /* we could simply get the first_list_entry since write-only entries
1802 are always at the end of the list but since the first entry might
1803 have a close pending, we go through the whole list */
1804 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
fef59fd7 1805 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
6508d904 1806 continue;
2e396b83 1807 if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) {
630f3f0c
SF
1808 if (!open_file->invalidHandle) {
1809 /* found a good file */
1810 /* lock it so it will not be closed on us */
3afca265
SF
1811 cifsFileInfo_get(open_file);
1812 spin_unlock(&tcon->open_file_lock);
630f3f0c
SF
1813 return open_file;
1814 } /* else might as well continue, and look for
1815 another, or simply have the caller reopen it
1816 again rather than trying to fix this handle */
1817 } else /* write only file */
1818 break; /* write only files are last so must be done */
1819 }
3afca265 1820 spin_unlock(&tcon->open_file_lock);
630f3f0c
SF
1821 return NULL;
1822}
630f3f0c 1823
6508d904
JL
1824struct cifsFileInfo *find_writable_file(struct cifsInodeInfo *cifs_inode,
1825 bool fsuid_only)
6148a742 1826{
2c0c2a08 1827 struct cifsFileInfo *open_file, *inv_file = NULL;
d3892294 1828 struct cifs_sb_info *cifs_sb;
3afca265 1829 struct cifs_tcon *tcon;
2846d386 1830 bool any_available = false;
dd99cd80 1831 int rc;
2c0c2a08 1832 unsigned int refind = 0;
6148a742 1833
60808233
SF
1834 /* Having a null inode here (because mapping->host was set to zero by
1835 the VFS or MM) should not happen but we had reports of on oops (due to
1836 it being zero) during stress testcases so we need to check for it */
1837
fb8c4b14 1838 if (cifs_inode == NULL) {
f96637be 1839 cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n");
60808233
SF
1840 dump_stack();
1841 return NULL;
1842 }
1843
d3892294 1844 cifs_sb = CIFS_SB(cifs_inode->vfs_inode.i_sb);
3afca265 1845 tcon = cifs_sb_master_tcon(cifs_sb);
d3892294 1846
6508d904
JL
1847 /* only filter by fsuid on multiuser mounts */
1848 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
1849 fsuid_only = false;
1850
3afca265 1851 spin_lock(&tcon->open_file_lock);
9b22b0b7 1852refind_writable:
2c0c2a08 1853 if (refind > MAX_REOPEN_ATT) {
3afca265 1854 spin_unlock(&tcon->open_file_lock);
2c0c2a08
SP
1855 return NULL;
1856 }
6148a742 1857 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
6508d904
JL
1858 if (!any_available && open_file->pid != current->tgid)
1859 continue;
fef59fd7 1860 if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
6148a742 1861 continue;
2e396b83 1862 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
9b22b0b7
SF
1863 if (!open_file->invalidHandle) {
1864 /* found a good writable file */
3afca265
SF
1865 cifsFileInfo_get(open_file);
1866 spin_unlock(&tcon->open_file_lock);
9b22b0b7 1867 return open_file;
2c0c2a08
SP
1868 } else {
1869 if (!inv_file)
1870 inv_file = open_file;
9b22b0b7 1871 }
6148a742
SF
1872 }
1873 }
2846d386
JL
1874 /* couldn't find useable FH with same pid, try any available */
1875 if (!any_available) {
1876 any_available = true;
1877 goto refind_writable;
1878 }
2c0c2a08
SP
1879
1880 if (inv_file) {
1881 any_available = false;
3afca265 1882 cifsFileInfo_get(inv_file);
2c0c2a08
SP
1883 }
1884
3afca265 1885 spin_unlock(&tcon->open_file_lock);
2c0c2a08
SP
1886
1887 if (inv_file) {
1888 rc = cifs_reopen_file(inv_file, false);
1889 if (!rc)
1890 return inv_file;
1891 else {
3afca265 1892 spin_lock(&tcon->open_file_lock);
2c0c2a08
SP
1893 list_move_tail(&inv_file->flist,
1894 &cifs_inode->openFileList);
3afca265 1895 spin_unlock(&tcon->open_file_lock);
2c0c2a08 1896 cifsFileInfo_put(inv_file);
2c0c2a08 1897 ++refind;
e1e9bda2 1898 inv_file = NULL;
3afca265 1899 spin_lock(&tcon->open_file_lock);
2c0c2a08
SP
1900 goto refind_writable;
1901 }
1902 }
1903
6148a742
SF
1904 return NULL;
1905}
1906
1da177e4
LT
1907static int cifs_partialpagewrite(struct page *page, unsigned from, unsigned to)
1908{
1909 struct address_space *mapping = page->mapping;
09cbfeaf 1910 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
1da177e4
LT
1911 char *write_data;
1912 int rc = -EFAULT;
1913 int bytes_written = 0;
1da177e4 1914 struct inode *inode;
6148a742 1915 struct cifsFileInfo *open_file;
1da177e4
LT
1916
1917 if (!mapping || !mapping->host)
1918 return -EFAULT;
1919
1920 inode = page->mapping->host;
1da177e4
LT
1921
1922 offset += (loff_t)from;
1923 write_data = kmap(page);
1924 write_data += from;
1925
09cbfeaf 1926 if ((to > PAGE_SIZE) || (from > to)) {
1da177e4
LT
1927 kunmap(page);
1928 return -EIO;
1929 }
1930
1931 /* racing with truncate? */
1932 if (offset > mapping->host->i_size) {
1933 kunmap(page);
1934 return 0; /* don't care */
1935 }
1936
1937 /* check to make sure that we are not extending the file */
1938 if (mapping->host->i_size - offset < (loff_t)to)
fb8c4b14 1939 to = (unsigned)(mapping->host->i_size - offset);
1da177e4 1940
6508d904 1941 open_file = find_writable_file(CIFS_I(mapping->host), false);
6148a742 1942 if (open_file) {
fa2989f4
PS
1943 bytes_written = cifs_write(open_file, open_file->pid,
1944 write_data, to - from, &offset);
6ab409b5 1945 cifsFileInfo_put(open_file);
1da177e4 1946 /* Does mm or vfs already set times? */
c2050a45 1947 inode->i_atime = inode->i_mtime = current_time(inode);
bb5a9a04 1948 if ((bytes_written > 0) && (offset))
6148a742 1949 rc = 0;
bb5a9a04
SF
1950 else if (bytes_written < 0)
1951 rc = bytes_written;
6148a742 1952 } else {
f96637be 1953 cifs_dbg(FYI, "No writeable filehandles for inode\n");
1da177e4
LT
1954 rc = -EIO;
1955 }
1956
1957 kunmap(page);
1958 return rc;
1959}
1960
90ac1387
PS
1961static struct cifs_writedata *
1962wdata_alloc_and_fillpages(pgoff_t tofind, struct address_space *mapping,
1963 pgoff_t end, pgoff_t *index,
1964 unsigned int *found_pages)
1965{
90ac1387
PS
1966 struct cifs_writedata *wdata;
1967
1968 wdata = cifs_writedata_alloc((unsigned int)tofind,
1969 cifs_writev_complete);
1970 if (!wdata)
1971 return NULL;
1972
9c19a9cb
JK
1973 *found_pages = find_get_pages_range_tag(mapping, index, end,
1974 PAGECACHE_TAG_DIRTY, tofind, wdata->pages);
90ac1387
PS
1975 return wdata;
1976}
1977
7e48ff82
PS
1978static unsigned int
1979wdata_prepare_pages(struct cifs_writedata *wdata, unsigned int found_pages,
1980 struct address_space *mapping,
1981 struct writeback_control *wbc,
1982 pgoff_t end, pgoff_t *index, pgoff_t *next, bool *done)
1983{
1984 unsigned int nr_pages = 0, i;
1985 struct page *page;
1986
1987 for (i = 0; i < found_pages; i++) {
1988 page = wdata->pages[i];
1989 /*
b93b0163
MW
1990 * At this point we hold neither the i_pages lock nor the
1991 * page lock: the page may be truncated or invalidated
1992 * (changing page->mapping to NULL), or even swizzled
1993 * back from swapper_space to tmpfs file mapping
7e48ff82
PS
1994 */
1995
1996 if (nr_pages == 0)
1997 lock_page(page);
1998 else if (!trylock_page(page))
1999 break;
2000
2001 if (unlikely(page->mapping != mapping)) {
2002 unlock_page(page);
2003 break;
2004 }
2005
2006 if (!wbc->range_cyclic && page->index > end) {
2007 *done = true;
2008 unlock_page(page);
2009 break;
2010 }
2011
2012 if (*next && (page->index != *next)) {
2013 /* Not next consecutive page */
2014 unlock_page(page);
2015 break;
2016 }
2017
2018 if (wbc->sync_mode != WB_SYNC_NONE)
2019 wait_on_page_writeback(page);
2020
2021 if (PageWriteback(page) ||
2022 !clear_page_dirty_for_io(page)) {
2023 unlock_page(page);
2024 break;
2025 }
2026
2027 /*
2028 * This actually clears the dirty bit in the radix tree.
2029 * See cifs_writepage() for more commentary.
2030 */
2031 set_page_writeback(page);
2032 if (page_offset(page) >= i_size_read(mapping->host)) {
2033 *done = true;
2034 unlock_page(page);
2035 end_page_writeback(page);
2036 break;
2037 }
2038
2039 wdata->pages[i] = page;
2040 *next = page->index + 1;
2041 ++nr_pages;
2042 }
2043
2044 /* reset index to refind any pages skipped */
2045 if (nr_pages == 0)
2046 *index = wdata->pages[0]->index + 1;
2047
2048 /* put any pages we aren't going to use */
2049 for (i = nr_pages; i < found_pages; i++) {
09cbfeaf 2050 put_page(wdata->pages[i]);
7e48ff82
PS
2051 wdata->pages[i] = NULL;
2052 }
2053
2054 return nr_pages;
2055}
2056
619aa48e
PS
2057static int
2058wdata_send_pages(struct cifs_writedata *wdata, unsigned int nr_pages,
2059 struct address_space *mapping, struct writeback_control *wbc)
2060{
2061 int rc = 0;
2062 struct TCP_Server_Info *server;
2063 unsigned int i;
2064
2065 wdata->sync_mode = wbc->sync_mode;
2066 wdata->nr_pages = nr_pages;
2067 wdata->offset = page_offset(wdata->pages[0]);
09cbfeaf 2068 wdata->pagesz = PAGE_SIZE;
619aa48e
PS
2069 wdata->tailsz = min(i_size_read(mapping->host) -
2070 page_offset(wdata->pages[nr_pages - 1]),
09cbfeaf
KS
2071 (loff_t)PAGE_SIZE);
2072 wdata->bytes = ((nr_pages - 1) * PAGE_SIZE) + wdata->tailsz;
619aa48e 2073
66231a47
PS
2074 if (wdata->cfile != NULL)
2075 cifsFileInfo_put(wdata->cfile);
2076 wdata->cfile = find_writable_file(CIFS_I(mapping->host), false);
2077 if (!wdata->cfile) {
2078 cifs_dbg(VFS, "No writable handles for inode\n");
2079 rc = -EBADF;
2080 } else {
619aa48e
PS
2081 wdata->pid = wdata->cfile->pid;
2082 server = tlink_tcon(wdata->cfile->tlink)->ses->server;
2083 rc = server->ops->async_writev(wdata, cifs_writedata_release);
66231a47 2084 }
619aa48e
PS
2085
2086 for (i = 0; i < nr_pages; ++i)
2087 unlock_page(wdata->pages[i]);
2088
2089 return rc;
2090}
2091
1da177e4 2092static int cifs_writepages(struct address_space *mapping,
37c0eb46 2093 struct writeback_control *wbc)
1da177e4 2094{
c3d17b63 2095 struct cifs_sb_info *cifs_sb = CIFS_SB(mapping->host->i_sb);
cb7e9eab 2096 struct TCP_Server_Info *server;
c3d17b63
JL
2097 bool done = false, scanned = false, range_whole = false;
2098 pgoff_t end, index;
2099 struct cifs_writedata *wdata;
37c0eb46 2100 int rc = 0;
50c2f753 2101
37c0eb46 2102 /*
c3d17b63 2103 * If wsize is smaller than the page cache size, default to writing
37c0eb46
SF
2104 * one page at a time via cifs_writepage
2105 */
09cbfeaf 2106 if (cifs_sb->wsize < PAGE_SIZE)
37c0eb46
SF
2107 return generic_writepages(mapping, wbc);
2108
111ebb6e 2109 if (wbc->range_cyclic) {
37c0eb46 2110 index = mapping->writeback_index; /* Start from prev offset */
111ebb6e
OH
2111 end = -1;
2112 } else {
09cbfeaf
KS
2113 index = wbc->range_start >> PAGE_SHIFT;
2114 end = wbc->range_end >> PAGE_SHIFT;
111ebb6e 2115 if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX)
c3d17b63
JL
2116 range_whole = true;
2117 scanned = true;
37c0eb46 2118 }
cb7e9eab 2119 server = cifs_sb_master_tcon(cifs_sb)->ses->server;
37c0eb46 2120retry:
c3d17b63 2121 while (!done && index <= end) {
cb7e9eab 2122 unsigned int i, nr_pages, found_pages, wsize, credits;
66231a47 2123 pgoff_t next = 0, tofind, saved_index = index;
c3d17b63 2124
cb7e9eab
PS
2125 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
2126 &wsize, &credits);
2127 if (rc)
2128 break;
c3d17b63 2129
09cbfeaf 2130 tofind = min((wsize / PAGE_SIZE) - 1, end - index) + 1;
c3d17b63 2131
90ac1387
PS
2132 wdata = wdata_alloc_and_fillpages(tofind, mapping, end, &index,
2133 &found_pages);
c3d17b63
JL
2134 if (!wdata) {
2135 rc = -ENOMEM;
cb7e9eab 2136 add_credits_and_wake_if(server, credits, 0);
c3d17b63
JL
2137 break;
2138 }
2139
c3d17b63
JL
2140 if (found_pages == 0) {
2141 kref_put(&wdata->refcount, cifs_writedata_release);
cb7e9eab 2142 add_credits_and_wake_if(server, credits, 0);
c3d17b63
JL
2143 break;
2144 }
2145
7e48ff82
PS
2146 nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc,
2147 end, &index, &next, &done);
37c0eb46 2148
c3d17b63
JL
2149 /* nothing to write? */
2150 if (nr_pages == 0) {
2151 kref_put(&wdata->refcount, cifs_writedata_release);
cb7e9eab 2152 add_credits_and_wake_if(server, credits, 0);
c3d17b63 2153 continue;
37c0eb46 2154 }
fbec9ab9 2155
cb7e9eab 2156 wdata->credits = credits;
941b853d 2157
619aa48e 2158 rc = wdata_send_pages(wdata, nr_pages, mapping, wbc);
f3983c21 2159
c3d17b63
JL
2160 /* send failure -- clean up the mess */
2161 if (rc != 0) {
cb7e9eab 2162 add_credits_and_wake_if(server, wdata->credits, 0);
c3d17b63 2163 for (i = 0; i < nr_pages; ++i) {
941b853d 2164 if (rc == -EAGAIN)
c3d17b63
JL
2165 redirty_page_for_writepage(wbc,
2166 wdata->pages[i]);
2167 else
2168 SetPageError(wdata->pages[i]);
2169 end_page_writeback(wdata->pages[i]);
09cbfeaf 2170 put_page(wdata->pages[i]);
37c0eb46 2171 }
941b853d
JL
2172 if (rc != -EAGAIN)
2173 mapping_set_error(mapping, rc);
c3d17b63
JL
2174 }
2175 kref_put(&wdata->refcount, cifs_writedata_release);
941b853d 2176
66231a47
PS
2177 if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN) {
2178 index = saved_index;
2179 continue;
2180 }
2181
c3d17b63
JL
2182 wbc->nr_to_write -= nr_pages;
2183 if (wbc->nr_to_write <= 0)
2184 done = true;
b066a48c 2185
c3d17b63 2186 index = next;
37c0eb46 2187 }
c3d17b63 2188
37c0eb46
SF
2189 if (!scanned && !done) {
2190 /*
2191 * We hit the last page and there is more work to be done: wrap
2192 * back to the start of the file
2193 */
c3d17b63 2194 scanned = true;
37c0eb46
SF
2195 index = 0;
2196 goto retry;
2197 }
c3d17b63 2198
111ebb6e 2199 if (wbc->range_cyclic || (range_whole && wbc->nr_to_write > 0))
37c0eb46
SF
2200 mapping->writeback_index = index;
2201
1da177e4
LT
2202 return rc;
2203}
1da177e4 2204
9ad1506b
PS
2205static int
2206cifs_writepage_locked(struct page *page, struct writeback_control *wbc)
1da177e4 2207{
9ad1506b 2208 int rc;
6d5786a3 2209 unsigned int xid;
1da177e4 2210
6d5786a3 2211 xid = get_xid();
1da177e4 2212/* BB add check for wbc flags */
09cbfeaf 2213 get_page(page);
ad7a2926 2214 if (!PageUptodate(page))
f96637be 2215 cifs_dbg(FYI, "ppw - page not up to date\n");
cb876f45
LT
2216
2217 /*
2218 * Set the "writeback" flag, and clear "dirty" in the radix tree.
2219 *
2220 * A writepage() implementation always needs to do either this,
2221 * or re-dirty the page with "redirty_page_for_writepage()" in
2222 * the case of a failure.
2223 *
2224 * Just unlocking the page will cause the radix tree tag-bits
2225 * to fail to update with the state of the page correctly.
2226 */
fb8c4b14 2227 set_page_writeback(page);
9ad1506b 2228retry_write:
09cbfeaf 2229 rc = cifs_partialpagewrite(page, 0, PAGE_SIZE);
97b37f24
JL
2230 if (rc == -EAGAIN) {
2231 if (wbc->sync_mode == WB_SYNC_ALL)
2232 goto retry_write;
9ad1506b 2233 redirty_page_for_writepage(wbc, page);
97b37f24 2234 } else if (rc != 0) {
9ad1506b 2235 SetPageError(page);
97b37f24
JL
2236 mapping_set_error(page->mapping, rc);
2237 } else {
9ad1506b 2238 SetPageUptodate(page);
97b37f24 2239 }
cb876f45 2240 end_page_writeback(page);
09cbfeaf 2241 put_page(page);
6d5786a3 2242 free_xid(xid);
1da177e4
LT
2243 return rc;
2244}
2245
9ad1506b
PS
2246static int cifs_writepage(struct page *page, struct writeback_control *wbc)
2247{
2248 int rc = cifs_writepage_locked(page, wbc);
2249 unlock_page(page);
2250 return rc;
2251}
2252
d9414774
NP
2253static int cifs_write_end(struct file *file, struct address_space *mapping,
2254 loff_t pos, unsigned len, unsigned copied,
2255 struct page *page, void *fsdata)
1da177e4 2256{
d9414774
NP
2257 int rc;
2258 struct inode *inode = mapping->host;
d4ffff1f
PS
2259 struct cifsFileInfo *cfile = file->private_data;
2260 struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
2261 __u32 pid;
2262
2263 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2264 pid = cfile->pid;
2265 else
2266 pid = current->tgid;
1da177e4 2267
f96637be 2268 cifs_dbg(FYI, "write_end for page %p from pos %lld with %d bytes\n",
b6b38f70 2269 page, pos, copied);
d9414774 2270
a98ee8c1
JL
2271 if (PageChecked(page)) {
2272 if (copied == len)
2273 SetPageUptodate(page);
2274 ClearPageChecked(page);
09cbfeaf 2275 } else if (!PageUptodate(page) && copied == PAGE_SIZE)
d9414774 2276 SetPageUptodate(page);
ad7a2926 2277
1da177e4 2278 if (!PageUptodate(page)) {
d9414774 2279 char *page_data;
09cbfeaf 2280 unsigned offset = pos & (PAGE_SIZE - 1);
6d5786a3 2281 unsigned int xid;
d9414774 2282
6d5786a3 2283 xid = get_xid();
1da177e4
LT
2284 /* this is probably better than directly calling
2285 partialpage_write since in this function the file handle is
2286 known which we might as well leverage */
2287 /* BB check if anything else missing out of ppw
2288 such as updating last write time */
2289 page_data = kmap(page);
d4ffff1f 2290 rc = cifs_write(cfile, pid, page_data + offset, copied, &pos);
d9414774 2291 /* if (rc < 0) should we set writebehind rc? */
1da177e4 2292 kunmap(page);
d9414774 2293
6d5786a3 2294 free_xid(xid);
fb8c4b14 2295 } else {
d9414774
NP
2296 rc = copied;
2297 pos += copied;
ca8aa29c 2298 set_page_dirty(page);
1da177e4
LT
2299 }
2300
d9414774
NP
2301 if (rc > 0) {
2302 spin_lock(&inode->i_lock);
2303 if (pos > inode->i_size)
2304 i_size_write(inode, pos);
2305 spin_unlock(&inode->i_lock);
2306 }
2307
2308 unlock_page(page);
09cbfeaf 2309 put_page(page);
d9414774 2310
1da177e4
LT
2311 return rc;
2312}
2313
02c24a82
JB
2314int cifs_strict_fsync(struct file *file, loff_t start, loff_t end,
2315 int datasync)
1da177e4 2316{
6d5786a3 2317 unsigned int xid;
1da177e4 2318 int rc = 0;
96daf2b0 2319 struct cifs_tcon *tcon;
1d8c4c00 2320 struct TCP_Server_Info *server;
c21dfb69 2321 struct cifsFileInfo *smbfile = file->private_data;
496ad9aa 2322 struct inode *inode = file_inode(file);
8be7e6ba 2323 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1da177e4 2324
3b49c9a1 2325 rc = file_write_and_wait_range(file, start, end);
02c24a82
JB
2326 if (rc)
2327 return rc;
5955102c 2328 inode_lock(inode);
02c24a82 2329
6d5786a3 2330 xid = get_xid();
1da177e4 2331
35c265e0
AV
2332 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2333 file, datasync);
50c2f753 2334
18cceb6a 2335 if (!CIFS_CACHE_READ(CIFS_I(inode))) {
4f73c7d3 2336 rc = cifs_zap_mapping(inode);
6feb9891 2337 if (rc) {
f96637be 2338 cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc);
6feb9891
PS
2339 rc = 0; /* don't care about it in fsync */
2340 }
2341 }
eb4b756b 2342
8be7e6ba 2343 tcon = tlink_tcon(smbfile->tlink);
1d8c4c00
PS
2344 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2345 server = tcon->ses->server;
2346 if (server->ops->flush)
2347 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2348 else
2349 rc = -ENOSYS;
2350 }
8be7e6ba 2351
6d5786a3 2352 free_xid(xid);
5955102c 2353 inode_unlock(inode);
8be7e6ba
PS
2354 return rc;
2355}
2356
02c24a82 2357int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
8be7e6ba 2358{
6d5786a3 2359 unsigned int xid;
8be7e6ba 2360 int rc = 0;
96daf2b0 2361 struct cifs_tcon *tcon;
1d8c4c00 2362 struct TCP_Server_Info *server;
8be7e6ba 2363 struct cifsFileInfo *smbfile = file->private_data;
7119e220 2364 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
02c24a82
JB
2365 struct inode *inode = file->f_mapping->host;
2366
3b49c9a1 2367 rc = file_write_and_wait_range(file, start, end);
02c24a82
JB
2368 if (rc)
2369 return rc;
5955102c 2370 inode_lock(inode);
8be7e6ba 2371
6d5786a3 2372 xid = get_xid();
8be7e6ba 2373
35c265e0
AV
2374 cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
2375 file, datasync);
8be7e6ba
PS
2376
2377 tcon = tlink_tcon(smbfile->tlink);
1d8c4c00
PS
2378 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
2379 server = tcon->ses->server;
2380 if (server->ops->flush)
2381 rc = server->ops->flush(xid, tcon, &smbfile->fid);
2382 else
2383 rc = -ENOSYS;
2384 }
b298f223 2385
6d5786a3 2386 free_xid(xid);
5955102c 2387 inode_unlock(inode);
1da177e4
LT
2388 return rc;
2389}
2390
1da177e4
LT
2391/*
2392 * As file closes, flush all cached write data for this inode checking
2393 * for write behind errors.
2394 */
75e1fcc0 2395int cifs_flush(struct file *file, fl_owner_t id)
1da177e4 2396{
496ad9aa 2397 struct inode *inode = file_inode(file);
1da177e4
LT
2398 int rc = 0;
2399
eb4b756b 2400 if (file->f_mode & FMODE_WRITE)
d3f1322a 2401 rc = filemap_write_and_wait(inode->i_mapping);
50c2f753 2402
f96637be 2403 cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc);
1da177e4
LT
2404
2405 return rc;
2406}
2407
72432ffc
PS
2408static int
2409cifs_write_allocate_pages(struct page **pages, unsigned long num_pages)
2410{
2411 int rc = 0;
2412 unsigned long i;
2413
2414 for (i = 0; i < num_pages; i++) {
e94f7ba1 2415 pages[i] = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
72432ffc
PS
2416 if (!pages[i]) {
2417 /*
2418 * save number of pages we have already allocated and
2419 * return with ENOMEM error
2420 */
2421 num_pages = i;
2422 rc = -ENOMEM;
e94f7ba1 2423 break;
72432ffc
PS
2424 }
2425 }
2426
e94f7ba1
JL
2427 if (rc) {
2428 for (i = 0; i < num_pages; i++)
2429 put_page(pages[i]);
2430 }
72432ffc
PS
2431 return rc;
2432}
2433
2434static inline
2435size_t get_numpages(const size_t wsize, const size_t len, size_t *cur_len)
2436{
2437 size_t num_pages;
2438 size_t clen;
2439
2440 clen = min_t(const size_t, len, wsize);
a7103b99 2441 num_pages = DIV_ROUND_UP(clen, PAGE_SIZE);
72432ffc
PS
2442
2443 if (cur_len)
2444 *cur_len = clen;
2445
2446 return num_pages;
2447}
2448
da82f7e7 2449static void
4a5c80d7 2450cifs_uncached_writedata_release(struct kref *refcount)
da82f7e7
JL
2451{
2452 int i;
4a5c80d7
SF
2453 struct cifs_writedata *wdata = container_of(refcount,
2454 struct cifs_writedata, refcount);
2455
c610c4b6 2456 kref_put(&wdata->ctx->refcount, cifs_aio_ctx_release);
4a5c80d7
SF
2457 for (i = 0; i < wdata->nr_pages; i++)
2458 put_page(wdata->pages[i]);
2459 cifs_writedata_release(refcount);
2460}
2461
c610c4b6
PS
2462static void collect_uncached_write_data(struct cifs_aio_ctx *ctx);
2463
4a5c80d7
SF
2464static void
2465cifs_uncached_writev_complete(struct work_struct *work)
2466{
da82f7e7
JL
2467 struct cifs_writedata *wdata = container_of(work,
2468 struct cifs_writedata, work);
2b0143b5 2469 struct inode *inode = d_inode(wdata->cfile->dentry);
da82f7e7
JL
2470 struct cifsInodeInfo *cifsi = CIFS_I(inode);
2471
2472 spin_lock(&inode->i_lock);
2473 cifs_update_eof(cifsi, wdata->offset, wdata->bytes);
2474 if (cifsi->server_eof > inode->i_size)
2475 i_size_write(inode, cifsi->server_eof);
2476 spin_unlock(&inode->i_lock);
2477
2478 complete(&wdata->done);
c610c4b6
PS
2479 collect_uncached_write_data(wdata->ctx);
2480 /* the below call can possibly free the last ref to aio ctx */
4a5c80d7 2481 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
da82f7e7
JL
2482}
2483
da82f7e7 2484static int
66386c08
PS
2485wdata_fill_from_iovec(struct cifs_writedata *wdata, struct iov_iter *from,
2486 size_t *len, unsigned long *num_pages)
da82f7e7 2487{
66386c08
PS
2488 size_t save_len, copied, bytes, cur_len = *len;
2489 unsigned long i, nr_pages = *num_pages;
c9de5c80 2490
66386c08
PS
2491 save_len = cur_len;
2492 for (i = 0; i < nr_pages; i++) {
2493 bytes = min_t(const size_t, cur_len, PAGE_SIZE);
2494 copied = copy_page_from_iter(wdata->pages[i], 0, bytes, from);
2495 cur_len -= copied;
2496 /*
2497 * If we didn't copy as much as we expected, then that
2498 * may mean we trod into an unmapped area. Stop copying
2499 * at that point. On the next pass through the big
2500 * loop, we'll likely end up getting a zero-length
2501 * write and bailing out of it.
2502 */
2503 if (copied < bytes)
2504 break;
2505 }
2506 cur_len = save_len - cur_len;
2507 *len = cur_len;
da82f7e7 2508
66386c08
PS
2509 /*
2510 * If we have no data to send, then that probably means that
2511 * the copy above failed altogether. That's most likely because
2512 * the address in the iovec was bogus. Return -EFAULT and let
2513 * the caller free anything we allocated and bail out.
2514 */
2515 if (!cur_len)
2516 return -EFAULT;
da82f7e7 2517
66386c08
PS
2518 /*
2519 * i + 1 now represents the number of pages we actually used in
2520 * the copy phase above.
2521 */
2522 *num_pages = i + 1;
2523 return 0;
da82f7e7
JL
2524}
2525
43de94ea
PS
2526static int
2527cifs_write_from_iter(loff_t offset, size_t len, struct iov_iter *from,
2528 struct cifsFileInfo *open_file,
c610c4b6
PS
2529 struct cifs_sb_info *cifs_sb, struct list_head *wdata_list,
2530 struct cifs_aio_ctx *ctx)
72432ffc 2531{
43de94ea
PS
2532 int rc = 0;
2533 size_t cur_len;
66386c08 2534 unsigned long nr_pages, num_pages, i;
43de94ea 2535 struct cifs_writedata *wdata;
fc56b983 2536 struct iov_iter saved_from = *from;
6ec0b01b 2537 loff_t saved_offset = offset;
da82f7e7 2538 pid_t pid;
6ec0b01b 2539 struct TCP_Server_Info *server;
d4ffff1f
PS
2540
2541 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
2542 pid = open_file->pid;
2543 else
2544 pid = current->tgid;
2545
6ec0b01b 2546 server = tlink_tcon(open_file->tlink)->ses->server;
6ec0b01b 2547
72432ffc 2548 do {
cb7e9eab
PS
2549 unsigned int wsize, credits;
2550
2551 rc = server->ops->wait_mtu_credits(server, cifs_sb->wsize,
2552 &wsize, &credits);
2553 if (rc)
2554 break;
da82f7e7 2555
cb7e9eab 2556 nr_pages = get_numpages(wsize, len, &cur_len);
da82f7e7
JL
2557 wdata = cifs_writedata_alloc(nr_pages,
2558 cifs_uncached_writev_complete);
2559 if (!wdata) {
2560 rc = -ENOMEM;
cb7e9eab 2561 add_credits_and_wake_if(server, credits, 0);
da82f7e7
JL
2562 break;
2563 }
2564
2565 rc = cifs_write_allocate_pages(wdata->pages, nr_pages);
2566 if (rc) {
2567 kfree(wdata);
cb7e9eab 2568 add_credits_and_wake_if(server, credits, 0);
da82f7e7
JL
2569 break;
2570 }
2571
66386c08
PS
2572 num_pages = nr_pages;
2573 rc = wdata_fill_from_iovec(wdata, from, &cur_len, &num_pages);
2574 if (rc) {
5d81de8e
JL
2575 for (i = 0; i < nr_pages; i++)
2576 put_page(wdata->pages[i]);
2577 kfree(wdata);
cb7e9eab 2578 add_credits_and_wake_if(server, credits, 0);
5d81de8e
JL
2579 break;
2580 }
2581
2582 /*
66386c08
PS
2583 * Bring nr_pages down to the number of pages we actually used,
2584 * and free any pages that we didn't use.
5d81de8e 2585 */
66386c08 2586 for ( ; nr_pages > num_pages; nr_pages--)
5d81de8e
JL
2587 put_page(wdata->pages[nr_pages - 1]);
2588
da82f7e7
JL
2589 wdata->sync_mode = WB_SYNC_ALL;
2590 wdata->nr_pages = nr_pages;
2591 wdata->offset = (__u64)offset;
2592 wdata->cfile = cifsFileInfo_get(open_file);
2593 wdata->pid = pid;
2594 wdata->bytes = cur_len;
eddb079d
JL
2595 wdata->pagesz = PAGE_SIZE;
2596 wdata->tailsz = cur_len - ((nr_pages - 1) * PAGE_SIZE);
cb7e9eab 2597 wdata->credits = credits;
c610c4b6
PS
2598 wdata->ctx = ctx;
2599 kref_get(&ctx->refcount);
6ec0b01b
PS
2600
2601 if (!wdata->cfile->invalidHandle ||
1fa839b4 2602 !(rc = cifs_reopen_file(wdata->cfile, false)))
6ec0b01b
PS
2603 rc = server->ops->async_writev(wdata,
2604 cifs_uncached_writedata_release);
da82f7e7 2605 if (rc) {
cb7e9eab 2606 add_credits_and_wake_if(server, wdata->credits, 0);
4a5c80d7
SF
2607 kref_put(&wdata->refcount,
2608 cifs_uncached_writedata_release);
6ec0b01b 2609 if (rc == -EAGAIN) {
fc56b983 2610 *from = saved_from;
6ec0b01b
PS
2611 iov_iter_advance(from, offset - saved_offset);
2612 continue;
2613 }
72432ffc
PS
2614 break;
2615 }
2616
43de94ea 2617 list_add_tail(&wdata->list, wdata_list);
da82f7e7
JL
2618 offset += cur_len;
2619 len -= cur_len;
72432ffc
PS
2620 } while (len > 0);
2621
43de94ea
PS
2622 return rc;
2623}
2624
c610c4b6 2625static void collect_uncached_write_data(struct cifs_aio_ctx *ctx)
43de94ea 2626{
c610c4b6 2627 struct cifs_writedata *wdata, *tmp;
43de94ea
PS
2628 struct cifs_tcon *tcon;
2629 struct cifs_sb_info *cifs_sb;
c610c4b6
PS
2630 struct dentry *dentry = ctx->cfile->dentry;
2631 unsigned int i;
43de94ea
PS
2632 int rc;
2633
c610c4b6
PS
2634 tcon = tlink_tcon(ctx->cfile->tlink);
2635 cifs_sb = CIFS_SB(dentry->d_sb);
43de94ea 2636
c610c4b6 2637 mutex_lock(&ctx->aio_mutex);
43de94ea 2638
c610c4b6
PS
2639 if (list_empty(&ctx->list)) {
2640 mutex_unlock(&ctx->aio_mutex);
2641 return;
2642 }
da82f7e7 2643
c610c4b6 2644 rc = ctx->rc;
da82f7e7
JL
2645 /*
2646 * Wait for and collect replies for any successful sends in order of
c610c4b6
PS
2647 * increasing offset. Once an error is hit, then return without waiting
2648 * for any more replies.
da82f7e7
JL
2649 */
2650restart_loop:
c610c4b6 2651 list_for_each_entry_safe(wdata, tmp, &ctx->list, list) {
da82f7e7 2652 if (!rc) {
c610c4b6
PS
2653 if (!try_wait_for_completion(&wdata->done)) {
2654 mutex_unlock(&ctx->aio_mutex);
2655 return;
2656 }
2657
2658 if (wdata->result)
da82f7e7
JL
2659 rc = wdata->result;
2660 else
c610c4b6 2661 ctx->total_len += wdata->bytes;
da82f7e7
JL
2662
2663 /* resend call if it's a retryable error */
2664 if (rc == -EAGAIN) {
6ec0b01b 2665 struct list_head tmp_list;
c610c4b6 2666 struct iov_iter tmp_from = ctx->iter;
6ec0b01b
PS
2667
2668 INIT_LIST_HEAD(&tmp_list);
2669 list_del_init(&wdata->list);
2670
6ec0b01b 2671 iov_iter_advance(&tmp_from,
c610c4b6 2672 wdata->offset - ctx->pos);
6ec0b01b
PS
2673
2674 rc = cifs_write_from_iter(wdata->offset,
2675 wdata->bytes, &tmp_from,
c610c4b6
PS
2676 ctx->cfile, cifs_sb, &tmp_list,
2677 ctx);
6ec0b01b 2678
c610c4b6 2679 list_splice(&tmp_list, &ctx->list);
6ec0b01b
PS
2680
2681 kref_put(&wdata->refcount,
2682 cifs_uncached_writedata_release);
da82f7e7
JL
2683 goto restart_loop;
2684 }
2685 }
2686 list_del_init(&wdata->list);
4a5c80d7 2687 kref_put(&wdata->refcount, cifs_uncached_writedata_release);
72432ffc
PS
2688 }
2689
c610c4b6
PS
2690 for (i = 0; i < ctx->npages; i++)
2691 put_page(ctx->bv[i].bv_page);
2692
2693 cifs_stats_bytes_written(tcon, ctx->total_len);
2694 set_bit(CIFS_INO_INVALID_MAPPING, &CIFS_I(dentry->d_inode)->flags);
2695
2696 ctx->rc = (rc == 0) ? ctx->total_len : rc;
2697
2698 mutex_unlock(&ctx->aio_mutex);
2699
2700 if (ctx->iocb && ctx->iocb->ki_complete)
2701 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
2702 else
2703 complete(&ctx->done);
2704}
2705
2706ssize_t cifs_user_writev(struct kiocb *iocb, struct iov_iter *from)
2707{
2708 struct file *file = iocb->ki_filp;
2709 ssize_t total_written = 0;
2710 struct cifsFileInfo *cfile;
2711 struct cifs_tcon *tcon;
2712 struct cifs_sb_info *cifs_sb;
2713 struct cifs_aio_ctx *ctx;
2714 struct iov_iter saved_from = *from;
2715 int rc;
2716
2717 /*
2718 * BB - optimize the way when signing is disabled. We can drop this
2719 * extra memory-to-memory copying and use iovec buffers for constructing
2720 * write request.
2721 */
2722
2723 rc = generic_write_checks(iocb, from);
2724 if (rc <= 0)
2725 return rc;
2726
2727 cifs_sb = CIFS_FILE_SB(file);
2728 cfile = file->private_data;
2729 tcon = tlink_tcon(cfile->tlink);
2730
2731 if (!tcon->ses->server->ops->async_writev)
2732 return -ENOSYS;
2733
2734 ctx = cifs_aio_ctx_alloc();
2735 if (!ctx)
2736 return -ENOMEM;
2737
2738 ctx->cfile = cifsFileInfo_get(cfile);
2739
2740 if (!is_sync_kiocb(iocb))
2741 ctx->iocb = iocb;
2742
2743 ctx->pos = iocb->ki_pos;
2744
2745 rc = setup_aio_ctx_iter(ctx, from, WRITE);
2746 if (rc) {
2747 kref_put(&ctx->refcount, cifs_aio_ctx_release);
2748 return rc;
2749 }
2750
2751 /* grab a lock here due to read response handlers can access ctx */
2752 mutex_lock(&ctx->aio_mutex);
2753
2754 rc = cifs_write_from_iter(iocb->ki_pos, ctx->len, &saved_from,
2755 cfile, cifs_sb, &ctx->list, ctx);
2756
2757 /*
2758 * If at least one write was successfully sent, then discard any rc
2759 * value from the later writes. If the other write succeeds, then
2760 * we'll end up returning whatever was written. If it fails, then
2761 * we'll get a new rc value from that.
2762 */
2763 if (!list_empty(&ctx->list))
2764 rc = 0;
2765
2766 mutex_unlock(&ctx->aio_mutex);
2767
2768 if (rc) {
2769 kref_put(&ctx->refcount, cifs_aio_ctx_release);
2770 return rc;
2771 }
2772
2773 if (!is_sync_kiocb(iocb)) {
2774 kref_put(&ctx->refcount, cifs_aio_ctx_release);
2775 return -EIOCBQUEUED;
2776 }
2777
2778 rc = wait_for_completion_killable(&ctx->done);
2779 if (rc) {
2780 mutex_lock(&ctx->aio_mutex);
2781 ctx->rc = rc = -EINTR;
2782 total_written = ctx->total_len;
2783 mutex_unlock(&ctx->aio_mutex);
2784 } else {
2785 rc = ctx->rc;
2786 total_written = ctx->total_len;
2787 }
2788
2789 kref_put(&ctx->refcount, cifs_aio_ctx_release);
2790
e9d1593d
AV
2791 if (unlikely(!total_written))
2792 return rc;
72432ffc 2793
e9d1593d 2794 iocb->ki_pos += total_written;
e9d1593d 2795 return total_written;
72432ffc
PS
2796}
2797
579f9053 2798static ssize_t
3dae8750 2799cifs_writev(struct kiocb *iocb, struct iov_iter *from)
72432ffc 2800{
579f9053
PS
2801 struct file *file = iocb->ki_filp;
2802 struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
2803 struct inode *inode = file->f_mapping->host;
2804 struct cifsInodeInfo *cinode = CIFS_I(inode);
2805 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
5f380c7f 2806 ssize_t rc;
72432ffc 2807
966681c9 2808 inode_lock(inode);
579f9053
PS
2809 /*
2810 * We need to hold the sem to be sure nobody modifies lock list
2811 * with a brlock that prevents writing.
2812 */
2813 down_read(&cinode->lock_sem);
5f380c7f 2814
3309dd04
AV
2815 rc = generic_write_checks(iocb, from);
2816 if (rc <= 0)
5f380c7f
AV
2817 goto out;
2818
5f380c7f 2819 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from),
579f9053 2820 server->vals->exclusive_lock_type, NULL,
5f380c7f 2821 CIFS_WRITE_OP))
3dae8750 2822 rc = __generic_file_write_iter(iocb, from);
5f380c7f
AV
2823 else
2824 rc = -EACCES;
2825out:
966681c9 2826 up_read(&cinode->lock_sem);
5955102c 2827 inode_unlock(inode);
19dfc1f5 2828
e2592217
CH
2829 if (rc > 0)
2830 rc = generic_write_sync(iocb, rc);
579f9053
PS
2831 return rc;
2832}
2833
2834ssize_t
3dae8750 2835cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from)
579f9053 2836{
496ad9aa 2837 struct inode *inode = file_inode(iocb->ki_filp);
579f9053
PS
2838 struct cifsInodeInfo *cinode = CIFS_I(inode);
2839 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2840 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
2841 iocb->ki_filp->private_data;
2842 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
88cf75aa 2843 ssize_t written;
ca8aa29c 2844
c11f1df5
SP
2845 written = cifs_get_writer(cinode);
2846 if (written)
2847 return written;
2848
18cceb6a 2849 if (CIFS_CACHE_WRITE(cinode)) {
88cf75aa
PS
2850 if (cap_unix(tcon->ses) &&
2851 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability))
c11f1df5 2852 && ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
3dae8750 2853 written = generic_file_write_iter(iocb, from);
c11f1df5
SP
2854 goto out;
2855 }
3dae8750 2856 written = cifs_writev(iocb, from);
c11f1df5 2857 goto out;
25078105 2858 }
25078105 2859 /*
ca8aa29c
PS
2860 * For non-oplocked files in strict cache mode we need to write the data
2861 * to the server exactly from the pos to pos+len-1 rather than flush all
2862 * affected pages because it may cause a error with mandatory locks on
2863 * these pages but not on the region from pos to ppos+len-1.
72432ffc 2864 */
3dae8750 2865 written = cifs_user_writev(iocb, from);
18cceb6a 2866 if (written > 0 && CIFS_CACHE_READ(cinode)) {
88cf75aa
PS
2867 /*
2868 * Windows 7 server can delay breaking level2 oplock if a write
2869 * request comes - break it on the client to prevent reading
2870 * an old data.
2871 */
4f73c7d3 2872 cifs_zap_mapping(inode);
f96637be
JP
2873 cifs_dbg(FYI, "Set no oplock for inode=%p after a write operation\n",
2874 inode);
18cceb6a 2875 cinode->oplock = 0;
88cf75aa 2876 }
c11f1df5
SP
2877out:
2878 cifs_put_writer(cinode);
88cf75aa 2879 return written;
72432ffc
PS
2880}
2881
0471ca3f 2882static struct cifs_readdata *
f9f5aca1 2883cifs_readdata_direct_alloc(struct page **pages, work_func_t complete)
0471ca3f
JL
2884{
2885 struct cifs_readdata *rdata;
f4e49cd2 2886
f9f5aca1 2887 rdata = kzalloc(sizeof(*rdata), GFP_KERNEL);
0471ca3f 2888 if (rdata != NULL) {
f9f5aca1 2889 rdata->pages = pages;
6993f74a 2890 kref_init(&rdata->refcount);
1c892549
JL
2891 INIT_LIST_HEAD(&rdata->list);
2892 init_completion(&rdata->done);
0471ca3f 2893 INIT_WORK(&rdata->work, complete);
0471ca3f 2894 }
f4e49cd2 2895
0471ca3f
JL
2896 return rdata;
2897}
2898
f9f5aca1
LL
2899static struct cifs_readdata *
2900cifs_readdata_alloc(unsigned int nr_pages, work_func_t complete)
2901{
2902 struct page **pages =
6396bb22 2903 kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL);
f9f5aca1
LL
2904 struct cifs_readdata *ret = NULL;
2905
2906 if (pages) {
2907 ret = cifs_readdata_direct_alloc(pages, complete);
2908 if (!ret)
2909 kfree(pages);
2910 }
2911
2912 return ret;
2913}
2914
6993f74a
JL
2915void
2916cifs_readdata_release(struct kref *refcount)
0471ca3f 2917{
6993f74a
JL
2918 struct cifs_readdata *rdata = container_of(refcount,
2919 struct cifs_readdata, refcount);
bd3dcc6a
LL
2920#ifdef CONFIG_CIFS_SMB_DIRECT
2921 if (rdata->mr) {
2922 smbd_deregister_mr(rdata->mr);
2923 rdata->mr = NULL;
2924 }
2925#endif
6993f74a
JL
2926 if (rdata->cfile)
2927 cifsFileInfo_put(rdata->cfile);
2928
f9f5aca1 2929 kvfree(rdata->pages);
0471ca3f
JL
2930 kfree(rdata);
2931}
2932
1c892549 2933static int
c5fab6f4 2934cifs_read_allocate_pages(struct cifs_readdata *rdata, unsigned int nr_pages)
1c892549
JL
2935{
2936 int rc = 0;
c5fab6f4 2937 struct page *page;
1c892549
JL
2938 unsigned int i;
2939
c5fab6f4 2940 for (i = 0; i < nr_pages; i++) {
1c892549
JL
2941 page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
2942 if (!page) {
2943 rc = -ENOMEM;
2944 break;
2945 }
c5fab6f4 2946 rdata->pages[i] = page;
1c892549
JL
2947 }
2948
2949 if (rc) {
c5fab6f4
JL
2950 for (i = 0; i < nr_pages; i++) {
2951 put_page(rdata->pages[i]);
2952 rdata->pages[i] = NULL;
1c892549
JL
2953 }
2954 }
2955 return rc;
2956}
2957
2958static void
2959cifs_uncached_readdata_release(struct kref *refcount)
2960{
1c892549
JL
2961 struct cifs_readdata *rdata = container_of(refcount,
2962 struct cifs_readdata, refcount);
c5fab6f4 2963 unsigned int i;
1c892549 2964
6685c5e2 2965 kref_put(&rdata->ctx->refcount, cifs_aio_ctx_release);
c5fab6f4
JL
2966 for (i = 0; i < rdata->nr_pages; i++) {
2967 put_page(rdata->pages[i]);
2968 rdata->pages[i] = NULL;
1c892549
JL
2969 }
2970 cifs_readdata_release(refcount);
2971}
2972
1c892549
JL
2973/**
2974 * cifs_readdata_to_iov - copy data from pages in response to an iovec
2975 * @rdata: the readdata response with list of pages holding data
7f25bba8 2976 * @iter: destination for our data
1c892549
JL
2977 *
2978 * This function copies data from a list of pages in a readdata response into
2979 * an array of iovecs. It will first calculate where the data should go
2980 * based on the info in the readdata and then copy the data into that spot.
2981 */
7f25bba8
AV
2982static int
2983cifs_readdata_to_iov(struct cifs_readdata *rdata, struct iov_iter *iter)
1c892549 2984{
34a54d61 2985 size_t remaining = rdata->got_bytes;
c5fab6f4 2986 unsigned int i;
1c892549 2987
c5fab6f4 2988 for (i = 0; i < rdata->nr_pages; i++) {
c5fab6f4 2989 struct page *page = rdata->pages[i];
e686bd8d 2990 size_t copy = min_t(size_t, remaining, PAGE_SIZE);
9c25702c
PS
2991 size_t written;
2992
2993 if (unlikely(iter->type & ITER_PIPE)) {
2994 void *addr = kmap_atomic(page);
2995
2996 written = copy_to_iter(addr, copy, iter);
2997 kunmap_atomic(addr);
2998 } else
2999 written = copy_page_to_iter(page, 0, copy, iter);
7f25bba8
AV
3000 remaining -= written;
3001 if (written < copy && iov_iter_count(iter) > 0)
3002 break;
1c892549 3003 }
7f25bba8 3004 return remaining ? -EFAULT : 0;
1c892549
JL
3005}
3006
6685c5e2
PS
3007static void collect_uncached_read_data(struct cifs_aio_ctx *ctx);
3008
1c892549
JL
3009static void
3010cifs_uncached_readv_complete(struct work_struct *work)
3011{
3012 struct cifs_readdata *rdata = container_of(work,
3013 struct cifs_readdata, work);
1c892549
JL
3014
3015 complete(&rdata->done);
6685c5e2
PS
3016 collect_uncached_read_data(rdata->ctx);
3017 /* the below call can possibly free the last ref to aio ctx */
1c892549
JL
3018 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
3019}
3020
3021static int
d70b9104
PS
3022uncached_fill_pages(struct TCP_Server_Info *server,
3023 struct cifs_readdata *rdata, struct iov_iter *iter,
3024 unsigned int len)
1c892549 3025{
b3160aeb 3026 int result = 0;
c5fab6f4
JL
3027 unsigned int i;
3028 unsigned int nr_pages = rdata->nr_pages;
1dbe3466 3029 unsigned int page_offset = rdata->page_offset;
1c892549 3030
b3160aeb 3031 rdata->got_bytes = 0;
8321fec4 3032 rdata->tailsz = PAGE_SIZE;
c5fab6f4
JL
3033 for (i = 0; i < nr_pages; i++) {
3034 struct page *page = rdata->pages[i];
71335664 3035 size_t n;
1dbe3466
LL
3036 unsigned int segment_size = rdata->pagesz;
3037
3038 if (i == 0)
3039 segment_size -= page_offset;
3040 else
3041 page_offset = 0;
3042
c5fab6f4 3043
71335664 3044 if (len <= 0) {
1c892549 3045 /* no need to hold page hostage */
c5fab6f4
JL
3046 rdata->pages[i] = NULL;
3047 rdata->nr_pages--;
1c892549 3048 put_page(page);
8321fec4 3049 continue;
1c892549 3050 }
1dbe3466 3051
71335664 3052 n = len;
1dbe3466 3053 if (len >= segment_size)
71335664 3054 /* enough data to fill the page */
1dbe3466
LL
3055 n = segment_size;
3056 else
71335664 3057 rdata->tailsz = len;
1dbe3466
LL
3058 len -= n;
3059
d70b9104 3060 if (iter)
1dbe3466
LL
3061 result = copy_page_from_iter(
3062 page, page_offset, n, iter);
bd3dcc6a
LL
3063#ifdef CONFIG_CIFS_SMB_DIRECT
3064 else if (rdata->mr)
3065 result = n;
3066#endif
d70b9104 3067 else
1dbe3466
LL
3068 result = cifs_read_page_from_socket(
3069 server, page, page_offset, n);
8321fec4
JL
3070 if (result < 0)
3071 break;
3072
b3160aeb 3073 rdata->got_bytes += result;
1c892549
JL
3074 }
3075
b3160aeb
PS
3076 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
3077 rdata->got_bytes : result;
1c892549
JL
3078}
3079
d70b9104
PS
3080static int
3081cifs_uncached_read_into_pages(struct TCP_Server_Info *server,
3082 struct cifs_readdata *rdata, unsigned int len)
3083{
3084 return uncached_fill_pages(server, rdata, NULL, len);
3085}
3086
3087static int
3088cifs_uncached_copy_into_pages(struct TCP_Server_Info *server,
3089 struct cifs_readdata *rdata,
3090 struct iov_iter *iter)
3091{
3092 return uncached_fill_pages(server, rdata, iter, iter->count);
3093}
3094
0ada36b2
PS
3095static int
3096cifs_send_async_read(loff_t offset, size_t len, struct cifsFileInfo *open_file,
6685c5e2
PS
3097 struct cifs_sb_info *cifs_sb, struct list_head *rdata_list,
3098 struct cifs_aio_ctx *ctx)
1da177e4 3099{
0ada36b2 3100 struct cifs_readdata *rdata;
bed9da02 3101 unsigned int npages, rsize, credits;
0ada36b2
PS
3102 size_t cur_len;
3103 int rc;
1c892549 3104 pid_t pid;
25f40259 3105 struct TCP_Server_Info *server;
a70307ee 3106
25f40259 3107 server = tlink_tcon(open_file->tlink)->ses->server;
fc9c5966 3108
d4ffff1f
PS
3109 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3110 pid = open_file->pid;
3111 else
3112 pid = current->tgid;
3113
1c892549 3114 do {
bed9da02
PS
3115 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
3116 &rsize, &credits);
3117 if (rc)
3118 break;
3119
3120 cur_len = min_t(const size_t, len, rsize);
1c892549 3121 npages = DIV_ROUND_UP(cur_len, PAGE_SIZE);
a70307ee 3122
1c892549
JL
3123 /* allocate a readdata struct */
3124 rdata = cifs_readdata_alloc(npages,
3125 cifs_uncached_readv_complete);
3126 if (!rdata) {
bed9da02 3127 add_credits_and_wake_if(server, credits, 0);
1c892549 3128 rc = -ENOMEM;
bae9f746 3129 break;
1da177e4 3130 }
a70307ee 3131
c5fab6f4 3132 rc = cifs_read_allocate_pages(rdata, npages);
1c892549
JL
3133 if (rc)
3134 goto error;
3135
3136 rdata->cfile = cifsFileInfo_get(open_file);
c5fab6f4 3137 rdata->nr_pages = npages;
1c892549
JL
3138 rdata->offset = offset;
3139 rdata->bytes = cur_len;
3140 rdata->pid = pid;
8321fec4 3141 rdata->pagesz = PAGE_SIZE;
1dbe3466 3142 rdata->tailsz = PAGE_SIZE;
8321fec4 3143 rdata->read_into_pages = cifs_uncached_read_into_pages;
d70b9104 3144 rdata->copy_into_pages = cifs_uncached_copy_into_pages;
bed9da02 3145 rdata->credits = credits;
6685c5e2
PS
3146 rdata->ctx = ctx;
3147 kref_get(&ctx->refcount);
1c892549 3148
25f40259 3149 if (!rdata->cfile->invalidHandle ||
1fa839b4 3150 !(rc = cifs_reopen_file(rdata->cfile, true)))
25f40259 3151 rc = server->ops->async_readv(rdata);
1c892549
JL
3152error:
3153 if (rc) {
bed9da02 3154 add_credits_and_wake_if(server, rdata->credits, 0);
1c892549
JL
3155 kref_put(&rdata->refcount,
3156 cifs_uncached_readdata_release);
25f40259
PS
3157 if (rc == -EAGAIN)
3158 continue;
1c892549
JL
3159 break;
3160 }
3161
0ada36b2 3162 list_add_tail(&rdata->list, rdata_list);
1c892549
JL
3163 offset += cur_len;
3164 len -= cur_len;
3165 } while (len > 0);
3166
0ada36b2
PS
3167 return rc;
3168}
3169
6685c5e2
PS
3170static void
3171collect_uncached_read_data(struct cifs_aio_ctx *ctx)
0ada36b2 3172{
6685c5e2
PS
3173 struct cifs_readdata *rdata, *tmp;
3174 struct iov_iter *to = &ctx->iter;
0ada36b2
PS
3175 struct cifs_sb_info *cifs_sb;
3176 struct cifs_tcon *tcon;
6685c5e2
PS
3177 unsigned int i;
3178 int rc;
0ada36b2 3179
6685c5e2
PS
3180 tcon = tlink_tcon(ctx->cfile->tlink);
3181 cifs_sb = CIFS_SB(ctx->cfile->dentry->d_sb);
0ada36b2 3182
6685c5e2 3183 mutex_lock(&ctx->aio_mutex);
0ada36b2 3184
6685c5e2
PS
3185 if (list_empty(&ctx->list)) {
3186 mutex_unlock(&ctx->aio_mutex);
3187 return;
3188 }
1c892549 3189
6685c5e2 3190 rc = ctx->rc;
1c892549 3191 /* the loop below should proceed in the order of increasing offsets */
25f40259 3192again:
6685c5e2 3193 list_for_each_entry_safe(rdata, tmp, &ctx->list, list) {
1c892549 3194 if (!rc) {
6685c5e2
PS
3195 if (!try_wait_for_completion(&rdata->done)) {
3196 mutex_unlock(&ctx->aio_mutex);
3197 return;
3198 }
3199
3200 if (rdata->result == -EAGAIN) {
74027f4a 3201 /* resend call if it's a retryable error */
fb8a3e52 3202 struct list_head tmp_list;
d913ed17 3203 unsigned int got_bytes = rdata->got_bytes;
25f40259 3204
fb8a3e52
PS
3205 list_del_init(&rdata->list);
3206 INIT_LIST_HEAD(&tmp_list);
25f40259 3207
d913ed17
PS
3208 /*
3209 * Got a part of data and then reconnect has
3210 * happened -- fill the buffer and continue
3211 * reading.
3212 */
3213 if (got_bytes && got_bytes < rdata->bytes) {
3214 rc = cifs_readdata_to_iov(rdata, to);
3215 if (rc) {
3216 kref_put(&rdata->refcount,
3217 cifs_uncached_readdata_release);
3218 continue;
3219 }
74027f4a 3220 }
d913ed17
PS
3221
3222 rc = cifs_send_async_read(
3223 rdata->offset + got_bytes,
3224 rdata->bytes - got_bytes,
3225 rdata->cfile, cifs_sb,
6685c5e2 3226 &tmp_list, ctx);
25f40259 3227
6685c5e2 3228 list_splice(&tmp_list, &ctx->list);
25f40259 3229
fb8a3e52
PS
3230 kref_put(&rdata->refcount,
3231 cifs_uncached_readdata_release);
3232 goto again;
3233 } else if (rdata->result)
3234 rc = rdata->result;
3235 else
e6a7bcb4 3236 rc = cifs_readdata_to_iov(rdata, to);
1c892549 3237
2e8a05d8
PS
3238 /* if there was a short read -- discard anything left */
3239 if (rdata->got_bytes && rdata->got_bytes < rdata->bytes)
3240 rc = -ENODATA;
1da177e4 3241 }
1c892549
JL
3242 list_del_init(&rdata->list);
3243 kref_put(&rdata->refcount, cifs_uncached_readdata_release);
1da177e4 3244 }
a70307ee 3245
6685c5e2
PS
3246 for (i = 0; i < ctx->npages; i++) {
3247 if (ctx->should_dirty)
3248 set_page_dirty(ctx->bv[i].bv_page);
3249 put_page(ctx->bv[i].bv_page);
3250 }
7f25bba8 3251
6685c5e2
PS
3252 ctx->total_len = ctx->len - iov_iter_count(to);
3253
3254 cifs_stats_bytes_read(tcon, ctx->total_len);
1c892549 3255
09a4707e
PS
3256 /* mask nodata case */
3257 if (rc == -ENODATA)
3258 rc = 0;
3259
6685c5e2
PS
3260 ctx->rc = (rc == 0) ? ctx->total_len : rc;
3261
3262 mutex_unlock(&ctx->aio_mutex);
3263
3264 if (ctx->iocb && ctx->iocb->ki_complete)
3265 ctx->iocb->ki_complete(ctx->iocb, ctx->rc, 0);
3266 else
3267 complete(&ctx->done);
3268}
3269
3270ssize_t cifs_user_readv(struct kiocb *iocb, struct iov_iter *to)
3271{
3272 struct file *file = iocb->ki_filp;
3273 ssize_t rc;
3274 size_t len;
3275 ssize_t total_read = 0;
3276 loff_t offset = iocb->ki_pos;
3277 struct cifs_sb_info *cifs_sb;
3278 struct cifs_tcon *tcon;
3279 struct cifsFileInfo *cfile;
3280 struct cifs_aio_ctx *ctx;
3281
3282 len = iov_iter_count(to);
3283 if (!len)
3284 return 0;
3285
3286 cifs_sb = CIFS_FILE_SB(file);
3287 cfile = file->private_data;
3288 tcon = tlink_tcon(cfile->tlink);
3289
3290 if (!tcon->ses->server->ops->async_readv)
3291 return -ENOSYS;
3292
3293 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
3294 cifs_dbg(FYI, "attempting read on write only file instance\n");
3295
3296 ctx = cifs_aio_ctx_alloc();
3297 if (!ctx)
3298 return -ENOMEM;
3299
3300 ctx->cfile = cifsFileInfo_get(cfile);
3301
3302 if (!is_sync_kiocb(iocb))
3303 ctx->iocb = iocb;
3304
8a7b0d8e 3305 if (to->type == ITER_IOVEC)
6685c5e2
PS
3306 ctx->should_dirty = true;
3307
3308 rc = setup_aio_ctx_iter(ctx, to, READ);
3309 if (rc) {
3310 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3311 return rc;
3312 }
3313
3314 len = ctx->len;
3315
3316 /* grab a lock here due to read response handlers can access ctx */
3317 mutex_lock(&ctx->aio_mutex);
3318
3319 rc = cifs_send_async_read(offset, len, cfile, cifs_sb, &ctx->list, ctx);
3320
3321 /* if at least one read request send succeeded, then reset rc */
3322 if (!list_empty(&ctx->list))
3323 rc = 0;
3324
3325 mutex_unlock(&ctx->aio_mutex);
3326
3327 if (rc) {
3328 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3329 return rc;
3330 }
3331
3332 if (!is_sync_kiocb(iocb)) {
3333 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3334 return -EIOCBQUEUED;
3335 }
3336
3337 rc = wait_for_completion_killable(&ctx->done);
3338 if (rc) {
3339 mutex_lock(&ctx->aio_mutex);
3340 ctx->rc = rc = -EINTR;
3341 total_read = ctx->total_len;
3342 mutex_unlock(&ctx->aio_mutex);
3343 } else {
3344 rc = ctx->rc;
3345 total_read = ctx->total_len;
3346 }
3347
3348 kref_put(&ctx->refcount, cifs_aio_ctx_release);
3349
0165e810 3350 if (total_read) {
e6a7bcb4 3351 iocb->ki_pos += total_read;
0165e810
AV
3352 return total_read;
3353 }
3354 return rc;
a70307ee
PS
3355}
3356
579f9053 3357ssize_t
e6a7bcb4 3358cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to)
a70307ee 3359{
496ad9aa 3360 struct inode *inode = file_inode(iocb->ki_filp);
579f9053
PS
3361 struct cifsInodeInfo *cinode = CIFS_I(inode);
3362 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3363 struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3364 iocb->ki_filp->private_data;
3365 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
3366 int rc = -EACCES;
a70307ee
PS
3367
3368 /*
3369 * In strict cache mode we need to read from the server all the time
3370 * if we don't have level II oplock because the server can delay mtime
3371 * change - so we can't make a decision about inode invalidating.
3372 * And we can also fail with pagereading if there are mandatory locks
3373 * on pages affected by this read but not on the region from pos to
3374 * pos+len-1.
3375 */
18cceb6a 3376 if (!CIFS_CACHE_READ(cinode))
e6a7bcb4 3377 return cifs_user_readv(iocb, to);
a70307ee 3378
579f9053
PS
3379 if (cap_unix(tcon->ses) &&
3380 (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
3381 ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
e6a7bcb4 3382 return generic_file_read_iter(iocb, to);
579f9053
PS
3383
3384 /*
3385 * We need to hold the sem to be sure nobody modifies lock list
3386 * with a brlock that prevents reading.
3387 */
3388 down_read(&cinode->lock_sem);
e6a7bcb4 3389 if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(to),
579f9053 3390 tcon->ses->server->vals->shared_lock_type,
081c0414 3391 NULL, CIFS_READ_OP))
e6a7bcb4 3392 rc = generic_file_read_iter(iocb, to);
579f9053
PS
3393 up_read(&cinode->lock_sem);
3394 return rc;
a70307ee 3395}
1da177e4 3396
f9c6e234
PS
3397static ssize_t
3398cifs_read(struct file *file, char *read_data, size_t read_size, loff_t *offset)
1da177e4
LT
3399{
3400 int rc = -EACCES;
3401 unsigned int bytes_read = 0;
3402 unsigned int total_read;
3403 unsigned int current_read_size;
5eba8ab3 3404 unsigned int rsize;
1da177e4 3405 struct cifs_sb_info *cifs_sb;
29e20f9c 3406 struct cifs_tcon *tcon;
f9c6e234 3407 struct TCP_Server_Info *server;
6d5786a3 3408 unsigned int xid;
f9c6e234 3409 char *cur_offset;
1da177e4 3410 struct cifsFileInfo *open_file;
d4ffff1f 3411 struct cifs_io_parms io_parms;
ec637e3f 3412 int buf_type = CIFS_NO_BUFFER;
d4ffff1f 3413 __u32 pid;
1da177e4 3414
6d5786a3 3415 xid = get_xid();
7119e220 3416 cifs_sb = CIFS_FILE_SB(file);
1da177e4 3417
5eba8ab3
JL
3418 /* FIXME: set up handlers for larger reads and/or convert to async */
3419 rsize = min_t(unsigned int, cifs_sb->rsize, CIFSMaxBufSize);
3420
1da177e4 3421 if (file->private_data == NULL) {
0f3bc09e 3422 rc = -EBADF;
6d5786a3 3423 free_xid(xid);
0f3bc09e 3424 return rc;
1da177e4 3425 }
c21dfb69 3426 open_file = file->private_data;
29e20f9c 3427 tcon = tlink_tcon(open_file->tlink);
f9c6e234
PS
3428 server = tcon->ses->server;
3429
3430 if (!server->ops->sync_read) {
3431 free_xid(xid);
3432 return -ENOSYS;
3433 }
1da177e4 3434
d4ffff1f
PS
3435 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3436 pid = open_file->pid;
3437 else
3438 pid = current->tgid;
3439
1da177e4 3440 if ((file->f_flags & O_ACCMODE) == O_WRONLY)
f96637be 3441 cifs_dbg(FYI, "attempting read on write only file instance\n");
1da177e4 3442
f9c6e234
PS
3443 for (total_read = 0, cur_offset = read_data; read_size > total_read;
3444 total_read += bytes_read, cur_offset += bytes_read) {
e374d90f
PS
3445 do {
3446 current_read_size = min_t(uint, read_size - total_read,
3447 rsize);
3448 /*
3449 * For windows me and 9x we do not want to request more
3450 * than it negotiated since it will refuse the read
3451 * then.
3452 */
3453 if ((tcon->ses) && !(tcon->ses->capabilities &
29e20f9c 3454 tcon->ses->server->vals->cap_large_files)) {
e374d90f
PS
3455 current_read_size = min_t(uint,
3456 current_read_size, CIFSMaxBufSize);
3457 }
cdff08e7 3458 if (open_file->invalidHandle) {
15886177 3459 rc = cifs_reopen_file(open_file, true);
1da177e4
LT
3460 if (rc != 0)
3461 break;
3462 }
d4ffff1f 3463 io_parms.pid = pid;
29e20f9c 3464 io_parms.tcon = tcon;
f9c6e234 3465 io_parms.offset = *offset;
d4ffff1f 3466 io_parms.length = current_read_size;
db8b631d 3467 rc = server->ops->sync_read(xid, &open_file->fid, &io_parms,
f9c6e234
PS
3468 &bytes_read, &cur_offset,
3469 &buf_type);
e374d90f
PS
3470 } while (rc == -EAGAIN);
3471
1da177e4
LT
3472 if (rc || (bytes_read == 0)) {
3473 if (total_read) {
3474 break;
3475 } else {
6d5786a3 3476 free_xid(xid);
1da177e4
LT
3477 return rc;
3478 }
3479 } else {
29e20f9c 3480 cifs_stats_bytes_read(tcon, total_read);
f9c6e234 3481 *offset += bytes_read;
1da177e4
LT
3482 }
3483 }
6d5786a3 3484 free_xid(xid);
1da177e4
LT
3485 return total_read;
3486}
3487
ca83ce3d
JL
3488/*
3489 * If the page is mmap'ed into a process' page tables, then we need to make
3490 * sure that it doesn't change while being written back.
3491 */
a5240cbd 3492static vm_fault_t
11bac800 3493cifs_page_mkwrite(struct vm_fault *vmf)
ca83ce3d
JL
3494{
3495 struct page *page = vmf->page;
3496
3497 lock_page(page);
3498 return VM_FAULT_LOCKED;
3499}
3500
7cbea8dc 3501static const struct vm_operations_struct cifs_file_vm_ops = {
ca83ce3d 3502 .fault = filemap_fault,
f1820361 3503 .map_pages = filemap_map_pages,
ca83ce3d
JL
3504 .page_mkwrite = cifs_page_mkwrite,
3505};
3506
7a6a19b1
PS
3507int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma)
3508{
f04a703c 3509 int xid, rc = 0;
496ad9aa 3510 struct inode *inode = file_inode(file);
7a6a19b1 3511
6d5786a3 3512 xid = get_xid();
7a6a19b1 3513
f04a703c 3514 if (!CIFS_CACHE_READ(CIFS_I(inode)))
4f73c7d3 3515 rc = cifs_zap_mapping(inode);
f04a703c
MW
3516 if (!rc)
3517 rc = generic_file_mmap(file, vma);
3518 if (!rc)
ca83ce3d 3519 vma->vm_ops = &cifs_file_vm_ops;
f04a703c 3520
6d5786a3 3521 free_xid(xid);
7a6a19b1
PS
3522 return rc;
3523}
3524
1da177e4
LT
3525int cifs_file_mmap(struct file *file, struct vm_area_struct *vma)
3526{
1da177e4
LT
3527 int rc, xid;
3528
6d5786a3 3529 xid = get_xid();
f04a703c 3530
abab095d 3531 rc = cifs_revalidate_file(file);
f04a703c 3532 if (rc)
f96637be
JP
3533 cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n",
3534 rc);
f04a703c
MW
3535 if (!rc)
3536 rc = generic_file_mmap(file, vma);
3537 if (!rc)
ca83ce3d 3538 vma->vm_ops = &cifs_file_vm_ops;
f04a703c 3539
6d5786a3 3540 free_xid(xid);
1da177e4
LT
3541 return rc;
3542}
3543
0471ca3f
JL
3544static void
3545cifs_readv_complete(struct work_struct *work)
3546{
b770ddfa 3547 unsigned int i, got_bytes;
0471ca3f
JL
3548 struct cifs_readdata *rdata = container_of(work,
3549 struct cifs_readdata, work);
0471ca3f 3550
b770ddfa 3551 got_bytes = rdata->got_bytes;
c5fab6f4
JL
3552 for (i = 0; i < rdata->nr_pages; i++) {
3553 struct page *page = rdata->pages[i];
3554
0471ca3f
JL
3555 lru_cache_add_file(page);
3556
b770ddfa
PS
3557 if (rdata->result == 0 ||
3558 (rdata->result == -EAGAIN && got_bytes)) {
0471ca3f
JL
3559 flush_dcache_page(page);
3560 SetPageUptodate(page);
3561 }
3562
3563 unlock_page(page);
3564
b770ddfa
PS
3565 if (rdata->result == 0 ||
3566 (rdata->result == -EAGAIN && got_bytes))
0471ca3f
JL
3567 cifs_readpage_to_fscache(rdata->mapping->host, page);
3568
09cbfeaf 3569 got_bytes -= min_t(unsigned int, PAGE_SIZE, got_bytes);
b770ddfa 3570
09cbfeaf 3571 put_page(page);
c5fab6f4 3572 rdata->pages[i] = NULL;
0471ca3f 3573 }
6993f74a 3574 kref_put(&rdata->refcount, cifs_readdata_release);
0471ca3f
JL
3575}
3576
8d5ce4d2 3577static int
d70b9104
PS
3578readpages_fill_pages(struct TCP_Server_Info *server,
3579 struct cifs_readdata *rdata, struct iov_iter *iter,
3580 unsigned int len)
8d5ce4d2 3581{
b3160aeb 3582 int result = 0;
c5fab6f4 3583 unsigned int i;
8d5ce4d2
JL
3584 u64 eof;
3585 pgoff_t eof_index;
c5fab6f4 3586 unsigned int nr_pages = rdata->nr_pages;
1dbe3466 3587 unsigned int page_offset = rdata->page_offset;
8d5ce4d2
JL
3588
3589 /* determine the eof that the server (probably) has */
3590 eof = CIFS_I(rdata->mapping->host)->server_eof;
09cbfeaf 3591 eof_index = eof ? (eof - 1) >> PAGE_SHIFT : 0;
f96637be 3592 cifs_dbg(FYI, "eof=%llu eof_index=%lu\n", eof, eof_index);
8d5ce4d2 3593
b3160aeb 3594 rdata->got_bytes = 0;
09cbfeaf 3595 rdata->tailsz = PAGE_SIZE;
c5fab6f4
JL
3596 for (i = 0; i < nr_pages; i++) {
3597 struct page *page = rdata->pages[i];
1dbe3466
LL
3598 unsigned int to_read = rdata->pagesz;
3599 size_t n;
3600
3601 if (i == 0)
3602 to_read -= page_offset;
3603 else
3604 page_offset = 0;
3605
3606 n = to_read;
c5fab6f4 3607
1dbe3466
LL
3608 if (len >= to_read) {
3609 len -= to_read;
8321fec4 3610 } else if (len > 0) {
8d5ce4d2 3611 /* enough for partial page, fill and zero the rest */
1dbe3466 3612 zero_user(page, len + page_offset, to_read - len);
71335664 3613 n = rdata->tailsz = len;
8321fec4 3614 len = 0;
8d5ce4d2
JL
3615 } else if (page->index > eof_index) {
3616 /*
3617 * The VFS will not try to do readahead past the
3618 * i_size, but it's possible that we have outstanding
3619 * writes with gaps in the middle and the i_size hasn't
3620 * caught up yet. Populate those with zeroed out pages
3621 * to prevent the VFS from repeatedly attempting to
3622 * fill them until the writes are flushed.
3623 */
09cbfeaf 3624 zero_user(page, 0, PAGE_SIZE);
8d5ce4d2
JL
3625 lru_cache_add_file(page);
3626 flush_dcache_page(page);
3627 SetPageUptodate(page);
3628 unlock_page(page);
09cbfeaf 3629 put_page(page);
c5fab6f4
JL
3630 rdata->pages[i] = NULL;
3631 rdata->nr_pages--;
8321fec4 3632 continue;
8d5ce4d2
JL
3633 } else {
3634 /* no need to hold page hostage */
8d5ce4d2
JL
3635 lru_cache_add_file(page);
3636 unlock_page(page);
09cbfeaf 3637 put_page(page);
c5fab6f4
JL
3638 rdata->pages[i] = NULL;
3639 rdata->nr_pages--;
8321fec4 3640 continue;
8d5ce4d2 3641 }
8321fec4 3642
d70b9104 3643 if (iter)
1dbe3466
LL
3644 result = copy_page_from_iter(
3645 page, page_offset, n, iter);
bd3dcc6a
LL
3646#ifdef CONFIG_CIFS_SMB_DIRECT
3647 else if (rdata->mr)
3648 result = n;
3649#endif
d70b9104 3650 else
1dbe3466
LL
3651 result = cifs_read_page_from_socket(
3652 server, page, page_offset, n);
8321fec4
JL
3653 if (result < 0)
3654 break;
3655
b3160aeb 3656 rdata->got_bytes += result;
8d5ce4d2
JL
3657 }
3658
b3160aeb
PS
3659 return rdata->got_bytes > 0 && result != -ECONNABORTED ?
3660 rdata->got_bytes : result;
8d5ce4d2
JL
3661}
3662
d70b9104
PS
3663static int
3664cifs_readpages_read_into_pages(struct TCP_Server_Info *server,
3665 struct cifs_readdata *rdata, unsigned int len)
3666{
3667 return readpages_fill_pages(server, rdata, NULL, len);
3668}
3669
3670static int
3671cifs_readpages_copy_into_pages(struct TCP_Server_Info *server,
3672 struct cifs_readdata *rdata,
3673 struct iov_iter *iter)
3674{
3675 return readpages_fill_pages(server, rdata, iter, iter->count);
3676}
3677
387eb92a
PS
3678static int
3679readpages_get_pages(struct address_space *mapping, struct list_head *page_list,
3680 unsigned int rsize, struct list_head *tmplist,
3681 unsigned int *nr_pages, loff_t *offset, unsigned int *bytes)
3682{
3683 struct page *page, *tpage;
3684 unsigned int expected_index;
3685 int rc;
8a5c743e 3686 gfp_t gfp = readahead_gfp_mask(mapping);
387eb92a 3687
69cebd75
PS
3688 INIT_LIST_HEAD(tmplist);
3689
387eb92a
PS
3690 page = list_entry(page_list->prev, struct page, lru);
3691
3692 /*
3693 * Lock the page and put it in the cache. Since no one else
3694 * should have access to this page, we're safe to simply set
3695 * PG_locked without checking it first.
3696 */
48c935ad 3697 __SetPageLocked(page);
387eb92a 3698 rc = add_to_page_cache_locked(page, mapping,
063d99b4 3699 page->index, gfp);
387eb92a
PS
3700
3701 /* give up if we can't stick it in the cache */
3702 if (rc) {
48c935ad 3703 __ClearPageLocked(page);
387eb92a
PS
3704 return rc;
3705 }
3706
3707 /* move first page to the tmplist */
09cbfeaf
KS
3708 *offset = (loff_t)page->index << PAGE_SHIFT;
3709 *bytes = PAGE_SIZE;
387eb92a
PS
3710 *nr_pages = 1;
3711 list_move_tail(&page->lru, tmplist);
3712
3713 /* now try and add more pages onto the request */
3714 expected_index = page->index + 1;
3715 list_for_each_entry_safe_reverse(page, tpage, page_list, lru) {
3716 /* discontinuity ? */
3717 if (page->index != expected_index)
3718 break;
3719
3720 /* would this page push the read over the rsize? */
09cbfeaf 3721 if (*bytes + PAGE_SIZE > rsize)
387eb92a
PS
3722 break;
3723
48c935ad 3724 __SetPageLocked(page);
063d99b4 3725 if (add_to_page_cache_locked(page, mapping, page->index, gfp)) {
48c935ad 3726 __ClearPageLocked(page);
387eb92a
PS
3727 break;
3728 }
3729 list_move_tail(&page->lru, tmplist);
09cbfeaf 3730 (*bytes) += PAGE_SIZE;
387eb92a
PS
3731 expected_index++;
3732 (*nr_pages)++;
3733 }
3734 return rc;
8d5ce4d2
JL
3735}
3736
1da177e4
LT
3737static int cifs_readpages(struct file *file, struct address_space *mapping,
3738 struct list_head *page_list, unsigned num_pages)
3739{
690c5e31
JL
3740 int rc;
3741 struct list_head tmplist;
3742 struct cifsFileInfo *open_file = file->private_data;
7119e220 3743 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
69cebd75 3744 struct TCP_Server_Info *server;
690c5e31 3745 pid_t pid;
1da177e4 3746
56698236
SJ
3747 /*
3748 * Reads as many pages as possible from fscache. Returns -ENOBUFS
3749 * immediately if the cookie is negative
54afa990
DH
3750 *
3751 * After this point, every page in the list might have PG_fscache set,
3752 * so we will need to clean that up off of every page we don't use.
56698236
SJ
3753 */
3754 rc = cifs_readpages_from_fscache(mapping->host, mapping, page_list,
3755 &num_pages);
3756 if (rc == 0)
690c5e31 3757 return rc;
56698236 3758
d4ffff1f
PS
3759 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3760 pid = open_file->pid;
3761 else
3762 pid = current->tgid;
3763
690c5e31 3764 rc = 0;
69cebd75 3765 server = tlink_tcon(open_file->tlink)->ses->server;
1da177e4 3766
f96637be
JP
3767 cifs_dbg(FYI, "%s: file=%p mapping=%p num_pages=%u\n",
3768 __func__, file, mapping, num_pages);
690c5e31
JL
3769
3770 /*
3771 * Start with the page at end of list and move it to private
3772 * list. Do the same with any following pages until we hit
3773 * the rsize limit, hit an index discontinuity, or run out of
3774 * pages. Issue the async read and then start the loop again
3775 * until the list is empty.
3776 *
3777 * Note that list order is important. The page_list is in
3778 * the order of declining indexes. When we put the pages in
3779 * the rdata->pages, then we want them in increasing order.
3780 */
3781 while (!list_empty(page_list)) {
bed9da02 3782 unsigned int i, nr_pages, bytes, rsize;
690c5e31
JL
3783 loff_t offset;
3784 struct page *page, *tpage;
3785 struct cifs_readdata *rdata;
bed9da02 3786 unsigned credits;
1da177e4 3787
bed9da02
PS
3788 rc = server->ops->wait_mtu_credits(server, cifs_sb->rsize,
3789 &rsize, &credits);
3790 if (rc)
3791 break;
690c5e31
JL
3792
3793 /*
69cebd75
PS
3794 * Give up immediately if rsize is too small to read an entire
3795 * page. The VFS will fall back to readpage. We should never
3796 * reach this point however since we set ra_pages to 0 when the
3797 * rsize is smaller than a cache page.
690c5e31 3798 */
09cbfeaf 3799 if (unlikely(rsize < PAGE_SIZE)) {
bed9da02 3800 add_credits_and_wake_if(server, credits, 0);
69cebd75 3801 return 0;
bed9da02 3802 }
690c5e31 3803
bed9da02
PS
3804 rc = readpages_get_pages(mapping, page_list, rsize, &tmplist,
3805 &nr_pages, &offset, &bytes);
690c5e31 3806 if (rc) {
bed9da02 3807 add_credits_and_wake_if(server, credits, 0);
690c5e31
JL
3808 break;
3809 }
3810
0471ca3f 3811 rdata = cifs_readdata_alloc(nr_pages, cifs_readv_complete);
690c5e31
JL
3812 if (!rdata) {
3813 /* best to give up if we're out of mem */
3814 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
3815 list_del(&page->lru);
3816 lru_cache_add_file(page);
3817 unlock_page(page);
09cbfeaf 3818 put_page(page);
690c5e31
JL
3819 }
3820 rc = -ENOMEM;
bed9da02 3821 add_credits_and_wake_if(server, credits, 0);
690c5e31
JL
3822 break;
3823 }
3824
6993f74a 3825 rdata->cfile = cifsFileInfo_get(open_file);
690c5e31
JL
3826 rdata->mapping = mapping;
3827 rdata->offset = offset;
3828 rdata->bytes = bytes;
3829 rdata->pid = pid;
09cbfeaf 3830 rdata->pagesz = PAGE_SIZE;
1dbe3466 3831 rdata->tailsz = PAGE_SIZE;
8321fec4 3832 rdata->read_into_pages = cifs_readpages_read_into_pages;
d70b9104 3833 rdata->copy_into_pages = cifs_readpages_copy_into_pages;
bed9da02 3834 rdata->credits = credits;
c5fab6f4
JL
3835
3836 list_for_each_entry_safe(page, tpage, &tmplist, lru) {
3837 list_del(&page->lru);
3838 rdata->pages[rdata->nr_pages++] = page;
3839 }
690c5e31 3840
69cebd75 3841 if (!rdata->cfile->invalidHandle ||
1fa839b4 3842 !(rc = cifs_reopen_file(rdata->cfile, true)))
69cebd75
PS
3843 rc = server->ops->async_readv(rdata);
3844 if (rc) {
bed9da02 3845 add_credits_and_wake_if(server, rdata->credits, 0);
c5fab6f4
JL
3846 for (i = 0; i < rdata->nr_pages; i++) {
3847 page = rdata->pages[i];
690c5e31
JL
3848 lru_cache_add_file(page);
3849 unlock_page(page);
09cbfeaf 3850 put_page(page);
1da177e4 3851 }
1209bbdf 3852 /* Fallback to the readpage in error/reconnect cases */
6993f74a 3853 kref_put(&rdata->refcount, cifs_readdata_release);
1da177e4
LT
3854 break;
3855 }
6993f74a
JL
3856
3857 kref_put(&rdata->refcount, cifs_readdata_release);
1da177e4
LT
3858 }
3859
54afa990
DH
3860 /* Any pages that have been shown to fscache but didn't get added to
3861 * the pagecache must be uncached before they get returned to the
3862 * allocator.
3863 */
3864 cifs_fscache_readpages_cancel(mapping->host, page_list);
1da177e4
LT
3865 return rc;
3866}
3867
a9e9b7bc
SP
3868/*
3869 * cifs_readpage_worker must be called with the page pinned
3870 */
1da177e4
LT
3871static int cifs_readpage_worker(struct file *file, struct page *page,
3872 loff_t *poffset)
3873{
3874 char *read_data;
3875 int rc;
3876
56698236 3877 /* Is the page cached? */
496ad9aa 3878 rc = cifs_readpage_from_fscache(file_inode(file), page);
56698236
SJ
3879 if (rc == 0)
3880 goto read_complete;
3881
1da177e4
LT
3882 read_data = kmap(page);
3883 /* for reads over a certain size could initiate async read ahead */
fb8c4b14 3884
09cbfeaf 3885 rc = cifs_read(file, read_data, PAGE_SIZE, poffset);
fb8c4b14 3886
1da177e4
LT
3887 if (rc < 0)
3888 goto io_error;
3889 else
f96637be 3890 cifs_dbg(FYI, "Bytes read %d\n", rc);
fb8c4b14 3891
496ad9aa 3892 file_inode(file)->i_atime =
c2050a45 3893 current_time(file_inode(file));
fb8c4b14 3894
09cbfeaf
KS
3895 if (PAGE_SIZE > rc)
3896 memset(read_data + rc, 0, PAGE_SIZE - rc);
1da177e4
LT
3897
3898 flush_dcache_page(page);
3899 SetPageUptodate(page);
9dc06558
SJ
3900
3901 /* send this page to the cache */
496ad9aa 3902 cifs_readpage_to_fscache(file_inode(file), page);
9dc06558 3903
1da177e4 3904 rc = 0;
fb8c4b14 3905
1da177e4 3906io_error:
fb8c4b14 3907 kunmap(page);
466bd31b 3908 unlock_page(page);
56698236
SJ
3909
3910read_complete:
1da177e4
LT
3911 return rc;
3912}
3913
3914static int cifs_readpage(struct file *file, struct page *page)
3915{
09cbfeaf 3916 loff_t offset = (loff_t)page->index << PAGE_SHIFT;
1da177e4 3917 int rc = -EACCES;
6d5786a3 3918 unsigned int xid;
1da177e4 3919
6d5786a3 3920 xid = get_xid();
1da177e4
LT
3921
3922 if (file->private_data == NULL) {
0f3bc09e 3923 rc = -EBADF;
6d5786a3 3924 free_xid(xid);
0f3bc09e 3925 return rc;
1da177e4
LT
3926 }
3927
f96637be 3928 cifs_dbg(FYI, "readpage %p at offset %d 0x%x\n",
b6b38f70 3929 page, (int)offset, (int)offset);
1da177e4
LT
3930
3931 rc = cifs_readpage_worker(file, page, &offset);
3932
6d5786a3 3933 free_xid(xid);
1da177e4
LT
3934 return rc;
3935}
3936
a403a0a3
SF
3937static int is_inode_writable(struct cifsInodeInfo *cifs_inode)
3938{
3939 struct cifsFileInfo *open_file;
3afca265
SF
3940 struct cifs_tcon *tcon =
3941 cifs_sb_master_tcon(CIFS_SB(cifs_inode->vfs_inode.i_sb));
a403a0a3 3942
3afca265 3943 spin_lock(&tcon->open_file_lock);
a403a0a3 3944 list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2e396b83 3945 if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
3afca265 3946 spin_unlock(&tcon->open_file_lock);
a403a0a3
SF
3947 return 1;
3948 }
3949 }
3afca265 3950 spin_unlock(&tcon->open_file_lock);
a403a0a3
SF
3951 return 0;
3952}
3953
1da177e4
LT
3954/* We do not want to update the file size from server for inodes
3955 open for write - to avoid races with writepage extending
3956 the file - in the future we could consider allowing
fb8c4b14 3957 refreshing the inode only on increases in the file size
1da177e4
LT
3958 but this is tricky to do without racing with writebehind
3959 page caching in the current Linux kernel design */
4b18f2a9 3960bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file)
1da177e4 3961{
a403a0a3 3962 if (!cifsInode)
4b18f2a9 3963 return true;
50c2f753 3964
a403a0a3
SF
3965 if (is_inode_writable(cifsInode)) {
3966 /* This inode is open for write at least once */
c32a0b68
SF
3967 struct cifs_sb_info *cifs_sb;
3968
c32a0b68 3969 cifs_sb = CIFS_SB(cifsInode->vfs_inode.i_sb);
ad7a2926 3970 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) {
fb8c4b14 3971 /* since no page cache to corrupt on directio
c32a0b68 3972 we can change size safely */
4b18f2a9 3973 return true;
c32a0b68
SF
3974 }
3975
fb8c4b14 3976 if (i_size_read(&cifsInode->vfs_inode) < end_of_file)
4b18f2a9 3977 return true;
7ba52631 3978
4b18f2a9 3979 return false;
23e7dd7d 3980 } else
4b18f2a9 3981 return true;
1da177e4
LT
3982}
3983
d9414774
NP
3984static int cifs_write_begin(struct file *file, struct address_space *mapping,
3985 loff_t pos, unsigned len, unsigned flags,
3986 struct page **pagep, void **fsdata)
1da177e4 3987{
466bd31b 3988 int oncethru = 0;
09cbfeaf
KS
3989 pgoff_t index = pos >> PAGE_SHIFT;
3990 loff_t offset = pos & (PAGE_SIZE - 1);
a98ee8c1
JL
3991 loff_t page_start = pos & PAGE_MASK;
3992 loff_t i_size;
3993 struct page *page;
3994 int rc = 0;
d9414774 3995
f96637be 3996 cifs_dbg(FYI, "write_begin from %lld len %d\n", (long long)pos, len);
d9414774 3997
466bd31b 3998start:
54566b2c 3999 page = grab_cache_page_write_begin(mapping, index, flags);
a98ee8c1
JL
4000 if (!page) {
4001 rc = -ENOMEM;
4002 goto out;
4003 }
8a236264 4004
a98ee8c1
JL
4005 if (PageUptodate(page))
4006 goto out;
8a236264 4007
a98ee8c1
JL
4008 /*
4009 * If we write a full page it will be up to date, no need to read from
4010 * the server. If the write is short, we'll end up doing a sync write
4011 * instead.
4012 */
09cbfeaf 4013 if (len == PAGE_SIZE)
a98ee8c1 4014 goto out;
8a236264 4015
a98ee8c1
JL
4016 /*
4017 * optimize away the read when we have an oplock, and we're not
4018 * expecting to use any of the data we'd be reading in. That
4019 * is, when the page lies beyond the EOF, or straddles the EOF
4020 * and the write will cover all of the existing data.
4021 */
18cceb6a 4022 if (CIFS_CACHE_READ(CIFS_I(mapping->host))) {
a98ee8c1
JL
4023 i_size = i_size_read(mapping->host);
4024 if (page_start >= i_size ||
4025 (offset == 0 && (pos + len) >= i_size)) {
4026 zero_user_segments(page, 0, offset,
4027 offset + len,
09cbfeaf 4028 PAGE_SIZE);
a98ee8c1
JL
4029 /*
4030 * PageChecked means that the parts of the page
4031 * to which we're not writing are considered up
4032 * to date. Once the data is copied to the
4033 * page, it can be set uptodate.
4034 */
4035 SetPageChecked(page);
4036 goto out;
4037 }
4038 }
d9414774 4039
466bd31b 4040 if ((file->f_flags & O_ACCMODE) != O_WRONLY && !oncethru) {
a98ee8c1
JL
4041 /*
4042 * might as well read a page, it is fast enough. If we get
4043 * an error, we don't need to return it. cifs_write_end will
4044 * do a sync write instead since PG_uptodate isn't set.
4045 */
4046 cifs_readpage_worker(file, page, &page_start);
09cbfeaf 4047 put_page(page);
466bd31b
SP
4048 oncethru = 1;
4049 goto start;
8a236264
SF
4050 } else {
4051 /* we could try using another file handle if there is one -
4052 but how would we lock it to prevent close of that handle
4053 racing with this read? In any case
d9414774 4054 this will be written out by write_end so is fine */
1da177e4 4055 }
a98ee8c1
JL
4056out:
4057 *pagep = page;
4058 return rc;
1da177e4
LT
4059}
4060
85f2d6b4
SJ
4061static int cifs_release_page(struct page *page, gfp_t gfp)
4062{
4063 if (PagePrivate(page))
4064 return 0;
4065
4066 return cifs_fscache_release_page(page, gfp);
4067}
4068
d47992f8
LC
4069static void cifs_invalidate_page(struct page *page, unsigned int offset,
4070 unsigned int length)
85f2d6b4
SJ
4071{
4072 struct cifsInodeInfo *cifsi = CIFS_I(page->mapping->host);
4073
09cbfeaf 4074 if (offset == 0 && length == PAGE_SIZE)
85f2d6b4
SJ
4075 cifs_fscache_invalidate_page(page, &cifsi->vfs_inode);
4076}
4077
9ad1506b
PS
4078static int cifs_launder_page(struct page *page)
4079{
4080 int rc = 0;
4081 loff_t range_start = page_offset(page);
09cbfeaf 4082 loff_t range_end = range_start + (loff_t)(PAGE_SIZE - 1);
9ad1506b
PS
4083 struct writeback_control wbc = {
4084 .sync_mode = WB_SYNC_ALL,
4085 .nr_to_write = 0,
4086 .range_start = range_start,
4087 .range_end = range_end,
4088 };
4089
f96637be 4090 cifs_dbg(FYI, "Launder page: %p\n", page);
9ad1506b
PS
4091
4092 if (clear_page_dirty_for_io(page))
4093 rc = cifs_writepage_locked(page, &wbc);
4094
4095 cifs_fscache_invalidate_page(page, page->mapping->host);
4096 return rc;
4097}
4098
9b646972 4099void cifs_oplock_break(struct work_struct *work)
3bc303c2
JL
4100{
4101 struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo,
4102 oplock_break);
2b0143b5 4103 struct inode *inode = d_inode(cfile->dentry);
3bc303c2 4104 struct cifsInodeInfo *cinode = CIFS_I(inode);
95a3f2f3 4105 struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
c11f1df5 4106 struct TCP_Server_Info *server = tcon->ses->server;
eb4b756b 4107 int rc = 0;
3bc303c2 4108
c11f1df5 4109 wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS,
74316201 4110 TASK_UNINTERRUPTIBLE);
c11f1df5
SP
4111
4112 server->ops->downgrade_oplock(server, cinode,
4113 test_bit(CIFS_INODE_DOWNGRADE_OPLOCK_TO_L2, &cinode->flags));
4114
18cceb6a 4115 if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) &&
63b7d3a4 4116 cifs_has_mand_locks(cinode)) {
f96637be
JP
4117 cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n",
4118 inode);
18cceb6a 4119 cinode->oplock = 0;
63b7d3a4
PS
4120 }
4121
3bc303c2 4122 if (inode && S_ISREG(inode->i_mode)) {
18cceb6a 4123 if (CIFS_CACHE_READ(cinode))
8737c930 4124 break_lease(inode, O_RDONLY);
d54ff732 4125 else
8737c930 4126 break_lease(inode, O_WRONLY);
3bc303c2 4127 rc = filemap_fdatawrite(inode->i_mapping);
18cceb6a 4128 if (!CIFS_CACHE_READ(cinode)) {
eb4b756b
JL
4129 rc = filemap_fdatawait(inode->i_mapping);
4130 mapping_set_error(inode->i_mapping, rc);
4f73c7d3 4131 cifs_zap_mapping(inode);
3bc303c2 4132 }
f96637be 4133 cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc);
3bc303c2
JL
4134 }
4135
85160e03
PS
4136 rc = cifs_push_locks(cfile);
4137 if (rc)
f96637be 4138 cifs_dbg(VFS, "Push locks rc = %d\n", rc);
85160e03 4139
3bc303c2
JL
4140 /*
4141 * releasing stale oplock after recent reconnect of smb session using
4142 * a now incorrect file handle is not a data integrity issue but do
4143 * not bother sending an oplock release if session to server still is
4144 * disconnected since oplock already released by the server
4145 */
cdff08e7 4146 if (!cfile->oplock_break_cancelled) {
95a3f2f3
PS
4147 rc = tcon->ses->server->ops->oplock_response(tcon, &cfile->fid,
4148 cinode);
f96637be 4149 cifs_dbg(FYI, "Oplock release rc = %d\n", rc);
3bc303c2 4150 }
c11f1df5 4151 cifs_done_oplock_break(cinode);
3bc303c2
JL
4152}
4153
dca69288
SF
4154/*
4155 * The presence of cifs_direct_io() in the address space ops vector
4156 * allowes open() O_DIRECT flags which would have failed otherwise.
4157 *
4158 * In the non-cached mode (mount with cache=none), we shunt off direct read and write requests
4159 * so this method should never be called.
4160 *
4161 * Direct IO is not yet supported in the cached mode.
4162 */
4163static ssize_t
c8b8e32d 4164cifs_direct_io(struct kiocb *iocb, struct iov_iter *iter)
dca69288
SF
4165{
4166 /*
4167 * FIXME
4168 * Eventually need to support direct IO for non forcedirectio mounts
4169 */
4170 return -EINVAL;
4171}
4172
4173
f5e54d6e 4174const struct address_space_operations cifs_addr_ops = {
1da177e4
LT
4175 .readpage = cifs_readpage,
4176 .readpages = cifs_readpages,
4177 .writepage = cifs_writepage,
37c0eb46 4178 .writepages = cifs_writepages,
d9414774
NP
4179 .write_begin = cifs_write_begin,
4180 .write_end = cifs_write_end,
1da177e4 4181 .set_page_dirty = __set_page_dirty_nobuffers,
85f2d6b4 4182 .releasepage = cifs_release_page,
dca69288 4183 .direct_IO = cifs_direct_io,
85f2d6b4 4184 .invalidatepage = cifs_invalidate_page,
9ad1506b 4185 .launder_page = cifs_launder_page,
1da177e4 4186};
273d81d6
DK
4187
4188/*
4189 * cifs_readpages requires the server to support a buffer large enough to
4190 * contain the header plus one complete page of data. Otherwise, we need
4191 * to leave cifs_readpages out of the address space operations.
4192 */
f5e54d6e 4193const struct address_space_operations cifs_addr_ops_smallbuf = {
273d81d6
DK
4194 .readpage = cifs_readpage,
4195 .writepage = cifs_writepage,
4196 .writepages = cifs_writepages,
d9414774
NP
4197 .write_begin = cifs_write_begin,
4198 .write_end = cifs_write_end,
273d81d6 4199 .set_page_dirty = __set_page_dirty_nobuffers,
85f2d6b4
SJ
4200 .releasepage = cifs_release_page,
4201 .invalidatepage = cifs_invalidate_page,
9ad1506b 4202 .launder_page = cifs_launder_page,
273d81d6 4203};