/drivers/staging/lustre/lustre/lov/ |
H A D | lov_page.c | 158 struct cl_page *page, struct page *vmpage) 190 cl_index(subobj, suboff), vmpage, page); 218 struct cl_page *page, struct page *vmpage) 224 addr = kmap(vmpage); 226 kunmap(vmpage); 157 lov_page_init_raid0(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument 217 lov_page_init_empty(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument
|
H A D | lov_object.c | 70 struct cl_page *page, struct page *vmpage); 817 struct cl_page *page, struct page *vmpage) 820 llo_page_init, env, obj, page, vmpage); 816 lov_page_init(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument
|
/drivers/staging/lustre/lustre/llite/ |
H A D | llite_mmap.c | 169 static int ll_page_mkwrite0(struct vm_area_struct *vma, struct page *vmpage, argument 181 LASSERT(vmpage != NULL); 183 io = ll_fault_io_init(vma, &env, &nest, vmpage->index, NULL); 198 vio->u.fault.ft_vmpage = vmpage; 219 lock_page(vmpage); 220 if (vmpage->mapping == NULL) { 221 unlock_page(vmpage); 228 } else if (!PageDirty(vmpage)) { 235 unlock_page(vmpage); 239 vmpage, vmpag 299 struct page *vmpage; local 359 struct page *vmpage = vmf->page; local [all...] |
H A D | vvp_page.c | 58 struct page *vmpage = cp->cpg_page; local 60 LASSERT(vmpage != NULL); 61 page_cache_release(vmpage); 68 struct page *vmpage = cp->cpg_page; local 71 * vmpage->private was already cleared when page was moved into 74 LASSERT((struct cl_page *)vmpage->private != slice->cpl_page); 83 struct page *vmpage = vpg->cpg_page; local 85 LASSERT(vmpage != NULL); 87 if (!trylock_page(vmpage)) 90 if (unlikely(PageWriteback(vmpage))) { 107 struct page *vmpage = cl2vm_page(slice); local 118 struct page *vmpage = cl2vm_page(slice); local 127 struct page *vmpage = cl2vm_page(slice); local 139 struct page *vmpage = cl2vm_page(slice); local 162 struct page *vmpage = cl2vm_page(slice); local 180 struct page *vmpage = cl2vm_page(slice); local 201 struct page *vmpage = cl2vm_page(slice); local 229 struct page *vmpage = cl2vm_page(slice); local 246 vvp_vmpage_error(struct inode *inode, struct page *vmpage, int ioret) argument 273 struct page *vmpage = cp->cpg_page; local 299 struct page *vmpage = cp->cpg_page; local 346 struct page *vmpage = cl2vm_page(slice); local 377 struct page *vmpage = vp->cpg_page; local 527 vvp_page_init(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument [all...] |
H A D | rw26.c | 72 static void ll_invalidatepage(struct page *vmpage, unsigned int offset, argument 82 LASSERT(PageLocked(vmpage)); 83 LASSERT(!PageWriteback(vmpage)); 93 inode = vmpage->mapping->host; 96 page = cl_vmpage_page(vmpage, obj); 99 "delete", vmpage); 102 "delete", vmpage); 106 LASSERT(vmpage->private == 0); 117 static int ll_releasepage(struct page *vmpage, RELEASEPAGE_ARG_TYPE gfp_mask) argument 126 LASSERT(PageLocked(vmpage)); 164 ll_set_page_dirty(struct page *vmpage) argument 271 struct page *vmpage = cl_page_vmpage(env, clp); local [all...] |
H A D | vvp_dev.c | 403 struct page *vmpage; local 407 vmpage = cpg->cpg_page; 414 PageWriteback(vmpage) ? "wb" : "-", 415 vmpage, vmpage->mapping->host->i_ino, 416 vmpage->mapping->host->i_generation, 417 vmpage->mapping->host, vmpage->index, 418 page_count(vmpage)); 420 seq_page_flag(seq, vmpage, locke [all...] |
H A D | vvp_io.c | 664 struct page *vmpage = NULL; local 692 vmpage = cfio->ft_vmpage; 693 LASSERT(PageLocked(vmpage)); 696 ll_invalidate_page(vmpage); 701 if (unlikely((vmpage->mapping != inode->i_mapping) || 702 (page_offset(vmpage) > size))) { 725 vmpage->mapping, fio->ft_index, last_index); 732 * converted to 0. If the vmpage->mapping is null 742 page = cl_page_find(env, obj, fio->ft_index, vmpage, CPT_CACHEABLE); 751 wait_on_page_writeback(vmpage); 832 struct page *vmpage = cp->cpg_page; local 949 struct page *vmpage = cp->cpg_page; local 987 struct page *vmpage = cp->cpg_page; local [all...] |
H A D | rw.c | 88 struct page *vmpage, int create) 99 clob = ll_i2info(vmpage->mapping->host)->lli_clob; 115 struct inode *inode = vmpage->mapping->host; 148 pos = (vmpage->index << PAGE_CACHE_SHIFT); 174 page = cl_page_find(env, clob, vmpage->index, vmpage, 189 vmpage->index, PFID(lu_object_fid(&clob->co_lu)), result, 215 int ll_prepare_write(struct file *file, struct page *vmpage, unsigned from, argument 221 lcc = ll_cl_init(file, vmpage, 1); 250 int ll_commit_write(struct file *file, struct page *vmpage, unsigne argument 87 ll_cl_init(struct file *file, struct page *vmpage, int create) argument 452 cl_read_ahead_page(const struct lu_env *env, struct cl_io *io, struct cl_page_list *queue, struct cl_page *page, struct page *vmpage) argument 497 struct page *vmpage; local 1122 ll_writepage(struct page *vmpage, struct writeback_control *wbc) argument 1263 ll_readpage(struct file *file, struct page *vmpage) argument [all...] |
H A D | llite_internal.h | 1004 static inline void ll_invalidate_page(struct page *vmpage) argument 1006 struct address_space *mapping = vmpage->mapping; 1007 loff_t offset = vmpage->index << PAGE_CACHE_SHIFT; 1009 LASSERT(PageLocked(vmpage)); 1014 truncate_complete_page(mapping, vmpage);
|
/drivers/staging/lustre/lustre/osc/ |
H A D | osc_io.c | 363 struct page *vmpage = cl_page_vmpage(env, page); local 365 if (PageLocked(vmpage))
|
H A D | osc_page.c | 507 struct cl_page *page, struct page *vmpage) 516 result = osc_prep_async_page(osc, opg, vmpage, 506 osc_page_init(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument
|
/drivers/staging/lustre/lustre/obdclass/ |
H A D | cl_page.c | 288 struct cl_object *o, pgoff_t ind, struct page *vmpage, 318 page, vmpage); 340 * the VM page \a vmpage. 351 pgoff_t idx, struct page *vmpage, 367 idx, PFID(&hdr->coh_lu.loh_fid), vmpage, vmpage->private, type); 370 /* vmpage lock is used to protect the child/parent 372 KLASSERT(PageLocked(vmpage)); 376 * - "vmpage" is locked (which prevents ->private from 382 page = cl_vmpage_page(vmpage, 287 cl_page_alloc(const struct lu_env *env, struct cl_object *o, pgoff_t ind, struct page *vmpage, enum cl_page_type type) argument 349 cl_page_find0(const struct lu_env *env, struct cl_object *o, pgoff_t idx, struct page *vmpage, enum cl_page_type type, struct cl_page *parent) argument 450 cl_page_find(const struct lu_env *env, struct cl_object *o, pgoff_t idx, struct page *vmpage, enum cl_page_type type) argument 459 cl_page_find_sub(const struct lu_env *env, struct cl_object *o, pgoff_t idx, struct page *vmpage, struct cl_page *parent) argument 641 cl_vmpage_page(struct page *vmpage, struct cl_object *obj) argument [all...] |
/drivers/staging/lustre/lustre/lclient/ |
H A D | lcommon_cl.c | 1063 * Returns a pointer to cl_page associated with \a vmpage, without acquiring 1065 * cl_vmpage_page() that can only be used under vmpage lock. 1067 struct cl_page *ccc_vmpage_page_transient(struct page *vmpage) argument 1069 KLASSERT(PageLocked(vmpage)); 1070 return (struct cl_page *)vmpage->private;
|
/drivers/staging/lustre/lustre/obdecho/ |
H A D | echo_client.c | 282 struct page *vmpage = ep->ep_vmpage; local 285 page_cache_release(vmpage); 373 struct cl_page *page, struct page *vmpage) 378 ep->ep_vmpage = vmpage; 379 page_cache_get(vmpage); 372 echo_page_init(const struct lu_env *env, struct cl_object *obj, struct cl_page *page, struct page *vmpage) argument
|