diff options
Diffstat (limited to 'fs/nfs/objlayout/objlayout.c')
| -rw-r--r-- | fs/nfs/objlayout/objlayout.c | 712 | 
1 files changed, 712 insertions, 0 deletions
diff --git a/fs/nfs/objlayout/objlayout.c b/fs/nfs/objlayout/objlayout.c new file mode 100644 index 00000000000..dc3956c0de8 --- /dev/null +++ b/fs/nfs/objlayout/objlayout.c @@ -0,0 +1,712 @@ +/* + *  pNFS Objects layout driver high level definitions + * + *  Copyright (C) 2007 Panasas Inc. [year of first publication] + *  All rights reserved. + * + *  Benny Halevy <bhalevy@panasas.com> + *  Boaz Harrosh <bharrosh@panasas.com> + * + *  This program is free software; you can redistribute it and/or modify + *  it under the terms of the GNU General Public License version 2 + *  See the file COPYING included with this distribution for more details. + * + *  Redistribution and use in source and binary forms, with or without + *  modification, are permitted provided that the following conditions + *  are met: + * + *  1. Redistributions of source code must retain the above copyright + *     notice, this list of conditions and the following disclaimer. + *  2. Redistributions in binary form must reproduce the above copyright + *     notice, this list of conditions and the following disclaimer in the + *     documentation and/or other materials provided with the distribution. + *  3. Neither the name of the Panasas company nor the names of its + *     contributors may be used to endorse or promote products derived + *     from this software without specific prior written permission. + * + *  THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED + *  WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF + *  MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + *  DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE + *  FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + *  BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF + *  LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING + *  NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS + *  SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include <scsi/osd_initiator.h> +#include "objlayout.h" + +#define NFSDBG_FACILITY         NFSDBG_PNFS_LD +/* + * Create a objlayout layout structure for the given inode and return it. + */ +struct pnfs_layout_hdr * +objlayout_alloc_layout_hdr(struct inode *inode, gfp_t gfp_flags) +{ +	struct objlayout *objlay; + +	objlay = kzalloc(sizeof(struct objlayout), gfp_flags); +	if (objlay) { +		spin_lock_init(&objlay->lock); +		INIT_LIST_HEAD(&objlay->err_list); +	} +	dprintk("%s: Return %p\n", __func__, objlay); +	return &objlay->pnfs_layout; +} + +/* + * Free an objlayout layout structure + */ +void +objlayout_free_layout_hdr(struct pnfs_layout_hdr *lo) +{ +	struct objlayout *objlay = OBJLAYOUT(lo); + +	dprintk("%s: objlay %p\n", __func__, objlay); + +	WARN_ON(!list_empty(&objlay->err_list)); +	kfree(objlay); +} + +/* + * Unmarshall layout and store it in pnfslay. + */ +struct pnfs_layout_segment * +objlayout_alloc_lseg(struct pnfs_layout_hdr *pnfslay, +		     struct nfs4_layoutget_res *lgr, +		     gfp_t gfp_flags) +{ +	int status = -ENOMEM; +	struct xdr_stream stream; +	struct xdr_buf buf = { +		.pages =  lgr->layoutp->pages, +		.page_len =  lgr->layoutp->len, +		.buflen =  lgr->layoutp->len, +		.len = lgr->layoutp->len, +	}; +	struct page *scratch; +	struct pnfs_layout_segment *lseg; + +	dprintk("%s: Begin pnfslay %p\n", __func__, pnfslay); + +	scratch = alloc_page(gfp_flags); +	if (!scratch) +		goto err_nofree; + +	xdr_init_decode(&stream, &buf, NULL); +	xdr_set_scratch_buffer(&stream, page_address(scratch), PAGE_SIZE); + +	status = objio_alloc_lseg(&lseg, pnfslay, &lgr->range, &stream, gfp_flags); +	if (unlikely(status)) { +		dprintk("%s: objio_alloc_lseg Return err %d\n", __func__, +			status); +		goto err; +	} + +	__free_page(scratch); + +	dprintk("%s: Return %p\n", __func__, lseg); +	return lseg; + +err: +	__free_page(scratch); +err_nofree: +	dprintk("%s: Err Return=>%d\n", __func__, status); +	return ERR_PTR(status); +} + +/* + * Free a layout segement + */ +void +objlayout_free_lseg(struct pnfs_layout_segment *lseg) +{ +	dprintk("%s: freeing layout segment %p\n", __func__, lseg); + +	if (unlikely(!lseg)) +		return; + +	objio_free_lseg(lseg); +} + +/* + * I/O Operations + */ +static inline u64 +end_offset(u64 start, u64 len) +{ +	u64 end; + +	end = start + len; +	return end >= start ? end : NFS4_MAX_UINT64; +} + +/* last octet in a range */ +static inline u64 +last_byte_offset(u64 start, u64 len) +{ +	u64 end; + +	BUG_ON(!len); +	end = start + len; +	return end > start ? end - 1 : NFS4_MAX_UINT64; +} + +static struct objlayout_io_state * +objlayout_alloc_io_state(struct pnfs_layout_hdr *pnfs_layout_type, +			struct page **pages, +			unsigned pgbase, +			loff_t offset, +			size_t count, +			struct pnfs_layout_segment *lseg, +			void *rpcdata, +			gfp_t gfp_flags) +{ +	struct objlayout_io_state *state; +	u64 lseg_end_offset; + +	dprintk("%s: allocating io_state\n", __func__); +	if (objio_alloc_io_state(lseg, &state, gfp_flags)) +		return NULL; + +	BUG_ON(offset < lseg->pls_range.offset); +	lseg_end_offset = end_offset(lseg->pls_range.offset, +				     lseg->pls_range.length); +	BUG_ON(offset >= lseg_end_offset); +	if (offset + count > lseg_end_offset) { +		count = lseg->pls_range.length - +				(offset - lseg->pls_range.offset); +		dprintk("%s: truncated count %Zd\n", __func__, count); +	} + +	if (pgbase > PAGE_SIZE) { +		pages += pgbase >> PAGE_SHIFT; +		pgbase &= ~PAGE_MASK; +	} + +	INIT_LIST_HEAD(&state->err_list); +	state->lseg = lseg; +	state->rpcdata = rpcdata; +	state->pages = pages; +	state->pgbase = pgbase; +	state->nr_pages = (pgbase + count + PAGE_SIZE - 1) >> PAGE_SHIFT; +	state->offset = offset; +	state->count = count; +	state->sync = 0; + +	return state; +} + +static void +objlayout_free_io_state(struct objlayout_io_state *state) +{ +	dprintk("%s: freeing io_state\n", __func__); +	if (unlikely(!state)) +		return; + +	objio_free_io_state(state); +} + +/* + * I/O done common code + */ +static void +objlayout_iodone(struct objlayout_io_state *state) +{ +	dprintk("%s: state %p status\n", __func__, state); + +	if (likely(state->status >= 0)) { +		objlayout_free_io_state(state); +	} else { +		struct objlayout *objlay = OBJLAYOUT(state->lseg->pls_layout); + +		spin_lock(&objlay->lock); +		objlay->delta_space_valid = OBJ_DSU_INVALID; +		list_add(&objlay->err_list, &state->err_list); +		spin_unlock(&objlay->lock); +	} +} + +/* + * objlayout_io_set_result - Set an osd_error code on a specific osd comp. + * + * The @index component IO failed (error returned from target). Register + * the error for later reporting at layout-return. + */ +void +objlayout_io_set_result(struct objlayout_io_state *state, unsigned index, +			struct pnfs_osd_objid *pooid, int osd_error, +			u64 offset, u64 length, bool is_write) +{ +	struct pnfs_osd_ioerr *ioerr = &state->ioerrs[index]; + +	BUG_ON(index >= state->num_comps); +	if (osd_error) { +		ioerr->oer_component = *pooid; +		ioerr->oer_comp_offset = offset; +		ioerr->oer_comp_length = length; +		ioerr->oer_iswrite = is_write; +		ioerr->oer_errno = osd_error; + +		dprintk("%s: err[%d]: errno=%d is_write=%d dev(%llx:%llx) " +			"par=0x%llx obj=0x%llx offset=0x%llx length=0x%llx\n", +			__func__, index, ioerr->oer_errno, +			ioerr->oer_iswrite, +			_DEVID_LO(&ioerr->oer_component.oid_device_id), +			_DEVID_HI(&ioerr->oer_component.oid_device_id), +			ioerr->oer_component.oid_partition_id, +			ioerr->oer_component.oid_object_id, +			ioerr->oer_comp_offset, +			ioerr->oer_comp_length); +	} else { +		/* User need not call if no error is reported */ +		ioerr->oer_errno = 0; +	} +} + +/* Function scheduled on rpc workqueue to call ->nfs_readlist_complete(). + * This is because the osd completion is called with ints-off from + * the block layer + */ +static void _rpc_read_complete(struct work_struct *work) +{ +	struct rpc_task *task; +	struct nfs_read_data *rdata; + +	dprintk("%s enter\n", __func__); +	task = container_of(work, struct rpc_task, u.tk_work); +	rdata = container_of(task, struct nfs_read_data, task); + +	pnfs_ld_read_done(rdata); +} + +void +objlayout_read_done(struct objlayout_io_state *state, ssize_t status, bool sync) +{ +	int eof = state->eof; +	struct nfs_read_data *rdata; + +	state->status = status; +	dprintk("%s: Begin status=%ld eof=%d\n", __func__, status, eof); +	rdata = state->rpcdata; +	rdata->task.tk_status = status; +	if (status >= 0) { +		rdata->res.count = status; +		rdata->res.eof = eof; +	} +	objlayout_iodone(state); +	/* must not use state after this point */ + +	if (sync) +		pnfs_ld_read_done(rdata); +	else { +		INIT_WORK(&rdata->task.u.tk_work, _rpc_read_complete); +		schedule_work(&rdata->task.u.tk_work); +	} +} + +/* + * Perform sync or async reads. + */ +enum pnfs_try_status +objlayout_read_pagelist(struct nfs_read_data *rdata) +{ +	loff_t offset = rdata->args.offset; +	size_t count = rdata->args.count; +	struct objlayout_io_state *state; +	ssize_t status = 0; +	loff_t eof; + +	dprintk("%s: Begin inode %p offset %llu count %d\n", +		__func__, rdata->inode, offset, (int)count); + +	eof = i_size_read(rdata->inode); +	if (unlikely(offset + count > eof)) { +		if (offset >= eof) { +			status = 0; +			rdata->res.count = 0; +			rdata->res.eof = 1; +			goto out; +		} +		count = eof - offset; +	} + +	state = objlayout_alloc_io_state(NFS_I(rdata->inode)->layout, +					 rdata->args.pages, rdata->args.pgbase, +					 offset, count, +					 rdata->lseg, rdata, +					 GFP_KERNEL); +	if (unlikely(!state)) { +		status = -ENOMEM; +		goto out; +	} + +	state->eof = state->offset + state->count >= eof; + +	status = objio_read_pagelist(state); + out: +	dprintk("%s: Return status %Zd\n", __func__, status); +	rdata->pnfs_error = status; +	return PNFS_ATTEMPTED; +} + +/* Function scheduled on rpc workqueue to call ->nfs_writelist_complete(). + * This is because the osd completion is called with ints-off from + * the block layer + */ +static void _rpc_write_complete(struct work_struct *work) +{ +	struct rpc_task *task; +	struct nfs_write_data *wdata; + +	dprintk("%s enter\n", __func__); +	task = container_of(work, struct rpc_task, u.tk_work); +	wdata = container_of(task, struct nfs_write_data, task); + +	pnfs_ld_write_done(wdata); +} + +void +objlayout_write_done(struct objlayout_io_state *state, ssize_t status, +		     bool sync) +{ +	struct nfs_write_data *wdata; + +	dprintk("%s: Begin\n", __func__); +	wdata = state->rpcdata; +	state->status = status; +	wdata->task.tk_status = status; +	if (status >= 0) { +		wdata->res.count = status; +		wdata->verf.committed = state->committed; +		dprintk("%s: Return status %d committed %d\n", +			__func__, wdata->task.tk_status, +			wdata->verf.committed); +	} else +		dprintk("%s: Return status %d\n", +			__func__, wdata->task.tk_status); +	objlayout_iodone(state); +	/* must not use state after this point */ + +	if (sync) +		pnfs_ld_write_done(wdata); +	else { +		INIT_WORK(&wdata->task.u.tk_work, _rpc_write_complete); +		schedule_work(&wdata->task.u.tk_work); +	} +} + +/* + * Perform sync or async writes. + */ +enum pnfs_try_status +objlayout_write_pagelist(struct nfs_write_data *wdata, +			 int how) +{ +	struct objlayout_io_state *state; +	ssize_t status; + +	dprintk("%s: Begin inode %p offset %llu count %u\n", +		__func__, wdata->inode, wdata->args.offset, wdata->args.count); + +	state = objlayout_alloc_io_state(NFS_I(wdata->inode)->layout, +					 wdata->args.pages, +					 wdata->args.pgbase, +					 wdata->args.offset, +					 wdata->args.count, +					 wdata->lseg, wdata, +					 GFP_NOFS); +	if (unlikely(!state)) { +		status = -ENOMEM; +		goto out; +	} + +	state->sync = how & FLUSH_SYNC; + +	status = objio_write_pagelist(state, how & FLUSH_STABLE); + out: +	dprintk("%s: Return status %Zd\n", __func__, status); +	wdata->pnfs_error = status; +	return PNFS_ATTEMPTED; +} + +void +objlayout_encode_layoutcommit(struct pnfs_layout_hdr *pnfslay, +			      struct xdr_stream *xdr, +			      const struct nfs4_layoutcommit_args *args) +{ +	struct objlayout *objlay = OBJLAYOUT(pnfslay); +	struct pnfs_osd_layoutupdate lou; +	__be32 *start; + +	dprintk("%s: Begin\n", __func__); + +	spin_lock(&objlay->lock); +	lou.dsu_valid = (objlay->delta_space_valid == OBJ_DSU_VALID); +	lou.dsu_delta = objlay->delta_space_used; +	objlay->delta_space_used = 0; +	objlay->delta_space_valid = OBJ_DSU_INIT; +	lou.olu_ioerr_flag = !list_empty(&objlay->err_list); +	spin_unlock(&objlay->lock); + +	start = xdr_reserve_space(xdr, 4); + +	BUG_ON(pnfs_osd_xdr_encode_layoutupdate(xdr, &lou)); + +	*start = cpu_to_be32((xdr->p - start - 1) * 4); + +	dprintk("%s: Return delta_space_used %lld err %d\n", __func__, +		lou.dsu_delta, lou.olu_ioerr_flag); +} + +static int +err_prio(u32 oer_errno) +{ +	switch (oer_errno) { +	case 0: +		return 0; + +	case PNFS_OSD_ERR_RESOURCE: +		return OSD_ERR_PRI_RESOURCE; +	case PNFS_OSD_ERR_BAD_CRED: +		return OSD_ERR_PRI_BAD_CRED; +	case PNFS_OSD_ERR_NO_ACCESS: +		return OSD_ERR_PRI_NO_ACCESS; +	case PNFS_OSD_ERR_UNREACHABLE: +		return OSD_ERR_PRI_UNREACHABLE; +	case PNFS_OSD_ERR_NOT_FOUND: +		return OSD_ERR_PRI_NOT_FOUND; +	case PNFS_OSD_ERR_NO_SPACE: +		return OSD_ERR_PRI_NO_SPACE; +	default: +		WARN_ON(1); +		/* fallthrough */ +	case PNFS_OSD_ERR_EIO: +		return OSD_ERR_PRI_EIO; +	} +} + +static void +merge_ioerr(struct pnfs_osd_ioerr *dest_err, +	    const struct pnfs_osd_ioerr *src_err) +{ +	u64 dest_end, src_end; + +	if (!dest_err->oer_errno) { +		*dest_err = *src_err; +		/* accumulated device must be blank */ +		memset(&dest_err->oer_component.oid_device_id, 0, +			sizeof(dest_err->oer_component.oid_device_id)); + +		return; +	} + +	if (dest_err->oer_component.oid_partition_id != +				src_err->oer_component.oid_partition_id) +		dest_err->oer_component.oid_partition_id = 0; + +	if (dest_err->oer_component.oid_object_id != +				src_err->oer_component.oid_object_id) +		dest_err->oer_component.oid_object_id = 0; + +	if (dest_err->oer_comp_offset > src_err->oer_comp_offset) +		dest_err->oer_comp_offset = src_err->oer_comp_offset; + +	dest_end = end_offset(dest_err->oer_comp_offset, +			      dest_err->oer_comp_length); +	src_end =  end_offset(src_err->oer_comp_offset, +			      src_err->oer_comp_length); +	if (dest_end < src_end) +		dest_end = src_end; + +	dest_err->oer_comp_length = dest_end - dest_err->oer_comp_offset; + +	if ((src_err->oer_iswrite == dest_err->oer_iswrite) && +	    (err_prio(src_err->oer_errno) > err_prio(dest_err->oer_errno))) { +			dest_err->oer_errno = src_err->oer_errno; +	} else if (src_err->oer_iswrite) { +		dest_err->oer_iswrite = true; +		dest_err->oer_errno = src_err->oer_errno; +	} +} + +static void +encode_accumulated_error(struct objlayout *objlay, __be32 *p) +{ +	struct objlayout_io_state *state, *tmp; +	struct pnfs_osd_ioerr accumulated_err = {.oer_errno = 0}; + +	list_for_each_entry_safe(state, tmp, &objlay->err_list, err_list) { +		unsigned i; + +		for (i = 0; i < state->num_comps; i++) { +			struct pnfs_osd_ioerr *ioerr = &state->ioerrs[i]; + +			if (!ioerr->oer_errno) +				continue; + +			printk(KERN_ERR "%s: err[%d]: errno=%d is_write=%d " +				"dev(%llx:%llx) par=0x%llx obj=0x%llx " +				"offset=0x%llx length=0x%llx\n", +				__func__, i, ioerr->oer_errno, +				ioerr->oer_iswrite, +				_DEVID_LO(&ioerr->oer_component.oid_device_id), +				_DEVID_HI(&ioerr->oer_component.oid_device_id), +				ioerr->oer_component.oid_partition_id, +				ioerr->oer_component.oid_object_id, +				ioerr->oer_comp_offset, +				ioerr->oer_comp_length); + +			merge_ioerr(&accumulated_err, ioerr); +		} +		list_del(&state->err_list); +		objlayout_free_io_state(state); +	} + +	pnfs_osd_xdr_encode_ioerr(p, &accumulated_err); +} + +void +objlayout_encode_layoutreturn(struct pnfs_layout_hdr *pnfslay, +			      struct xdr_stream *xdr, +			      const struct nfs4_layoutreturn_args *args) +{ +	struct objlayout *objlay = OBJLAYOUT(pnfslay); +	struct objlayout_io_state *state, *tmp; +	__be32 *start; + +	dprintk("%s: Begin\n", __func__); +	start = xdr_reserve_space(xdr, 4); +	BUG_ON(!start); + +	spin_lock(&objlay->lock); + +	list_for_each_entry_safe(state, tmp, &objlay->err_list, err_list) { +		__be32 *last_xdr = NULL, *p; +		unsigned i; +		int res = 0; + +		for (i = 0; i < state->num_comps; i++) { +			struct pnfs_osd_ioerr *ioerr = &state->ioerrs[i]; + +			if (!ioerr->oer_errno) +				continue; + +			dprintk("%s: err[%d]: errno=%d is_write=%d " +				"dev(%llx:%llx) par=0x%llx obj=0x%llx " +				"offset=0x%llx length=0x%llx\n", +				__func__, i, ioerr->oer_errno, +				ioerr->oer_iswrite, +				_DEVID_LO(&ioerr->oer_component.oid_device_id), +				_DEVID_HI(&ioerr->oer_component.oid_device_id), +				ioerr->oer_component.oid_partition_id, +				ioerr->oer_component.oid_object_id, +				ioerr->oer_comp_offset, +				ioerr->oer_comp_length); + +			p = pnfs_osd_xdr_ioerr_reserve_space(xdr); +			if (unlikely(!p)) { +				res = -E2BIG; +				break; /* accumulated_error */ +			} + +			last_xdr = p; +			pnfs_osd_xdr_encode_ioerr(p, &state->ioerrs[i]); +		} + +		/* TODO: use xdr_write_pages */ +		if (unlikely(res)) { +			/* no space for even one error descriptor */ +			BUG_ON(!last_xdr); + +			/* we've encountered a situation with lots and lots of +			 * errors and no space to encode them all. Use the last +			 * available slot to report the union of all the +			 * remaining errors. +			 */ +			encode_accumulated_error(objlay, last_xdr); +			goto loop_done; +		} +		list_del(&state->err_list); +		objlayout_free_io_state(state); +	} +loop_done: +	spin_unlock(&objlay->lock); + +	*start = cpu_to_be32((xdr->p - start - 1) * 4); +	dprintk("%s: Return\n", __func__); +} + + +/* + * Get Device Info API for io engines + */ +struct objlayout_deviceinfo { +	struct page *page; +	struct pnfs_osd_deviceaddr da; /* This must be last */ +}; + +/* Initialize and call nfs_getdeviceinfo, then decode and return a + * "struct pnfs_osd_deviceaddr *" Eventually objlayout_put_deviceinfo() + * should be called. + */ +int objlayout_get_deviceinfo(struct pnfs_layout_hdr *pnfslay, +	struct nfs4_deviceid *d_id, struct pnfs_osd_deviceaddr **deviceaddr, +	gfp_t gfp_flags) +{ +	struct objlayout_deviceinfo *odi; +	struct pnfs_device pd; +	struct super_block *sb; +	struct page *page, **pages; +	u32 *p; +	int err; + +	page = alloc_page(gfp_flags); +	if (!page) +		return -ENOMEM; + +	pages = &page; +	pd.pages = pages; + +	memcpy(&pd.dev_id, d_id, sizeof(*d_id)); +	pd.layout_type = LAYOUT_OSD2_OBJECTS; +	pd.pages = &page; +	pd.pgbase = 0; +	pd.pglen = PAGE_SIZE; +	pd.mincount = 0; + +	sb = pnfslay->plh_inode->i_sb; +	err = nfs4_proc_getdeviceinfo(NFS_SERVER(pnfslay->plh_inode), &pd); +	dprintk("%s nfs_getdeviceinfo returned %d\n", __func__, err); +	if (err) +		goto err_out; + +	p = page_address(page); +	odi = kzalloc(sizeof(*odi), gfp_flags); +	if (!odi) { +		err = -ENOMEM; +		goto err_out; +	} +	pnfs_osd_xdr_decode_deviceaddr(&odi->da, p); +	odi->page = page; +	*deviceaddr = &odi->da; +	return 0; + +err_out: +	__free_page(page); +	return err; +} + +void objlayout_put_deviceinfo(struct pnfs_osd_deviceaddr *deviceaddr) +{ +	struct objlayout_deviceinfo *odi = container_of(deviceaddr, +						struct objlayout_deviceinfo, +						da); + +	__free_page(odi->page); +	kfree(odi); +}  |