diff options
| author | Tony Lindgren <tony@atomide.com> | 2010-12-21 16:53:00 -0800 | 
|---|---|---|
| committer | Tony Lindgren <tony@atomide.com> | 2010-12-21 16:53:00 -0800 | 
| commit | bb3613aa34a81a5e2f1227ccdb801fde04a7da10 (patch) | |
| tree | bb79c15d5da41113bd7b83d9e74fbfc4a1bf8569 /include/linux/blkdev.h | |
| parent | 6971071cdda79cad5f53ba390e466d696e7e9006 (diff) | |
| parent | bb1c9034b3ce7f29d3d178a87b42b767611d6574 (diff) | |
| download | olio-linux-3.10-bb3613aa34a81a5e2f1227ccdb801fde04a7da10.tar.xz olio-linux-3.10-bb3613aa34a81a5e2f1227ccdb801fde04a7da10.zip  | |
Merge branch 'pm-next' of ssh://master.kernel.org/pub/scm/linux/kernel/git/khilman/linux-omap-pm into omap-for-linus
Diffstat (limited to 'include/linux/blkdev.h')
| -rw-r--r-- | include/linux/blkdev.h | 10 | 
1 files changed, 7 insertions, 3 deletions
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index aae86fd10c4..36ab42c9bb9 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h @@ -250,7 +250,7 @@ struct queue_limits {  	unsigned char		misaligned;  	unsigned char		discard_misaligned; -	unsigned char		no_cluster; +	unsigned char		cluster;  	signed char		discard_zeroes_data;  }; @@ -380,7 +380,6 @@ struct request_queue  #endif  }; -#define QUEUE_FLAG_CLUSTER	0	/* cluster several segments into 1 */  #define QUEUE_FLAG_QUEUED	1	/* uses generic tag queueing */  #define QUEUE_FLAG_STOPPED	2	/* queue is stopped */  #define	QUEUE_FLAG_SYNCFULL	3	/* read queue has been filled */ @@ -403,7 +402,6 @@ struct request_queue  #define QUEUE_FLAG_SECDISCARD  19	/* supports SECDISCARD */  #define QUEUE_FLAG_DEFAULT	((1 << QUEUE_FLAG_IO_STAT) |		\ -				 (1 << QUEUE_FLAG_CLUSTER) |		\  				 (1 << QUEUE_FLAG_STACKABLE)	|	\  				 (1 << QUEUE_FLAG_SAME_COMP)	|	\  				 (1 << QUEUE_FLAG_ADD_RANDOM)) @@ -510,6 +508,11 @@ static inline void queue_flag_clear(unsigned int flag, struct request_queue *q)  #define rq_data_dir(rq)		((rq)->cmd_flags & 1) +static inline unsigned int blk_queue_cluster(struct request_queue *q) +{ +	return q->limits.cluster; +} +  /*   * We regard a request as sync, if either a read or a sync write   */ @@ -805,6 +808,7 @@ extern struct request_queue *blk_init_allocated_queue(struct request_queue *,  extern void blk_cleanup_queue(struct request_queue *);  extern void blk_queue_make_request(struct request_queue *, make_request_fn *);  extern void blk_queue_bounce_limit(struct request_queue *, u64); +extern void blk_limits_max_hw_sectors(struct queue_limits *, unsigned int);  extern void blk_queue_max_hw_sectors(struct request_queue *, unsigned int);  extern void blk_queue_max_segments(struct request_queue *, unsigned short);  extern void blk_queue_max_segment_size(struct request_queue *, unsigned int);  |