diff options
Diffstat (limited to 'net/core/dev.c')
| -rw-r--r-- | net/core/dev.c | 96 | 
1 files changed, 48 insertions, 48 deletions
diff --git a/net/core/dev.c b/net/core/dev.c index 1e94a1b9a0f..85d58d79932 100644 --- a/net/core/dev.c +++ b/net/core/dev.c @@ -255,7 +255,7 @@ static int netdev_nit;   *	is linked into kernel lists and may not be freed until it has been   *	removed from the kernel lists.   * - *	This call does not sleep therefore it can not  + *	This call does not sleep therefore it can not   *	guarantee all CPU's that are in middle of receiving packets   *	will see the new packet type (until the next received packet).   */ @@ -282,7 +282,7 @@ void dev_add_pack(struct packet_type *pt)   *	Remove a protocol handler that was previously added to the kernel   *	protocol handlers by dev_add_pack(). The passed &packet_type is removed   *	from the kernel lists and can be freed or reused once this function - *	returns.  + *	returns.   *   *      The packet type might still be in use by receivers   *	and must not be freed until after all the CPU's have gone @@ -327,7 +327,7 @@ out:  void dev_remove_pack(struct packet_type *pt)  {  	__dev_remove_pack(pt); -	 +  	synchronize_net();  } @@ -607,7 +607,7 @@ EXPORT_SYMBOL(dev_getfirstbyhwtype);   *	@mask: bitmask of bits in if_flags to check   *   *	Search for any interface with the given flags. Returns NULL if a device - *	is not found or a pointer to the device. The device returned has  + *	is not found or a pointer to the device. The device returned has   *	had a reference added and the pointer is safe until the user calls   *	dev_put to indicate they have finished with it.   */ @@ -802,7 +802,7 @@ void netdev_state_change(struct net_device *dev)  void dev_load(const char *name)  { -	struct net_device *dev;   +	struct net_device *dev;  	read_lock(&dev_base_lock);  	dev = __dev_get_by_name(name); @@ -860,7 +860,7 @@ int dev_open(struct net_device *dev)  			clear_bit(__LINK_STATE_START, &dev->state);  	} - 	/* +	/*  	 *	If it went open OK then:  	 */ @@ -964,7 +964,7 @@ int dev_close(struct net_device *dev)   *	is returned on a failure.   *   * 	When registered all registration and up events are replayed - *	to the new notifier to allow device to have a race free  + *	to the new notifier to allow device to have a race free   *	view of the network device list.   */ @@ -979,7 +979,7 @@ int register_netdevice_notifier(struct notifier_block *nb)  		for (dev = dev_base; dev; dev = dev->next) {  			nb->notifier_call(nb, NETDEV_REGISTER, dev); -			if (dev->flags & IFF_UP)  +			if (dev->flags & IFF_UP)  				nb->notifier_call(nb, NETDEV_UP, dev);  		}  	} @@ -1157,7 +1157,7 @@ void netif_device_attach(struct net_device *dev)  	if (!test_and_set_bit(__LINK_STATE_PRESENT, &dev->state) &&  	    netif_running(dev)) {  		netif_wake_queue(dev); - 		__netdev_watchdog_up(dev); +		__netdev_watchdog_up(dev);  	}  }  EXPORT_SYMBOL(netif_device_attach); @@ -1197,7 +1197,7 @@ int skb_checksum_help(struct sk_buff *skb)  out_set_summed:  	skb->ip_summed = CHECKSUM_NONE; -out:	 +out:  	return ret;  } @@ -1258,7 +1258,7 @@ EXPORT_SYMBOL(skb_gso_segment);  void netdev_rx_csum_fault(struct net_device *dev)  {  	if (net_ratelimit()) { -		printk(KERN_ERR "%s: hw csum failure.\n",  +		printk(KERN_ERR "%s: hw csum failure.\n",  			dev ? dev->name : "<unknown>");  		dump_stack();  	} @@ -1372,7 +1372,7 @@ gso:  		if (unlikely(netif_queue_stopped(dev) && skb->next))  			return NETDEV_TX_BUSY;  	} while (skb->next); -	 +  	skb->destructor = DEV_GSO_CB(skb)->destructor;  out_kfree_skb: @@ -1449,25 +1449,25 @@ int dev_queue_xmit(struct sk_buff *skb)  	    (!(dev->features & NETIF_F_GEN_CSUM) &&  	     (!(dev->features & NETIF_F_IP_CSUM) ||  	      skb->protocol != htons(ETH_P_IP)))) -	      	if (skb_checksum_help(skb)) -	      		goto out_kfree_skb; +		if (skb_checksum_help(skb)) +			goto out_kfree_skb;  gso:  	spin_lock_prefetch(&dev->queue_lock); -	/* Disable soft irqs for various locks below. Also  -	 * stops preemption for RCU.  +	/* Disable soft irqs for various locks below. Also +	 * stops preemption for RCU.  	 */ -	rcu_read_lock_bh();  +	rcu_read_lock_bh(); -	/* Updates of qdisc are serialized by queue_lock.  -	 * The struct Qdisc which is pointed to by qdisc is now a  -	 * rcu structure - it may be accessed without acquiring  +	/* Updates of qdisc are serialized by queue_lock. +	 * The struct Qdisc which is pointed to by qdisc is now a +	 * rcu structure - it may be accessed without acquiring  	 * a lock (but the structure may be stale.) The freeing of the -	 * qdisc will be deferred until it's known that there are no  +	 * qdisc will be deferred until it's known that there are no  	 * more references to it. -	 *  -	 * If the qdisc has an enqueue function, we still need to  +	 * +	 * If the qdisc has an enqueue function, we still need to  	 * hold the queue_lock before calling it, since queue_lock  	 * also serializes access to the device queue.  	 */ @@ -1715,8 +1715,8 @@ static __inline__ int handle_bridge(struct sk_buff **pskb,  	if (*pt_prev) {  		*ret = deliver_skb(*pskb, *pt_prev, orig_dev);  		*pt_prev = NULL; -	}  -	 +	} +  	return br_handle_frame_hook(port, pskb);  }  #else @@ -1728,16 +1728,16 @@ static __inline__ int handle_bridge(struct sk_buff **pskb,   * when CONFIG_NET_CLS_ACT is? otherwise some useless instructions   * a compare and 2 stores extra right now if we dont have it on   * but have CONFIG_NET_CLS_ACT - * NOTE: This doesnt stop any functionality; if you dont have  + * NOTE: This doesnt stop any functionality; if you dont have   * the ingress scheduler, you just cant add policies on ingress.   *   */ -static int ing_filter(struct sk_buff *skb)  +static int ing_filter(struct sk_buff *skb)  {  	struct Qdisc *q;  	struct net_device *dev = skb->dev;  	int result = TC_ACT_OK; -	 +  	if (dev->qdisc_ingress) {  		__u32 ttl = (__u32) G_TC_RTTL(skb->tc_verd);  		if (MAX_RED_LOOP < ttl++) { @@ -1801,7 +1801,7 @@ int netif_receive_skb(struct sk_buff *skb)  	list_for_each_entry_rcu(ptype, &ptype_all, list) {  		if (!ptype->dev || ptype->dev == skb->dev) { -			if (pt_prev)  +			if (pt_prev)  				ret = deliver_skb(skb, pt_prev, orig_dev);  			pt_prev = ptype;  		} @@ -1833,7 +1833,7 @@ ncls:  	list_for_each_entry_rcu(ptype, &ptype_base[ntohs(type)&15], list) {  		if (ptype->type == type &&  		    (!ptype->dev || ptype->dev == skb->dev)) { -			if (pt_prev)  +			if (pt_prev)  				ret = deliver_skb(skb, pt_prev, orig_dev);  			pt_prev = ptype;  		} @@ -2061,7 +2061,7 @@ static int dev_ifconf(char __user *arg)  				total += done;  			}  		} -  	} +	}  	/*  	 *	All done.  Write the updated control block back to the caller. @@ -2154,7 +2154,7 @@ static struct netif_rx_stats *softnet_get_online(loff_t *pos)  	struct netif_rx_stats *rc = NULL;  	while (*pos < NR_CPUS) -	       	if (cpu_online(*pos)) { +		if (cpu_online(*pos)) {  			rc = &per_cpu(netdev_rx_stat, *pos);  			break;  		} else @@ -2282,7 +2282,7 @@ int netdev_set_master(struct net_device *slave, struct net_device *master)  	}  	slave->master = master; -	 +  	synchronize_net();  	if (old) @@ -2319,13 +2319,13 @@ void dev_set_promiscuity(struct net_device *dev, int inc)  		dev_mc_upload(dev);  		printk(KERN_INFO "device %s %s promiscuous mode\n",  		       dev->name, (dev->flags & IFF_PROMISC) ? "entered" : -		       					       "left"); +							       "left");  		audit_log(current->audit_context, GFP_ATOMIC,  			AUDIT_ANOM_PROMISCUOUS,  			"dev=%s prom=%d old_prom=%d auid=%u",  			dev->name, (dev->flags & IFF_PROMISC),  			(old_flags & IFF_PROMISC), -			audit_get_loginuid(current->audit_context));  +			audit_get_loginuid(current->audit_context));  	}  } @@ -2816,7 +2816,7 @@ int dev_ioctl(unsigned int cmd, void __user *arg)  				rtnl_unlock();  				if (IW_IS_GET(cmd) &&  				    copy_to_user(arg, &ifr, -					    	 sizeof(struct ifreq))) +						 sizeof(struct ifreq)))  					ret = -EFAULT;  				return ret;  			} @@ -2906,7 +2906,7 @@ int register_netdevice(struct net_device *dev)  			goto out;  		}  	} -  +  	if (!dev_valid_name(dev->name)) {  		ret = -EINVAL;  		goto out; @@ -2923,9 +2923,9 @@ int register_netdevice(struct net_device *dev)  			= hlist_entry(p, struct net_device, name_hlist);  		if (!strncmp(d->name, dev->name, IFNAMSIZ)) {  			ret = -EEXIST; - 			goto out; +			goto out;  		} - 	} +	}  	/* Fix illegal SG+CSUM combinations. */  	if ((dev->features & NETIF_F_SG) && @@ -3024,7 +3024,7 @@ int register_netdev(struct net_device *dev)  		if (err < 0)  			goto out;  	} -	 +  	err = register_netdevice(dev);  out:  	rtnl_unlock(); @@ -3041,7 +3041,7 @@ EXPORT_SYMBOL(register_netdev);   * for netdevice notification, and cleanup and put back the   * reference if they receive an UNREGISTER event.   * We can get stuck here if buggy protocols don't correctly - * call dev_put.  + * call dev_put.   */  static void netdev_wait_allrefs(struct net_device *dev)  { @@ -3205,8 +3205,8 @@ EXPORT_SYMBOL(alloc_netdev);   *	free_netdev - free network device   *	@dev: device   * - *	This function does the last stage of destroying an allocated device  - * 	interface. The reference to the device object is released.   + *	This function does the last stage of destroying an allocated device + * 	interface. The reference to the device object is released.   *	If this is the last reference then it will be freed.   */  void free_netdev(struct net_device *dev) @@ -3227,9 +3227,9 @@ void free_netdev(struct net_device *dev)  	kfree((char *)dev - dev->padded);  #endif  } -  +  /* Synchronize with packet receive processing. */ -void synchronize_net(void)  +void synchronize_net(void)  {  	might_sleep();  	synchronize_rcu(); @@ -3291,12 +3291,12 @@ void unregister_netdevice(struct net_device *dev)  	/* Shutdown queueing discipline. */  	dev_shutdown(dev); -	 +  	/* Notify protocols, that we are about to destroy  	   this device. They should clean all the things.  	*/  	raw_notifier_call_chain(&netdev_chain, NETDEV_UNREGISTER, dev); -	 +  	/*  	 *	Flush the multicast chain  	 */ @@ -3483,7 +3483,7 @@ static int __init net_dev_init(void)  		goto out;  	INIT_LIST_HEAD(&ptype_all); -	for (i = 0; i < 16; i++)  +	for (i = 0; i < 16; i++)  		INIT_LIST_HEAD(&ptype_base[i]);  	for (i = 0; i < ARRAY_SIZE(dev_name_head); i++)  |