diff options
Diffstat (limited to 'net/tipc/net.c')
| -rw-r--r-- | net/tipc/net.c | 126 | 
1 files changed, 63 insertions, 63 deletions
diff --git a/net/tipc/net.c b/net/tipc/net.c index 6826b493c1d..074891ad4f0 100644 --- a/net/tipc/net.c +++ b/net/tipc/net.c @@ -58,25 +58,25 @@   * 1: The routing hierarchy.   *    Comprises the structures 'zone', 'cluster', 'node', 'link'    *    and 'bearer'. The whole hierarchy is protected by a big  - *    read/write lock, net_lock, to enssure that nothing is added  + *    read/write lock, tipc_net_lock, to enssure that nothing is added    *    or removed while code is accessing any of these structures.    *    This layer must not be called from the two others while they    *    hold any of their own locks.   *    Neither must it itself do any upcalls to the other two before - *    it has released net_lock and other protective locks. + *    it has released tipc_net_lock and other protective locks.   * - *   Within the net_lock domain there are two sub-domains;'node' and  + *   Within the tipc_net_lock domain there are two sub-domains;'node' and    *   'bearer', where local write operations are permitted,   *   provided that those are protected by individual spin_locks - *   per instance. Code holding net_lock(read) and a node spin_lock  + *   per instance. Code holding tipc_net_lock(read) and a node spin_lock    *   is permitted to poke around in both the node itself and its   *   subordinate links. I.e, it can update link counters and queues,    *   change link state, send protocol messages, and alter the    *   "active_links" array in the node; but it can _not_ remove a link    *   or a node from the overall structure.   *   Correspondingly, individual bearers may change status within a  - *   net_lock(read), protected by an individual spin_lock ber bearer  - *   instance, but it needs net_lock(write) to remove/add any bearers. + *   tipc_net_lock(read), protected by an individual spin_lock ber bearer  + *   instance, but it needs tipc_net_lock(write) to remove/add any bearers.   *        *   *  2: The transport level of the protocol.  @@ -97,91 +97,91 @@   *       (Nobody is using read-only access to this, so it can just as    *       well be changed to a spin_lock)   *     - A spin lock to protect the registry of kernel/driver users (reg.c) - *     - A global spin_lock (port_lock), which only task is to ensure  + *     - A global spin_lock (tipc_port_lock), which only task is to ensure    *       consistency where more than one port is involved in an operation,   *       i.e., whe a port is part of a linked list of ports.   *       There are two such lists; 'port_list', which is used for management,   *       and 'wait_list', which is used to queue ports during congestion.   *        *  3: The name table (name_table.c, name_distr.c, subscription.c) - *     - There is one big read/write-lock (nametbl_lock) protecting the  + *     - There is one big read/write-lock (tipc_nametbl_lock) protecting the    *       overall name table structure. Nothing must be added/removed to    *       this structure without holding write access to it.   *     - There is one local spin_lock per sub_sequence, which can be seen - *       as a sub-domain to the nametbl_lock domain. It is used only + *       as a sub-domain to the tipc_nametbl_lock domain. It is used only   *       for translation operations, and is needed because a translation   *       steps the root of the 'publication' linked list between each lookup. - *       This is always used within the scope of a nametbl_lock(read). + *       This is always used within the scope of a tipc_nametbl_lock(read).   *     - A local spin_lock protecting the queue of subscriber events.  */ -rwlock_t net_lock = RW_LOCK_UNLOCKED; -struct network net = { 0 }; +rwlock_t tipc_net_lock = RW_LOCK_UNLOCKED; +struct network tipc_net = { 0 }; -struct node *net_select_remote_node(u32 addr, u32 ref)  +struct node *tipc_net_select_remote_node(u32 addr, u32 ref)   { -	return zone_select_remote_node(net.zones[tipc_zone(addr)], addr, ref); +	return tipc_zone_select_remote_node(tipc_net.zones[tipc_zone(addr)], addr, ref);  } -u32 net_select_router(u32 addr, u32 ref) +u32 tipc_net_select_router(u32 addr, u32 ref)  { -	return zone_select_router(net.zones[tipc_zone(addr)], addr, ref); +	return tipc_zone_select_router(tipc_net.zones[tipc_zone(addr)], addr, ref);  } -u32 net_next_node(u32 a) +u32 tipc_net_next_node(u32 a)  { -	if (net.zones[tipc_zone(a)]) -		return zone_next_node(a); +	if (tipc_net.zones[tipc_zone(a)]) +		return tipc_zone_next_node(a);  	return 0;  } -void net_remove_as_router(u32 router) +void tipc_net_remove_as_router(u32 router)  {  	u32 z_num;  	for (z_num = 1; z_num <= tipc_max_zones; z_num++) { -		if (!net.zones[z_num]) +		if (!tipc_net.zones[z_num])  			continue; -		zone_remove_as_router(net.zones[z_num], router); +		tipc_zone_remove_as_router(tipc_net.zones[z_num], router);  	}  } -void net_send_external_routes(u32 dest) +void tipc_net_send_external_routes(u32 dest)  {  	u32 z_num;  	for (z_num = 1; z_num <= tipc_max_zones; z_num++) { -		if (net.zones[z_num]) -			zone_send_external_routes(net.zones[z_num], dest); +		if (tipc_net.zones[z_num]) +			tipc_zone_send_external_routes(tipc_net.zones[z_num], dest);  	}  } -int net_init(void) +static int net_init(void)  {  	u32 sz = sizeof(struct _zone *) * (tipc_max_zones + 1); -	memset(&net, 0, sizeof(net)); -	net.zones = (struct _zone **)kmalloc(sz, GFP_ATOMIC); -	if (!net.zones) { +	memset(&tipc_net, 0, sizeof(tipc_net)); +	tipc_net.zones = (struct _zone **)kmalloc(sz, GFP_ATOMIC); +	if (!tipc_net.zones) {  		return -ENOMEM;  	} -	memset(net.zones, 0, sz); +	memset(tipc_net.zones, 0, sz);  	return TIPC_OK;  } -void net_stop(void) +static void net_stop(void)  {  	u32 z_num; -	if (!net.zones) +	if (!tipc_net.zones)  		return;  	for (z_num = 1; z_num <= tipc_max_zones; z_num++) { -		zone_delete(net.zones[z_num]); +		tipc_zone_delete(tipc_net.zones[z_num]);  	} -	kfree(net.zones); -	net.zones = 0; +	kfree(tipc_net.zones); +	tipc_net.zones = 0;  }  static void net_route_named_msg(struct sk_buff *buf) @@ -191,26 +191,26 @@ static void net_route_named_msg(struct sk_buff *buf)  	u32 dport;  	if (!msg_named(msg)) { -		msg_dbg(msg, "net->drop_nam:"); +		msg_dbg(msg, "tipc_net->drop_nam:");  		buf_discard(buf);  		return;  	}  	dnode = addr_domain(msg_lookup_scope(msg)); -	dport = nametbl_translate(msg_nametype(msg), msg_nameinst(msg), &dnode); -	dbg("net->lookup<%u,%u>-><%u,%x>\n", +	dport = tipc_nametbl_translate(msg_nametype(msg), msg_nameinst(msg), &dnode); +	dbg("tipc_net->lookup<%u,%u>-><%u,%x>\n",  	    msg_nametype(msg), msg_nameinst(msg), dport, dnode);  	if (dport) {  		msg_set_destnode(msg, dnode);  		msg_set_destport(msg, dport); -		net_route_msg(buf); +		tipc_net_route_msg(buf);  		return;  	} -	msg_dbg(msg, "net->rej:NO NAME: "); +	msg_dbg(msg, "tipc_net->rej:NO NAME: ");  	tipc_reject_msg(buf, TIPC_ERR_NO_NAME);  } -void net_route_msg(struct sk_buff *buf) +void tipc_net_route_msg(struct sk_buff *buf)  {  	struct tipc_msg *msg;  	u32 dnode; @@ -232,29 +232,29 @@ void net_route_msg(struct sk_buff *buf)  		return;  	} -	msg_dbg(msg, "net->rout: "); +	msg_dbg(msg, "tipc_net->rout: ");  	/* Handle message for this node */  	dnode = msg_short(msg) ? tipc_own_addr : msg_destnode(msg);  	if (in_scope(dnode, tipc_own_addr)) {  		if (msg_isdata(msg)) {  			if (msg_mcast(msg))  -				port_recv_mcast(buf, NULL); +				tipc_port_recv_mcast(buf, NULL);  			else if (msg_destport(msg)) -				port_recv_msg(buf); +				tipc_port_recv_msg(buf);  			else  				net_route_named_msg(buf);  			return;  		}  		switch (msg_user(msg)) {  		case ROUTE_DISTRIBUTOR: -			cluster_recv_routing_table(buf); +			tipc_cltr_recv_routing_table(buf);  			break;  		case NAME_DISTRIBUTOR: -			named_recv(buf); +			tipc_named_recv(buf);  			break;  		case CONN_MANAGER: -			port_recv_proto_msg(buf); +			tipc_port_recv_proto_msg(buf);  			break;  		default:  			msg_dbg(msg,"DROP/NET/<REC<"); @@ -265,10 +265,10 @@ void net_route_msg(struct sk_buff *buf)  	/* Handle message for another node */  	msg_dbg(msg, "NET>SEND>: "); -	link_send(buf, dnode, msg_link_selector(msg)); +	tipc_link_send(buf, dnode, msg_link_selector(msg));  } -int tipc_start_net(void) +int tipc_net_start(void)  {  	char addr_string[16];  	int res; @@ -277,35 +277,35 @@ int tipc_start_net(void)  		return -ENOPROTOOPT;  	tipc_mode = TIPC_NET_MODE; -	named_reinit(); -	port_reinit(); +	tipc_named_reinit(); +	tipc_port_reinit(); -	if ((res = bearer_init()) || +	if ((res = tipc_bearer_init()) ||  	    (res = net_init()) || -	    (res = cluster_init()) || -	    (res = bclink_init())) { +	    (res = tipc_cltr_init()) || +	    (res = tipc_bclink_init())) {  		return res;  	} -        subscr_stop(); -	cfg_stop(); -	k_signal((Handler)subscr_start, 0); -	k_signal((Handler)cfg_init, 0); +        tipc_subscr_stop(); +	tipc_cfg_stop(); +	tipc_k_signal((Handler)tipc_subscr_start, 0); +	tipc_k_signal((Handler)tipc_cfg_init, 0);  	info("Started in network mode\n");  	info("Own node address %s, network identity %u\n",  	     addr_string_fill(addr_string, tipc_own_addr), tipc_net_id);  	return TIPC_OK;  } -void tipc_stop_net(void) +void tipc_net_stop(void)  {  	if (tipc_mode != TIPC_NET_MODE)  		return; -        write_lock_bh(&net_lock); -	bearer_stop(); +        write_lock_bh(&tipc_net_lock); +	tipc_bearer_stop();  	tipc_mode = TIPC_NODE_MODE; -	bclink_stop(); +	tipc_bclink_stop();  	net_stop(); -        write_unlock_bh(&net_lock); +        write_unlock_bh(&tipc_net_lock);  	info("Left network mode \n");  }  |