mirror of
				https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
				synced 2025-10-31 06:18:54 +00:00 
			
		
		
		
	 a842ef297f
			
		
	
	
		a842ef297f
		
	
	
	
	
		
			
			This patch converts the combination of list_del(A) and list_add(A, B) to list_move(A, B) under net/rxrpc. Acked-by: David Howells <dhowells@redhat.com> Signed-off-by: Akinobu Mita <mita@miraclelinux.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
		
			
				
	
	
		
			2278 lines
		
	
	
		
			57 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			2278 lines
		
	
	
		
			57 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* call.c: Rx call routines
 | |
|  *
 | |
|  * Copyright (C) 2002 Red Hat, Inc. All Rights Reserved.
 | |
|  * Written by David Howells (dhowells@redhat.com)
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU General Public License
 | |
|  * as published by the Free Software Foundation; either version
 | |
|  * 2 of the License, or (at your option) any later version.
 | |
|  */
 | |
| 
 | |
| #include <linux/sched.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/module.h>
 | |
| #include <rxrpc/rxrpc.h>
 | |
| #include <rxrpc/transport.h>
 | |
| #include <rxrpc/peer.h>
 | |
| #include <rxrpc/connection.h>
 | |
| #include <rxrpc/call.h>
 | |
| #include <rxrpc/message.h>
 | |
| #include "internal.h"
 | |
| 
 | |
| __RXACCT_DECL(atomic_t rxrpc_call_count);
 | |
| __RXACCT_DECL(atomic_t rxrpc_message_count);
 | |
| 
 | |
| LIST_HEAD(rxrpc_calls);
 | |
| DECLARE_RWSEM(rxrpc_calls_sem);
 | |
| 
 | |
| unsigned rxrpc_call_rcv_timeout			= HZ/3;
 | |
| static unsigned rxrpc_call_acks_timeout		= HZ/3;
 | |
| static unsigned rxrpc_call_dfr_ack_timeout	= HZ/20;
 | |
| static unsigned short rxrpc_call_max_resend	= HZ/10;
 | |
| 
 | |
| const char *rxrpc_call_states[] = {
 | |
| 	"COMPLETE",
 | |
| 	"ERROR",
 | |
| 	"SRVR_RCV_OPID",
 | |
| 	"SRVR_RCV_ARGS",
 | |
| 	"SRVR_GOT_ARGS",
 | |
| 	"SRVR_SND_REPLY",
 | |
| 	"SRVR_RCV_FINAL_ACK",
 | |
| 	"CLNT_SND_ARGS",
 | |
| 	"CLNT_RCV_REPLY",
 | |
| 	"CLNT_GOT_REPLY"
 | |
| };
 | |
| 
 | |
| const char *rxrpc_call_error_states[] = {
 | |
| 	"NO_ERROR",
 | |
| 	"LOCAL_ABORT",
 | |
| 	"PEER_ABORT",
 | |
| 	"LOCAL_ERROR",
 | |
| 	"REMOTE_ERROR"
 | |
| };
 | |
| 
 | |
| const char *rxrpc_pkts[] = {
 | |
| 	"?00",
 | |
| 	"data", "ack", "busy", "abort", "ackall", "chall", "resp", "debug",
 | |
| 	"?09", "?10", "?11", "?12", "?13", "?14", "?15"
 | |
| };
 | |
| 
 | |
| static const char *rxrpc_acks[] = {
 | |
| 	"---", "REQ", "DUP", "SEQ", "WIN", "MEM", "PNG", "PNR", "DLY", "IDL",
 | |
| 	"-?-"
 | |
| };
 | |
| 
 | |
| static const char _acktype[] = "NA-";
 | |
| 
 | |
| static void rxrpc_call_receive_packet(struct rxrpc_call *call);
 | |
| static void rxrpc_call_receive_data_packet(struct rxrpc_call *call,
 | |
| 					   struct rxrpc_message *msg);
 | |
| static void rxrpc_call_receive_ack_packet(struct rxrpc_call *call,
 | |
| 					  struct rxrpc_message *msg);
 | |
| static void rxrpc_call_definitively_ACK(struct rxrpc_call *call,
 | |
| 					rxrpc_seq_t higest);
 | |
| static void rxrpc_call_resend(struct rxrpc_call *call, rxrpc_seq_t highest);
 | |
| static int __rxrpc_call_read_data(struct rxrpc_call *call);
 | |
| 
 | |
| static int rxrpc_call_record_ACK(struct rxrpc_call *call,
 | |
| 				 struct rxrpc_message *msg,
 | |
| 				 rxrpc_seq_t seq,
 | |
| 				 size_t count);
 | |
| 
 | |
| static int rxrpc_call_flush(struct rxrpc_call *call);
 | |
| 
 | |
| #define _state(call) \
 | |
| 	_debug("[[[ state %s ]]]", rxrpc_call_states[call->app_call_state]);
 | |
| 
 | |
| static void rxrpc_call_default_attn_func(struct rxrpc_call *call)
 | |
| {
 | |
| 	wake_up(&call->waitq);
 | |
| }
 | |
| 
 | |
| static void rxrpc_call_default_error_func(struct rxrpc_call *call)
 | |
| {
 | |
| 	wake_up(&call->waitq);
 | |
| }
 | |
| 
 | |
| static void rxrpc_call_default_aemap_func(struct rxrpc_call *call)
 | |
| {
 | |
| 	switch (call->app_err_state) {
 | |
| 	case RXRPC_ESTATE_LOCAL_ABORT:
 | |
| 		call->app_abort_code = -call->app_errno;
 | |
| 	case RXRPC_ESTATE_PEER_ABORT:
 | |
| 		call->app_errno = -ECONNABORTED;
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void __rxrpc_call_acks_timeout(unsigned long _call)
 | |
| {
 | |
| 	struct rxrpc_call *call = (struct rxrpc_call *) _call;
 | |
| 
 | |
| 	_debug("ACKS TIMEOUT %05lu", jiffies - call->cjif);
 | |
| 
 | |
| 	call->flags |= RXRPC_CALL_ACKS_TIMO;
 | |
| 	rxrpc_krxiod_queue_call(call);
 | |
| }
 | |
| 
 | |
| static void __rxrpc_call_rcv_timeout(unsigned long _call)
 | |
| {
 | |
| 	struct rxrpc_call *call = (struct rxrpc_call *) _call;
 | |
| 
 | |
| 	_debug("RCV TIMEOUT %05lu", jiffies - call->cjif);
 | |
| 
 | |
| 	call->flags |= RXRPC_CALL_RCV_TIMO;
 | |
| 	rxrpc_krxiod_queue_call(call);
 | |
| }
 | |
| 
 | |
| static void __rxrpc_call_ackr_timeout(unsigned long _call)
 | |
| {
 | |
| 	struct rxrpc_call *call = (struct rxrpc_call *) _call;
 | |
| 
 | |
| 	_debug("ACKR TIMEOUT %05lu",jiffies - call->cjif);
 | |
| 
 | |
| 	call->flags |= RXRPC_CALL_ACKR_TIMO;
 | |
| 	rxrpc_krxiod_queue_call(call);
 | |
| }
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * calculate a timeout based on an RTT value
 | |
|  */
 | |
| static inline unsigned long __rxrpc_rtt_based_timeout(struct rxrpc_call *call,
 | |
| 						      unsigned long val)
 | |
| {
 | |
| 	unsigned long expiry = call->conn->peer->rtt / (1000000 / HZ);
 | |
| 
 | |
| 	expiry += 10;
 | |
| 	if (expiry < HZ / 25)
 | |
| 		expiry = HZ / 25;
 | |
| 	if (expiry > HZ)
 | |
| 		expiry = HZ;
 | |
| 
 | |
| 	_leave(" = %lu jiffies", expiry);
 | |
| 	return jiffies + expiry;
 | |
| } /* end __rxrpc_rtt_based_timeout() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * create a new call record
 | |
|  */
 | |
| static inline int __rxrpc_create_call(struct rxrpc_connection *conn,
 | |
| 				      struct rxrpc_call **_call)
 | |
| {
 | |
| 	struct rxrpc_call *call;
 | |
| 
 | |
| 	_enter("%p", conn);
 | |
| 
 | |
| 	/* allocate and initialise a call record */
 | |
| 	call = (struct rxrpc_call *) get_zeroed_page(GFP_KERNEL);
 | |
| 	if (!call) {
 | |
| 		_leave(" ENOMEM");
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	atomic_set(&call->usage, 1);
 | |
| 
 | |
| 	init_waitqueue_head(&call->waitq);
 | |
| 	spin_lock_init(&call->lock);
 | |
| 	INIT_LIST_HEAD(&call->link);
 | |
| 	INIT_LIST_HEAD(&call->acks_pendq);
 | |
| 	INIT_LIST_HEAD(&call->rcv_receiveq);
 | |
| 	INIT_LIST_HEAD(&call->rcv_krxiodq_lk);
 | |
| 	INIT_LIST_HEAD(&call->app_readyq);
 | |
| 	INIT_LIST_HEAD(&call->app_unreadyq);
 | |
| 	INIT_LIST_HEAD(&call->app_link);
 | |
| 	INIT_LIST_HEAD(&call->app_attn_link);
 | |
| 
 | |
| 	init_timer(&call->acks_timeout);
 | |
| 	call->acks_timeout.data = (unsigned long) call;
 | |
| 	call->acks_timeout.function = __rxrpc_call_acks_timeout;
 | |
| 
 | |
| 	init_timer(&call->rcv_timeout);
 | |
| 	call->rcv_timeout.data = (unsigned long) call;
 | |
| 	call->rcv_timeout.function = __rxrpc_call_rcv_timeout;
 | |
| 
 | |
| 	init_timer(&call->ackr_dfr_timo);
 | |
| 	call->ackr_dfr_timo.data = (unsigned long) call;
 | |
| 	call->ackr_dfr_timo.function = __rxrpc_call_ackr_timeout;
 | |
| 
 | |
| 	call->conn = conn;
 | |
| 	call->ackr_win_bot = 1;
 | |
| 	call->ackr_win_top = call->ackr_win_bot + RXRPC_CALL_ACK_WINDOW_SIZE - 1;
 | |
| 	call->ackr_prev_seq = 0;
 | |
| 	call->app_mark = RXRPC_APP_MARK_EOF;
 | |
| 	call->app_attn_func = rxrpc_call_default_attn_func;
 | |
| 	call->app_error_func = rxrpc_call_default_error_func;
 | |
| 	call->app_aemap_func = rxrpc_call_default_aemap_func;
 | |
| 	call->app_scr_alloc = call->app_scratch;
 | |
| 
 | |
| 	call->cjif = jiffies;
 | |
| 
 | |
| 	_leave(" = 0 (%p)", call);
 | |
| 
 | |
| 	*_call = call;
 | |
| 
 | |
| 	return 0;
 | |
| } /* end __rxrpc_create_call() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * create a new call record for outgoing calls
 | |
|  */
 | |
| int rxrpc_create_call(struct rxrpc_connection *conn,
 | |
| 		      rxrpc_call_attn_func_t attn,
 | |
| 		      rxrpc_call_error_func_t error,
 | |
| 		      rxrpc_call_aemap_func_t aemap,
 | |
| 		      struct rxrpc_call **_call)
 | |
| {
 | |
| 	DECLARE_WAITQUEUE(myself, current);
 | |
| 
 | |
| 	struct rxrpc_call *call;
 | |
| 	int ret, cix, loop;
 | |
| 
 | |
| 	_enter("%p", conn);
 | |
| 
 | |
| 	/* allocate and initialise a call record */
 | |
| 	ret = __rxrpc_create_call(conn, &call);
 | |
| 	if (ret < 0) {
 | |
| 		_leave(" = %d", ret);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	call->app_call_state = RXRPC_CSTATE_CLNT_SND_ARGS;
 | |
| 	if (attn)
 | |
| 		call->app_attn_func = attn;
 | |
| 	if (error)
 | |
| 		call->app_error_func = error;
 | |
| 	if (aemap)
 | |
| 		call->app_aemap_func = aemap;
 | |
| 
 | |
| 	_state(call);
 | |
| 
 | |
| 	spin_lock(&conn->lock);
 | |
| 	set_current_state(TASK_INTERRUPTIBLE);
 | |
| 	add_wait_queue(&conn->chanwait, &myself);
 | |
| 
 | |
|  try_again:
 | |
| 	/* try to find an unused channel */
 | |
| 	for (cix = 0; cix < 4; cix++)
 | |
| 		if (!conn->channels[cix])
 | |
| 			goto obtained_chan;
 | |
| 
 | |
| 	/* no free channels - wait for one to become available */
 | |
| 	ret = -EINTR;
 | |
| 	if (signal_pending(current))
 | |
| 		goto error_unwait;
 | |
| 
 | |
| 	spin_unlock(&conn->lock);
 | |
| 
 | |
| 	schedule();
 | |
| 	set_current_state(TASK_INTERRUPTIBLE);
 | |
| 
 | |
| 	spin_lock(&conn->lock);
 | |
| 	goto try_again;
 | |
| 
 | |
| 	/* got a channel - now attach to the connection */
 | |
|  obtained_chan:
 | |
| 	remove_wait_queue(&conn->chanwait, &myself);
 | |
| 	set_current_state(TASK_RUNNING);
 | |
| 
 | |
| 	/* concoct a unique call number */
 | |
|  next_callid:
 | |
| 	call->call_id = htonl(++conn->call_counter);
 | |
| 	for (loop = 0; loop < 4; loop++)
 | |
| 		if (conn->channels[loop] &&
 | |
| 		    conn->channels[loop]->call_id == call->call_id)
 | |
| 			goto next_callid;
 | |
| 
 | |
| 	rxrpc_get_connection(conn);
 | |
| 	conn->channels[cix] = call; /* assign _after_ done callid check loop */
 | |
| 	do_gettimeofday(&conn->atime);
 | |
| 	call->chan_ix = htonl(cix);
 | |
| 
 | |
| 	spin_unlock(&conn->lock);
 | |
| 
 | |
| 	down_write(&rxrpc_calls_sem);
 | |
| 	list_add_tail(&call->call_link, &rxrpc_calls);
 | |
| 	up_write(&rxrpc_calls_sem);
 | |
| 
 | |
| 	__RXACCT(atomic_inc(&rxrpc_call_count));
 | |
| 	*_call = call;
 | |
| 
 | |
| 	_leave(" = 0 (call=%p cix=%u)", call, cix);
 | |
| 	return 0;
 | |
| 
 | |
|  error_unwait:
 | |
| 	remove_wait_queue(&conn->chanwait, &myself);
 | |
| 	set_current_state(TASK_RUNNING);
 | |
| 	spin_unlock(&conn->lock);
 | |
| 
 | |
| 	free_page((unsigned long) call);
 | |
| 	_leave(" = %d", ret);
 | |
| 	return ret;
 | |
| } /* end rxrpc_create_call() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * create a new call record for incoming calls
 | |
|  */
 | |
| int rxrpc_incoming_call(struct rxrpc_connection *conn,
 | |
| 			struct rxrpc_message *msg,
 | |
| 			struct rxrpc_call **_call)
 | |
| {
 | |
| 	struct rxrpc_call *call;
 | |
| 	unsigned cix;
 | |
| 	int ret;
 | |
| 
 | |
| 	cix = ntohl(msg->hdr.cid) & RXRPC_CHANNELMASK;
 | |
| 
 | |
| 	_enter("%p,%u,%u", conn, ntohl(msg->hdr.callNumber), cix);
 | |
| 
 | |
| 	/* allocate and initialise a call record */
 | |
| 	ret = __rxrpc_create_call(conn, &call);
 | |
| 	if (ret < 0) {
 | |
| 		_leave(" = %d", ret);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	call->pkt_rcv_count = 1;
 | |
| 	call->app_call_state = RXRPC_CSTATE_SRVR_RCV_OPID;
 | |
| 	call->app_mark = sizeof(uint32_t);
 | |
| 
 | |
| 	_state(call);
 | |
| 
 | |
| 	/* attach to the connection */
 | |
| 	ret = -EBUSY;
 | |
| 	call->chan_ix = htonl(cix);
 | |
| 	call->call_id = msg->hdr.callNumber;
 | |
| 
 | |
| 	spin_lock(&conn->lock);
 | |
| 
 | |
| 	if (!conn->channels[cix] ||
 | |
| 	    conn->channels[cix]->app_call_state == RXRPC_CSTATE_COMPLETE ||
 | |
| 	    conn->channels[cix]->app_call_state == RXRPC_CSTATE_ERROR
 | |
| 	    ) {
 | |
| 		conn->channels[cix] = call;
 | |
| 		rxrpc_get_connection(conn);
 | |
| 		ret = 0;
 | |
| 	}
 | |
| 
 | |
| 	spin_unlock(&conn->lock);
 | |
| 
 | |
| 	if (ret < 0) {
 | |
| 		free_page((unsigned long) call);
 | |
| 		call = NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (ret == 0) {
 | |
| 		down_write(&rxrpc_calls_sem);
 | |
| 		list_add_tail(&call->call_link, &rxrpc_calls);
 | |
| 		up_write(&rxrpc_calls_sem);
 | |
| 		__RXACCT(atomic_inc(&rxrpc_call_count));
 | |
| 		*_call = call;
 | |
| 	}
 | |
| 
 | |
| 	_leave(" = %d [%p]", ret, call);
 | |
| 	return ret;
 | |
| } /* end rxrpc_incoming_call() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * free a call record
 | |
|  */
 | |
| void rxrpc_put_call(struct rxrpc_call *call)
 | |
| {
 | |
| 	struct rxrpc_connection *conn = call->conn;
 | |
| 	struct rxrpc_message *msg;
 | |
| 
 | |
| 	_enter("%p{u=%d}",call,atomic_read(&call->usage));
 | |
| 
 | |
| 	/* sanity check */
 | |
| 	if (atomic_read(&call->usage) <= 0)
 | |
| 		BUG();
 | |
| 
 | |
| 	/* to prevent a race, the decrement and the de-list must be effectively
 | |
| 	 * atomic */
 | |
| 	spin_lock(&conn->lock);
 | |
| 	if (likely(!atomic_dec_and_test(&call->usage))) {
 | |
| 		spin_unlock(&conn->lock);
 | |
| 		_leave("");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (conn->channels[ntohl(call->chan_ix)] == call)
 | |
| 		conn->channels[ntohl(call->chan_ix)] = NULL;
 | |
| 
 | |
| 	spin_unlock(&conn->lock);
 | |
| 
 | |
| 	wake_up(&conn->chanwait);
 | |
| 
 | |
| 	rxrpc_put_connection(conn);
 | |
| 
 | |
| 	/* clear the timers and dequeue from krxiod */
 | |
| 	del_timer_sync(&call->acks_timeout);
 | |
| 	del_timer_sync(&call->rcv_timeout);
 | |
| 	del_timer_sync(&call->ackr_dfr_timo);
 | |
| 
 | |
| 	rxrpc_krxiod_dequeue_call(call);
 | |
| 
 | |
| 	/* clean up the contents of the struct */
 | |
| 	if (call->snd_nextmsg)
 | |
| 		rxrpc_put_message(call->snd_nextmsg);
 | |
| 
 | |
| 	if (call->snd_ping)
 | |
| 		rxrpc_put_message(call->snd_ping);
 | |
| 
 | |
| 	while (!list_empty(&call->acks_pendq)) {
 | |
| 		msg = list_entry(call->acks_pendq.next,
 | |
| 				 struct rxrpc_message, link);
 | |
| 		list_del(&msg->link);
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
| 	while (!list_empty(&call->rcv_receiveq)) {
 | |
| 		msg = list_entry(call->rcv_receiveq.next,
 | |
| 				 struct rxrpc_message, link);
 | |
| 		list_del(&msg->link);
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
| 	while (!list_empty(&call->app_readyq)) {
 | |
| 		msg = list_entry(call->app_readyq.next,
 | |
| 				 struct rxrpc_message, link);
 | |
| 		list_del(&msg->link);
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
| 	while (!list_empty(&call->app_unreadyq)) {
 | |
| 		msg = list_entry(call->app_unreadyq.next,
 | |
| 				 struct rxrpc_message, link);
 | |
| 		list_del(&msg->link);
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
| 	module_put(call->owner);
 | |
| 
 | |
| 	down_write(&rxrpc_calls_sem);
 | |
| 	list_del(&call->call_link);
 | |
| 	up_write(&rxrpc_calls_sem);
 | |
| 
 | |
| 	__RXACCT(atomic_dec(&rxrpc_call_count));
 | |
| 	free_page((unsigned long) call);
 | |
| 
 | |
| 	_leave(" [destroyed]");
 | |
| } /* end rxrpc_put_call() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * actually generate a normal ACK
 | |
|  */
 | |
| static inline int __rxrpc_call_gen_normal_ACK(struct rxrpc_call *call,
 | |
| 					      rxrpc_seq_t seq)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	struct kvec diov[3];
 | |
| 	__be32 aux[4];
 | |
| 	int delta, ret;
 | |
| 
 | |
| 	/* ACKs default to DELAY */
 | |
| 	if (!call->ackr.reason)
 | |
| 		call->ackr.reason = RXRPC_ACK_DELAY;
 | |
| 
 | |
| 	_proto("Rx %05lu Sending ACK { m=%hu f=#%u p=#%u s=%%%u r=%s n=%u }",
 | |
| 	       jiffies - call->cjif,
 | |
| 	       ntohs(call->ackr.maxSkew),
 | |
| 	       ntohl(call->ackr.firstPacket),
 | |
| 	       ntohl(call->ackr.previousPacket),
 | |
| 	       ntohl(call->ackr.serial),
 | |
| 	       rxrpc_acks[call->ackr.reason],
 | |
| 	       call->ackr.nAcks);
 | |
| 
 | |
| 	aux[0] = htonl(call->conn->peer->if_mtu);	/* interface MTU */
 | |
| 	aux[1] = htonl(1444);				/* max MTU */
 | |
| 	aux[2] = htonl(16);				/* rwind */
 | |
| 	aux[3] = htonl(4);				/* max packets */
 | |
| 
 | |
| 	diov[0].iov_len  = sizeof(struct rxrpc_ackpacket);
 | |
| 	diov[0].iov_base = &call->ackr;
 | |
| 	diov[1].iov_len  = call->ackr_pend_cnt + 3;
 | |
| 	diov[1].iov_base = call->ackr_array;
 | |
| 	diov[2].iov_len  = sizeof(aux);
 | |
| 	diov[2].iov_base = &aux;
 | |
| 
 | |
| 	/* build and send the message */
 | |
| 	ret = rxrpc_conn_newmsg(call->conn,call, RXRPC_PACKET_TYPE_ACK,
 | |
| 				3, diov, GFP_KERNEL, &msg);
 | |
| 	if (ret < 0)
 | |
| 		goto out;
 | |
| 
 | |
| 	msg->seq = seq;
 | |
| 	msg->hdr.seq = htonl(seq);
 | |
| 	msg->hdr.flags |= RXRPC_SLOW_START_OK;
 | |
| 
 | |
| 	ret = rxrpc_conn_sendmsg(call->conn, msg);
 | |
| 	rxrpc_put_message(msg);
 | |
| 	if (ret < 0)
 | |
| 		goto out;
 | |
| 	call->pkt_snd_count++;
 | |
| 
 | |
| 	/* count how many actual ACKs there were at the front */
 | |
| 	for (delta = 0; delta < call->ackr_pend_cnt; delta++)
 | |
| 		if (call->ackr_array[delta] != RXRPC_ACK_TYPE_ACK)
 | |
| 			break;
 | |
| 
 | |
| 	call->ackr_pend_cnt -= delta; /* all ACK'd to this point */
 | |
| 
 | |
| 	/* crank the ACK window around */
 | |
| 	if (delta == 0) {
 | |
| 		/* un-ACK'd window */
 | |
| 	}
 | |
| 	else if (delta < RXRPC_CALL_ACK_WINDOW_SIZE) {
 | |
| 		/* partially ACK'd window
 | |
| 		 * - shuffle down to avoid losing out-of-sequence packets
 | |
| 		 */
 | |
| 		call->ackr_win_bot += delta;
 | |
| 		call->ackr_win_top += delta;
 | |
| 
 | |
| 		memmove(&call->ackr_array[0],
 | |
| 			&call->ackr_array[delta],
 | |
| 			call->ackr_pend_cnt);
 | |
| 
 | |
| 		memset(&call->ackr_array[call->ackr_pend_cnt],
 | |
| 		       RXRPC_ACK_TYPE_NACK,
 | |
| 		       sizeof(call->ackr_array) - call->ackr_pend_cnt);
 | |
| 	}
 | |
| 	else {
 | |
| 		/* fully ACK'd window
 | |
| 		 * - just clear the whole thing
 | |
| 		 */
 | |
| 		memset(&call->ackr_array,
 | |
| 		       RXRPC_ACK_TYPE_NACK,
 | |
| 		       sizeof(call->ackr_array));
 | |
| 	}
 | |
| 
 | |
| 	/* clear this ACK */
 | |
| 	memset(&call->ackr, 0, sizeof(call->ackr));
 | |
| 
 | |
|  out:
 | |
| 	if (!call->app_call_state)
 | |
| 		printk("___ STATE 0 ___\n");
 | |
| 	return ret;
 | |
| } /* end __rxrpc_call_gen_normal_ACK() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * note the reception of a packet in the call's ACK records and generate an
 | |
|  * appropriate ACK packet if necessary
 | |
|  * - returns 0 if packet should be processed, 1 if packet should be ignored
 | |
|  *   and -ve on an error
 | |
|  */
 | |
| static int rxrpc_call_generate_ACK(struct rxrpc_call *call,
 | |
| 				   struct rxrpc_header *hdr,
 | |
| 				   struct rxrpc_ackpacket *ack)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	rxrpc_seq_t seq;
 | |
| 	unsigned offset;
 | |
| 	int ret = 0, err;
 | |
| 	u8 special_ACK, do_ACK, force;
 | |
| 
 | |
| 	_enter("%p,%p { seq=%d tp=%d fl=%02x }",
 | |
| 	       call, hdr, ntohl(hdr->seq), hdr->type, hdr->flags);
 | |
| 
 | |
| 	seq = ntohl(hdr->seq);
 | |
| 	offset = seq - call->ackr_win_bot;
 | |
| 	do_ACK = RXRPC_ACK_DELAY;
 | |
| 	special_ACK = 0;
 | |
| 	force = (seq == 1);
 | |
| 
 | |
| 	if (call->ackr_high_seq < seq)
 | |
| 		call->ackr_high_seq = seq;
 | |
| 
 | |
| 	/* deal with generation of obvious special ACKs first */
 | |
| 	if (ack && ack->reason == RXRPC_ACK_PING) {
 | |
| 		special_ACK = RXRPC_ACK_PING_RESPONSE;
 | |
| 		ret = 1;
 | |
| 		goto gen_ACK;
 | |
| 	}
 | |
| 
 | |
| 	if (seq < call->ackr_win_bot) {
 | |
| 		special_ACK = RXRPC_ACK_DUPLICATE;
 | |
| 		ret = 1;
 | |
| 		goto gen_ACK;
 | |
| 	}
 | |
| 
 | |
| 	if (seq >= call->ackr_win_top) {
 | |
| 		special_ACK = RXRPC_ACK_EXCEEDS_WINDOW;
 | |
| 		ret = 1;
 | |
| 		goto gen_ACK;
 | |
| 	}
 | |
| 
 | |
| 	if (call->ackr_array[offset] != RXRPC_ACK_TYPE_NACK) {
 | |
| 		special_ACK = RXRPC_ACK_DUPLICATE;
 | |
| 		ret = 1;
 | |
| 		goto gen_ACK;
 | |
| 	}
 | |
| 
 | |
| 	/* okay... it's a normal data packet inside the ACK window */
 | |
| 	call->ackr_array[offset] = RXRPC_ACK_TYPE_ACK;
 | |
| 
 | |
| 	if (offset < call->ackr_pend_cnt) {
 | |
| 	}
 | |
| 	else if (offset > call->ackr_pend_cnt) {
 | |
| 		do_ACK = RXRPC_ACK_OUT_OF_SEQUENCE;
 | |
| 		call->ackr_pend_cnt = offset;
 | |
| 		goto gen_ACK;
 | |
| 	}
 | |
| 
 | |
| 	if (hdr->flags & RXRPC_REQUEST_ACK) {
 | |
| 		do_ACK = RXRPC_ACK_REQUESTED;
 | |
| 	}
 | |
| 
 | |
| 	/* generate an ACK on the final packet of a reply just received */
 | |
| 	if (hdr->flags & RXRPC_LAST_PACKET) {
 | |
| 		if (call->conn->out_clientflag)
 | |
| 			force = 1;
 | |
| 	}
 | |
| 	else if (!(hdr->flags & RXRPC_MORE_PACKETS)) {
 | |
| 		do_ACK = RXRPC_ACK_REQUESTED;
 | |
| 	}
 | |
| 
 | |
| 	/* re-ACK packets previously received out-of-order */
 | |
| 	for (offset++; offset < RXRPC_CALL_ACK_WINDOW_SIZE; offset++)
 | |
| 		if (call->ackr_array[offset] != RXRPC_ACK_TYPE_ACK)
 | |
| 			break;
 | |
| 
 | |
| 	call->ackr_pend_cnt = offset;
 | |
| 
 | |
| 	/* generate an ACK if we fill up the window */
 | |
| 	if (call->ackr_pend_cnt >= RXRPC_CALL_ACK_WINDOW_SIZE)
 | |
| 		force = 1;
 | |
| 
 | |
|  gen_ACK:
 | |
| 	_debug("%05lu ACKs pend=%u norm=%s special=%s%s",
 | |
| 	       jiffies - call->cjif,
 | |
| 	       call->ackr_pend_cnt,
 | |
| 	       rxrpc_acks[do_ACK],
 | |
| 	       rxrpc_acks[special_ACK],
 | |
| 	       force ? " immediate" :
 | |
| 	       do_ACK == RXRPC_ACK_REQUESTED ? " merge-req" :
 | |
| 	       hdr->flags & RXRPC_LAST_PACKET ? " finalise" :
 | |
| 	       " defer"
 | |
| 	       );
 | |
| 
 | |
| 	/* send any pending normal ACKs if need be */
 | |
| 	if (call->ackr_pend_cnt > 0) {
 | |
| 		/* fill out the appropriate form */
 | |
| 		call->ackr.bufferSpace	= htons(RXRPC_CALL_ACK_WINDOW_SIZE);
 | |
| 		call->ackr.maxSkew	= htons(min(call->ackr_high_seq - seq,
 | |
| 						    65535U));
 | |
| 		call->ackr.firstPacket	= htonl(call->ackr_win_bot);
 | |
| 		call->ackr.previousPacket = call->ackr_prev_seq;
 | |
| 		call->ackr.serial	= hdr->serial;
 | |
| 		call->ackr.nAcks	= call->ackr_pend_cnt;
 | |
| 
 | |
| 		if (do_ACK == RXRPC_ACK_REQUESTED)
 | |
| 			call->ackr.reason = do_ACK;
 | |
| 
 | |
| 		/* generate the ACK immediately if necessary */
 | |
| 		if (special_ACK || force) {
 | |
| 			err = __rxrpc_call_gen_normal_ACK(
 | |
| 				call, do_ACK == RXRPC_ACK_DELAY ? 0 : seq);
 | |
| 			if (err < 0) {
 | |
| 				ret = err;
 | |
| 				goto out;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (call->ackr.reason == RXRPC_ACK_REQUESTED)
 | |
| 		call->ackr_dfr_seq = seq;
 | |
| 
 | |
| 	/* start the ACK timer if not running if there are any pending deferred
 | |
| 	 * ACKs */
 | |
| 	if (call->ackr_pend_cnt > 0 &&
 | |
| 	    call->ackr.reason != RXRPC_ACK_REQUESTED &&
 | |
| 	    !timer_pending(&call->ackr_dfr_timo)
 | |
| 	    ) {
 | |
| 		unsigned long timo;
 | |
| 
 | |
| 		timo = rxrpc_call_dfr_ack_timeout + jiffies;
 | |
| 
 | |
| 		_debug("START ACKR TIMER for cj=%lu", timo - call->cjif);
 | |
| 
 | |
| 		spin_lock(&call->lock);
 | |
| 		mod_timer(&call->ackr_dfr_timo, timo);
 | |
| 		spin_unlock(&call->lock);
 | |
| 	}
 | |
| 	else if ((call->ackr_pend_cnt == 0 ||
 | |
| 		  call->ackr.reason == RXRPC_ACK_REQUESTED) &&
 | |
| 		 timer_pending(&call->ackr_dfr_timo)
 | |
| 		 ) {
 | |
| 		/* stop timer if no pending ACKs */
 | |
| 		_debug("CLEAR ACKR TIMER");
 | |
| 		del_timer_sync(&call->ackr_dfr_timo);
 | |
| 	}
 | |
| 
 | |
| 	/* send a special ACK if one is required */
 | |
| 	if (special_ACK) {
 | |
| 		struct rxrpc_ackpacket ack;
 | |
| 		struct kvec diov[2];
 | |
| 		uint8_t acks[1] = { RXRPC_ACK_TYPE_ACK };
 | |
| 
 | |
| 		/* fill out the appropriate form */
 | |
| 		ack.bufferSpace	= htons(RXRPC_CALL_ACK_WINDOW_SIZE);
 | |
| 		ack.maxSkew	= htons(min(call->ackr_high_seq - seq,
 | |
| 					    65535U));
 | |
| 		ack.firstPacket	= htonl(call->ackr_win_bot);
 | |
| 		ack.previousPacket = call->ackr_prev_seq;
 | |
| 		ack.serial	= hdr->serial;
 | |
| 		ack.reason	= special_ACK;
 | |
| 		ack.nAcks	= 0;
 | |
| 
 | |
| 		_proto("Rx Sending s-ACK"
 | |
| 		       " { m=%hu f=#%u p=#%u s=%%%u r=%s n=%u }",
 | |
| 		       ntohs(ack.maxSkew),
 | |
| 		       ntohl(ack.firstPacket),
 | |
| 		       ntohl(ack.previousPacket),
 | |
| 		       ntohl(ack.serial),
 | |
| 		       rxrpc_acks[ack.reason],
 | |
| 		       ack.nAcks);
 | |
| 
 | |
| 		diov[0].iov_len  = sizeof(struct rxrpc_ackpacket);
 | |
| 		diov[0].iov_base = &ack;
 | |
| 		diov[1].iov_len  = sizeof(acks);
 | |
| 		diov[1].iov_base = acks;
 | |
| 
 | |
| 		/* build and send the message */
 | |
| 		err = rxrpc_conn_newmsg(call->conn,call, RXRPC_PACKET_TYPE_ACK,
 | |
| 					hdr->seq ? 2 : 1, diov,
 | |
| 					GFP_KERNEL,
 | |
| 					&msg);
 | |
| 		if (err < 0) {
 | |
| 			ret = err;
 | |
| 			goto out;
 | |
| 		}
 | |
| 
 | |
| 		msg->seq = seq;
 | |
| 		msg->hdr.seq = htonl(seq);
 | |
| 		msg->hdr.flags |= RXRPC_SLOW_START_OK;
 | |
| 
 | |
| 		err = rxrpc_conn_sendmsg(call->conn, msg);
 | |
| 		rxrpc_put_message(msg);
 | |
| 		if (err < 0) {
 | |
| 			ret = err;
 | |
| 			goto out;
 | |
| 		}
 | |
| 		call->pkt_snd_count++;
 | |
| 	}
 | |
| 
 | |
|  out:
 | |
| 	if (hdr->seq)
 | |
| 		call->ackr_prev_seq = hdr->seq;
 | |
| 
 | |
| 	_leave(" = %d", ret);
 | |
| 	return ret;
 | |
| } /* end rxrpc_call_generate_ACK() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * handle work to be done on a call
 | |
|  * - includes packet reception and timeout processing
 | |
|  */
 | |
| void rxrpc_call_do_stuff(struct rxrpc_call *call)
 | |
| {
 | |
| 	_enter("%p{flags=%lx}", call, call->flags);
 | |
| 
 | |
| 	/* handle packet reception */
 | |
| 	if (call->flags & RXRPC_CALL_RCV_PKT) {
 | |
| 		_debug("- receive packet");
 | |
| 		call->flags &= ~RXRPC_CALL_RCV_PKT;
 | |
| 		rxrpc_call_receive_packet(call);
 | |
| 	}
 | |
| 
 | |
| 	/* handle overdue ACKs */
 | |
| 	if (call->flags & RXRPC_CALL_ACKS_TIMO) {
 | |
| 		_debug("- overdue ACK timeout");
 | |
| 		call->flags &= ~RXRPC_CALL_ACKS_TIMO;
 | |
| 		rxrpc_call_resend(call, call->snd_seq_count);
 | |
| 	}
 | |
| 
 | |
| 	/* handle lack of reception */
 | |
| 	if (call->flags & RXRPC_CALL_RCV_TIMO) {
 | |
| 		_debug("- reception timeout");
 | |
| 		call->flags &= ~RXRPC_CALL_RCV_TIMO;
 | |
| 		rxrpc_call_abort(call, -EIO);
 | |
| 	}
 | |
| 
 | |
| 	/* handle deferred ACKs */
 | |
| 	if (call->flags & RXRPC_CALL_ACKR_TIMO ||
 | |
| 	    (call->ackr.nAcks > 0 && call->ackr.reason == RXRPC_ACK_REQUESTED)
 | |
| 	    ) {
 | |
| 		_debug("- deferred ACK timeout: cj=%05lu r=%s n=%u",
 | |
| 		       jiffies - call->cjif,
 | |
| 		       rxrpc_acks[call->ackr.reason],
 | |
| 		       call->ackr.nAcks);
 | |
| 
 | |
| 		call->flags &= ~RXRPC_CALL_ACKR_TIMO;
 | |
| 
 | |
| 		if (call->ackr.nAcks > 0 &&
 | |
| 		    call->app_call_state != RXRPC_CSTATE_ERROR) {
 | |
| 			/* generate ACK */
 | |
| 			__rxrpc_call_gen_normal_ACK(call, call->ackr_dfr_seq);
 | |
| 			call->ackr_dfr_seq = 0;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	_leave("");
 | |
| 
 | |
| } /* end rxrpc_call_do_stuff() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * send an abort message at call or connection level
 | |
|  * - must be called with call->lock held
 | |
|  * - the supplied error code is sent as the packet data
 | |
|  */
 | |
| static int __rxrpc_call_abort(struct rxrpc_call *call, int errno)
 | |
| {
 | |
| 	struct rxrpc_connection *conn = call->conn;
 | |
| 	struct rxrpc_message *msg;
 | |
| 	struct kvec diov[1];
 | |
| 	int ret;
 | |
| 	__be32 _error;
 | |
| 
 | |
| 	_enter("%p{%08x},%p{%d},%d",
 | |
| 	       conn, ntohl(conn->conn_id), call, ntohl(call->call_id), errno);
 | |
| 
 | |
| 	/* if this call is already aborted, then just wake up any waiters */
 | |
| 	if (call->app_call_state == RXRPC_CSTATE_ERROR) {
 | |
| 		spin_unlock(&call->lock);
 | |
| 		call->app_error_func(call);
 | |
| 		_leave(" = 0");
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	rxrpc_get_call(call);
 | |
| 
 | |
| 	/* change the state _with_ the lock still held */
 | |
| 	call->app_call_state	= RXRPC_CSTATE_ERROR;
 | |
| 	call->app_err_state	= RXRPC_ESTATE_LOCAL_ABORT;
 | |
| 	call->app_errno		= errno;
 | |
| 	call->app_mark		= RXRPC_APP_MARK_EOF;
 | |
| 	call->app_read_buf	= NULL;
 | |
| 	call->app_async_read	= 0;
 | |
| 
 | |
| 	_state(call);
 | |
| 
 | |
| 	/* ask the app to translate the error code */
 | |
| 	call->app_aemap_func(call);
 | |
| 
 | |
| 	spin_unlock(&call->lock);
 | |
| 
 | |
| 	/* flush any outstanding ACKs */
 | |
| 	del_timer_sync(&call->acks_timeout);
 | |
| 	del_timer_sync(&call->rcv_timeout);
 | |
| 	del_timer_sync(&call->ackr_dfr_timo);
 | |
| 
 | |
| 	if (rxrpc_call_is_ack_pending(call))
 | |
| 		__rxrpc_call_gen_normal_ACK(call, 0);
 | |
| 
 | |
| 	/* send the abort packet only if we actually traded some other
 | |
| 	 * packets */
 | |
| 	ret = 0;
 | |
| 	if (call->pkt_snd_count || call->pkt_rcv_count) {
 | |
| 		/* actually send the abort */
 | |
| 		_proto("Rx Sending Call ABORT { data=%d }",
 | |
| 		       call->app_abort_code);
 | |
| 
 | |
| 		_error = htonl(call->app_abort_code);
 | |
| 
 | |
| 		diov[0].iov_len  = sizeof(_error);
 | |
| 		diov[0].iov_base = &_error;
 | |
| 
 | |
| 		ret = rxrpc_conn_newmsg(conn, call, RXRPC_PACKET_TYPE_ABORT,
 | |
| 					1, diov, GFP_KERNEL, &msg);
 | |
| 		if (ret == 0) {
 | |
| 			ret = rxrpc_conn_sendmsg(conn, msg);
 | |
| 			rxrpc_put_message(msg);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* tell the app layer to let go */
 | |
| 	call->app_error_func(call);
 | |
| 
 | |
| 	rxrpc_put_call(call);
 | |
| 
 | |
| 	_leave(" = %d", ret);
 | |
| 	return ret;
 | |
| } /* end __rxrpc_call_abort() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * send an abort message at call or connection level
 | |
|  * - the supplied error code is sent as the packet data
 | |
|  */
 | |
| int rxrpc_call_abort(struct rxrpc_call *call, int error)
 | |
| {
 | |
| 	spin_lock(&call->lock);
 | |
| 
 | |
| 	return __rxrpc_call_abort(call, error);
 | |
| 
 | |
| } /* end rxrpc_call_abort() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * process packets waiting for this call
 | |
|  */
 | |
| static void rxrpc_call_receive_packet(struct rxrpc_call *call)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	struct list_head *_p;
 | |
| 
 | |
| 	_enter("%p", call);
 | |
| 
 | |
| 	rxrpc_get_call(call); /* must not go away too soon if aborted by
 | |
| 			       * app-layer */
 | |
| 
 | |
| 	while (!list_empty(&call->rcv_receiveq)) {
 | |
| 		/* try to get next packet */
 | |
| 		_p = NULL;
 | |
| 		spin_lock(&call->lock);
 | |
| 		if (!list_empty(&call->rcv_receiveq)) {
 | |
| 			_p = call->rcv_receiveq.next;
 | |
| 			list_del_init(_p);
 | |
| 		}
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		if (!_p)
 | |
| 			break;
 | |
| 
 | |
| 		msg = list_entry(_p, struct rxrpc_message, link);
 | |
| 
 | |
| 		_proto("Rx %05lu Received %s packet (%%%u,#%u,%c%c%c%c%c)",
 | |
| 		       jiffies - call->cjif,
 | |
| 		       rxrpc_pkts[msg->hdr.type],
 | |
| 		       ntohl(msg->hdr.serial),
 | |
| 		       msg->seq,
 | |
| 		       msg->hdr.flags & RXRPC_JUMBO_PACKET	? 'j' : '-',
 | |
| 		       msg->hdr.flags & RXRPC_MORE_PACKETS	? 'm' : '-',
 | |
| 		       msg->hdr.flags & RXRPC_LAST_PACKET	? 'l' : '-',
 | |
| 		       msg->hdr.flags & RXRPC_REQUEST_ACK	? 'r' : '-',
 | |
| 		       msg->hdr.flags & RXRPC_CLIENT_INITIATED	? 'C' : 'S'
 | |
| 		       );
 | |
| 
 | |
| 		switch (msg->hdr.type) {
 | |
| 			/* deal with data packets */
 | |
| 		case RXRPC_PACKET_TYPE_DATA:
 | |
| 			/* ACK the packet if necessary */
 | |
| 			switch (rxrpc_call_generate_ACK(call, &msg->hdr,
 | |
| 							NULL)) {
 | |
| 			case 0: /* useful packet */
 | |
| 				rxrpc_call_receive_data_packet(call, msg);
 | |
| 				break;
 | |
| 			case 1: /* duplicate or out-of-window packet */
 | |
| 				break;
 | |
| 			default:
 | |
| 				rxrpc_put_message(msg);
 | |
| 				goto out;
 | |
| 			}
 | |
| 			break;
 | |
| 
 | |
| 			/* deal with ACK packets */
 | |
| 		case RXRPC_PACKET_TYPE_ACK:
 | |
| 			rxrpc_call_receive_ack_packet(call, msg);
 | |
| 			break;
 | |
| 
 | |
| 			/* deal with abort packets */
 | |
| 		case RXRPC_PACKET_TYPE_ABORT: {
 | |
| 			__be32 _dbuf, *dp;
 | |
| 
 | |
| 			dp = skb_header_pointer(msg->pkt, msg->offset,
 | |
| 						sizeof(_dbuf), &_dbuf);
 | |
| 			if (dp == NULL)
 | |
| 				printk("Rx Received short ABORT packet\n");
 | |
| 
 | |
| 			_proto("Rx Received Call ABORT { data=%d }",
 | |
| 			       (dp ? ntohl(*dp) : 0));
 | |
| 
 | |
| 			spin_lock(&call->lock);
 | |
| 			call->app_call_state	= RXRPC_CSTATE_ERROR;
 | |
| 			call->app_err_state	= RXRPC_ESTATE_PEER_ABORT;
 | |
| 			call->app_abort_code	= (dp ? ntohl(*dp) : 0);
 | |
| 			call->app_errno		= -ECONNABORTED;
 | |
| 			call->app_mark		= RXRPC_APP_MARK_EOF;
 | |
| 			call->app_read_buf	= NULL;
 | |
| 			call->app_async_read	= 0;
 | |
| 
 | |
| 			/* ask the app to translate the error code */
 | |
| 			call->app_aemap_func(call);
 | |
| 			_state(call);
 | |
| 			spin_unlock(&call->lock);
 | |
| 			call->app_error_func(call);
 | |
| 			break;
 | |
| 		}
 | |
| 		default:
 | |
| 			/* deal with other packet types */
 | |
| 			_proto("Rx Unsupported packet type %u (#%u)",
 | |
| 			       msg->hdr.type, msg->seq);
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
|  out:
 | |
| 	rxrpc_put_call(call);
 | |
| 	_leave("");
 | |
| } /* end rxrpc_call_receive_packet() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * process next data packet
 | |
|  * - as the next data packet arrives:
 | |
|  *   - it is queued on app_readyq _if_ it is the next one expected
 | |
|  *     (app_ready_seq+1)
 | |
|  *   - it is queued on app_unreadyq _if_ it is not the next one expected
 | |
|  *   - if a packet placed on app_readyq completely fills a hole leading up to
 | |
|  *     the first packet on app_unreadyq, then packets now in sequence are
 | |
|  *     tranferred to app_readyq
 | |
|  * - the application layer can only see packets on app_readyq
 | |
|  *   (app_ready_qty bytes)
 | |
|  * - the application layer is prodded every time a new packet arrives
 | |
|  */
 | |
| static void rxrpc_call_receive_data_packet(struct rxrpc_call *call,
 | |
| 					   struct rxrpc_message *msg)
 | |
| {
 | |
| 	const struct rxrpc_operation *optbl, *op;
 | |
| 	struct rxrpc_message *pmsg;
 | |
| 	struct list_head *_p;
 | |
| 	int ret, lo, hi, rmtimo;
 | |
| 	__be32 opid;
 | |
| 
 | |
| 	_enter("%p{%u},%p{%u}", call, ntohl(call->call_id), msg, msg->seq);
 | |
| 
 | |
| 	rxrpc_get_message(msg);
 | |
| 
 | |
| 	/* add to the unready queue if we'd have to create a hole in the ready
 | |
| 	 * queue otherwise */
 | |
| 	if (msg->seq != call->app_ready_seq + 1) {
 | |
| 		_debug("Call add packet %d to unreadyq", msg->seq);
 | |
| 
 | |
| 		/* insert in seq order */
 | |
| 		list_for_each(_p, &call->app_unreadyq) {
 | |
| 			pmsg = list_entry(_p, struct rxrpc_message, link);
 | |
| 			if (pmsg->seq > msg->seq)
 | |
| 				break;
 | |
| 		}
 | |
| 
 | |
| 		list_add_tail(&msg->link, _p);
 | |
| 
 | |
| 		_leave(" [unreadyq]");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* next in sequence - simply append into the call's ready queue */
 | |
| 	_debug("Call add packet %d to readyq (+%Zd => %Zd bytes)",
 | |
| 	       msg->seq, msg->dsize, call->app_ready_qty);
 | |
| 
 | |
| 	spin_lock(&call->lock);
 | |
| 	call->app_ready_seq = msg->seq;
 | |
| 	call->app_ready_qty += msg->dsize;
 | |
| 	list_add_tail(&msg->link, &call->app_readyq);
 | |
| 
 | |
| 	/* move unready packets to the readyq if we got rid of a hole */
 | |
| 	while (!list_empty(&call->app_unreadyq)) {
 | |
| 		pmsg = list_entry(call->app_unreadyq.next,
 | |
| 				  struct rxrpc_message, link);
 | |
| 
 | |
| 		if (pmsg->seq != call->app_ready_seq + 1)
 | |
| 			break;
 | |
| 
 | |
| 		/* next in sequence - just move list-to-list */
 | |
| 		_debug("Call transfer packet %d to readyq (+%Zd => %Zd bytes)",
 | |
| 		       pmsg->seq, pmsg->dsize, call->app_ready_qty);
 | |
| 
 | |
| 		call->app_ready_seq = pmsg->seq;
 | |
| 		call->app_ready_qty += pmsg->dsize;
 | |
| 		list_move_tail(&pmsg->link, &call->app_readyq);
 | |
| 	}
 | |
| 
 | |
| 	/* see if we've got the last packet yet */
 | |
| 	if (!list_empty(&call->app_readyq)) {
 | |
| 		pmsg = list_entry(call->app_readyq.prev,
 | |
| 				  struct rxrpc_message, link);
 | |
| 		if (pmsg->hdr.flags & RXRPC_LAST_PACKET) {
 | |
| 			call->app_last_rcv = 1;
 | |
| 			_debug("Last packet on readyq");
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	switch (call->app_call_state) {
 | |
| 		/* do nothing if call already aborted */
 | |
| 	case RXRPC_CSTATE_ERROR:
 | |
| 		spin_unlock(&call->lock);
 | |
| 		_leave(" [error]");
 | |
| 		return;
 | |
| 
 | |
| 		/* extract the operation ID from an incoming call if that's not
 | |
| 		 * yet been done */
 | |
| 	case RXRPC_CSTATE_SRVR_RCV_OPID:
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		/* handle as yet insufficient data for the operation ID */
 | |
| 		if (call->app_ready_qty < 4) {
 | |
| 			if (call->app_last_rcv)
 | |
| 				/* trouble - last packet seen */
 | |
| 				rxrpc_call_abort(call, -EINVAL);
 | |
| 
 | |
| 			_leave("");
 | |
| 			return;
 | |
| 		}
 | |
| 
 | |
| 		/* pull the operation ID out of the buffer */
 | |
| 		ret = rxrpc_call_read_data(call, &opid, sizeof(opid), 0);
 | |
| 		if (ret < 0) {
 | |
| 			printk("Unexpected error from read-data: %d\n", ret);
 | |
| 			if (call->app_call_state != RXRPC_CSTATE_ERROR)
 | |
| 				rxrpc_call_abort(call, ret);
 | |
| 			_leave("");
 | |
| 			return;
 | |
| 		}
 | |
| 		call->app_opcode = ntohl(opid);
 | |
| 
 | |
| 		/* locate the operation in the available ops table */
 | |
| 		optbl = call->conn->service->ops_begin;
 | |
| 		lo = 0;
 | |
| 		hi = call->conn->service->ops_end - optbl;
 | |
| 
 | |
| 		while (lo < hi) {
 | |
| 			int mid = (hi + lo) / 2;
 | |
| 			op = &optbl[mid];
 | |
| 			if (call->app_opcode == op->id)
 | |
| 				goto found_op;
 | |
| 			if (call->app_opcode > op->id)
 | |
| 				lo = mid + 1;
 | |
| 			else
 | |
| 				hi = mid;
 | |
| 		}
 | |
| 
 | |
| 		/* search failed */
 | |
| 		kproto("Rx Client requested operation %d from %s service",
 | |
| 		       call->app_opcode, call->conn->service->name);
 | |
| 		rxrpc_call_abort(call, -EINVAL);
 | |
| 		_leave(" [inval]");
 | |
| 		return;
 | |
| 
 | |
| 	found_op:
 | |
| 		_proto("Rx Client requested operation %s from %s service",
 | |
| 		       op->name, call->conn->service->name);
 | |
| 
 | |
| 		/* we're now waiting for the argument block (unless the call
 | |
| 		 * was aborted) */
 | |
| 		spin_lock(&call->lock);
 | |
| 		if (call->app_call_state == RXRPC_CSTATE_SRVR_RCV_OPID ||
 | |
| 		    call->app_call_state == RXRPC_CSTATE_SRVR_SND_REPLY) {
 | |
| 			if (!call->app_last_rcv)
 | |
| 				call->app_call_state =
 | |
| 					RXRPC_CSTATE_SRVR_RCV_ARGS;
 | |
| 			else if (call->app_ready_qty > 0)
 | |
| 				call->app_call_state =
 | |
| 					RXRPC_CSTATE_SRVR_GOT_ARGS;
 | |
| 			else
 | |
| 				call->app_call_state =
 | |
| 					RXRPC_CSTATE_SRVR_SND_REPLY;
 | |
| 			call->app_mark = op->asize;
 | |
| 			call->app_user = op->user;
 | |
| 		}
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		_state(call);
 | |
| 		break;
 | |
| 
 | |
| 	case RXRPC_CSTATE_SRVR_RCV_ARGS:
 | |
| 		/* change state if just received last packet of arg block */
 | |
| 		if (call->app_last_rcv)
 | |
| 			call->app_call_state = RXRPC_CSTATE_SRVR_GOT_ARGS;
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		_state(call);
 | |
| 		break;
 | |
| 
 | |
| 	case RXRPC_CSTATE_CLNT_RCV_REPLY:
 | |
| 		/* change state if just received last packet of reply block */
 | |
| 		rmtimo = 0;
 | |
| 		if (call->app_last_rcv) {
 | |
| 			call->app_call_state = RXRPC_CSTATE_CLNT_GOT_REPLY;
 | |
| 			rmtimo = 1;
 | |
| 		}
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		if (rmtimo) {
 | |
| 			del_timer_sync(&call->acks_timeout);
 | |
| 			del_timer_sync(&call->rcv_timeout);
 | |
| 			del_timer_sync(&call->ackr_dfr_timo);
 | |
| 		}
 | |
| 
 | |
| 		_state(call);
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		/* deal with data reception in an unexpected state */
 | |
| 		printk("Unexpected state [[[ %u ]]]\n", call->app_call_state);
 | |
| 		__rxrpc_call_abort(call, -EBADMSG);
 | |
| 		_leave("");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if (call->app_call_state == RXRPC_CSTATE_CLNT_RCV_REPLY &&
 | |
| 	    call->app_last_rcv)
 | |
| 		BUG();
 | |
| 
 | |
| 	/* otherwise just invoke the data function whenever we can satisfy its desire for more
 | |
| 	 * data
 | |
| 	 */
 | |
| 	_proto("Rx Received Op Data: st=%u qty=%Zu mk=%Zu%s",
 | |
| 	       call->app_call_state, call->app_ready_qty, call->app_mark,
 | |
| 	       call->app_last_rcv ? " last-rcvd" : "");
 | |
| 
 | |
| 	spin_lock(&call->lock);
 | |
| 
 | |
| 	ret = __rxrpc_call_read_data(call);
 | |
| 	switch (ret) {
 | |
| 	case 0:
 | |
| 		spin_unlock(&call->lock);
 | |
| 		call->app_attn_func(call);
 | |
| 		break;
 | |
| 	case -EAGAIN:
 | |
| 		spin_unlock(&call->lock);
 | |
| 		break;
 | |
| 	case -ECONNABORTED:
 | |
| 		spin_unlock(&call->lock);
 | |
| 		break;
 | |
| 	default:
 | |
| 		__rxrpc_call_abort(call, ret);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	_state(call);
 | |
| 
 | |
| 	_leave("");
 | |
| 
 | |
| } /* end rxrpc_call_receive_data_packet() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * received an ACK packet
 | |
|  */
 | |
| static void rxrpc_call_receive_ack_packet(struct rxrpc_call *call,
 | |
| 					  struct rxrpc_message *msg)
 | |
| {
 | |
| 	struct rxrpc_ackpacket _ack, *ap;
 | |
| 	rxrpc_serial_net_t serial;
 | |
| 	rxrpc_seq_t seq;
 | |
| 	int ret;
 | |
| 
 | |
| 	_enter("%p{%u},%p{%u}", call, ntohl(call->call_id), msg, msg->seq);
 | |
| 
 | |
| 	/* extract the basic ACK record */
 | |
| 	ap = skb_header_pointer(msg->pkt, msg->offset, sizeof(_ack), &_ack);
 | |
| 	if (ap == NULL) {
 | |
| 		printk("Rx Received short ACK packet\n");
 | |
| 		return;
 | |
| 	}
 | |
| 	msg->offset += sizeof(_ack);
 | |
| 
 | |
| 	serial = ap->serial;
 | |
| 	seq = ntohl(ap->firstPacket);
 | |
| 
 | |
| 	_proto("Rx Received ACK %%%d { b=%hu m=%hu f=%u p=%u s=%u r=%s n=%u }",
 | |
| 	       ntohl(msg->hdr.serial),
 | |
| 	       ntohs(ap->bufferSpace),
 | |
| 	       ntohs(ap->maxSkew),
 | |
| 	       seq,
 | |
| 	       ntohl(ap->previousPacket),
 | |
| 	       ntohl(serial),
 | |
| 	       rxrpc_acks[ap->reason],
 | |
| 	       call->ackr.nAcks
 | |
| 	       );
 | |
| 
 | |
| 	/* check the other side isn't ACK'ing a sequence number I haven't sent
 | |
| 	 * yet */
 | |
| 	if (ap->nAcks > 0 &&
 | |
| 	    (seq > call->snd_seq_count ||
 | |
| 	     seq + ap->nAcks - 1 > call->snd_seq_count)) {
 | |
| 		printk("Received ACK (#%u-#%u) for unsent packet\n",
 | |
| 		       seq, seq + ap->nAcks - 1);
 | |
| 		rxrpc_call_abort(call, -EINVAL);
 | |
| 		_leave("");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* deal with RTT calculation */
 | |
| 	if (serial) {
 | |
| 		struct rxrpc_message *rttmsg;
 | |
| 
 | |
| 		/* find the prompting packet */
 | |
| 		spin_lock(&call->lock);
 | |
| 		if (call->snd_ping && call->snd_ping->hdr.serial == serial) {
 | |
| 			/* it was a ping packet */
 | |
| 			rttmsg = call->snd_ping;
 | |
| 			call->snd_ping = NULL;
 | |
| 			spin_unlock(&call->lock);
 | |
| 
 | |
| 			if (rttmsg) {
 | |
| 				rttmsg->rttdone = 1;
 | |
| 				rxrpc_peer_calculate_rtt(call->conn->peer,
 | |
| 							 rttmsg, msg);
 | |
| 				rxrpc_put_message(rttmsg);
 | |
| 			}
 | |
| 		}
 | |
| 		else {
 | |
| 			struct list_head *_p;
 | |
| 
 | |
| 			/* it ought to be a data packet - look in the pending
 | |
| 			 * ACK list */
 | |
| 			list_for_each(_p, &call->acks_pendq) {
 | |
| 				rttmsg = list_entry(_p, struct rxrpc_message,
 | |
| 						    link);
 | |
| 				if (rttmsg->hdr.serial == serial) {
 | |
| 					if (rttmsg->rttdone)
 | |
| 						/* never do RTT twice without
 | |
| 						 * resending */
 | |
| 						break;
 | |
| 
 | |
| 					rttmsg->rttdone = 1;
 | |
| 					rxrpc_peer_calculate_rtt(
 | |
| 						call->conn->peer, rttmsg, msg);
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 			spin_unlock(&call->lock);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	switch (ap->reason) {
 | |
| 		/* deal with negative/positive acknowledgement of data
 | |
| 		 * packets */
 | |
| 	case RXRPC_ACK_REQUESTED:
 | |
| 	case RXRPC_ACK_DELAY:
 | |
| 	case RXRPC_ACK_IDLE:
 | |
| 		rxrpc_call_definitively_ACK(call, seq - 1);
 | |
| 
 | |
| 	case RXRPC_ACK_DUPLICATE:
 | |
| 	case RXRPC_ACK_OUT_OF_SEQUENCE:
 | |
| 	case RXRPC_ACK_EXCEEDS_WINDOW:
 | |
| 		call->snd_resend_cnt = 0;
 | |
| 		ret = rxrpc_call_record_ACK(call, msg, seq, ap->nAcks);
 | |
| 		if (ret < 0)
 | |
| 			rxrpc_call_abort(call, ret);
 | |
| 		break;
 | |
| 
 | |
| 		/* respond to ping packets immediately */
 | |
| 	case RXRPC_ACK_PING:
 | |
| 		rxrpc_call_generate_ACK(call, &msg->hdr, ap);
 | |
| 		break;
 | |
| 
 | |
| 		/* only record RTT on ping response packets */
 | |
| 	case RXRPC_ACK_PING_RESPONSE:
 | |
| 		if (call->snd_ping) {
 | |
| 			struct rxrpc_message *rttmsg;
 | |
| 
 | |
| 			/* only do RTT stuff if the response matches the
 | |
| 			 * retained ping */
 | |
| 			rttmsg = NULL;
 | |
| 			spin_lock(&call->lock);
 | |
| 			if (call->snd_ping &&
 | |
| 			    call->snd_ping->hdr.serial == ap->serial) {
 | |
| 				rttmsg = call->snd_ping;
 | |
| 				call->snd_ping = NULL;
 | |
| 			}
 | |
| 			spin_unlock(&call->lock);
 | |
| 
 | |
| 			if (rttmsg) {
 | |
| 				rttmsg->rttdone = 1;
 | |
| 				rxrpc_peer_calculate_rtt(call->conn->peer,
 | |
| 							 rttmsg, msg);
 | |
| 				rxrpc_put_message(rttmsg);
 | |
| 			}
 | |
| 		}
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		printk("Unsupported ACK reason %u\n", ap->reason);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	_leave("");
 | |
| } /* end rxrpc_call_receive_ack_packet() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * record definitive ACKs for all messages up to and including the one with the
 | |
|  * 'highest' seq
 | |
|  */
 | |
| static void rxrpc_call_definitively_ACK(struct rxrpc_call *call,
 | |
| 					rxrpc_seq_t highest)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	int now_complete;
 | |
| 
 | |
| 	_enter("%p{ads=%u},%u", call, call->acks_dftv_seq, highest);
 | |
| 
 | |
| 	while (call->acks_dftv_seq < highest) {
 | |
| 		call->acks_dftv_seq++;
 | |
| 
 | |
| 		_proto("Definitive ACK on packet #%u", call->acks_dftv_seq);
 | |
| 
 | |
| 		/* discard those at front of queue until message with highest
 | |
| 		 * ACK is found */
 | |
| 		spin_lock(&call->lock);
 | |
| 		msg = NULL;
 | |
| 		if (!list_empty(&call->acks_pendq)) {
 | |
| 			msg = list_entry(call->acks_pendq.next,
 | |
| 					 struct rxrpc_message, link);
 | |
| 			list_del_init(&msg->link); /* dequeue */
 | |
| 			if (msg->state == RXRPC_MSG_SENT)
 | |
| 				call->acks_pend_cnt--;
 | |
| 		}
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		/* insanity check */
 | |
| 		if (!msg)
 | |
| 			panic("%s(): acks_pendq unexpectedly empty\n",
 | |
| 			      __FUNCTION__);
 | |
| 
 | |
| 		if (msg->seq != call->acks_dftv_seq)
 | |
| 			panic("%s(): Packet #%u expected at front of acks_pendq"
 | |
| 			      " (#%u found)\n",
 | |
| 			      __FUNCTION__, call->acks_dftv_seq, msg->seq);
 | |
| 
 | |
| 		/* discard the message */
 | |
| 		msg->state = RXRPC_MSG_DONE;
 | |
| 		rxrpc_put_message(msg);
 | |
| 	}
 | |
| 
 | |
| 	/* if all sent packets are definitively ACK'd then prod any sleepers just in case */
 | |
| 	now_complete = 0;
 | |
| 	spin_lock(&call->lock);
 | |
| 	if (call->acks_dftv_seq == call->snd_seq_count) {
 | |
| 		if (call->app_call_state != RXRPC_CSTATE_COMPLETE) {
 | |
| 			call->app_call_state = RXRPC_CSTATE_COMPLETE;
 | |
| 			_state(call);
 | |
| 			now_complete = 1;
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&call->lock);
 | |
| 
 | |
| 	if (now_complete) {
 | |
| 		del_timer_sync(&call->acks_timeout);
 | |
| 		del_timer_sync(&call->rcv_timeout);
 | |
| 		del_timer_sync(&call->ackr_dfr_timo);
 | |
| 		call->app_attn_func(call);
 | |
| 	}
 | |
| 
 | |
| 	_leave("");
 | |
| } /* end rxrpc_call_definitively_ACK() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * record the specified amount of ACKs/NAKs
 | |
|  */
 | |
| static int rxrpc_call_record_ACK(struct rxrpc_call *call,
 | |
| 				 struct rxrpc_message *msg,
 | |
| 				 rxrpc_seq_t seq,
 | |
| 				 size_t count)
 | |
| {
 | |
| 	struct rxrpc_message *dmsg;
 | |
| 	struct list_head *_p;
 | |
| 	rxrpc_seq_t highest;
 | |
| 	unsigned ix;
 | |
| 	size_t chunk;
 | |
| 	char resend, now_complete;
 | |
| 	u8 acks[16];
 | |
| 
 | |
| 	_enter("%p{apc=%u ads=%u},%p,%u,%Zu",
 | |
| 	       call, call->acks_pend_cnt, call->acks_dftv_seq,
 | |
| 	       msg, seq, count);
 | |
| 
 | |
| 	/* handle re-ACK'ing of definitively ACK'd packets (may be out-of-order
 | |
| 	 * ACKs) */
 | |
| 	if (seq <= call->acks_dftv_seq) {
 | |
| 		unsigned delta = call->acks_dftv_seq - seq;
 | |
| 
 | |
| 		if (count <= delta) {
 | |
| 			_leave(" = 0 [all definitively ACK'd]");
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		seq += delta;
 | |
| 		count -= delta;
 | |
| 		msg->offset += delta;
 | |
| 	}
 | |
| 
 | |
| 	highest = seq + count - 1;
 | |
| 	resend = 0;
 | |
| 	while (count > 0) {
 | |
| 		/* extract up to 16 ACK slots at a time */
 | |
| 		chunk = min(count, sizeof(acks));
 | |
| 		count -= chunk;
 | |
| 
 | |
| 		memset(acks, 2, sizeof(acks));
 | |
| 
 | |
| 		if (skb_copy_bits(msg->pkt, msg->offset, &acks, chunk) < 0) {
 | |
| 			printk("Rx Received short ACK packet\n");
 | |
| 			_leave(" = -EINVAL");
 | |
| 			return -EINVAL;
 | |
| 		}
 | |
| 		msg->offset += chunk;
 | |
| 
 | |
| 		/* check that the ACK set is valid */
 | |
| 		for (ix = 0; ix < chunk; ix++) {
 | |
| 			switch (acks[ix]) {
 | |
| 			case RXRPC_ACK_TYPE_ACK:
 | |
| 				break;
 | |
| 			case RXRPC_ACK_TYPE_NACK:
 | |
| 				resend = 1;
 | |
| 				break;
 | |
| 			default:
 | |
| 				printk("Rx Received unsupported ACK state"
 | |
| 				       " %u\n", acks[ix]);
 | |
| 				_leave(" = -EINVAL");
 | |
| 				return -EINVAL;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		_proto("Rx ACK of packets #%u-#%u "
 | |
| 		       "[%c%c%c%c%c%c%c%c%c%c%c%c%c%c%c%c] (pend=%u)",
 | |
| 		       seq, (unsigned) (seq + chunk - 1),
 | |
| 		       _acktype[acks[0x0]],
 | |
| 		       _acktype[acks[0x1]],
 | |
| 		       _acktype[acks[0x2]],
 | |
| 		       _acktype[acks[0x3]],
 | |
| 		       _acktype[acks[0x4]],
 | |
| 		       _acktype[acks[0x5]],
 | |
| 		       _acktype[acks[0x6]],
 | |
| 		       _acktype[acks[0x7]],
 | |
| 		       _acktype[acks[0x8]],
 | |
| 		       _acktype[acks[0x9]],
 | |
| 		       _acktype[acks[0xA]],
 | |
| 		       _acktype[acks[0xB]],
 | |
| 		       _acktype[acks[0xC]],
 | |
| 		       _acktype[acks[0xD]],
 | |
| 		       _acktype[acks[0xE]],
 | |
| 		       _acktype[acks[0xF]],
 | |
| 		       call->acks_pend_cnt
 | |
| 		       );
 | |
| 
 | |
| 		/* mark the packets in the ACK queue as being provisionally
 | |
| 		 * ACK'd */
 | |
| 		ix = 0;
 | |
| 		spin_lock(&call->lock);
 | |
| 
 | |
| 		/* find the first packet ACK'd/NAK'd here */
 | |
| 		list_for_each(_p, &call->acks_pendq) {
 | |
| 			dmsg = list_entry(_p, struct rxrpc_message, link);
 | |
| 			if (dmsg->seq == seq)
 | |
| 				goto found_first;
 | |
| 			_debug("- %u: skipping #%u", ix, dmsg->seq);
 | |
| 		}
 | |
| 		goto bad_queue;
 | |
| 
 | |
| 	found_first:
 | |
| 		do {
 | |
| 			_debug("- %u: processing #%u (%c) apc=%u",
 | |
| 			       ix, dmsg->seq, _acktype[acks[ix]],
 | |
| 			       call->acks_pend_cnt);
 | |
| 
 | |
| 			if (acks[ix] == RXRPC_ACK_TYPE_ACK) {
 | |
| 				if (dmsg->state == RXRPC_MSG_SENT)
 | |
| 					call->acks_pend_cnt--;
 | |
| 				dmsg->state = RXRPC_MSG_ACKED;
 | |
| 			}
 | |
| 			else {
 | |
| 				if (dmsg->state == RXRPC_MSG_ACKED)
 | |
| 					call->acks_pend_cnt++;
 | |
| 				dmsg->state = RXRPC_MSG_SENT;
 | |
| 			}
 | |
| 			ix++;
 | |
| 			seq++;
 | |
| 
 | |
| 			_p = dmsg->link.next;
 | |
| 			dmsg = list_entry(_p, struct rxrpc_message, link);
 | |
| 		} while(ix < chunk &&
 | |
| 			_p != &call->acks_pendq &&
 | |
| 			dmsg->seq == seq);
 | |
| 
 | |
| 		if (ix < chunk)
 | |
| 			goto bad_queue;
 | |
| 
 | |
| 		spin_unlock(&call->lock);
 | |
| 	}
 | |
| 
 | |
| 	if (resend)
 | |
| 		rxrpc_call_resend(call, highest);
 | |
| 
 | |
| 	/* if all packets are provisionally ACK'd, then wake up anyone who's
 | |
| 	 * waiting for that */
 | |
| 	now_complete = 0;
 | |
| 	spin_lock(&call->lock);
 | |
| 	if (call->acks_pend_cnt == 0) {
 | |
| 		if (call->app_call_state == RXRPC_CSTATE_SRVR_RCV_FINAL_ACK) {
 | |
| 			call->app_call_state = RXRPC_CSTATE_COMPLETE;
 | |
| 			_state(call);
 | |
| 		}
 | |
| 		now_complete = 1;
 | |
| 	}
 | |
| 	spin_unlock(&call->lock);
 | |
| 
 | |
| 	if (now_complete) {
 | |
| 		_debug("- wake up waiters");
 | |
| 		del_timer_sync(&call->acks_timeout);
 | |
| 		del_timer_sync(&call->rcv_timeout);
 | |
| 		del_timer_sync(&call->ackr_dfr_timo);
 | |
| 		call->app_attn_func(call);
 | |
| 	}
 | |
| 
 | |
| 	_leave(" = 0 (apc=%u)", call->acks_pend_cnt);
 | |
| 	return 0;
 | |
| 
 | |
|  bad_queue:
 | |
| 	panic("%s(): acks_pendq in bad state (packet #%u absent)\n",
 | |
| 	      __FUNCTION__, seq);
 | |
| 
 | |
| } /* end rxrpc_call_record_ACK() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * transfer data from the ready packet queue to the asynchronous read buffer
 | |
|  * - since this func is the only one going to look at packets queued on
 | |
|  *   app_readyq, we don't need a lock to modify or access them, only to modify
 | |
|  *   the queue pointers
 | |
|  * - called with call->lock held
 | |
|  * - the buffer must be in kernel space
 | |
|  * - returns:
 | |
|  *	0 if buffer filled
 | |
|  *	-EAGAIN if buffer not filled and more data to come
 | |
|  *	-EBADMSG if last packet received and insufficient data left
 | |
|  *	-ECONNABORTED if the call has in an error state
 | |
|  */
 | |
| static int __rxrpc_call_read_data(struct rxrpc_call *call)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	size_t qty;
 | |
| 	int ret;
 | |
| 
 | |
| 	_enter("%p{as=%d buf=%p qty=%Zu/%Zu}",
 | |
| 	       call,
 | |
| 	       call->app_async_read, call->app_read_buf,
 | |
| 	       call->app_ready_qty, call->app_mark);
 | |
| 
 | |
| 	/* check the state */
 | |
| 	switch (call->app_call_state) {
 | |
| 	case RXRPC_CSTATE_SRVR_RCV_ARGS:
 | |
| 	case RXRPC_CSTATE_CLNT_RCV_REPLY:
 | |
| 		if (call->app_last_rcv) {
 | |
| 			printk("%s(%p,%p,%Zd):"
 | |
| 			       " Inconsistent call state (%s, last pkt)",
 | |
| 			       __FUNCTION__,
 | |
| 			       call, call->app_read_buf, call->app_mark,
 | |
| 			       rxrpc_call_states[call->app_call_state]);
 | |
| 			BUG();
 | |
| 		}
 | |
| 		break;
 | |
| 
 | |
| 	case RXRPC_CSTATE_SRVR_RCV_OPID:
 | |
| 	case RXRPC_CSTATE_SRVR_GOT_ARGS:
 | |
| 	case RXRPC_CSTATE_CLNT_GOT_REPLY:
 | |
| 		break;
 | |
| 
 | |
| 	case RXRPC_CSTATE_SRVR_SND_REPLY:
 | |
| 		if (!call->app_last_rcv) {
 | |
| 			printk("%s(%p,%p,%Zd):"
 | |
| 			       " Inconsistent call state (%s, not last pkt)",
 | |
| 			       __FUNCTION__,
 | |
| 			       call, call->app_read_buf, call->app_mark,
 | |
| 			       rxrpc_call_states[call->app_call_state]);
 | |
| 			BUG();
 | |
| 		}
 | |
| 		_debug("Trying to read data from call in SND_REPLY state");
 | |
| 		break;
 | |
| 
 | |
| 	case RXRPC_CSTATE_ERROR:
 | |
| 		_leave(" = -ECONNABORTED");
 | |
| 		return -ECONNABORTED;
 | |
| 
 | |
| 	default:
 | |
| 		printk("reading in unexpected state [[[ %u ]]]\n",
 | |
| 		       call->app_call_state);
 | |
| 		BUG();
 | |
| 	}
 | |
| 
 | |
| 	/* handle the case of not having an async buffer */
 | |
| 	if (!call->app_async_read) {
 | |
| 		if (call->app_mark == RXRPC_APP_MARK_EOF) {
 | |
| 			ret = call->app_last_rcv ? 0 : -EAGAIN;
 | |
| 		}
 | |
| 		else {
 | |
| 			if (call->app_mark >= call->app_ready_qty) {
 | |
| 				call->app_mark = RXRPC_APP_MARK_EOF;
 | |
| 				ret = 0;
 | |
| 			}
 | |
| 			else {
 | |
| 				ret = call->app_last_rcv ? -EBADMSG : -EAGAIN;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		_leave(" = %d [no buf]", ret);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	while (!list_empty(&call->app_readyq) && call->app_mark > 0) {
 | |
| 		msg = list_entry(call->app_readyq.next,
 | |
| 				 struct rxrpc_message, link);
 | |
| 
 | |
| 		/* drag as much data as we need out of this packet */
 | |
| 		qty = min(call->app_mark, msg->dsize);
 | |
| 
 | |
| 		_debug("reading %Zu from skb=%p off=%lu",
 | |
| 		       qty, msg->pkt, msg->offset);
 | |
| 
 | |
| 		if (call->app_read_buf)
 | |
| 			if (skb_copy_bits(msg->pkt, msg->offset,
 | |
| 					  call->app_read_buf, qty) < 0)
 | |
| 				panic("%s: Failed to copy data from packet:"
 | |
| 				      " (%p,%p,%Zd)",
 | |
| 				      __FUNCTION__,
 | |
| 				      call, call->app_read_buf, qty);
 | |
| 
 | |
| 		/* if that packet is now empty, discard it */
 | |
| 		call->app_ready_qty -= qty;
 | |
| 		msg->dsize -= qty;
 | |
| 
 | |
| 		if (msg->dsize == 0) {
 | |
| 			list_del_init(&msg->link);
 | |
| 			rxrpc_put_message(msg);
 | |
| 		}
 | |
| 		else {
 | |
| 			msg->offset += qty;
 | |
| 		}
 | |
| 
 | |
| 		call->app_mark -= qty;
 | |
| 		if (call->app_read_buf)
 | |
| 			call->app_read_buf += qty;
 | |
| 	}
 | |
| 
 | |
| 	if (call->app_mark == 0) {
 | |
| 		call->app_async_read = 0;
 | |
| 		call->app_mark = RXRPC_APP_MARK_EOF;
 | |
| 		call->app_read_buf = NULL;
 | |
| 
 | |
| 		/* adjust the state if used up all packets */
 | |
| 		if (list_empty(&call->app_readyq) && call->app_last_rcv) {
 | |
| 			switch (call->app_call_state) {
 | |
| 			case RXRPC_CSTATE_SRVR_RCV_OPID:
 | |
| 				call->app_call_state = RXRPC_CSTATE_SRVR_SND_REPLY;
 | |
| 				call->app_mark = RXRPC_APP_MARK_EOF;
 | |
| 				_state(call);
 | |
| 				del_timer_sync(&call->rcv_timeout);
 | |
| 				break;
 | |
| 			case RXRPC_CSTATE_SRVR_GOT_ARGS:
 | |
| 				call->app_call_state = RXRPC_CSTATE_SRVR_SND_REPLY;
 | |
| 				_state(call);
 | |
| 				del_timer_sync(&call->rcv_timeout);
 | |
| 				break;
 | |
| 			default:
 | |
| 				call->app_call_state = RXRPC_CSTATE_COMPLETE;
 | |
| 				_state(call);
 | |
| 				del_timer_sync(&call->acks_timeout);
 | |
| 				del_timer_sync(&call->ackr_dfr_timo);
 | |
| 				del_timer_sync(&call->rcv_timeout);
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		_leave(" = 0");
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	if (call->app_last_rcv) {
 | |
| 		_debug("Insufficient data (%Zu/%Zu)",
 | |
| 		       call->app_ready_qty, call->app_mark);
 | |
| 		call->app_async_read = 0;
 | |
| 		call->app_mark = RXRPC_APP_MARK_EOF;
 | |
| 		call->app_read_buf = NULL;
 | |
| 
 | |
| 		_leave(" = -EBADMSG");
 | |
| 		return -EBADMSG;
 | |
| 	}
 | |
| 
 | |
| 	_leave(" = -EAGAIN");
 | |
| 	return -EAGAIN;
 | |
| } /* end __rxrpc_call_read_data() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * attempt to read the specified amount of data from the call's ready queue
 | |
|  * into the buffer provided
 | |
|  * - since this func is the only one going to look at packets queued on
 | |
|  *   app_readyq, we don't need a lock to modify or access them, only to modify
 | |
|  *   the queue pointers
 | |
|  * - if the buffer pointer is NULL, then data is merely drained, not copied
 | |
|  * - if flags&RXRPC_CALL_READ_BLOCK, then the function will wait until there is
 | |
|  *   enough data or an error will be generated
 | |
|  *   - note that the caller must have added the calling task to the call's wait
 | |
|  *     queue beforehand
 | |
|  * - if flags&RXRPC_CALL_READ_ALL, then an error will be generated if this
 | |
|  *   function doesn't read all available data
 | |
|  */
 | |
| int rxrpc_call_read_data(struct rxrpc_call *call,
 | |
| 			 void *buffer, size_t size, int flags)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	_enter("%p{arq=%Zu},%p,%Zd,%x",
 | |
| 	       call, call->app_ready_qty, buffer, size, flags);
 | |
| 
 | |
| 	spin_lock(&call->lock);
 | |
| 
 | |
| 	if (unlikely(!!call->app_read_buf)) {
 | |
| 		spin_unlock(&call->lock);
 | |
| 		_leave(" = -EBUSY");
 | |
| 		return -EBUSY;
 | |
| 	}
 | |
| 
 | |
| 	call->app_mark = size;
 | |
| 	call->app_read_buf = buffer;
 | |
| 	call->app_async_read = 1;
 | |
| 	call->app_read_count++;
 | |
| 
 | |
| 	/* read as much data as possible */
 | |
| 	ret = __rxrpc_call_read_data(call);
 | |
| 	switch (ret) {
 | |
| 	case 0:
 | |
| 		if (flags & RXRPC_CALL_READ_ALL &&
 | |
| 		    (!call->app_last_rcv || call->app_ready_qty > 0)) {
 | |
| 			_leave(" = -EBADMSG");
 | |
| 			__rxrpc_call_abort(call, -EBADMSG);
 | |
| 			return -EBADMSG;
 | |
| 		}
 | |
| 
 | |
| 		spin_unlock(&call->lock);
 | |
| 		call->app_attn_func(call);
 | |
| 		_leave(" = 0");
 | |
| 		return ret;
 | |
| 
 | |
| 	case -ECONNABORTED:
 | |
| 		spin_unlock(&call->lock);
 | |
| 		_leave(" = %d [aborted]", ret);
 | |
| 		return ret;
 | |
| 
 | |
| 	default:
 | |
| 		__rxrpc_call_abort(call, ret);
 | |
| 		_leave(" = %d", ret);
 | |
| 		return ret;
 | |
| 
 | |
| 	case -EAGAIN:
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		if (!(flags & RXRPC_CALL_READ_BLOCK)) {
 | |
| 			_leave(" = -EAGAIN");
 | |
| 			return -EAGAIN;
 | |
| 		}
 | |
| 
 | |
| 		/* wait for the data to arrive */
 | |
| 		_debug("blocking for data arrival");
 | |
| 
 | |
| 		for (;;) {
 | |
| 			set_current_state(TASK_INTERRUPTIBLE);
 | |
| 			if (!call->app_async_read || signal_pending(current))
 | |
| 				break;
 | |
| 			schedule();
 | |
| 		}
 | |
| 		set_current_state(TASK_RUNNING);
 | |
| 
 | |
| 		if (signal_pending(current)) {
 | |
| 			_leave(" = -EINTR");
 | |
| 			return -EINTR;
 | |
| 		}
 | |
| 
 | |
| 		if (call->app_call_state == RXRPC_CSTATE_ERROR) {
 | |
| 			_leave(" = -ECONNABORTED");
 | |
| 			return -ECONNABORTED;
 | |
| 		}
 | |
| 
 | |
| 		_leave(" = 0");
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| } /* end rxrpc_call_read_data() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * write data to a call
 | |
|  * - the data may not be sent immediately if it doesn't fill a buffer
 | |
|  * - if we can't queue all the data for buffering now, siov[] will have been
 | |
|  *   adjusted to take account of what has been sent
 | |
|  */
 | |
| int rxrpc_call_write_data(struct rxrpc_call *call,
 | |
| 			  size_t sioc,
 | |
| 			  struct kvec *siov,
 | |
| 			  u8 rxhdr_flags,
 | |
| 			  gfp_t alloc_flags,
 | |
| 			  int dup_data,
 | |
| 			  size_t *size_sent)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	struct kvec *sptr;
 | |
| 	size_t space, size, chunk, tmp;
 | |
| 	char *buf;
 | |
| 	int ret;
 | |
| 
 | |
| 	_enter("%p,%Zu,%p,%02x,%x,%d,%p",
 | |
| 	       call, sioc, siov, rxhdr_flags, alloc_flags, dup_data,
 | |
| 	       size_sent);
 | |
| 
 | |
| 	*size_sent = 0;
 | |
| 	size = 0;
 | |
| 	ret = -EINVAL;
 | |
| 
 | |
| 	/* can't send more if we've sent last packet from this end */
 | |
| 	switch (call->app_call_state) {
 | |
| 	case RXRPC_CSTATE_SRVR_SND_REPLY:
 | |
| 	case RXRPC_CSTATE_CLNT_SND_ARGS:
 | |
| 		break;
 | |
| 	case RXRPC_CSTATE_ERROR:
 | |
| 		ret = call->app_errno;
 | |
| 	default:
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	/* calculate how much data we've been given */
 | |
| 	sptr = siov;
 | |
| 	for (; sioc > 0; sptr++, sioc--) {
 | |
| 		if (!sptr->iov_len)
 | |
| 			continue;
 | |
| 
 | |
| 		if (!sptr->iov_base)
 | |
| 			goto out;
 | |
| 
 | |
| 		size += sptr->iov_len;
 | |
| 	}
 | |
| 
 | |
| 	_debug("- size=%Zu mtu=%Zu", size, call->conn->mtu_size);
 | |
| 
 | |
| 	do {
 | |
| 		/* make sure there's a message under construction */
 | |
| 		if (!call->snd_nextmsg) {
 | |
| 			/* no - allocate a message with no data yet attached */
 | |
| 			ret = rxrpc_conn_newmsg(call->conn, call,
 | |
| 						RXRPC_PACKET_TYPE_DATA,
 | |
| 						0, NULL, alloc_flags,
 | |
| 						&call->snd_nextmsg);
 | |
| 			if (ret < 0)
 | |
| 				goto out;
 | |
| 			_debug("- allocated new message [ds=%Zu]",
 | |
| 			       call->snd_nextmsg->dsize);
 | |
| 		}
 | |
| 
 | |
| 		msg = call->snd_nextmsg;
 | |
| 		msg->hdr.flags |= rxhdr_flags;
 | |
| 
 | |
| 		/* deal with zero-length terminal packet */
 | |
| 		if (size == 0) {
 | |
| 			if (rxhdr_flags & RXRPC_LAST_PACKET) {
 | |
| 				ret = rxrpc_call_flush(call);
 | |
| 				if (ret < 0)
 | |
| 					goto out;
 | |
| 			}
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		/* work out how much space current packet has available */
 | |
| 		space = call->conn->mtu_size - msg->dsize;
 | |
| 		chunk = min(space, size);
 | |
| 
 | |
| 		_debug("- [before] space=%Zu chunk=%Zu", space, chunk);
 | |
| 
 | |
| 		while (!siov->iov_len)
 | |
| 			siov++;
 | |
| 
 | |
| 		/* if we are going to have to duplicate the data then coalesce
 | |
| 		 * it too */
 | |
| 		if (dup_data) {
 | |
| 			/* don't allocate more that 1 page at a time */
 | |
| 			if (chunk > PAGE_SIZE)
 | |
| 				chunk = PAGE_SIZE;
 | |
| 
 | |
| 			/* allocate a data buffer and attach to the message */
 | |
| 			buf = kmalloc(chunk, alloc_flags);
 | |
| 			if (unlikely(!buf)) {
 | |
| 				if (msg->dsize ==
 | |
| 				    sizeof(struct rxrpc_header)) {
 | |
| 					/* discard an empty msg and wind back
 | |
| 					 * the seq counter */
 | |
| 					rxrpc_put_message(msg);
 | |
| 					call->snd_nextmsg = NULL;
 | |
| 					call->snd_seq_count--;
 | |
| 				}
 | |
| 
 | |
| 				ret = -ENOMEM;
 | |
| 				goto out;
 | |
| 			}
 | |
| 
 | |
| 			tmp = msg->dcount++;
 | |
| 			set_bit(tmp, &msg->dfree);
 | |
| 			msg->data[tmp].iov_base = buf;
 | |
| 			msg->data[tmp].iov_len = chunk;
 | |
| 			msg->dsize += chunk;
 | |
| 			*size_sent += chunk;
 | |
| 			size -= chunk;
 | |
| 
 | |
| 			/* load the buffer with data */
 | |
| 			while (chunk > 0) {
 | |
| 				tmp = min(chunk, siov->iov_len);
 | |
| 				memcpy(buf, siov->iov_base, tmp);
 | |
| 				buf += tmp;
 | |
| 				siov->iov_base += tmp;
 | |
| 				siov->iov_len -= tmp;
 | |
| 				if (!siov->iov_len)
 | |
| 					siov++;
 | |
| 				chunk -= tmp;
 | |
| 			}
 | |
| 		}
 | |
| 		else {
 | |
| 			/* we want to attach the supplied buffers directly */
 | |
| 			while (chunk > 0 &&
 | |
| 			       msg->dcount < RXRPC_MSG_MAX_IOCS) {
 | |
| 				tmp = msg->dcount++;
 | |
| 				msg->data[tmp].iov_base = siov->iov_base;
 | |
| 				msg->data[tmp].iov_len = siov->iov_len;
 | |
| 				msg->dsize += siov->iov_len;
 | |
| 				*size_sent += siov->iov_len;
 | |
| 				size -= siov->iov_len;
 | |
| 				chunk -= siov->iov_len;
 | |
| 				siov++;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		_debug("- [loaded] chunk=%Zu size=%Zu", chunk, size);
 | |
| 
 | |
| 		/* dispatch the message when full, final or requesting ACK */
 | |
| 		if (msg->dsize >= call->conn->mtu_size || rxhdr_flags) {
 | |
| 			ret = rxrpc_call_flush(call);
 | |
| 			if (ret < 0)
 | |
| 				goto out;
 | |
| 		}
 | |
| 
 | |
| 	} while(size > 0);
 | |
| 
 | |
| 	ret = 0;
 | |
|  out:
 | |
| 	_leave(" = %d (%Zd queued, %Zd rem)", ret, *size_sent, size);
 | |
| 	return ret;
 | |
| 
 | |
| } /* end rxrpc_call_write_data() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * flush outstanding packets to the network
 | |
|  */
 | |
| static int rxrpc_call_flush(struct rxrpc_call *call)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	_enter("%p", call);
 | |
| 
 | |
| 	rxrpc_get_call(call);
 | |
| 
 | |
| 	/* if there's a packet under construction, then dispatch it now */
 | |
| 	if (call->snd_nextmsg) {
 | |
| 		msg = call->snd_nextmsg;
 | |
| 		call->snd_nextmsg = NULL;
 | |
| 
 | |
| 		if (msg->hdr.flags & RXRPC_LAST_PACKET) {
 | |
| 			msg->hdr.flags &= ~RXRPC_MORE_PACKETS;
 | |
| 			if (call->app_call_state != RXRPC_CSTATE_CLNT_SND_ARGS)
 | |
| 				msg->hdr.flags |= RXRPC_REQUEST_ACK;
 | |
| 		}
 | |
| 		else {
 | |
| 			msg->hdr.flags |= RXRPC_MORE_PACKETS;
 | |
| 		}
 | |
| 
 | |
| 		_proto("Sending DATA message { ds=%Zu dc=%u df=%02lu }",
 | |
| 		       msg->dsize, msg->dcount, msg->dfree);
 | |
| 
 | |
| 		/* queue and adjust call state */
 | |
| 		spin_lock(&call->lock);
 | |
| 		list_add_tail(&msg->link, &call->acks_pendq);
 | |
| 
 | |
| 		/* decide what to do depending on current state and if this is
 | |
| 		 * the last packet */
 | |
| 		ret = -EINVAL;
 | |
| 		switch (call->app_call_state) {
 | |
| 		case RXRPC_CSTATE_SRVR_SND_REPLY:
 | |
| 			if (msg->hdr.flags & RXRPC_LAST_PACKET) {
 | |
| 				call->app_call_state =
 | |
| 					RXRPC_CSTATE_SRVR_RCV_FINAL_ACK;
 | |
| 				_state(call);
 | |
| 			}
 | |
| 			break;
 | |
| 
 | |
| 		case RXRPC_CSTATE_CLNT_SND_ARGS:
 | |
| 			if (msg->hdr.flags & RXRPC_LAST_PACKET) {
 | |
| 				call->app_call_state =
 | |
| 					RXRPC_CSTATE_CLNT_RCV_REPLY;
 | |
| 				_state(call);
 | |
| 			}
 | |
| 			break;
 | |
| 
 | |
| 		case RXRPC_CSTATE_ERROR:
 | |
| 			ret = call->app_errno;
 | |
| 		default:
 | |
| 			spin_unlock(&call->lock);
 | |
| 			goto out;
 | |
| 		}
 | |
| 
 | |
| 		call->acks_pend_cnt++;
 | |
| 
 | |
| 		mod_timer(&call->acks_timeout,
 | |
| 			  __rxrpc_rtt_based_timeout(call,
 | |
| 						    rxrpc_call_acks_timeout));
 | |
| 
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		ret = rxrpc_conn_sendmsg(call->conn, msg);
 | |
| 		if (ret == 0)
 | |
| 			call->pkt_snd_count++;
 | |
| 	}
 | |
| 
 | |
|  out:
 | |
| 	rxrpc_put_call(call);
 | |
| 
 | |
| 	_leave(" = %d", ret);
 | |
| 	return ret;
 | |
| 
 | |
| } /* end rxrpc_call_flush() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * resend NAK'd or unacknowledged packets up to the highest one specified
 | |
|  */
 | |
| static void rxrpc_call_resend(struct rxrpc_call *call, rxrpc_seq_t highest)
 | |
| {
 | |
| 	struct rxrpc_message *msg;
 | |
| 	struct list_head *_p;
 | |
| 	rxrpc_seq_t seq = 0;
 | |
| 
 | |
| 	_enter("%p,%u", call, highest);
 | |
| 
 | |
| 	_proto("Rx Resend required");
 | |
| 
 | |
| 	/* handle too many resends */
 | |
| 	if (call->snd_resend_cnt >= rxrpc_call_max_resend) {
 | |
| 		_debug("Aborting due to too many resends (rcv=%d)",
 | |
| 		       call->pkt_rcv_count);
 | |
| 		rxrpc_call_abort(call,
 | |
| 				 call->pkt_rcv_count > 0 ? -EIO : -ETIMEDOUT);
 | |
| 		_leave("");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	spin_lock(&call->lock);
 | |
| 	call->snd_resend_cnt++;
 | |
| 	for (;;) {
 | |
| 		/* determine which the next packet we might need to ACK is */
 | |
| 		if (seq <= call->acks_dftv_seq)
 | |
| 			seq = call->acks_dftv_seq;
 | |
| 		seq++;
 | |
| 
 | |
| 		if (seq > highest)
 | |
| 			break;
 | |
| 
 | |
| 		/* look for the packet in the pending-ACK queue */
 | |
| 		list_for_each(_p, &call->acks_pendq) {
 | |
| 			msg = list_entry(_p, struct rxrpc_message, link);
 | |
| 			if (msg->seq == seq)
 | |
| 				goto found_msg;
 | |
| 		}
 | |
| 
 | |
| 		panic("%s(%p,%d):"
 | |
| 		      " Inconsistent pending-ACK queue (ds=%u sc=%u sq=%u)\n",
 | |
| 		      __FUNCTION__, call, highest,
 | |
| 		      call->acks_dftv_seq, call->snd_seq_count, seq);
 | |
| 
 | |
| 	found_msg:
 | |
| 		if (msg->state != RXRPC_MSG_SENT)
 | |
| 			continue; /* only un-ACK'd packets */
 | |
| 
 | |
| 		rxrpc_get_message(msg);
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		/* send each message again (and ignore any errors we might
 | |
| 		 * incur) */
 | |
| 		_proto("Resending DATA message { ds=%Zu dc=%u df=%02lu }",
 | |
| 		       msg->dsize, msg->dcount, msg->dfree);
 | |
| 
 | |
| 		if (rxrpc_conn_sendmsg(call->conn, msg) == 0)
 | |
| 			call->pkt_snd_count++;
 | |
| 
 | |
| 		rxrpc_put_message(msg);
 | |
| 
 | |
| 		spin_lock(&call->lock);
 | |
| 	}
 | |
| 
 | |
| 	/* reset the timeout */
 | |
| 	mod_timer(&call->acks_timeout,
 | |
| 		  __rxrpc_rtt_based_timeout(call, rxrpc_call_acks_timeout));
 | |
| 
 | |
| 	spin_unlock(&call->lock);
 | |
| 
 | |
| 	_leave("");
 | |
| } /* end rxrpc_call_resend() */
 | |
| 
 | |
| /*****************************************************************************/
 | |
| /*
 | |
|  * handle an ICMP error being applied to a call
 | |
|  */
 | |
| void rxrpc_call_handle_error(struct rxrpc_call *call, int local, int errno)
 | |
| {
 | |
| 	_enter("%p{%u},%d", call, ntohl(call->call_id), errno);
 | |
| 
 | |
| 	/* if this call is already aborted, then just wake up any waiters */
 | |
| 	if (call->app_call_state == RXRPC_CSTATE_ERROR) {
 | |
| 		call->app_error_func(call);
 | |
| 	}
 | |
| 	else {
 | |
| 		/* tell the app layer what happened */
 | |
| 		spin_lock(&call->lock);
 | |
| 		call->app_call_state = RXRPC_CSTATE_ERROR;
 | |
| 		_state(call);
 | |
| 		if (local)
 | |
| 			call->app_err_state = RXRPC_ESTATE_LOCAL_ERROR;
 | |
| 		else
 | |
| 			call->app_err_state = RXRPC_ESTATE_REMOTE_ERROR;
 | |
| 		call->app_errno		= errno;
 | |
| 		call->app_mark		= RXRPC_APP_MARK_EOF;
 | |
| 		call->app_read_buf	= NULL;
 | |
| 		call->app_async_read	= 0;
 | |
| 
 | |
| 		/* map the error */
 | |
| 		call->app_aemap_func(call);
 | |
| 
 | |
| 		del_timer_sync(&call->acks_timeout);
 | |
| 		del_timer_sync(&call->rcv_timeout);
 | |
| 		del_timer_sync(&call->ackr_dfr_timo);
 | |
| 
 | |
| 		spin_unlock(&call->lock);
 | |
| 
 | |
| 		call->app_error_func(call);
 | |
| 	}
 | |
| 
 | |
| 	_leave("");
 | |
| } /* end rxrpc_call_handle_error() */
 |