X-Git-Url: http://git.osdn.net/view?a=blobdiff_plain;f=ndk%2Fandroid-ndk-r4%2Fdarwin%2Fplatforms%2Fandroid-8%2Farch-arm%2Fusr%2Finclude%2Flinux%2Fsunrpc%2Fsched.h;fp=ndk%2Fandroid-ndk-r4%2Fdarwin%2Fplatforms%2Fandroid-8%2Farch-arm%2Fusr%2Finclude%2Flinux%2Fsunrpc%2Fsched.h;h=859b2d8390a2d0704d61c36f1284980f9785b645;hb=72eddb87bfc38a2f6fae40566ec6380eb838489a;hp=0000000000000000000000000000000000000000;hpb=e19e12d257815bef7bfe4f3d81056a5b2464c464;p=android-x86%2Fprebuilt.git diff --git a/ndk/android-ndk-r4/darwin/platforms/android-8/arch-arm/usr/include/linux/sunrpc/sched.h b/ndk/android-ndk-r4/darwin/platforms/android-8/arch-arm/usr/include/linux/sunrpc/sched.h new file mode 100644 index 0000000..859b2d8 --- /dev/null +++ b/ndk/android-ndk-r4/darwin/platforms/android-8/arch-arm/usr/include/linux/sunrpc/sched.h @@ -0,0 +1,182 @@ +/**************************************************************************** + **************************************************************************** + *** + *** This header was automatically generated from a Linux kernel header + *** of the same name, to make information necessary for userspace to + *** call into the kernel available to libc. It contains only constants, + *** structures, and macros generated from the original header, and thus, + *** contains no copyrightable information. + *** + **************************************************************************** + ****************************************************************************/ +#ifndef _LINUX_SUNRPC_SCHED_H_ +#define _LINUX_SUNRPC_SCHED_H_ + +#include +#include +#include +#include +#include +#include + +struct rpc_procinfo; +struct rpc_message { + struct rpc_procinfo * rpc_proc; + void * rpc_argp; + void * rpc_resp; + struct rpc_cred * rpc_cred; +}; + +struct rpc_call_ops; +struct rpc_wait_queue; +struct rpc_wait { + struct list_head list; + struct list_head links; + struct rpc_wait_queue * rpc_waitq; +}; + +struct rpc_task { +#ifdef RPC_DEBUG + unsigned long tk_magic; +#endif + atomic_t tk_count; + struct list_head tk_task; + struct rpc_clnt * tk_client; + struct rpc_rqst * tk_rqstp; + int tk_status; + + struct rpc_message tk_msg; + __u8 tk_garb_retry; + __u8 tk_cred_retry; + + unsigned long tk_cookie; + + void (*tk_timeout_fn)(struct rpc_task *); + void (*tk_callback)(struct rpc_task *); + void (*tk_action)(struct rpc_task *); + const struct rpc_call_ops *tk_ops; + void * tk_calldata; + + struct timer_list tk_timer; + unsigned long tk_timeout; + unsigned short tk_flags; + unsigned char tk_priority : 2; + unsigned long tk_runstate; + struct workqueue_struct *tk_workqueue; + union { + struct work_struct tk_work; + struct rpc_wait tk_wait; + } u; + + unsigned short tk_timeouts; + size_t tk_bytes_sent; + unsigned long tk_start; + long tk_rtt; + +#ifdef RPC_DEBUG + unsigned short tk_pid; +#endif +}; +#define tk_auth tk_client->cl_auth +#define tk_xprt tk_client->cl_xprt + +#define task_for_each(task, pos, head) list_for_each(pos, head) if ((task=list_entry(pos, struct rpc_task, u.tk_wait.list)),1) + +#define task_for_first(task, head) if (!list_empty(head) && ((task=list_entry((head)->next, struct rpc_task, u.tk_wait.list)),1)) + +#define alltask_for_each(task, pos, head) list_for_each(pos, head) if ((task=list_entry(pos, struct rpc_task, tk_task)),1) + +typedef void (*rpc_action)(struct rpc_task *); + +struct rpc_call_ops { + void (*rpc_call_prepare)(struct rpc_task *, void *); + void (*rpc_call_done)(struct rpc_task *, void *); + void (*rpc_release)(void *); +}; + +#define RPC_TASK_ASYNC 0x0001 +#define RPC_TASK_SWAPPER 0x0002 +#define RPC_TASK_CHILD 0x0008 +#define RPC_CALL_MAJORSEEN 0x0020 +#define RPC_TASK_ROOTCREDS 0x0040 +#define RPC_TASK_DYNAMIC 0x0080 +#define RPC_TASK_KILLED 0x0100 +#define RPC_TASK_SOFT 0x0200 +#define RPC_TASK_NOINTR 0x0400 + +#define RPC_IS_ASYNC(t) ((t)->tk_flags & RPC_TASK_ASYNC) +#define RPC_IS_CHILD(t) ((t)->tk_flags & RPC_TASK_CHILD) +#define RPC_IS_SWAPPER(t) ((t)->tk_flags & RPC_TASK_SWAPPER) +#define RPC_DO_ROOTOVERRIDE(t) ((t)->tk_flags & RPC_TASK_ROOTCREDS) +#define RPC_ASSASSINATED(t) ((t)->tk_flags & RPC_TASK_KILLED) +#define RPC_DO_CALLBACK(t) ((t)->tk_callback != NULL) +#define RPC_IS_SOFT(t) ((t)->tk_flags & RPC_TASK_SOFT) +#define RPC_TASK_UNINTERRUPTIBLE(t) ((t)->tk_flags & RPC_TASK_NOINTR) + +#define RPC_TASK_RUNNING 0 +#define RPC_TASK_QUEUED 1 +#define RPC_TASK_WAKEUP 2 +#define RPC_TASK_HAS_TIMER 3 +#define RPC_TASK_ACTIVE 4 + +#define RPC_IS_RUNNING(t) (test_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)) +#define rpc_set_running(t) (set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)) +#define rpc_test_and_set_running(t) (test_and_set_bit(RPC_TASK_RUNNING, &(t)->tk_runstate)) +#define rpc_clear_running(t) do { smp_mb__before_clear_bit(); clear_bit(RPC_TASK_RUNNING, &(t)->tk_runstate); smp_mb__after_clear_bit(); } while (0) + +#define RPC_IS_QUEUED(t) (test_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)) +#define rpc_set_queued(t) (set_bit(RPC_TASK_QUEUED, &(t)->tk_runstate)) +#define rpc_clear_queued(t) do { smp_mb__before_clear_bit(); clear_bit(RPC_TASK_QUEUED, &(t)->tk_runstate); smp_mb__after_clear_bit(); } while (0) + +#define rpc_start_wakeup(t) (test_and_set_bit(RPC_TASK_WAKEUP, &(t)->tk_runstate) == 0) +#define rpc_finish_wakeup(t) do { smp_mb__before_clear_bit(); clear_bit(RPC_TASK_WAKEUP, &(t)->tk_runstate); smp_mb__after_clear_bit(); } while (0) + +#define RPC_IS_ACTIVATED(t) (test_bit(RPC_TASK_ACTIVE, &(t)->tk_runstate)) +#define rpc_set_active(t) (set_bit(RPC_TASK_ACTIVE, &(t)->tk_runstate)) +#define rpc_clear_active(t) do { smp_mb__before_clear_bit(); clear_bit(RPC_TASK_ACTIVE, &(t)->tk_runstate); smp_mb__after_clear_bit(); } while(0) + +#define RPC_PRIORITY_LOW 0 +#define RPC_PRIORITY_NORMAL 1 +#define RPC_PRIORITY_HIGH 2 +#define RPC_NR_PRIORITY (RPC_PRIORITY_HIGH+1) + +struct rpc_wait_queue { + spinlock_t lock; + struct list_head tasks[RPC_NR_PRIORITY]; + unsigned long cookie; + unsigned char maxpriority; + unsigned char priority; + unsigned char count; + unsigned char nr; + unsigned short qlen; +#ifdef RPC_DEBUG + const char * name; +#endif +}; + +#define RPC_BATCH_COUNT 16 + +#ifndef RPC_DEBUG +#define RPC_WAITQ_INIT(var,qname) { .lock = SPIN_LOCK_UNLOCKED, .tasks = { [0] = LIST_HEAD_INIT(var.tasks[0]), [1] = LIST_HEAD_INIT(var.tasks[1]), [2] = LIST_HEAD_INIT(var.tasks[2]), }, } +#else +#define RPC_WAITQ_INIT(var,qname) { .lock = SPIN_LOCK_UNLOCKED, .tasks = { [0] = LIST_HEAD_INIT(var.tasks[0]), [1] = LIST_HEAD_INIT(var.tasks[1]), [2] = LIST_HEAD_INIT(var.tasks[2]), }, .name = qname, } +#endif +#define RPC_WAITQ(var,qname) struct rpc_wait_queue var = RPC_WAITQ_INIT(var,qname) + +#define RPC_IS_PRIORITY(q) ((q)->maxpriority > 0) + +struct rpc_task *rpc_new_task(struct rpc_clnt *, int flags, + const struct rpc_call_ops *ops, void *data); +struct rpc_task *rpc_run_task(struct rpc_clnt *clnt, int flags, + const struct rpc_call_ops *ops, void *data); +struct rpc_task *rpc_new_child(struct rpc_clnt *, struct rpc_task *parent); + +struct rpc_task *rpc_wake_up_next(struct rpc_wait_queue *); + +#ifdef RPC_DEBUG + +#endif + +#ifdef RPC_DEBUG +#endif +#endif