/* * GPL HEADER START * * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 only, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License version 2 for more details (a copy is included * in the LICENSE file that accompanied this code). * * You should have received a copy of the GNU General Public License * version 2 along with this program; If not, see * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf * * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara, * CA 95054 USA or visit www.sun.com if you need additional information or * have any questions. * * GPL HEADER END */ /* * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved. * Use is subject to license terms. * * Copyright (c) 2011, 2012, Intel Corporation. */ /* * This file is part of Lustre, http://www.lustre.org/ * Lustre is a trademark of Sun Microsystems, Inc. */ #define DEBUG_SUBSYSTEM S_LNET #include #include #include #include #include "../../../include/linux/libcfs/libcfs.h" #if defined(CONFIG_KGDB) #include #endif /** * wait_queue_t of Linux (version < 2.6.34) is a FIFO list for exclusively * waiting threads, which is not always desirable because all threads will * be waken up again and again, even user only needs a few of them to be * active most time. This is not good for performance because cache can * be polluted by different threads. * * LIFO list can resolve this problem because we always wakeup the most * recent active thread by default. * * NB: please don't call non-exclusive & exclusive wait on the same * waitq if add_wait_queue_exclusive_head is used. */ void add_wait_queue_exclusive_head(wait_queue_head_t *waitq, wait_queue_t *link) { unsigned long flags; spin_lock_irqsave(&waitq->lock, flags); __add_wait_queue_exclusive(waitq, link); spin_unlock_irqrestore(&waitq->lock, flags); } EXPORT_SYMBOL(add_wait_queue_exclusive_head); void cfs_init_timer(struct timer_list *t) { init_timer(t); } EXPORT_SYMBOL(cfs_init_timer); void cfs_timer_init(struct timer_list *t, cfs_timer_func_t *func, void *arg) { init_timer(t); t->function = func; t->data = (unsigned long)arg; } EXPORT_SYMBOL(cfs_timer_init); void cfs_timer_done(struct timer_list *t) { return; } EXPORT_SYMBOL(cfs_timer_done); void cfs_timer_arm(struct timer_list *t, unsigned long deadline) { mod_timer(t, deadline); } EXPORT_SYMBOL(cfs_timer_arm); void cfs_timer_disarm(struct timer_list *t) { del_timer(t); } EXPORT_SYMBOL(cfs_timer_disarm); int cfs_timer_is_armed(struct timer_list *t) { return timer_pending(t); } EXPORT_SYMBOL(cfs_timer_is_armed); unsigned long cfs_timer_deadline(struct timer_list *t) { return t->expires; } EXPORT_SYMBOL(cfs_timer_deadline); void cfs_enter_debugger(void) { #if defined(CONFIG_KGDB) /* BREAKPOINT(); */ #else /* nothing */ #endif } EXPORT_SYMBOL(cfs_enter_debugger); sigset_t cfs_block_allsigs(void) { unsigned long flags; sigset_t old; spin_lock_irqsave(¤t->sighand->siglock, flags); old = current->blocked; sigfillset(¤t->blocked); recalc_sigpending(); spin_unlock_irqrestore(¤t->sighand->siglock, flags); return old; } EXPORT_SYMBOL(cfs_block_allsigs); sigset_t cfs_block_sigs(unsigned long sigs) { unsigned long flags; sigset_t old; spin_lock_irqsave(¤t->sighand->siglock, flags); old = current->blocked; sigaddsetmask(¤t->blocked, sigs); recalc_sigpending(); spin_unlock_irqrestore(¤t->sighand->siglock, flags); return old; } EXPORT_SYMBOL(cfs_block_sigs); /* Block all signals except for the @sigs */ sigset_t cfs_block_sigsinv(unsigned long sigs) { unsigned long flags; sigset_t old; spin_lock_irqsave(¤t->sighand->siglock, flags); old = current->blocked; sigaddsetmask(¤t->blocked, ~sigs); recalc_sigpending(); spin_unlock_irqrestore(¤t->sighand->siglock, flags); return old; } EXPORT_SYMBOL(cfs_block_sigsinv); void cfs_restore_sigs(sigset_t old) { unsigned long flags; spin_lock_irqsave(¤t->sighand->siglock, flags); current->blocked = old; recalc_sigpending(); spin_unlock_irqrestore(¤t->sighand->siglock, flags); } EXPORT_SYMBOL(cfs_restore_sigs); int cfs_signal_pending(void) { return signal_pending(current); } EXPORT_SYMBOL(cfs_signal_pending); void cfs_clear_sigpending(void) { unsigned long flags; spin_lock_irqsave(¤t->sighand->siglock, flags); clear_tsk_thread_flag(current, TIF_SIGPENDING); spin_unlock_irqrestore(¤t->sighand->siglock, flags); } EXPORT_SYMBOL(cfs_clear_sigpending); int libcfs_arch_init(void) { return 0; } EXPORT_SYMBOL(libcfs_arch_init); void libcfs_arch_cleanup(void) { return; } EXPORT_SYMBOL(libcfs_arch_cleanup);