2005-09-11 20:15:07 -06:00
|
|
|
/*
|
|
|
|
* connector.h
|
|
|
|
*
|
|
|
|
* 2004-2005 Copyright (c) Evgeniy Polyakov <johnpol@2ka.mipt.ru>
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License as published by
|
|
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
|
|
* (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef __CONNECTOR_H
|
|
|
|
#define __CONNECTOR_H
|
|
|
|
|
2009-01-15 14:51:26 -07:00
|
|
|
#include <linux/types.h>
|
2005-09-11 20:15:07 -06:00
|
|
|
|
|
|
|
#define CN_IDX_CONNECTOR 0xffffffff
|
|
|
|
#define CN_VAL_CONNECTOR 0xffffffff
|
|
|
|
|
2005-11-07 01:59:16 -07:00
|
|
|
/*
|
|
|
|
* Process Events connector unique ids -- used for message routing
|
|
|
|
*/
|
|
|
|
#define CN_IDX_PROC 0x1
|
|
|
|
#define CN_VAL_PROC 0x1
|
2005-11-10 16:33:38 -07:00
|
|
|
#define CN_IDX_CIFS 0x2
|
|
|
|
#define CN_VAL_CIFS 0x1
|
2006-03-23 09:11:58 -07:00
|
|
|
#define CN_W1_IDX 0x3 /* w1 communication */
|
|
|
|
#define CN_W1_VAL 0x1
|
2007-10-16 02:28:25 -06:00
|
|
|
#define CN_IDX_V86D 0x4
|
|
|
|
#define CN_VAL_V86D_UVESAFB 0x1
|
2008-07-29 23:33:33 -06:00
|
|
|
#define CN_IDX_BB 0x5 /* BlackBoard, from the TSP GPL sampling framework */
|
2009-01-13 16:05:27 -07:00
|
|
|
#define CN_DST_IDX 0x6
|
|
|
|
#define CN_DST_VAL 0x1
|
2009-06-22 03:12:35 -06:00
|
|
|
#define CN_IDX_DM 0x7 /* Device Mapper */
|
|
|
|
#define CN_VAL_DM_USERSPACE_LOG 0x1
|
2005-11-07 01:59:16 -07:00
|
|
|
|
2009-06-22 03:12:35 -06:00
|
|
|
#define CN_NETLINK_USERS 8
|
2005-09-11 20:15:07 -06:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Maximum connector's message size.
|
|
|
|
*/
|
2007-10-16 02:28:24 -06:00
|
|
|
#define CONNECTOR_MAX_MSG_SIZE 16384
|
2005-09-11 20:15:07 -06:00
|
|
|
|
|
|
|
/*
|
|
|
|
* idx and val are unique identifiers which
|
|
|
|
* are used for message routing and
|
|
|
|
* must be registered in connector.h for in-kernel usage.
|
|
|
|
*/
|
|
|
|
|
|
|
|
struct cb_id {
|
|
|
|
__u32 idx;
|
|
|
|
__u32 val;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct cn_msg {
|
|
|
|
struct cb_id id;
|
|
|
|
|
|
|
|
__u32 seq;
|
|
|
|
__u32 ack;
|
|
|
|
|
|
|
|
__u16 len; /* Length of the following data */
|
|
|
|
__u16 flags;
|
|
|
|
__u8 data[0];
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Notify structure - requests notification about
|
|
|
|
* registering/unregistering idx/val in range [first, first+range].
|
|
|
|
*/
|
|
|
|
struct cn_notify_req {
|
|
|
|
__u32 first;
|
|
|
|
__u32 range;
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Main notification control message
|
|
|
|
* *_notify_num - number of appropriate cn_notify_req structures after
|
|
|
|
* this struct.
|
|
|
|
* group - notification receiver's idx.
|
|
|
|
* len - total length of the attached data.
|
|
|
|
*/
|
|
|
|
struct cn_ctl_msg {
|
|
|
|
__u32 idx_notify_num;
|
|
|
|
__u32 val_notify_num;
|
|
|
|
__u32 group;
|
|
|
|
__u32 len;
|
|
|
|
__u8 data[0];
|
|
|
|
};
|
|
|
|
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
|
|
|
|
#include <asm/atomic.h>
|
|
|
|
|
|
|
|
#include <linux/list.h>
|
|
|
|
#include <linux/workqueue.h>
|
|
|
|
|
|
|
|
#include <net/sock.h>
|
|
|
|
|
|
|
|
#define CN_CBQ_NAMELEN 32
|
|
|
|
|
|
|
|
struct cn_queue_dev {
|
|
|
|
atomic_t refcnt;
|
|
|
|
unsigned char name[CN_CBQ_NAMELEN];
|
|
|
|
|
|
|
|
struct workqueue_struct *cn_queue;
|
connector: create connector workqueue only while needed once
The netlink connector uses its own workqueue to relay the datas sent
from userspace to the appropriate callback. If you launch the test
from Documentation/connector and change it a bit to send a high flow
of data, you will see thousands of events coming to the "cqueue"
workqueue by looking at the workqueue tracer.
This flow of events can be sent very quickly. So, to not encumber the
kevent workqueue and delay other jobs, the "cqueue" workqueue should
remain.
But this workqueue is pointless most of the time, it will always be
created (assuming you have built it of course) although only
developpers with specific needs will use it.
So avoid this "most of the time useless task", this patch proposes to
create this workqueue only when needed once. The first jobs to be
sent to connector callbacks will be sent to kevent while the "cqueue"
thread creation will be scheduled to kevent too.
The following jobs will continue to be scheduled to keventd until the
cqueue workqueue is created, and then the rest of the jobs will
continue to perform as usual, through this dedicated workqueue.
Each time I tested this patch, only the first event was sent to
keventd, the rest has been sent to cqueue which have been created
quickly.
Also, this patch fixes some trailing whitespaces on the connector files.
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Acked-by: Evgeniy Polyakov <zbr@ioremap.net>
Signed-off-by: David S. Miller <davem@davemloft.net>
2009-02-03 00:22:04 -07:00
|
|
|
/* Sent to kevent to create cn_queue only when needed */
|
|
|
|
struct work_struct wq_creation;
|
|
|
|
/* Tell if the wq_creation job is pending/completed */
|
|
|
|
atomic_t wq_requested;
|
|
|
|
/* Wait for cn_queue to be created */
|
|
|
|
wait_queue_head_t wq_created;
|
2005-09-11 20:15:07 -06:00
|
|
|
|
|
|
|
struct list_head queue_list;
|
|
|
|
spinlock_t queue_lock;
|
|
|
|
|
|
|
|
struct sock *nls;
|
|
|
|
};
|
|
|
|
|
2005-09-26 16:06:50 -06:00
|
|
|
struct cn_callback_id {
|
2005-09-11 20:15:07 -06:00
|
|
|
unsigned char name[CN_CBQ_NAMELEN];
|
|
|
|
struct cb_id id;
|
2005-09-26 16:06:50 -06:00
|
|
|
};
|
|
|
|
|
|
|
|
struct cn_callback_data {
|
2009-10-01 20:40:04 -06:00
|
|
|
struct sk_buff *skb;
|
2009-10-01 20:40:05 -06:00
|
|
|
void (*callback) (struct cn_msg *, struct netlink_skb_parms *);
|
2005-09-26 16:06:50 -06:00
|
|
|
|
|
|
|
void *free;
|
2005-09-11 20:15:07 -06:00
|
|
|
};
|
|
|
|
|
|
|
|
struct cn_callback_entry {
|
|
|
|
struct list_head callback_entry;
|
2006-12-18 02:53:58 -07:00
|
|
|
struct work_struct work;
|
2005-09-11 20:15:07 -06:00
|
|
|
struct cn_queue_dev *pdev;
|
|
|
|
|
2005-09-26 16:06:50 -06:00
|
|
|
struct cn_callback_id id;
|
|
|
|
struct cn_callback_data data;
|
2005-09-11 20:15:07 -06:00
|
|
|
|
2008-01-04 02:59:42 -07:00
|
|
|
u32 seq, group;
|
2005-09-11 20:15:07 -06:00
|
|
|
};
|
|
|
|
|
|
|
|
struct cn_ctl_entry {
|
|
|
|
struct list_head notify_entry;
|
|
|
|
struct cn_ctl_msg *msg;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct cn_dev {
|
|
|
|
struct cb_id id;
|
|
|
|
|
|
|
|
u32 seq, groups;
|
|
|
|
struct sock *nls;
|
2007-10-10 22:15:29 -06:00
|
|
|
void (*input) (struct sk_buff *skb);
|
2005-09-11 20:15:07 -06:00
|
|
|
|
|
|
|
struct cn_queue_dev *cbdev;
|
|
|
|
};
|
|
|
|
|
2009-10-01 20:40:05 -06:00
|
|
|
int cn_add_callback(struct cb_id *, char *, void (*callback) (struct cn_msg *, struct netlink_skb_parms *));
|
2005-09-11 20:15:07 -06:00
|
|
|
void cn_del_callback(struct cb_id *);
|
2005-10-07 00:46:04 -06:00
|
|
|
int cn_netlink_send(struct cn_msg *, u32, gfp_t);
|
2005-09-11 20:15:07 -06:00
|
|
|
|
2009-10-01 20:40:05 -06:00
|
|
|
int cn_queue_add_callback(struct cn_queue_dev *dev, char *name, struct cb_id *id, void (*callback)(struct cn_msg *, struct netlink_skb_parms *));
|
2005-09-11 20:15:07 -06:00
|
|
|
void cn_queue_del_callback(struct cn_queue_dev *dev, struct cb_id *id);
|
|
|
|
|
connector: create connector workqueue only while needed once
The netlink connector uses its own workqueue to relay the datas sent
from userspace to the appropriate callback. If you launch the test
from Documentation/connector and change it a bit to send a high flow
of data, you will see thousands of events coming to the "cqueue"
workqueue by looking at the workqueue tracer.
This flow of events can be sent very quickly. So, to not encumber the
kevent workqueue and delay other jobs, the "cqueue" workqueue should
remain.
But this workqueue is pointless most of the time, it will always be
created (assuming you have built it of course) although only
developpers with specific needs will use it.
So avoid this "most of the time useless task", this patch proposes to
create this workqueue only when needed once. The first jobs to be
sent to connector callbacks will be sent to kevent while the "cqueue"
thread creation will be scheduled to kevent too.
The following jobs will continue to be scheduled to keventd until the
cqueue workqueue is created, and then the rest of the jobs will
continue to perform as usual, through this dedicated workqueue.
Each time I tested this patch, only the first event was sent to
keventd, the rest has been sent to cqueue which have been created
quickly.
Also, this patch fixes some trailing whitespaces on the connector files.
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Acked-by: Evgeniy Polyakov <zbr@ioremap.net>
Signed-off-by: David S. Miller <davem@davemloft.net>
2009-02-03 00:22:04 -07:00
|
|
|
int queue_cn_work(struct cn_callback_entry *cbq, struct work_struct *work);
|
|
|
|
|
2005-09-11 20:15:07 -06:00
|
|
|
struct cn_queue_dev *cn_queue_alloc_dev(char *name, struct sock *);
|
|
|
|
void cn_queue_free_dev(struct cn_queue_dev *dev);
|
|
|
|
|
|
|
|
int cn_cb_equal(struct cb_id *, struct cb_id *);
|
|
|
|
|
2006-11-22 07:57:56 -07:00
|
|
|
void cn_queue_wrapper(struct work_struct *work);
|
2005-09-26 16:06:50 -06:00
|
|
|
|
2005-09-11 20:15:07 -06:00
|
|
|
#endif /* __KERNEL__ */
|
|
|
|
#endif /* __CONNECTOR_H */
|