c6ef1e7fdc
Unfortunelly (my fault) the kernel thread that now handles IR processing has classical races in regard to wakeup and stop. This patch hopefully closes them all. Tested with module reload running in a loop, while receiver is blasted with IR data for 10 minutes. Signed-off-by: Maxim Levitsky <maximlevitsky@gmail.com> Acked-by: Jarod Wilson <jarod@redhat.com> Signed-off-by: Mauro Carvalho Chehab <mchehab@redhat.com>
395 lines
10 KiB
C
395 lines
10 KiB
C
/* ir-raw-event.c - handle IR Pulse/Space event
|
|
*
|
|
* Copyright (C) 2010 by Mauro Carvalho Chehab <mchehab@redhat.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation version 2 of the License.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*/
|
|
|
|
#include <linux/kthread.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/freezer.h>
|
|
#include "ir-core-priv.h"
|
|
|
|
/* Define the max number of pulse/space transitions to buffer */
|
|
#define MAX_IR_EVENT_SIZE 512
|
|
|
|
/* Used to keep track of IR raw clients, protected by ir_raw_handler_lock */
|
|
static LIST_HEAD(ir_raw_client_list);
|
|
|
|
/* Used to handle IR raw handler extensions */
|
|
static DEFINE_MUTEX(ir_raw_handler_lock);
|
|
static LIST_HEAD(ir_raw_handler_list);
|
|
static u64 available_protocols;
|
|
|
|
#ifdef MODULE
|
|
/* Used to load the decoders */
|
|
static struct work_struct wq_load;
|
|
#endif
|
|
|
|
static int ir_raw_event_thread(void *data)
|
|
{
|
|
struct ir_raw_event ev;
|
|
struct ir_raw_handler *handler;
|
|
struct ir_raw_event_ctrl *raw = (struct ir_raw_event_ctrl *)data;
|
|
int retval;
|
|
|
|
while (!kthread_should_stop()) {
|
|
|
|
spin_lock_irq(&raw->lock);
|
|
retval = kfifo_out(&raw->kfifo, &ev, sizeof(ev));
|
|
|
|
if (!retval) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
|
|
if (kthread_should_stop())
|
|
set_current_state(TASK_RUNNING);
|
|
|
|
spin_unlock_irq(&raw->lock);
|
|
schedule();
|
|
continue;
|
|
}
|
|
|
|
spin_unlock_irq(&raw->lock);
|
|
|
|
|
|
BUG_ON(retval != sizeof(ev));
|
|
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
list_for_each_entry(handler, &ir_raw_handler_list, list)
|
|
handler->decode(raw->input_dev, ev);
|
|
raw->prev_ev = ev;
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ir_raw_event_store() - pass a pulse/space duration to the raw ir decoders
|
|
* @input_dev: the struct input_dev device descriptor
|
|
* @ev: the struct ir_raw_event descriptor of the pulse/space
|
|
*
|
|
* This routine (which may be called from an interrupt context) stores a
|
|
* pulse/space duration for the raw ir decoding state machines. Pulses are
|
|
* signalled as positive values and spaces as negative values. A zero value
|
|
* will reset the decoding state machines.
|
|
*/
|
|
int ir_raw_event_store(struct input_dev *input_dev, struct ir_raw_event *ev)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
|
|
if (!ir->raw)
|
|
return -EINVAL;
|
|
|
|
IR_dprintk(2, "sample: (05%dus %s)\n",
|
|
TO_US(ev->duration), TO_STR(ev->pulse));
|
|
|
|
if (kfifo_in(&ir->raw->kfifo, ev, sizeof(*ev)) != sizeof(*ev))
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ir_raw_event_store);
|
|
|
|
/**
|
|
* ir_raw_event_store_edge() - notify raw ir decoders of the start of a pulse/space
|
|
* @input_dev: the struct input_dev device descriptor
|
|
* @type: the type of the event that has occurred
|
|
*
|
|
* This routine (which may be called from an interrupt context) is used to
|
|
* store the beginning of an ir pulse or space (or the start/end of ir
|
|
* reception) for the raw ir decoding state machines. This is used by
|
|
* hardware which does not provide durations directly but only interrupts
|
|
* (or similar events) on state change.
|
|
*/
|
|
int ir_raw_event_store_edge(struct input_dev *input_dev, enum raw_event_type type)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
ktime_t now;
|
|
s64 delta; /* ns */
|
|
struct ir_raw_event ev;
|
|
int rc = 0;
|
|
|
|
if (!ir->raw)
|
|
return -EINVAL;
|
|
|
|
now = ktime_get();
|
|
delta = ktime_to_ns(ktime_sub(now, ir->raw->last_event));
|
|
|
|
/* Check for a long duration since last event or if we're
|
|
* being called for the first time, note that delta can't
|
|
* possibly be negative.
|
|
*/
|
|
ev.duration = 0;
|
|
if (delta > IR_MAX_DURATION || !ir->raw->last_type)
|
|
type |= IR_START_EVENT;
|
|
else
|
|
ev.duration = delta;
|
|
|
|
if (type & IR_START_EVENT)
|
|
ir_raw_event_reset(input_dev);
|
|
else if (ir->raw->last_type & IR_SPACE) {
|
|
ev.pulse = false;
|
|
rc = ir_raw_event_store(input_dev, &ev);
|
|
} else if (ir->raw->last_type & IR_PULSE) {
|
|
ev.pulse = true;
|
|
rc = ir_raw_event_store(input_dev, &ev);
|
|
} else
|
|
return 0;
|
|
|
|
ir->raw->last_event = now;
|
|
ir->raw->last_type = type;
|
|
return rc;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ir_raw_event_store_edge);
|
|
|
|
/**
|
|
* ir_raw_event_store_with_filter() - pass next pulse/space to decoders with some processing
|
|
* @input_dev: the struct input_dev device descriptor
|
|
* @type: the type of the event that has occurred
|
|
*
|
|
* This routine (which may be called from an interrupt context) works
|
|
* in similiar manner to ir_raw_event_store_edge.
|
|
* This routine is intended for devices with limited internal buffer
|
|
* It automerges samples of same type, and handles timeouts
|
|
*/
|
|
int ir_raw_event_store_with_filter(struct input_dev *input_dev,
|
|
struct ir_raw_event *ev)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
struct ir_raw_event_ctrl *raw = ir->raw;
|
|
|
|
if (!raw || !ir->props)
|
|
return -EINVAL;
|
|
|
|
/* Ignore spaces in idle mode */
|
|
if (ir->idle && !ev->pulse)
|
|
return 0;
|
|
else if (ir->idle)
|
|
ir_raw_event_set_idle(input_dev, 0);
|
|
|
|
if (!raw->this_ev.duration) {
|
|
raw->this_ev = *ev;
|
|
} else if (ev->pulse == raw->this_ev.pulse) {
|
|
raw->this_ev.duration += ev->duration;
|
|
} else {
|
|
ir_raw_event_store(input_dev, &raw->this_ev);
|
|
raw->this_ev = *ev;
|
|
}
|
|
|
|
/* Enter idle mode if nessesary */
|
|
if (!ev->pulse && ir->props->timeout &&
|
|
raw->this_ev.duration >= ir->props->timeout)
|
|
ir_raw_event_set_idle(input_dev, 1);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ir_raw_event_store_with_filter);
|
|
|
|
void ir_raw_event_set_idle(struct input_dev *input_dev, int idle)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
struct ir_raw_event_ctrl *raw = ir->raw;
|
|
ktime_t now;
|
|
u64 delta;
|
|
|
|
if (!ir->props)
|
|
return;
|
|
|
|
if (!ir->raw)
|
|
goto out;
|
|
|
|
if (idle) {
|
|
IR_dprintk(2, "enter idle mode\n");
|
|
raw->last_event = ktime_get();
|
|
} else {
|
|
IR_dprintk(2, "exit idle mode\n");
|
|
|
|
now = ktime_get();
|
|
delta = ktime_to_ns(ktime_sub(now, ir->raw->last_event));
|
|
|
|
WARN_ON(raw->this_ev.pulse);
|
|
|
|
raw->this_ev.duration =
|
|
min(raw->this_ev.duration + delta,
|
|
(u64)IR_MAX_DURATION);
|
|
|
|
ir_raw_event_store(input_dev, &raw->this_ev);
|
|
|
|
if (raw->this_ev.duration == IR_MAX_DURATION)
|
|
ir_raw_event_reset(input_dev);
|
|
|
|
raw->this_ev.duration = 0;
|
|
}
|
|
out:
|
|
if (ir->props->s_idle)
|
|
ir->props->s_idle(ir->props->priv, idle);
|
|
ir->idle = idle;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ir_raw_event_set_idle);
|
|
|
|
/**
|
|
* ir_raw_event_handle() - schedules the decoding of stored ir data
|
|
* @input_dev: the struct input_dev device descriptor
|
|
*
|
|
* This routine will signal the workqueue to start decoding stored ir data.
|
|
*/
|
|
void ir_raw_event_handle(struct input_dev *input_dev)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
unsigned long flags;
|
|
|
|
if (!ir->raw)
|
|
return;
|
|
|
|
spin_lock_irqsave(&ir->raw->lock, flags);
|
|
wake_up_process(ir->raw->thread);
|
|
spin_unlock_irqrestore(&ir->raw->lock, flags);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ir_raw_event_handle);
|
|
|
|
/* used internally by the sysfs interface */
|
|
u64
|
|
ir_raw_get_allowed_protocols()
|
|
{
|
|
u64 protocols;
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
protocols = available_protocols;
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
return protocols;
|
|
}
|
|
|
|
/*
|
|
* Used to (un)register raw event clients
|
|
*/
|
|
int ir_raw_event_register(struct input_dev *input_dev)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
int rc;
|
|
struct ir_raw_handler *handler;
|
|
|
|
ir->raw = kzalloc(sizeof(*ir->raw), GFP_KERNEL);
|
|
if (!ir->raw)
|
|
return -ENOMEM;
|
|
|
|
ir->raw->input_dev = input_dev;
|
|
|
|
ir->raw->enabled_protocols = ~0;
|
|
rc = kfifo_alloc(&ir->raw->kfifo, sizeof(s64) * MAX_IR_EVENT_SIZE,
|
|
GFP_KERNEL);
|
|
if (rc < 0) {
|
|
kfree(ir->raw);
|
|
ir->raw = NULL;
|
|
return rc;
|
|
}
|
|
|
|
spin_lock_init(&ir->raw->lock);
|
|
ir->raw->thread = kthread_run(ir_raw_event_thread, ir->raw,
|
|
"rc%u", (unsigned int)ir->devno);
|
|
|
|
if (IS_ERR(ir->raw->thread)) {
|
|
int ret = PTR_ERR(ir->raw->thread);
|
|
|
|
kfree(ir->raw);
|
|
ir->raw = NULL;
|
|
return ret;
|
|
}
|
|
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
list_add_tail(&ir->raw->list, &ir_raw_client_list);
|
|
list_for_each_entry(handler, &ir_raw_handler_list, list)
|
|
if (handler->raw_register)
|
|
handler->raw_register(ir->raw->input_dev);
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void ir_raw_event_unregister(struct input_dev *input_dev)
|
|
{
|
|
struct ir_input_dev *ir = input_get_drvdata(input_dev);
|
|
struct ir_raw_handler *handler;
|
|
|
|
if (!ir->raw)
|
|
return;
|
|
|
|
kthread_stop(ir->raw->thread);
|
|
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
list_del(&ir->raw->list);
|
|
list_for_each_entry(handler, &ir_raw_handler_list, list)
|
|
if (handler->raw_unregister)
|
|
handler->raw_unregister(ir->raw->input_dev);
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
|
|
kfifo_free(&ir->raw->kfifo);
|
|
kfree(ir->raw);
|
|
ir->raw = NULL;
|
|
}
|
|
|
|
/*
|
|
* Extension interface - used to register the IR decoders
|
|
*/
|
|
|
|
int ir_raw_handler_register(struct ir_raw_handler *ir_raw_handler)
|
|
{
|
|
struct ir_raw_event_ctrl *raw;
|
|
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
list_add_tail(&ir_raw_handler->list, &ir_raw_handler_list);
|
|
if (ir_raw_handler->raw_register)
|
|
list_for_each_entry(raw, &ir_raw_client_list, list)
|
|
ir_raw_handler->raw_register(raw->input_dev);
|
|
available_protocols |= ir_raw_handler->protocols;
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(ir_raw_handler_register);
|
|
|
|
void ir_raw_handler_unregister(struct ir_raw_handler *ir_raw_handler)
|
|
{
|
|
struct ir_raw_event_ctrl *raw;
|
|
|
|
mutex_lock(&ir_raw_handler_lock);
|
|
list_del(&ir_raw_handler->list);
|
|
if (ir_raw_handler->raw_unregister)
|
|
list_for_each_entry(raw, &ir_raw_client_list, list)
|
|
ir_raw_handler->raw_unregister(raw->input_dev);
|
|
available_protocols &= ~ir_raw_handler->protocols;
|
|
mutex_unlock(&ir_raw_handler_lock);
|
|
}
|
|
EXPORT_SYMBOL(ir_raw_handler_unregister);
|
|
|
|
#ifdef MODULE
|
|
static void init_decoders(struct work_struct *work)
|
|
{
|
|
/* Load the decoder modules */
|
|
|
|
load_nec_decode();
|
|
load_rc5_decode();
|
|
load_rc6_decode();
|
|
load_jvc_decode();
|
|
load_sony_decode();
|
|
load_lirc_codec();
|
|
|
|
/* If needed, we may later add some init code. In this case,
|
|
it is needed to change the CONFIG_MODULE test at ir-core.h
|
|
*/
|
|
}
|
|
#endif
|
|
|
|
void ir_raw_init(void)
|
|
{
|
|
#ifdef MODULE
|
|
INIT_WORK(&wq_load, init_decoders);
|
|
schedule_work(&wq_load);
|
|
#endif
|
|
}
|