|
@@ -13,11 +13,12 @@
|
|
|
#include <rtdevice.h>
|
|
|
|
|
|
#ifdef RT_USING_HEAP
|
|
|
+
|
|
|
rt_inline rt_err_t _workqueue_work_completion(struct rt_workqueue *queue)
|
|
|
{
|
|
|
rt_err_t result;
|
|
|
-
|
|
|
- rt_enter_critical();
|
|
|
+
|
|
|
+ rt_enter_critical();
|
|
|
while (1)
|
|
|
{
|
|
|
/* try to take condition semaphore */
|
|
@@ -40,17 +41,17 @@ rt_inline rt_err_t _workqueue_work_completion(struct rt_workqueue *queue)
|
|
|
}
|
|
|
}
|
|
|
rt_exit_critical();
|
|
|
-
|
|
|
+
|
|
|
return result;
|
|
|
}
|
|
|
|
|
|
-static void _workqueue_thread_entry(void* parameter)
|
|
|
+static void _workqueue_thread_entry(void *parameter)
|
|
|
{
|
|
|
rt_base_t level;
|
|
|
- struct rt_work* work;
|
|
|
- struct rt_workqueue* queue;
|
|
|
+ struct rt_work *work;
|
|
|
+ struct rt_workqueue *queue;
|
|
|
|
|
|
- queue = (struct rt_workqueue*) parameter;
|
|
|
+ queue = (struct rt_workqueue *) parameter;
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
|
|
|
|
while (1)
|
|
@@ -67,6 +68,7 @@ static void _workqueue_thread_entry(void* parameter)
|
|
|
work = rt_list_entry(queue->work_list.next, struct rt_work, list);
|
|
|
rt_list_remove(&(work->list));
|
|
|
queue->work_current = work;
|
|
|
+ work->flags &= ~RT_WORK_STATE_PENDING;
|
|
|
rt_hw_interrupt_enable(level);
|
|
|
|
|
|
/* do work */
|
|
@@ -81,11 +83,156 @@ static void _workqueue_thread_entry(void* parameter)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-struct rt_workqueue *rt_workqueue_create(const char* name, rt_uint16_t stack_size, rt_uint8_t priority)
|
|
|
+static rt_err_t _workqueue_submit_work(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
+{
|
|
|
+ rt_base_t level;
|
|
|
+
|
|
|
+ level = rt_hw_interrupt_disable();
|
|
|
+ if (work->flags & RT_WORK_STATE_PENDING)
|
|
|
+ {
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+ return -RT_EBUSY;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (queue->work_current == work)
|
|
|
+ {
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+ return -RT_EBUSY;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* NOTE: the work MUST be initialized firstly */
|
|
|
+ rt_list_remove(&(work->list));
|
|
|
+
|
|
|
+ rt_list_insert_after(queue->work_list.prev, &(work->list));
|
|
|
+ work->flags |= RT_WORK_STATE_PENDING;
|
|
|
+
|
|
|
+ /* whether the workqueue is doing work */
|
|
|
+ if (queue->work_current == RT_NULL)
|
|
|
+ {
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+ /* resume work thread */
|
|
|
+ rt_thread_resume(queue->work_thread);
|
|
|
+ rt_schedule();
|
|
|
+ }
|
|
|
+ else
|
|
|
+ {
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+ }
|
|
|
+
|
|
|
+ return RT_EOK;
|
|
|
+}
|
|
|
+
|
|
|
+static rt_err_t _workqueue_cancel_work(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
+{
|
|
|
+ rt_base_t level;
|
|
|
+
|
|
|
+ level = rt_hw_interrupt_disable();
|
|
|
+ if (queue->work_current == work)
|
|
|
+ {
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+ return -RT_EBUSY;
|
|
|
+ }
|
|
|
+ rt_list_remove(&(work->list));
|
|
|
+ work->flags &= ~RT_WORK_STATE_PENDING;
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+
|
|
|
+ return RT_EOK;
|
|
|
+}
|
|
|
+
|
|
|
+static rt_err_t _workqueue_cancel_delayed_work(struct rt_delayed_work *work)
|
|
|
+{
|
|
|
+ rt_base_t level;
|
|
|
+ int ret = RT_EOK;
|
|
|
+
|
|
|
+ if (!work->workqueue)
|
|
|
+ {
|
|
|
+ ret = -EINVAL;
|
|
|
+ goto __exit;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (work->work.flags & RT_WORK_STATE_PENDING)
|
|
|
+ {
|
|
|
+ /* Remove from the queue if already submitted */
|
|
|
+ ret = rt_workqueue_cancel_work(work->workqueue, &(work->work));
|
|
|
+ if (ret)
|
|
|
+ {
|
|
|
+ goto __exit;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ else
|
|
|
+ {
|
|
|
+ rt_timer_stop(&(work->timer));
|
|
|
+ }
|
|
|
+
|
|
|
+ level = rt_hw_interrupt_disable();
|
|
|
+ /* Detach from workqueue */
|
|
|
+ work->workqueue = RT_NULL;
|
|
|
+ work->work.flags &= ~(RT_WORK_STATE_PENDING);
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+
|
|
|
+__exit:
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+static rt_err_t _workqueue_submit_delayed_work(struct rt_workqueue *queue,
|
|
|
+ struct rt_delayed_work *work, rt_tick_t ticks)
|
|
|
+{
|
|
|
+ rt_base_t level;
|
|
|
+ int ret = RT_EOK;
|
|
|
+
|
|
|
+
|
|
|
+ /* Work cannot be active in multiple queues */
|
|
|
+ if (work->workqueue && work->workqueue != queue)
|
|
|
+ {
|
|
|
+ ret = -RT_EINVAL;
|
|
|
+ goto __exit;
|
|
|
+ }
|
|
|
+
|
|
|
+ /* Cancel if work has been submitted */
|
|
|
+ if (work->workqueue == queue)
|
|
|
+ {
|
|
|
+ ret = _workqueue_cancel_delayed_work(work);
|
|
|
+ if (ret < 0)
|
|
|
+ {
|
|
|
+ goto __exit;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ level = rt_hw_interrupt_disable();
|
|
|
+ /* Attach workqueue so the timeout callback can submit it */
|
|
|
+ work->workqueue = queue;
|
|
|
+ rt_hw_interrupt_enable(level);
|
|
|
+
|
|
|
+ if (!ticks)
|
|
|
+ {
|
|
|
+ /* Submit work if no ticks is 0 */
|
|
|
+ _workqueue_submit_work(work->workqueue, &(work->work));
|
|
|
+ }
|
|
|
+ else
|
|
|
+ {
|
|
|
+ /* Add timeout */
|
|
|
+ rt_timer_control(&(work->timer), RT_TIMER_CTRL_SET_TIME, &ticks);
|
|
|
+ rt_timer_start(&(work->timer));
|
|
|
+ }
|
|
|
+
|
|
|
+__exit:
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+static void _delayed_work_timeout_handler(void *parameter)
|
|
|
+{
|
|
|
+ struct rt_delayed_work *delayed_work;
|
|
|
+
|
|
|
+ delayed_work = (struct rt_delayed_work *)parameter;
|
|
|
+ rt_timer_stop(&(delayed_work->timer));
|
|
|
+ _workqueue_submit_work(delayed_work->workqueue, &(delayed_work->work));
|
|
|
+}
|
|
|
+
|
|
|
+struct rt_workqueue *rt_workqueue_create(const char *name, rt_uint16_t stack_size, rt_uint8_t priority)
|
|
|
{
|
|
|
struct rt_workqueue *queue = RT_NULL;
|
|
|
|
|
|
- queue = (struct rt_workqueue*)RT_KERNEL_MALLOC(sizeof(struct rt_workqueue));
|
|
|
+ queue = (struct rt_workqueue *)RT_KERNEL_MALLOC(sizeof(struct rt_workqueue));
|
|
|
if (queue != RT_NULL)
|
|
|
{
|
|
|
/* initialize work list */
|
|
@@ -107,7 +254,7 @@ struct rt_workqueue *rt_workqueue_create(const char* name, rt_uint16_t stack_siz
|
|
|
return queue;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_destroy(struct rt_workqueue* queue)
|
|
|
+rt_err_t rt_workqueue_destroy(struct rt_workqueue *queue)
|
|
|
{
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
|
|
|
@@ -117,37 +264,30 @@ rt_err_t rt_workqueue_destroy(struct rt_workqueue* queue)
|
|
|
return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_dowork(struct rt_workqueue* queue, struct rt_work* work)
|
|
|
+rt_err_t rt_workqueue_dowork(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
{
|
|
|
- rt_base_t level;
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
|
RT_ASSERT(work != RT_NULL);
|
|
|
|
|
|
- level = rt_hw_interrupt_disable();
|
|
|
- if (queue->work_current == work)
|
|
|
- {
|
|
|
- rt_hw_interrupt_enable(level);
|
|
|
- return -RT_EBUSY;
|
|
|
- }
|
|
|
+ return _workqueue_submit_work(queue, work);
|
|
|
+}
|
|
|
|
|
|
- /* NOTE: the work MUST be initialized firstly */
|
|
|
- rt_list_remove(&(work->list));
|
|
|
+rt_err_t rt_workqueue_submit_work(struct rt_workqueue *queue, struct rt_work *work, rt_tick_t time)
|
|
|
+{
|
|
|
+ RT_ASSERT(queue != RT_NULL);
|
|
|
+ RT_ASSERT(work != RT_NULL);
|
|
|
|
|
|
- rt_list_insert_after(queue->work_list.prev, &(work->list));
|
|
|
- /* whether the workqueue is doing work */
|
|
|
- if (queue->work_current == RT_NULL)
|
|
|
+ if (work->type & RT_WORK_TYPE_DELAYED)
|
|
|
{
|
|
|
- rt_hw_interrupt_enable(level);
|
|
|
- /* resume work thread */
|
|
|
- rt_thread_resume(queue->work_thread);
|
|
|
- rt_schedule();
|
|
|
+ return _workqueue_submit_delayed_work(queue, (struct rt_delayed_work *)work, time);
|
|
|
+ }
|
|
|
+ else
|
|
|
+ {
|
|
|
+ return _workqueue_submit_work(queue, work);
|
|
|
}
|
|
|
- else rt_hw_interrupt_enable(level);
|
|
|
-
|
|
|
- return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_critical_work(struct rt_workqueue* queue, struct rt_work* work)
|
|
|
+rt_err_t rt_workqueue_critical_work(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
{
|
|
|
rt_base_t level;
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
@@ -176,26 +316,22 @@ rt_err_t rt_workqueue_critical_work(struct rt_workqueue* queue, struct rt_work*
|
|
|
return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_cancel_work(struct rt_workqueue* queue, struct rt_work* work)
|
|
|
+rt_err_t rt_workqueue_cancel_work(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
{
|
|
|
- rt_base_t level;
|
|
|
-
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
|
RT_ASSERT(work != RT_NULL);
|
|
|
|
|
|
- level = rt_hw_interrupt_disable();
|
|
|
- if (queue->work_current == work)
|
|
|
+ if (work->type & RT_WORK_TYPE_DELAYED)
|
|
|
{
|
|
|
- rt_hw_interrupt_enable(level);
|
|
|
- return -RT_EBUSY;
|
|
|
+ return _workqueue_cancel_delayed_work((struct rt_delayed_work *)work);
|
|
|
+ }
|
|
|
+ else
|
|
|
+ {
|
|
|
+ return _workqueue_cancel_work(queue, work);
|
|
|
}
|
|
|
- rt_list_remove(&(work->list));
|
|
|
- rt_hw_interrupt_enable(level);
|
|
|
-
|
|
|
- return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_cancel_work_sync(struct rt_workqueue* queue, struct rt_work* work)
|
|
|
+rt_err_t rt_workqueue_cancel_work_sync(struct rt_workqueue *queue, struct rt_work *work)
|
|
|
{
|
|
|
rt_base_t level;
|
|
|
|
|
@@ -212,12 +348,13 @@ rt_err_t rt_workqueue_cancel_work_sync(struct rt_workqueue* queue, struct rt_wor
|
|
|
{
|
|
|
rt_list_remove(&(work->list));
|
|
|
}
|
|
|
+ work->flags &= ~RT_WORK_STATE_PENDING;
|
|
|
rt_hw_interrupt_enable(level);
|
|
|
|
|
|
return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-rt_err_t rt_workqueue_cancel_all_work(struct rt_workqueue* queue)
|
|
|
+rt_err_t rt_workqueue_cancel_all_work(struct rt_workqueue *queue)
|
|
|
{
|
|
|
struct rt_list_node *node, *next;
|
|
|
RT_ASSERT(queue != RT_NULL);
|
|
@@ -233,5 +370,36 @@ rt_err_t rt_workqueue_cancel_all_work(struct rt_workqueue* queue)
|
|
|
return RT_EOK;
|
|
|
}
|
|
|
|
|
|
-#endif
|
|
|
+void rt_delayed_work_init(struct rt_delayed_work *work, void (*work_func)(struct rt_work *work,
|
|
|
+ void *work_data), void *work_data)
|
|
|
+{
|
|
|
+ rt_work_init(&(work->work), work_func, work_data);
|
|
|
+ work->work.type = RT_WORK_TYPE_DELAYED;
|
|
|
+ rt_timer_init(&(work->timer), "work", _delayed_work_timeout_handler, work, 0,
|
|
|
+ RT_TIMER_FLAG_ONE_SHOT | RT_TIMER_FLAG_SOFT_TIMER);
|
|
|
+}
|
|
|
|
|
|
+#ifdef RT_USING_SYSTEM_WORKQUEUE
|
|
|
+static struct rt_workqueue *sys_workq;
|
|
|
+
|
|
|
+rt_err_t rt_work_submit(struct rt_work *work, rt_tick_t time)
|
|
|
+{
|
|
|
+ return rt_workqueue_submit_work(sys_workq, work, time);
|
|
|
+}
|
|
|
+
|
|
|
+rt_err_t rt_work_cancel(struct rt_work *work)
|
|
|
+{
|
|
|
+ return rt_workqueue_cancel_work(sys_workq, work);
|
|
|
+}
|
|
|
+
|
|
|
+static int rt_work_sys_workqueue_init(void)
|
|
|
+{
|
|
|
+ sys_workq = rt_workqueue_create("sys_work", RT_SYSTEM_WORKQUEUE_STACKSIZE * 4,
|
|
|
+ RT_SYSTEM_WORKQUEUE_PRIORITY);
|
|
|
+
|
|
|
+ return RT_EOK;
|
|
|
+}
|
|
|
+
|
|
|
+INIT_DEVICE_EXPORT(rt_work_sys_workqueue_init);
|
|
|
+#endif
|
|
|
+#endif
|