50aeea2dc0
- `g_iob_sem.semcount` must be equal to the total number of free IOBs. It can also be negative if there are no free IOBs and there are threads waiting for an IOB. - g_throttle_sem.semcount represents the number of IOBs available for throttled IOB allocations. Like any other semaphore, it should only go negative if there is a thread waiting for it. - Both semaphores are related to the same resource (free IOBs), hence, they must be incremented/decremented simultaneously: - Whenever a IOB buffer is freed, if a thread is waiting for a non-throttled IOB or a thread is waiting for a throttled IOB and we have at least `CONFIG_IOB_THROTTLE` buffers available, the IOB is put in the committed list (`g_iob_committed`). Otherwise, it is put in the common free list (`g_iob_freelist`). - `g_iob_sem` is always incremented when an IOB buffer is freed, but `g_throttle_sem` is incremented only if we have at least CONFIG_IOB_THROTTLE buffers free. - Both semaphores are posted with the schedule locked to avoid any mismatches in the semaphores count. - If a task is waiting for an IOB semaphore (`iob_allocwait`) is awakened and would check the `g_iob_committed`. The highest priority task waiting for a semaphore will be awakened first.
240 lines
7.1 KiB
C
240 lines
7.1 KiB
C
/****************************************************************************
|
|
* mm/iob/iob_free.c
|
|
*
|
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
|
* contributor license agreements. See the NOTICE file distributed with
|
|
* this work for additional information regarding copyright ownership. The
|
|
* ASF licenses this file to you under the Apache License, Version 2.0 (the
|
|
* "License"); you may not use this file except in compliance with the
|
|
* License. You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
* License for the specific language governing permissions and limitations
|
|
* under the License.
|
|
*
|
|
****************************************************************************/
|
|
|
|
/****************************************************************************
|
|
* Included Files
|
|
****************************************************************************/
|
|
|
|
#include <nuttx/config.h>
|
|
|
|
#include <stdbool.h>
|
|
#include <assert.h>
|
|
#include <debug.h>
|
|
|
|
#include <nuttx/irq.h>
|
|
#include <nuttx/arch.h>
|
|
#ifdef CONFIG_IOB_ALLOC
|
|
# include <nuttx/kmalloc.h>
|
|
#endif
|
|
#include <nuttx/mm/iob.h>
|
|
|
|
#include "iob.h"
|
|
|
|
/****************************************************************************
|
|
* Pre-processor Definitions
|
|
****************************************************************************/
|
|
|
|
#ifdef CONFIG_IOB_NOTIFIER
|
|
# if !defined(CONFIG_IOB_NOTIFIER_DIV) || CONFIG_IOB_NOTIFIER_DIV < 2
|
|
# define IOB_DIVIDER 1
|
|
# elif CONFIG_IOB_NOTIFIER_DIV < 4
|
|
# define IOB_DIVIDER 2
|
|
# elif CONFIG_IOB_NOTIFIER_DIV < 8
|
|
# define IOB_DIVIDER 4
|
|
# elif CONFIG_IOB_NOTIFIER_DIV < 16
|
|
# define IOB_DIVIDER 8
|
|
# elif CONFIG_IOB_NOTIFIER_DIV < 32
|
|
# define IOB_DIVIDER 16
|
|
# elif CONFIG_IOB_NOTIFIER_DIV < 64
|
|
# define IOB_DIVIDER 32
|
|
# else
|
|
# define IOB_DIVIDER 64
|
|
# endif
|
|
#endif
|
|
|
|
#define IOB_MASK (IOB_DIVIDER - 1)
|
|
|
|
/****************************************************************************
|
|
* Public Functions
|
|
****************************************************************************/
|
|
|
|
/****************************************************************************
|
|
* Name: iob_free
|
|
*
|
|
* Description:
|
|
* Free the I/O buffer at the head of a buffer chain returning it to the
|
|
* free list. The link to the next I/O buffer in the chain is return.
|
|
*
|
|
****************************************************************************/
|
|
|
|
FAR struct iob_s *iob_free(FAR struct iob_s *iob)
|
|
{
|
|
FAR struct iob_s *next = iob->io_flink;
|
|
irqstate_t flags;
|
|
#ifdef CONFIG_IOB_NOTIFIER
|
|
int16_t navail;
|
|
#endif
|
|
#if CONFIG_IOB_THROTTLE > 0
|
|
bool committed_thottled = false;
|
|
#endif
|
|
|
|
iobinfo("iob=%p io_pktlen=%u io_len=%u next=%p\n",
|
|
iob, iob->io_pktlen, iob->io_len, next);
|
|
|
|
/* Copy the data that only exists in the head of a I/O buffer chain into
|
|
* the next entry.
|
|
*/
|
|
|
|
if (next != NULL)
|
|
{
|
|
/* Copy and decrement the total packet length, being careful to
|
|
* do nothing too crazy.
|
|
*/
|
|
|
|
if (iob->io_pktlen > iob->io_len)
|
|
{
|
|
/* Adjust packet length and move it to the next entry */
|
|
|
|
next->io_pktlen = iob->io_pktlen - iob->io_len;
|
|
DEBUGASSERT(next->io_pktlen >= next->io_len);
|
|
}
|
|
else
|
|
{
|
|
/* This can only happen if the free entry isn't first entry in the
|
|
* chain...
|
|
*/
|
|
|
|
next->io_pktlen = 0;
|
|
}
|
|
|
|
iobinfo("next=%p io_pktlen=%u io_len=%u\n",
|
|
next, next->io_pktlen, next->io_len);
|
|
}
|
|
|
|
#ifdef CONFIG_IOB_ALLOC
|
|
if (iob->io_free != NULL)
|
|
{
|
|
iob->io_free(iob->io_data);
|
|
kmm_free(iob);
|
|
return next;
|
|
}
|
|
#endif
|
|
|
|
/* Free the I/O buffer by adding it to the head of the free or the
|
|
* committed list. We don't know what context we are called from so
|
|
* we use extreme measures to protect the free list: We disable
|
|
* interrupts very briefly.
|
|
*/
|
|
|
|
flags = spin_lock_irqsave(&g_iob_lock);
|
|
|
|
/* Which list? If there is a task waiting for an IOB, then put
|
|
* the IOB on either the free list or on the committed list where
|
|
* it is reserved for that allocation (and not available to
|
|
* iob_tryalloc()). This is true for both throttled and non-throttled
|
|
* cases.
|
|
*/
|
|
|
|
#if CONFIG_IOB_THROTTLE > 0
|
|
if ((g_iob_sem.semcount < 0) ||
|
|
((g_iob_sem.semcount >= CONFIG_IOB_THROTTLE) &&
|
|
(g_throttle_sem.semcount < 0)))
|
|
#else
|
|
if (g_iob_sem.semcount < 0)
|
|
#endif
|
|
{
|
|
iob->io_flink = g_iob_committed;
|
|
g_iob_committed = iob;
|
|
#if CONFIG_IOB_THROTTLE > 0
|
|
if ((g_iob_sem.semcount >= CONFIG_IOB_THROTTLE) &&
|
|
(g_throttle_sem.semcount < 0))
|
|
{
|
|
committed_thottled = true;
|
|
}
|
|
#endif
|
|
}
|
|
else
|
|
{
|
|
iob->io_flink = g_iob_freelist;
|
|
g_iob_freelist = iob;
|
|
}
|
|
|
|
spin_unlock_irqrestore(&g_iob_lock, flags);
|
|
|
|
/* Signal that an IOB is available. This is done with schedule locked
|
|
* to make sure that both g_iob_sem and g_throttle_sem are incremented
|
|
* together (if applicable). After the schedule is unlocked, if there
|
|
* is a thread blocked, waiting for an IOB, this will wake up exactly
|
|
* one thread. The semaphore count will correctly indicate that the
|
|
* awakened task owns an IOB and should find it in the committed list.
|
|
*/
|
|
|
|
sched_lock();
|
|
|
|
nxsem_post(&g_iob_sem);
|
|
DEBUGASSERT(g_iob_sem.semcount <= CONFIG_IOB_NBUFFERS);
|
|
|
|
#if CONFIG_IOB_THROTTLE > 0
|
|
flags = spin_lock_irqsave(&g_iob_lock);
|
|
|
|
if (g_iob_sem.semcount > CONFIG_IOB_THROTTLE)
|
|
{
|
|
/* If posting to the the throttled semaphore is going to awake a
|
|
* waiting task, then the g_iob_sem count should be decremented
|
|
* because an I/O buffer (from the head of the g_iob_committed list)
|
|
* will be allocated to this waiting task.
|
|
* Decrementing the g_throttled_sem (when posting to the g_iob_sem)
|
|
* is not necessary because this condition can only occur when the
|
|
* g_throttled_sem is less or equal to zero. On the other hand, if
|
|
* the g_iob_sem is greater than the CONFIG_IOB_THROTTLE and there
|
|
* is a waiting thread, then the I/O buffer just freed will be
|
|
* committed to a waiting task and is not available for general use.
|
|
*/
|
|
|
|
if (committed_thottled)
|
|
{
|
|
g_iob_sem.semcount--;
|
|
}
|
|
|
|
spin_unlock_irqrestore(&g_iob_lock, flags);
|
|
|
|
nxsem_post(&g_throttle_sem);
|
|
DEBUGASSERT(g_throttle_sem.semcount <=
|
|
(CONFIG_IOB_NBUFFERS - CONFIG_IOB_THROTTLE));
|
|
}
|
|
else
|
|
{
|
|
spin_unlock_irqrestore(&g_iob_lock, flags);
|
|
}
|
|
#endif
|
|
|
|
sched_unlock();
|
|
|
|
#ifdef CONFIG_IOB_NOTIFIER
|
|
/* Check if the IOB was claimed by a thread that is blocked waiting
|
|
* for an IOB.
|
|
*/
|
|
|
|
navail = iob_navail(false);
|
|
if (navail > 0 && (navail & IOB_MASK) == 0)
|
|
{
|
|
/* Signal any threads that have requested a signal notification
|
|
* when an IOB becomes available.
|
|
*/
|
|
|
|
iob_notifier_signal();
|
|
}
|
|
#endif
|
|
|
|
/* And return the I/O buffer after the one that was freed */
|
|
|
|
return next;
|
|
}
|