nuttx/mm/map/mm_map.c
Alin Jerpelea 339457dda3 mm: migrate to SPDX identifier
Most tools used for compliance and SBOM generation use SPDX identifiers
This change brings us a step closer to an easy SBOM generation.

Signed-off-by: Alin Jerpelea <alin.jerpelea@sony.com>
2024-09-13 08:49:05 +08:00

382 lines
9.9 KiB
C

/****************************************************************************
* mm/map/mm_map.c
*
* SPDX-License-Identifier: Apache-2.0
*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership. The
* ASF licenses this file to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance with the
* License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
* License for the specific language governing permissions and limitations
* under the License.
*
****************************************************************************/
/****************************************************************************
* Included Files
****************************************************************************/
#include <nuttx/mm/map.h>
#include <nuttx/pgalloc.h>
#include <nuttx/addrenv.h>
#include <stdbool.h>
#include <stddef.h>
#include <nuttx/sched.h>
#include <nuttx/kmalloc.h>
#include <assert.h>
#include <debug.h>
#include "sched/sched.h"
#if defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__)
/****************************************************************************
* Private Functions
****************************************************************************/
static bool in_range(FAR const void *start, size_t length,
FAR const void *range_start, size_t range_length)
{
FAR const char *u_start = (FAR const char *)start;
FAR const char *u_end = u_start + length;
FAR const char *r_start = (FAR const char *)range_start;
FAR const char *r_end = r_start + range_length;
return (u_start >= r_start && u_start < r_end && /* Start is in range. */
u_end >= r_start && u_end <= r_end); /* End is in range. */
}
/****************************************************************************
* Public Functions
****************************************************************************/
/****************************************************************************
* Name: mm_map_lock
*
* Description:
* Get exclusive access to task_group's mm_map
*
****************************************************************************/
int mm_map_lock(void)
{
FAR struct tcb_s *tcb = this_task();
FAR struct task_group_s *group = tcb->group;
if (group == NULL)
{
return -EINVAL;
}
return nxrmutex_lock(&group->tg_mm_map.mm_map_mutex);
}
/****************************************************************************
* Name: mm_map_unlock
*
* Description:
* Relinquish exclusive access to task_group's mm_map
*
****************************************************************************/
void mm_map_unlock(void)
{
FAR struct tcb_s *tcb = this_task();
FAR struct task_group_s *group = tcb->group;
if (group == NULL)
{
return;
}
DEBUGVERIFY(nxrmutex_unlock(&group->tg_mm_map.mm_map_mutex));
}
/****************************************************************************
* Name: mm_map_initialize
*
* Description:
* Allocates a task group specific mm_map structure. Called when the group
* is initialized
*
****************************************************************************/
void mm_map_initialize(FAR struct mm_map_s *mm, bool kernel)
{
sq_init(&mm->mm_map_sq);
nxrmutex_init(&mm->mm_map_mutex);
mm->map_count = 0;
/* Create the virtual pages allocator for user process */
#ifdef CONFIG_ARCH_VMA_MAPPING
if (!kernel)
{
mm->mm_map_vpages = gran_initialize((FAR void *)CONFIG_ARCH_SHM_VBASE,
ARCH_SHM_SIZE, MM_PGSHIFT, MM_PGSHIFT);
if (!mm->mm_map_vpages)
{
merr("gran_initialize() failed\n");
}
}
else
{
mm->mm_map_vpages = NULL;
}
#endif
}
/****************************************************************************
* Name: mm_map_destroy
*
* Description:
* De-allocates a task group specific mm_map structure and the mm_map_mutex
*
****************************************************************************/
void mm_map_destroy(FAR struct mm_map_s *mm)
{
FAR struct mm_map_entry_s *entry;
while ((entry = (FAR struct mm_map_entry_s *)sq_remfirst(&mm->mm_map_sq)))
{
/* Pass null as group argument to indicate that actual MMU mappings
* must not be touched. The process is being deleted and we don't
* know in which context we are. Only kernel memory allocations
* need to be freed by drivers
*/
/* Unmap the whole region */
if (entry->munmap)
{
if (entry->munmap(NULL, entry, entry->vaddr, entry->length) < 0)
{
/* This would be an error in the driver. It has defined munmap,
* but is not able to munmap the full area which it has mapped
*/
merr("Driver munmap failed\n");
}
}
mm->map_count--;
kmm_free(entry);
}
DEBUGASSERT(mm->map_count == 0);
nxrmutex_destroy(&mm->mm_map_mutex);
/* Release the virtual pages allocator */
#ifdef CONFIG_ARCH_VMA_MAPPING
if (mm->mm_map_vpages)
{
gran_release(mm->mm_map_vpages);
}
#endif
}
/****************************************************************************
* Name: mm_map_add
*
* Description:
* Add a mapping to task group's mm_map list
*
****************************************************************************/
int mm_map_add(FAR struct mm_map_s *mm, FAR struct mm_map_entry_s *entry)
{
FAR struct mm_map_entry_s *new_entry;
int ret;
if (!entry)
{
return -EINVAL;
}
/* Copy the provided mapping and add to the list */
new_entry = kmm_malloc(sizeof(struct mm_map_entry_s));
if (!new_entry)
{
return -EINVAL;
}
*new_entry = *entry;
ret = nxrmutex_lock(&mm->mm_map_mutex);
if (ret < 0)
{
kmm_free(new_entry);
return ret;
}
/* Too many mappings? */
if (mm->map_count >= CONFIG_MM_MAP_COUNT_MAX)
{
kmm_free(new_entry);
nxrmutex_unlock(&mm->mm_map_mutex);
return -ENOMEM;
}
mm->map_count++;
sq_addfirst((sq_entry_t *)new_entry, &mm->mm_map_sq);
nxrmutex_unlock(&mm->mm_map_mutex);
return OK;
}
/****************************************************************************
* Name: mm_map_next
*
* Description:
* Returns the next mapping in the list.
*
****************************************************************************/
FAR struct mm_map_entry_s *mm_map_next(FAR struct mm_map_s *mm,
FAR const struct mm_map_entry_s *entry)
{
FAR struct mm_map_entry_s *next_entry = NULL;
if (nxrmutex_lock(&mm->mm_map_mutex) == OK)
{
if (entry == NULL)
{
next_entry = (struct mm_map_entry_s *)sq_peek(&mm->mm_map_sq);
}
else
{
next_entry = (struct mm_map_entry_s *)
sq_next(((sq_entry_t *)entry));
}
nxrmutex_unlock(&mm->mm_map_mutex);
}
return next_entry;
}
/****************************************************************************
* Name: mm_map_find
*
* Description:
* Find the first mapping containing the range from the task group's list
*
****************************************************************************/
FAR struct mm_map_entry_s *mm_map_find(FAR struct mm_map_s *mm,
FAR const void *vaddr,
size_t length)
{
FAR struct mm_map_entry_s *found_entry = NULL;
if (nxrmutex_lock(&mm->mm_map_mutex) == OK)
{
found_entry = (struct mm_map_entry_s *)sq_peek(&mm->mm_map_sq);
while (found_entry && !in_range(vaddr, length,
found_entry->vaddr,
found_entry->length))
{
found_entry = (struct mm_map_entry_s *)
sq_next(((sq_entry_t *)found_entry));
}
nxrmutex_unlock(&mm->mm_map_mutex);
}
return found_entry;
}
/****************************************************************************
* Name: mm_map_remove
*
* Description:
* Remove a mapping from the task group's list
*
****************************************************************************/
int mm_map_remove(FAR struct mm_map_s *mm,
FAR struct mm_map_entry_s *entry)
{
FAR struct mm_map_entry_s *prev_entry;
FAR struct mm_map_entry_s *removed_entry = NULL;
int ret;
if (!mm || !entry)
{
return OK;
}
ret = nxrmutex_lock(&mm->mm_map_mutex);
if (ret < 0)
{
return ret;
}
prev_entry = (struct mm_map_entry_s *)sq_peek(&mm->mm_map_sq);
/* Check if the list was empty */
if (!prev_entry)
{
nxrmutex_unlock(&mm->mm_map_mutex);
return -ENOENT;
}
/* Check if removing the first item */
if (entry == prev_entry)
{
sq_remfirst(&mm->mm_map_sq);
mm->map_count--;
removed_entry = prev_entry;
}
else
{
/* Loop through the remaining items to find the one to be removed */
while ((removed_entry = (struct mm_map_entry_s *)
sq_next(((sq_entry_t *)prev_entry))))
{
if (entry == removed_entry)
{
sq_remafter((sq_entry_t *)prev_entry, &mm->mm_map_sq);
mm->map_count--;
break;
}
prev_entry = removed_entry;
}
}
nxrmutex_unlock(&mm->mm_map_mutex);
/* If the item was removed, also delete the entry struct */
if (removed_entry)
{
kmm_free(removed_entry);
return OK;
}
return -ENOENT;
}
#endif /* defined(CONFIG_BUILD_FLAT) || defined(__KERNEL__) */