diff --git a/README.md b/README.md index d5e25782..d767052c 100644 --- a/README.md +++ b/README.md @@ -1,2 +1,8 @@ +<<<<<<< HEAD # Iot_Cs_best +======= +# Brief Intro + +LMFAO, we do this shit just for fun :) +>>>>>>> remotes/origin/main diff --git a/doc/cache/UML用例图.vsdx b/doc/cache/UML用例图.vsdx new file mode 100644 index 00000000..73024512 Binary files /dev/null and b/doc/cache/UML用例图.vsdx differ diff --git a/doc/cache/image-1.png b/doc/cache/image-1.png new file mode 100644 index 00000000..18765b3d Binary files /dev/null and b/doc/cache/image-1.png differ diff --git a/doc/cache/image-2.png b/doc/cache/image-2.png new file mode 100644 index 00000000..18765b3d Binary files /dev/null and b/doc/cache/image-2.png differ diff --git a/doc/cache/image-3.png b/doc/cache/image-3.png new file mode 100644 index 00000000..fc5179eb Binary files /dev/null and b/doc/cache/image-3.png differ diff --git a/doc/cache/image-4.png b/doc/cache/image-4.png new file mode 100644 index 00000000..fc5179eb Binary files /dev/null and b/doc/cache/image-4.png differ diff --git a/doc/cache/image-5.png b/doc/cache/image-5.png new file mode 100644 index 00000000..7bff255f Binary files /dev/null and b/doc/cache/image-5.png differ diff --git a/doc/cache/image-6.png b/doc/cache/image-6.png new file mode 100644 index 00000000..7bff255f Binary files /dev/null and b/doc/cache/image-6.png differ diff --git a/doc/cache/image.png b/doc/cache/image.png new file mode 100644 index 00000000..04b5dd2e Binary files /dev/null and b/doc/cache/image.png differ diff --git a/doc/cache/~$$包图.~vsdx b/doc/cache/~$$包图.~vsdx new file mode 100644 index 00000000..b139c4fc Binary files /dev/null and b/doc/cache/~$$包图.~vsdx differ diff --git a/doc/cache/~$$进程控制顺序图.~vsdx b/doc/cache/~$$进程控制顺序图.~vsdx new file mode 100644 index 00000000..f3ad3b1e Binary files /dev/null and b/doc/cache/~$$进程控制顺序图.~vsdx differ diff --git a/doc/cache/包图.vsdx b/doc/cache/包图.vsdx new file mode 100644 index 00000000..01b1908b Binary files /dev/null and b/doc/cache/包图.vsdx differ diff --git a/doc/cache/学习报告.md b/doc/cache/学习报告.md new file mode 100644 index 00000000..533c3037 --- /dev/null +++ b/doc/cache/学习报告.md @@ -0,0 +1,44 @@ +# base 下的代码阅读 + +根据官方的README.md,kernel/base下的文件主要为**基础内核包括调度、内存等模块** + +# 1. brief intro + +OpenHarmony liteos 的进程实现方式与 linux 类似,都是通过父进程和子进程的一系列操作来达到实现一些功能的目的. + +## 1.1 描述各个子功能 + +* los_bitmap.c 主要提供一些位操作 +* los_info.c 提供一些基本的方法来对进程的信息等内容进行操作 +* los_process.c 提供对于进程操作的方法包括创建切换等.. + + +# 2. natural expression of requirement of software + +作为操作系统当中的子系统,以软件的角度来看进程管理: + +## 2.1 一个完备的进程管理系统应当具备 + +### (1) 进程调度 + +鸿蒙内核的进程采用抢占式调度机制,支持时间片轮转调度方式和FIFO调度机制 + +### (2) 进程创建 + +![Alt text](image-4.png) + + +### (3) 进程回收 +### (4) +### (5) +### (6) + +# 3. user case diagram and user case description + +![Alt text](image.png) + + +# 4. software architecture + + +![Alt text](image-5.png) \ No newline at end of file diff --git a/doc/cache/泛读报告.md b/doc/cache/泛读报告.md new file mode 100644 index 00000000..769db50e --- /dev/null +++ b/doc/cache/泛读报告.md @@ -0,0 +1,42 @@ + + +# LiteOS_kernel_a 泛读报告 + +## 一、 软件功能的自然语言描述 + +### 1.1 简介 + +Huawei LiteOS是华为面向IoT领域,构建的轻量级物联网操作系统,可广泛应用于智能家居、个人穿戴、车联网、城市公共服务、制造业等领域。 + +Huawei LiteOS发布于2015年5月的华为网络大会上。自开源社区发布以来,围绕 NB-IoT 物联网市场从技术、生态、解决方案、商用支持等多维度使能合作伙伴,构建开源的物联网生态。目前已经聚合了 50+ MCU 和解决方案合作伙伴,共同推出一批开源开发套件和行业解决方案,帮助众多行业客户快速的推出物联网产品和服务。客户涵盖抄表、停车、路灯、环保、共享单车、物流等众多行业,为开发者提供 “一站式” 完整软件平台,可大幅降低设备布置及维护成本,有效降低开发门槛、缩短开发周期。 + +Huawei LiteOS开源项目目前支持 **ARM64、ARM Cortex-A、ARM Cortex-M0,Cortex-M3,Cortex-M4,Cortex-M7** 等芯片架构。 + +### 1.2 软件功能 + +#### 1.2.1 提供一个简单轻便的用户交互界面以供使用。 + +#### 1.2.2 能够调用系统软硬件资源来完成某些操作。 + +#### 1.2.3 进程管理 + +#### 1.2.4 内存分配 + +#### 1.2.5 提供并行编程 + +## 二、 软件的用例图和用例描述 + +![Alt text](image.png) + +## 三、 软件的体系结构 + +![Alt text](image-2.png) + +![Alt text](image-6.png) \ No newline at end of file diff --git a/doc/cache/绘图1.vsdx b/doc/cache/绘图1.vsdx new file mode 100644 index 00000000..5cec30dc Binary files /dev/null and b/doc/cache/绘图1.vsdx differ diff --git a/doc/cache/进程控制顺序图.vsdx b/doc/cache/进程控制顺序图.vsdx new file mode 100644 index 00000000..9d0a1f31 Binary files /dev/null and b/doc/cache/进程控制顺序图.vsdx differ diff --git a/doc/泛读报告.docx b/doc/泛读报告.docx new file mode 100644 index 00000000..42d226e8 Binary files /dev/null and b/doc/泛读报告.docx differ diff --git a/src/kernel_liteos_a/.gitignore b/src/kernel_liteos_a/.gitignore new file mode 100644 index 00000000..c796d20f --- /dev/null +++ b/src/kernel_liteos_a/.gitignore @@ -0,0 +1,24 @@ +# General ignored file types +*.o +*.a +*.so +*.swp + +# IDE settings +.vscode +.idea +.settings +.cproject +.project + +# VIM files +cscope* +tags + +# Menuconfig temp files +/config.h +/.config +/.config.old + +# Build temp files +/out diff --git a/src/kernel_liteos_a/apps/init/src/init.c b/src/kernel_liteos_a/apps/init/src/init.c index a0420394..20953ed2 100644 --- a/src/kernel_liteos_a/apps/init/src/init.c +++ b/src/kernel_liteos_a/apps/init/src/init.c @@ -54,13 +54,21 @@ int main(int argc, char * const *argv) const char *shellPath = "/bin/mksh"; #ifdef LOSCFG_QUICK_START +<<<<<<< HEAD const char *samplePath = "/dev/shm/sample_quickstart"; +======= + const char *samplePath = "/dev/shm/sample_quickstart"; +>>>>>>> remotes/origin/main ret = fork(); if (ret < 0) { printf("Failed to fork for sample_quickstart\n"); } else if (ret == 0) { +<<<<<<< HEAD (void)execve(samplePath, NULL, NULL); +======= + (void)execve(samplePath, NULL, NULL); // 执行可执行文件 +>>>>>>> remotes/origin/main exit(0); } @@ -72,24 +80,43 @@ int main(int argc, char * const *argv) close(fd); } #endif +<<<<<<< HEAD ret = fork(); if (ret < 0) { printf("Failed to fork for shell\n"); } else if (ret == 0) { gid = getpgrp(); +======= + ret = fork(); // 创建第一个程序来跑shell + if (ret < 0) { + printf("Failed to fork for shell\n"); + } else if (ret == 0) { + gid = getpgrp(); // 返回进程组ID +>>>>>>> remotes/origin/main if (gid < 0) { printf("get group id failed, pgrpid %d, errno %d\n", gid, errno); exit(0); } +<<<<<<< HEAD ret = tcsetpgrp(STDIN_FILENO, gid); +======= + ret = tcsetpgrp(STDIN_FILENO, gid); +>>>>>>> remotes/origin/main if (ret != 0) { printf("tcsetpgrp failed, errno %d\n", errno); exit(0); } +<<<<<<< HEAD (void)execve(shellPath, NULL, NULL); exit(0); } +======= + (void)execve(shellPath, NULL, NULL); // 正常执行命令行程序可执行文件 + exit(0); + } + // ret > 0 +>>>>>>> remotes/origin/main while (1) { ret = waitpid(-1, 0, WNOHANG); if (ret == 0) { diff --git a/src/kernel_liteos_a/fs/jffs2/jffs2.patch b/src/kernel_liteos_a/fs/jffs2/jffs2.patch index 073aa696..7b8ab9a4 100644 --- a/src/kernel_liteos_a/fs/jffs2/jffs2.patch +++ b/src/kernel_liteos_a/fs/jffs2/jffs2.patch @@ -2311,7 +2311,11 @@ diff -Nupr old/fs/jffs2/erase.c new/fs/jffs2/erase.c diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c --- old/fs/jffs2/file.c 2022-05-09 17:22:53.000000000 +0800 +++ new/fs/jffs2/file.c 2022-05-10 09:43:14.250000000 +0800 +<<<<<<< HEAD @@ -9,335 +9,30 @@ +======= +@@ -9,334 +9,31 @@ +>>>>>>> remotes/origin/main * For licensing information, see the file 'LICENCE' in this directory. * */ @@ -2348,7 +2352,11 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c - struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb); + /* FIXME: This works only with one file system mounted at a time */ int ret; +<<<<<<< HEAD - +======= + +>>>>>>> remotes/origin/main - ret = file_write_and_wait_range(filp, start, end); + ret = jffs2_read_inode_range(c, f, gc_buffer, + offset & ~(PAGE_SIZE-1), PAGE_SIZE); @@ -2361,8 +2369,15 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c - inode_unlock(inode); - - return 0; +<<<<<<< HEAD -} - +======= ++ return ERR_PTR(ret); ++ return gc_buffer; + } + +>>>>>>> remotes/origin/main -const struct file_operations jffs2_file_operations = -{ - .llseek = generic_file_llseek, @@ -2387,7 +2402,14 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c -}; - -const struct address_space_operations jffs2_file_address_operations = +<<<<<<< HEAD -{ +======= ++void jffs2_gc_release_page(struct jffs2_sb_info *c, ++ unsigned char *ptr, ++ unsigned long *priv) + { +>>>>>>> remotes/origin/main - .readpage = jffs2_readpage, - .write_begin = jffs2_write_begin, - .write_end = jffs2_write_end, @@ -2443,6 +2465,7 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c - ret = jffs2_do_readpage_unlock(pg->mapping->host, pg); - mutex_unlock(&f->sem); - return ret; +<<<<<<< HEAD + return ERR_PTR(ret); + return gc_buffer; } @@ -2454,24 +2477,45 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c + unsigned char *ptr, + unsigned long *priv) { +======= +-} +- +-static int jffs2_write_begin(struct file *filp, struct address_space *mapping, +- loff_t pos, unsigned len, unsigned flags, +- struct page **pagep, void **fsdata) +-{ +>>>>>>> remotes/origin/main - struct page *pg; - struct inode *inode = mapping->host; - struct jffs2_inode_info *f = JFFS2_INODE_INFO(inode); - struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb); - pgoff_t index = pos >> PAGE_SHIFT; +<<<<<<< HEAD - uint32_t pageofs = index << PAGE_SHIFT; +======= +>>>>>>> remotes/origin/main - int ret = 0; - - jffs2_dbg(1, "%s()\n", __func__); - +<<<<<<< HEAD - if (pageofs > inode->i_size) { - /* Make new hole frag from old EOF to new page */ +======= +- if (pos > inode->i_size) { +- /* Make new hole frag from old EOF to new position */ +>>>>>>> remotes/origin/main - struct jffs2_raw_inode ri; - struct jffs2_full_dnode *fn; - uint32_t alloc_len; - +<<<<<<< HEAD - jffs2_dbg(1, "Writing new hole frag 0x%x-0x%x between current EOF and new page\n", - (unsigned int)inode->i_size, pageofs); +======= +- jffs2_dbg(1, "Writing new hole frag 0x%x-0x%x between current EOF and new position\n", +- (unsigned int)inode->i_size, (uint32_t)pos); +>>>>>>> remotes/origin/main - - ret = jffs2_reserve_space(c, sizeof(ri), &alloc_len, - ALLOC_NORMAL, JFFS2_SUMMARY_INODE_SIZE); @@ -2491,10 +2535,17 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c - ri.mode = cpu_to_jemode(inode->i_mode); - ri.uid = cpu_to_je16(i_uid_read(inode)); - ri.gid = cpu_to_je16(i_gid_read(inode)); +<<<<<<< HEAD - ri.isize = cpu_to_je32(max((uint32_t)inode->i_size, pageofs)); - ri.atime = ri.ctime = ri.mtime = cpu_to_je32(JFFS2_NOW()); - ri.offset = cpu_to_je32(inode->i_size); - ri.dsize = cpu_to_je32(pageofs - inode->i_size); +======= +- ri.isize = cpu_to_je32((uint32_t)pos); +- ri.atime = ri.ctime = ri.mtime = cpu_to_je32(JFFS2_NOW()); +- ri.offset = cpu_to_je32(inode->i_size); +- ri.dsize = cpu_to_je32((uint32_t)pos - inode->i_size); +>>>>>>> remotes/origin/main - ri.csize = cpu_to_je32(0); - ri.compr = JFFS2_COMPR_ZERO; - ri.node_crc = cpu_to_je32(crc32(0, &ri, sizeof(ri)-8)); @@ -2524,7 +2575,11 @@ diff -Nupr old/fs/jffs2/file.c new/fs/jffs2/file.c - goto out_err; - } - jffs2_complete_reservation(c); +<<<<<<< HEAD - inode->i_size = pageofs; +======= +- inode->i_size = pos; +>>>>>>> remotes/origin/main - mutex_unlock(&f->sem); - } - diff --git a/src/kernel_liteos_a/kernel/base/core/los_bitmap.c b/src/kernel_liteos_a/kernel/base/core/los_bitmap.c index 4e8b8600..63c7bbd3 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_bitmap.c +++ b/src/kernel_liteos_a/kernel/base/core/los_bitmap.c @@ -33,6 +33,7 @@ #include "los_printf.h" #include "los_toolchain.h" +<<<<<<< HEAD #define OS_BITMAP_MASK 0x1FU #define OS_BITMAP_WORD_MASK ~0UL @@ -43,15 +44,54 @@ STATIC INLINE UINT16 Ffz(UINTPTR x) return __builtin_ffsl(~x) - 1; } +======= +/* + function: 这部分的代码提供了位操作 + + interpretation: 位操作为可以提供对标志位的修改等功能 +*/ +#define OS_BITMAP_MASK 0x1FU +#define OS_BITMAP_WORD_MASK ~0UL + +/* + function: 返回从右开始第一个为1的下标 + + interpretation: +*/ + +/* find first zero bit starting from LSB */ +STATIC INLINE UINT16 Ffz(UINTPTR x) +{ + // gcc 自带的内建函数 + return __builtin_ffsl(~x) - 1; +} + +/* + funciton: 对状态字某一标志位进行置1操作 + + interpretation: +*/ +>>>>>>> remotes/origin/main VOID LOS_BitmapSet(UINT32 *bitmap, UINT16 pos) { if (bitmap == NULL) { return; } +<<<<<<< HEAD *bitmap |= 1U << (pos & OS_BITMAP_MASK); } +======= + *bitmap |= 1U << (pos & OS_BITMAP_MASK); //00011111 +} + +/* + function: 对某一标志位进行清0操作 + + interpretation: +*/ +>>>>>>> remotes/origin/main VOID LOS_BitmapClr(UINT32 *bitmap, UINT16 pos) { if (bitmap == NULL) { @@ -61,6 +101,12 @@ VOID LOS_BitmapClr(UINT32 *bitmap, UINT16 pos) *bitmap &= ~(1U << (pos & OS_BITMAP_MASK)); } +<<<<<<< HEAD +======= +/* + function: 用于返回参数位图当中最高的索引位 +*/ +>>>>>>> remotes/origin/main UINT16 LOS_HighBitGet(UINT32 bitmap) { if (bitmap == 0) { @@ -70,6 +116,12 @@ UINT16 LOS_HighBitGet(UINT32 bitmap) return (OS_BITMAP_MASK - CLZ(bitmap)); } +<<<<<<< HEAD +======= +/* + function: 获取参数位图中最低位为1的索引位 +*/ +>>>>>>> remotes/origin/main UINT16 LOS_LowBitGet(UINT32 bitmap) { if (bitmap == 0) { @@ -79,6 +131,12 @@ UINT16 LOS_LowBitGet(UINT32 bitmap) return CTZ(bitmap); } +<<<<<<< HEAD +======= +/* + function: 从start位置开始设置numsSet个bit位 置1 +*/ +>>>>>>> remotes/origin/main VOID LOS_BitmapSetNBits(UINTPTR *bitmap, UINT32 start, UINT32 numsSet) { UINTPTR *p = bitmap + BITMAP_WORD(start); @@ -99,6 +157,12 @@ VOID LOS_BitmapSetNBits(UINTPTR *bitmap, UINT32 start, UINT32 numsSet) } } +<<<<<<< HEAD +======= +/* + fuction: 从start位置开始 清除numsSet个bit位置0 ,对状态字的连续标志位进行清0操作 +*/ +>>>>>>> remotes/origin/main VOID LOS_BitmapClrNBits(UINTPTR *bitmap, UINT32 start, UINT32 numsClear) { UINTPTR *p = bitmap + BITMAP_WORD(start); @@ -118,7 +182,13 @@ VOID LOS_BitmapClrNBits(UINTPTR *bitmap, UINT32 start, UINT32 numsClear) *p &= ~maskToClear; } } +<<<<<<< HEAD +======= +/* + fuction: 从numBits位置开始找到第一个0位 +*/ +>>>>>>> remotes/origin/main INT32 LOS_BitmapFfz(UINTPTR *bitmap, UINT32 numBits) { INT32 bit, i; diff --git a/src/kernel_liteos_a/kernel/base/core/los_info.c b/src/kernel_liteos_a/kernel/base/core/los_info.c index f03bac81..cd8f70f6 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_info.c +++ b/src/kernel_liteos_a/kernel/base/core/los_info.c @@ -32,8 +32,15 @@ #include "los_task_pri.h" #include "los_vm_dump.h" +<<<<<<< HEAD STATIC UINT32 GetCurrParentPid(UINT32 pid, const LosProcessCB *processCB) { +======= +// 得到父进程的pid +STATIC UINT32 GetCurrParentPid(UINT32 pid, const LosProcessCB *processCB) +{ + // 如果没有父进程 +>>>>>>> remotes/origin/main if (processCB->parentProcess == NULL) { return 0; } @@ -50,6 +57,10 @@ STATIC UINT32 GetCurrParentPid(UINT32 pid, const LosProcessCB *processCB) return processCB->parentProcess->processID; } +<<<<<<< HEAD +======= +// 得到当前任务ID +>>>>>>> remotes/origin/main STATIC INLINE UINT32 GetCurrTid(const LosTaskCB *taskCB) { #ifdef LOSCFG_PID_CONTAINER @@ -60,6 +71,10 @@ STATIC INLINE UINT32 GetCurrTid(const LosTaskCB *taskCB) return taskCB->taskID; } +<<<<<<< HEAD +======= +// 得到进程的状态 +>>>>>>> remotes/origin/main STATIC UINT16 GetProcessStatus(LosProcessCB *processCB) { UINT16 status; @@ -76,6 +91,10 @@ STATIC UINT16 GetProcessStatus(LosProcessCB *processCB) return status; } +<<<<<<< HEAD +======= +// 得到进程的信息 +>>>>>>> remotes/origin/main STATIC VOID GetProcessInfo(ProcessInfo *pcbInfo, const LosProcessCB *processCB) { SchedParam param = {0}; @@ -133,6 +152,10 @@ STATIC VOID GetProcessMemInfo(ProcessInfo *pcbInfo, const LosProcessCB *processC } #endif +<<<<<<< HEAD +======= +// 得到线程的信息 +>>>>>>> remotes/origin/main STATIC VOID GetThreadInfo(ProcessThreadInfo *threadInfo, LosProcessCB *processCB) { SchedParam param = {0}; diff --git a/src/kernel_liteos_a/kernel/base/core/los_process.c b/src/kernel_liteos_a/kernel/base/core/los_process.c index 724ecf66..332b466c 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_process.c +++ b/src/kernel_liteos_a/kernel/base/core/los_process.c @@ -1,4 +1,16 @@ /* +<<<<<<< HEAD +======= +进程模块主文件 + 并发(Concurrent):多个线程在单个核心运行,同一时间只能一个线程运行,内核不停切换线程, + 看起来像同时运行,实际上是线程不停切换 + 并行(Parallel)每个线程分配给独立的CPU核心,线程同时运行 + 单核CPU多个进程或多个线程内能实现并发(微观上的串行,宏观上的并行) + 多核CPU线程间可以实现宏观和微观上的并行 + LITE_OS_SEC_BSS 和 LITE_OS_SEC_DATA_INIT 是告诉编译器这些全局变量放在哪个数据段 + */ +/* +>>>>>>> remotes/origin/main * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. * @@ -59,12 +71,21 @@ #include "los_vm_phys.h" #include "los_vm_syscall.h" +<<<<<<< HEAD LITE_OS_SEC_BSS LosProcessCB *g_processCBArray = NULL; LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_freeProcess; LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_processRecycleList; LITE_OS_SEC_BSS UINT32 g_processMaxNum; #ifndef LOSCFG_PID_CONTAINER LITE_OS_SEC_BSS ProcessGroup *g_processGroup = NULL; +======= +LITE_OS_SEC_BSS LosProcessCB *g_processCBArray = NULL;//进程池数组 +LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_freeProcess;//空闲状态下的进程链表, +LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_processRecycleList;//需要收回的进程列表 +LITE_OS_SEC_BSS UINT32 g_processMaxNum;//进程最大数量,默认64个 +#ifndef LOSCFG_PID_CONTAINER +LITE_OS_SEC_BSS ProcessGroup *g_processGroup = NULL;//全局进程组,负责管理所有的进程组 +>>>>>>> remotes/origin/main #define OS_ROOT_PGRP(processCB) (g_processGroup) #endif @@ -73,12 +94,21 @@ STATIC INLINE VOID OsInsertPCBToFreeList(LosProcessCB *processCB) #ifdef LOSCFG_PID_CONTAINER OsPidContainerDestroy(processCB->container, processCB); #endif +<<<<<<< HEAD UINT32 pid = processCB->processID; (VOID)memset_s(processCB, sizeof(LosProcessCB), 0, sizeof(LosProcessCB)); processCB->processID = pid; processCB->processStatus = OS_PROCESS_FLAG_UNUSED; processCB->timerID = (timer_t)(UINTPTR)MAX_INVALID_TIMER_VID; LOS_ListTailInsert(&g_freeProcess, &processCB->pendList); +======= + UINT32 pid = processCB->processID;//获取进程ID + (VOID)memset_s(processCB, sizeof(LosProcessCB), 0, sizeof(LosProcessCB));//进程描述符数据清0 + processCB->processID = pid;//进程ID + processCB->processStatus = OS_PROCESS_FLAG_UNUSED;//设置为进程未使用 + processCB->timerID = (timer_t)(UINTPTR)MAX_INVALID_TIMER_VID;//timeID初始化值 + LOS_ListTailInsert(&g_freeProcess, &processCB->pendList);//进程节点挂入g_freeProcess以分配给后续进程使用 +>>>>>>> remotes/origin/main } VOID OsDeleteTaskFromProcess(LosTaskCB *taskCB) @@ -132,14 +162,29 @@ UINT32 OsProcessAddNewTask(UINTPTR processID, LosTaskCB *taskCB, SchedParam *par SCHEDULER_UNLOCK(intSave); return LOS_OK; } +<<<<<<< HEAD ProcessGroup *OsCreateProcessGroup(LosProcessCB *processCB) { ProcessGroup *pgroup = LOS_MemAlloc(m_aucSysMem1, sizeof(ProcessGroup)); +======= +/** + * @brief 创建进程组 + * @details + * @param argc 1 + * @param[UINT32] pid 进程ID + * @return 函数执行结果 + * - ProcessGroup 返回进程组 +*/ +ProcessGroup *OsCreateProcessGroup(LosProcessCB *processCB) +{ + ProcessGroup *pgroup = LOS_MemAlloc(m_aucSysMem1, sizeof(ProcessGroup));//分配一个进程组 +>>>>>>> remotes/origin/main if (pgroup == NULL) { return NULL; } +<<<<<<< HEAD pgroup->pgroupLeader = (UINTPTR)processCB; LOS_ListInit(&pgroup->processList); LOS_ListInit(&pgroup->exitProcessList); @@ -147,6 +192,15 @@ ProcessGroup *OsCreateProcessGroup(LosProcessCB *processCB) LOS_ListTailInsert(&pgroup->processList, &processCB->subordinateGroupList); processCB->pgroup = pgroup; processCB->processStatus |= OS_PROCESS_FLAG_GROUP_LEADER; +======= + pgroup->pgroupLeader = (UINTPTR)processCB;//指定进程负责人 + LOS_ListInit(&pgroup->processList);//初始化组员链表 + LOS_ListInit(&pgroup->exitProcessList);//初始化僵死进程链表 + + LOS_ListTailInsert(&pgroup->processList, &processCB->subordinateGroupList); + processCB->pgroup = pgroup; + processCB->processStatus |= OS_PROCESS_FLAG_GROUP_LEADER;//进程状态贴上当老大的标签 +>>>>>>> remotes/origin/main ProcessGroup *rootPGroup = OS_ROOT_PGRP(processCB); if (rootPGroup == NULL) { @@ -157,30 +211,52 @@ ProcessGroup *OsCreateProcessGroup(LosProcessCB *processCB) } return pgroup; } +<<<<<<< HEAD STATIC VOID ExitProcessGroup(LosProcessCB *processCB, ProcessGroup **pgroup) { LosProcessCB *pgroupCB = OS_GET_PGROUP_LEADER(processCB->pgroup); LOS_ListDelete(&processCB->subordinateGroupList); +======= +//退出进程组,参数是进程地址和进程组地址的地址 +STATIC VOID ExitProcessGroup(LosProcessCB *processCB, ProcessGroup **pgroup) +{ + LosProcessCB *pgroupCB = OS_GET_PGROUP_LEADER(processCB->pgroup); + LOS_ListDelete(&processCB->subordinateGroupList);//从进程组进程链表上摘出去 +>>>>>>> remotes/origin/main if (LOS_ListEmpty(&processCB->pgroup->processList) && LOS_ListEmpty(&processCB->pgroup->exitProcessList)) { #ifdef LOSCFG_PID_CONTAINER if (processCB->pgroup != OS_ROOT_PGRP(processCB)) { #endif +<<<<<<< HEAD LOS_ListDelete(&processCB->pgroup->groupList); +======= + LOS_ListDelete(&processCB->pgroup->groupList);//从全局进程组链表上把自己摘出去 记住它是 LOS_ListTailInsert(&g_processGroup->groupList, &group->groupList) 挂上去的 +>>>>>>> remotes/origin/main *pgroup = processCB->pgroup; #ifdef LOSCFG_PID_CONTAINER } #endif pgroupCB->processStatus &= ~OS_PROCESS_FLAG_GROUP_LEADER; +<<<<<<< HEAD if (OsProcessIsUnused(pgroupCB) && !(pgroupCB->processStatus & OS_PROCESS_FLAG_EXIT)) { LOS_ListDelete(&pgroupCB->pendList); OsInsertPCBToFreeList(pgroupCB); +======= + if (OsProcessIsUnused(pgroupCB) && !(pgroupCB->processStatus & OS_PROCESS_FLAG_EXIT)) {//组长进程时退出的标签 + LOS_ListDelete(&pgroupCB->pendList);//进程从全局进程链表上摘除 + OsInsertPCBToFreeList(pgroupCB);//释放进程的资源,回到freelist再利用 +>>>>>>> remotes/origin/main } } processCB->pgroup = NULL; } +<<<<<<< HEAD +======= +/*! 通过指定组ID找到进程组 */ +>>>>>>> remotes/origin/main STATIC ProcessGroup *OsFindProcessGroup(UINT32 gid) { ProcessGroup *pgroup = NULL; @@ -200,7 +276,11 @@ STATIC ProcessGroup *OsFindProcessGroup(UINT32 gid) PRINT_INFO("%s failed! pgroup id = %u\n", __FUNCTION__, gid); return NULL; } +<<<<<<< HEAD +======= +/*! 给指定进程组发送信号 */ +>>>>>>> remotes/origin/main STATIC INT32 OsSendSignalToSpecifyProcessGroup(ProcessGroup *pgroup, siginfo_t *info, INT32 permission) { INT32 ret, success, err; @@ -293,16 +373,27 @@ STATIC LosProcessCB *OsFindExitChildProcess(const LosProcessCB *processCB, const return NULL; } +<<<<<<< HEAD +======= +/*! 唤醒等待wakePID结束的任务 */ +>>>>>>> remotes/origin/main VOID OsWaitWakeTask(LosTaskCB *taskCB, UINTPTR wakePID) { taskCB->waitID = wakePID; taskCB->ops->wake(taskCB); #ifdef LOSCFG_KERNEL_SMP +<<<<<<< HEAD LOS_MpSchedule(OS_MP_CPU_ALL); #endif } +======= + LOS_MpSchedule(OS_MP_CPU_ALL);//向所有cpu发送调度指令 +#endif +} +/*! 唤醒等待参数进程结束的任务 */ +>>>>>>> remotes/origin/main STATIC BOOL OsWaitWakeSpecifiedProcess(LOS_DL_LIST *head, const LosProcessCB *processCB, LOS_DL_LIST **anyList) { LOS_DL_LIST *list = head; @@ -310,17 +401,30 @@ STATIC BOOL OsWaitWakeSpecifiedProcess(LOS_DL_LIST *head, const LosProcessCB *pr UINTPTR processID = 0; BOOL find = FALSE; +<<<<<<< HEAD while (list->pstNext != head) { taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(list)); if ((taskCB->waitFlag == OS_PROCESS_WAIT_PRO) && (taskCB->waitID == (UINTPTR)processCB)) { if (processID == 0) { processID = taskCB->waitID; find = TRUE; +======= + while (list->pstNext != head) {//遍历等待链表 processCB->waitList + taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(list));//一个一个来 + if ((taskCB->waitFlag == OS_PROCESS_WAIT_PRO) && (taskCB->waitID == (UINTPTR)processCB)) { + if (processID == 0) { + processID = taskCB->waitID; + find = TRUE;//找到了 +>>>>>>> remotes/origin/main } else { processID = OS_INVALID_VALUE; } +<<<<<<< HEAD OsWaitWakeTask(taskCB, processID); +======= + OsWaitWakeTask(taskCB, processID);//唤醒这个任务,此时会切到 LOS_Wait runTask->waitFlag = 0;处运行 +>>>>>>> remotes/origin/main continue; } @@ -333,7 +437,11 @@ STATIC BOOL OsWaitWakeSpecifiedProcess(LOS_DL_LIST *head, const LosProcessCB *pr return find; } +<<<<<<< HEAD +======= +/*! 检查父进程的等待任务并唤醒父进程去处理等待任务 */ +>>>>>>> remotes/origin/main STATIC VOID OsWaitCheckAndWakeParentProcess(LosProcessCB *parentCB, const LosProcessCB *processCB) { LOS_DL_LIST *head = &parentCB->waitList; @@ -341,6 +449,7 @@ STATIC VOID OsWaitCheckAndWakeParentProcess(LosProcessCB *parentCB, const LosPro LosTaskCB *taskCB = NULL; BOOL findSpecified = FALSE; +<<<<<<< HEAD if (LOS_ListEmpty(&parentCB->waitList)) { return; } @@ -354,17 +463,40 @@ STATIC VOID OsWaitCheckAndWakeParentProcess(LosProcessCB *parentCB, const LosPro /* Other child processes exist, and other threads that are waiting * for the child to finish continue to wait */ +======= + if (LOS_ListEmpty(&parentCB->waitList)) {//父进程中是否有在等待子进程退出的任务? + return;//没有就退出 + } + + findSpecified = OsWaitWakeSpecifiedProcess(head, processCB, &list);//找到指定的任务 + if (findSpecified == TRUE) { + /* No thread is waiting for any child process to finish */ + if (LOS_ListEmpty(&parentCB->waitList)) {//没有线程正在等待任何子进程结束 + return;//已经处理完了,注意在OsWaitWakeSpecifiedProcess中做了频繁的任务切换 + } else if (!LOS_ListEmpty(&parentCB->childrenList)) { + /* Other child processes exist, and other threads that are waiting + * for the child to finish continue to wait + *///存在其他子进程,正在等待它们的子进程结束而将继续等待 +>>>>>>> remotes/origin/main return; } } /* Waiting threads are waiting for a specified child process to finish */ +<<<<<<< HEAD if (list == NULL) { +======= + if (list == NULL) {//等待线程正在等待指定的子进程结束 +>>>>>>> remotes/origin/main return; } /* No child processes exist and all waiting threads are awakened */ +<<<<<<< HEAD if (findSpecified == TRUE) { +======= + if (findSpecified == TRUE) {//所有等待的任务都被一一唤醒 +>>>>>>> remotes/origin/main while (list->pstNext != head) { taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(list)); OsWaitWakeTask(taskCB, OS_INVALID_VALUE); @@ -372,7 +504,11 @@ STATIC VOID OsWaitCheckAndWakeParentProcess(LosProcessCB *parentCB, const LosPro return; } +<<<<<<< HEAD while (list->pstNext != head) { +======= + while (list->pstNext != head) {//处理 OS_PROCESS_WAIT_GID 标签 +>>>>>>> remotes/origin/main taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(list)); if (taskCB->waitFlag == OS_PROCESS_WAIT_GID) { if (taskCB->waitID != (UINTPTR)OS_GET_PGROUP_LEADER(processCB->pgroup)) { @@ -395,7 +531,11 @@ STATIC VOID OsWaitCheckAndWakeParentProcess(LosProcessCB *parentCB, const LosPro return; } +<<<<<<< HEAD +======= +/*! 回收指定进程的资源 */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT VOID OsProcessResourcesToFree(LosProcessCB *processCB) { #ifdef LOSCFG_KERNEL_VM @@ -404,10 +544,17 @@ LITE_OS_SEC_TEXT VOID OsProcessResourcesToFree(LosProcessCB *processCB) } #endif +<<<<<<< HEAD #ifdef LOSCFG_SECURITY_CAPABILITY if (processCB->user != NULL) { (VOID)LOS_MemFree(m_aucSysMem1, processCB->user); processCB->user = NULL; +======= +#ifdef LOSCFG_SECURITY_CAPABILITY//安全开关 + if (processCB->user != NULL) { + (VOID)LOS_MemFree(m_aucSysMem1, processCB->user);//删除用户 + processCB->user = NULL;//重置指针为空 +>>>>>>> remotes/origin/main } #endif @@ -464,6 +611,7 @@ LITE_OS_SEC_TEXT VOID OsProcessResourcesToFree(LosProcessCB *processCB) processCB->resourceLimit = NULL; } } +<<<<<<< HEAD STATIC VOID OsRecycleZombiesProcess(LosProcessCB *childCB, ProcessGroup **pgroup) { @@ -485,6 +633,29 @@ STATIC VOID OsRecycleZombiesProcess(LosProcessCB *childCB, ProcessGroup **pgroup } } +======= +/*! 回收僵死状态进程的资源 */ +STATIC VOID OsRecycleZombiesProcess(LosProcessCB *childCB, ProcessGroup **pgroup) +{ + ExitProcessGroup(childCB, pgroup);//退出进程组 + LOS_ListDelete(&childCB->siblingList);//从父亲大人的子孙链表上摘除 + if (OsProcessIsDead(childCB)) { + OsDeleteTaskFromProcess(childCB->threadGroup); + childCB->processStatus &= ~OS_PROCESS_STATUS_ZOMBIES;//去掉僵死标签 + childCB->processStatus |= OS_PROCESS_FLAG_UNUSED;//贴上没使用标签,进程由进程池分配,进程退出后重新回到空闲进程池 + } + + LOS_ListDelete(&childCB->pendList);//将自己从阻塞链表上摘除,注意有很多原因引起阻塞,pendList挂在哪里就以为这属于哪类阻塞 + if (childCB->processStatus & OS_PROCESS_FLAG_EXIT) {//如果有退出标签 + LOS_ListHeadInsert(&g_processRecycleList, &childCB->pendList);//从头部插入,注意g_processRecyleList挂的是pendList节点,所以要通过OS_PCB_FROM_PENDLIST找. + } else if (OsProcessIsPGroupLeader(childCB)) { + LOS_ListTailInsert(&g_processRecycleList, &childCB->pendList);//从尾部插入,意思就是组长尽量最后一个处理 + } else { + OsInsertPCBToFreeList(childCB);//直接插到freeList中去,可用于重新分配了 + } +} +/*! 当一个进程自然退出的时候,它的孩子进程由两位老祖宗收养 */ +>>>>>>> remotes/origin/main STATIC VOID OsDealAliveChildProcess(LosProcessCB *processCB) { LosProcessCB *childCB = NULL; @@ -498,35 +669,59 @@ STATIC VOID OsDealAliveChildProcess(LosProcessCB *processCB) } #endif +<<<<<<< HEAD if (!LOS_ListEmpty(&processCB->childrenList)) { childHead = processCB->childrenList.pstNext; LOS_ListDelete(&(processCB->childrenList)); if (OsProcessIsUserMode(processCB)) { +======= + if (!LOS_ListEmpty(&processCB->childrenList)) {//如果存在孩子进程 + childHead = processCB->childrenList.pstNext;//获取孩子链表 + LOS_ListDelete(&(processCB->childrenList));//清空自己的孩子链表 + if (OsProcessIsUserMode(processCB)) {//是用户态进程 +>>>>>>> remotes/origin/main parentCB = OS_PCB_FROM_PID(OS_USER_ROOT_PROCESS_ID); } else { parentCB = OsGetKernelInitProcess(); } +<<<<<<< HEAD for (nextList = childHead; ;) { childCB = OS_PCB_FROM_SIBLIST(nextList); childCB->parentProcess = parentCB; nextList = nextList->pstNext; if (nextList == childHead) { +======= + for (nextList = childHead; ;) {//遍历孩子链表 + childCB = OS_PCB_FROM_SIBLIST(nextList);//找到孩子的真身 + childCB->parentProcess = parentCB; + nextList = nextList->pstNext;//找下一个孩子进程 + if (nextList == childHead) {//一圈下来,孩子们都磕完头了 +>>>>>>> remotes/origin/main break; } } +<<<<<<< HEAD LOS_ListTailInsertList(&parentCB->childrenList, childHead); +======= + LOS_ListTailInsertList(&parentCB->childrenList, childHead);//挂到老祖宗的孩子链表上 +>>>>>>> remotes/origin/main } return; } +<<<<<<< HEAD +======= +/*! 回收指定进程的已经退出(死亡)的孩子进程所占资源 */ +>>>>>>> remotes/origin/main STATIC VOID OsChildProcessResourcesFree(const LosProcessCB *processCB) { LosProcessCB *childCB = NULL; ProcessGroup *pgroup = NULL; +<<<<<<< HEAD while (!LOS_ListEmpty(&((LosProcessCB *)processCB)->exitChildList)) { childCB = LOS_DL_LIST_ENTRY(processCB->exitChildList.pstNext, LosProcessCB, siblingList); OsRecycleZombiesProcess(childCB, &pgroup); @@ -537,10 +732,23 @@ STATIC VOID OsChildProcessResourcesFree(const LosProcessCB *processCB) VOID OsProcessNaturalExit(LosProcessCB *processCB, UINT32 status) { OsChildProcessResourcesFree(processCB); +======= + while (!LOS_ListEmpty(&((LosProcessCB *)processCB)->exitChildList)) {//遍历直到没有了退出(死亡)的孩子进程 + childCB = LOS_DL_LIST_ENTRY(processCB->exitChildList.pstNext, LosProcessCB, siblingList);//获取孩子进程, + OsRecycleZombiesProcess(childCB, &pgroup);//其中会将childCB从exitChildList链表上摘出去 + (VOID)LOS_MemFree(m_aucSysMem1, pgroup); + } +} +/*! 一个进程的自然消亡过程,参数是当前运行的任务*/ +VOID OsProcessNaturalExit(LosProcessCB *processCB, UINT32 status) +{ + OsChildProcessResourcesFree(processCB);//释放孩子进程的资源 +>>>>>>> remotes/origin/main /* is a child process */ if (processCB->parentProcess != NULL) { LosProcessCB *parentCB = processCB->parentProcess; +<<<<<<< HEAD LOS_ListDelete(&processCB->siblingList); if (!OsProcessExitCodeSignalIsSet(processCB)) { OsProcessExitCodeSet(processCB, status); @@ -558,6 +766,26 @@ VOID OsProcessNaturalExit(LosProcessCB *processCB, UINT32 status) (VOID)OsSendSigToProcess(parentCB, SIGCHLD, OS_KERNEL_KILL_PERMISSION); #endif LOS_ListHeadInsert(&g_processRecycleList, &processCB->pendList); +======= + LOS_ListDelete(&processCB->siblingList);//将自己从兄弟链表中摘除,家人们,永别了! + if (!OsProcessExitCodeSignalIsSet(processCB)) {//是否设置了退出码? + OsProcessExitCodeSet(processCB, status);//将进程状态设为退出码 + } + LOS_ListTailInsert(&parentCB->exitChildList, &processCB->siblingList);//挂到父进程的孩子消亡链表,家人中,永别的可不止我一个. + LOS_ListDelete(&processCB->subordinateGroupList);//和志同道合的朋友们永别了,注意家里可不一定是朋友的,所有各有链表. + LOS_ListTailInsert(&processCB->pgroup->exitProcessList, &processCB->subordinateGroupList);//挂到进程组消亡链表,朋友中,永别的可不止我一个. + + OsWaitCheckAndWakeParentProcess(parentCB, processCB);//检查父进程的等待任务链表并唤醒对应的任务,此处将会频繁的切到其他任务运行. + + OsDealAliveChildProcess(processCB);//孩子们要怎么处理,移交给(用户态和内核态)根进程 + + + processCB->processStatus |= OS_PROCESS_STATUS_ZOMBIES;//贴上僵死进程的标签 +#ifdef LOSCFG_KERNEL_VM + (VOID)OsSendSigToProcess(parentCB, SIGCHLD, OS_KERNEL_KILL_PERMISSION); +#endif + LOS_ListHeadInsert(&g_processRecycleList, &processCB->pendList);//将进程通过其阻塞节点挂入全局进程回收链表 +>>>>>>> remotes/origin/main return; } @@ -571,17 +799,26 @@ STATIC VOID SystemProcessEarlyInit(LosProcessCB *processCB) #ifdef LOSCFG_KERNEL_CONTAINER OsContainerInitSystemProcess(processCB); #endif +<<<<<<< HEAD if (processCB == OsGetKernelInitProcess()) { OsSetMainTaskProcess((UINTPTR)processCB); } } +======= + if (processCB == OsGetKernelInitProcess()) {//2号进程 + OsSetMainTaskProcess((UINTPTR)processCB);//将内核根进程设为主任务所属进程 + } +} +/*! 进程模块初始化,被编译放在代码段 .init 中*/ +>>>>>>> remotes/origin/main UINT32 OsProcessInit(VOID) { UINT32 index; UINT32 size; UINT32 ret; +<<<<<<< HEAD g_processMaxNum = LOSCFG_BASE_CORE_PROCESS_LIMIT; size = (g_processMaxNum + 1) * sizeof(LosProcessCB); @@ -598,6 +835,24 @@ UINT32 OsProcessInit(VOID) g_processCBArray[index].processID = index; g_processCBArray[index].processStatus = OS_PROCESS_FLAG_UNUSED; LOS_ListTailInsert(&g_freeProcess, &g_processCBArray[index].pendList); +======= + g_processMaxNum = LOSCFG_BASE_CORE_PROCESS_LIMIT;//默认支持64个进程 + size = (g_processMaxNum + 1) * sizeof(LosProcessCB); + + g_processCBArray = (LosProcessCB *)LOS_MemAlloc(m_aucSysMem1, size);// 进程池,占用内核堆,内存池分配 + if (g_processCBArray == NULL) { + return LOS_NOK; + } + (VOID)memset_s(g_processCBArray, size, 0, size);//安全方式重置清0 + + LOS_ListInit(&g_freeProcess);//进程空闲链表初始化,创建一个进程时从g_freeProcess中申请一个进程描述符使用 + LOS_ListInit(&g_processRecycleList);//进程回收链表初始化,回收完成后进入g_freeProcess等待再次被申请使用 + + for (index = 0; index < g_processMaxNum; index++) {//进程池循环创建 + g_processCBArray[index].processID = index;//进程ID[0-g_processMaxNum-1]赋值 + g_processCBArray[index].processStatus = OS_PROCESS_FLAG_UNUSED;// 默认都是白纸一张,贴上未使用标签 + LOS_ListTailInsert(&g_freeProcess, &g_processCBArray[index].pendList);//注意g_freeProcess挂的是pendList节点,所以使用要通过OS_PCB_FROM_PENDLIST找到进程实体. +>>>>>>> remotes/origin/main } /* Default process to prevent thread PCB from being empty */ @@ -616,21 +871,36 @@ UINT32 OsProcessInit(VOID) #ifdef LOSCFG_KERNEL_PLIMITS OsProcLimiterSetInit(); #endif +<<<<<<< HEAD SystemProcessEarlyInit(OsGetIdleProcess()); +======= + SystemProcessEarlyInit(OsGetIdleProcess());//初始化 0,1,2号进程 +>>>>>>> remotes/origin/main SystemProcessEarlyInit(OsGetUserInitProcess()); SystemProcessEarlyInit(OsGetKernelInitProcess()); return LOS_OK; } +<<<<<<< HEAD +======= +/*! 进程回收再利用过程*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT VOID OsProcessCBRecycleToFree(VOID) { UINT32 intSave; LosProcessCB *processCB = NULL; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD while (!LOS_ListEmpty(&g_processRecycleList)) { processCB = OS_PCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&g_processRecycleList)); if (!(processCB->processStatus & OS_PROCESS_FLAG_EXIT)) { +======= + while (!LOS_ListEmpty(&g_processRecycleList)) {//循环任务回收链表,直到为空 + processCB = OS_PCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&g_processRecycleList));//找到回收链表中第一个进程实体 + //OS_PCB_FROM_PENDLIST 代表的是通过pendlist节点找到 PCB实体,因为g_processRecyleList上面挂的是pendlist节点位置 + if (!(processCB->processStatus & OS_PROCESS_FLAG_EXIT)) {//进程没有退出标签 +>>>>>>> remotes/origin/main break; } SCHEDULER_UNLOCK(intSave); @@ -638,17 +908,26 @@ LITE_OS_SEC_TEXT VOID OsProcessCBRecycleToFree(VOID) OsTaskCBRecycleToFree(); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD processCB->processStatus &= ~OS_PROCESS_FLAG_EXIT; #ifdef LOSCFG_KERNEL_VM LosVmSpace *space = NULL; if (OsProcessIsUserMode(processCB)) { space = processCB->vmSpace; +======= + processCB->processStatus &= ~OS_PROCESS_FLAG_EXIT;//给进程撕掉退出标签,(可能进程并没有这个标签) +#ifdef LOSCFG_KERNEL_VM + LosVmSpace *space = NULL; + if (OsProcessIsUserMode(processCB)) {//进程是否是用户态进程 + space = processCB->vmSpace;//只有用户态的进程才需要释放虚拟内存空间 +>>>>>>> remotes/origin/main } processCB->vmSpace = NULL; #endif /* OS_PROCESS_FLAG_GROUP_LEADER: The lead process group cannot be recycled without destroying the PCB. * !OS_PROCESS_FLAG_UNUSED: Parent process does not reclaim child process resources. */ +<<<<<<< HEAD LOS_ListDelete(&processCB->pendList); if (OsProcessIsPGroupLeader(processCB) || OsProcessIsDead(processCB)) { LOS_ListTailInsert(&g_processRecycleList, &processCB->pendList); @@ -659,13 +938,29 @@ LITE_OS_SEC_TEXT VOID OsProcessCBRecycleToFree(VOID) #ifdef LOSCFG_KERNEL_VM SCHEDULER_UNLOCK(intSave); (VOID)LOS_VmSpaceFree(space); +======= + LOS_ListDelete(&processCB->pendList);//将进程从进程链表上摘除 + if (OsProcessIsPGroupLeader(processCB) || OsProcessIsDead(processCB)) { + LOS_ListTailInsert(&g_processRecycleList, &processCB->pendList);//将进程挂到进程回收链表上,因为组长不能走啊 + } else { + /* Clear the bottom 4 bits of process status */ + OsInsertPCBToFreeList(processCB);//进程回到可分配池中,再分配利用 + } +#ifdef LOSCFG_KERNEL_VM + SCHEDULER_UNLOCK(intSave); + (VOID)LOS_VmSpaceFree(space);//释放用户态进程的虚拟内存空间,因为内核只有一个虚拟空间,因此不需要释放虚拟空间. +>>>>>>> remotes/origin/main SCHEDULER_LOCK(intSave); #endif } SCHEDULER_UNLOCK(intSave); } +<<<<<<< HEAD +======= +/*! 删除PCB块 其实是 PCB块回归进程池,先进入回收链表*/ +>>>>>>> remotes/origin/main STATIC VOID OsDeInitPCB(LosProcessCB *processCB) { UINT32 intSave; @@ -681,11 +976,19 @@ STATIC VOID OsDeInitPCB(LosProcessCB *processCB) } #endif +<<<<<<< HEAD OsProcessResourcesToFree(processCB); SCHEDULER_LOCK(intSave); if (processCB->parentProcess != NULL) { LOS_ListDelete(&processCB->siblingList); +======= + OsProcessResourcesToFree(processCB);//释放进程所占用的资源 + + SCHEDULER_LOCK(intSave); + if (processCB->parentProcess != NULL) { + LOS_ListDelete(&processCB->siblingList);//将进程从兄弟链表中摘除 +>>>>>>> remotes/origin/main processCB->parentProcess = NULL; } @@ -693,8 +996,13 @@ STATIC VOID OsDeInitPCB(LosProcessCB *processCB) ExitProcessGroup(processCB, &pgroup); } +<<<<<<< HEAD processCB->processStatus &= ~OS_PROCESS_STATUS_INIT; processCB->processStatus |= OS_PROCESS_FLAG_EXIT; +======= + processCB->processStatus &= ~OS_PROCESS_STATUS_INIT;//设置进程状态为非初始化 + processCB->processStatus |= OS_PROCESS_FLAG_EXIT;//设置进程状态为退出 +>>>>>>> remotes/origin/main LOS_ListHeadInsert(&g_processRecycleList, &processCB->pendList); SCHEDULER_UNLOCK(intSave); @@ -703,6 +1011,10 @@ STATIC VOID OsDeInitPCB(LosProcessCB *processCB) return; } +<<<<<<< HEAD +======= +/*! 设置进程的名字*/ +>>>>>>> remotes/origin/main UINT32 OsSetProcessName(LosProcessCB *processCB, const CHAR *name) { errno_t errRet; @@ -735,6 +1047,7 @@ UINT32 OsSetProcessName(LosProcessCB *processCB, const CHAR *name) return LOS_OK; } +<<<<<<< HEAD STATIC UINT32 OsInitPCB(LosProcessCB *processCB, UINT32 mode, const CHAR *name) { processCB->processMode = mode; @@ -752,13 +1065,38 @@ STATIC UINT32 OsInitPCB(LosProcessCB *processCB, UINT32 mode, const CHAR *name) #ifdef LOSCFG_KERNEL_VM if (OsProcessIsUserMode(processCB)) { processCB->vmSpace = OsCreateUserVmSpace(); +======= +/*! 初始化PCB(进程控制块)*/ +STATIC UINT32 OsInitPCB(LosProcessCB *processCB, UINT32 mode, const CHAR *name) +{ + processCB->processMode = mode;//用户态进程还是内核态进程 + processCB->processStatus = OS_PROCESS_STATUS_INIT;//进程初始状态 + processCB->parentProcess = NULL; + processCB->threadGroup = NULL; + processCB->umask = OS_PROCESS_DEFAULT_UMASK;//掩码 + processCB->timerID = (timer_t)(UINTPTR)MAX_INVALID_TIMER_VID; + + LOS_ListInit(&processCB->threadSiblingList);//初始化孩子任务/线程链表,上面挂的都是由此fork的孩子线程 见于 OsTaskCBInit LOS_ListTailInsert(&(processCB->threadSiblingList), &(taskCB->threadList)); + LOS_ListInit(&processCB->childrenList);//初始化孩子进程链表,上面挂的都是由此fork的孩子进程 见于 OsCopyParent LOS_ListTailInsert(&parentProcessCB->childrenList, &childProcessCB->siblingList); + LOS_ListInit(&processCB->exitChildList);//初始化记录退出孩子进程链表,上面挂的是哪些exit 见于 OsProcessNaturalExit LOS_ListTailInsert(&parentCB->exitChildList, &processCB->siblingList); + LOS_ListInit(&(processCB->waitList));//初始化等待任务链表 上面挂的是处于等待的 见于 OsWaitInsertWaitLIstInOrder LOS_ListHeadInsert(&processCB->waitList, &runTask->pendList); + +#ifdef LOSCFG_KERNEL_VM + if (OsProcessIsUserMode(processCB)) {//如果是用户态进程 + processCB->vmSpace = OsCreateUserVmSpace();//创建用户空间 +>>>>>>> remotes/origin/main if (processCB->vmSpace == NULL) { processCB->processStatus = OS_PROCESS_FLAG_UNUSED; return LOS_ENOMEM; } } else { +<<<<<<< HEAD processCB->vmSpace = LOS_GetKVmSpace(); } +======= + processCB->vmSpace = LOS_GetKVmSpace();//从这里也可以看出,所有内核态进程是共享一个进程空间的 + }//在鸿蒙内核态进程只有kprocess 和 kidle 两个 +>>>>>>> remotes/origin/main #endif #ifdef LOSCFG_KERNEL_CPUP @@ -777,7 +1115,11 @@ STATIC UINT32 OsInitPCB(LosProcessCB *processCB, UINT32 mode, const CHAR *name) #endif #ifdef LOSCFG_SECURITY_CAPABILITY +<<<<<<< HEAD OsInitCapability(processCB); +======= + OsInitCapability(processCB);//初始化进程安全相关功能 +>>>>>>> remotes/origin/main #endif if (OsSetProcessName(processCB, name) != LOS_OK) { @@ -786,10 +1128,17 @@ STATIC UINT32 OsInitPCB(LosProcessCB *processCB, UINT32 mode, const CHAR *name) return LOS_OK; } +<<<<<<< HEAD #ifdef LOSCFG_SECURITY_CAPABILITY STATIC User *OsCreateUser(UINT32 userID, UINT32 gid, UINT32 size) { +======= +//创建用户 +#ifdef LOSCFG_SECURITY_CAPABILITY +STATIC User *OsCreateUser(UINT32 userID, UINT32 gid, UINT32 size)//参数size 表示组数量 +{//(size - 1) * sizeof(UINT32) 用于 user->groups[..],这种设计节约了内存,不造成不需要的浪费 +>>>>>>> remotes/origin/main User *user = LOS_MemAlloc(m_aucSysMem1, sizeof(User) + (size - 1) * sizeof(UINT32)); if (user == NULL) { return NULL; @@ -799,11 +1148,20 @@ STATIC User *OsCreateUser(UINT32 userID, UINT32 gid, UINT32 size) user->effUserID = userID; user->gid = gid; user->effGid = gid; +<<<<<<< HEAD user->groupNumber = size; user->groups[0] = gid; return user; } +======= + user->groupNumber = size;//用户组数量 + user->groups[0] = gid;//用户组列表,一个用户可以属于多个用户组 + return user; +} + +/*! 检查参数群组ID是否在当前用户所属群组中*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT BOOL LOS_CheckInGroups(UINT32 gid) { UINT32 intSave; @@ -811,8 +1169,13 @@ LITE_OS_SEC_TEXT BOOL LOS_CheckInGroups(UINT32 gid) User *user = NULL; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD user = OsCurrUserGet(); for (count = 0; count < user->groupNumber; count++) { +======= + user = OsCurrUserGet();//当前进程所属用户 + for (count = 0; count < user->groupNumber; count++) {//循环对比 +>>>>>>> remotes/origin/main if (user->groups[count] == gid) { SCHEDULER_UNLOCK(intSave); return TRUE; @@ -824,6 +1187,10 @@ LITE_OS_SEC_TEXT BOOL LOS_CheckInGroups(UINT32 gid) } #endif +<<<<<<< HEAD +======= +/*! 获取当前进程的用户ID*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetUserID(VOID) { #ifdef LOSCFG_SECURITY_CAPABILITY @@ -843,6 +1210,10 @@ LITE_OS_SEC_TEXT INT32 LOS_GetUserID(VOID) #endif } +<<<<<<< HEAD +======= +/*! 获取当前进程的用户组ID*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetGroupID(VOID) { #ifdef LOSCFG_SECURITY_CAPABILITY @@ -863,6 +1234,10 @@ LITE_OS_SEC_TEXT INT32 LOS_GetGroupID(VOID) #endif } +<<<<<<< HEAD +======= +/*! 进程创建初始化*/ +>>>>>>> remotes/origin/main STATIC UINT32 OsSystemProcessInit(LosProcessCB *processCB, UINT32 flags, const CHAR *name) { UINT32 ret = OsInitPCB(processCB, flags, name); @@ -871,7 +1246,11 @@ STATIC UINT32 OsSystemProcessInit(LosProcessCB *processCB, UINT32 flags, const C } #ifdef LOSCFG_FS_VFS +<<<<<<< HEAD processCB->files = alloc_files(); +======= + processCB->files = alloc_files();//分配进程的文件的管理器 +>>>>>>> remotes/origin/main if (processCB->files == NULL) { ret = LOS_ENOMEM; goto EXIT; @@ -884,8 +1263,13 @@ STATIC UINT32 OsSystemProcessInit(LosProcessCB *processCB, UINT32 flags, const C goto EXIT; } +<<<<<<< HEAD #ifdef LOSCFG_SECURITY_CAPABILITY processCB->user = OsCreateUser(0, 0, 1); +======= +#ifdef LOSCFG_SECURITY_CAPABILITY //用户安全宏 + processCB->user = OsCreateUser(0, 0, 1); //创建用户 +>>>>>>> remotes/origin/main if (processCB->user == NULL) { ret = LOS_ENOMEM; goto EXIT; @@ -893,7 +1277,11 @@ STATIC UINT32 OsSystemProcessInit(LosProcessCB *processCB, UINT32 flags, const C #endif #ifdef LOSCFG_KERNEL_PLIMITS +<<<<<<< HEAD ret = OsPLimitsAddProcess(NULL, processCB); +======= + ret = OsPLimitsAddProcess(NULL, processCB);//删除进程控制块,归还内存 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { ret = LOS_ENOMEM; goto EXIT; @@ -905,7 +1293,11 @@ EXIT: OsDeInitPCB(processCB); return ret; } +<<<<<<< HEAD +======= +/*! 创建2,0号进程,即内核态进程的老祖宗*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsSystemProcessCreate(VOID) { LosProcessCB *kerInitProcess = OsGetKernelInitProcess(); @@ -913,14 +1305,22 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSystemProcessCreate(VOID) if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD kerInitProcess->processStatus &= ~OS_PROCESS_STATUS_INIT; LosProcessCB *idleProcess = OsGetIdleProcess(); ret = OsInitPCB(idleProcess, OS_KERNEL_MODE, "KIdle"); +======= + kerInitProcess->processStatus &= ~OS_PROCESS_STATUS_INIT;//去掉初始化标签 + + LosProcessCB *idleProcess = OsGetIdleProcess(); + ret = OsInitPCB(idleProcess, OS_KERNEL_MODE, "KIdle");//创建内核态0号进程 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } idleProcess->parentProcess = kerInitProcess; +<<<<<<< HEAD LOS_ListTailInsert(&kerInitProcess->childrenList, &idleProcess->siblingList); idleProcess->pgroup = kerInitProcess->pgroup; LOS_ListTailInsert(&kerInitProcess->pgroup->processList, &idleProcess->subordinateGroupList); @@ -929,6 +1329,16 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSystemProcessCreate(VOID) #endif #ifdef LOSCFG_FS_VFS idleProcess->files = kerInitProcess->files; +======= + LOS_ListTailInsert(&kerInitProcess->childrenList, &idleProcess->siblingList);//挂到内核态祖宗进程的子孙链接上 + idleProcess->pgroup = kerInitProcess->pgroup; + LOS_ListTailInsert(&kerInitProcess->pgroup->processList, &idleProcess->subordinateGroupList); +#ifdef LOSCFG_SECURITY_CAPABILITY + idleProcess->user = kerInitProcess->user;//共享用户 +#endif +#ifdef LOSCFG_FS_VFS + idleProcess->files = kerInitProcess->files;//共享文件 +>>>>>>> remotes/origin/main #endif idleProcess->processStatus &= ~OS_PROCESS_STATUS_INIT; @@ -938,7 +1348,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSystemProcessCreate(VOID) } return LOS_OK; } +<<<<<<< HEAD +======= +// 进程调度参数检查 +>>>>>>> remotes/origin/main INT32 OsSchedulerParamCheck(UINT16 policy, BOOL isThread, const LosSchedParam *param) { if (param == NULL) { @@ -971,17 +1385,26 @@ INT32 OsSchedulerParamCheck(UINT16 policy, BOOL isThread, const LosSchedParam *p STATIC INLINE INT32 ProcessSchedulerParamCheck(INT32 which, INT32 pid, UINT16 policy, const LosSchedParam *param) { +<<<<<<< HEAD if (OS_PID_CHECK_INVALID(pid)) { return LOS_EINVAL; } if (which != LOS_PRIO_PROCESS) { +======= + if (OS_PID_CHECK_INVALID(pid)) {//进程ID是否有效,默认 g_processMaxNum = 64 + return LOS_EINVAL; + } + + if (which != LOS_PRIO_PROCESS) {//进程标识 +>>>>>>> remotes/origin/main return LOS_EINVAL; } return OsSchedulerParamCheck(policy, FALSE, param); } +<<<<<<< HEAD #ifdef LOSCFG_SECURITY_CAPABILITY STATIC BOOL OsProcessCapPermitCheck(const LosProcessCB *processCB, const SchedParam *param, UINT16 policy, UINT16 prio) { @@ -989,6 +1412,15 @@ STATIC BOOL OsProcessCapPermitCheck(const LosProcessCB *processCB, const SchedPa /* always trust kernel process */ if (!OsProcessIsUserMode(runProcess)) { +======= +#ifdef LOSCFG_SECURITY_CAPABILITY//检查进程的安全许可证 +STATIC BOOL OsProcessCapPermitCheck(const LosProcessCB *processCB, const SchedParam *param, UINT16 policy, UINT16 prio) +{ + LosProcessCB *runProcess = OsCurrProcessGet();//获得当前进程 + + /* always trust kernel process */ + if (!OsProcessIsUserMode(runProcess)) {//进程必须在内核模式下,也就是说在内核模式下是安全的. +>>>>>>> remotes/origin/main return TRUE; } @@ -1004,7 +1436,11 @@ STATIC BOOL OsProcessCapPermitCheck(const LosProcessCB *processCB, const SchedPa return FALSE; } #endif +<<<<<<< HEAD +======= +// 设置进程调度计划 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 OsSetProcessScheduler(INT32 which, INT32 pid, UINT16 policy, const LosSchedParam *schedParam) { SchedParam param = { 0 }; @@ -1017,8 +1453,13 @@ LITE_OS_SEC_TEXT INT32 OsSetProcessScheduler(INT32 which, INT32 pid, UINT16 poli } LosProcessCB *processCB = OS_PCB_FROM_PID(pid); +<<<<<<< HEAD SCHEDULER_LOCK(intSave); if (OsProcessIsInactive(processCB)) { +======= + SCHEDULER_LOCK(intSave);//持有调度自旋锁,多CPU情况下调度期间需要原子处理 + if (OsProcessIsInactive(processCB)) {//进程未活动的处理 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return -LOS_ESRCH; } @@ -1058,6 +1499,7 @@ LITE_OS_SEC_TEXT INT32 OsSetProcessScheduler(INT32 which, INT32 pid, UINT16 poli needSched = taskCB->ops->schedParamModify(taskCB, ¶m); TO_SCHED: +<<<<<<< HEAD SCHEDULER_UNLOCK(intSave); LOS_MpSchedule(OS_MP_CPU_ALL); @@ -1067,11 +1509,26 @@ TO_SCHED: return LOS_OK; } +======= + SCHEDULER_UNLOCK(intSave);//还锁 + + LOS_MpSchedule(OS_MP_CPU_ALL);//核间中断 + if (needSched && OS_SCHEDULER_ACTIVE) { + LOS_Schedule();//发起调度 + } + return LOS_OK; +} +// 设置指定进程的调度参数,包括优先级和调度策略 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_SetProcessScheduler(INT32 pid, UINT16 policy, const LosSchedParam *schedParam) { return OsSetProcessScheduler(LOS_PRIO_PROCESS, pid, policy, schedParam); } +<<<<<<< HEAD +======= +// 获得指定进程的调度策略 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetProcessScheduler(INT32 pid, INT32 *policy, LosSchedParam *schedParam) { UINT32 intSave; @@ -1115,7 +1572,11 @@ LITE_OS_SEC_TEXT INT32 LOS_GetProcessScheduler(INT32 pid, INT32 *policy, LosSche } return LOS_OK; } +<<<<<<< HEAD +======= +// 接口封装 - 设置进程优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_SetProcessPriority(INT32 pid, INT32 prio) { INT32 ret; @@ -1135,7 +1596,11 @@ LITE_OS_SEC_TEXT INT32 LOS_SetProcessPriority(INT32 pid, INT32 prio) return OsSetProcessScheduler(LOS_PRIO_PROCESS, pid, (UINT16)policy, ¶m); } +<<<<<<< HEAD +======= +// 接口封装 - 获取进程优先级 which:标识进程,进程组,用户 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 OsGetProcessPriority(INT32 which, INT32 pid) { UINT32 intSave; @@ -1168,12 +1633,23 @@ LITE_OS_SEC_TEXT INT32 OsGetProcessPriority(INT32 which, INT32 pid) SCHEDULER_UNLOCK(intSave); return param.basePrio; } +<<<<<<< HEAD +======= +// 接口封装 - 获取指定进程优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetProcessPriority(INT32 pid) { return OsGetProcessPriority(LOS_PRIO_PROCESS, pid); } +<<<<<<< HEAD +======= +/*! +* 将任务挂入进程的waitList链表,表示这个任务在等待某个进程的退出 +* 当被等待进程退出时候会将自己挂到父进程的退出子进程链表和进程组的退出进程链表. +*/ +>>>>>>> remotes/origin/main STATIC VOID OsWaitInsertWaitListInOrder(LosTaskCB *runTask, LosProcessCB *processCB) { LOS_DL_LIST *head = &processCB->waitList; @@ -1205,7 +1681,11 @@ STATIC VOID OsWaitInsertWaitListInOrder(LosTaskCB *runTask, LosProcessCB *proces (VOID)runTask->ops->wait(runTask, list->pstNext, LOS_WAIT_FOREVER); return; } +<<<<<<< HEAD +======= +// 设置等待子进程退出方式方法 +>>>>>>> remotes/origin/main STATIC UINT32 WaitFindSpecifiedProcess(UINT32 pid, LosTaskCB *runTask, const LosProcessCB *processCB, LosProcessCB **childCB) { @@ -1226,7 +1706,11 @@ STATIC UINT32 WaitFindSpecifiedProcess(UINT32 pid, LosTaskCB *runTask, #endif /* Wait for the child process whose process number is pid. */ *childCB = OsFindExitChildProcess(processCB, waitProcess); +<<<<<<< HEAD if (*childCB != NULL) { +======= + if (*childCB != NULL) {//找到了,确实有一个已经退出的PID,注意一个进程退出时会挂到父进程的exitChildList上 +>>>>>>> remotes/origin/main return LOS_OK; } @@ -1234,7 +1718,11 @@ STATIC UINT32 WaitFindSpecifiedProcess(UINT32 pid, LosTaskCB *runTask, return LOS_ECHILD; } +<<<<<<< HEAD runTask->waitFlag = OS_PROCESS_WAIT_PRO; +======= + runTask->waitFlag = OS_PROCESS_WAIT_PRO;//设置当前任务的等待类型 +>>>>>>> remotes/origin/main runTask->waitID = (UINTPTR)waitProcess; return LOS_OK; } @@ -1253,6 +1741,7 @@ STATIC UINT32 OsWaitSetFlag(const LosProcessCB *processCB, INT32 pid, LosProcess if (childCB != NULL) { goto WAIT_BACK; } +<<<<<<< HEAD } else if (pid == 0) { /* Wait for any child process in the same process group */ childCB = OsFindGroupExitProcess(processCB->pgroup, OS_INVALID_VALUE); @@ -1265,6 +1754,20 @@ STATIC UINT32 OsWaitSetFlag(const LosProcessCB *processCB, INT32 pid, LosProcess /* Wait for any child process */ childCB = OsFindExitChildProcess(processCB, NULL); if (childCB != NULL) { +======= + } else if (pid == 0) {//等待同一进程组中的任何子进程 + /* Wait for any child process in the same process group */ + childCB = OsFindGroupExitProcess(processCB->pgroup, OS_INVALID_VALUE); + if (childCB != NULL) {//找到了,确实有一个已经退出的PID + goto WAIT_BACK;//直接成功返回 + } + runTask->waitID = (UINTPTR)OS_GET_PGROUP_LEADER(processCB->pgroup); + runTask->waitFlag = OS_PROCESS_WAIT_GID;//设置当前任务的等待类型 + } else if (pid == -1) {//等待任意子进程 + /* Wait for any child process */ + childCB = OsFindExitChildProcess(processCB, NULL); + if (childCB != NULL) {//找到了,确实有一个已经退出的PID +>>>>>>> remotes/origin/main goto WAIT_BACK; } runTask->waitID = pid; @@ -1282,14 +1785,22 @@ STATIC UINT32 OsWaitSetFlag(const LosProcessCB *processCB, INT32 pid, LosProcess } runTask->waitID = (UINTPTR)OS_GET_PGROUP_LEADER(pgroup); +<<<<<<< HEAD runTask->waitFlag = OS_PROCESS_WAIT_GID; +======= + runTask->waitFlag = OS_PROCESS_WAIT_GID;//设置当前任务的等待类型 +>>>>>>> remotes/origin/main } WAIT_BACK: *child = childCB; return LOS_OK; } +<<<<<<< HEAD +======= +// 等待回收孩子进程 +>>>>>>> remotes/origin/main STATIC UINT32 OsWaitRecycleChildProcess(const LosProcessCB *childCB, UINT32 intSave, INT32 *status, siginfo_t *info) { ProcessGroup *pgroup = NULL; @@ -1308,7 +1819,11 @@ STATIC UINT32 OsWaitRecycleChildProcess(const LosProcessCB *childCB, UINT32 intS SCHEDULER_UNLOCK(intSave); if (status != NULL) { +<<<<<<< HEAD if (mode == OS_USER_MODE) { +======= + if (mode == OS_USER_MODE) {//孩子为用户态进程 +>>>>>>> remotes/origin/main (VOID)LOS_ArchCopyToUser((VOID *)status, (const VOID *)(&(exitCode)), sizeof(INT32)); } else { *status = exitCode; @@ -1344,14 +1859,24 @@ STATIC UINT32 OsWaitRecycleChildProcess(const LosProcessCB *childCB, UINT32 intS (VOID)LOS_MemFree(m_aucSysMem1, pgroup); return pid; } +<<<<<<< HEAD STATIC UINT32 OsWaitChildProcessCheck(LosProcessCB *processCB, INT32 pid, LosProcessCB **childCB) { +======= +// 检查要等待的孩子进程 +STATIC UINT32 OsWaitChildProcessCheck(LosProcessCB *processCB, INT32 pid, LosProcessCB **childCB) +{//当进程没有孩子且没有退出的孩子进程 +>>>>>>> remotes/origin/main if (LOS_ListEmpty(&(processCB->childrenList)) && LOS_ListEmpty(&(processCB->exitChildList))) { return LOS_ECHILD; } +<<<<<<< HEAD return OsWaitSetFlag(processCB, pid, childCB); +======= + return OsWaitSetFlag(processCB, pid, childCB);//设置等待子进程退出方式方法 +>>>>>>> remotes/origin/main } STATIC UINT32 OsWaitOptionsCheck(UINT32 options) @@ -1359,6 +1884,7 @@ STATIC UINT32 OsWaitOptionsCheck(UINT32 options) UINT32 flag = LOS_WAIT_WNOHANG | LOS_WAIT_WUNTRACED | LOS_WAIT_WCONTINUED; flag = ~flag & options; +<<<<<<< HEAD if (flag != 0) { return LOS_EINVAL; } @@ -1369,11 +1895,27 @@ STATIC UINT32 OsWaitOptionsCheck(UINT32 options) if (OS_INT_ACTIVE) { return LOS_EINTR; +======= + if (flag != 0) {//三种方式中一种都没有 + return LOS_EINVAL;//无效参数 + } + + if ((options & (LOS_WAIT_WCONTINUED | LOS_WAIT_WUNTRACED)) != 0) {//暂不支持这两种方式. + return LOS_EOPNOTSUPP;//不支持 + } + + if (OS_INT_ACTIVE) {//中断发生期间 + return LOS_EINTR;//中断提示 +>>>>>>> remotes/origin/main } return LOS_OK; } +<<<<<<< HEAD +======= +//等待子进程结束并回收子进程,返回已经终止的子进程的进程ID号,并清除僵死进程。 +>>>>>>> remotes/origin/main STATIC INT32 OsWait(INT32 pid, USER INT32 *status, USER siginfo_t *info, UINT32 options, VOID *rusage) { (VOID)rusage; @@ -1384,13 +1926,21 @@ STATIC INT32 OsWait(INT32 pid, USER INT32 *status, USER siginfo_t *info, UINT32 LosProcessCB *processCB = OsCurrProcessGet(); LosTaskCB *runTask = OsCurrTaskGet(); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsWaitChildProcessCheck(processCB, pid, &childCB); +======= + ret = OsWaitChildProcessCheck(processCB, pid, &childCB);//先检查下看能不能找到参数要求的退出子进程 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { pid = -ret; goto ERROR; } +<<<<<<< HEAD if (childCB != NULL) { +======= + if (childCB != NULL) {//找到了进程 +>>>>>>> remotes/origin/main #ifdef LOSCFG_PID_CONTAINER if (childCB == processCB) { SCHEDULER_UNLOCK(intSave); @@ -1402,6 +1952,7 @@ STATIC INT32 OsWait(INT32 pid, USER INT32 *status, USER siginfo_t *info, UINT32 #endif return (INT32)OsWaitRecycleChildProcess(childCB, intSave, status, info); } +<<<<<<< HEAD if ((options & LOS_WAIT_WNOHANG) != 0) { runTask->waitFlag = 0; @@ -1417,6 +1968,24 @@ STATIC INT32 OsWait(INT32 pid, USER INT32 *status, USER siginfo_t *info, UINT32 goto ERROR; } +======= + //没有找到,看是否要返回还是去做个登记 + if ((options & LOS_WAIT_WNOHANG) != 0) {//有LOS_WAIT_WNOHANG标签 + runTask->waitFlag = 0;//等待标识置0 + pid = 0;//这里置0,是为了 return 0 + goto ERROR; + } + //等待孩子进程退出 + OsWaitInsertWaitListInOrder(runTask, processCB); + //发起调度的目的是为了让出CPU,让其他进程/任务运行 + + runTask->waitFlag = 0; + if (runTask->waitID == OS_INVALID_VALUE) { + pid = -LOS_ECHILD;//没有此子进程 + goto ERROR; + } + //回收僵死进程 +>>>>>>> remotes/origin/main childCB = (LosProcessCB *)runTask->waitID; if (!OsProcessIsDead(childCB)) { pid = -LOS_ESRCH; @@ -1503,12 +2072,21 @@ UINT32 OsGetProcessGroupCB(UINT32 pid, UINTPTR *ppgroupLeader) STATIC UINT32 OsSetProcessGroupCheck(const LosProcessCB *processCB, LosProcessCB *pgroupCB) { +<<<<<<< HEAD LosProcessCB *runProcessCB = OsCurrProcessGet(); if (OsProcessIsInactive(processCB)) { return LOS_ESRCH; } +======= + LosProcessCB *runProcessCB = OsCurrProcessGet();//拿到当前运行进程 + + if (OsProcessIsInactive(processCB)) {//进程是否活动 + return LOS_ESRCH; + } + //参数进程不在用户态或者组长不在用户态 +>>>>>>> remotes/origin/main #ifdef LOSCFG_PID_CONTAINER if ((processCB->processID == OS_USER_ROOT_PROCESS_ID) || OS_PROCESS_CONTAINER_CHECK(processCB, runProcessCB)) { return LOS_EPERM; @@ -1625,12 +2203,20 @@ EXIT: SCHEDULER_UNLOCK(intSave); return gid; } +<<<<<<< HEAD +======= +// 获取当前进程的组ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetCurrProcessGroupID(VOID) { return LOS_GetProcessGroupID(OsCurrProcessGet()->processID); } +<<<<<<< HEAD +======= +// 为用户态任务分配栈空间 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_VM STATIC LosProcessCB *OsGetFreePCB(VOID) { @@ -1654,19 +2240,34 @@ STATIC LosProcessCB *OsGetFreePCB(VOID) STATIC VOID *OsUserInitStackAlloc(LosProcessCB *processCB, UINT32 *size) { LosVmMapRegion *region = NULL; +<<<<<<< HEAD UINT32 stackSize = ALIGN(OS_USER_TASK_STACK_SIZE, PAGE_SIZE); region = LOS_RegionAlloc(processCB->vmSpace, 0, stackSize, VM_MAP_REGION_FLAG_PERM_USER | VM_MAP_REGION_FLAG_PERM_READ | VM_MAP_REGION_FLAG_PERM_WRITE, 0); +======= + UINT32 stackSize = ALIGN(OS_USER_TASK_STACK_SIZE, PAGE_SIZE);//1M栈空间 按页对齐 + //线性区分配虚拟内存 + region = LOS_RegionAlloc(processCB->vmSpace, 0, stackSize, + VM_MAP_REGION_FLAG_PERM_USER | VM_MAP_REGION_FLAG_PERM_READ | + VM_MAP_REGION_FLAG_PERM_WRITE, 0);//可使用可读可写区 +>>>>>>> remotes/origin/main if (region == NULL) { return NULL; } +<<<<<<< HEAD LOS_SetRegionTypeAnon(region); region->regionFlags |= VM_MAP_REGION_FLAG_STACK; *size = stackSize; +======= + LOS_SetRegionTypeAnon(region);//匿名映射 + region->regionFlags |= VM_MAP_REGION_FLAG_STACK;//标记该线性区为栈区 + + *size = stackSize;//记录栈大小 +>>>>>>> remotes/origin/main return (VOID *)(UINTPTR)region->range.base; } @@ -1701,6 +2302,17 @@ LITE_OS_SEC_TEXT LosVmSpace *OsExecProcessVmSpaceReplace(LosVmSpace *newSpace, U return oldSpace; } +<<<<<<< HEAD +======= +/** + * @brief 进程的回收再利用,被LOS_DoExecveFile调用 + * @param processCB + * @param name + * @param oldSpace + * @param oldFiles + * @return LITE_OS_SEC_TEXT + */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsExecRecycleAndInit(LosProcessCB *processCB, const CHAR *name, LosVmSpace *oldSpace, UINTPTR oldFiles) { @@ -1750,7 +2362,11 @@ LITE_OS_SEC_TEXT UINT32 OsExecRecycleAndInit(LosProcessCB *processCB, const CHAR return LOS_OK; } +<<<<<<< HEAD +======= +// 执行用户态任务, entry为入口函数 ,其中 创建好task,task上下文 等待调度真正执行, sp:栈指针 mapBase:栈底 mapSize:栈大小 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsExecStart(const TSK_ENTRY_FUNC entry, UINTPTR sp, UINTPTR mapBase, UINT32 mapSize) { UINT32 intSave; @@ -1759,6 +2375,7 @@ LITE_OS_SEC_TEXT UINT32 OsExecStart(const TSK_ENTRY_FUNC entry, UINTPTR sp, UINT return LOS_NOK; } +<<<<<<< HEAD if ((sp == 0) || (LOS_Align(sp, LOSCFG_STACK_POINT_ALIGN_SIZE) != sp)) { return LOS_NOK; } @@ -1782,6 +2399,32 @@ LITE_OS_SEC_TEXT UINT32 OsExecStart(const TSK_ENTRY_FUNC entry, UINTPTR sp, UINT } #endif +======= + if ((sp == 0) || (LOS_Align(sp, LOSCFG_STACK_POINT_ALIGN_SIZE) != sp)) {//对齐 + return LOS_NOK; + } + //注意 sp此时指向栈底,栈底地址要大于栈顶 + if ((mapBase == 0) || (mapSize == 0) || (sp <= mapBase) || (sp > (mapBase + mapSize))) {//参数检查 + return LOS_NOK; + } + + LosTaskCB *taskCB = OsCurrTaskGet();//获取当前任务 + + SCHEDULER_LOCK(intSave);//拿自旋锁 + taskCB->userMapBase = mapBase;//用户态栈顶位置 + taskCB->userMapSize = mapSize;//用户态栈 + taskCB->taskEntry = (TSK_ENTRY_FUNC)entry;//任务的入口函数 + //初始化内核态栈 + TaskContext *taskContext = (TaskContext *)OsTaskStackInit(taskCB->taskID, taskCB->stackSize, + (VOID *)taskCB->topOfStack, FALSE); + OsUserTaskStackInit(taskContext, (UINTPTR)taskCB->taskEntry, sp);//初始化用户栈,将内核栈中上下文的 context->R[0] = sp ,context->sp = sp + //这样做的目的是将用户栈SP保存到内核栈中, + SCHEDULER_UNLOCK(intSave);//解锁 + return LOS_OK; +} +#endif +// 用户进程开始初始化 +>>>>>>> remotes/origin/main STATIC UINT32 OsUserInitProcessStart(LosProcessCB *processCB, TSK_INIT_PARAM_S *param) { UINT32 intSave; @@ -1802,7 +2445,11 @@ STATIC UINT32 OsUserInitProcessStart(LosProcessCB *processCB, TSK_INIT_PARAM_S * processCB->processStatus &= ~OS_PROCESS_STATUS_INIT; SCHEDULER_UNLOCK(intSave); +<<<<<<< HEAD ret = LOS_SetTaskScheduler(taskID, LOS_SCHED_RR, OS_TASK_PRIORITY_LOWEST); +======= + ret = LOS_SetTaskScheduler(taskID, LOS_SCHED_RR, OS_TASK_PRIORITY_LOWEST);//调度器:设置为抢占式调度和最低任务优先级(31级) +>>>>>>> remotes/origin/main if (ret != LOS_OK) { PRINT_ERR("User init process set scheduler failed! ERROR:%d \n", ret); goto EXIT; @@ -1921,7 +2568,11 @@ ERROR: OsDeInitPCB(processCB); return ret; } +<<<<<<< HEAD +======= +// 拷贝用户信息 直接用memcpy_s +>>>>>>> remotes/origin/main STATIC UINT32 OsCopyUser(LosProcessCB *childCB, LosProcessCB *parentCB) { #ifdef LOSCFG_SECURITY_CAPABILITY @@ -1936,6 +2587,10 @@ STATIC UINT32 OsCopyUser(LosProcessCB *childCB, LosProcessCB *parentCB) return LOS_OK; } +<<<<<<< HEAD +======= +//拷贝一个Task过程 +>>>>>>> remotes/origin/main STATIC VOID GetCopyTaskParam(LosProcessCB *childProcessCB, UINTPTR entry, UINT32 size, TSK_INIT_PARAM_S *taskParam, SchedParam *param) { @@ -1943,6 +2598,7 @@ STATIC VOID GetCopyTaskParam(LosProcessCB *childProcessCB, UINTPTR entry, UINT32 LosTaskCB *runTask = OsCurrTaskGet(); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (OsProcessIsUserMode(childProcessCB)) { taskParam->pfnTaskEntry = runTask->taskEntry; taskParam->uwStackSize = runTask->stackSize; @@ -1952,6 +2608,17 @@ STATIC VOID GetCopyTaskParam(LosProcessCB *childProcessCB, UINTPTR entry, UINT32 } else { taskParam->pfnTaskEntry = (TSK_ENTRY_FUNC)entry; taskParam->uwStackSize = size; +======= + if (OsProcessIsUserMode(childProcessCB)) {//用户态进程 + taskParam->pfnTaskEntry = runTask->taskEntry;//拷贝当前任务入口地址 + taskParam->uwStackSize = runTask->stackSize;//栈空间大小 + taskParam->userParam.userArea = runTask->userArea;//用户态栈区栈顶位置 + taskParam->userParam.userMapBase = runTask->userMapBase;//用户态栈底 + taskParam->userParam.userMapSize = runTask->userMapSize;//用户态栈大小 + } else {//注意内核态进程创建任务的入口由外界指定,例如 OsCreateIdleProcess 指定了OsIdleTask + taskParam->pfnTaskEntry = (TSK_ENTRY_FUNC)entry;//参数(sp)为内核态入口地址 + taskParam->uwStackSize = size;//参数(size)为内核态栈大小 +>>>>>>> remotes/origin/main } if (runTask->taskStatus & OS_TASK_FLAG_PTHREAD_JOIN) { taskParam->uwResved = LOS_TASK_ATTR_JOINABLE; @@ -1987,6 +2654,7 @@ STATIC UINT32 OsCopyTask(UINT32 flags, LosProcessCB *childProcessCB, const CHAR } LosTaskCB *childTaskCB = childProcessCB->threadGroup; +<<<<<<< HEAD childTaskCB->taskStatus = runTask->taskStatus; childTaskCB->ops->schedParamModify(childTaskCB, ¶m); if (childTaskCB->taskStatus & OS_TASK_STATUS_RUNNING) { @@ -1999,13 +2667,31 @@ STATIC UINT32 OsCopyTask(UINT32 flags, LosProcessCB *childProcessCB, const CHAR } if (OsProcessIsUserMode(childProcessCB)) { +======= + childTaskCB->taskStatus = runTask->taskStatus;//任务状态先同步,注意这里是赋值操作. ...01101001 + childTaskCB->ops->schedParamModify(childTaskCB, ¶m); + if (childTaskCB->taskStatus & OS_TASK_STATUS_RUNNING) {//因只能有一个运行的task,所以如果一样要改4号位 + childTaskCB->taskStatus &= ~OS_TASK_STATUS_RUNNING;//将四号位清0 ,变成 ...01100001 + } else {//非运行状态下会发生什么? + if (OS_SCHEDULER_ACTIVE) {//克隆线程发生错误未运行 + LOS_Panic("Clone thread status not running error status: 0x%x\n", childTaskCB->taskStatus); + } + childTaskCB->taskStatus &= ~OS_TASK_STATUS_UNUSED;//干净的Task + } + + if (OsProcessIsUserMode(childProcessCB)) {//是否是用户进程 +>>>>>>> remotes/origin/main SCHEDULER_LOCK(intSave); OsUserCloneParentStack(childTaskCB->stackPointer, entry, runTask->topOfStack, runTask->stackSize); SCHEDULER_UNLOCK(intSave); } return LOS_OK; } +<<<<<<< HEAD +======= +//拷贝父亲大人的遗传基因信息 +>>>>>>> remotes/origin/main STATIC UINT32 OsCopyParent(UINT32 flags, LosProcessCB *childProcessCB, LosProcessCB *runProcessCB) { UINT32 intSave; @@ -2013,6 +2699,7 @@ STATIC UINT32 OsCopyParent(UINT32 flags, LosProcessCB *childProcessCB, LosProces SCHEDULER_LOCK(intSave); if (childProcessCB->parentProcess == NULL) { +<<<<<<< HEAD if (flags & CLONE_PARENT) { parentProcessCB = runProcessCB->parentProcess; } else { @@ -2020,6 +2707,16 @@ STATIC UINT32 OsCopyParent(UINT32 flags, LosProcessCB *childProcessCB, LosProces } childProcessCB->parentProcess = parentProcessCB; LOS_ListTailInsert(&parentProcessCB->childrenList, &childProcessCB->siblingList); +======= + if (flags & CLONE_PARENT) { //这里指明 childProcessCB 和 runProcessCB 有同一个父亲,是兄弟关系 + parentProcessCB = runProcessCB->parentProcess; + } else { + parentProcessCB = runProcessCB;//指认父亲,这个赋值代表从此是你儿了 + } + childProcessCB->parentProcess = parentProcessCB; + LOS_ListTailInsert(&parentProcessCB->childrenList, &childProcessCB->siblingList);//通过我的兄弟姐妹节点,挂到父亲的孩子链表上,于我而言,父亲的这个链表上挂的都是我的兄弟姐妹 + //不会被排序,老大,老二,老三 老天爷指定了。 +>>>>>>> remotes/origin/main } if (childProcessCB->pgroup == NULL) { @@ -2029,12 +2726,17 @@ STATIC UINT32 OsCopyParent(UINT32 flags, LosProcessCB *childProcessCB, LosProces SCHEDULER_UNLOCK(intSave); return LOS_OK; } +<<<<<<< HEAD +======= +//拷贝虚拟空间 +>>>>>>> remotes/origin/main STATIC UINT32 OsCopyMM(UINT32 flags, LosProcessCB *childProcessCB, LosProcessCB *runProcessCB) { status_t status; UINT32 intSave; +<<<<<<< HEAD if (!OsProcessIsUserMode(childProcessCB)) { return LOS_OK; } @@ -2043,6 +2745,16 @@ STATIC UINT32 OsCopyMM(UINT32 flags, LosProcessCB *childProcessCB, LosProcessCB SCHEDULER_LOCK(intSave); childProcessCB->vmSpace->archMmu.virtTtb = runProcessCB->vmSpace->archMmu.virtTtb; childProcessCB->vmSpace->archMmu.physTtb = runProcessCB->vmSpace->archMmu.physTtb; +======= + if (!OsProcessIsUserMode(childProcessCB)) {//不是用户模式,直接返回,什么意思?内核虚拟空间只有一个,无需COPY !!! + return LOS_OK; + } + + if (flags & CLONE_VM) {//贴有虚拟内存的标签 + SCHEDULER_LOCK(intSave); + childProcessCB->vmSpace->archMmu.virtTtb = runProcessCB->vmSpace->archMmu.virtTtb;//TTB虚拟地址基地址,即L1表存放位置,virtTtb是个指针,进程的虚拟空间是指定的范围的 + childProcessCB->vmSpace->archMmu.physTtb = runProcessCB->vmSpace->archMmu.physTtb;//TTB物理地址基地址,physTtb是个值,取决于运行时映射到物理内存的具体哪个位置. +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return LOS_OK; } @@ -2053,7 +2765,11 @@ STATIC UINT32 OsCopyMM(UINT32 flags, LosProcessCB *childProcessCB, LosProcessCB } return LOS_OK; } +<<<<<<< HEAD +======= +// 拷贝进程文件描述符(proc_fd)信息 +>>>>>>> remotes/origin/main STATIC UINT32 OsCopyFile(UINT32 flags, LosProcessCB *childProcessCB, LosProcessCB *runProcessCB) { #ifdef LOSCFG_FS_VFS @@ -2083,7 +2799,11 @@ STATIC UINT32 OsCopyFile(UINT32 flags, LosProcessCB *childProcessCB, LosProcessC #endif #endif +<<<<<<< HEAD childProcessCB->consoleID = runProcessCB->consoleID; +======= + childProcessCB->consoleID = runProcessCB->consoleID;//控制台也是文件 +>>>>>>> remotes/origin/main childProcessCB->umask = runProcessCB->umask; return LOS_OK; } @@ -2091,16 +2811,28 @@ STATIC UINT32 OsCopyFile(UINT32 flags, LosProcessCB *childProcessCB, LosProcessC STATIC UINT32 OsForkInitPCB(UINT32 flags, LosProcessCB *child, const CHAR *name, UINTPTR sp, UINT32 size) { UINT32 ret; +<<<<<<< HEAD LosProcessCB *run = OsCurrProcessGet(); ret = OsCopyParent(flags, child, run); +======= + LosProcessCB *run = OsCurrProcessGet();//获取当前进程 + + ret = OsCopyParent(flags, child, run);//拷贝父亲大人的基因信息 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD return OsCopyTask(flags, child, name, sp, size); } +======= + return OsCopyTask(flags, child, name, sp, size);//拷贝任务,设置任务入口函数,栈大小 +} +//设置进程组和加入进程调度就绪队列 +>>>>>>> remotes/origin/main STATIC UINT32 OsChildSetProcessGroupAndSched(LosProcessCB *child, LosProcessCB *run) { UINT32 intSave; @@ -2124,7 +2856,11 @@ STATIC UINT32 OsChildSetProcessGroupAndSched(LosProcessCB *child, LosProcessCB * (VOID)LOS_MemFree(m_aucSysMem1, pgroup); return LOS_OK; } +<<<<<<< HEAD +======= +// 拷贝进程资源 +>>>>>>> remotes/origin/main STATIC UINT32 OsCopyProcessResources(UINT32 flags, LosProcessCB *child, LosProcessCB *run) { UINT32 ret; @@ -2134,26 +2870,42 @@ STATIC UINT32 OsCopyProcessResources(UINT32 flags, LosProcessCB *child, LosProce return ret; } +<<<<<<< HEAD ret = OsCopyMM(flags, child, run); +======= + ret = OsCopyMM(flags, child, run);//拷贝虚拟空间 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD ret = OsCopyFile(flags, child, run); +======= + ret = OsCopyFile(flags, child, run);//拷贝文件信息 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } #ifdef LOSCFG_KERNEL_LITEIPC +<<<<<<< HEAD if (run->ipcInfo != NULL) { child->ipcInfo = LiteIpcPoolReInit((const ProcIpcInfo *)(run->ipcInfo)); if (child->ipcInfo == NULL) { return LOS_ENOMEM; +======= + if (run->ipcInfo != NULL) {//重新初始化IPC池 + child->ipcInfo = LiteIpcPoolReInit((const ProcIpcInfo *)(run->ipcInfo));//@note_good 将沿用用户态空间地址(即线性区地址) + if (child->ipcInfo == NULL) {//因为整个进程虚拟空间都是拷贝的,ipc的用户态虚拟地址当然可以拷贝,但因进程不同了,所以需要重新申请ipc池和重新 + return LOS_ENOMEM;//映射池中两个地址. +>>>>>>> remotes/origin/main } } #endif #ifdef LOSCFG_SECURITY_CAPABILITY +<<<<<<< HEAD OsCopyCapability(run, child); #endif return LOS_OK; @@ -2165,6 +2917,19 @@ STATIC INT32 OsCopyProcess(UINT32 flags, const CHAR *name, UINTPTR sp, UINT32 si LosProcessCB *run = OsCurrProcessGet(); LosProcessCB *child = OsGetFreePCB(); +======= + OsCopyCapability(run, child);//拷贝安全能力 +#endif + return LOS_OK; +} +// 拷贝进程 +STATIC INT32 OsCopyProcess(UINT32 flags, const CHAR *name, UINTPTR sp, UINT32 size) +{ + UINT32 ret, processID; + LosProcessCB *run = OsCurrProcessGet();//获取当前进程 + + LosProcessCB *child = OsGetFreePCB();//从进程池中申请一个进程控制块,鸿蒙进程池默认64 +>>>>>>> remotes/origin/main if (child == NULL) { return -LOS_EAGAIN; } @@ -2189,24 +2954,42 @@ STATIC INT32 OsCopyProcess(UINT32 flags, const CHAR *name, UINTPTR sp, UINT32 si #endif #endif +<<<<<<< HEAD ret = OsForkInitPCB(flags, child, name, sp, size); +======= + ret = OsForkInitPCB(flags, child, name, sp, size);//初始化进程控制块 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { goto ERROR_INIT; } +<<<<<<< HEAD ret = OsCopyProcessResources(flags, child, run); +======= + ret = OsCopyProcessResources(flags, child, run);//拷贝进程的资源,包括虚拟空间,文件,安全,IPC == +>>>>>>> remotes/origin/main if (ret != LOS_OK) { goto ERROR_TASK; } +<<<<<<< HEAD ret = OsChildSetProcessGroupAndSched(child, run); +======= + ret = OsChildSetProcessGroupAndSched(child, run);//设置进程组和加入进程调度就绪队列 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { goto ERROR_TASK; } +<<<<<<< HEAD LOS_MpSchedule(OS_MP_CPU_ALL); if (OS_SCHEDULER_ACTIVE) { LOS_Schedule(); +======= + LOS_MpSchedule(OS_MP_CPU_ALL);//给各CPU发送准备接受调度信号 + if (OS_SCHEDULER_ACTIVE) {//当前CPU core处于活动状态 + LOS_Schedule();// 申请调度 +>>>>>>> remotes/origin/main } return processID; @@ -2218,6 +3001,19 @@ ERROR_INIT: return -ret; } +<<<<<<< HEAD +======= +/*! + * @brief OsClone 进程克隆 + * + * @param flags + * @param size 进程主任务内核栈大小 + * @param sp 进程主任务的入口函数 + * @return + * + * @see + */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 OsClone(UINT32 flags, UINTPTR sp, UINT32 size) { UINT32 cloneFlag = CLONE_PARENT | CLONE_THREAD | SIGCHLD; @@ -2266,7 +3062,10 @@ LITE_OS_SEC_TEXT INT32 OsClone(UINT32 flags, UINTPTR sp, UINT32 size) return OsCopyProcess(cloneFlag & flags, NULL, sp, size); } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_Fork(UINT32 flags, const CHAR *name, const TSK_ENTRY_FUNC entry, UINT32 stackSize) { UINT32 cloneFlag = CLONE_PARENT | CLONE_THREAD | CLONE_VFORK | CLONE_FILES; @@ -2276,7 +3075,11 @@ LITE_OS_SEC_TEXT INT32 LOS_Fork(UINT32 flags, const CHAR *name, const TSK_ENTRY_ } flags |= CLONE_FILES; +<<<<<<< HEAD return OsCopyProcess(cloneFlag & flags, name, (UINTPTR)entry, stackSize); +======= + return OsCopyProcess(cloneFlag & flags, name, (UINTPTR)entry, stackSize);//拷贝一个进程 +>>>>>>> remotes/origin/main } #else LITE_OS_SEC_TEXT_INIT UINT32 OsUserInitProcess(VOID) @@ -2285,6 +3088,17 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsUserInitProcess(VOID) } #endif +<<<<<<< HEAD +======= +/*! + * @brief LOS_Exit + * 进程退出 + * @param status + * @return + * + * @see + */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT VOID LOS_Exit(INT32 status) { UINT32 intSave; @@ -2294,7 +3108,11 @@ LITE_OS_SEC_TEXT VOID LOS_Exit(INT32 status) /* The exit of a kernel - state process must be kernel - state and all threads must actively exit */ LosProcessCB *processCB = OsCurrProcessGet(); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (!OsProcessIsUserMode(processCB) && (processCB->threadNumber != 1)) { +======= + if (!OsProcessIsUserMode(processCB) && (processCB->threadNumber != 1)) {//内核态下进程的退出方式,必须是所有的任务都退出了 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); PRINT_ERR("Kernel-state processes with multiple threads are not allowed to exit directly\n"); return; @@ -2305,6 +3123,18 @@ LITE_OS_SEC_TEXT VOID LOS_Exit(INT32 status) OsRunningTaskToExit(OsCurrTaskGet(), OS_PRO_EXIT_OK); } +<<<<<<< HEAD +======= +/*! + * @brief LOS_GetUsedPIDList + * 获取使用中的进程列表 + * @param pidList + * @param pidMaxNum + * @return + * + * @see + */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetUsedPIDList(UINT32 *pidList, INT32 pidMaxNum) { LosProcessCB *pcb = NULL; @@ -2316,6 +3146,7 @@ LITE_OS_SEC_TEXT INT32 LOS_GetUsedPIDList(UINT32 *pidList, INT32 pidMaxNum) return 0; } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD while (OsProcessIDUserCheckInvalid(pid) == false) { pcb = OS_PCB_FROM_PID(pid); pid++; @@ -2323,6 +3154,15 @@ LITE_OS_SEC_TEXT INT32 LOS_GetUsedPIDList(UINT32 *pidList, INT32 pidMaxNum) continue; } pidList[num] = pcb->processID; +======= + while (OsProcessIDUserCheckInvalid(pid) == false) {//遍历进程池 + pcb = OS_PCB_FROM_PID(pid); + pid++; + if (OsProcessIsUnused(pcb)) {//未使用的不算 + continue; + } + pidList[num] = pcb->processID;//由参数带走 +>>>>>>> remotes/origin/main num++; if (num >= pidMaxNum) { break; @@ -2348,12 +3188,20 @@ LITE_OS_SEC_TEXT struct fd_table_s *LOS_GetFdTable(UINT32 pid) return files->fdt; } #endif +<<<<<<< HEAD +======= +// 获取当前进程的进程ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_GetCurrProcessID(VOID) { return OsCurrProcessGet()->processID; } +<<<<<<< HEAD +======= +// 按指定状态退出指定进程 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_VM STATIC VOID ThreadGroupActiveTaskKilled(LosTaskCB *taskCB) { @@ -2426,12 +3274,20 @@ LITE_OS_SEC_TEXT VOID OsProcessThreadGroupDestroy(VOID) #endif return; } +<<<<<<< HEAD +======= +// 获取系统支持的最大进程数目 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_GetSystemProcessMaximum(VOID) { return g_processMaxNum; } +<<<<<<< HEAD +======= +// 获取用户态进程的根进程,所有用户进程都是g_processCBArray[g_userInitProcess] fork来的 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT LosProcessCB *OsGetUserInitProcess(VOID) { return &g_processCBArray[OS_USER_ROOT_PROCESS_ID]; @@ -2441,7 +3297,11 @@ LITE_OS_SEC_TEXT LosProcessCB *OsGetKernelInitProcess(VOID) { return &g_processCBArray[OS_KERNEL_ROOT_PROCESS_ID]; } +<<<<<<< HEAD +======= +// 获取空闲进程,0号进程为空闲进程,该进程不干活,专给CPU休息的。 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT LosProcessCB *OsGetIdleProcess(VOID) { return &g_processCBArray[OS_KERNEL_IDLE_PROCESS_ID]; diff --git a/src/kernel_liteos_a/kernel/base/core/los_smp.c b/src/kernel_liteos_a/kernel/base/core/los_smp.c index bbeb9c22..8c26e82c 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_smp.c +++ b/src/kernel_liteos_a/kernel/base/core/los_smp.c @@ -40,7 +40,11 @@ #ifdef LOSCFG_KERNEL_SMP STATIC struct SmpOps *g_smpOps = NULL; +<<<<<<< HEAD +======= +/// 多核中次级CPU核初始化,每个核都会调用一次 +>>>>>>> remotes/origin/main STATIC VOID OsSmpSecondaryInit(VOID *arg) { UNUSED(arg); @@ -56,7 +60,11 @@ STATIC VOID OsSmpSecondaryInit(VOID *arg) OsSchedStart(); } +<<<<<<< HEAD +======= +/// 设置多核操作接口, 通过外部注册 +>>>>>>> remotes/origin/main VOID LOS_SmpOpsSet(struct SmpOps *ops) { g_smpOps = ops; diff --git a/src/kernel_liteos_a/kernel/base/core/los_swtmr.c b/src/kernel_liteos_a/kernel/base/core/los_swtmr.c index 285753c3..ebd40614 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_swtmr.c +++ b/src/kernel_liteos_a/kernel/base/core/los_swtmr.c @@ -28,7 +28,66 @@ * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ - +<<<<<<< HEAD + +======= +/*基本概念 + 软件定时器,是基于系统Tick时钟中断且由软件来模拟的定时器。当经过设定的Tick数后,会触发用户自定义的回调函数。 + 硬件定时器受硬件的限制,数量上不足以满足用户的实际需求。因此为了满足用户需求,提供更多的定时器, + 软件定时器功能,支持如下特性: + 创建软件定时器。 + 启动软件定时器。 + 停止软件定时器。 + 删除软件定时器。 + 获取软件定时器剩余Tick数。 + 可配置支持的软件定时器个数。 + + 运作机制 + 软件定时器是系统资源,在模块初始化的时候已经分配了一块连续内存。 + 软件定时器使用了系统的一个队列和一个任务资源,软件定时器的触发遵循队列规则, + 先进先出。定时时间短的定时器总是比定时时间长的靠近队列头,满足优先触发的准则。 + 软件定时器以Tick为基本计时单位,当创建并启动一个软件定时器时,Huawei LiteOS会根据 + 当前系统Tick时间及设置的定时时长确定该定时器的到期Tick时间,并将该定时器控制结构挂入计时全局链表。 + 当Tick中断到来时,在Tick中断处理函数中扫描软件定时器的计时全局链表,检查是否有定时器超时, + 若有则将超时的定时器记录下来。Tick中断处理函数结束后,软件定时器任务(优先级为最高) + 被唤醒,在该任务中调用已经记录下来的定时器的回调函数。 + + 定时器状态 + OS_SWTMR_STATUS_UNUSED(定时器未使用) + 系统在定时器模块初始化时,会将系统中所有定时器资源初始化成该状态。 + + OS_SWTMR_STATUS_TICKING(定时器处于计数状态) + 在定时器创建后调用LOS_SwtmrStart接口启动,定时器将变成该状态,是定时器运行时的状态。 + + OS_SWTMR_STATUS_CREATED(定时器创建后未启动,或已停止) + 定时器创建后,不处于计数状态时,定时器将变成该状态。 + + 软件定时器提供了三类模式: + 单次触发定时器,这类定时器在启动后只会触发一次定时器事件,然后定时器自动删除。 + 周期触发定时器,这类定时器会周期性的触发定时器事件,直到用户手动停止定时器,否则将永远持续执行下去。 + 单次触发定时器,但这类定时器超时触发后不会自动删除,需要调用定时器删除接口删除定时器。 + + 使用场景 + 创建一个单次触发的定时器,超时后执行用户自定义的回调函数。 + 创建一个周期性触发的定时器,超时后执行用户自定义的回调函数。 + + 软件定时器的典型开发流程 + 通过make menuconfig配置软件定时器 + 创建定时器LOS_SwtmrCreate,设置定时器的定时时长、定时器模式、超时后的回调函数。 + 启动定时器LOS_SwtmrStart。 + 获得软件定时器剩余Tick数LOS_SwtmrTimeGet。 + 停止定时器LOS_SwtmrStop。 + 删除定时器LOS_SwtmrDelete。 + + 注意事项 + 软件定时器的回调函数中不应执行过多操作,不建议使用可能引起任务挂起或者阻塞的接口或操作, + 如果使用会导致软件定时器响应不及时,造成的影响无法确定。 + 软件定时器使用了系统的一个队列和一个任务资源。软件定时器任务的优先级设定为0,且不允许修改 。 + 系统可配置的软件定时器个数是指:整个系统可使用的软件定时器总个数,并非用户可使用的软件定时器个数。 + 例如:系统多占用一个软件定时器,那么用户能使用的软件定时器资源就会减少一个。 + 创建单次不自删除属性的定时器,用户需要自行调用定时器删除接口删除定时器,回收定时器资源,避免资源泄露。 + 软件定时器的定时精度与系统Tick时钟的周期有关。*/ +>>>>>>> remotes/origin/main #include "los_swtmr_pri.h" #include "los_init.h" #include "los_process_pri.h" @@ -51,6 +110,7 @@ LITE_OS_SEC_BSS UINT8 *g_swtmrHandlerPool = NULL; /* Pool of Swtmr Han LITE_OS_SEC_BSS LOS_DL_LIST g_swtmrFreeList; /* Free list of Software Timer */ /* spinlock for swtmr module, only available on SMP mode */ +<<<<<<< HEAD LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_swtmrSpin); #define SWTMR_LOCK(state) LOS_SpinLockSave(&g_swtmrSpin, &(state)) #define SWTMR_UNLOCK(state) LOS_SpinUnlockRestore(&g_swtmrSpin, (state)) @@ -59,6 +119,16 @@ typedef struct { SortLinkAttribute swtmrSortLink; LosTaskCB *swtmrTask; /* software timer task id */ LOS_DL_LIST swtmrHandlerQueue; /* software timer timeout queue id */ +======= +LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_swtmrSpin);//初始化软件钟自旋锁,只有SMP情况才需要,只要是自旋锁都是由于CPU多核的同步 +#define SWTMR_LOCK(state) LOS_SpinLockSave(&g_swtmrSpin, &(state))//持有软时钟自旋锁 +#define SWTMR_UNLOCK(state) LOS_SpinUnlockRestore(&g_swtmrSpin, (state))//释放软时钟自旋锁 + +typedef struct { + SortLinkAttribute swtmrSortLink; + LosTaskCB *swtmrTask; /* software timer task id *///定时器任务ID + LOS_DL_LIST swtmrHandlerQueue; /* software timer timeout queue id *///定时器超时队列 +>>>>>>> remotes/origin/main } SwtmrRunqueue; STATIC SwtmrRunqueue g_swtmrRunqueue[LOSCFG_KERNEL_CORE_NUM]; @@ -254,7 +324,11 @@ STATIC INLINE VOID ScanSwtmrTimeList(SwtmrRunqueue *srq) LOS_SpinUnlockRestore(&swtmrSortLink->spinLock, intSave); return; } +<<<<<<< HEAD +======= +//软时钟的入口函数,拥有任务的最高优先级0级 +>>>>>>> remotes/origin/main STATIC VOID SwtmrTask(VOID) { SwtmrHandlerItem swtmrHandle; @@ -263,7 +337,11 @@ STATIC VOID SwtmrTask(VOID) SwtmrRunqueue *srq = &g_swtmrRunqueue[ArchCurrCpuid()]; LOS_DL_LIST *head = &srq->swtmrHandlerQueue; +<<<<<<< HEAD for (;;) { +======= + for (;;) {//死循环获取队列item,一直读干净为止 +>>>>>>> remotes/origin/main waitTime = OsSortLinkGetNextExpireTime(OsGetCurrSchedTimeCycle(), &srq->swtmrSortLink); if (waitTime != 0) { SCHEDULER_LOCK(intSave); @@ -279,17 +357,26 @@ STATIC VOID SwtmrTask(VOID) LOS_ListDelete(&swtmrHandlePtr->node); (VOID)memcpy_s(&swtmrHandle, sizeof(SwtmrHandlerItem), swtmrHandlePtr, sizeof(SwtmrHandlerItem)); +<<<<<<< HEAD (VOID)LOS_MemboxFree(g_swtmrHandlerPool, swtmrHandlePtr); +======= + (VOID)LOS_MemboxFree(g_swtmrHandlerPool, swtmrHandlePtr);//静态释放内存,注意在鸿蒙内核只有软时钟注册用到了静态内存 +>>>>>>> remotes/origin/main SwtmrHandler(&swtmrHandle); } } } +<<<<<<< HEAD +======= +//创建软时钟任务,每个cpu core都可以拥有自己的软时钟任务 +>>>>>>> remotes/origin/main STATIC UINT32 SwtmrTaskCreate(UINT16 cpuid, UINT32 *swtmrTaskID) { UINT32 ret; TSK_INIT_PARAM_S swtmrTask; +<<<<<<< HEAD (VOID)memset_s(&swtmrTask, sizeof(TSK_INIT_PARAM_S), 0, sizeof(TSK_INIT_PARAM_S)); swtmrTask.pfnTaskEntry = (TSK_ENTRY_FUNC)SwtmrTask; swtmrTask.uwStackSize = LOSCFG_BASE_CORE_TSK_DEFAULT_STACK_SIZE; @@ -302,6 +389,20 @@ STATIC UINT32 SwtmrTaskCreate(UINT16 cpuid, UINT32 *swtmrTaskID) ret = LOS_TaskCreate(swtmrTaskID, &swtmrTask); if (ret == LOS_OK) { OS_TCB_FROM_TID(*swtmrTaskID)->taskStatus |= OS_TASK_FLAG_SYSTEM_TASK; +======= + (VOID)memset_s(&swtmrTask, sizeof(TSK_INIT_PARAM_S), 0, sizeof(TSK_INIT_PARAM_S));//清0 + swtmrTask.pfnTaskEntry = (TSK_ENTRY_FUNC)SwtmrTask;//入口函数 + swtmrTask.uwStackSize = LOSCFG_BASE_CORE_TSK_DEFAULT_STACK_SIZE;//16k默认内核任务栈 + swtmrTask.pcName = "Swt_Task";//任务名称 + swtmrTask.usTaskPrio = 0;// + swtmrTask.uwResved = LOS_TASK_STATUS_DETACHED;//分离模式 +#ifdef LOSCFG_KERNEL_SMP + swtmrTask.usCpuAffiMask = CPUID_TO_AFFI_MASK(cpuid);//交给当前CPU执行这个任务 +#endif + ret = LOS_TaskCreate(swtmrTaskID, &swtmrTask);//创建任务并申请调度 + if (ret == LOS_OK) { + OS_TCB_FROM_TID(*swtmrTaskID)->taskStatus |= OS_TASK_FLAG_SYSTEM_TASK;//告知这是个任务系统 +>>>>>>> remotes/origin/main } return ret; @@ -319,6 +420,7 @@ BOOL OsIsSwtmrTask(const LosTaskCB *taskCB) } return FALSE; } +<<<<<<< HEAD LITE_OS_SEC_TEXT_INIT VOID OsSwtmrRecycle(UINTPTR ownerID) { @@ -329,6 +431,18 @@ LITE_OS_SEC_TEXT_INIT VOID OsSwtmrRecycle(UINTPTR ownerID) } } +======= +//回收指定进程的软时钟 +LITE_OS_SEC_TEXT_INIT VOID OsSwtmrRecycle(UINTPTR ownerID) +{ + for (UINT16 index = 0; index < LOSCFG_BASE_CORE_SWTMR_LIMIT; index++) {//一个进程往往会有多个定时器 + if (g_swtmrCBArray[index].uwOwnerPid == ownerID) { + LOS_SwtmrDelete(index);//删除定时器 + } + } +} +//软时钟初始化,注意函数在多CPU情况下会执行多次 +>>>>>>> remotes/origin/main STATIC UINT32 SwtmrBaseInit(VOID) { UINT32 ret; @@ -338,6 +452,7 @@ STATIC UINT32 SwtmrBaseInit(VOID) return LOS_ERRNO_SWTMR_NO_MEMORY; } +<<<<<<< HEAD (VOID)memset_s(swtmr, size, 0, size); g_swtmrCBArray = swtmr; LOS_ListInit(&g_swtmrFreeList); @@ -347,6 +462,17 @@ STATIC UINT32 SwtmrBaseInit(VOID) } size = LOS_MEMBOX_SIZE(sizeof(SwtmrHandlerItem), OS_SWTMR_HANDLE_QUEUE_SIZE); +======= + (VOID)memset_s(swtmr, size, 0, size);//清0 + g_swtmrCBArray = swtmr;//软时钟 + LOS_ListInit(&g_swtmrFreeList);//初始化空间链表 + for (UINT16 index = 0; index < LOSCFG_BASE_CORE_SWTMR_LIMIT; index++, swtmr++) { + swtmr->usTimerID = index;//按顺序赋值 + LOS_ListTailInsert(&g_swtmrFreeList, &swtmr->stSortList.sortLinkNode);//通过sortLinkNode将节点挂到空闲链表 + } + //想要用静态内存池管理,就必须要使用LOS_MEMBOX_SIZE来计算申请列表的内存大小,因为需要点前缀内存承载头部信息 + size = LOS_MEMBOX_SIZE(sizeof(SwtmrHandlerItem), OS_SWTMR_HANDLE_QUEUE_SIZE);//规划一片内存区域作为软时钟处理函数的静态内存池 +>>>>>>> remotes/origin/main g_swtmrHandlerPool = (UINT8 *)LOS_MemAlloc(m_aucSysMem1, size); /* system resident resource */ if (g_swtmrHandlerPool == NULL) { return LOS_ERRNO_SWTMR_NO_MEMORY; @@ -474,6 +600,7 @@ STATIC UINT64 SwtmrToStart(SWTMR_CTRL_S *swtmr, UINT16 cpuid) if ((swtmr->uwOverrun == 0) && ((swtmr->ucMode == LOS_SWTMR_MODE_ONCE) || (swtmr->ucMode == LOS_SWTMR_MODE_OPP) || +<<<<<<< HEAD (swtmr->ucMode == LOS_SWTMR_MODE_NO_SELFDELETE))) { ticks = swtmr->uwExpiry; } else { @@ -482,6 +609,16 @@ STATIC UINT64 SwtmrToStart(SWTMR_CTRL_S *swtmr, UINT16 cpuid) swtmr->ucState = OS_SWTMR_STATUS_TICKING; UINT64 period = (UINT64)ticks * OS_CYCLE_PER_TICK; +======= + (swtmr->ucMode == LOS_SWTMR_MODE_NO_SELFDELETE))) {//如果是一次性的定时器 + ticks = swtmr->uwExpiry;//获取定时间隔 + } else { + ticks = swtmr->uwInterval; + } + swtmr->ucState = OS_SWTMR_STATUS_TICKING;//获取周期性定时器时间间隔 + + UINT64 period = (UINT64)ticks * OS_CYCLE_PER_TICK;//计数状态 +>>>>>>> remotes/origin/main UINT64 responseTime = swtmr->startTime + period; UINT64 currTime = OsGetCurrSchedTimeCycle(); if (responseTime < currTime) { @@ -522,7 +659,11 @@ STATIC INLINE VOID SwtmrStart(SWTMR_CTRL_S *swtmr) STATIC INLINE VOID SwtmrDelete(SWTMR_CTRL_S *swtmr) { /* insert to free list */ +<<<<<<< HEAD LOS_ListTailInsert(&g_swtmrFreeList, &swtmr->stSortList.sortLinkNode); +======= + LOS_ListTailInsert(&g_swtmrFreeList, &swtmr->stSortList.sortLinkNode);//直接插入空闲链表中,回收再利用 +>>>>>>> remotes/origin/main swtmr->ucState = OS_SWTMR_STATUS_UNUSED; swtmr->uwOwnerPid = OS_INVALID_VALUE; @@ -641,7 +782,11 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsSwtmrTimeGet(const SWTMR_CTRL_S *swtmr) } return (UINT32)time; } +<<<<<<< HEAD +======= +//创建定时器,设置定时器的定时时长、定时模式、回调函数、并返回定时器ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_SwtmrCreate(UINT32 interval, UINT8 mode, SWTMR_PROC_FUNC handler, @@ -670,7 +815,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SwtmrCreate(UINT32 interval, } SWTMR_LOCK(intSave); +<<<<<<< HEAD if (LOS_ListEmpty(&g_swtmrFreeList)) { +======= + if (LOS_ListEmpty(&g_swtmrFreeList)) {//空闲列表不能为空 +>>>>>>> remotes/origin/main SWTMR_UNLOCK(intSave); return LOS_ERRNO_SWTMR_MAXSIZE; } @@ -681,6 +830,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SwtmrCreate(UINT32 interval, SWTMR_UNLOCK(intSave); swtmr->uwOwnerPid = (UINTPTR)OsCurrProcessGet(); +<<<<<<< HEAD swtmr->pfnHandler = handler; swtmr->ucMode = mode; swtmr->uwOverrun = 0; @@ -688,12 +838,25 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SwtmrCreate(UINT32 interval, swtmr->uwExpiry = interval; swtmr->uwArg = arg; swtmr->ucState = OS_SWTMR_STATUS_CREATED; +======= + swtmr->pfnHandler = handler;//时间到了的回调函数 + swtmr->ucMode = mode;//定时模式 + swtmr->uwOverrun = 0; + swtmr->uwInterval = interval;//周期性超时间隔 + swtmr->uwExpiry = interval;//一次性超时间隔 + swtmr->uwArg = arg;//回调函数的参数 + swtmr->ucState = OS_SWTMR_STATUS_CREATED;//已创建状态 +>>>>>>> remotes/origin/main SET_SORTLIST_VALUE(&swtmr->stSortList, OS_SORT_LINK_INVALID_TIME); *swtmrID = swtmr->usTimerID; OsHookCall(LOS_HOOK_TYPE_SWTMR_CREATE, swtmr); return LOS_OK; } +<<<<<<< HEAD +======= +//接口函数 启动定时器 参数定时任务ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStart(UINT16 swtmrID) { SWTMR_CTRL_S *swtmr = NULL; @@ -705,15 +868,24 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStart(UINT16 swtmrID) return LOS_ERRNO_SWTMR_ID_INVALID; } +<<<<<<< HEAD swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT; swtmr = g_swtmrCBArray + swtmrCBID; SWTMR_LOCK(intSave); if (swtmr->usTimerID != swtmrID) { +======= + swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT;//取模 + swtmr = g_swtmrCBArray + swtmrCBID;//获取定时器控制结构体 + + SWTMR_LOCK(intSave); + if (swtmr->usTimerID != swtmrID) {//ID必须一样 +>>>>>>> remotes/origin/main SWTMR_UNLOCK(intSave); return LOS_ERRNO_SWTMR_ID_INVALID; } +<<<<<<< HEAD switch (swtmr->ucState) { case OS_SWTMR_STATUS_UNUSED: ret = LOS_ERRNO_SWTMR_NOT_CREATED; @@ -726,6 +898,20 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStart(UINT16 swtmrID) SwtmrStop(swtmr); /* fall-through */ case OS_SWTMR_STATUS_CREATED: +======= + switch (swtmr->ucState) {//判断定时器状态 + case OS_SWTMR_STATUS_UNUSED: + ret = LOS_ERRNO_SWTMR_NOT_CREATED; + break; + /*如果定时器的状态为启动中,应先停止定时器再重新启动 + * If the status of swtmr is timing, it should stop the swtmr first, + * then start the swtmr again. + */ + case OS_SWTMR_STATUS_TICKING://正在计数的定时器 + SwtmrStop(swtmr);//先停止定时器,注意这里没有break;在OsSwtmrStop中状态将会回到了OS_SWTMR_STATUS_CRWEATED接下来就是执行启动了 + /* fall-through */ + case OS_SWTMR_STATUS_CREATED://已经创建好了 +>>>>>>> remotes/origin/main SwtmrStart(swtmr); break; default: @@ -737,7 +923,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStart(UINT16 swtmrID) OsHookCall(LOS_HOOK_TYPE_SWTMR_START, swtmr); return ret; } +<<<<<<< HEAD +======= +//接口函数 停止计时器 参加定时任务ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStop(UINT16 swtmrID) { SWTMR_CTRL_S *swtmr = NULL; @@ -749,15 +939,24 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStop(UINT16 swtmrID) return LOS_ERRNO_SWTMR_ID_INVALID; } +<<<<<<< HEAD swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT; swtmr = g_swtmrCBArray + swtmrCBID; SWTMR_LOCK(intSave); if (swtmr->usTimerID != swtmrID) { +======= + swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT;//取模 + swtmr = g_swtmrCBArray + swtmrCBID;//获取定时器控制结构体 + SWTMR_LOCK(intSave); + + if (swtmr->usTimerID != swtmrID) {//ID必须一样 +>>>>>>> remotes/origin/main SWTMR_UNLOCK(intSave); return LOS_ERRNO_SWTMR_ID_INVALID; } +<<<<<<< HEAD switch (swtmr->ucState) { case OS_SWTMR_STATUS_UNUSED: ret = LOS_ERRNO_SWTMR_NOT_CREATED; @@ -767,6 +966,17 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStop(UINT16 swtmrID) break; case OS_SWTMR_STATUS_TICKING: SwtmrStop(swtmr); +======= + switch (swtmr->ucState) {//判断定时器状态 + case OS_SWTMR_STATUS_UNUSED: + ret = LOS_ERRNO_SWTMR_NOT_CREATED;//返回没有创建 + break; + case OS_SWTMR_STATUS_CREATED: + ret = LOS_ERRNO_SWTMR_NOT_STARTED;//返回没有开始 + break; + case OS_SWTMR_STATUS_TICKING://正在计数 + SwtmrStop(swtmr);//执行正在停止计时器操作 +>>>>>>> remotes/origin/main break; default: ret = LOS_ERRNO_SWTMR_STATUS_INVALID; @@ -777,7 +987,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrStop(UINT16 swtmrID) OsHookCall(LOS_HOOK_TYPE_SWTMR_STOP, swtmr); return ret; } +<<<<<<< HEAD +======= +//接口函数 获得软件定时器剩余Tick数 通过 *tick 带走 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_SwtmrTimeGet(UINT16 swtmrID, UINT32 *tick) { SWTMR_CTRL_S *swtmr = NULL; @@ -793,11 +1007,19 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrTimeGet(UINT16 swtmrID, UINT32 *tick) return LOS_ERRNO_SWTMR_TICK_PTR_NULL; } +<<<<<<< HEAD swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT; swtmr = g_swtmrCBArray + swtmrCBID; SWTMR_LOCK(intSave); if (swtmr->usTimerID != swtmrID) { +======= + swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT;//取模 + swtmr = g_swtmrCBArray + swtmrCBID;//获取定时器控制结构体 + SWTMR_LOCK(intSave); + + if (swtmr->usTimerID != swtmrID) {//ID必须一样 +>>>>>>> remotes/origin/main SWTMR_UNLOCK(intSave); return LOS_ERRNO_SWTMR_ID_INVALID; } @@ -808,8 +1030,13 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrTimeGet(UINT16 swtmrID, UINT32 *tick) case OS_SWTMR_STATUS_CREATED: ret = LOS_ERRNO_SWTMR_NOT_STARTED; break; +<<<<<<< HEAD case OS_SWTMR_STATUS_TICKING: *tick = OsSwtmrTimeGet(swtmr); +======= + case OS_SWTMR_STATUS_TICKING://正在计数的定时器 + *tick = OsSwtmrTimeGet(swtmr);//获取 +>>>>>>> remotes/origin/main break; default: ret = LOS_ERRNO_SWTMR_STATUS_INVALID; @@ -818,7 +1045,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrTimeGet(UINT16 swtmrID, UINT32 *tick) SWTMR_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +//接口函数 删除定时器 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_SwtmrDelete(UINT16 swtmrID) { SWTMR_CTRL_S *swtmr = NULL; @@ -830,11 +1061,19 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrDelete(UINT16 swtmrID) return LOS_ERRNO_SWTMR_ID_INVALID; } +<<<<<<< HEAD swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT; swtmr = g_swtmrCBArray + swtmrCBID; SWTMR_LOCK(intSave); if (swtmr->usTimerID != swtmrID) { +======= + swtmrCBID = swtmrID % LOSCFG_BASE_CORE_SWTMR_LIMIT;//取模 + swtmr = g_swtmrCBArray + swtmrCBID;//获取定时器控制结构体 + SWTMR_LOCK(intSave); + + if (swtmr->usTimerID != swtmrID) {//ID必须一样 +>>>>>>> remotes/origin/main SWTMR_UNLOCK(intSave); return LOS_ERRNO_SWTMR_ID_INVALID; } @@ -843,10 +1082,17 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrDelete(UINT16 swtmrID) case OS_SWTMR_STATUS_UNUSED: ret = LOS_ERRNO_SWTMR_NOT_CREATED; break; +<<<<<<< HEAD case OS_SWTMR_STATUS_TICKING: SwtmrStop(swtmr); /* fall-through */ case OS_SWTMR_STATUS_CREATED: +======= + case OS_SWTMR_STATUS_TICKING://正在计数就先停止在删除,这里没有break + SwtmrStop(swtmr); + /* fall-through */ + case OS_SWTMR_STATUS_CREATED://再删除定时器 +>>>>>>> remotes/origin/main SwtmrDelete(swtmr); break; default: @@ -859,4 +1105,8 @@ LITE_OS_SEC_TEXT UINT32 LOS_SwtmrDelete(UINT16 swtmrID) return ret; } +<<<<<<< HEAD +#endif /* LOSCFG_BASE_CORE_SWTMR_ENABLE */ +======= #endif /* LOSCFG_BASE_CORE_SWTMR_ENABLE */ +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/core/los_task.c b/src/kernel_liteos_a/kernel/base/core/los_task.c index 2bd80726..7d78a1d7 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_task.c +++ b/src/kernel_liteos_a/kernel/base/core/los_task.c @@ -68,6 +68,7 @@ #if (LOSCFG_BASE_CORE_TSK_LIMIT <= 0) #error "task maxnum cannot be zero" #endif /* LOSCFG_BASE_CORE_TSK_LIMIT <= 0 */ +<<<<<<< HEAD LITE_OS_SEC_BSS LosTaskCB *g_taskCBArray; LITE_OS_SEC_BSS LOS_DL_LIST g_losFreeTask; @@ -75,6 +76,85 @@ LITE_OS_SEC_BSS LOS_DL_LIST g_taskRecycleList; LITE_OS_SEC_BSS UINT32 g_taskMaxNum; LITE_OS_SEC_BSS UINT32 g_taskScheduled; /* one bit for each cores */ LITE_OS_SEC_BSS EVENT_CB_S g_resourceEvent; +======= +/* +基本概念 + 从系统角度看,任务是竞争系统资源的最小运行单元。任务可以使用或等待CPU、 + 使用内存空间等系统资源,并独立于其它任务运行。 + 任务模块可以给用户提供多个任务,实现任务间的切换,帮助用户管理业务程序流程。具有如下特性: + 支持多任务。 + 一个任务表示一个线程。 + 抢占式调度机制,高优先级的任务可打断低优先级任务,低优先级任务必须在高优先级任务阻塞或结束后才能得到调度。 + 相同优先级任务支持时间片轮转调度方式。 + 共有32个优先级[0-31],最高优先级为0,最低优先级为31。 + + 任务状态通常分为以下四种: + 就绪(Ready):该任务在就绪队列中,只等待CPU。 + 运行(Running):该任务正在执行。 + 阻塞(Blocked):该任务不在就绪队列中。包含任务被挂起(suspend状态)、任务被延时(delay状态)、 + 任务正在等待信号量、读写队列或者等待事件等。 + 退出态(Dead):该任务运行结束,等待系统回收资源。 + + 任务状态迁移说明 + 就绪态→运行态 + 任务创建后进入就绪态,发生任务切换时,就绪队列中最高优先级的任务被执行, + 从而进入运行态,但此刻该任务依旧在就绪队列中。 + 运行态→阻塞态 + 正在运行的任务发生阻塞(挂起、延时、读信号量等)时,该任务会从就绪队列中删除, + 任务状态由运行态变成阻塞态,然后发生任务切换,运行就绪队列中最高优先级任务。 + 阻塞态→就绪态(阻塞态→运行态) + 阻塞的任务被恢复后(任务恢复、延时时间超时、读信号量超时或读到信号量等),此时被 + 恢复的任务会被加入就绪队列,从而由阻塞态变成就绪态;此时如果被恢复任务的优先级高于 + 正在运行任务的优先级,则会发生任务切换,该任务由就绪态变成运行态。 + 就绪态→阻塞态 + 任务也有可能在就绪态时被阻塞(挂起),此时任务状态由就绪态变为阻塞态,该任务 + 从就绪队列中删除,不会参与任务调度,直到该任务被恢复。 + 运行态→就绪态 + 有更高优先级任务创建或者恢复后,会发生任务调度,此刻就绪队列中最高优先级任务 + 变为运行态,那么原先运行的任务由运行态变为就绪态,依然在就绪队列中。 + 运行态→退出态 + 运行中的任务运行结束,任务状态由运行态变为退出态。退出态包含任务运行结束的正常退出状态 + 以及Invalid状态。例如,任务运行结束但是没有自删除,对外呈现的就是Invalid状态,即退出态。 + 阻塞态→退出态 + 阻塞的任务调用删除接口,任务状态由阻塞态变为退出态。 + + 主要术语 + 任务ID + 任务ID,在任务创建时通过参数返回给用户,是任务的重要标识。系统中的ID号是唯一的。用户可以 + 通过任务ID对指定任务进行任务挂起、任务恢复、查询任务名等操作。 + + 任务优先级 + 优先级表示任务执行的优先顺序。任务的优先级决定了在发生任务切换时即将要执行的任务, + 就绪队列中最高优先级的任务将得到执行。 + + 任务入口函数 + 新任务得到调度后将执行的函数。该函数由用户实现,在任务创建时,通过任务创建结构体设置。 + + 任务栈 + 每个任务都拥有一个独立的栈空间,我们称为任务栈。栈空间里保存的信息包含局部变量、寄存器、函数参数、函数返回地址等。 + + 任务上下文 + 任务在运行过程中使用的一些资源,如寄存器等,称为任务上下文。当这个任务挂起时,其他任务继续执行, + 可能会修改寄存器等资源中的值。如果任务切换时没有保存任务上下文,可能会导致任务恢复后出现未知错误。 + 因此,Huawei LiteOS在任务切换时会将切出任务的任务上下文信息,保存在自身的任务栈中,以便任务恢复后, + 从栈空间中恢复挂起时的上下文信息,从而继续执行挂起时被打断的代码。 + 任务控制块TCB + 每个任务都含有一个任务控制块(TCB)。TCB包含了任务上下文栈指针(stack pointer)、任务状态、 + 任务优先级、任务ID、任务名、任务栈大小等信息。TCB可以反映出每个任务运行情况。 + 任务切换 + 任务切换包含获取就绪队列中最高优先级任务、切出任务上下文保存、切入任务上下文恢复等动作。 + + 运作机制 + 用户创建任务时,系统会初始化任务栈,预置上下文。此外,系统还会将“任务入口函数” + 地址放在相应位置。这样在任务第一次启动进入运行态时,将会执行“任务入口函数”。 +*/ +LITE_OS_SEC_BSS LosTaskCB *g_taskCBArray;//任务池 128个 +LITE_OS_SEC_BSS LOS_DL_LIST g_losFreeTask;//空闲任务链表 +LITE_OS_SEC_BSS LOS_DL_LIST g_taskRecycleList;//回收任务链表 +LITE_OS_SEC_BSS UINT32 g_taskMaxNum;//任务最大个数 +LITE_OS_SEC_BSS UINT32 g_taskScheduled; /* one bit for each cores */ +LITE_OS_SEC_BSS EVENT_CB_S g_resourceEvent;//资源的事件 +>>>>>>> remotes/origin/main /* spinlock for task module, only available on SMP mode */ LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_taskSpin); @@ -82,7 +162,11 @@ STATIC VOID OsConsoleIDSetHook(UINT32 param1, UINT32 param2) __attribute__((weakref("OsSetConsoleID"))); /* temp task blocks for booting procedure */ +<<<<<<< HEAD LITE_OS_SEC_BSS STATIC LosTaskCB g_mainTask[LOSCFG_KERNEL_CORE_NUM]; +======= +LITE_OS_SEC_BSS STATIC LosTaskCB g_mainTask[LOSCFG_KERNEL_CORE_NUM];//启动引导过程中使用的临时任务 +>>>>>>> remotes/origin/main LosTaskCB *OsGetMainTask(VOID) { @@ -92,23 +176,38 @@ LosTaskCB *OsGetMainTask(VOID) VOID OsSetMainTask(VOID) { UINT32 i; +<<<<<<< HEAD CHAR *name = "osMain"; +======= + CHAR *name = "osMain";//任务名称 +>>>>>>> remotes/origin/main SchedParam schedParam = { 0 }; schedParam.policy = LOS_SCHED_RR; schedParam.basePrio = OS_PROCESS_PRIORITY_HIGHEST; schedParam.priority = OS_TASK_PRIORITY_LOWEST; +<<<<<<< HEAD for (i = 0; i < LOSCFG_KERNEL_CORE_NUM; i++) { g_mainTask[i].taskStatus = OS_TASK_STATUS_UNUSED; g_mainTask[i].taskID = LOSCFG_BASE_CORE_TSK_LIMIT; +======= + //为每个CPU core 设置mainTask + for (i = 0; i < LOSCFG_KERNEL_CORE_NUM; i++) { + g_mainTask[i].taskStatus = OS_TASK_STATUS_UNUSED; + g_mainTask[i].taskID = LOSCFG_BASE_CORE_TSK_LIMIT;//128 +>>>>>>> remotes/origin/main g_mainTask[i].processCB = OS_KERNEL_PROCESS_GROUP; #ifdef LOSCFG_KERNEL_SMP_LOCKDEP g_mainTask[i].lockDep.lockDepth = 0; g_mainTask[i].lockDep.waitLock = NULL; #endif (VOID)strncpy_s(g_mainTask[i].taskName, OS_TCB_NAME_LEN, name, OS_TCB_NAME_LEN - 1); +<<<<<<< HEAD LOS_ListInit(&g_mainTask[i].lockList); +======= + LOS_ListInit(&g_mainTask[i].lockList);//初始化任务锁链表,上面挂的是任务已申请到的互斥锁 +>>>>>>> remotes/origin/main (VOID)OsSchedParamInit(&g_mainTask[i], schedParam.policy, &schedParam, NULL); } } @@ -122,16 +221,25 @@ VOID OsSetMainTaskProcess(UINTPTR processCB) #endif } } +<<<<<<< HEAD LITE_OS_SEC_TEXT WEAK VOID OsIdleTask(VOID) { while (1) { WFI; +======= +//空闲任务,每个CPU都有自己的空闲任务 +LITE_OS_SEC_TEXT WEAK VOID OsIdleTask(VOID) +{ + while (1) {//只有一个死循环 + WFI;//WFI指令:arm core立即进入low-power standly state,进入休眠模式,等待中断 +>>>>>>> remotes/origin/main } } VOID OsTaskInsertToRecycleList(LosTaskCB *taskCB) { +<<<<<<< HEAD LOS_ListTailInsert(&g_taskRecycleList, &taskCB->pendList); } @@ -147,6 +255,26 @@ LITE_OS_SEC_TEXT_INIT VOID OsTaskJoinPostUnsafe(LosTaskCB *taskCB) taskCB->taskStatus |= OS_TASK_STATUS_EXIT; } +======= + LOS_ListTailInsert(&g_taskRecycleList, &taskCB->pendList);//将任务挂入回收链表,等待回收 +} +/* +查找task通过OS_TCB_FROM_PENDLIST来实现,相当于由LOS_DL_LIST找到LosTaskCB, +将这些和参数任务绑在一起的task唤醒 +*/ +LITE_OS_SEC_TEXT_INIT VOID OsTaskJoinPostUnsafe(LosTaskCB *taskCB) +{ + if (taskCB->taskStatus & OS_TASK_FLAG_PTHREAD_JOIN) {//join任务处理 + if (!LOS_ListEmpty(&taskCB->joinList)) {//joinList中的节点身上都有阻塞标签 + LosTaskCB *resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(taskCB->joinList)));//通过贴有JOIN标签链表的第一个节点找到Task + OsTaskWakeClearPendMask(resumedTask);//清除任务的挂起标记 + resumedTask->ops->wake(resumedTask); + } + } + taskCB->taskStatus |= OS_TASK_STATUS_EXIT;//贴上任务退出标签 +} +//挂起任务,任务进入等待链表,Join代表是支持通过的第一个任务去唤醒其他的任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsTaskJoinPendUnsafe(LosTaskCB *taskCB) { if (taskCB->taskStatus & OS_TASK_STATUS_INIT) { @@ -165,6 +293,7 @@ LITE_OS_SEC_TEXT UINT32 OsTaskJoinPendUnsafe(LosTaskCB *taskCB) return LOS_EINVAL; } +<<<<<<< HEAD LITE_OS_SEC_TEXT UINT32 OsTaskSetDetachUnsafe(LosTaskCB *taskCB) { @@ -172,6 +301,15 @@ LITE_OS_SEC_TEXT UINT32 OsTaskSetDetachUnsafe(LosTaskCB *taskCB) if (LOS_ListEmpty(&(taskCB->joinList))) { LOS_ListDelete(&(taskCB->joinList)); taskCB->taskStatus &= ~OS_TASK_FLAG_PTHREAD_JOIN; +======= +//任务设置分离模式Deatch和JOIN是一对有你没我的状态 +LITE_OS_SEC_TEXT UINT32 OsTaskSetDetachUnsafe(LosTaskCB *taskCB) +{ + if (taskCB->taskStatus & OS_TASK_FLAG_PTHREAD_JOIN) {//join状态时 + if (LOS_ListEmpty(&(taskCB->joinList))) {//joinlist中没有数据了 + LOS_ListDelete(&(taskCB->joinList));//所谓删除就是自己指向自己 + taskCB->taskStatus &= ~OS_TASK_FLAG_PTHREAD_JOIN;//去掉JOIN标签 +>>>>>>> remotes/origin/main return LOS_OK; } /* This error code has a special purpose and is not allowed to appear again on the interface */ @@ -181,25 +319,39 @@ LITE_OS_SEC_TEXT UINT32 OsTaskSetDetachUnsafe(LosTaskCB *taskCB) return LOS_EINVAL; } +<<<<<<< HEAD +======= +//初始化任务模块 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsTaskInit(UINTPTR processCB) { UINT32 index; UINT32 size; UINT32 ret; +<<<<<<< HEAD g_taskMaxNum = LOSCFG_BASE_CORE_TSK_LIMIT; size = (g_taskMaxNum + 1) * sizeof(LosTaskCB); +======= + g_taskMaxNum = LOSCFG_BASE_CORE_TSK_LIMIT;//任务池中最多默认128个,可谓铁打的任务池流水的线程 + size = (g_taskMaxNum + 1) * sizeof(LosTaskCB);//计算需分配内存总大小 +>>>>>>> remotes/origin/main /* * This memory is resident memory and is used to save the system resources * of task control block and will not be freed. */ +<<<<<<< HEAD g_taskCBArray = (LosTaskCB *)LOS_MemAlloc(m_aucSysMem0, size); +======= + g_taskCBArray = (LosTaskCB *)LOS_MemAlloc(m_aucSysMem0, size);//任务池常驻内存,不被释放 +>>>>>>> remotes/origin/main if (g_taskCBArray == NULL) { ret = LOS_ERRNO_TSK_NO_MEMORY; goto EXIT; } (VOID)memset_s(g_taskCBArray, size, 0, size); +<<<<<<< HEAD LOS_ListInit(&g_losFreeTask); LOS_ListInit(&g_taskRecycleList); for (index = 0; index < g_taskMaxNum; index++) { @@ -208,12 +360,26 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsTaskInit(UINTPTR processCB) g_taskCBArray[index].processCB = processCB; LOS_ListTailInsert(&g_losFreeTask, &g_taskCBArray[index].pendList); } +======= + LOS_ListInit(&g_losFreeTask);//初始化空闲任务链表 + LOS_ListInit(&g_taskRecycleList);//初始化回收任务链表 + for (index = 0; index < g_taskMaxNum; index++) {//任务挨个初始化 + g_taskCBArray[index].taskStatus = OS_TASK_STATUS_UNUSED;//默认未使用, + g_taskCBArray[index].taskID = index;//任务ID[0~g_taskMaxNum-1] + g_taskCBArray[index].processCB = processCB; + LOS_ListTailInsert(&g_losFreeTask, &g_taskCBArray[index].pendList);//通过pendlist节点插入空闲任务列表 + }//注意:这里挂的是pendList节点,可以取TCB也要通过OS_TCB_FROM-PENDLIST取 +>>>>>>> remotes/origin/main g_taskCBArray[index].taskStatus = OS_TASK_STATUS_UNUSED; g_taskCBArray[index].taskID = index; g_taskCBArray[index].processCB = processCB; +<<<<<<< HEAD ret = OsSchedInit(); +======= + ret = OsSchedInit();//调度器初始化 +>>>>>>> remotes/origin/main EXIT: if (ret != LOS_OK) { @@ -221,18 +387,27 @@ EXIT: } return ret; } +<<<<<<< HEAD +======= +//获取IdletaskId,每个CPU核都对Task进行了内部管理,做到真正的并行处理 +>>>>>>> remotes/origin/main UINT32 OsGetIdleTaskId(VOID) { return OsSchedRunqueueIdleGet()->taskID; } +<<<<<<< HEAD +======= +//创建一个空闲任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsIdleTaskCreate(UINTPTR processID) { UINT32 ret; TSK_INIT_PARAM_S taskInitParam; UINT32 idleTaskID; +<<<<<<< HEAD (VOID)memset_s((VOID *)(&taskInitParam), sizeof(TSK_INIT_PARAM_S), 0, sizeof(TSK_INIT_PARAM_S)); taskInitParam.pfnTaskEntry = (TSK_ENTRY_FUNC)OsIdleTask; taskInitParam.uwStackSize = LOSCFG_BASE_CORE_TSK_IDLE_STACK_SIZE; @@ -242,20 +417,39 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsIdleTaskCreate(UINTPTR processID) taskInitParam.processID = processID; #ifdef LOSCFG_KERNEL_SMP taskInitParam.usCpuAffiMask = CPUID_TO_AFFI_MASK(ArchCurrCpuid()); +======= + (VOID)memset_s((VOID *)(&taskInitParam), sizeof(TSK_INIT_PARAM_S), 0, sizeof(TSK_INIT_PARAM_S));//任务初始参数清零 + taskInitParam.pfnTaskEntry = (TSK_ENTRY_FUNC)OsIdleTask;//入口函数 + taskInitParam.uwStackSize = LOSCFG_BASE_CORE_TSK_IDLE_STACK_SIZE;//任务栈大小 2K + taskInitParam.pcName = "Idle";//任务名称叫pcNAME + taskInitParam.policy = LOS_SCHED_IDLE; + taskInitParam.usTaskPrio = OS_TASK_PRIORITY_LOWEST;//默认最低优先级31 + taskInitParam.processID = processID; +#ifdef LOSCFG_KERNEL_SMP + taskInitParam.usCpuAffiMask = CPUID_TO_AFFI_MASK(ArchCurrCpuid());//每个idle任务只在单独的CPU上运行 +>>>>>>> remotes/origin/main #endif ret = LOS_TaskCreateOnly(&idleTaskID, &taskInitParam); if (ret != LOS_OK) { return ret; } LosTaskCB *idleTask = OS_TCB_FROM_TID(idleTaskID); +<<<<<<< HEAD idleTask->taskStatus |= OS_TASK_FLAG_SYSTEM_TASK; +======= + idleTask->taskStatus |= OS_TASK_FLAG_SYSTEM_TASK;//标记为系统任务,idle任务是给CPU休息用的,当然是个系统任务 +>>>>>>> remotes/origin/main OsSchedRunqueueIdleInit(idleTask); return LOS_TaskResume(idleTaskID); } /* +<<<<<<< HEAD * Description : get id of current running task. +======= + * Description : get id of current running task. |获取当前CPU正在执行的任务ID +>>>>>>> remotes/origin/main * Return : task id */ LITE_OS_SEC_TEXT UINT32 LOS_CurTaskIDGet(VOID) @@ -267,7 +461,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_CurTaskIDGet(VOID) } return runTask->taskID; } +<<<<<<< HEAD +======= +//创建指定任务同步信号量 +>>>>>>> remotes/origin/main STATIC INLINE UINT32 TaskSyncCreate(LosTaskCB *taskCB) { #ifdef LOSCFG_KERNEL_SMP_TASK_SYNC @@ -280,7 +478,11 @@ STATIC INLINE UINT32 TaskSyncCreate(LosTaskCB *taskCB) #endif return LOS_OK; } +<<<<<<< HEAD +======= +//销毁指定任务同步信号量 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsTaskSyncDestroy(UINT32 syncSignal) { #ifdef LOSCFG_KERNEL_SMP_TASK_SYNC @@ -291,6 +493,10 @@ STATIC INLINE VOID OsTaskSyncDestroy(UINT32 syncSignal) } #ifdef LOSCFG_KERNEL_SMP +<<<<<<< HEAD +======= +//任务同步等待,通过信号量保持同步 +>>>>>>> remotes/origin/main STATIC INLINE UINT32 OsTaskSyncWait(const LosTaskCB *taskCB) { #ifdef LOSCFG_KERNEL_SMP_TASK_SYNC @@ -316,7 +522,11 @@ STATIC INLINE UINT32 OsTaskSyncWait(const LosTaskCB *taskCB) #endif } #endif +<<<<<<< HEAD +======= +//同步唤醒 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsTaskSyncWake(const LosTaskCB *taskCB) { #ifdef LOSCFG_KERNEL_SMP_TASK_SYNC @@ -338,6 +548,7 @@ STATIC INLINE VOID OsInsertTCBToFreeList(LosTaskCB *taskCB) taskCB->taskStatus = OS_TASK_STATUS_UNUSED; LOS_ListAdd(&g_losFreeTask, &taskCB->pendList); } +<<<<<<< HEAD STATIC VOID OsTaskKernelResourcesToFree(UINT32 syncSignal, UINTPTR topOfStack) { @@ -346,6 +557,16 @@ STATIC VOID OsTaskKernelResourcesToFree(UINT32 syncSignal, UINTPTR topOfStack) (VOID)LOS_MemFree((VOID *)m_aucSysMem1, (VOID *)topOfStack); } +======= +//释放任务在内核状态下占用的资源 +STATIC VOID OsTaskKernelResourcesToFree(UINT32 syncSignal, UINTPTR topOfStack) +{ + OsTaskSyncDestroy(syncSignal);//任务销毁,同步信息 + + (VOID)LOS_MemFree((VOID *)m_aucSysMem1, (VOID *)topOfStack);//释放内核态空间 +} +//释放任务资源 +>>>>>>> remotes/origin/main STATIC VOID OsTaskResourcesToFree(LosTaskCB *taskCB) { UINT32 syncSignal = LOSCFG_BASE_IPC_SEM_LIMIT; @@ -353,7 +574,11 @@ STATIC VOID OsTaskResourcesToFree(LosTaskCB *taskCB) UINTPTR topOfStack; #ifdef LOSCFG_KERNEL_VM +<<<<<<< HEAD if ((taskCB->taskStatus & OS_TASK_FLAG_USER_MODE) && (taskCB->userMapBase != 0)) { +======= + if ((taskCB->taskStatus & OS_TASK_FLAG_USER_MODE) && (taskCB->userMapBase != 0)) {//释放用户态栈 +>>>>>>> remotes/origin/main SCHEDULER_LOCK(intSave); UINT32 mapBase = (UINTPTR)taskCB->userMapBase; UINT32 mapSize = taskCB->userMapSize; @@ -363,7 +588,11 @@ STATIC VOID OsTaskResourcesToFree(LosTaskCB *taskCB) LosProcessCB *processCB = OS_PCB_FROM_TCB(taskCB); LOS_ASSERT(!(OsProcessVmSpaceGet(processCB) == NULL)); +<<<<<<< HEAD UINT32 ret = OsUnMMap(OsProcessVmSpaceGet(processCB), (UINTPTR)mapBase, mapSize); +======= + UINT32 ret = OsUnMMap(OsProcessVmSpaceGet(processCB), (UINTPTR)mapBase, mapSize);//解除映射 +>>>>>>> remotes/origin/main if ((ret != LOS_OK) && (mapBase != 0) && !OsProcessIsInit(processCB)) { PRINT_ERR("process(%u) unmmap user task(%u) stack failed! mapbase: 0x%x size :0x%x, error: %d\n", processCB->processID, taskCB->taskID, mapBase, mapSize, ret); @@ -375,36 +604,61 @@ STATIC VOID OsTaskResourcesToFree(LosTaskCB *taskCB) } #endif +<<<<<<< HEAD if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) { +======= + if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) {//任务还未使用情况 +>>>>>>> remotes/origin/main topOfStack = taskCB->topOfStack; taskCB->topOfStack = 0; #ifdef LOSCFG_KERNEL_SMP_TASK_SYNC syncSignal = taskCB->syncSignal; taskCB->syncSignal = LOSCFG_BASE_IPC_SEM_LIMIT; #endif +<<<<<<< HEAD OsTaskKernelResourcesToFree(syncSignal, topOfStack); SCHEDULER_LOCK(intSave); #ifdef LOSCFG_KERNEL_VM OsClearSigInfoTmpList(&(taskCB->sig)); +======= + OsTaskKernelResourcesToFree(syncSignal, topOfStack);//释放内核所占内存,即内核栈的栈空间 + + SCHEDULER_LOCK(intSave); +#ifdef LOSCFG_KERNEL_VM + OsClearSigInfoTmpList(&(taskCB->sig));//归还信号控制块的内存 +>>>>>>> remotes/origin/main #endif OsInsertTCBToFreeList(taskCB); SCHEDULER_UNLOCK(intSave); } return; } +<<<<<<< HEAD +======= +//批量回收任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT VOID OsTaskCBRecycleToFree(void) { UINT32 intSave; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD while (!LOS_ListEmpty(&g_taskRecycleList)) { LosTaskCB *taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&g_taskRecycleList)); LOS_ListDelete(&taskCB->pendList); SCHEDULER_UNLOCK(intSave); OsTaskResourcesToFree(taskCB); +======= + while (!LOS_ListEmpty(&g_taskRecycleList)) {//遍历回收链表 + LosTaskCB *taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&g_taskRecycleList));//取出任务 + LOS_ListDelete(&taskCB->pendList);//重置节点 + SCHEDULER_UNLOCK(intSave); + + OsTaskResourcesToFree(taskCB);//释放任务所占资源 +>>>>>>> remotes/origin/main SCHEDULER_LOCK(intSave); } @@ -414,7 +668,11 @@ LITE_OS_SEC_TEXT VOID OsTaskCBRecycleToFree(void) /* * Description : All task entry * Input : taskID --- The ID of the task to be run +<<<<<<< HEAD */ +======= + *///所有任务的入口函数,OsTaskEntry是new task OsTaskStackInit时指定的 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT VOID OsTaskEntry(UINT32 taskID) { LOS_ASSERT(!OS_TID_CHECK_INVALID(taskID)); @@ -424,6 +682,7 @@ LITE_OS_SEC_TEXT_INIT VOID OsTaskEntry(UINT32 taskID) * from interrupt and other cores. release task spinlock and enable * interrupt in sequence at the task entry. */ +<<<<<<< HEAD LOS_SpinUnlock(&g_taskSpin); (VOID)LOS_IntUnLock(); @@ -432,11 +691,25 @@ LITE_OS_SEC_TEXT_INIT VOID OsTaskEntry(UINT32 taskID) taskCB->args[2], taskCB->args[3]); /* 2 & 3: just for args array index */ if (!(taskCB->taskStatus & OS_TASK_FLAG_PTHREAD_JOIN)) { taskCB->joinRetval = 0; +======= + LOS_SpinUnlock(&g_taskSpin);//释放任务自旋锁 + (VOID)LOS_IntUnLock();//恢复中断 + + LosTaskCB *taskCB = OS_TCB_FROM_TID(taskID); + taskCB->joinRetval = taskCB->taskEntry(taskCB->args[0], taskCB->args[1],//调出任务的入口函数 + taskCB->args[2], taskCB->args[3]); /* 2 & 3: just for args array index */ + if (!(taskCB->taskStatus & OS_TASK_FLAG_PTHREAD_JOIN)) { + taskCB->joinRetval = 0;//结合数为0 +>>>>>>> remotes/origin/main } OsRunningTaskToExit(taskCB, 0); } +<<<<<<< HEAD +======= +//任务创建参数检查 +>>>>>>> remotes/origin/main STATIC UINT32 TaskCreateParamCheck(const UINT32 *taskID, TSK_INIT_PARAM_S *initParam) { UINT32 poolSize = OS_SYS_MEM_SIZE; @@ -455,6 +728,7 @@ STATIC UINT32 TaskCreateParamCheck(const UINT32 *taskID, TSK_INIT_PARAM_S *initP } } +<<<<<<< HEAD if (initParam->pfnTaskEntry == NULL) { return LOS_ERRNO_TSK_ENTRY_NULL; } @@ -468,17 +742,40 @@ STATIC UINT32 TaskCreateParamCheck(const UINT32 *taskID, TSK_INIT_PARAM_S *initP } if (initParam->uwStackSize == 0) { +======= + if (initParam->pfnTaskEntry == NULL) {//入口函数不能为空 + return LOS_ERRNO_TSK_ENTRY_NULL; + } + + if (initParam->usTaskPrio > OS_TASK_PRIORITY_LOWEST) {//优先级必须大于31 + return LOS_ERRNO_TSK_PRIOR_ERROR; + } + + if (initParam->uwStackSize > poolSize) {//希望申请的栈大小不能大于总池子 + return LOS_ERRNO_TSK_STKSZ_TOO_LARGE; + } + + if (initParam->uwStackSize == 0) {//任何任务都必须由内核态栈,所以uwStackSize不能为0 +>>>>>>> remotes/origin/main initParam->uwStackSize = LOSCFG_BASE_CORE_TSK_DEFAULT_STACK_SIZE; } initParam->uwStackSize = (UINT32)ALIGN(initParam->uwStackSize, LOSCFG_STACK_POINT_ALIGN_SIZE); +<<<<<<< HEAD if (initParam->uwStackSize < LOS_TASK_MIN_STACK_SIZE) { +======= + if (initParam->uwStackSize < LOS_TASK_MIN_STACK_SIZE) {//运行栈空间不能低于最低值 +>>>>>>> remotes/origin/main return LOS_ERRNO_TSK_STKSZ_TOO_SMALL; } return LOS_OK; } +<<<<<<< HEAD +======= +//任务栈(内核态)内存分配,由内核态进程空间提供,即KProcess进程空间 +>>>>>>> remotes/origin/main STATIC VOID TaskCBDeInit(LosTaskCB *taskCB) { UINT32 intSave; @@ -531,13 +828,21 @@ STATIC VOID TaskCBBaseInit(LosTaskCB *taskCB, const TSK_INIT_PARAM_S *initParam) LOS_ListInit(&taskCB->joinList); } +<<<<<<< HEAD LOS_ListInit(&taskCB->lockList); +======= + LOS_ListInit(&taskCB->lockList);//初始化互斥锁链表 +>>>>>>> remotes/origin/main SET_SORTLIST_VALUE(&taskCB->sortList, OS_SORT_LINK_INVALID_TIME); #ifdef LOSCFG_KERNEL_VM taskCB->futex.index = OS_INVALID_VALUE; #endif } +<<<<<<< HEAD +======= +//任务初始化 +>>>>>>> remotes/origin/main STATIC UINT32 TaskCBInit(LosTaskCB *taskCB, const TSK_INIT_PARAM_S *initParam) { UINT32 ret; @@ -546,7 +851,11 @@ STATIC UINT32 TaskCBInit(LosTaskCB *taskCB, const TSK_INIT_PARAM_S *initParam) LosSchedParam initSchedParam = {0}; UINT16 policy = (initParam->policy == LOS_SCHED_NORMAL) ? LOS_SCHED_RR : initParam->policy; +<<<<<<< HEAD TaskCBBaseInit(taskCB, initParam); +======= + TaskCBBaseInit(taskCB, initParam);//初始化任务的基本信息,task->stackPointer指向内核态栈sp位置,该位置存着任务初始上下文 +>>>>>>> remotes/origin/main schedParam.policy = policy; ret = OsProcessAddNewTask(initParam->processID, taskCB, &schedParam, &numCount); @@ -598,25 +907,43 @@ STATIC UINT32 TaskStackInit(LosTaskCB *taskCB, const TSK_INIT_PARAM_S *initParam #endif return LOS_OK; } +<<<<<<< HEAD +======= +//获取一个空闲TCB +>>>>>>> remotes/origin/main STATIC LosTaskCB *GetFreeTaskCB(VOID) { UINT32 intSave; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (LOS_ListEmpty(&g_losFreeTask)) { +======= + if (LOS_ListEmpty(&g_losFreeTask)) {//全局空闲task为空 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); PRINT_ERR("No idle TCB in the system!\n"); return NULL; } LosTaskCB *taskCB = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&g_losFreeTask)); +<<<<<<< HEAD LOS_ListDelete(LOS_DL_LIST_FIRST(&g_losFreeTask)); +======= + LOS_ListDelete(LOS_DL_LIST_FIRST(&g_losFreeTask));//从g_losFreeTask链表中摘除自己 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return taskCB; } +<<<<<<< HEAD +======= +/* +创建任务,并使该任务进入suspend状态,不对该状态进行调度。如果需要调度,可以调用LOS_TaskResume使该任务进入ready状态 +*/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskCreateOnly(UINT32 *taskID, TSK_INIT_PARAM_S *initParam) { UINT32 errRet = TaskCreateParamCheck(taskID, initParam); @@ -656,7 +983,11 @@ DEINIT_TCB: TaskCBDeInit(taskCB); return errRet; } +<<<<<<< HEAD +======= +//创建任务,并使该任务进入ready状态,如果就绪队列中没有更高优先级的任务,则运行该任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskCreate(UINT32 *taskID, TSK_INIT_PARAM_S *initParam) { UINT32 ret; @@ -670,7 +1001,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskCreate(UINT32 *taskID, TSK_INIT_PARAM_S *in return LOS_ERRNO_TSK_YIELD_IN_INT; } +<<<<<<< HEAD if (OsProcessIsUserMode(OsCurrProcessGet())) { +======= + if (OsProcessIsUserMode(OsCurrProcessGet())) {//当前进程为用户进程 +>>>>>>> remotes/origin/main initParam->processID = (UINTPTR)OsGetKernelInitProcess(); } else { initParam->processID = (UINTPTR)OsCurrProcessGet(); @@ -696,7 +1031,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskCreate(UINT32 *taskID, TSK_INIT_PARAM_S *in return LOS_OK; } +<<<<<<< HEAD +======= +//恢复挂起的任务,是该任务进入ready状态 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskResume(UINT32 taskID) { UINT32 intSave; @@ -711,7 +1050,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskResume(UINT32 taskID) SCHEDULER_LOCK(intSave); /* clear pending signal */ +<<<<<<< HEAD taskCB->signal &= ~SIGNAL_SUSPEND; +======= + taskCB->signal &= ~SIGNAL_SUSPEND;//清楚挂起信号 +>>>>>>> remotes/origin/main if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) { errRet = LOS_ERRNO_TSK_NOT_CREATED; @@ -737,6 +1080,7 @@ LOS_ERREND: } /* +<<<<<<< HEAD * Check if needs to do the suspend operation on the running task. * Return TRUE, if needs to do the suspension. * Return FALSE, if meets following circumstances: @@ -744,6 +1088,15 @@ LOS_ERREND: * 2. Do the suspension when preemption is disabled * 3. Do the suspension in hard-irq * then LOS_TaskSuspend will directly return with 'ret' value. +======= + * Check if needs to do the suspend operation on the running task.//检查是否需要对正在运行的任务执行挂起操作。 + * Return TRUE, if needs to do the suspension. //如果需要暂停,返回TRUE。 + * Return FALSE, if meets following circumstances: //如果满足一下情况,则返回FALSE: + * 1. Do the suspension across cores, if SMP is enabled //1.如果启用了SMP,则跨CPU核执行挂起操作 + * 2. Do the suspension when preemption is disabled //2.当禁用抢占时则挂起 + * 3. Do the suspension in hard-irq //3.在硬中断时则挂起 + * then LOS_TaskSuspend will directly return with 'ret' value. //那么LOS_taskssuspend将直接返回ret值。 +>>>>>>> remotes/origin/main */ LITE_OS_SEC_TEXT_INIT STATIC BOOL OsTaskSuspendCheckOnRun(LosTaskCB *taskCB, UINT32 *ret) { @@ -752,20 +1105,34 @@ LITE_OS_SEC_TEXT_INIT STATIC BOOL OsTaskSuspendCheckOnRun(LosTaskCB *taskCB, UIN #ifdef LOSCFG_KERNEL_SMP /* ASYNCHRONIZED. No need to do task lock checking */ +<<<<<<< HEAD if (taskCB->currCpu != ArchCurrCpuid()) { taskCB->signal = SIGNAL_SUSPEND; LOS_MpSchedule(taskCB->currCpu); +======= + if (taskCB->currCpu != ArchCurrCpuid()) {//跨CPU核的情况 + taskCB->signal = SIGNAL_SUSPEND; + LOS_MpSchedule(taskCB->currCpu);//task所属CPU执行调度 +>>>>>>> remotes/origin/main return FALSE; } #endif +<<<<<<< HEAD if (!OsPreemptableInSched()) { +======= + if (!OsPreemptableInSched()) {//不能抢占时 +>>>>>>> remotes/origin/main /* Suspending the current core's running task */ *ret = LOS_ERRNO_TSK_SUSPEND_LOCKED; return FALSE; } +<<<<<<< HEAD if (OS_INT_ACTIVE) { +======= + if (OS_INT_ACTIVE) {//正在硬抢断时 +>>>>>>> remotes/origin/main /* suspend running task in interrupt */ taskCB->signal = SIGNAL_SUSPEND; return FALSE; @@ -773,7 +1140,11 @@ LITE_OS_SEC_TEXT_INIT STATIC BOOL OsTaskSuspendCheckOnRun(LosTaskCB *taskCB, UIN return TRUE; } +<<<<<<< HEAD +======= +//任务暂停,参数可以不是当前任务,也就是说A任务可以让B任务处于阻塞状态,挂起指定的任务,然后切换任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsTaskSuspend(LosTaskCB *taskCB) { UINT32 errRet; @@ -786,14 +1157,23 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsTaskSuspend(LosTaskCB *taskCB) return LOS_ERRNO_TSK_ALREADY_SUSPENDED; } +<<<<<<< HEAD if ((tempStatus & OS_TASK_STATUS_RUNNING) && !OsTaskSuspendCheckOnRun(taskCB, &errRet)) { +======= + if ((tempStatus & OS_TASK_STATUS_RUNNING) &&//如果参数任务正在运行,注意多Cpu core情况,贴着正在运行标签的任务并不一定是当前CPU的执行任务, + !OsTaskSuspendCheckOnRun(taskCB, &errRet)) {//很有可能是别的CPU core在跑的任务 +>>>>>>> remotes/origin/main return errRet; } return taskCB->ops->suspend(taskCB); } +<<<<<<< HEAD +======= +//外部接口,对OsTaskSuspend的封装 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskSuspend(UINT32 taskID) { UINT32 intSave; @@ -813,7 +1193,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_TaskSuspend(UINT32 taskID) SCHEDULER_UNLOCK(intSave); return errRet; } +<<<<<<< HEAD +======= +//设置任务为不使用状态 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsTaskStatusUnusedSet(LosTaskCB *taskCB) { taskCB->taskStatus |= OS_TASK_STATUS_UNUSED; @@ -959,7 +1343,11 @@ LOS_ERREND: } return ret; } +<<<<<<< HEAD +======= +//任务延时等待,释放CPU,等待时间到期后该任务会重新进入ready状态 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_TaskDelay(UINT32 tick) { UINT32 intSave; @@ -989,7 +1377,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_TaskDelay(UINT32 tick) SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +//获取任务的优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT16 LOS_TaskPriGet(UINT32 taskID) { UINT32 intSave; @@ -1010,7 +1402,11 @@ LITE_OS_SEC_TEXT_MINOR UINT16 LOS_TaskPriGet(UINT32 taskID) SCHEDULER_UNLOCK(intSave); return param.priority; } +<<<<<<< HEAD +======= +//设置指定任务的优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskPriSet(UINT32 taskID, UINT16 taskPrio) { UINT32 intSave; @@ -1048,12 +1444,20 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskPriSet(UINT32 taskID, UINT16 taskPrio) } return LOS_OK; } +<<<<<<< HEAD +======= +//设置当前任务的优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_CurTaskPriSet(UINT16 taskPrio) { return LOS_TaskPriSet(OsCurrTaskGet()->taskID, taskPrio); } +<<<<<<< HEAD +======= +//当前任务释放CPU,并将其移到具有相同优先级的就绪任务队列的末尾。 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskYield(VOID) { UINT32 intSave; @@ -1099,7 +1503,11 @@ LITE_OS_SEC_TEXT_MINOR VOID LOS_TaskUnlock(VOID) LOS_Schedule(); } } +<<<<<<< HEAD +======= +//获取任务信息,给shell使用的 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskInfoGet(UINT32 taskID, TSK_INFO_S *taskInfo) { UINT32 intSave; @@ -1129,8 +1537,13 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskInfoGet(UINT32 taskID, TSK_INFO_S *taskInf taskCB->ops->schedParamGet(taskCB, ¶m); taskInfo->usTaskStatus = taskCB->taskStatus; taskInfo->usTaskPrio = param.priority; +<<<<<<< HEAD taskInfo->uwStackSize = taskCB->stackSize; taskInfo->uwTopOfStack = taskCB->topOfStack; +======= + taskInfo->uwStackSize = taskCB->stackSize;//内核态栈大小 + taskInfo->uwTopOfStack = taskCB->topOfStack;//内核态栈顶位置 +>>>>>>> remotes/origin/main taskInfo->uwEventMask = taskCB->eventMask; taskInfo->taskEvent = taskCB->taskEvent; taskInfo->pTaskMux = taskCB->taskMux; @@ -1141,16 +1554,28 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskInfoGet(UINT32 taskID, TSK_INFO_S *taskInf } taskInfo->acName[LOS_TASK_NAMELEN - 1] = '\0'; +<<<<<<< HEAD taskInfo->uwBottomOfStack = TRUNCATE(((UINTPTR)taskCB->topOfStack + taskCB->stackSize), OS_TASK_STACK_ADDR_ALIGN); taskInfo->uwCurrUsed = (UINT32)(taskInfo->uwBottomOfStack - taskInfo->uwSP); taskInfo->bOvf = OsStackWaterLineGet((const UINTPTR *)taskInfo->uwBottomOfStack, +======= + taskInfo->uwBottomOfStack = TRUNCATE(((UINTPTR)taskCB->topOfStack + taskCB->stackSize),//这里可以看出栈顶地址是高于栈顶 + OS_TASK_STACK_ADDR_ALIGN); + taskInfo->uwCurrUsed = (UINT32)(taskInfo->uwBottomOfStack - taskInfo->uwSP);//当前任务栈已使用了多少 + + taskInfo->bOvf = OsStackWaterLineGet((const UINTPTR *)taskInfo->uwBottomOfStack,//获取栈的使用情况 +>>>>>>> remotes/origin/main (const UINTPTR *)taskInfo->uwTopOfStack, &taskInfo->uwPeakUsed); SCHEDULER_UNLOCK(intSave); return LOS_OK; } +<<<<<<< HEAD +======= +//CPU亲和性(affinity)将任务绑定在指定CPU上,用于多核CPU情况,(该函数仅在SMP模式下支持) +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT BOOL OsTaskCpuAffiSetUnsafe(UINT32 taskID, UINT16 newCpuAffiMask, UINT16 *oldCpuAffiMask) { #ifdef LOSCFG_KERNEL_SMP @@ -1176,17 +1601,29 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskCpuAffiSet(UINT32 taskID, UINT16 cpuAffiMa UINT32 intSave; UINT16 currCpuMask; +<<<<<<< HEAD if (OS_TID_CHECK_INVALID(taskID)) { return LOS_ERRNO_TSK_ID_INVALID; } if (!(cpuAffiMask & LOSCFG_KERNEL_CPU_MASK)) { +======= + if (OS_TID_CHECK_INVALID(taskID)) {//检测taskid是否有效,task由task池分配,鸿蒙默认128个任务 ID范围[0:127] + return LOS_ERRNO_TSK_ID_INVALID; + } + + if (!(cpuAffiMask & LOSCFG_KERNEL_CPU_MASK)) {//检测cpu亲和力 +>>>>>>> remotes/origin/main return LOS_ERRNO_TSK_CPU_AFFINITY_MASK_ERR; } LosTaskCB *taskCB = OS_TCB_FROM_TID(taskID); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) { +======= + if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) {//贴有未使用标签的处理 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return LOS_ERRNO_TSK_NOT_CREATED; } @@ -1194,13 +1631,22 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_TaskCpuAffiSet(UINT32 taskID, UINT16 cpuAffiMa SCHEDULER_UNLOCK(intSave); if (needSched && OS_SCHEDULER_ACTIVE) { +<<<<<<< HEAD LOS_MpSchedule(currCpuMask); LOS_Schedule(); +======= + LOS_MpSchedule(currCpuMask);//发送信号调度信号给目标cpu + LOS_Schedule();//申请调度 +>>>>>>> remotes/origin/main } return LOS_OK; } +<<<<<<< HEAD +======= +//查询任务被绑在哪个cpu上 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT16 LOS_TaskCpuAffiGet(UINT32 taskID) { #ifdef LOSCFG_KERNEL_SMP @@ -1214,18 +1660,30 @@ LITE_OS_SEC_TEXT_MINOR UINT16 LOS_TaskCpuAffiGet(UINT32 taskID) LosTaskCB *taskCB = OS_TCB_FROM_TID(taskID); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) { +======= + if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) {//任务必须在使用 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return INVALID_CPU_AFFI_MASK; } +<<<<<<< HEAD cpuAffiMask = taskCB->cpuAffiMask; +======= + cpuAffiMask = taskCB->cpuAffiMask;//获取亲和力掩码 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return cpuAffiMask; #else (VOID)taskID; +<<<<<<< HEAD return 1; +======= + return 1;//单核情况直接返回1,0号cpu对应0x01 +>>>>>>> remotes/origin/main #endif } @@ -1235,14 +1693,22 @@ LITE_OS_SEC_TEXT_MINOR UINT16 LOS_TaskCpuAffiGet(UINT32 taskID) LITE_OS_SEC_TEXT_MINOR VOID OsTaskProcSignal(VOID) { UINT32 ret; +<<<<<<< HEAD +======= + //私有且不可中断,无需保护。这个任务在其他cpu核看到它时总是在运行,所以它在执行代码的同时也可以继续接收信号 +>>>>>>> remotes/origin/main /* * private and uninterruptable, no protection needed. * while this task is always running when others cores see it, * so it keeps receiving signals while follow code executing. */ LosTaskCB *runTask = OsCurrTaskGet(); +<<<<<<< HEAD if (runTask->signal == SIGNAL_NONE) { +======= + if (runTask->signal == SIGNAL_NONE) {//意思是其他cpu发起了要干掉你的信号 +>>>>>>> remotes/origin/main return; } @@ -1250,23 +1716,41 @@ LITE_OS_SEC_TEXT_MINOR VOID OsTaskProcSignal(VOID) /* * clear the signal, and do the task deletion. if the signaled task has been * scheduled out, then this deletion will wait until next run. +<<<<<<< HEAD */ runTask->signal = SIGNAL_NONE; +======= + *///如果发出信号的任务以出调度就绪队列,则此删除将等待下次运行 + runTask->signal = SIGNAL_NONE;//清除信号 +>>>>>>> remotes/origin/main ret = LOS_TaskDelete(runTask->taskID); if (ret != LOS_OK) { PRINT_ERR("Task proc signal delete task(%u) failed err:0x%x\n", runTask->taskID, ret); } +<<<<<<< HEAD } else if (runTask->signal & SIGNAL_SUSPEND) { runTask->signal &= ~SIGNAL_SUSPEND; +======= + } else if (runTask->signal & SIGNAL_SUSPEND) {//意思是其他cpu发起了要挂起你的信号 + runTask->signal &= ~SIGNAL_SUSPEND;//任务贴在被其他cpu挂起的标签 +>>>>>>> remotes/origin/main /* suspend killed task may fail, ignore the result */ (VOID)LOS_TaskSuspend(runTask->taskID); #ifdef LOSCFG_KERNEL_SMP +<<<<<<< HEAD } else if (runTask->signal & SIGNAL_AFFI) { runTask->signal &= ~SIGNAL_AFFI; /* priority queue has updated, notify the target cpu */ LOS_MpSchedule((UINT32)runTask->cpuAffiMask); +======= + } else if (runTask->signal & SIGNAL_AFFI) {//意思是下次调度其他cpu要媾和你 + runTask->signal &= ~SIGNAL_AFFI;//任务贴上被其他CPU媾和的标签 + + /* priority queue has updated, notify the target cpu */ + LOS_MpSchedule((UINT32)runTask->cpuAffiMask);//发生调度,此任务将移交给媾和CPU运行. +>>>>>>> remotes/origin/main #endif } } @@ -1339,7 +1823,11 @@ INT32 OsUserProcessOperatePermissionsCheck(const LosTaskCB *taskCB, UINTPTR proc return LOS_OK; } +<<<<<<< HEAD +======= +//创建任务之前,检查用户态任务栈的参数,是否地址在用户空间 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT STATIC UINT32 OsCreateUserTaskParamCheck(UINT32 processID, TSK_INIT_PARAM_S *param) { UserTaskParam *userParam = NULL; @@ -1349,6 +1837,7 @@ LITE_OS_SEC_TEXT_INIT STATIC UINT32 OsCreateUserTaskParamCheck(UINT32 processID, } userParam = ¶m->userParam; +<<<<<<< HEAD if ((processID == OS_INVALID_VALUE) && !LOS_IsUserAddress(userParam->userArea)) { return OS_INVALID_VALUE; } @@ -1361,13 +1850,31 @@ LITE_OS_SEC_TEXT_INIT STATIC UINT32 OsCreateUserTaskParamCheck(UINT32 processID, return OS_INVALID_VALUE; } +======= + if ((processID == OS_INVALID_VALUE) && !LOS_IsUserAddress(userParam->userArea)) {//堆地址必须在用户空间 + return OS_INVALID_VALUE; + } + + if (!LOS_IsUserAddress((UINTPTR)param->pfnTaskEntry)) {//入口函数必须在用户空间 + return OS_INVALID_VALUE; + } + //堆栈必须在用户空间 + if (userParam->userMapBase && !LOS_IsUserAddressRange(userParam->userMapBase, userParam->userMapSize)) { + return OS_INVALID_VALUE; + } + //检查堆,栈范围 +>>>>>>> remotes/origin/main if (!LOS_IsUserAddress(userParam->userSP)) { return OS_INVALID_VALUE; } return LOS_OK; } +<<<<<<< HEAD +======= +//创建一个用户态任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsCreateUserTask(UINTPTR processID, TSK_INIT_PARAM_S *initParam) { UINT32 taskID; @@ -1376,6 +1883,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsCreateUserTask(UINTPTR processID, TSK_INIT_PARAM_ INT32 policy; SchedParam param; +<<<<<<< HEAD ret = OsCreateUserTaskParamCheck(processID, initParam); if (ret != LOS_OK) { return ret; @@ -1388,6 +1896,20 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsCreateUserTask(UINTPTR processID, TSK_INIT_PARAM_ LosProcessCB *processCB = OsCurrProcessGet(); initParam->processID = (UINTPTR)processCB; initParam->consoleID = processCB->consoleID; +======= + ret = OsCreateUserTaskParamCheck(processID, initParam);//检查参数,堆栈,入口地址必须在用户空间 + if (ret != LOS_OK) { + return ret; + } + //这里可看出一个任务有两个栈,内核态栈(内核指定栈大小)和用户态栈(用户指定栈大小) + initParam->uwStackSize = OS_USER_TASK_SYSCALL_STACK_SIZE; + initParam->usTaskPrio = OS_TASK_PRIORITY_LOWEST;//设置最低优先级31级 + if (processID == OS_INVALID_VALUE) {//外面没指定进程ID的处理 + SCHEDULER_LOCK(intSave); + LosProcessCB *processCB = OsCurrProcessGet(); + initParam->processID = (UINTPTR)processCB; + initParam->consoleID = processCB->consoleID;//任务控制台ID归属 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); ret = LOS_GetProcessScheduler(processCB->processID, &policy, NULL); if (ret != LOS_OK) { @@ -1400,6 +1922,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsCreateUserTask(UINTPTR processID, TSK_INIT_PARAM_ initParam->deadlineUs = param.deadlineUs; initParam->periodUs = param.periodUs; } +<<<<<<< HEAD } else { initParam->policy = LOS_SCHED_RR; initParam->processID = processID; @@ -1407,13 +1930,26 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsCreateUserTask(UINTPTR processID, TSK_INIT_PARAM_ } ret = LOS_TaskCreateOnly(&taskID, initParam); +======= + } else {//进程已经创建 + initParam->policy = LOS_SCHED_RR;//调度方式为抢占式,注意鸿蒙不仅仅只支持抢占式调度方式 + initParam->processID = processID;//进程ID赋值 + initParam->consoleID = 0;//默认0号控制台 + } + + ret = LOS_TaskCreateOnly(&taskID, initParam);//只创建task实体,不申请调度 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return OS_INVALID_VALUE; } return taskID; } +<<<<<<< HEAD +======= +//获取任务的调度方式 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_GetTaskScheduler(INT32 taskID) { UINT32 intSave; @@ -1426,7 +1962,11 @@ LITE_OS_SEC_TEXT INT32 LOS_GetTaskScheduler(INT32 taskID) LosTaskCB *taskCB = OS_TCB_FROM_TID(taskID); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) { +======= + if (taskCB->taskStatus & OS_TASK_STATUS_UNUSED) {//任务不能是没有在使用 +>>>>>>> remotes/origin/main policy = -LOS_EINVAL; OS_GOTO_ERREND(); } @@ -1438,7 +1978,11 @@ LOS_ERREND: SCHEDULER_UNLOCK(intSave); return policy; } +<<<<<<< HEAD +======= +//设置任务的调度信息 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT INT32 LOS_SetTaskScheduler(INT32 taskID, UINT16 policy, UINT16 priority) { SchedParam param = { 0 }; @@ -1582,12 +2126,20 @@ UINT32 LOS_TaskDetach(UINT32 taskID) SCHEDULER_UNLOCK(intSave); return errRet; } +<<<<<<< HEAD +======= +//获取最大任务数 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_GetSystemTaskMaximum(VOID) { return g_taskMaxNum; } +<<<<<<< HEAD +======= +//任务池中最后一个 +>>>>>>> remotes/origin/main LosTaskCB *OsGetDefaultTaskCB(VOID) { return &g_taskCBArray[g_taskMaxNum]; @@ -1602,11 +2154,16 @@ LITE_OS_SEC_TEXT VOID OsWriteResourceEventUnsafe(UINT32 events) { (VOID)OsEventWriteUnsafe(&g_resourceEvent, events, FALSE, NULL); } +<<<<<<< HEAD +======= +//资源回收任务 +>>>>>>> remotes/origin/main STATIC VOID OsResourceRecoveryTask(VOID) { UINT32 ret; +<<<<<<< HEAD while (1) { ret = LOS_EventRead(&g_resourceEvent, OS_RESOURCE_EVENT_MASK, LOS_WAITMODE_OR | LOS_WAITMODE_CLR, LOS_WAIT_FOREVER); @@ -1619,27 +2176,56 @@ STATIC VOID OsResourceRecoveryTask(VOID) #ifdef LOSCFG_ENABLE_OOM_LOOP_TASK if (ret & OS_RESOURCE_EVENT_OOM) { (VOID)OomCheckProcess(); +======= + while (1) {//死循环,回收资源不存在退出情况,只要系统在运行资源就需要回收 + ret = LOS_EventRead(&g_resourceEvent, OS_RESOURCE_EVENT_MASK, + LOS_WAITMODE_OR | LOS_WAITMODE_CLR, LOS_WAIT_FOREVER);//读取资源事件 + if (ret & (OS_RESOURCE_EVENT_FREE | OS_RESOURCE_EVENT_OOM)) {//收到资源释放或内存异常情况 + OsTaskCBRecycleToFree(); + + OsProcessCBRecycleToFree();//回收进程到空闲进程池 + } + +#ifdef LOSCFG_ENABLE_OOM_LOOP_TASK//内存溢出检测任务开关 + if (ret & OS_RESOURCE_EVENT_OOM) {//触发了这个事件 + (VOID)OomCheckProcess();//检查进程的内存溢出情况 +>>>>>>> remotes/origin/main } #endif } } +<<<<<<< HEAD +======= +//创建一个回收资源的任务 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsResourceFreeTaskCreate(VOID) { UINT32 ret; UINT32 taskID; TSK_INIT_PARAM_S taskInitParam; +<<<<<<< HEAD ret = LOS_EventInit((PEVENT_CB_S)&g_resourceEvent); +======= + ret = LOS_EventInit((PEVENT_CB_S)&g_resourceEvent);//初始化资源事件 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return LOS_NOK; } (VOID)memset_s((VOID *)(&taskInitParam), sizeof(TSK_INIT_PARAM_S), 0, sizeof(TSK_INIT_PARAM_S)); +<<<<<<< HEAD taskInitParam.pfnTaskEntry = (TSK_ENTRY_FUNC)OsResourceRecoveryTask; taskInitParam.uwStackSize = OS_TASK_RESOURCE_STATIC_SIZE; taskInitParam.pcName = "ResourcesTask"; taskInitParam.usTaskPrio = OS_TASK_RESOURCE_FREE_PRIORITY; +======= + taskInitParam.pfnTaskEntry = (TSK_ENTRY_FUNC)OsResourceRecoveryTask;//入口函数 + taskInitParam.uwStackSize = OS_TASK_RESOURCE_STATIC_SIZE; + taskInitParam.pcName = "ResourcesTask"; + taskInitParam.usTaskPrio = OS_TASK_RESOURCE_FREE_PRIORITY;//5,优先级很高 +>>>>>>> remotes/origin/main ret = LOS_TaskCreate(&taskID, &taskInitParam); if (ret == LOS_OK) { OS_TCB_FROM_TID(taskID)->taskStatus |= OS_TASK_FLAG_NO_DELETE; @@ -1647,5 +2233,9 @@ LITE_OS_SEC_TEXT UINT32 OsResourceFreeTaskCreate(VOID) return ret; } +<<<<<<< HEAD LOS_MODULE_INIT(OsResourceFreeTaskCreate, LOS_INIT_LEVEL_KMOD_TASK); +======= +LOS_MODULE_INIT(OsResourceFreeTaskCreate, LOS_INIT_LEVEL_KMOD_TASK);//资源回收任务初始化 +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/core/los_tick.c b/src/kernel_liteos_a/kernel/base/core/los_tick.c index b5c2794e..45f030b6 100644 --- a/src/kernel_liteos_a/kernel/base/core/los_tick.c +++ b/src/kernel_liteos_a/kernel/base/core/los_tick.c @@ -37,30 +37,51 @@ #endif +<<<<<<< HEAD LITE_OS_SEC_DATA_INIT UINT32 g_sysClock; LITE_OS_SEC_DATA_INIT UINT32 g_tickPerSecond; LITE_OS_SEC_BSS DOUBLE g_cycle2NsScale; /* spinlock for task module */ LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_tickSpin); +======= +LITE_OS_SEC_DATA_INIT UINT32 g_sysClock; //系统时钟,绝大部分部件工作的时钟源,以及所有外设的始终来源 +LITE_OS_SEC_DATA_INIT UINT32 g_tickPerSecond; //每秒Tick数,harmony默认为每秒100次即10ms +LITE_OS_SEC_BSS DOUBLE g_cycle2NsScale; //将周期转为纳秒级 + +/* spinlock for task module */ +LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_tickSpin); //节拍器自旋锁 +>>>>>>> remotes/origin/main /* * Description : Tick interruption handler */ +<<<<<<< HEAD LITE_OS_SEC_TEXT VOID OsTickHandler(VOID) +======= +LITE_OS_SEC_TEXT VOID OsTickHandler(VOID)//节拍中断处理函数,harmony默认1ms触发一次 +>>>>>>> remotes/origin/main { #ifdef LOSCFG_SCHED_TICK_DEBUG OsSchedDebugRecordData(); #endif #ifdef LOSCFG_KERNEL_VDSO +<<<<<<< HEAD OsVdsoTimevalUpdate(); +======= + OsVdsoTimevalUpdate(); //更新vdso数据页时间,vsdo可以直接在用户进程空间绕过系统调用获取系统时间 +>>>>>>> remotes/origin/main #endif #ifdef LOSCFG_BASE_CORE_TICK_HW_TIME HalClockIrqClear(); /* diff from every platform */ #endif +<<<<<<< HEAD OsSchedTick(); +======= + OsSchedTick();//由时钟发起的调度 +>>>>>>> remotes/origin/main } diff --git a/src/kernel_liteos_a/kernel/base/include/los_container_pri.h b/src/kernel_liteos_a/kernel/base/include/los_container_pri.h index d7b72c70..c1c235c6 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_container_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_container_pri.h @@ -57,10 +57,17 @@ typedef enum { CONTAINER = 0, +<<<<<<< HEAD PID_CONTAINER, PID_CHILD_CONTAINER, UTS_CONTAINER, MNT_CONTAINER, +======= + PID_CONTAINER, //进程容器 + PID_CHILD_CONTAINER, //子进程容器 + UTS_CONTAINER, // + MNT_CONTAINER, //挂载容器 +>>>>>>> remotes/origin/main IPC_CONTAINER, USER_CONTAINER, TIME_CONTAINER, @@ -70,6 +77,7 @@ typedef enum { } ContainerType; typedef struct Container { +<<<<<<< HEAD Atomic rc; #ifdef LOSCFG_PID_CONTAINER struct PidContainer *pidContainer; @@ -93,6 +101,31 @@ typedef struct Container { #endif } Container; +======= + Atomic rc; //原子操作 +#ifdef LOSCFG_PID_CONTAINER + struct PidContainer *pidContainer; //进程容器 + struct PidContainer *pidForChildContainer;//进程的孩子容器 +#endif +#ifdef LOSCFG_UTS_CONTAINER + struct UtsContainer *utsContainer; // +#endif +#ifdef LOSCFG_MNT_CONTAINER + struct MntContainer *mntContainer; //挂载容器 +#endif +#ifdef LOSCFG_IPC_CONTAINER + struct IpcContainer *ipcContainer; //IPC容器 +#endif +#ifdef LOSCFG_TIME_CONTAINER + struct TimeContainer *timeContainer; //时间容器 + struct TimeContainer *timeForChildContainer; +#endif +#ifdef LOSCFG_NET_CONTAINER + struct NetContainer *netContainer; //网络容器 +#endif +} Container; +//容器数量上限 +>>>>>>> remotes/origin/main typedef struct TagContainerLimit { #ifdef LOSCFG_PID_CONTAINER UINT32 pidLimit; diff --git a/src/kernel_liteos_a/kernel/base/include/los_err_pri.h b/src/kernel_liteos_a/kernel/base/include/los_err_pri.h index 40a667ae..c47405e8 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_err_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_err_pri.h @@ -40,6 +40,7 @@ extern "C" { #endif /* __cplusplus */ #endif /* __cplusplus */ +<<<<<<< HEAD /** * @ingroup los_err * Define the error magic word. @@ -64,11 +65,17 @@ extern "C" { * * @see None. */ +======= + +#define OS_ERR_MAGIC_WORD 0xa1b2c3f8 + +>>>>>>> remotes/origin/main #define OS_RETURN_ERROR(errNo) do { \ (VOID)LOS_ErrHandle("os_unspecific_file", OS_ERR_MAGIC_WORD, errNo, 0, NULL); \ return errNo; \ } while (0) +<<<<<<< HEAD /** * @ingroup los_err * @brief Error handling macro capable of returning error codes. @@ -89,11 +96,15 @@ extern "C" { * * @see None. */ +======= + +>>>>>>> remotes/origin/main #define OS_RETURN_ERROR_P2(errLine, errNo) do { \ (VOID)LOS_ErrHandle("os_unspecific_file", errLine, errNo, 0, NULL); \ return errNo; \ } while (0) +<<<<<<< HEAD /** * @ingroup los_err * @brief Macro for jumping to error handler. @@ -112,6 +123,9 @@ extern "C" { * * @see None. */ +======= + +>>>>>>> remotes/origin/main #define OS_GOTO_ERR_HANDLER(errorNo) do { \ errNo = errorNo; \ errLine = OS_ERR_MAGIC_WORD; \ diff --git a/src/kernel_liteos_a/kernel/base/include/los_futex_pri.h b/src/kernel_liteos_a/kernel/base/include/los_futex_pri.h index 7671115b..bd150db2 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_futex_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_futex_pri.h @@ -33,15 +33,24 @@ #define _LOS_FUTEX_PRI_H #include "los_list.h" +<<<<<<< HEAD #define FUTEX_WAIT 0 #define FUTEX_WAKE 1 #define FUTEX_REQUEUE 3 #define FUTEX_WAKE_OP 5 +======= +#define FUTEX_WAIT 0 ///< 原子性的检查 uaddr 中计数器的值是否为 val,如果是则让任务休眠,直到 FUTEX_WAKE 或者超时(time-out)。 + //也就是把任务挂到 uaddr 相对应的等待队列上去。 +#define FUTEX_WAKE 1 ///< 最多唤醒 val 个等待在 uaddr 上任务。 +#define FUTEX_REQUEUE 3 ///< 调整指定锁在Futex表中的位置 +#define FUTEX_WAKE_OP 5 +>>>>>>> remotes/origin/main #define FUTEX_LOCK_PI 6 #define FUTEX_UNLOCK_PI 7 #define FUTEX_TRYLOCK_PI 8 #define FUTEX_WAIT_BITSET 9 +<<<<<<< HEAD #define FUTEX_PRIVATE 128 #define FUTEX_MASK 0x3U @@ -52,6 +61,19 @@ typedef struct { LOS_DL_LIST pendList; /* point to pendList in TCB struct */ LOS_DL_LIST queueList; /* thread list blocked by this lock */ LOS_DL_LIST futexList; /* point to the next FutexNode */ +======= +#define FUTEX_PRIVATE 128 //私有快锁(以虚拟地址进行哈希) +#define FUTEX_MASK 0x3U +/// 每个futex node对应一个被挂起的task ,key值唯一标识一把用户态锁,具有相同key值的node被queue_list串联起来表示被同一把锁阻塞的task队列。 +typedef struct { + UINTPTR key; /* private:uvaddr | 私有锁,用虚拟地址 shared:paddr | 共享锁,用物理地址*/ + UINT32 index; /* hash bucket index | 哈希桶索引 OsFutexKeyToIndex */ + UINT32 pid; /* private:process id shared:OS_INVALID(-1) | 私有锁:进程ID , 共享锁为 -1 */ + LOS_DL_LIST pendList; /* point to pendList in TCB struct | 指向 TCB 结构中的 pendList, 通过它找到任务*/ + LOS_DL_LIST queueList; /* thread list blocked by this lock | 挂等待这把锁的任务,其实这里挂到是FutexNode.queueList , + 通过 queueList 可以找到 pendList ,通过 pendList又可以找到真正的任务*/ + LOS_DL_LIST futexList; /* point to the next FutexNode | 下一把Futex锁*/ +>>>>>>> remotes/origin/main } FutexNode; extern UINT32 OsFutexInit(VOID); diff --git a/src/kernel_liteos_a/kernel/base/include/los_ipc_container_pri.h b/src/kernel_liteos_a/kernel/base/include/los_ipc_container_pri.h index 42ccbf3e..d36d15f3 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_ipc_container_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_ipc_container_pri.h @@ -42,6 +42,7 @@ typedef struct TagQueueCB LosQueueCB; typedef struct OsMux LosMux; typedef LosMux pthread_mutex_t; typedef struct ProcessCB LosProcessCB; +<<<<<<< HEAD typedef struct IpcContainer { Atomic rc; @@ -49,6 +50,15 @@ typedef struct IpcContainer { LOS_DL_LIST freeQueueList; fd_set queueFdSet; struct mqarray queueTable[LOSCFG_BASE_IPC_QUEUE_LIMIT]; +======= +//IPC容器 +typedef struct IpcContainer { + Atomic rc; + LosQueueCB *allQueue; //队列控制块(读写分离模式) + LOS_DL_LIST freeQueueList;//空闲队列链表 + fd_set queueFdSet; + struct mqarray queueTable[LOSCFG_BASE_IPC_QUEUE_LIMIT];//队列池 +>>>>>>> remotes/origin/main pthread_mutex_t mqueueMutex; struct mqpersonal *mqPrivBuf[MAX_MQ_FD]; struct shminfo shmInfo; diff --git a/src/kernel_liteos_a/kernel/base/include/los_memstat_pri.h b/src/kernel_liteos_a/kernel/base/include/los_memstat_pri.h index a2a9f581..4e056499 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_memstat_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_memstat_pri.h @@ -42,7 +42,11 @@ extern "C" { #endif /* __cplusplus */ typedef struct { +<<<<<<< HEAD UINT32 memUsed; +======= + UINT32 memUsed; ///< 记录任务内存使用量 +>>>>>>> remotes/origin/main } TskMemUsedInfo; extern VOID OsTaskMemUsedInc(UINT32 usedSize, UINT32 taskID); @@ -53,7 +57,11 @@ extern VOID OsTaskMemClear(UINT32 taskID); #ifdef LOS_MEM_SLAB typedef struct { +<<<<<<< HEAD UINT32 slabUsed; +======= + UINT32 slabUsed; ///< 任务占用以slab分配方式内存量 +>>>>>>> remotes/origin/main } TskSlabUsedInfo; extern VOID OsTaskSlabUsedInc(UINT32 usedSize, UINT32 taskID); diff --git a/src/kernel_liteos_a/kernel/base/include/los_mux_pri.h b/src/kernel_liteos_a/kernel/base/include/los_mux_pri.h index 8837114c..90b4afb7 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_mux_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_mux_pri.h @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: diff --git a/src/kernel_liteos_a/kernel/base/include/los_percpu_pri.h b/src/kernel_liteos_a/kernel/base/include/los_percpu_pri.h index 8a824ffd..aeacd772 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_percpu_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_percpu_pri.h @@ -43,9 +43,15 @@ extern "C" { #ifdef LOSCFG_KERNEL_SMP typedef enum { +<<<<<<< HEAD CPU_RUNNING = 0, /* cpu is running */ CPU_HALT, /* cpu in the halt */ CPU_EXC /* cpu in the exc */ +======= + CPU_RUNNING = 0, ///< cpu is running | CPU正在运行状态 + CPU_HALT, ///< cpu in the halt | CPU处于暂停状态 + CPU_EXC ///< cpu in the exc | CPU处于异常状态 +>>>>>>> remotes/origin/main } ExcFlag; typedef struct { @@ -55,6 +61,7 @@ typedef struct { #endif } Percpu; +<<<<<<< HEAD /* the kernel per-cpu structure */ extern Percpu g_percpu[LOSCFG_KERNEL_CORE_NUM]; @@ -63,6 +70,16 @@ STATIC INLINE Percpu *OsPercpuGet(VOID) return &g_percpu[ArchCurrCpuid()]; } +======= +/*! the kernel per-cpu structure | 每个cpu的内核描述符 */ +extern Percpu g_percpu[LOSCFG_KERNEL_CORE_NUM]; +/*! 获得当前运行CPU的信息 */ +STATIC INLINE Percpu *OsPercpuGet(VOID) +{ + return &g_percpu[ArchCurrCpuid()]; +} +/*! 获得参数CPU的信息 */ +>>>>>>> remotes/origin/main STATIC INLINE Percpu *OsPercpuGetByID(UINT32 cpuid) { return &g_percpu[cpuid]; diff --git a/src/kernel_liteos_a/kernel/base/include/los_pid_container_pri.h b/src/kernel_liteos_a/kernel/base/include/los_pid_container_pri.h index cc047e68..44ef9ab9 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_pid_container_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_pid_container_pri.h @@ -38,6 +38,7 @@ typedef struct TagTaskCB LosTaskCB; typedef struct ProcessCB LosProcessCB; struct ProcessGroup; struct Container; +<<<<<<< HEAD typedef struct { UINT32 vid; /* Virtual ID */ @@ -61,6 +62,31 @@ typedef struct PidContainer { ProcessVid tidArray[LOSCFG_BASE_CORE_TSK_LIMIT]; LOS_DL_LIST pidFreeList; ProcessVid pidArray[LOSCFG_BASE_CORE_PROCESS_LIMIT]; +======= +//虚拟进程/任务 信息 +typedef struct { + UINT32 vid; /* Virtual ID | 虚拟ID*/ + UINT32 vpid; /* Virtual parent ID | 父进程虚拟ID*/ + UINTPTR cb; /* Control block | 控制块*/ + LosProcessCB *realParent; /* process real parent | 进程真实的父进程 */ + LOS_DL_LIST node;//用于挂入 PidContainer.pidFreeList | tidFreeList +} ProcessVid; + +#define PID_CONTAINER_LEVEL_LIMIT 3 +//进程容器 +typedef struct PidContainer { + Atomic rc; //原子操作 + Atomic level; //等级,0为最高级,父比子高一级 + Atomic lock; //锁 + BOOL referenced; //是否被引用 + UINT32 containerID; //容器ID + struct PidContainer *parent; //父进程容器 + struct ProcessGroup *rootPGroup; //进程组 + LOS_DL_LIST tidFreeList; //任务空闲链表 + ProcessVid tidArray[LOSCFG_BASE_CORE_TSK_LIMIT];//虚拟任务池 + LOS_DL_LIST pidFreeList; //进程空闲链表 + ProcessVid pidArray[LOSCFG_BASE_CORE_PROCESS_LIMIT];//虚拟进程池 +>>>>>>> remotes/origin/main } PidContainer; #define OS_PID_CONTAINER_FROM_PCB(processCB) ((processCB)->container->pidContainer) diff --git a/src/kernel_liteos_a/kernel/base/include/los_process_pri.h b/src/kernel_liteos_a/kernel/base/include/los_process_pri.h index accd618c..f22a76a5 100644 --- a/src/kernel_liteos_a/kernel/base/include/los_process_pri.h +++ b/src/kernel_liteos_a/kernel/base/include/los_process_pri.h @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -65,6 +69,7 @@ extern "C" { #ifdef LOSCFG_SECURITY_CAPABILITY #define OS_GROUPS_NUMBER_MAX 256 +<<<<<<< HEAD typedef struct { UINT32 userID; UINT32 effUserID; @@ -141,6 +146,88 @@ typedef struct ProcessCB { #endif #ifdef LOSCFG_PROC_PROCESS_DIR struct ProcDirEntry *procDir; +======= +/*! 用户描述体*/ +typedef struct { + UINT32 userID; ///<用户ID [0,60000],0为root用户 + UINT32 effUserID; + UINT32 gid; ///<用户组ID [0,60000],0为root用户组 + UINT32 effGid; + UINT32 groupNumber;///< 用户组数量 + UINT32 groups[1]; //所属用户组列表,一个用户可属多个用户组 +} User; +#endif +/*! 进程组结构体*/ +typedef struct ProcessGroup { + UINTPTR pgroupLeader; /**< Process group leader is the the process that created the group | 负责创建进程组的进程首地址*/ + LOS_DL_LIST processList; /**< List of processes under this process group | 属于该进程组的进程链表*/ + LOS_DL_LIST exitProcessList; /**< List of closed processes (zombie processes) under this group | 进程组的僵死进程链表*/ + LOS_DL_LIST groupList; /**< Process group list | 进程组链表,上面挂的都是进程组*/ +} ProcessGroup; + +/** + * 进程控制块. + */ +typedef struct ProcessCB { + CHAR processName[OS_PCB_NAME_LEN]; /**< Process name | 进程名称 */ + UINT32 processID; /**< Process ID = leader thread ID | 进程ID,由进程池分配,范围[0,64] */ + UINT16 processStatus; /**< [15:4] Process Status; [3:0] The number of threads currently + running in the process | 这里设计很巧妙.用一个变量表示了两层逻辑 数量和状态,点赞! @note_good 从这里也可以看出一个进程可以有多个正在运行的任务*/ + UINT16 consoleID; /**< The console id of task belongs | 任务的控制台id归属 */ + UINT16 processMode; /**< Kernel Mode:0; User Mode:1; | 模式指定为内核还是用户进程 */ + struct ProcessCB *parentProcess; /**< Parent process */ + UINT32 exitCode; /**< Process exit status | 进程退出状态码*/ + LOS_DL_LIST pendList; /**< Block list to which the process belongs | 进程所在的阻塞列表,进程因阻塞挂入相应的链表.*/ + LOS_DL_LIST childrenList; /**< Children process list | 孩子进程都挂到这里,形成双循环链表*/ + LOS_DL_LIST exitChildList; /**< Exit children process list | 要退出的孩子进程链表,白发人要送黑发人.*/ + LOS_DL_LIST siblingList; /**< Linkage in parent's children list | 兄弟进程链表, 56个民族是一家,来自同一个父进程.*/ + ProcessGroup *pgroup; /**< Process group to which a process belongs | 所属进程组*/ + LOS_DL_LIST subordinateGroupList; /**< Linkage in group list | 进程组员链表*/ + LosTaskCB *threadGroup; + LOS_DL_LIST threadSiblingList; /**< List of threads under this process | 进程的线程(任务)列表 */ + volatile UINT32 threadNumber; /**< Number of threads alive under this process | 此进程下的活动线程数*/ + UINT32 threadCount; /**< Total number of threads created under this process | 在此进程下创建的线程总数*/ // + LOS_DL_LIST waitList; /**< The process holds the waitLits to support wait/waitpid | 父进程通过进程等待的方式,回收子进程资源,获取子进程退出信息*/ +#ifdef LOSCFG_KERNEL_SMP + UINT32 timerCpu; /**< CPU core number of this task is delayed or pended | 统计各线程被延期或阻塞的时间*/ +#endif + UINTPTR sigHandler; /**< Signal handler | 信号处理函数,处理如 SIGSYS 等信号*/ + sigset_t sigShare; /**< Signal share bit | 信号共享位 sigset_t是个64位的变量,对应64种信号*/ +#ifdef LOSCFG_KERNEL_LITEIPC + ProcIpcInfo *ipcInfo; /**< Memory pool for lite ipc | 用于进程间通讯的虚拟设备文件系统,设备装载点为 /dev/lite_ipc*/ +#endif +#ifdef LOSCFG_KERNEL_VM + LosVmSpace *vmSpace; /**< VMM space for processes | 虚拟空间,描述进程虚拟内存的数据结构,linux称为内存描述符 */ +#endif +#ifdef LOSCFG_FS_VFS + struct files_struct *files; /**< Files held by the process | 进程所持有的所有文件,注者称之为进程的文件管理器*/ +#endif //每个进程都有属于自己的文件管理器,记录对文件的操作. 注意:一个文件可以被多个进程操作 + timer_t timerID; /**< iTimer */ + +#ifdef LOSCFG_SECURITY_CAPABILITY //安全能力 + User *user; ///< 进程的拥有者 + UINT32 capability; ///< 安全能力范围 对应 CAP_SETGID +#endif +#ifdef LOSCFG_SECURITY_VID //虚拟ID映射功能 + TimerIdMap timerIdMap; +#endif +#ifdef LOSCFG_DRIVERS_TZDRIVER + struct Vnode *execVnode; /**< Exec bin of the process | 进程的可执行文件 */ +#endif + mode_t umask; ///< umask(user file-creatiopn mode mask)为用户文件创建掩码,是创建文件或文件夹时默认权限的基础。 +#ifdef LOSCFG_KERNEL_CPUP + OsCpupBase *processCpup; /**< Process cpu usage | 进程占用CPU情况统计*/ +#endif + struct rlimit *resourceLimit; ///< 每个进程在运行时系统不会无限制的允许单个进程不断的消耗资源,因此都会设置资源限制。 +#ifdef LOSCFG_KERNEL_CONTAINER + Container *container; ///< 内核容器 +#ifdef LOSCFG_USER_CONTAINER + struct Credentials *credentials; ///< 用户身份证 +#endif +#endif +#ifdef LOSCFG_PROC_PROCESS_DIR + struct ProcDirEntry *procDir; ///< 目录文件项 +>>>>>>> remotes/origin/main #endif #ifdef LOSCFG_KERNEL_PLIMITS ProcLimiterSet *plimits; @@ -161,8 +248,13 @@ extern UINT32 g_processMaxNum; #define OS_PCB_FROM_TCB(taskCB) ((LosProcessCB *)((taskCB)->processCB)) #define OS_PCB_FROM_TID(taskID) ((LosProcessCB *)(OS_TCB_FROM_TID(taskID)->processCB)) #define OS_GET_PGROUP_LEADER(pgroup) ((LosProcessCB *)((pgroup)->pgroupLeader)) +<<<<<<< HEAD #define OS_PCB_FROM_SIBLIST(ptr) LOS_DL_LIST_ENTRY((ptr), LosProcessCB, siblingList) #define OS_PCB_FROM_PENDLIST(ptr) LOS_DL_LIST_ENTRY((ptr), LosProcessCB, pendList) +======= +#define OS_PCB_FROM_SIBLIST(ptr) LOS_DL_LIST_ENTRY((ptr), LosProcessCB, siblingList)///< 通过siblingList节点找到 LosProcessCB +#define OS_PCB_FROM_PENDLIST(ptr) LOS_DL_LIST_ENTRY((ptr), LosProcessCB, pendList) ///< 通过pendlist节点找到 LosProcessCB +>>>>>>> remotes/origin/main /** * @ingroup los_process @@ -202,7 +294,11 @@ extern UINT32 g_processMaxNum; * * The process is run out but the resources occupied by the process are not recovered. */ +<<<<<<< HEAD #define OS_PROCESS_STATUS_ZOMBIES 0x0100U +======= +#define OS_PROCESS_STATUS_ZOMBIES 0x0100U ///< 进程状态: 僵死 +>>>>>>> remotes/origin/main /** * @ingroup los_process @@ -211,7 +307,11 @@ extern UINT32 g_processMaxNum; * The process status equal this is process control block unused, * coexisting with OS_PROCESS_STATUS_ZOMBIES means that the control block is not recovered. */ +<<<<<<< HEAD #define OS_PROCESS_FLAG_UNUSED 0x0200U +======= +#define OS_PROCESS_FLAG_UNUSED 0x0200U ///< 进程未使用标签,一般用于进程的初始状态 freelist里面都是这种标签 +>>>>>>> remotes/origin/main /** * @ingroup los_process @@ -219,7 +319,11 @@ extern UINT32 g_processMaxNum; * * The process has been call exit, it only works with multiple cores. */ +<<<<<<< HEAD #define OS_PROCESS_FLAG_EXIT 0x0400U +======= +#define OS_PROCESS_FLAG_EXIT 0x0400U ///< 进程退出标签,退出的进程进入回收链表等待回收资源 +>>>>>>> remotes/origin/main /** * @ingroup los_process @@ -227,7 +331,11 @@ extern UINT32 g_processMaxNum; * * The process is the leader of the process group. */ +<<<<<<< HEAD #define OS_PROCESS_FLAG_GROUP_LEADER 0x0800U +======= +#define OS_PROCESS_FLAG_GROUP_LEADER 0x0800U ///< 进程当了进程组领导标签 +>>>>>>> remotes/origin/main /** * @ingroup los_process @@ -235,21 +343,34 @@ extern UINT32 g_processMaxNum; * * The process has performed the exec operation. */ +<<<<<<< HEAD #define OS_PROCESS_FLAG_ALREADY_EXEC 0x1000U +======= +#define OS_PROCESS_FLAG_ALREADY_EXEC 0x1000U ///< 进程已执行exec操作 load elf时使用 +>>>>>>> remotes/origin/main /** * @ingroup los_process * Flag that indicates the process or process control block status. * * The process is dying or already dying. +<<<<<<< HEAD */ #define OS_PROCESS_STATUS_INACTIVE (OS_PROCESS_FLAG_EXIT | OS_PROCESS_STATUS_ZOMBIES) +======= + */ /// 进程不活跃状态定义: 身上贴有退出便签且状态为僵死的进程 +#define OS_PROCESS_STATUS_INACTIVE (OS_PROCESS_FLAG_EXIT | OS_PROCESS_STATUS_ZOMBIES) +>>>>>>> remotes/origin/main /** * @ingroup los_process * Used to check if the process control block is unused. */ +<<<<<<< HEAD STATIC INLINE BOOL OsProcessIsUnused(const LosProcessCB *processCB) +======= +STATIC INLINE BOOL OsProcessIsUnused(const LosProcessCB *processCB)//查下进程是否还在使用? +>>>>>>> remotes/origin/main { return ((processCB->processStatus & OS_PROCESS_FLAG_UNUSED) != 0); } @@ -257,8 +378,13 @@ STATIC INLINE BOOL OsProcessIsUnused(const LosProcessCB *processCB) /** * @ingroup los_process * Used to check if the process is inactive. +<<<<<<< HEAD */ STATIC INLINE BOOL OsProcessIsInactive(const LosProcessCB *processCB) +======= + */ /// 进程不活跃函数定义:身上贴有不使用且不活跃标签的进程 +STATIC INLINE BOOL OsProcessIsInactive(const LosProcessCB *processCB)//查下进程是否不活跃? +>>>>>>> remotes/origin/main { return ((processCB->processStatus & (OS_PROCESS_FLAG_UNUSED | OS_PROCESS_STATUS_INACTIVE)) != 0); } @@ -266,8 +392,13 @@ STATIC INLINE BOOL OsProcessIsInactive(const LosProcessCB *processCB) /** * @ingroup los_process * Used to check if the process is dead. +<<<<<<< HEAD */ STATIC INLINE BOOL OsProcessIsDead(const LosProcessCB *processCB) +======= + */ /// 进程死啦死啦的定义: 身上贴有不使用且状态为僵死的进程 +STATIC INLINE BOOL OsProcessIsDead(const LosProcessCB *processCB)//查下进程是否死啦死啦滴? +>>>>>>> remotes/origin/main { return ((processCB->processStatus & OS_PROCESS_STATUS_ZOMBIES) != 0); } @@ -282,6 +413,7 @@ STATIC INLINE BOOL OsProcessIsPGroupLeader(const LosProcessCB *processCB) return ((processCB->processStatus & OS_PROCESS_FLAG_GROUP_LEADER) != 0); } +<<<<<<< HEAD /** * @ingroup los_process * The highest priority of a kernel mode process. @@ -336,14 +468,52 @@ STATIC INLINE BOOL OsProcessIsPGroupLeader(const LosProcessCB *processCB) #define OS_KERNEL_MODE 0x0U #define OS_USER_MODE 0x1U +======= + +#define OS_PROCESS_PRIORITY_HIGHEST 0 ///< 进程最高优先级 + + +#define OS_PROCESS_PRIORITY_LOWEST 31 ///< 进程最低优先级 + + +#define OS_USER_PROCESS_PRIORITY_HIGHEST 10 ///< 内核模式和用户模式的优先级分割线 10-31 用户级, 0-9内核级 + + +#define OS_USER_PROCESS_PRIORITY_LOWEST OS_PROCESS_PRIORITY_LOWEST ///< 用户进程的最低优先级 + + +#define OS_PROCESS_USERINIT_PRIORITY 28 ///< 用户进程默认的优先级,28级好低啊 + + +#define OS_KERNEL_IDLE_PROCESS_ID 0U //0号进程为空闲进程 + + +#define OS_USER_ROOT_PROCESS_ID 1U //1号为用户态根进程 + + +#define OS_KERNEL_ROOT_PROCESS_ID 2U //1号为内核态根进程 +#define OS_TASK_DEFAULT_STACK_SIZE 0x2000 ///< task默认栈大小 8K +#define OS_USER_TASK_SYSCALL_STACK_SIZE 0x3000 ///< 用户通过系统调用的栈大小 12K ,这时是运行在内核模式下 +#define OS_USER_TASK_STACK_SIZE 0x100000 ///< 用户任务运行在用户空间的栈大小 1M + +#define OS_KERNEL_MODE 0x0U ///< 内核态 +#define OS_USER_MODE 0x1U ///< 用户态 +/*! 用户态进程*/ +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsProcessIsUserMode(const LosProcessCB *processCB) { return (processCB->processMode == OS_USER_MODE); } +<<<<<<< HEAD #define LOS_PRIO_PROCESS 0U #define LOS_PRIO_PGRP 1U #define LOS_PRIO_USER 2U +======= +#define LOS_PRIO_PROCESS 0U ///< 进程标识 +#define LOS_PRIO_PGRP 1U ///< 进程组标识 +#define LOS_PRIO_USER 2U ///< 用户标识 +>>>>>>> remotes/origin/main #define OS_USER_PRIVILEGE_PROCESS_GROUP ((UINTPTR)OsGetUserInitProcess()) #define OS_KERNEL_PROCESS_GROUP ((UINTPTR)OsGetKernelInitProcess()) @@ -353,6 +523,7 @@ STATIC INLINE BOOL OsProcessIsUserMode(const LosProcessCB *processCB) * 31 15 8 7 0 * | | exit code | core dump | signal | */ +<<<<<<< HEAD #define OS_PRO_EXIT_OK 0 STATIC INLINE VOID OsProcessExitCodeCoreDumpSet(LosProcessCB *processCB) @@ -370,23 +541,54 @@ STATIC INLINE VOID OsProcessExitCodeSignalClear(LosProcessCB *processCB) processCB->exitCode &= (~0x7FU); } +======= +#define OS_PRO_EXIT_OK 0 ///< 进程正常退出 +/// 置进程退出码第七位为1 +STATIC INLINE VOID OsProcessExitCodeCoreDumpSet(LosProcessCB *processCB) +{ + processCB->exitCode |= 0x80U; // 0b10000000 +} +/// 设置进程退出信号(0 ~ 7) +STATIC INLINE VOID OsProcessExitCodeSignalSet(LosProcessCB *processCB, UINT32 signal) +{ + processCB->exitCode |= signal & 0x7FU;// 0b01111111 +} +/// 清除进程退出信号(0 ~ 7) +STATIC INLINE VOID OsProcessExitCodeSignalClear(LosProcessCB *processCB) +{ + processCB->exitCode &= (~0x7FU);// 低7位全部清0 +} +/// 进程退出码是否被设置过,默认是 0 ,如果 & 0x7FU 还是 0 ,说明没有被设置过. +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsProcessExitCodeSignalIsSet(LosProcessCB *processCB) { return (processCB->exitCode) & 0x7FU; } +<<<<<<< HEAD +======= +/// 设置进程退出号(8 ~ 15) +>>>>>>> remotes/origin/main STATIC INLINE VOID OsProcessExitCodeSet(LosProcessCB *processCB, UINT32 code) { processCB->exitCode |= ((code & 0x000000FFU) << 8U) & 0x0000FF00U; /* 8: Move 8 bits to the left, exitCode */ } #define OS_PID_CHECK_INVALID(pid) (((UINT32)(pid)) >= g_processMaxNum) +<<<<<<< HEAD +======= +/*! 内联函数 进程ID是否有效 */ +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsProcessIDUserCheckInvalid(UINT32 pid) { return ((pid >= g_processMaxNum) || (pid == 0)); } +<<<<<<< HEAD +======= +/*! 获取当前进程PCB */ +>>>>>>> remotes/origin/main STATIC INLINE LosProcessCB *OsCurrProcessGet(VOID) { UINT32 intSave; @@ -398,6 +600,10 @@ STATIC INLINE LosProcessCB *OsCurrProcessGet(VOID) } #ifdef LOSCFG_SECURITY_CAPABILITY +<<<<<<< HEAD +======= +/*! 获取当前进程的所属用户 */ +>>>>>>> remotes/origin/main STATIC INLINE User *OsCurrUserGet(VOID) { User *user = NULL; @@ -469,14 +675,22 @@ STATIC INLINE UINT32 OsGetRootPid(const LosProcessCB *processCB) /* * return immediately if no child has exited. */ +<<<<<<< HEAD #define LOS_WAIT_WNOHANG (1 << 0U) +======= +#define LOS_WAIT_WNOHANG (1 << 0U) ///< 如果没有孩子进程退出,则立即返回,而不是阻塞在这个函数上等待;如果结束了,则返回该子进程的进程号。 +>>>>>>> remotes/origin/main /* * return if a child has stopped (but not traced via ptrace(2)). * Status for traced children which have stopped is provided even * if this option is not specified. */ +<<<<<<< HEAD #define LOS_WAIT_WUNTRACED (1 << 1U) +======= +#define LOS_WAIT_WUNTRACED (1 << 1U) ///< 如果子进程进入暂停情况则马上返回,不予以理会结束状态。untraced +>>>>>>> remotes/origin/main #define LOS_WAIT_WSTOPPED (1 << 1U) /* @@ -488,7 +702,11 @@ STATIC INLINE UINT32 OsGetRootPid(const LosProcessCB *processCB) * return if a stopped child has been resumed by delivery of SIGCONT. * (For Linux-only options, see below.) */ +<<<<<<< HEAD #define LOS_WAIT_WCONTINUED (1 << 3U) +======= +#define LOS_WAIT_WCONTINUED (1 << 3U) ///< 可获取子进程恢复执行的状态,也就是可获取continued状态 continued +>>>>>>> remotes/origin/main /* * Leave the child in a waitable state; @@ -499,21 +717,34 @@ STATIC INLINE UINT32 OsGetRootPid(const LosProcessCB *processCB) /* * Indicates that you are already in a wait state */ +<<<<<<< HEAD #define OS_PROCESS_WAIT (1 << 15U) +======= +#define OS_PROCESS_WAIT (1 << 15U) ///< 表示已经处于等待状态 +>>>>>>> remotes/origin/main /* * Wait for any child process to finish */ +<<<<<<< HEAD #define OS_PROCESS_WAIT_ANY OS_TASK_WAIT_ANYPROCESS +======= +#define OS_PROCESS_WAIT_ANY OS_TASK_WAIT_ANYPROCESS ///< 等待任意子进程完成 +>>>>>>> remotes/origin/main /* * Wait for the child process specified by the pid to finish */ +<<<<<<< HEAD #define OS_PROCESS_WAIT_PRO OS_TASK_WAIT_PROCESS +======= +#define OS_PROCESS_WAIT_PRO OS_TASK_WAIT_PROCESS ///< 等待pid指定的子进程完成 +>>>>>>> remotes/origin/main /* * Waits for any child process in the specified process group to finish. */ +<<<<<<< HEAD #define OS_PROCESS_WAIT_GID OS_TASK_WAIT_GID #define OS_PROCESS_INFO_ALL 1 @@ -523,6 +754,17 @@ extern UINTPTR __user_init_entry; extern UINTPTR __user_init_bss; extern UINTPTR __user_init_end; extern UINTPTR __user_init_load_addr; +======= +#define OS_PROCESS_WAIT_GID OS_TASK_WAIT_GID ///< 等待指定进程组中的任意子进程完成 + +#define OS_PROCESS_INFO_ALL 1 +#define OS_PROCESS_DEFAULT_UMASK 0022 ///< 系统默认的用户掩码(umask),大多数的Linux系统的默认掩码为022。 +//用户掩码的作用是用户在创建文件时从文件的默认权限中去除掩码中的权限。所以文件创建之后的权限实际为:创建文件的权限为:0666-0022=0644。创建文件夹的权限为:0777-0022=0755 +extern UINTPTR __user_init_entry; ///< 第一个用户态进程(init)的入口地址 查看 LITE_USER_SEC_ENTRY +extern UINTPTR __user_init_bss; ///< 查看 LITE_USER_SEC_BSS ,赋值由liteos.ld完成 +extern UINTPTR __user_init_end; ///< init 进程的用户空间初始化结束地址 +extern UINTPTR __user_init_load_addr;///< init 进程的加载地址 ,由链接器赋值 +>>>>>>> remotes/origin/main extern UINT32 OsProcessInit(VOID); extern UINT32 OsSystemProcessCreate(VOID); extern VOID OsProcessNaturalExit(LosProcessCB *processCB, UINT32 status); diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_event.c b/src/kernel_liteos_a/kernel/base/ipc/los_event.c index dc55aa25..42a02f6f 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_event.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_event.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -40,6 +44,10 @@ #include "los_exc.h" #endif +<<<<<<< HEAD +======= +/// 初始化一个事件控制块 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_EventInit(PEVENT_CB_S eventCB) { UINT32 intSave; @@ -48,6 +56,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_EventInit(PEVENT_CB_S eventCB) return LOS_ERRNO_EVENT_PTR_NULL; } +<<<<<<< HEAD intSave = LOS_IntLock(); eventCB->uwEventID = 0; LOS_ListInit(&eventCB->stEventList); @@ -56,6 +65,16 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_EventInit(PEVENT_CB_S eventCB) return LOS_OK; } +======= + intSave = LOS_IntLock();//锁中断 + eventCB->uwEventID = 0;//事件类型初始化 + LOS_ListInit(&eventCB->stEventList);//事件链表初始化 + LOS_IntRestore(intSave);//恢复中断 + OsHookCall(LOS_HOOK_TYPE_EVENT_INIT, eventCB); + return LOS_OK; +} +///事件参数检查 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsEventParamCheck(const VOID *ptr, UINT32 eventMask, UINT32 mode) { if (ptr == NULL) { @@ -77,11 +96,16 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventParamCheck(const VOID *ptr, UINT32 eventMa } return LOS_OK; } +<<<<<<< HEAD +======= +///根据用户传入的事件值、事件掩码及校验模式,返回用户传入的事件是否符合预期 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsEventPoll(UINT32 *eventID, UINT32 eventMask, UINT32 mode) { UINT32 ret = 0; +<<<<<<< HEAD LOS_ASSERT(OsIntLocked()); LOS_ASSERT(LOS_SpinHeld(&g_taskSpin)); @@ -91,26 +115,51 @@ LITE_OS_SEC_TEXT UINT32 OsEventPoll(UINT32 *eventID, UINT32 eventMask, UINT32 mo } } else { if ((eventMask != 0) && (eventMask == (*eventID & eventMask))) { +======= + LOS_ASSERT(OsIntLocked());//断言不允许中断了 + LOS_ASSERT(LOS_SpinHeld(&g_taskSpin));//任务自旋锁 + + if (mode & LOS_WAITMODE_OR) {//如果模式是读取掩码中任意事件 + if ((*eventID & eventMask) != 0) { + ret = *eventID & eventMask; + } + } else {//等待全部事件发生 + if ((eventMask != 0) && (eventMask == (*eventID & eventMask))) {//必须满足全部事件发生 +>>>>>>> remotes/origin/main ret = *eventID & eventMask; } } +<<<<<<< HEAD if (ret && (mode & LOS_WAITMODE_CLR)) { +======= + if (ret && (mode & LOS_WAITMODE_CLR)) {//读取完成后清除事件 +>>>>>>> remotes/origin/main *eventID = *eventID & ~ret; } return ret; } +<<<<<<< HEAD +======= +///检查读事件 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsEventReadCheck(const PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode) { UINT32 ret; LosTaskCB *runTask = NULL; +<<<<<<< HEAD ret = OsEventParamCheck(eventCB, eventMask, mode); +======= + + ret = OsEventParamCheck(eventCB, eventMask, mode);//事件参数检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD if (OS_INT_ACTIVE) { return LOS_ERRNO_EVENT_READ_IN_INTERRUPT; } @@ -123,6 +172,20 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventReadCheck(const PEVENT_CB_S eventCB, UINT3 return LOS_OK; } +======= + if (OS_INT_ACTIVE) {//中断正在进行 + return LOS_ERRNO_EVENT_READ_IN_INTERRUPT;//不能在中断发送时读事件 + } + + runTask = OsCurrTaskGet();//获取当前任务 + if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) {//任务属于系统任务 + OsBackTrace(); + return LOS_ERRNO_EVENT_READ_IN_SYSTEM_TASK;//不能在系统任务中读取事件 + } + return LOS_OK; +} +/// 读取指定事件类型的实现函数,超时时间为相对时间:单位为Tick +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsEventReadImp(PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode, UINT32 timeout, BOOL once) { @@ -131,6 +194,7 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventReadImp(PEVENT_CB_S eventCB, UINT32 eventM OsHookCall(LOS_HOOK_TYPE_EVENT_READ, eventCB, eventMask, mode, timeout); if (once == FALSE) { +<<<<<<< HEAD ret = OsEventPoll(&eventCB->uwEventID, eventMask, mode); } @@ -147,28 +211,59 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventReadImp(PEVENT_CB_S eventCB, UINT32 eventM runTask->eventMode = mode; runTask->taskEvent = eventCB; OsTaskWaitSetPendMask(OS_TASK_WAIT_EVENT, eventMask, timeout); +======= + ret = OsEventPoll(&eventCB->uwEventID, eventMask, mode);//检测事件是否符合预期 + } + + if (ret == 0) {//不符合预期时 + if (timeout == 0) {//不等待的情况 + return ret; + } + + if (!OsPreemptableInSched()) {//不能抢占式调度 + return LOS_ERRNO_EVENT_READ_IN_LOCK; + } + + runTask->eventMask = eventMask; //等待事件 + runTask->eventMode = mode; //事件模式 + runTask->taskEvent = eventCB; //事件控制块 + OsTaskWaitSetPendMask(OS_TASK_WAIT_EVENT, eventMask, timeout);//任务进入等待状态,等待事件的到来并设置时长和掩码 +>>>>>>> remotes/origin/main ret = runTask->ops->wait(runTask, &eventCB->stEventList, timeout); if (ret == LOS_ERRNO_TSK_TIMEOUT) { return LOS_ERRNO_EVENT_READ_TIMEOUT; } +<<<<<<< HEAD ret = OsEventPoll(&eventCB->uwEventID, eventMask, mode); } return ret; } +======= + ret = OsEventPoll(&eventCB->uwEventID, eventMask, mode);//检测事件是否符合预期 + } + return ret; +} +///读取指定事件类型,超时时间为相对时间:单位为Tick +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsEventRead(PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode, UINT32 timeout, BOOL once) { UINT32 ret; UINT32 intSave; +<<<<<<< HEAD ret = OsEventReadCheck(eventCB, eventMask, mode); +======= + ret = OsEventReadCheck(eventCB, eventMask, mode);//读取事件检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsEventReadImp(eventCB, eventMask, mode, timeout, once); SCHEDULER_UNLOCK(intSave); return ret; @@ -182,6 +277,21 @@ LITE_OS_SEC_TEXT STATIC UINT8 OsEventResume(LosTaskCB *resumedTask, const PEVENT ((resumedTask->eventMode & LOS_WAITMODE_AND) && ((resumedTask->eventMask & eventCB->uwEventID) == resumedTask->eventMask))) { exitFlag = 1; +======= + ret = OsEventReadImp(eventCB, eventMask, mode, timeout, once);//读事件实现函数 + SCHEDULER_UNLOCK(intSave); + return ret; +} +///事件恢复操作 +LITE_OS_SEC_TEXT STATIC UINT8 OsEventResume(LosTaskCB *resumedTask, const PEVENT_CB_S eventCB, UINT32 events) +{ + UINT8 exitFlag = 0;//是否唤醒 + + if (((resumedTask->eventMode & LOS_WAITMODE_OR) && ((resumedTask->eventMask & events) != 0)) || + ((resumedTask->eventMode & LOS_WAITMODE_AND) && + ((resumedTask->eventMask & eventCB->uwEventID) == resumedTask->eventMask))) {//逻辑与 和 逻辑或 的处理 + exitFlag = 1; +>>>>>>> remotes/origin/main resumedTask->taskEvent = NULL; OsTaskWakeClearPendMask(resumedTask); @@ -190,13 +300,18 @@ LITE_OS_SEC_TEXT STATIC UINT8 OsEventResume(LosTaskCB *resumedTask, const PEVENT return exitFlag; } +<<<<<<< HEAD +======= +///以不安全的方式写事件 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT VOID OsEventWriteUnsafe(PEVENT_CB_S eventCB, UINT32 events, BOOL once, UINT8 *exitFlag) { LosTaskCB *resumedTask = NULL; LosTaskCB *nextTask = NULL; BOOL schedFlag = FALSE; OsHookCall(LOS_HOOK_TYPE_EVENT_WRITE, eventCB, events); +<<<<<<< HEAD eventCB->uwEventID |= events; if (!LOS_ListEmpty(&eventCB->stEventList)) { for (resumedTask = LOS_DL_LIST_ENTRY((&eventCB->stEventList)->pstNext, LosTaskCB, pendList); @@ -217,6 +332,28 @@ LITE_OS_SEC_TEXT VOID OsEventWriteUnsafe(PEVENT_CB_S eventCB, UINT32 events, BOO } } +======= + eventCB->uwEventID |= events;//对应位贴上标签 + if (!LOS_ListEmpty(&eventCB->stEventList)) {//等待事件链表判断,处理等待事件的任务 + for (resumedTask = LOS_DL_LIST_ENTRY((&eventCB->stEventList)->pstNext, LosTaskCB, pendList); + &resumedTask->pendList != &eventCB->stEventList;) {//循环获取任务链表 + nextTask = LOS_DL_LIST_ENTRY(resumedTask->pendList.pstNext, LosTaskCB, pendList);//获取任务实体 + if (OsEventResume(resumedTask, eventCB, events)) {//是否恢复任务 + schedFlag = TRUE;//任务已加至就绪队列,申请发生一次调度 + } + if (once == TRUE) {//是否只处理一次任务 + break;//退出循环 + } + resumedTask = nextTask;//检查链表中下一个任务 + } + } + + if ((exitFlag != NULL) && (schedFlag == TRUE)) {//是否让外面调度 + *exitFlag = 1; + } +} +///写入事件 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT STATIC UINT32 OsEventWrite(PEVENT_CB_S eventCB, UINT32 events, BOOL once) { UINT32 intSave; @@ -230,6 +367,7 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventWrite(PEVENT_CB_S eventCB, UINT32 events, return LOS_ERRNO_EVENT_SETBIT_INVALID; } +<<<<<<< HEAD SCHEDULER_LOCK(intSave); OsEventWriteUnsafe(eventCB, events, once, &exitFlag); SCHEDULER_UNLOCK(intSave); @@ -241,43 +379,84 @@ LITE_OS_SEC_TEXT STATIC UINT32 OsEventWrite(PEVENT_CB_S eventCB, UINT32 events, return LOS_OK; } +======= + SCHEDULER_LOCK(intSave); //禁止调度 + OsEventWriteUnsafe(eventCB, events, once, &exitFlag);//写入事件 + SCHEDULER_UNLOCK(intSave); //允许调度 + + if (exitFlag == 1) { //需要发生调度 + LOS_MpSchedule(OS_MP_CPU_ALL);//通知所有CPU调度 + LOS_Schedule();//执行调度 + } + return LOS_OK; +} +///根据用户传入的事件值、事件掩码及校验模式,返回用户传入的事件是否符合预期 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_EventPoll(UINT32 *eventID, UINT32 eventMask, UINT32 mode) { UINT32 ret; UINT32 intSave; +<<<<<<< HEAD +======= + //事件参数检查 +>>>>>>> remotes/origin/main ret = OsEventParamCheck((VOID *)eventID, eventMask, mode); if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD SCHEDULER_LOCK(intSave); +======= + SCHEDULER_LOCK(intSave);//申请任务自旋锁 +>>>>>>> remotes/origin/main ret = OsEventPoll(eventID, eventMask, mode); SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///读取指定事件类型,超时时间为相对时间:单位为Tick +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_EventRead(PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode, UINT32 timeout) { return OsEventRead(eventCB, eventMask, mode, timeout, FALSE); } +<<<<<<< HEAD +======= +///写指定的事件类型 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_EventWrite(PEVENT_CB_S eventCB, UINT32 events) { return OsEventWrite(eventCB, events, FALSE); } +<<<<<<< HEAD +======= +///只读一次事件 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsEventReadOnce(PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode, UINT32 timeout) { return OsEventRead(eventCB, eventMask, mode, timeout, TRUE); } +<<<<<<< HEAD +======= +///只写一次事件 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsEventWriteOnce(PEVENT_CB_S eventCB, UINT32 events) { return OsEventWrite(eventCB, events, TRUE); } +<<<<<<< HEAD +======= +///销毁指定的事件控制块 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_EventDestroy(PEVENT_CB_S eventCB) { UINT32 intSave; @@ -298,7 +477,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_EventDestroy(PEVENT_CB_S eventCB) OsHookCall(LOS_HOOK_TYPE_EVENT_DESTROY, eventCB); return LOS_OK; } +<<<<<<< HEAD +======= +///清除指定的事件类型 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_EventClear(PEVENT_CB_S eventCB, UINT32 eventMask) { UINT32 intSave; @@ -313,7 +496,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_EventClear(PEVENT_CB_S eventCB, UINT32 eventMa return LOS_OK; } +<<<<<<< HEAD +======= +///有条件式读事件 +>>>>>>> remotes/origin/main #ifdef LOSCFG_COMPAT_POSIX LITE_OS_SEC_TEXT UINT32 OsEventReadWithCond(const EventCond *cond, PEVENT_CB_S eventCB, UINT32 eventMask, UINT32 mode, UINT32 timeout) diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_futex.c b/src/kernel_liteos_a/kernel/base/ipc/los_futex.c index 384bb483..ce2f7288 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_futex.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_futex.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -43,6 +47,7 @@ #ifdef LOSCFG_KERNEL_VM +<<<<<<< HEAD #define OS_FUTEX_FROM_FUTEXLIST(ptr) LOS_DL_LIST_ENTRY(ptr, FutexNode, futexList) #define OS_FUTEX_FROM_QUEUELIST(ptr) LOS_DL_LIST_ENTRY(ptr, FutexNode, queueList) #define OS_FUTEX_KEY_BASE USER_ASPACE_BASE @@ -65,6 +70,32 @@ typedef struct { FutexHash g_futexHash[FUTEX_INDEX_MAX]; +======= +#define OS_FUTEX_FROM_FUTEXLIST(ptr) LOS_DL_LIST_ENTRY(ptr, FutexNode, futexList) // 通过快锁节点找到结构体 +#define OS_FUTEX_FROM_QUEUELIST(ptr) LOS_DL_LIST_ENTRY(ptr, FutexNode, queueList) // 通过队列节点找到结构体 +#define OS_FUTEX_KEY_BASE USER_ASPACE_BASE ///< 进程用户空间基址 +#define OS_FUTEX_KEY_MAX (USER_ASPACE_BASE + USER_ASPACE_SIZE) ///< 进程用户空间尾址 + +/* private: 0~63 hash index_num + * shared: 64~79 hash index_num */ +#define FUTEX_INDEX_PRIVATE_MAX 64 ///< 0~63号桶用于存放私有锁(以虚拟地址进行哈希),同一进程不同线程共享futex变量,表明变量在进程地址空间中的位置 +///< 它告诉内核,这个futex是进程专有的,不可以与其他进程共享。它仅仅用作同一进程的线程间同步。 +#define FUTEX_INDEX_SHARED_MAX 16 ///< 64~79号桶用于存放共享锁(以物理地址进行哈希),不同进程间通过文件共享futex变量,表明该变量在文件中的位置 +#define FUTEX_INDEX_MAX (FUTEX_INDEX_PRIVATE_MAX + FUTEX_INDEX_SHARED_MAX) ///< 80个哈希桶 + +#define FUTEX_INDEX_SHARED_POS FUTEX_INDEX_PRIVATE_MAX ///< 共享锁开始位置 +#define FUTEX_HASH_PRIVATE_MASK (FUTEX_INDEX_PRIVATE_MAX - 1) +#define FUTEX_HASH_SHARED_MASK (FUTEX_INDEX_SHARED_MAX - 1) +/// 单独哈希桶,上面挂了一个个 FutexNode +typedef struct { + LosMux listLock;///< 内核操作lockList的互斥锁 + LOS_DL_LIST lockList;///< 用于挂载 FutexNode (Fast userspace mutex,用户态快速互斥锁) +} FutexHash; + +FutexHash g_futexHash[FUTEX_INDEX_MAX];///< 80个哈希桶 + +/// 对互斥锁封装 +>>>>>>> remotes/origin/main STATIC INT32 OsFutexLock(LosMux *lock) { UINT32 ret = LOS_MuxLock(lock, LOS_WAIT_FOREVER); @@ -84,15 +115,26 @@ STATIC INT32 OsFutexUnlock(LosMux *lock) } return LOS_OK; } +<<<<<<< HEAD +======= +///< 初始化Futex(Fast userspace mutex,用户态快速互斥锁)模块 +>>>>>>> remotes/origin/main UINT32 OsFutexInit(VOID) { INT32 count; UINT32 ret; +<<<<<<< HEAD for (count = 0; count < FUTEX_INDEX_MAX; count++) { LOS_ListInit(&g_futexHash[count].lockList); ret = LOS_MuxInit(&(g_futexHash[count].listLock), NULL); +======= + // 初始化 80个哈希桶 + for (count = 0; count < FUTEX_INDEX_MAX; count++) { + LOS_ListInit(&g_futexHash[count].lockList); // 初始化双向链表,上面挂 FutexNode + ret = LOS_MuxInit(&(g_futexHash[count].listLock), NULL);//初始化互斥锁 +>>>>>>> remotes/origin/main if (ret) { return ret; } @@ -101,7 +143,11 @@ UINT32 OsFutexInit(VOID) return LOS_OK; } +<<<<<<< HEAD LOS_MODULE_INIT(OsFutexInit, LOS_INIT_LEVEL_KMOD_EXTENDED); +======= +LOS_MODULE_INIT(OsFutexInit, LOS_INIT_LEVEL_KMOD_EXTENDED);///< 注册Futex模块 +>>>>>>> remotes/origin/main #ifdef LOS_FUTEX_DEBUG STATIC VOID OsFutexShowTaskNodeAttr(const LOS_DL_LIST *futexList) @@ -152,19 +198,30 @@ VOID OsFutexHashShow(VOID) } } #endif +<<<<<<< HEAD +======= +/// 通过用户空间地址获取哈希key +>>>>>>> remotes/origin/main STATIC INLINE UINTPTR OsFutexFlagsToKey(const UINT32 *userVaddr, const UINT32 flags) { UINTPTR futexKey; if (flags & FUTEX_PRIVATE) { +<<<<<<< HEAD futexKey = (UINTPTR)userVaddr; } else { futexKey = (UINTPTR)LOS_PaddrQuery((UINT32 *)userVaddr); +======= + futexKey = (UINTPTR)userVaddr;//私有锁(以虚拟地址进行哈希) + } else { + futexKey = (UINTPTR)LOS_PaddrQuery((UINT32 *)userVaddr);//共享锁(以物理地址进行哈希) +>>>>>>> remotes/origin/main } return futexKey; } +<<<<<<< HEAD STATIC INLINE UINT32 OsFutexKeyToIndex(const UINTPTR futexKey, const UINT32 flags) { @@ -175,10 +232,23 @@ STATIC INLINE UINT32 OsFutexKeyToIndex(const UINTPTR futexKey, const UINT32 flag } else { index &= FUTEX_HASH_SHARED_MASK; index += FUTEX_INDEX_SHARED_POS; +======= +/// 通过哈希key获取索引 +STATIC INLINE UINT32 OsFutexKeyToIndex(const UINTPTR futexKey, const UINT32 flags) +{ + UINT32 index = LOS_HashFNV32aBuf(&futexKey, sizeof(UINTPTR), FNV1_32A_INIT);//获取哈希桶索引 + + if (flags & FUTEX_PRIVATE) { + index &= FUTEX_HASH_PRIVATE_MASK;//将index锁定在 0 ~ 63号 + } else { + index &= FUTEX_HASH_SHARED_MASK; + index += FUTEX_INDEX_SHARED_POS;//共享锁索引,将index锁定在 64 ~ 79号 +>>>>>>> remotes/origin/main } return index; } +<<<<<<< HEAD STATIC INLINE VOID OsFutexSetKey(UINTPTR futexKey, UINT32 flags, FutexNode *node) { @@ -187,12 +257,23 @@ STATIC INLINE VOID OsFutexSetKey(UINTPTR futexKey, UINT32 flags, FutexNode *node node->pid = (flags & FUTEX_PRIVATE) ? LOS_GetCurrProcessID() : OS_INVALID; } +======= +/// 设置快锁哈希key +STATIC INLINE VOID OsFutexSetKey(UINTPTR futexKey, UINT32 flags, FutexNode *node) +{ + node->key = futexKey;//哈希key + node->index = OsFutexKeyToIndex(futexKey, flags);//哈希桶索引 + node->pid = (flags & FUTEX_PRIVATE) ? LOS_GetCurrProcessID() : OS_INVALID;//获取进程ID,共享快锁时 快锁节点没有进程ID +} +//析构参数节点 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsFutexDeinitFutexNode(FutexNode *node) { node->index = OS_INVALID_VALUE; node->pid = 0; LOS_ListDelete(&node->queueList); } +<<<<<<< HEAD STATIC INLINE VOID OsFutexReplaceQueueListHeadNode(FutexNode *oldHeadNode, FutexNode *newHeadNode) { @@ -204,11 +285,28 @@ STATIC INLINE VOID OsFutexReplaceQueueListHeadNode(FutexNode *oldHeadNode, Futex } } +======= +/// 新旧两个节点交换 futexList 位置 +STATIC INLINE VOID OsFutexReplaceQueueListHeadNode(FutexNode *oldHeadNode, FutexNode *newHeadNode) +{ + LOS_DL_LIST *futexList = oldHeadNode->futexList.pstPrev; + LOS_ListDelete(&oldHeadNode->futexList);//将旧节点从futexList链表上摘除 + LOS_ListHeadInsert(futexList, &newHeadNode->futexList);//将新节点从头部插入futexList链表 + if ((newHeadNode->queueList.pstNext == NULL) || (newHeadNode->queueList.pstPrev == NULL)) {//新节点前后没有等待这把锁的任务 + LOS_ListInit(&newHeadNode->queueList);//初始化等锁任务链表 + } +} +/// 将参数节点从futexList上摘除 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsFutexDeleteKeyFromFutexList(FutexNode *node) { LOS_ListDelete(&node->futexList); } +<<<<<<< HEAD +======= +/// 从哈希桶中删除快锁节点 +>>>>>>> remotes/origin/main STATIC VOID OsFutexDeleteKeyNodeFromHash(FutexNode *node, BOOL isDeleteHead, FutexNode **headNode, BOOL *queueFlags) { FutexNode *nextNode = NULL; @@ -217,8 +315,13 @@ STATIC VOID OsFutexDeleteKeyNodeFromHash(FutexNode *node, BOOL isDeleteHead, Fut return; } +<<<<<<< HEAD if (LOS_ListEmpty(&node->queueList)) { OsFutexDeleteKeyFromFutexList(node); +======= + if (LOS_ListEmpty(&node->queueList)) {//如果没有任务在等锁 + OsFutexDeleteKeyFromFutexList(node);//从快锁链表上摘除 +>>>>>>> remotes/origin/main if (queueFlags != NULL) { *queueFlags = TRUE; } @@ -226,10 +329,17 @@ STATIC VOID OsFutexDeleteKeyNodeFromHash(FutexNode *node, BOOL isDeleteHead, Fut } /* FutexList is not NULL, but the header node of queueList */ +<<<<<<< HEAD if (node->futexList.pstNext != NULL) { if (isDeleteHead == TRUE) { nextNode = OS_FUTEX_FROM_QUEUELIST(LOS_DL_LIST_FIRST(&node->queueList)); OsFutexReplaceQueueListHeadNode(node, nextNode); +======= + if (node->futexList.pstNext != NULL) {//是头节点 + if (isDeleteHead == TRUE) {//是否要删除头节点 + nextNode = OS_FUTEX_FROM_QUEUELIST(LOS_DL_LIST_FIRST(&node->queueList));//取出第一个快锁节点 + OsFutexReplaceQueueListHeadNode(node, nextNode);//两个节点交换位置 +>>>>>>> remotes/origin/main if (headNode != NULL) { *headNode = nextNode; } @@ -242,22 +352,38 @@ EXIT: OsFutexDeinitFutexNode(node); return; } +<<<<<<< HEAD VOID OsFutexNodeDeleteFromFutexHash(FutexNode *node, BOOL isDeleteHead, FutexNode **headNode, BOOL *queueFlags) { FutexHash *hashNode = NULL; +======= +/// 从哈希桶上删除快锁 +VOID OsFutexNodeDeleteFromFutexHash(FutexNode *node, BOOL isDeleteHead, FutexNode **headNode, BOOL *queueFlags) +{ + FutexHash *hashNode = NULL; + //通过key找到桶号 +>>>>>>> remotes/origin/main UINT32 index = OsFutexKeyToIndex(node->key, (node->pid == OS_INVALID) ? 0 : FUTEX_PRIVATE); if (index >= FUTEX_INDEX_MAX) { return; } +<<<<<<< HEAD hashNode = &g_futexHash[index]; +======= + hashNode = &g_futexHash[index];//找到hash桶 +>>>>>>> remotes/origin/main if (OsMuxLockUnsafe(&hashNode->listLock, LOS_WAIT_FOREVER)) { return; } +<<<<<<< HEAD if (node->index != index) { +======= + if (node->index != index) {//快锁节点桶号需和哈希桶号一致 +>>>>>>> remotes/origin/main goto EXIT; } @@ -270,7 +396,10 @@ EXIT: return; } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main STATIC FutexNode *OsFutexDeleteAlreadyWakeTaskAndGetNext(const FutexNode *node, FutexNode **headNode, BOOL isDeleteHead) { FutexNode *tempNode = (FutexNode *)node; @@ -292,7 +421,11 @@ STATIC FutexNode *OsFutexDeleteAlreadyWakeTaskAndGetNext(const FutexNode *node, return tempNode; } +<<<<<<< HEAD +======= +/// 插入一把新Futex锁到哈希桶中,只有是新的key时才会插入,因为其实存在多个FutexNode是一个key +>>>>>>> remotes/origin/main STATIC VOID OsFutexInsertNewFutexKeyToHash(FutexNode *node) { FutexNode *headNode = NULL; @@ -322,16 +455,27 @@ STATIC VOID OsFutexInsertNewFutexKeyToHash(FutexNode *node) futexList != &(hashNode->lockList); futexList = futexList->pstNext) { headNode = OS_FUTEX_FROM_FUTEXLIST(futexList); +<<<<<<< HEAD if (node->key <= headNode->key) { LOS_ListTailInsert(&(headNode->futexList), &(node->futexList)); break; } +======= + if (node->key <= headNode->key) { + LOS_ListTailInsert(&(headNode->futexList), &(node->futexList)); + break; + } + +>>>>>>> remotes/origin/main } EXIT: return; } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main STATIC INT32 OsFutexInsertFindFormBackToFront(LOS_DL_LIST *queueList, const LosTaskCB *runTask, FutexNode *node) { LOS_DL_LIST *listHead = queueList; @@ -407,15 +551,23 @@ STATIC INT32 OsFutexRecycleAndFindHeadNode(FutexNode *headNode, FutexNode *node, return LOS_OK; } +<<<<<<< HEAD STATIC INT32 OsFutexInsertTasktoPendList(FutexNode **firstNode, FutexNode *node, const LosTaskCB *run) { LosTaskCB *taskHead = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&((*firstNode)->pendList))); +======= +///< 将快锁挂到任务的阻塞链表上 +STATIC INT32 OsFutexInsertTasktoPendList(FutexNode **firstNode, FutexNode *node, const LosTaskCB *run) +{ + LosTaskCB *taskHead = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&((*firstNode)->pendList)));//获取阻塞链表首个任务 +>>>>>>> remotes/origin/main LOS_DL_LIST *queueList = &((*firstNode)->queueList); INT32 ret1 = OsSchedParamCompare(run, taskHead); if (ret1 < 0) { /* The one with the highest priority is inserted at the top of the queue */ +<<<<<<< HEAD LOS_ListTailInsert(queueList, &(node->queueList)); OsFutexReplaceQueueListHeadNode(*firstNode, node); *firstNode = node; @@ -441,21 +593,60 @@ STATIC INT32 OsFutexInsertTasktoPendList(FutexNode **firstNode, FutexNode *node, STATIC FutexNode *OsFindFutexNode(const FutexNode *node) { FutexHash *hashNode = &g_futexHash[node->index]; +======= + LOS_ListTailInsert(queueList, &(node->queueList));//查到queueList的尾部 + OsFutexReplaceQueueListHeadNode(*firstNode, node);//同时交换futexList链表上的位置 + *firstNode = node; + return LOS_OK; + } + //如果等锁链表上没有任务或者当前任务大于链表首个任务 + if (LOS_ListEmpty(queueList) && (ret1 >= 0)) { + /* Insert the next position in the queue with equal priority */ + LOS_ListHeadInsert(queueList, &(node->queueList));//从头部插入当前任务,当前任务是要被挂起的 + return LOS_OK; + } + + FutexNode *tailNode = OS_FUTEX_FROM_QUEUELIST(LOS_DL_LIST_LAST(queueList));//获取尾部节点 + LosTaskCB *taskTail = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(tailNode->pendList)));//获取阻塞任务的最后一个 + INT32 ret2 = OsSchedParamCompare(taskTail, run); + if ((ret2 <= 0) || (ret1 > ret2)) { + return OsFutexInsertFindFormBackToFront(queueList, run, node);//从后往前插入 + } + + return OsFutexInsertFindFromFrontToBack(queueList, run, node);//否则从前往后插入 +} +/// 由指定快锁找到对应哈希桶 +STATIC FutexNode *OsFindFutexNode(const FutexNode *node) +{ + FutexHash *hashNode = &g_futexHash[node->index];//先找到所在哈希桶 +>>>>>>> remotes/origin/main LOS_DL_LIST *futexList = &(hashNode->lockList); FutexNode *headNode = NULL; for (futexList = futexList->pstNext; +<<<<<<< HEAD futexList != &(hashNode->lockList); futexList = futexList->pstNext) { headNode = OS_FUTEX_FROM_FUTEXLIST(futexList); if ((headNode->key == node->key) && (headNode->pid == node->pid)) { return headNode; } +======= + futexList != &(hashNode->lockList);//判断循环结束条件,相等时说明跑完一轮了 + futexList = futexList->pstNext) { + headNode = OS_FUTEX_FROM_FUTEXLIST(futexList);//拿到快锁节点实体 + if ((headNode->key == node->key) && (headNode->pid == node->pid)) {//已经存在这个节点,注意这里的比较 + return headNode;//是key和pid 一起比较,因为只有这样才能确定唯一性 +>>>>>>> remotes/origin/main } return NULL; } +<<<<<<< HEAD +======= +///< 查找快锁并插入哈希桶中 +>>>>>>> remotes/origin/main STATIC INT32 OsFindAndInsertToHash(FutexNode *node) { FutexNode *headNode = NULL; @@ -464,7 +655,11 @@ STATIC INT32 OsFindAndInsertToHash(FutexNode *node) INT32 ret; headNode = OsFindFutexNode(node); +<<<<<<< HEAD if (headNode == NULL) { +======= + if (headNode == NULL) {//没有找到,说明这是一把新锁 +>>>>>>> remotes/origin/main OsFutexInsertNewFutexKeyToHash(node); LOS_ListInit(&(node->queueList)); return LOS_OK; @@ -483,14 +678,23 @@ STATIC INT32 OsFindAndInsertToHash(FutexNode *node) return ret; } +<<<<<<< HEAD +======= +/// 共享内存检查 +>>>>>>> remotes/origin/main STATIC INT32 OsFutexKeyShmPermCheck(const UINT32 *userVaddr, const UINT32 flags) { PADDR_T paddr; /* Check whether the futexKey is a shared lock */ +<<<<<<< HEAD if (!(flags & FUTEX_PRIVATE)) { paddr = (UINTPTR)LOS_PaddrQuery((UINT32 *)userVaddr); +======= + if (!(flags & FUTEX_PRIVATE)) {//非私有快锁 + paddr = (UINTPTR)LOS_PaddrQuery((UINT32 *)userVaddr);//能否查询到物理地址 +>>>>>>> remotes/origin/main if (paddr == 0) return LOS_NOK; } @@ -549,6 +753,7 @@ STATIC INT32 OsFutexDeleteTimeoutTaskNode(FutexHash *hashNode, FutexNode *node) } return LOS_ETIMEDOUT; } +<<<<<<< HEAD STATIC INT32 OsFutexInsertTaskToHash(LosTaskCB **taskCB, FutexNode **node, const UINTPTR futexKey, const UINT32 flags) { @@ -556,6 +761,15 @@ STATIC INT32 OsFutexInsertTaskToHash(LosTaskCB **taskCB, FutexNode **node, const *taskCB = OsCurrTaskGet(); *node = &((*taskCB)->futex); OsFutexSetKey(futexKey, flags, *node); +======= +/// 将快锁节点插入任务 +STATIC INT32 OsFutexInsertTaskToHash(LosTaskCB **taskCB, FutexNode **node, const UINTPTR futexKey, const UINT32 flags) +{ + INT32 ret; + *taskCB = OsCurrTaskGet(); //获取当前任务 + *node = &((*taskCB)->futex); //获取当前任务的快锁节点 + OsFutexSetKey(futexKey, flags, *node);//设置参数 key index pid +>>>>>>> remotes/origin/main ret = OsFindAndInsertToHash(*node); if (ret) { @@ -565,13 +779,18 @@ STATIC INT32 OsFutexInsertTaskToHash(LosTaskCB **taskCB, FutexNode **node, const LOS_ListInit(&((*node)->pendList)); return LOS_OK; } +<<<<<<< HEAD +======= +/// 将当前任务挂入等待链表中 +>>>>>>> remotes/origin/main STATIC INT32 OsFutexWaitTask(const UINT32 *userVaddr, const UINT32 flags, const UINT32 val, const UINT32 timeout) { INT32 futexRet; UINT32 intSave, lockVal; LosTaskCB *taskCB = NULL; FutexNode *node = NULL; +<<<<<<< HEAD UINTPTR futexKey = OsFutexFlagsToKey(userVaddr, flags); UINT32 index = OsFutexKeyToIndex(futexKey, flags); FutexHash *hashNode = &g_futexHash[index]; @@ -581,17 +800,37 @@ STATIC INT32 OsFutexWaitTask(const UINT32 *userVaddr, const UINT32 flags, const } if (LOS_ArchCopyFromUser(&lockVal, userVaddr, sizeof(UINT32))) { +======= + UINTPTR futexKey = OsFutexFlagsToKey(userVaddr, flags);//通过地址和flags 找到 key + UINT32 index = OsFutexKeyToIndex(futexKey, flags);//通过key找到哈希桶 + FutexHash *hashNode = &g_futexHash[index]; + + if (OsFutexLock(&hashNode->listLock)) {//操作快锁节点链表前先上互斥锁 + return LOS_EINVAL; + } + //userVaddr必须是用户空间虚拟地址 + if (LOS_ArchCopyFromUser(&lockVal, userVaddr, sizeof(UINT32))) {//将值拷贝到内核空间 +>>>>>>> remotes/origin/main PRINT_ERR("Futex wait param check failed! copy from user failed!\n"); futexRet = LOS_EINVAL; goto EXIT_ERR; } +<<<<<<< HEAD if (lockVal != val) { futexRet = LOS_EBADF; goto EXIT_ERR; } if (OsFutexInsertTaskToHash(&taskCB, &node, futexKey, flags)) { +======= + if (lockVal != val) {//对参数内部逻辑检查 + futexRet = LOS_EBADF; + goto EXIT_ERR; + } + //注意第二个参数 FutexNode *node = NULL + if (OsFutexInsertTaskToHash(&taskCB, &node, futexKey, flags)) {// node = taskCB->futex +>>>>>>> remotes/origin/main futexRet = LOS_NOK; goto EXIT_ERR; } @@ -602,7 +841,11 @@ STATIC INT32 OsFutexWaitTask(const UINT32 *userVaddr, const UINT32 flags, const taskCB->ops->wait(taskCB, &(node->pendList), timeout); LOS_SpinUnlock(&g_taskSpin); +<<<<<<< HEAD futexRet = OsFutexUnlock(&hashNode->listLock); +======= + futexRet = OsFutexUnlock(&hashNode->listLock);// +>>>>>>> remotes/origin/main if (futexRet) { OsSchedUnlock(); LOS_IntRestore(intSave); @@ -632,12 +875,17 @@ EXIT_ERR: EXIT_UNLOCK_ERR: return futexRet; } +<<<<<<< HEAD +======= +/// 设置线程等待 | 向Futex表中插入代表被阻塞的线程的node +>>>>>>> remotes/origin/main INT32 OsFutexWait(const UINT32 *userVaddr, UINT32 flags, UINT32 val, UINT32 absTime) { INT32 ret; UINT32 timeout = LOS_WAIT_FOREVER; +<<<<<<< HEAD ret = OsFutexWaitParamCheck(userVaddr, flags, absTime); if (ret) { return ret; @@ -647,6 +895,17 @@ INT32 OsFutexWait(const UINT32 *userVaddr, UINT32 flags, UINT32 val, UINT32 absT } return OsFutexWaitTask(userVaddr, flags, val, timeout); +======= + ret = OsFutexWaitParamCheck(userVaddr, flags, absTime);//参数检查 + if (ret) { + return ret; + } + if (absTime != LOS_WAIT_FOREVER) {//转换时间 , 内核的时间单位是 tick + timeout = OsNS2Tick((UINT64)absTime * OS_SYS_NS_PER_US); //转成 tick + } + + return OsFutexWaitTask(userVaddr, flags, val, timeout);//将任务挂起 timeOut 时长 +>>>>>>> remotes/origin/main } STATIC INT32 OsFutexWakeParamCheck(const UINT32 *userVaddr, UINT32 flags) @@ -657,12 +916,20 @@ STATIC INT32 OsFutexWakeParamCheck(const UINT32 *userVaddr, UINT32 flags) PRINT_ERR("Futex wake param check failed! error flags: 0x%x\n", flags); return LOS_EINVAL; } +<<<<<<< HEAD +======= + //地址必须在用户空间 +>>>>>>> remotes/origin/main if ((vaddr % sizeof(INT32)) || (vaddr < OS_FUTEX_KEY_BASE) || (vaddr >= OS_FUTEX_KEY_MAX)) { PRINT_ERR("Futex wake param check failed! error userVaddr: 0x%x\n", userVaddr); return LOS_EINVAL; } +<<<<<<< HEAD +======= + //必须得是个共享内存地址 +>>>>>>> remotes/origin/main if (flags && (OsFutexKeyShmPermCheck(userVaddr, flags) != LOS_OK)) { PRINT_ERR("Futex wake param check failed! error shared memory perm userVaddr: 0x%x\n", userVaddr); return LOS_EINVAL; @@ -672,7 +939,12 @@ STATIC INT32 OsFutexWakeParamCheck(const UINT32 *userVaddr, UINT32 flags) } /* Check to see if the task to be awakened has timed out +<<<<<<< HEAD * if time out, to weak next pend task. +======= + * if time out, to weak next pend task. + * | 查看要唤醒的任务是否超时,如果超时,就唤醒,并查看下一个挂起的任务。 +>>>>>>> remotes/origin/main */ STATIC VOID OsFutexCheckAndWakePendTask(FutexNode *headNode, const INT32 wakeNumber, FutexHash *hashNode, FutexNode **nextNode, BOOL *wakeAny) @@ -707,6 +979,10 @@ STATIC VOID OsFutexCheckAndWakePendTask(FutexNode *headNode, const INT32 wakeNum } return; } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main STATIC INT32 OsFutexWakeTask(UINTPTR futexKey, UINT32 flags, INT32 wakeNumber, FutexNode **newHeadNode, BOOL *wakeAny) { @@ -715,13 +991,21 @@ STATIC INT32 OsFutexWakeTask(UINTPTR futexKey, UINT32 flags, INT32 wakeNumber, F FutexNode *headNode = NULL; UINT32 index = OsFutexKeyToIndex(futexKey, flags); FutexHash *hashNode = &g_futexHash[index]; +<<<<<<< HEAD FutexNode tempNode = { +======= + FutexNode tempNode = { //先组成一个临时快锁节点,目的是为了找到哈希桶中是否有这个节点 +>>>>>>> remotes/origin/main .key = futexKey, .index = index, .pid = (flags & FUTEX_PRIVATE) ? LOS_GetCurrProcessID() : OS_INVALID, }; +<<<<<<< HEAD node = OsFindFutexNode(&tempNode); +======= + node = OsFindFutexNode(&tempNode);//找快锁节点 +>>>>>>> remotes/origin/main if (node == NULL) { return LOS_EBADF; } @@ -729,7 +1013,11 @@ STATIC INT32 OsFutexWakeTask(UINTPTR futexKey, UINT32 flags, INT32 wakeNumber, F headNode = node; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD OsFutexCheckAndWakePendTask(headNode, wakeNumber, hashNode, newHeadNode, wakeAny); +======= + OsFutexCheckAndWakePendTask(headNode, wakeNumber, hashNode, newHeadNode, wakeAny);//再找到等这把锁的唤醒指向数量的任务 +>>>>>>> remotes/origin/main if ((*newHeadNode) != NULL) { OsFutexReplaceQueueListHeadNode(headNode, *newHeadNode); OsFutexDeinitFutexNode(headNode); @@ -741,7 +1029,11 @@ STATIC INT32 OsFutexWakeTask(UINTPTR futexKey, UINT32 flags, INT32 wakeNumber, F return LOS_OK; } +<<<<<<< HEAD +======= +/// 唤醒一个被指定锁阻塞的线程 +>>>>>>> remotes/origin/main INT32 OsFutexWake(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber) { INT32 ret, futexRet; @@ -750,11 +1042,19 @@ INT32 OsFutexWake(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber) FutexHash *hashNode = NULL; FutexNode *headNode = NULL; BOOL wakeAny = FALSE; +<<<<<<< HEAD if (OsFutexWakeParamCheck(userVaddr, flags)) { return LOS_EINVAL; } +======= + //1.检查参数 + if (OsFutexWakeParamCheck(userVaddr, flags)) { + return LOS_EINVAL; + } + //2.找到指定用户空间地址对应的桶 +>>>>>>> remotes/origin/main futexKey = OsFutexFlagsToKey(userVaddr, flags); index = OsFutexKeyToIndex(futexKey, flags); @@ -762,7 +1062,11 @@ INT32 OsFutexWake(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber) if (OsFutexLock(&hashNode->listLock)) { return LOS_EINVAL; } +<<<<<<< HEAD +======= + //3.换起等待该锁的进程 +>>>>>>> remotes/origin/main ret = OsFutexWakeTask(futexKey, flags, wakeNumber, &headNode, &wakeAny); if (ret) { goto EXIT_ERR; @@ -776,7 +1080,11 @@ INT32 OsFutexWake(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber) if (futexRet) { goto EXIT_UNLOCK_ERR; } +<<<<<<< HEAD +======= + //4.根据指定参数决定是否发起调度 +>>>>>>> remotes/origin/main if (wakeAny == TRUE) { LOS_MpSchedule(OS_MP_CPU_ALL); LOS_Schedule(); @@ -885,7 +1193,11 @@ STATIC VOID OsFutexRequeueSplitTwoLists(FutexHash *oldHashNode, FutexNode *oldHe tailNode->queueList.pstNext = &newHeadNode->queueList; return; } +<<<<<<< HEAD +======= +/// 删除旧key并获取头节点 +>>>>>>> remotes/origin/main STATIC FutexNode *OsFutexRequeueRemoveOldKeyAndGetHead(UINTPTR oldFutexKey, UINT32 flags, INT32 wakeNumber, UINTPTR newFutexKey, INT32 requeueCount, BOOL *wakeAny) { @@ -921,7 +1233,11 @@ STATIC FutexNode *OsFutexRequeueRemoveOldKeyAndGetHead(UINTPTR oldFutexKey, UINT return oldHeadNode; } +<<<<<<< HEAD +======= +/// 检查锁在Futex表中的状态 +>>>>>>> remotes/origin/main STATIC INT32 OsFutexRequeueParamCheck(const UINT32 *oldUserVaddr, UINT32 flags, const UINT32 *newUserVaddr) { VADDR_T oldVaddr = (VADDR_T)(UINTPTR)oldUserVaddr; @@ -930,12 +1246,20 @@ STATIC INT32 OsFutexRequeueParamCheck(const UINT32 *oldUserVaddr, UINT32 flags, if (oldVaddr == newVaddr) { return LOS_EINVAL; } +<<<<<<< HEAD +======= + //检查标记 +>>>>>>> remotes/origin/main if ((flags & (~FUTEX_PRIVATE)) != FUTEX_REQUEUE) { PRINT_ERR("Futex requeue param check failed! error flags: 0x%x\n", flags); return LOS_EINVAL; } +<<<<<<< HEAD +======= + //检查地址范围,必须在用户空间 +>>>>>>> remotes/origin/main if ((oldVaddr % sizeof(INT32)) || (oldVaddr < OS_FUTEX_KEY_BASE) || (oldVaddr >= OS_FUTEX_KEY_MAX)) { PRINT_ERR("Futex requeue param check failed! error old userVaddr: 0x%x\n", oldUserVaddr); return LOS_EINVAL; @@ -948,7 +1272,11 @@ STATIC INT32 OsFutexRequeueParamCheck(const UINT32 *oldUserVaddr, UINT32 flags, return LOS_OK; } +<<<<<<< HEAD +======= +/// 调整指定锁在Futex表中的位置 +>>>>>>> remotes/origin/main INT32 OsFutexRequeue(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber, INT32 count, const UINT32 *newUserVaddr) { INT32 ret; @@ -965,12 +1293,21 @@ INT32 OsFutexRequeue(const UINT32 *userVaddr, UINT32 flags, INT32 wakeNumber, IN return LOS_EINVAL; } +<<<<<<< HEAD oldFutexKey = OsFutexFlagsToKey(userVaddr, flags); newFutexKey = OsFutexFlagsToKey(newUserVaddr, flags); oldIndex = OsFutexKeyToIndex(oldFutexKey, flags); newIndex = OsFutexKeyToIndex(newFutexKey, flags); oldHashNode = &g_futexHash[oldIndex]; +======= + oldFutexKey = OsFutexFlagsToKey(userVaddr, flags);//先拿key + newFutexKey = OsFutexFlagsToKey(newUserVaddr, flags); + oldIndex = OsFutexKeyToIndex(oldFutexKey, flags);//再拿所在哈希桶位置,共有80个哈希桶 + newIndex = OsFutexKeyToIndex(newFutexKey, flags); + + oldHashNode = &g_futexHash[oldIndex];//拿到对应哈希桶实体 +>>>>>>> remotes/origin/main if (OsFutexLock(&oldHashNode->listLock)) { return LOS_EINVAL; } diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_mux.c b/src/kernel_liteos_a/kernel/base/ipc/los_mux.c index fa4a67b8..55ed73c5 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_mux.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_mux.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -40,19 +44,32 @@ #ifdef LOSCFG_BASE_IPC_MUX #define MUTEXATTR_TYPE_MASK 0x0FU +<<<<<<< HEAD +======= +///互斥属性初始化 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrInit(LosMuxAttr *attr) { if (attr == NULL) { return LOS_EINVAL; } +<<<<<<< HEAD attr->protocol = LOS_MUX_PRIO_INHERIT; attr->prioceiling = OS_TASK_PRIORITY_LOWEST; attr->type = LOS_MUX_DEFAULT; return LOS_OK; } +======= + attr->protocol = LOS_MUX_PRIO_INHERIT; //协议默认用继承方式, A(4)task等B(19)释放锁时,B的调度优先级直接升到(4) + attr->prioceiling = OS_TASK_PRIORITY_LOWEST;//最低优先级 + attr->type = LOS_MUX_DEFAULT; //默认 LOS_MUX_RECURSIVE + return LOS_OK; +} +/// ????? 销毁互斥属 ,这里啥也没干呀 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrDestroy(LosMuxAttr *attr) { if (attr == NULL) { @@ -61,7 +78,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrDestroy(LosMuxAttr *attr) return LOS_OK; } +<<<<<<< HEAD +======= +///获取互斥锁的类型属性,由outType接走,不送! +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetType(const LosMuxAttr *attr, INT32 *outType) { INT32 type; @@ -79,7 +100,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetType(const LosMuxAttr *attr, INT32 *outTyp return LOS_OK; } +<<<<<<< HEAD +======= +///设置互斥锁的类型属性 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetType(LosMuxAttr *attr, INT32 type) { if ((attr == NULL) || (type < LOS_MUX_NORMAL) || (type > LOS_MUX_ERRORCHECK)) { @@ -89,7 +114,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetType(LosMuxAttr *attr, INT32 type) attr->type = (UINT8)((attr->type & ~MUTEXATTR_TYPE_MASK) | (UINT32)type); return LOS_OK; } +<<<<<<< HEAD +======= +///获取互斥锁的类型属性 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetProtocol(const LosMuxAttr *attr, INT32 *protocol) { if ((attr != NULL) && (protocol != NULL)) { @@ -100,7 +129,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetProtocol(const LosMuxAttr *attr, INT32 *pr return LOS_OK; } +<<<<<<< HEAD +======= +///设置互斥锁属性的协议 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetProtocol(LosMuxAttr *attr, INT32 protocol) { if (attr == NULL) { @@ -117,7 +150,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetProtocol(LosMuxAttr *attr, INT32 protocol) return LOS_EINVAL; } } +<<<<<<< HEAD +======= +///获取互斥锁属性优先级 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetPrioceiling(const LosMuxAttr *attr, INT32 *prioceiling) { if (attr == NULL) { @@ -130,7 +167,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrGetPrioceiling(const LosMuxAttr *attr, INT32 return LOS_OK; } +<<<<<<< HEAD +======= +///设置互斥锁属性的优先级的上限 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetPrioceiling(LosMuxAttr *attr, INT32 prioceiling) { if ((attr == NULL) || @@ -143,7 +184,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxAttrSetPrioceiling(LosMuxAttr *attr, INT32 prioce return LOS_OK; } +<<<<<<< HEAD +======= +///设置互斥锁的优先级的上限,老优先级由oldPrioceiling带走 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxSetPrioceiling(LosMux *mutex, INT32 prioceiling, INT32 *oldPrioceiling) { INT32 ret; @@ -172,7 +217,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxSetPrioceiling(LosMux *mutex, INT32 prioceiling, return ret; } +<<<<<<< HEAD +======= +///获取互斥锁的优先级的上限 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxGetPrioceiling(const LosMux *mutex, INT32 *prioceiling) { if ((mutex != NULL) && (prioceiling != NULL) && (mutex->magic == OS_MUX_MAGIC)) { @@ -182,7 +231,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxGetPrioceiling(const LosMux *mutex, INT32 *prioce return LOS_EINVAL; } +<<<<<<< HEAD +======= +///互斥锁是否有效 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT BOOL LOS_MuxIsValid(const LosMux *mutex) { if ((mutex != NULL) && (mutex->magic == OS_MUX_MAGIC)) { @@ -191,7 +244,11 @@ LITE_OS_SEC_TEXT BOOL LOS_MuxIsValid(const LosMux *mutex) return FALSE; } +<<<<<<< HEAD +======= +///检查互斥锁属性是否OK,否则 no ok :|) +>>>>>>> remotes/origin/main STATIC UINT32 OsCheckMutexAttr(const LosMuxAttr *attr) { if (((INT8)(attr->type) < LOS_MUX_NORMAL) || (attr->type > LOS_MUX_ERRORCHECK)) { @@ -205,7 +262,11 @@ STATIC UINT32 OsCheckMutexAttr(const LosMuxAttr *attr) } return LOS_OK; } +<<<<<<< HEAD +======= +/// 初始化互斥锁 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxInit(LosMux *mutex, const LosMuxAttr *attr) { UINT32 intSave; @@ -215,6 +276,7 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxInit(LosMux *mutex, const LosMuxAttr *attr) } if (attr == NULL) { +<<<<<<< HEAD (VOID)LOS_MuxAttrInit(&mutex->attr); } else { (VOID)memcpy_s(&mutex->attr, sizeof(LosMuxAttr), attr, sizeof(LosMuxAttr)); @@ -233,6 +295,26 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxInit(LosMux *mutex, const LosMuxAttr *attr) return LOS_OK; } +======= + (VOID)LOS_MuxAttrInit(&mutex->attr);//属性初始化 + } else { + (VOID)memcpy_s(&mutex->attr, sizeof(LosMuxAttr), attr, sizeof(LosMuxAttr));//把attr 拷贝到 mutex->attr + } + + if (OsCheckMutexAttr(&mutex->attr) != LOS_OK) {//检查属性 + return LOS_EINVAL; + } + + SCHEDULER_LOCK(intSave); //拿到调度自旋锁 + mutex->muxCount = 0; //锁定互斥量的次数 + mutex->owner = NULL; //谁持有该锁 + LOS_ListInit(&mutex->muxList); //互斥量双循环链表 + mutex->magic = OS_MUX_MAGIC; //固定标识,互斥锁的魔法数字 + SCHEDULER_UNLOCK(intSave); //释放调度自旋锁 + return LOS_OK; +} +///销毁互斥锁 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxDestroy(LosMux *mutex) { UINT32 intSave; @@ -241,13 +323,20 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxDestroy(LosMux *mutex) return LOS_EINVAL; } +<<<<<<< HEAD SCHEDULER_LOCK(intSave); if (mutex->magic != OS_MUX_MAGIC) { SCHEDULER_UNLOCK(intSave); +======= + SCHEDULER_LOCK(intSave); //保存调度自旋锁 + if (mutex->magic != OS_MUX_MAGIC) { + SCHEDULER_UNLOCK(intSave);//释放调度自旋锁 +>>>>>>> remotes/origin/main return LOS_EBADF; } if (mutex->muxCount != 0) { +<<<<<<< HEAD SCHEDULER_UNLOCK(intSave); return LOS_EBUSY; } @@ -257,6 +346,17 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxDestroy(LosMux *mutex) return LOS_OK; } +======= + SCHEDULER_UNLOCK(intSave);//释放调度自旋锁 + return LOS_EBUSY; + } + + (VOID)memset_s(mutex, sizeof(LosMux), 0, sizeof(LosMux));//很简单,全部清0处理. + SCHEDULER_UNLOCK(intSave); //释放调度自旋锁 + return LOS_OK; +} +///设置互斥锁位图 +>>>>>>> remotes/origin/main STATIC VOID OsMuxBitmapSet(const LosMux *mutex, const LosTaskCB *runTask) { if (mutex->attr.protocol != LOS_MUX_PRIO_INHERIT) { @@ -271,7 +371,11 @@ STATIC VOID OsMuxBitmapSet(const LosMux *mutex, const LosTaskCB *runTask) owner->ops->priorityInheritance(owner, ¶m); } } +<<<<<<< HEAD +======= +///恢复互斥锁位图 +>>>>>>> remotes/origin/main VOID OsMuxBitmapRestore(const LosMux *mutex, const LOS_DL_LIST *list, const LosTaskCB *runTask) { if (mutex->attr.protocol != LOS_MUX_PRIO_INHERIT) { @@ -284,10 +388,15 @@ VOID OsMuxBitmapRestore(const LosMux *mutex, const LOS_DL_LIST *list, const LosT owner->ops->priorityRestore(owner, list, ¶m); } +<<<<<<< HEAD +======= +/// 最坏情况就是拿锁失败,让出CPU,变成阻塞任务,等别的任务释放锁后排到自己了接着执行. +>>>>>>> remotes/origin/main STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) { UINT32 ret; +<<<<<<< HEAD if ((mutex->muxList.pstPrev == NULL) || (mutex->muxList.pstNext == NULL)) { /* This is for mutex macro initialization. */ mutex->muxCount = 0; @@ -298,6 +407,18 @@ STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) if (mutex->muxCount == 0) { mutex->muxCount++; mutex->owner = (VOID *)runTask; +======= + if ((mutex->muxList.pstPrev == NULL) || (mutex->muxList.pstNext == NULL)) {//列表为空时的处理 + /* This is for mutex macro initialization. */ + mutex->muxCount = 0;//锁计数器清0 + mutex->owner = NULL;//锁没有归属任务 + LOS_ListInit(&mutex->muxList);//初始化锁的任务链表,后续申请这把锁任务都会挂上去 + } + + if (mutex->muxCount == 0) {//无task用锁时,肯定能拿到锁了.在里面返回 + mutex->muxCount++; //互斥锁计数器加1 + mutex->owner = (VOID *)runTask; //当前任务拿到锁 +>>>>>>> remotes/origin/main LOS_ListTailInsert(&runTask->lockList, &mutex->holdList); if (mutex->attr.protocol == LOS_MUX_PRIO_PROTECT) { SchedParam param = { 0 }; @@ -307,6 +428,7 @@ STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) } return LOS_OK; } +<<<<<<< HEAD if (((LosTaskCB *)mutex->owner == runTask) && (mutex->attr.type == LOS_MUX_RECURSIVE)) { mutex->muxCount++; @@ -324,6 +446,25 @@ STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) OsMuxBitmapSet(mutex, runTask); runTask->taskMux = (VOID *)mutex; +======= + //递归锁muxCount>0 如果是递归锁就要处理两种情况 1.runtask持有锁 2.锁被别的任务拿走了 + if (((LosTaskCB *)mutex->owner == runTask) && (mutex->attr.type == LOS_MUX_RECURSIVE)) {//第一种情况 runtask是锁持有方 + mutex->muxCount++; //递归锁计数器加1,递归锁的目的是防止死锁,鸿蒙默认用的就是递归锁(LOS_MUX_DEFAULT = LOS_MUX_RECURSIVE) + return LOS_OK; //成功退出 + } + //到了这里说明锁在别的任务那里,当前任务只能被阻塞了. + if (!timeout) {//参数timeout表示等待多久再来拿锁 + return LOS_EINVAL;//timeout = 0表示不等了,没拿到锁就返回不纠结,返回错误.见于LOS_MuxTrylock + } + //自己要被阻塞,只能申请调度,让出CPU core 让别的任务上 + if (!OsPreemptableInSched()) {//不能申请调度 (不能调度的原因是因为没有持有调度任务自旋锁) + return LOS_EDEADLK;//返回错误,自旋锁被别的CPU core 持有 + } + + OsMuxBitmapSet(mutex, runTask);//设置锁位图,尽可能的提高锁持有任务的优先级 + + runTask->taskMux = (VOID *)mutex; //记下当前任务在等待这把锁 +>>>>>>> remotes/origin/main LOS_DL_LIST *node = OsSchedLockPendFindPos(runTask, &mutex->muxList); if (node == NULL) { ret = LOS_NOK; @@ -332,10 +473,17 @@ STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) OsTaskWaitSetPendMask(OS_TASK_WAIT_MUTEX, (UINTPTR)mutex, timeout); ret = runTask->ops->wait(runTask, node, timeout); +<<<<<<< HEAD if (ret == LOS_ERRNO_TSK_TIMEOUT) { OsMuxBitmapRestore(mutex, NULL, runTask); runTask->taskMux = NULL; ret = LOS_ETIMEDOUT; +======= + if (ret == LOS_ERRNO_TSK_TIMEOUT) {//这行代码虽和OsTaskWait挨在一起,但要过很久才会执行到,因为在OsTaskWait中CPU切换了任务上下文 + OsMuxBitmapRestore(mutex, NULL, runTask); + runTask->taskMux = NULL;// 所以重新回到这里时可能已经超时了 + ret = LOS_ETIMEDOUT;//返回超时 +>>>>>>> remotes/origin/main } return ret; @@ -343,7 +491,11 @@ STATIC UINT32 OsMuxPendOp(LosTaskCB *runTask, LosMux *mutex, UINT32 timeout) UINT32 OsMuxLockUnsafe(LosMux *mutex, UINT32 timeout) { +<<<<<<< HEAD LosTaskCB *runTask = OsCurrTaskGet(); +======= + LosTaskCB *runTask = OsCurrTaskGet();//获取当前任务 +>>>>>>> remotes/origin/main if (mutex->magic != OS_MUX_MAGIC) { return LOS_EBADF; @@ -352,13 +504,18 @@ UINT32 OsMuxLockUnsafe(LosMux *mutex, UINT32 timeout) if (OsCheckMutexAttr(&mutex->attr) != LOS_OK) { return LOS_EINVAL; } +<<<<<<< HEAD +======= + //LOS_MUX_ERRORCHECK 时 muxCount是要等于0 ,当前任务持有锁就不能再lock了. 鸿蒙默认用的是递归锁LOS_MUX_RECURSIVE +>>>>>>> remotes/origin/main if ((mutex->attr.type == LOS_MUX_ERRORCHECK) && (mutex->owner == (VOID *)runTask)) { return LOS_EDEADLK; } return OsMuxPendOp(runTask, mutex, timeout); } +<<<<<<< HEAD UINT32 OsMuxTrylockUnsafe(LosMux *mutex, UINT32 timeout) { @@ -369,6 +526,18 @@ UINT32 OsMuxTrylockUnsafe(LosMux *mutex, UINT32 timeout) } if (OsCheckMutexAttr(&mutex->attr) != LOS_OK) { +======= +/// 尝试加锁, +UINT32 OsMuxTrylockUnsafe(LosMux *mutex, UINT32 timeout) +{ + LosTaskCB *runTask = OsCurrTaskGet();//获取当前任务 + + if (mutex->magic != OS_MUX_MAGIC) {//检查MAGIC有没有被改变 + return LOS_EBADF; + } + + if (OsCheckMutexAttr(&mutex->attr) != LOS_OK) {//检查互斥锁属性 +>>>>>>> remotes/origin/main return LOS_EINVAL; } @@ -377,9 +546,15 @@ UINT32 OsMuxTrylockUnsafe(LosMux *mutex, UINT32 timeout) return LOS_EBUSY; } +<<<<<<< HEAD return OsMuxPendOp(runTask, mutex, timeout); } +======= + return OsMuxPendOp(runTask, mutex, timeout);//当前任务去拿锁,拿不到就等timeout +} +/// 拿互斥锁, +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxLock(LosMux *mutex, UINT32 timeout) { LosTaskCB *runTask = NULL; @@ -394,6 +569,7 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxLock(LosMux *mutex, UINT32 timeout) return LOS_EINTR; } +<<<<<<< HEAD runTask = (LosTaskCB *)OsCurrTaskGet(); /* DO NOT Call blocking API in system tasks */ if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) { @@ -407,6 +583,21 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxLock(LosMux *mutex, UINT32 timeout) return ret; } +======= + runTask = (LosTaskCB *)OsCurrTaskGet();//获取当前任务 + /* DO NOT Call blocking API in system tasks */ + if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) {//不要在内核任务里用mux锁 + PRINTK("Warning: DO NOT call %s in system tasks.\n", __FUNCTION__); + OsBackTrace();//打印task信息 + } + + SCHEDULER_LOCK(intSave);//调度自旋锁 + ret = OsMuxLockUnsafe(mutex, timeout);//如果任务没拿到锁,将进入阻塞队列一直等待,直到timeout或者持锁任务释放锁时唤醒它 + SCHEDULER_UNLOCK(intSave); + return ret; +} +///尝试要锁,没拿到也不等,直接返回,不纠结 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxTrylock(LosMux *mutex) { LosTaskCB *runTask = NULL; @@ -421,27 +612,46 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxTrylock(LosMux *mutex) return LOS_EINTR; } +<<<<<<< HEAD runTask = (LosTaskCB *)OsCurrTaskGet(); /* DO NOT Call blocking API in system tasks */ if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) { +======= + runTask = (LosTaskCB *)OsCurrTaskGet();//获取当前执行的任务 + /* DO NOT Call blocking API in system tasks */ + if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) {//系统任务不能 +>>>>>>> remotes/origin/main PRINTK("Warning: DO NOT call %s in system tasks.\n", __FUNCTION__); OsBackTrace(); } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsMuxTrylockUnsafe(mutex, 0); +======= + ret = OsMuxTrylockUnsafe(mutex, 0);//timeout = 0,不等待,没拿到锁就算了 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD STATIC UINT32 OsMuxPostOp(LosTaskCB *taskCB, LosMux *mutex, BOOL *needSched) { if (LOS_ListEmpty(&mutex->muxList)) { LOS_ListDelete(&mutex->holdList); +======= + +STATIC UINT32 OsMuxPostOp(LosTaskCB *taskCB, LosMux *mutex, BOOL *needSched) +{ + if (LOS_ListEmpty(&mutex->muxList)) {//如果互斥锁列表为空 + LOS_ListDelete(&mutex->holdList);//把持有互斥锁的节点摘掉 +>>>>>>> remotes/origin/main mutex->owner = NULL; return LOS_OK; } +<<<<<<< HEAD LosTaskCB *resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(mutex->muxList))); OsMuxBitmapRestore(mutex, &mutex->muxList, resumedTask); @@ -454,6 +664,20 @@ STATIC UINT32 OsMuxPostOp(LosTaskCB *taskCB, LosMux *mutex, BOOL *needSched) resumedTask->taskMux = NULL; if (needSched != NULL) { *needSched = TRUE; +======= + LosTaskCB *resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(mutex->muxList)));//拿到等待互斥锁链表的第一个任务实体,接下来要唤醒任务 + OsMuxBitmapRestore(mutex, &mutex->muxList, resumedTask); + + mutex->muxCount = 1;//互斥锁数量为1 + mutex->owner = (VOID *)resumedTask;//互斥锁的持有人换了 + LOS_ListDelete(&mutex->holdList);//自然要从等锁链表中把自己摘出去 + LOS_ListTailInsert(&resumedTask->lockList, &mutex->holdList);//把锁挂到恢复任务的锁链表上,lockList是任务持有的所有锁记录 + OsTaskWakeClearPendMask(resumedTask); + resumedTask->ops->wake(resumedTask); + resumedTask->taskMux = NULL; + if (needSched != NULL) {//如果不为空 + *needSched = TRUE;//就走起再次调度流程 +>>>>>>> remotes/origin/main } return LOS_OK; @@ -476,21 +700,37 @@ UINT32 OsMuxUnlockUnsafe(LosTaskCB *taskCB, LosMux *mutex, BOOL *needSched) if (mutex->muxCount == 0) { return LOS_EPERM; } +<<<<<<< HEAD if ((--mutex->muxCount != 0) && (mutex->attr.type == LOS_MUX_RECURSIVE)) { return LOS_OK; } if (mutex->attr.protocol == LOS_MUX_PRIO_PROTECT) { +======= + //注意 --mutex->muxCount 先执行了-- 操作. + if ((--mutex->muxCount != 0) && (mutex->attr.type == LOS_MUX_RECURSIVE)) {//属性类型为LOS_MUX_RECURSIVE时,muxCount是可以不为0的 + return LOS_OK; + } + + if (mutex->attr.protocol == LOS_MUX_PRIO_PROTECT) {//属性协议为保护时 +>>>>>>> remotes/origin/main SchedParam param = { 0 }; taskCB->ops->schedParamGet(taskCB, ¶m); taskCB->ops->priorityRestore(taskCB, NULL, ¶m); } +<<<<<<< HEAD /* Whether a task block the mutex lock. */ return OsMuxPostOp(taskCB, mutex, needSched); } +======= + /* Whether a task block the mutex lock. *///任务是否阻塞互斥锁 + return OsMuxPostOp(taskCB, mutex, needSched);//一个任务去唤醒另一个在等锁的任务 +} +///释放锁 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_MuxUnlock(LosMux *mutex) { LosTaskCB *runTask = NULL; @@ -506,9 +746,15 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxUnlock(LosMux *mutex) return LOS_EINTR; } +<<<<<<< HEAD runTask = (LosTaskCB *)OsCurrTaskGet(); /* DO NOT Call blocking API in system tasks */ if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) { +======= + runTask = (LosTaskCB *)OsCurrTaskGet();//获取当前任务 + /* DO NOT Call blocking API in system tasks */ + if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) {//不能在系统任务里调用,因为很容易让系统任务发生死锁 +>>>>>>> remotes/origin/main PRINTK("Warning: DO NOT call %s in system tasks.\n", __FUNCTION__); OsBackTrace(); } @@ -516,12 +762,22 @@ LITE_OS_SEC_TEXT UINT32 LOS_MuxUnlock(LosMux *mutex) SCHEDULER_LOCK(intSave); ret = OsMuxUnlockUnsafe(runTask, mutex, &needSched); SCHEDULER_UNLOCK(intSave); +<<<<<<< HEAD if (needSched == TRUE) { LOS_MpSchedule(OS_MP_CPU_ALL); LOS_Schedule(); +======= + if (needSched == TRUE) {//需要调度的情况 + LOS_MpSchedule(OS_MP_CPU_ALL);//向所有CPU发送调度指令 + LOS_Schedule();//发起调度 +>>>>>>> remotes/origin/main } return ret; } +<<<<<<< HEAD #endif /* LOSCFG_BASE_IPC_MUX */ +======= +#endif /* (LOSCFG_BASE_IPC_MUX == YES) */ +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_queue.c b/src/kernel_liteos_a/kernel/base/ipc/los_queue.c index 8fe06ba2..9924eb73 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_queue.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_queue.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -47,8 +51,13 @@ #endif /* LOSCFG_BASE_IPC_QUEUE_LIMIT <= 0 */ #ifndef LOSCFG_IPC_CONTAINER +<<<<<<< HEAD LITE_OS_SEC_BSS LosQueueCB *g_allQueue = NULL; LITE_OS_SEC_BSS STATIC LOS_DL_LIST g_freeQueueList; +======= +LITE_OS_SEC_BSS LosQueueCB *g_allQueue = NULL;///< 消息队列池 +LITE_OS_SEC_BSS STATIC LOS_DL_LIST g_freeQueueList;///< 空闲队列链表,管分配的,需要队列从这里申请 +>>>>>>> remotes/origin/main #define FREE_QUEUE_LIST g_freeQueueList #endif @@ -83,10 +92,13 @@ LITE_OS_SEC_TEXT_INIT LosQueueCB *OsAllQueueCBInit(LOS_DL_LIST *freeQueueList) return allQueue; } +<<<<<<< HEAD /* * Description : queue initial * Return : LOS_OK on success or error code on failure */ +======= +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsQueueInit(VOID) { #ifndef LOSCFG_IPC_CONTAINER @@ -97,7 +109,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsQueueInit(VOID) #endif return LOS_OK; } +<<<<<<< HEAD +======= +///创建一个队列,根据用户传入队列长度和消息节点大小来开辟相应的内存空间以供该队列使用,参数queueID带走队列ID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueCreate(CHAR *queueName, UINT16 len, UINT32 *queueID, UINT32 flags, UINT16 maxMsgSize) { @@ -114,7 +130,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueCreate(CHAR *queueName, UINT16 len, UINT32 return LOS_ERRNO_QUEUE_CREAT_PTR_NULL; } +<<<<<<< HEAD if (maxMsgSize > (OS_NULL_SHORT - sizeof(UINT32))) { +======= + if (maxMsgSize > (OS_NULL_SHORT - sizeof(UINT32))) {// maxMsgSize上限 为啥要减去 sizeof(UINT32) ,因为前面存的是队列的大小 +>>>>>>> remotes/origin/main return LOS_ERRNO_QUEUE_SIZE_TOO_BIG; } @@ -122,6 +142,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueCreate(CHAR *queueName, UINT16 len, UINT32 return LOS_ERRNO_QUEUE_PARA_ISZERO; } +<<<<<<< HEAD msgSize = maxMsgSize + sizeof(UINT32); /* * Memory allocation is time-consuming, to shorten the time of disable interrupt, @@ -175,17 +196,79 @@ STATIC LITE_OS_SEC_TEXT UINT32 OsQueueReadParameterCheck(UINT32 queueID, const V if ((*bufferSize == 0) || (*bufferSize > (OS_NULL_SHORT - sizeof(UINT32)))) { return LOS_ERRNO_QUEUE_READSIZE_IS_INVALID; +======= + msgSize = maxMsgSize + sizeof(UINT32);//总size = 消息体内容长度 + 消息大小(UINT32) + /* + * Memory allocation is time-consuming, to shorten the time of disable interrupt, + * move the memory allocation to here. + *///内存分配非常耗时,为了缩短禁用中断的时间,将内存分配移到此处,用的时候分配队列内存 + queue = (UINT8 *)LOS_MemAlloc(m_aucSysMem1, (UINT32)len * msgSize);//从系统内存池中分配,由这里提供读写队列的内存 + if (queue == NULL) {//这里是一次把队列要用到的所有最大内存都申请下来了,能保证不会出现后续使用过程中内存不够的问题出现 + return LOS_ERRNO_QUEUE_CREATE_NO_MEMORY;//调用处有 OsSwtmrInit sys_mbox_new DoMqueueCreate == + } + + SCHEDULER_LOCK(intSave); + if (LOS_ListEmpty(&FREE_QUEUE_LIST)) {//没有空余的队列ID的处理,注意软时钟定时器是由 g_swtmrCBArray统一管理的,里面有正在使用和可分配空闲的队列 + SCHEDULER_UNLOCK(intSave);//g_freeQueueList是管理可用于分配的队列链表,申请消息队列的ID需要向它要 + OsQueueCheckHook(); + (VOID)LOS_MemFree(m_aucSysMem1, queue);//没有就要释放 queue申请的内存 + return LOS_ERRNO_QUEUE_CB_UNAVAILABLE; + } + + unusedQueue = LOS_DL_LIST_FIRST(&FREE_QUEUE_LIST);//找到一个没有被使用的队列 + LOS_ListDelete(unusedQueue);//将自己从g_freeQueueList中摘除, unusedQueue只是个 LOS_DL_LIST 结点. + queueCB = GET_QUEUE_LIST(unusedQueue);//通过unusedQueue找到整个消息队列(LosQueueCB) + queueCB->queueLen = len; //队列中消息的总个数,注意这个一旦创建是不能变的. + queueCB->queueSize = msgSize;//消息节点的大小,注意这个一旦创建也是不能变的. + queueCB->queueHandle = queue; //队列句柄,队列内容存储区. + queueCB->queueState = OS_QUEUE_INUSED; //队列状态使用中 + queueCB->readWriteableCnt[OS_QUEUE_READ] = 0;//可读资源计数,OS_QUEUE_READ(0):可读. + queueCB->readWriteableCnt[OS_QUEUE_WRITE] = len;//可些资源计数 OS_QUEUE_WRITE(1):可写, 默认len可写. + queueCB->queueHead = 0;//队列头节点 + queueCB->queueTail = 0;//队列尾节点 + LOS_ListInit(&queueCB->readWriteList[OS_QUEUE_READ]);//初始化可读队列任务链表 + LOS_ListInit(&queueCB->readWriteList[OS_QUEUE_WRITE]);//初始化可写队列任务链表 + LOS_ListInit(&queueCB->memList);// + + OsQueueDbgUpdateHook(queueCB->queueID, OsCurrTaskGet()->taskEntry);//在创建或删除队列调试信息时更新任务条目 + SCHEDULER_UNLOCK(intSave); + + *queueID = queueCB->queueID;//带走队列ID + OsHookCall(LOS_HOOK_TYPE_QUEUE_CREATE, queueCB); + return LOS_OK; +} +///读队列参数检查 +STATIC LITE_OS_SEC_TEXT UINT32 OsQueueReadParameterCheck(UINT32 queueID, const VOID *bufferAddr, + const UINT32 *bufferSize, UINT32 timeout) +{ + if (GET_QUEUE_INDEX(queueID) >= LOSCFG_BASE_IPC_QUEUE_LIMIT) {//队列ID不能超上限 + return LOS_ERRNO_QUEUE_INVALID; + } + if ((bufferAddr == NULL) || (bufferSize == NULL)) {//缓存地址和大小参数判断 + return LOS_ERRNO_QUEUE_READ_PTR_NULL; + } + + if ((*bufferSize == 0) || (*bufferSize > (OS_NULL_SHORT - sizeof(UINT32)))) {//限制了读取数据的上限64K, sizeof(UINT32)代表的是队列的长度 + return LOS_ERRNO_QUEUE_READSIZE_IS_INVALID; //所以要减去 +>>>>>>> remotes/origin/main } OsQueueDbgTimeUpdateHook(queueID); +<<<<<<< HEAD if (timeout != LOS_NO_WAIT) { if (OS_INT_ACTIVE) { return LOS_ERRNO_QUEUE_READ_IN_INTERRUPT; +======= + if (timeout != LOS_NO_WAIT) {//等待一定时间再读取 + if (OS_INT_ACTIVE) {//如果碰上了硬中断 + return LOS_ERRNO_QUEUE_READ_IN_INTERRUPT;//意思是:硬中断发生时是不能读消息队列的 +>>>>>>> remotes/origin/main } } return LOS_OK; } +<<<<<<< HEAD STATIC LITE_OS_SEC_TEXT UINT32 OsQueueWriteParameterCheck(UINT32 queueID, const VOID *bufferAddr, const UINT32 *bufferSize, UINT32 timeout) @@ -199,6 +282,21 @@ STATIC LITE_OS_SEC_TEXT UINT32 OsQueueWriteParameterCheck(UINT32 queueID, const } if (*bufferSize == 0) { +======= +///写队列参数检查 +STATIC LITE_OS_SEC_TEXT UINT32 OsQueueWriteParameterCheck(UINT32 queueID, const VOID *bufferAddr, + const UINT32 *bufferSize, UINT32 timeout) +{ + if (GET_QUEUE_INDEX(queueID) >= LOSCFG_BASE_IPC_QUEUE_LIMIT) {//队列ID不能超上限 + return LOS_ERRNO_QUEUE_INVALID; + } + + if (bufferAddr == NULL) {//没有数据源 + return LOS_ERRNO_QUEUE_WRITE_PTR_NULL; + } + + if (*bufferSize == 0) {//这里没有限制写队列的大小,如果写入一个很大buf 会怎样? +>>>>>>> remotes/origin/main return LOS_ERRNO_QUEUE_WRITESIZE_ISZERO; } @@ -211,13 +309,18 @@ STATIC LITE_OS_SEC_TEXT UINT32 OsQueueWriteParameterCheck(UINT32 queueID, const } return LOS_OK; } +<<<<<<< HEAD +======= +///队列buf操作,注意队列数据是按顺序来读取的,要不从头,要不从尾部,不会出现从中间读写,所有可由 head 和 tail 来管理队列. +>>>>>>> remotes/origin/main STATIC VOID OsQueueBufferOperate(LosQueueCB *queueCB, UINT32 operateType, VOID *bufferAddr, UINT32 *bufferSize) { UINT8 *queueNode = NULL; UINT32 msgDataSize; UINT16 queuePosition; +<<<<<<< HEAD /* get the queue position */ switch (OS_QUEUE_OPERATE_GET(operateType)) { case OS_QUEUE_READ_HEAD: @@ -231,26 +334,55 @@ STATIC VOID OsQueueBufferOperate(LosQueueCB *queueCB, UINT32 operateType, VOID * case OS_QUEUE_WRITE_TAIL: queuePosition = queueCB->queueTail; ((queueCB->queueTail + 1) == queueCB->queueLen) ? (queueCB->queueTail = 0) : (queueCB->queueTail++); +======= + /* get the queue position | 先找到队列的位置*/ + switch (OS_QUEUE_OPERATE_GET(operateType)) {//获取操作类型 + case OS_QUEUE_READ_HEAD://从列队头开始读 + queuePosition = queueCB->queueHead;//拿到头部位置 + ((queueCB->queueHead + 1) == queueCB->queueLen) ? (queueCB->queueHead = 0) : (queueCB->queueHead++);//调整队列头部位置 + break; + case OS_QUEUE_WRITE_HEAD://从列队头开始写 + (queueCB->queueHead == 0) ? (queueCB->queueHead = queueCB->queueLen - 1) : (--queueCB->queueHead);//调整队列头部位置 + queuePosition = queueCB->queueHead;//拿到头部位置 + break; + case OS_QUEUE_WRITE_TAIL://从列队尾部开始写 + queuePosition = queueCB->queueTail;//设置队列位置为尾部位置 + ((queueCB->queueTail + 1) == queueCB->queueLen) ? (queueCB->queueTail = 0) : (queueCB->queueTail++);//调整队列尾部位置 +>>>>>>> remotes/origin/main break; default: /* read tail, reserved. */ PRINT_ERR("invalid queue operate type!\n"); return; } +<<<<<<< HEAD queueNode = &(queueCB->queueHandle[(queuePosition * (queueCB->queueSize))]); if (OS_QUEUE_IS_READ(operateType)) { if (memcpy_s(&msgDataSize, sizeof(UINT32), queueNode + queueCB->queueSize - sizeof(UINT32), sizeof(UINT32)) != EOK) { +======= + //queueHandle是create队列时,由外界参数申请的一块内存. 用于copy 使用 + queueNode = &(queueCB->queueHandle[(queuePosition * (queueCB->queueSize))]);//拿到队列节点 + + if (OS_QUEUE_IS_READ(operateType)) {//读操作处理,读队列分两步走 + if (memcpy_s(&msgDataSize, sizeof(UINT32), queueNode + queueCB->queueSize - sizeof(UINT32), + sizeof(UINT32)) != EOK) {//1.先读出队列大小,由队列头四个字节表示 +>>>>>>> remotes/origin/main PRINT_ERR("get msgdatasize failed\n"); return; } msgDataSize = (*bufferSize < msgDataSize) ? *bufferSize : msgDataSize; +<<<<<<< HEAD if (memcpy_s(bufferAddr, *bufferSize, queueNode, msgDataSize) != EOK) { +======= + if (memcpy_s(bufferAddr, *bufferSize, queueNode, msgDataSize) != EOK) {//2.读表示读走已有数据,所以相当于bufferAddr接着了queueNode的数据 +>>>>>>> remotes/origin/main PRINT_ERR("copy message to buffer failed\n"); return; } +<<<<<<< HEAD *bufferSize = msgDataSize; } else { if (memcpy_s(queueNode, queueCB->queueSize, bufferAddr, *bufferSize) != EOK) { @@ -259,11 +391,22 @@ STATIC VOID OsQueueBufferOperate(LosQueueCB *queueCB, UINT32 operateType, VOID * } if (memcpy_s(queueNode + queueCB->queueSize - sizeof(UINT32), sizeof(UINT32), bufferSize, sizeof(UINT32)) != EOK) { +======= + *bufferSize = msgDataSize;//通过入参 带走消息的大小 + } else {//只有读写两种操作,这里就是写队列了.写也分两步走 , @note_thinking 这里建议鸿蒙加上 OS_QUEUE_IS_WRITE 判断 + if (memcpy_s(queueNode, queueCB->queueSize, bufferAddr, *bufferSize) != EOK) {//1.写入消息内容 + PRINT_ERR("store message failed\n");//表示把外面数据写进来,所以相当于queueNode接着了bufferAddr的数据 + return; + } + if (memcpy_s(queueNode + queueCB->queueSize - sizeof(UINT32), sizeof(UINT32), bufferSize, + sizeof(UINT32)) != EOK) {//2.写入消息数据的长度,sizeof(UINT32) +>>>>>>> remotes/origin/main PRINT_ERR("store message size failed\n"); return; } } } +<<<<<<< HEAD STATIC UINT32 OsQueueOperateParamCheck(const LosQueueCB *queueCB, UINT32 queueID, UINT32 operateType, const UINT32 *bufferSize) @@ -274,30 +417,61 @@ STATIC UINT32 OsQueueOperateParamCheck(const LosQueueCB *queueCB, UINT32 queueID if (OS_QUEUE_IS_WRITE(operateType) && (*bufferSize > (queueCB->queueSize - sizeof(UINT32)))) { return LOS_ERRNO_QUEUE_WRITE_SIZE_TOO_BIG; +======= +///队列操作参数检查 +STATIC UINT32 OsQueueOperateParamCheck(const LosQueueCB *queueCB, UINT32 queueID, + UINT32 operateType, const UINT32 *bufferSize) +{ + if ((queueCB->queueID != queueID) || (queueCB->queueState == OS_QUEUE_UNUSED)) {//队列ID和状态判断 + return LOS_ERRNO_QUEUE_NOT_CREATE; + } + + if (OS_QUEUE_IS_WRITE(operateType) && (*bufferSize > (queueCB->queueSize - sizeof(UINT32)))) {//写时判断 + return LOS_ERRNO_QUEUE_WRITE_SIZE_TOO_BIG;//塞进来的数据太大,大于队列节点能承受的范围 +>>>>>>> remotes/origin/main } return LOS_OK; } +<<<<<<< HEAD UINT32 OsQueueOperate(UINT32 queueID, UINT32 operateType, VOID *bufferAddr, UINT32 *bufferSize, UINT32 timeout) { UINT32 ret; UINT32 readWrite = OS_QUEUE_READ_WRITE_GET(operateType); +======= + +UINT32 OsQueueOperate(UINT32 queueID, UINT32 operateType, VOID *bufferAddr, UINT32 *bufferSize, UINT32 timeout) +{ + UINT32 ret; + UINT32 readWrite = OS_QUEUE_READ_WRITE_GET(operateType);//获取读/写操作标识 +>>>>>>> remotes/origin/main UINT32 intSave; OsHookCall(LOS_HOOK_TYPE_QUEUE_READ, (LosQueueCB *)GET_QUEUE_HANDLE(queueID), operateType, *bufferSize, timeout); SCHEDULER_LOCK(intSave); +<<<<<<< HEAD LosQueueCB *queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID); ret = OsQueueOperateParamCheck(queueCB, queueID, operateType, bufferSize); +======= + LosQueueCB *queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID);//获取对应的队列控制块 + ret = OsQueueOperateParamCheck(queueCB, queueID, operateType, bufferSize);//参数检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { goto QUEUE_END; } +<<<<<<< HEAD if (queueCB->readWriteableCnt[readWrite] == 0) { if (timeout == LOS_NO_WAIT) { +======= + if (queueCB->readWriteableCnt[readWrite] == 0) {//根据readWriteableCnt判断队列是否有消息读/写 + if (timeout == LOS_NO_WAIT) {//不等待直接退出 +>>>>>>> remotes/origin/main ret = OS_QUEUE_IS_READ(operateType) ? LOS_ERRNO_QUEUE_ISEMPTY : LOS_ERRNO_QUEUE_ISFULL; goto QUEUE_END; } +<<<<<<< HEAD if (!OsPreemptableInSched()) { ret = LOS_ERRNO_QUEUE_PEND_IN_LOCK; goto QUEUE_END; @@ -326,13 +500,47 @@ UINT32 OsQueueOperate(UINT32 queueID, UINT32 operateType, VOID *bufferAddr, UINT return LOS_OK; } else { queueCB->readWriteableCnt[!readWrite]++; +======= + if (!OsPreemptableInSched()) {//不支持抢占式调度 + ret = LOS_ERRNO_QUEUE_PEND_IN_LOCK; + goto QUEUE_END; + } + //任务等待,这里很重要啊,将自己从就绪列表摘除,让出了CPU并发起了调度,并挂在readWriteList[readWrite]上,挂的都等待读/写消息的task + LosTaskCB *runTask = OsCurrTaskGet(); + OsTaskWaitSetPendMask(OS_TASK_WAIT_QUEUE, queueCB->queueID, timeout); + ret = runTask->ops->wait(runTask, &queueCB->readWriteList[readWrite], timeout); + if (ret == LOS_ERRNO_TSK_TIMEOUT) {//唤醒后如果超时了,返回读/写消息失败 + ret = LOS_ERRNO_QUEUE_TIMEOUT; + goto QUEUE_END;// + } + } else { + queueCB->readWriteableCnt[readWrite]--;//对应队列中计数器--,说明一条消息只能被读/写一次 + } + + OsQueueBufferOperate(queueCB, operateType, bufferAddr, bufferSize);//发起读或写队列操作 + + if (!LOS_ListEmpty(&queueCB->readWriteList[!readWrite])) {//如果还有任务在排着队等待读/写入消息(当时不能读/写的原因有可能当时队列满了==) + LosTaskCB *resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&queueCB->readWriteList[!readWrite]));//取出要读/写消息的任务 + OsTaskWakeClearPendMask(resumedTask); + resumedTask->ops->wake(resumedTask); + SCHEDULER_UNLOCK(intSave); + LOS_MpSchedule(OS_MP_CPU_ALL);//让所有CPU发出调度申请,因为很可能那个要读/写消息的队列是由其他CPU执行 + LOS_Schedule();//申请调度 + return LOS_OK; + } else { + queueCB->readWriteableCnt[!readWrite]++;//对应队列读/写中计数器++ +>>>>>>> remotes/origin/main } QUEUE_END: SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///接口函数定时读取消息队列 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_QueueReadCopy(UINT32 queueID, VOID *bufferAddr, UINT32 *bufferSize, @@ -341,15 +549,26 @@ LITE_OS_SEC_TEXT UINT32 LOS_QueueReadCopy(UINT32 queueID, UINT32 ret; UINT32 operateType; +<<<<<<< HEAD ret = OsQueueReadParameterCheck(queueID, bufferAddr, bufferSize, timeout); +======= + ret = OsQueueReadParameterCheck(queueID, bufferAddr, bufferSize, timeout);//参数检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_READ, OS_QUEUE_HEAD); return OsQueueOperate(queueID, operateType, bufferAddr, bufferSize, timeout); } +======= + operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_READ, OS_QUEUE_HEAD);//从头开始读 + return OsQueueOperate(queueID, operateType, bufferAddr, bufferSize, timeout);//定时执行读操作 +} +///接口函数从队列头开始写 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteHeadCopy(UINT32 queueID, VOID *bufferAddr, UINT32 bufferSize, @@ -358,15 +577,26 @@ LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteHeadCopy(UINT32 queueID, UINT32 ret; UINT32 operateType; +<<<<<<< HEAD ret = OsQueueWriteParameterCheck(queueID, bufferAddr, &bufferSize, timeout); +======= + ret = OsQueueWriteParameterCheck(queueID, bufferAddr, &bufferSize, timeout);//参数检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_WRITE, OS_QUEUE_HEAD); return OsQueueOperate(queueID, operateType, bufferAddr, &bufferSize, timeout); } +======= + operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_WRITE, OS_QUEUE_HEAD);//从头开始写 + return OsQueueOperate(queueID, operateType, bufferAddr, &bufferSize, timeout);//执行写操作 +} +///接口函数 从队列尾部开始写 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteCopy(UINT32 queueID, VOID *bufferAddr, UINT32 bufferSize, @@ -375,15 +605,27 @@ LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteCopy(UINT32 queueID, UINT32 ret; UINT32 operateType; +<<<<<<< HEAD ret = OsQueueWriteParameterCheck(queueID, bufferAddr, &bufferSize, timeout); +======= + ret = OsQueueWriteParameterCheck(queueID, bufferAddr, &bufferSize, timeout);//参数检查 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_WRITE, OS_QUEUE_TAIL); return OsQueueOperate(queueID, operateType, bufferAddr, &bufferSize, timeout); } +======= + operateType = OS_QUEUE_OPERATE_TYPE(OS_QUEUE_WRITE, OS_QUEUE_TAIL);//从尾部开始写 + return OsQueueOperate(queueID, operateType, bufferAddr, &bufferSize, timeout);//执行写操作 +} + + +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_QueueRead(UINT32 queueID, VOID *bufferAddr, UINT32 bufferSize, UINT32 timeout) { return LOS_QueueReadCopy(queueID, bufferAddr, &bufferSize, timeout); @@ -398,6 +640,10 @@ LITE_OS_SEC_TEXT UINT32 LOS_QueueWrite(UINT32 queueID, VOID *bufferAddr, UINT32 return LOS_QueueWriteCopy(queueID, &bufferAddr, bufferSize, timeout); } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteHead(UINT32 queueID, VOID *bufferAddr, UINT32 bufferSize, @@ -410,6 +656,10 @@ LITE_OS_SEC_TEXT UINT32 LOS_QueueWriteHead(UINT32 queueID, return LOS_QueueWriteHeadCopy(queueID, &bufferAddr, bufferSize, timeout); } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueDelete(UINT32 queueID) { LosQueueCB *queueCB = NULL; @@ -422,33 +672,54 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueDelete(UINT32 queueID) } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID); +======= + queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID);//拿到队列实体 +>>>>>>> remotes/origin/main if ((queueCB->queueID != queueID) || (queueCB->queueState == OS_QUEUE_UNUSED)) { ret = LOS_ERRNO_QUEUE_NOT_CREATE; goto QUEUE_END; } +<<<<<<< HEAD if (!LOS_ListEmpty(&queueCB->readWriteList[OS_QUEUE_READ])) { +======= + if (!LOS_ListEmpty(&queueCB->readWriteList[OS_QUEUE_READ])) {//尚有任务要读数据 +>>>>>>> remotes/origin/main ret = LOS_ERRNO_QUEUE_IN_TSKUSE; goto QUEUE_END; } +<<<<<<< HEAD if (!LOS_ListEmpty(&queueCB->readWriteList[OS_QUEUE_WRITE])) { +======= + if (!LOS_ListEmpty(&queueCB->readWriteList[OS_QUEUE_WRITE])) {//尚有任务要写数据 +>>>>>>> remotes/origin/main ret = LOS_ERRNO_QUEUE_IN_TSKUSE; goto QUEUE_END; } +<<<<<<< HEAD if (!LOS_ListEmpty(&queueCB->memList)) { +======= + if (!LOS_ListEmpty(&queueCB->memList)) {// +>>>>>>> remotes/origin/main ret = LOS_ERRNO_QUEUE_IN_TSKUSE; goto QUEUE_END; } if ((queueCB->readWriteableCnt[OS_QUEUE_WRITE] + queueCB->readWriteableCnt[OS_QUEUE_READ]) != +<<<<<<< HEAD queueCB->queueLen) { +======= + queueCB->queueLen) {//读写队列的内容长度不等于总长度 +>>>>>>> remotes/origin/main ret = LOS_ERRNO_QUEUE_IN_TSKWRITE; goto QUEUE_END; } +<<<<<<< HEAD queue = queueCB->queueHandle; queueCB->queueHandle = NULL; queueCB->queueState = OS_QUEUE_UNUSED; @@ -459,13 +730,29 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_QueueDelete(UINT32 queueID) SCHEDULER_UNLOCK(intSave); OsHookCall(LOS_HOOK_TYPE_QUEUE_DELETE, queueCB); ret = LOS_MemFree(m_aucSysMem1, (VOID *)queue); +======= + queue = queueCB->queueHandle; //队列buf + queueCB->queueHandle = NULL; // + queueCB->queueState = OS_QUEUE_UNUSED;//重置队列状态 + queueCB->queueID = SET_QUEUE_ID(GET_QUEUE_COUNT(queueCB->queueID) + 1, GET_QUEUE_INDEX(queueCB->queueID));//@note_why 这里需要这样做吗? + OsQueueDbgUpdateHook(queueCB->queueID, NULL); + + LOS_ListTailInsert(&FREE_QUEUE_LIST, &queueCB->readWriteList[OS_QUEUE_WRITE]);//回收,将节点挂入可分配链表,等待重新被分配再利用 + SCHEDULER_UNLOCK(intSave); + OsHookCall(LOS_HOOK_TYPE_QUEUE_DELETE, queueCB); + ret = LOS_MemFree(m_aucSysMem1, (VOID *)queue);//释放队列句柄 +>>>>>>> remotes/origin/main return ret; QUEUE_END: SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///外部接口, 获取队列信息,用queueInfo 把 LosQueueCB数据接走,QUEUE_INFO_S对内部数据的封装 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 LOS_QueueInfoGet(UINT32 queueID, QUEUE_INFO_S *queueInfo) { UINT32 intSave; @@ -477,6 +764,7 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_QueueInfoGet(UINT32 queueID, QUEUE_INFO_S *que return LOS_ERRNO_QUEUE_PTR_NULL; } +<<<<<<< HEAD if (GET_QUEUE_INDEX(queueID) >= LOSCFG_BASE_IPC_QUEUE_LIMIT) { return LOS_ERRNO_QUEUE_INVALID; } @@ -485,6 +773,16 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_QueueInfoGet(UINT32 queueID, QUEUE_INFO_S *que SCHEDULER_LOCK(intSave); queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID); +======= + if (GET_QUEUE_INDEX(queueID) >= LOSCFG_BASE_IPC_QUEUE_LIMIT) {//1024 + return LOS_ERRNO_QUEUE_INVALID; + } + + (VOID)memset_s((VOID *)queueInfo, sizeof(QUEUE_INFO_S), 0, sizeof(QUEUE_INFO_S));//接走数据之前先清0 + SCHEDULER_LOCK(intSave); + + queueCB = (LosQueueCB *)GET_QUEUE_HANDLE(queueID);//通过队列ID获取 QCB +>>>>>>> remotes/origin/main if ((queueCB->queueID != queueID) || (queueCB->queueState == OS_QUEUE_UNUSED)) { ret = LOS_ERRNO_QUEUE_NOT_CREATE; goto QUEUE_END; @@ -495,6 +793,7 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_QueueInfoGet(UINT32 queueID, QUEUE_INFO_S *que queueInfo->usQueueSize = queueCB->queueSize; queueInfo->usQueueHead = queueCB->queueHead; queueInfo->usQueueTail = queueCB->queueTail; +<<<<<<< HEAD queueInfo->usReadableCnt = queueCB->readWriteableCnt[OS_QUEUE_READ]; queueInfo->usWritableCnt = queueCB->readWriteableCnt[OS_QUEUE_WRITE]; @@ -508,6 +807,21 @@ LITE_OS_SEC_TEXT_MINOR UINT32 LOS_QueueInfoGet(UINT32 queueID, QUEUE_INFO_S *que LOS_DL_LIST_FOR_EACH_ENTRY(tskCB, &queueCB->memList, LosTaskCB, pendList) { queueInfo->uwWaitMemTask |= 1ULL << tskCB->taskID; +======= + queueInfo->usReadableCnt = queueCB->readWriteableCnt[OS_QUEUE_READ];//可读数 + queueInfo->usWritableCnt = queueCB->readWriteableCnt[OS_QUEUE_WRITE];//可写数 + + LOS_DL_LIST_FOR_EACH_ENTRY(tskCB, &queueCB->readWriteList[OS_QUEUE_READ], LosTaskCB, pendList) {//找出哪些task需要读消息 + queueInfo->uwWaitReadTask |= 1ULL << tskCB->taskID;//记录等待读消息的任务号, uwWaitReadTask 每一位代表一个任务编号 + }//0b..011011011 代表 0,1,3,4,6,7 号任务有数据等待读消息. + + LOS_DL_LIST_FOR_EACH_ENTRY(tskCB, &queueCB->readWriteList[OS_QUEUE_WRITE], LosTaskCB, pendList) {//找出哪些task需要写消息 + queueInfo->uwWaitWriteTask |= 1ULL << tskCB->taskID;//记录等待写消息的任务号, uwWaitWriteTask 每一位代表一个任务编号 + }////0b..011011011 代表 0,1,3,4,6,7 号任务有数据等待写消息. + + LOS_DL_LIST_FOR_EACH_ENTRY(tskCB, &queueCB->memList, LosTaskCB, pendList) {//同上 + queueInfo->uwWaitMemTask |= 1ULL << tskCB->taskID; //MailBox模块使用 +>>>>>>> remotes/origin/main } QUEUE_END: @@ -515,5 +829,9 @@ QUEUE_END: return ret; } +<<<<<<< HEAD #endif /* LOSCFG_BASE_IPC_QUEUE */ +======= +#endif /* (LOSCFG_BASE_IPC_QUEUE == YES) */ +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_rwlock.c b/src/kernel_liteos_a/kernel/base/ipc/los_rwlock.c index 546fe831..21b12e7e 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_rwlock.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_rwlock.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -37,9 +41,16 @@ #include "los_exc.h" #include "los_sched_pri.h" +<<<<<<< HEAD #ifdef LOSCFG_BASE_IPC_RWLOCK #define RWLOCK_COUNT_MASK 0x00FFFFFFU +======= + +#ifdef LOSCFG_BASE_IPC_RWLOCK +#define RWLOCK_COUNT_MASK 0x00FFFFFFU +/// 判断读写锁有效性 +>>>>>>> remotes/origin/main BOOL LOS_RwlockIsValid(const LosRwlock *rwlock) { if ((rwlock != NULL) && ((rwlock->magic & RWLOCK_COUNT_MASK) == OS_RWLOCK_MAGIC)) { @@ -48,7 +59,11 @@ BOOL LOS_RwlockIsValid(const LosRwlock *rwlock) return FALSE; } +<<<<<<< HEAD +======= +/// 创建读写锁,初始化锁信息 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockInit(LosRwlock *rwlock) { UINT32 intSave; @@ -71,7 +86,11 @@ UINT32 LOS_RwlockInit(LosRwlock *rwlock) SCHEDULER_UNLOCK(intSave); return LOS_OK; } +<<<<<<< HEAD +======= +/// 删除指定的读写锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockDestroy(LosRwlock *rwlock) { UINT32 intSave; @@ -95,18 +114,30 @@ UINT32 LOS_RwlockDestroy(LosRwlock *rwlock) SCHEDULER_UNLOCK(intSave); return LOS_OK; } +<<<<<<< HEAD +======= +/// 读写锁检查 +>>>>>>> remotes/origin/main STATIC UINT32 OsRwlockCheck(const LosRwlock *rwlock) { if (rwlock == NULL) { return LOS_EINVAL; } +<<<<<<< HEAD if (OS_INT_ACTIVE) { return LOS_EINTR; } /* DO NOT Call blocking API in system tasks */ +======= + if (OS_INT_ACTIVE) { // 读写锁不能在中断服务程序中使用。请想想为什么 ? + return LOS_EINTR; + } + + /* DO NOT Call blocking API in system tasks | 系统任务不能使用读写锁 */ +>>>>>>> remotes/origin/main LosTaskCB *runTask = (LosTaskCB *)OsCurrTaskGet(); if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) { return LOS_EPERM; @@ -114,19 +145,36 @@ STATIC UINT32 OsRwlockCheck(const LosRwlock *rwlock) return LOS_OK; } +<<<<<<< HEAD STATIC BOOL OsRwlockPriCompare(LosTaskCB *runTask, LOS_DL_LIST *rwList) { if (!LOS_ListEmpty(rwList)) { LosTaskCB *highestTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(rwList)); if (OsSchedParamCompare(runTask, highestTask) < 0) { +======= +/// 指定任务优先级优先级是否低于 写锁任务最高优先级 +STATIC BOOL OsRwlockPriCompare(LosTaskCB *runTask, LOS_DL_LIST *rwList) +{ + if (!LOS_ListEmpty(rwList)) { + LosTaskCB *highestTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(rwList));//首个写锁任务优先级是最高的 + if (OsSchedParamCompare(runTask, highestTask) < 0) {//如果当前任务优先级低于等待写锁任务 +>>>>>>> remotes/origin/main return TRUE; } return FALSE; } return TRUE; } - +<<<<<<< HEAD + +======= +/* 申请读模式下的锁,分三种情况: +1. 若无人持有锁,读任务可获得锁。 +2. 若有人持有锁,读任务可获得锁,读取顺序按照任务优先级。 +3. 若有人(非自己)持有写模式下的锁,则当前任务无法获得锁,直到写模式下的锁释放。 +*/ +>>>>>>> remotes/origin/main STATIC UINT32 OsRwlockRdPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 timeout) { UINT32 ret; @@ -135,12 +183,21 @@ STATIC UINT32 OsRwlockRdPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 tim * When the rwlock mode is read mode or free mode and the priority of the current read task * is higher than the first pended write task. current read task can obtain this rwlock. */ +<<<<<<< HEAD if (rwlock->rwCount >= 0) { if (OsRwlockPriCompare(runTask, &(rwlock->writeList))) { if (rwlock->rwCount == INT8_MAX) { return LOS_EINVAL; } rwlock->rwCount++; +======= + if (rwlock->rwCount >= 0) {//第一和第二种情况 + if (OsRwlockPriCompare(runTask, &(rwlock->writeList))) {//读优先级低于写优先级,意思就是必须先写再读 + if (rwlock->rwCount == INT8_MAX) {//读锁任务达到上限 + return LOS_EINVAL; + } + rwlock->rwCount++;//拿读锁成功 +>>>>>>> remotes/origin/main return LOS_OK; } } @@ -149,32 +206,56 @@ STATIC UINT32 OsRwlockRdPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 tim return LOS_EINVAL; } +<<<<<<< HEAD if (!OsPreemptableInSched()) { return LOS_EDEADLK; } /* The current task is not allowed to obtain the write lock when it obtains the read lock. */ if ((LosTaskCB *)(rwlock->writeOwner) == runTask) { +======= + if (!OsPreemptableInSched()) {//不可抢占时 + return LOS_EDEADLK; + } + + /* The current task is not allowed to obtain the write lock when it obtains the read lock. + | 当前任务在获得读锁时不允许获得写锁 */ + if ((LosTaskCB *)(rwlock->writeOwner) == runTask) { //拥有写锁任务是否为当前任务 +>>>>>>> remotes/origin/main return LOS_EINVAL; } /* * When the rwlock mode is write mode or the priority of the current read task * is lower than the first pended write task, current read task will be pended. +<<<<<<< HEAD */ LOS_DL_LIST *node = OsSchedLockPendFindPos(runTask, &(rwlock->readList)); ret = runTask->ops->wait(runTask, node, timeout); +======= + | 当 rwlock 模式为写模式或当前读任务的优先级低于第一个挂起的写任务时,当前读任务将被挂起。 + 反正就是写锁任务优先 + */ + LOS_DL_LIST *node = OsSchedLockPendFindPos(runTask, &(rwlock->readList));//找到要挂入的位置 + //例如现有链表内任务优先级为 0 3 8 9 23 当前为 10 时, 返回的是 9 这个节点 + ret = runTask->ops->wait(runTask, node, timeout);//从尾部插入读锁链表 由此变成了 0 3 8 9 10 23 +>>>>>>> remotes/origin/main if (ret == LOS_ERRNO_TSK_TIMEOUT) { return LOS_ETIMEDOUT; } return ret; } +<<<<<<< HEAD +======= +/// 申请写模式下的锁 +>>>>>>> remotes/origin/main STATIC UINT32 OsRwlockWrPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 timeout) { UINT32 ret; +<<<<<<< HEAD /* When the rwlock is free mode, current write task can obtain this rwlock. */ if (rwlock->rwCount == 0) { rwlock->rwCount = -1; @@ -183,11 +264,27 @@ STATIC UINT32 OsRwlockWrPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 tim } /* Current write task can use one rwlock once again if the rwlock owner is it. */ +======= + /* When the rwlock is free mode, current write task can obtain this rwlock. + | 若该锁当前没有任务持有,或者持有该读模式下的锁的任务和申请该锁的任务为同一个任务,则申请成功,可立即获得写模式下的锁。*/ + if (rwlock->rwCount == 0) { + rwlock->rwCount = -1; + rwlock->writeOwner = (VOID *)runTask;//直接给当前进程锁 + return LOS_OK; + } + + /* Current write task can use one rwlock once again if the rwlock owner is it. + | 如果 rwlock 拥有者是当前写入任务,则它可以再次使用该锁。*/ +>>>>>>> remotes/origin/main if ((rwlock->rwCount < 0) && ((LosTaskCB *)(rwlock->writeOwner) == runTask)) { if (rwlock->rwCount == INT8_MIN) { return LOS_EINVAL; } +<<<<<<< HEAD rwlock->rwCount--; +======= + rwlock->rwCount--;//注意再次拥有算是两把写锁了. +>>>>>>> remotes/origin/main return LOS_OK; } @@ -201,9 +298,15 @@ STATIC UINT32 OsRwlockWrPendOp(LosTaskCB *runTask, LosRwlock *rwlock, UINT32 tim /* * When the rwlock is read mode or other write task obtains this rwlock, current +<<<<<<< HEAD * write task will be pended. */ LOS_DL_LIST *node = OsSchedLockPendFindPos(runTask, &(rwlock->writeList)); +======= + * write task will be pended. | 当 rwlock 为读模式或其他写任务获得该 rwlock 时,当前的写任务将被挂起。直到读模式下的锁释放 + */ + LOS_DL_LIST *node = OsSchedLockPendFindPos(runTask, &(rwlock->writeList));//找到要挂入的位置 +>>>>>>> remotes/origin/main ret = runTask->ops->wait(runTask, node, timeout); if (ret == LOS_ERRNO_TSK_TIMEOUT) { ret = LOS_ETIMEDOUT; @@ -265,20 +368,36 @@ UINT32 OsRwlockTryWrUnsafe(LosRwlock *rwlock, UINT32 timeout) return LOS_EBADF; } +<<<<<<< HEAD /* When the rwlock is read mode, current write task will be pended. */ +======= + /* When the rwlock is read mode, current write task will be pended. + | 当 rwlock 为读模式时,当前的写任务将被挂起。*/ +>>>>>>> remotes/origin/main if (rwlock->rwCount > 0) { return LOS_EBUSY; } +<<<<<<< HEAD /* When other write task obtains this rwlock, current write task will be pended. */ +======= + /* When other write task obtains this rwlock, current write task will be pended. + | 当其他写任务获得这个rwlock时,当前的写任务将被挂起。*/ +>>>>>>> remotes/origin/main LosTaskCB *runTask = OsCurrTaskGet(); if ((rwlock->rwCount < 0) && ((LosTaskCB *)(rwlock->writeOwner) != runTask)) { return LOS_EBUSY; } +<<<<<<< HEAD return OsRwlockWrPendOp(runTask, rwlock, timeout); } +======= + return OsRwlockWrPendOp(runTask, rwlock, timeout);// +} +/// 申请指定的读模式下的锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockRdLock(LosRwlock *rwlock, UINT32 timeout) { UINT32 intSave; @@ -293,7 +412,11 @@ UINT32 LOS_RwlockRdLock(LosRwlock *rwlock, UINT32 timeout) SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +/// 尝试申请指定的读模式下的锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockTryRdLock(LosRwlock *rwlock) { UINT32 intSave; @@ -304,11 +427,19 @@ UINT32 LOS_RwlockTryRdLock(LosRwlock *rwlock) } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsRwlockTryRdUnsafe(rwlock, 0); SCHEDULER_UNLOCK(intSave); return ret; } +======= + ret = OsRwlockTryRdUnsafe(rwlock, 0);//所谓尝试就是没锁爷就返回,不等待,不纠结.当前任务也不会被挂起 + SCHEDULER_UNLOCK(intSave); + return ret; +} +/// 申请指定的写模式下的锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockWrLock(LosRwlock *rwlock, UINT32 timeout) { UINT32 intSave; @@ -323,7 +454,11 @@ UINT32 LOS_RwlockWrLock(LosRwlock *rwlock, UINT32 timeout) SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +/// 尝试申请指定的写模式下的锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockTryWrLock(LosRwlock *rwlock) { UINT32 intSave; @@ -334,15 +469,24 @@ UINT32 LOS_RwlockTryWrLock(LosRwlock *rwlock) } SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsRwlockTryWrUnsafe(rwlock, 0); SCHEDULER_UNLOCK(intSave); return ret; } +======= + ret = OsRwlockTryWrUnsafe(rwlock, 0);//所谓尝试就是没锁爷就返回,不等待,不纠结.当前任务也不会被挂起 + SCHEDULER_UNLOCK(intSave); + return ret; +} +/// 获取读写锁模式 +>>>>>>> remotes/origin/main STATIC UINT32 OsRwlockGetMode(LOS_DL_LIST *readList, LOS_DL_LIST *writeList) { BOOL isReadEmpty = LOS_ListEmpty(readList); BOOL isWriteEmpty = LOS_ListEmpty(writeList); +<<<<<<< HEAD if (isReadEmpty && isWriteEmpty) { return RWLOCK_NONE_MODE; } @@ -350,16 +494,33 @@ STATIC UINT32 OsRwlockGetMode(LOS_DL_LIST *readList, LOS_DL_LIST *writeList) return RWLOCK_READ_MODE; } if (isReadEmpty && !isWriteEmpty) { +======= + if (isReadEmpty && isWriteEmpty) { //读写链表都没有内容 + return RWLOCK_NONE_MODE; //自由模式 + } + if (!isReadEmpty && isWriteEmpty) { //读链表有数据,写链表没有数据 + return RWLOCK_READ_MODE; + } + if (isReadEmpty && !isWriteEmpty) { //写链表有数据,读链表没有数据 +>>>>>>> remotes/origin/main return RWLOCK_WRITE_MODE; } LosTaskCB *pendedReadTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(readList)); LosTaskCB *pendedWriteTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(writeList)); if (OsSchedParamCompare(pendedWriteTask, pendedReadTask) <= 0) { +<<<<<<< HEAD return RWLOCK_WRITEFIRST_MODE; } return RWLOCK_READFIRST_MODE; } +======= + return RWLOCK_WRITEFIRST_MODE; //写的优先级高时,为写优先模式 + } + return RWLOCK_READFIRST_MODE; //读的优先级高时,为读优先模式 +} +/// 释放锁 +>>>>>>> remotes/origin/main STATIC UINT32 OsRwlockPostOp(LosRwlock *rwlock, BOOL *needSched) { UINT32 rwlockMode; @@ -367,6 +528,7 @@ STATIC UINT32 OsRwlockPostOp(LosRwlock *rwlock, BOOL *needSched) rwlock->rwCount = 0; rwlock->writeOwner = NULL; +<<<<<<< HEAD rwlockMode = OsRwlockGetMode(&(rwlock->readList), &(rwlock->writeList)); if (rwlockMode == RWLOCK_NONE_MODE) { return LOS_OK; @@ -376,6 +538,17 @@ STATIC UINT32 OsRwlockPostOp(LosRwlock *rwlock, BOOL *needSched) resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->writeList))); rwlock->rwCount = -1; rwlock->writeOwner = (VOID *)resumedTask; +======= + rwlockMode = OsRwlockGetMode(&(rwlock->readList), &(rwlock->writeList));//先获取模式 + if (rwlockMode == RWLOCK_NONE_MODE) {//自由模式则正常返回 + return LOS_OK; + } + /* In this case, rwlock will wake the first pended write task. | 在这种情况下,rwlock 将唤醒第一个挂起的写任务。 */ + if ((rwlockMode == RWLOCK_WRITE_MODE) || (rwlockMode == RWLOCK_WRITEFIRST_MODE)) {//如果当前是写模式 (有任务在等写锁涅) + resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->writeList)));//获取任务实体 + rwlock->rwCount = -1;//直接干成-1,注意这里并不是 -- + rwlock->writeOwner = (VOID *)resumedTask;//有锁了则唤醒等锁的任务(写模式) +>>>>>>> remotes/origin/main resumedTask->ops->wake(resumedTask); if (needSched != NULL) { *needSched = TRUE; @@ -383,29 +556,51 @@ STATIC UINT32 OsRwlockPostOp(LosRwlock *rwlock, BOOL *needSched) return LOS_OK; } +<<<<<<< HEAD rwlock->rwCount = 1; resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->readList))); resumedTask->ops->wake(resumedTask); while (!LOS_ListEmpty(&(rwlock->readList))) { +======= + rwlock->rwCount = 1; //直接干成1,因为是释放操作 + resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->readList))); + resumedTask->ops->wake(resumedTask); + while (!LOS_ListEmpty(&(rwlock->readList))) {//遍历读链表,目的是要唤醒其他读模式的任务(优先级得要高于pendedWriteTaskPri才行) +>>>>>>> remotes/origin/main resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->readList))); if (rwlockMode == RWLOCK_READFIRST_MODE) { LosTaskCB *pendedWriteTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(rwlock->writeList))); if (OsSchedParamCompare(resumedTask, pendedWriteTask) >= 0) { +<<<<<<< HEAD break; } +======= + break;//跳出循环 + } +>>>>>>> remotes/origin/main } if (rwlock->rwCount == INT8_MAX) { return EINVAL; } +<<<<<<< HEAD rwlock->rwCount++; resumedTask->ops->wake(resumedTask); +======= + rwlock->rwCount++;//读锁任务数量增加 + resumedTask->ops->wake(resumedTask);//不断唤醒读锁任务,由此实现了允许多个读操作并发,因为在多核情况下resumedTask很大可能 + //与当前任务并不在同一个核上运行, 此处非常有意思,点赞! @note_good +>>>>>>> remotes/origin/main } if (needSched != NULL) { *needSched = TRUE; } return LOS_OK; } +<<<<<<< HEAD +======= +/// 释放锁,唤醒任务 +>>>>>>> remotes/origin/main UINT32 OsRwlockUnlockUnsafe(LosRwlock *rwlock, BOOL *needSched) { if ((rwlock->magic & RWLOCK_COUNT_MASK) != OS_RWLOCK_MAGIC) { @@ -417,27 +612,45 @@ UINT32 OsRwlockUnlockUnsafe(LosRwlock *rwlock, BOOL *needSched) } LosTaskCB *runTask = OsCurrTaskGet(); +<<<<<<< HEAD if ((rwlock->rwCount < 0) && ((LosTaskCB *)(rwlock->writeOwner) != runTask)) { +======= + if ((rwlock->rwCount < 0) && ((LosTaskCB *)(rwlock->writeOwner) != runTask)) {//写模式时,当前任务未持有锁 +>>>>>>> remotes/origin/main return LOS_EPERM; } /* * When the rwCount of the rwlock more than 1 or less than -1, the rwlock mode will * not changed after current unlock operation, so pended tasks can not be waken. +<<<<<<< HEAD */ if (rwlock->rwCount > 1) { +======= + | 当 rwlock 的 rwCount 大于 1 或小于 -1 时,当前解锁操作后 rwlock 模式不会改变,因此挂起的任务不能被唤醒。 + */ + if (rwlock->rwCount > 1) {//读模式 +>>>>>>> remotes/origin/main rwlock->rwCount--; return LOS_OK; } +<<<<<<< HEAD if (rwlock->rwCount < -1) { +======= + if (rwlock->rwCount < -1) {//写模式 +>>>>>>> remotes/origin/main rwlock->rwCount++; return LOS_OK; } return OsRwlockPostOp(rwlock, needSched); } +<<<<<<< HEAD +======= +/// 释放指定读写锁 +>>>>>>> remotes/origin/main UINT32 LOS_RwlockUnLock(LosRwlock *rwlock) { UINT32 intSave; @@ -451,9 +664,15 @@ UINT32 LOS_RwlockUnLock(LosRwlock *rwlock) SCHEDULER_LOCK(intSave); ret = OsRwlockUnlockUnsafe(rwlock, &needSched); SCHEDULER_UNLOCK(intSave); +<<<<<<< HEAD LOS_MpSchedule(OS_MP_CPU_ALL); if (needSched == TRUE) { LOS_Schedule(); +======= + LOS_MpSchedule(OS_MP_CPU_ALL);//设置调度CPU的方式,所有CPU参与调度 + if (needSched == TRUE) {//是否需要调度 + LOS_Schedule();//产生调度,切换任务执行 +>>>>>>> remotes/origin/main } return ret; } diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_sem.c b/src/kernel_liteos_a/kernel/base/ipc/los_sem.c index 31d484a1..bbc2f25c 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_sem.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_sem.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -40,17 +44,29 @@ #include "los_percpu_pri.h" #include "los_hook.h" +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main #ifdef LOSCFG_BASE_IPC_SEM #if (LOSCFG_BASE_IPC_SEM_LIMIT <= 0) #error "sem maxnum cannot be zero" #endif /* LOSCFG_BASE_IPC_SEM_LIMIT <= 0 */ +<<<<<<< HEAD LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_unusedSemList; LITE_OS_SEC_BSS LosSemCB *g_allSem = NULL; /* * Description : Initialize the semaphore doubly linked list +======= +LITE_OS_SEC_DATA_INIT STATIC LOS_DL_LIST g_unusedSemList; ///< 可用的信号量列表,干嘛不用freeList? 可以看出这里是另一个人写的代码 +LITE_OS_SEC_BSS LosSemCB *g_allSem = NULL; ///< 信号池,一次分配 LOSCFG_BASE_IPC_SEM_LIMIT 个信号量 + +/* + * Description : Initialize the semaphore doubly linked list | 信号量初始化 +>>>>>>> remotes/origin/main * Return : LOS_OK on success, or error code on failure */ LITE_OS_SEC_TEXT_INIT UINT32 OsSemInit(VOID) @@ -58,18 +74,31 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemInit(VOID) LosSemCB *semNode = NULL; UINT32 index; +<<<<<<< HEAD LOS_ListInit(&g_unusedSemList); /* system resident memory, don't free */ g_allSem = (LosSemCB *)LOS_MemAlloc(m_aucSysMem0, (LOSCFG_BASE_IPC_SEM_LIMIT * sizeof(LosSemCB))); +======= + LOS_ListInit(&g_unusedSemList);//初始化链表,链表上挂未使用的信号量,用于分配信号量,鸿蒙信号量的个数是有限的,默认1024个 + /* system resident memory, don't free */ + g_allSem = (LosSemCB *)LOS_MemAlloc(m_aucSysMem0, (LOSCFG_BASE_IPC_SEM_LIMIT * sizeof(LosSemCB)));//分配信号池 +>>>>>>> remotes/origin/main if (g_allSem == NULL) { return LOS_ERRNO_SEM_NO_MEMORY; } for (index = 0; index < LOSCFG_BASE_IPC_SEM_LIMIT; index++) { +<<<<<<< HEAD semNode = ((LosSemCB *)g_allSem) + index; semNode->semID = SET_SEM_ID(0, index); semNode->semStat = OS_SEM_UNUSED; LOS_ListTailInsert(&g_unusedSemList, &semNode->semList); +======= + semNode = ((LosSemCB *)g_allSem) + index;//拿信号控制块, 可以直接g_allSem[index]来嘛 + semNode->semID = SET_SEM_ID(0, index);//保存ID + semNode->semStat = OS_SEM_UNUSED;//标记未使用 + LOS_ListTailInsert(&g_unusedSemList, &semNode->semList);//通过semList把 信号块挂到空闲链表上 +>>>>>>> remotes/origin/main } if (OsSemDbgInitHook() != LOS_OK) { @@ -78,6 +107,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemInit(VOID) return LOS_OK; } +<<<<<<< HEAD /* * Description : Create a semaphore, * Input : count --- semaphore count, @@ -85,6 +115,9 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemInit(VOID) * semHandle --- Index of semaphore, * Return : LOS_OK on success ,or error code on failure */ +======= + +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 OsSemCreate(UINT16 count, UINT16 maxCount, UINT32 *semHandle) { UINT32 intSave; @@ -97,6 +130,7 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemCreate(UINT16 count, UINT16 maxCount, UINT32 * return LOS_ERRNO_SEM_PTR_NULL; } +<<<<<<< HEAD if (count > maxCount) { OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_OVERFLOW); } @@ -104,11 +138,21 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemCreate(UINT16 count, UINT16 maxCount, UINT32 * SCHEDULER_LOCK(intSave); if (LOS_ListEmpty(&g_unusedSemList)) { +======= + if (count > maxCount) {//信号量不能大于最大值,两参数都是外面给的 + OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_OVERFLOW); + } + + SCHEDULER_LOCK(intSave);//进入临界区,拿自旋锁 + + if (LOS_ListEmpty(&g_unusedSemList)) {//没有可分配的空闲信号提供 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); OsSemInfoGetFullDataHook(); OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_ALL_BUSY); } +<<<<<<< HEAD unusedSem = LOS_DL_LIST_FIRST(&g_unusedSemList); LOS_ListDelete(unusedSem); SCHEDULER_UNLOCK(intSave); @@ -120,22 +164,48 @@ LITE_OS_SEC_TEXT_INIT UINT32 OsSemCreate(UINT16 count, UINT16 maxCount, UINT32 * *semHandle = semCreated->semID; OsHookCall(LOS_HOOK_TYPE_SEM_CREATE, semCreated); OsSemDbgUpdateHook(semCreated->semID, OsCurrTaskGet()->taskEntry, count); +======= + unusedSem = LOS_DL_LIST_FIRST(&g_unusedSemList);//从未使用信号量池中取首个 + LOS_ListDelete(unusedSem);//从空闲链表上摘除 + SCHEDULER_UNLOCK(intSave); + semCreated = GET_SEM_LIST(unusedSem);//通过semList挂到链表上的,这里也要通过它把LosSemCB头查到. 进程,线程等结构体也都是这么干的. + semCreated->semCount = count;//设置数量 + semCreated->semStat = OS_SEM_USED;//设置可用状态 + semCreated->maxSemCount = maxCount;//设置最大信号数量 + LOS_ListInit(&semCreated->semList);//初始化链表,后续阻塞任务通过task->pendList挂到semList链表上,就知道哪些任务在等它了. + *semHandle = semCreated->semID;//参数带走 semID + OsHookCall(LOS_HOOK_TYPE_SEM_CREATE, semCreated); + OsSemDbgUpdateHook(semCreated->semID, OsCurrTaskGet()->taskEntry, count); + +>>>>>>> remotes/origin/main return LOS_OK; ERR_HANDLER: OS_RETURN_ERROR_P2(errLine, errNo); } +<<<<<<< HEAD +======= +///对外接口 创建信号量 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_SemCreate(UINT16 count, UINT32 *semHandle) { return OsSemCreate(count, OS_SEM_COUNT_MAX, semHandle); } +<<<<<<< HEAD +======= +///对外接口 创建二值信号量,其计数值最大为1,可以当互斥锁用 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_BinarySemCreate(UINT16 count, UINT32 *semHandle) { return OsSemCreate(count, OS_SEM_BINARY_COUNT_MAX, semHandle); } +<<<<<<< HEAD +======= +///对外接口 删除指定的信号量,参数就是 semID +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_SemDelete(UINT32 semHandle) { UINT32 intSave; @@ -147,15 +217,24 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SemDelete(UINT32 semHandle) OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_INVALID); } +<<<<<<< HEAD semDeleted = GET_SEM(semHandle); SCHEDULER_LOCK(intSave); if ((semDeleted->semStat == OS_SEM_UNUSED) || (semDeleted->semID != semHandle)) { +======= + semDeleted = GET_SEM(semHandle);//通过ID拿到信号量实体 + + SCHEDULER_LOCK(intSave); + + if ((semDeleted->semStat == OS_SEM_UNUSED) || (semDeleted->semID != semHandle)) {//参数判断 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_INVALID); } +<<<<<<< HEAD if (!LOS_ListEmpty(&semDeleted->semList)) { SCHEDULER_UNLOCK(intSave); OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_PENDED); @@ -164,6 +243,16 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SemDelete(UINT32 semHandle) LOS_ListTailInsert(&g_unusedSemList, &semDeleted->semList); semDeleted->semStat = OS_SEM_UNUSED; semDeleted->semID = SET_SEM_ID(GET_SEM_COUNT(semDeleted->semID) + 1, GET_SEM_INDEX(semDeleted->semID)); +======= + if (!LOS_ListEmpty(&semDeleted->semList)) {//当前还有任务挂在这个信号上面,当然不能删除 + SCHEDULER_UNLOCK(intSave); + OS_GOTO_ERR_HANDLER(LOS_ERRNO_SEM_PENDED);//这个宏很有意思,里面goto到ERR_HANDLER + } + + LOS_ListTailInsert(&g_unusedSemList, &semDeleted->semList);//通过semList从尾部插入空闲链表 + semDeleted->semStat = OS_SEM_UNUSED;//状态变成了未使用 + semDeleted->semID = SET_SEM_ID(GET_SEM_COUNT(semDeleted->semID) + 1, GET_SEM_INDEX(semDeleted->semID));//设置ID +>>>>>>> remotes/origin/main OsHookCall(LOS_HOOK_TYPE_SEM_DELETE, semDeleted); OsSemDbgUpdateHook(semDeleted->semID, NULL, 0); @@ -174,11 +263,19 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_SemDelete(UINT32 semHandle) ERR_HANDLER: OS_RETURN_ERROR_P2(errLine, errNo); } +<<<<<<< HEAD LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) { UINT32 intSave; LosSemCB *semPended = GET_SEM(semHandle); +======= +///对外接口 申请指定的信号量,并设置超时时间 +LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) +{ + UINT32 intSave; + LosSemCB *semPended = GET_SEM(semHandle);//通过ID拿到信号体 +>>>>>>> remotes/origin/main UINT32 retErr = LOS_OK; LosTaskCB *runTask = NULL; @@ -192,7 +289,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) return LOS_ERRNO_SEM_PEND_INTERR; } +<<<<<<< HEAD runTask = OsCurrTaskGet(); +======= + runTask = OsCurrTaskGet();//获取当前任务 +>>>>>>> remotes/origin/main if (runTask->taskStatus & OS_TASK_FLAG_SYSTEM_TASK) { OsBackTrace(); return LOS_ERRNO_SEM_PEND_IN_SYSTEM_TASK; @@ -204,6 +305,7 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) retErr = LOS_ERRNO_SEM_INVALID; goto OUT; } +<<<<<<< HEAD /* Update the operate time, no matter the actual Pend success or not */ OsSemDbgTimeUpdateHook(semHandle); @@ -211,12 +313,26 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) semPended->semCount--; OsHookCall(LOS_HOOK_TYPE_SEM_PEND, semPended, runTask, timeout); goto OUT; +======= + + /* Update the operate time, no matter the actual Pend success or not */ + OsSemDbgTimeUpdateHook(semHandle); + + if (semPended->semCount > 0) {//还有资源可用,返回肯定得成功,semCount=0时代表没资源了,task会必须去睡眠了 + semPended->semCount--;//资源少了一个 + OsHookCall(LOS_HOOK_TYPE_SEM_PEND, semPended, runTask, timeout); + goto OUT;//注意这里 retErr = LOS_OK ,所以返回是OK的 +>>>>>>> remotes/origin/main } else if (!timeout) { retErr = LOS_ERRNO_SEM_UNAVAILABLE; goto OUT; } +<<<<<<< HEAD if (!OsPreemptableInSched()) { +======= + if (!OsPreemptableInSched()) {//不能申请调度 (不能调度的原因是因为没有持有调度任务自旋锁) +>>>>>>> remotes/origin/main PRINT_ERR("!!!LOS_ERRNO_SEM_PEND_IN_LOCK!!!\n"); OsBackTrace(); retErr = LOS_ERRNO_SEM_PEND_IN_LOCK; @@ -226,7 +342,11 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPend(UINT32 semHandle, UINT32 timeout) OsHookCall(LOS_HOOK_TYPE_SEM_PEND, semPended, runTask, timeout); OsTaskWaitSetPendMask(OS_TASK_WAIT_SEM, semPended->semID, timeout); retErr = runTask->ops->wait(runTask, &semPended->semList, timeout); +<<<<<<< HEAD if (retErr == LOS_ERRNO_TSK_TIMEOUT) { +======= + if (retErr == LOS_ERRNO_TSK_TIMEOUT) {//注意:这里是涉及到task切换的,把自己挂起,唤醒其他task +>>>>>>> remotes/origin/main retErr = LOS_ERRNO_SEM_TIMEOUT; } @@ -234,7 +354,11 @@ OUT: SCHEDULER_UNLOCK(intSave); return retErr; } +<<<<<<< HEAD +======= +///以不安全的方式释放指定的信号量,所谓不安全指的是不用自旋锁 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 OsSemPostUnsafe(UINT32 semHandle, BOOL *needSched) { LosTaskCB *resumedTask = NULL; @@ -246,6 +370,7 @@ LITE_OS_SEC_TEXT UINT32 OsSemPostUnsafe(UINT32 semHandle, BOOL *needSched) /* Update the operate time, no matter the actual Post success or not */ OsSemDbgTimeUpdateHook(semHandle); +<<<<<<< HEAD if (semPosted->semCount == OS_SEM_COUNT_MAX) { return LOS_ERRNO_SEM_OVERFLOW; } @@ -258,11 +383,29 @@ LITE_OS_SEC_TEXT UINT32 OsSemPostUnsafe(UINT32 semHandle, BOOL *needSched) } } else { semPosted->semCount++; +======= + if (semPosted->semCount == OS_SEM_COUNT_MAX) {//当前信号资源不能大于最大资源量 + return LOS_ERRNO_SEM_OVERFLOW; + } + if (!LOS_ListEmpty(&semPosted->semList)) {//当前有任务挂在semList上,要去唤醒任务 + resumedTask = OS_TCB_FROM_PENDLIST(LOS_DL_LIST_FIRST(&(semPosted->semList)));//semList上面挂的都是task->pendlist节点,取第一个task下来唤醒 + OsTaskWakeClearPendMask(resumedTask); + resumedTask->ops->wake(resumedTask); + if (needSched != NULL) {//参数不为空,就返回需要调度的标签 + *needSched = TRUE;//TRUE代表需要调度 + } + } else {//当前没有任务挂在semList上, + semPosted->semCount++;//信号资源多一个 +>>>>>>> remotes/origin/main } OsHookCall(LOS_HOOK_TYPE_SEM_POST, semPosted, resumedTask); return LOS_OK; } +<<<<<<< HEAD +======= +///对外接口 释放指定的信号量 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINT32 LOS_SemPost(UINT32 semHandle) { UINT32 intSave; @@ -272,6 +415,7 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPost(UINT32 semHandle) if (GET_SEM_INDEX(semHandle) >= LOSCFG_BASE_IPC_SEM_LIMIT) { return LOS_ERRNO_SEM_INVALID; } +<<<<<<< HEAD SCHEDULER_LOCK(intSave); ret = OsSemPostUnsafe(semHandle, &needSched); @@ -279,9 +423,21 @@ LITE_OS_SEC_TEXT UINT32 LOS_SemPost(UINT32 semHandle) if (needSched) { LOS_MpSchedule(OS_MP_CPU_ALL); LOS_Schedule(); +======= + SCHEDULER_LOCK(intSave); + ret = OsSemPostUnsafe(semHandle, &needSched); + SCHEDULER_UNLOCK(intSave); + if (needSched) {//需要调度的情况 + LOS_MpSchedule(OS_MP_CPU_ALL);//向所有CPU发送调度指令 + LOS_Schedule();////发起调度 +>>>>>>> remotes/origin/main } return ret; } +<<<<<<< HEAD #endif /* LOSCFG_BASE_IPC_SEM */ +======= +#endif /* (LOSCFG_BASE_IPC_SEM == YES) */ +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_sem_debug.c b/src/kernel_liteos_a/kernel/base/ipc/los_sem_debug.c index 1ed42515..e19a272d 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_sem_debug.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_sem_debug.c @@ -78,11 +78,19 @@ STATIC VOID OsSemPendedTaskNamePrint(LosSemCB *semNode) #ifdef LOSCFG_DEBUG_SEMAPHORE typedef struct { +<<<<<<< HEAD UINT16 origSemCount; /* Number of original available semaphores */ UINT64 lastAccessTime; /* The last operation time */ TSK_ENTRY_FUNC creator; /* The task entry who created this sem */ } SemDebugCB; STATIC SemDebugCB *g_semDebugArray = NULL; +======= + UINT16 origSemCount; /* Number of orignal available semaphores *///原始可用信号量数 + UINT64 lastAccessTime; /* The last operation time */ //最后操作时间 + TSK_ENTRY_FUNC creator; /* The task entry who created this sem */ //由哪个task的入口函数创建了这个任务 +} SemDebugCB; +STATIC SemDebugCB *g_semDebugArray = NULL;//默认1024个SemDebugCB debug信号量池 +>>>>>>> remotes/origin/main STATIC BOOL SemCompareValue(const IpcSortParam *sortParam, UINT32 left, UINT32 right) { @@ -102,6 +110,7 @@ UINT32 OsSemDbgInit(VOID) (VOID)memset_s(g_semDebugArray, size, 0, size); return LOS_OK; } +<<<<<<< HEAD VOID OsSemDbgTimeUpdate(UINT32 semID) { @@ -119,6 +128,25 @@ VOID OsSemDbgUpdate(UINT32 semID, TSK_ENTRY_FUNC creator, UINT16 count) return; } +======= +///更新最后访问时间 +VOID OsSemDbgTimeUpdate(UINT32 semID) +{ + SemDebugCB *semDebug = &g_semDebugArray[GET_SEM_INDEX(semID)]; + semDebug->lastAccessTime = LOS_TickCountGet();//获取tick总数 + return; +} +///更新信号量 +VOID OsSemDbgUpdate(UINT32 semID, TSK_ENTRY_FUNC creator, UINT16 count) +{ + SemDebugCB *semDebug = &g_semDebugArray[GET_SEM_INDEX(semID)]; + semDebug->creator = creator; //改为由参数入口函数创建了这个任务 + semDebug->lastAccessTime = LOS_TickCountGet();//获取tick总数 + semDebug->origSemCount = count;//原始信号量改变 + return; +} +///按信号量访问时间排序 +>>>>>>> remotes/origin/main STATIC VOID OsSemSort(UINT32 *semIndexArray, UINT32 usedCount) { UINT32 i, intSave; @@ -296,6 +324,10 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdSemInfoGet(UINT32 argc, const CHAR **arg return ret; } +<<<<<<< HEAD SHELLCMD_ENTRY(sem_shellcmd, CMD_TYPE_EX, "sem", 1, (CmdCallBackFunc)OsShellCmdSemInfoGet); +======= +SHELLCMD_ENTRY(sem_shellcmd, CMD_TYPE_EX, "sem", 1, (CmdCallBackFunc)OsShellCmdSemInfoGet);//采用shell命令静态注册方式 +>>>>>>> remotes/origin/main #endif diff --git a/src/kernel_liteos_a/kernel/base/ipc/los_signal.c b/src/kernel_liteos_a/kernel/base/ipc/los_signal.c index bb7bd2f7..73c13860 100644 --- a/src/kernel_liteos_a/kernel/base/ipc/los_signal.c +++ b/src/kernel_liteos_a/kernel/base/ipc/los_signal.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -53,6 +57,7 @@ int raise(int sig) #define GETUNMASKSET(procmask, pendFlag) ((~(procmask)) & (sigset_t)(pendFlag)) #define UINT64_BIT_SIZE 64 +<<<<<<< HEAD int OsSigIsMember(const sigset_t *set, int signo) { int ret = LOS_NOK; @@ -62,6 +67,19 @@ int OsSigIsMember(const sigset_t *set, int signo) if (GOOD_SIGNO(signo)) { /* Check if the signal is in the set */ ret = ((*set & SIGNO2SET((unsigned int)signo)) != 0); +======= + +int OsSigIsMember(const sigset_t *set, int signo) +{ + int ret = LOS_NOK; + /* In musl, sig No bits 00000100 present sig No 3, but 1<< 3 = 00001000, so signo needs minus 1 */ + //在musl中,sig No bits 00000100表示sig No 3,但是在SIGNO2SET中 1<<3 = 00001000,因此signo需要减1 + signo -= 1; + /* Verify the signal */ + if (GOOD_SIGNO(signo)) {//有效信号判断 + /* Check if the signal is in the set */ + ret = ((*set & SIGNO2SET((unsigned int)signo)) != 0);//检查信号是否还在集合中 +>>>>>>> remotes/origin/main } return ret; @@ -120,7 +138,10 @@ VOID OsClearSigInfoTmpList(sig_cb *sigcb) (VOID)LOS_MemFree(m_aucSysMem0, tmpInfoNode); } } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main STATIC INLINE VOID OsSigWaitTaskWake(LosTaskCB *taskCB, INT32 signo) { sig_cb *sigcb = &taskCB->sig; @@ -132,14 +153,22 @@ STATIC INLINE VOID OsSigWaitTaskWake(LosTaskCB *taskCB, INT32 signo) OsSigEmptySet(&sigcb->sigwaitmask); } } +<<<<<<< HEAD +======= +///< 唤醒被挂起的处于等待指定信号的任务 +>>>>>>> remotes/origin/main STATIC UINT32 OsPendingTaskWake(LosTaskCB *taskCB, INT32 signo) { if (!OsTaskIsPending(taskCB) || !OsProcessIsUserMode(OS_PCB_FROM_TCB(taskCB))) { return 0; } +<<<<<<< HEAD if ((signo != SIGKILL) && (taskCB->waitFlag != OS_TASK_WAIT_SIGNAL)) { +======= + if ((signo != SIGKILL) && (taskCB->waitFlag != OS_TASK_WAIT_SIGNAL)) { // @note_thinking 这个判断会不会有问题 ? +>>>>>>> remotes/origin/main return 0; } @@ -153,6 +182,7 @@ STATIC UINT32 OsPendingTaskWake(LosTaskCB *taskCB, INT32 signo) OsTaskWakeClearPendMask(taskCB); taskCB->ops->wake(taskCB); break; +<<<<<<< HEAD case OS_TASK_WAIT_SIGNAL: OsSigWaitTaskWake(taskCB, signo); break; @@ -163,6 +193,18 @@ STATIC UINT32 OsPendingTaskWake(LosTaskCB *taskCB, INT32 signo) case OS_TASK_WAIT_FUTEX: OsFutexNodeDeleteFromFutexHash(&taskCB->futex, TRUE, NULL, NULL); OsTaskWakeClearPendMask(taskCB); +======= + case OS_TASK_WAIT_SIGNAL://等待普通信号 + OsSigWaitTaskWake(taskCB, signo); + break; + case OS_TASK_WAIT_LITEIPC://等待liteipc信号 + OsTaskWakeClearPendMask(taskCB);//重置任务的等待信息 + taskCB->ops->wake(taskCB); + break; + case OS_TASK_WAIT_FUTEX://等待快锁信号 + OsFutexNodeDeleteFromFutexHash(&taskCB->futex, TRUE, NULL, NULL);//从哈希桶中删除快锁 + OsTaskWakeClearPendMask(taskCB);//重置任务的等待信息 +>>>>>>> remotes/origin/main taskCB->ops->wake(taskCB); break; default: @@ -171,7 +213,11 @@ STATIC UINT32 OsPendingTaskWake(LosTaskCB *taskCB, INT32 signo) return 0; } +<<<<<<< HEAD +======= +///给任务(线程)发送一个信号 +>>>>>>> remotes/origin/main int OsTcbDispatch(LosTaskCB *stcb, siginfo_t *info) { bool masked = FALSE; @@ -179,6 +225,7 @@ int OsTcbDispatch(LosTaskCB *stcb, siginfo_t *info) OS_RETURN_IF_NULL(sigcb); /* If signo is 0, not send signal, just check process or pthread exist */ +<<<<<<< HEAD if (info->si_signo == 0) { return 0; } @@ -192,6 +239,21 @@ int OsTcbDispatch(LosTaskCB *stcb, siginfo_t *info) } else { /* unmasked signal actions */ OsSigAddSet(&sigcb->sigFlag, info->si_signo); +======= + if (info->si_signo == 0) {//如果信号为0,则不发送信号,只是作为检查进程和线程是否还存在. + return 0; + } + masked = (bool)OsSigIsMember(&sigcb->sigprocmask, info->si_signo);//@note_thinking 这里还有 masked= -1的情况要处理!!! + if (masked) {//如果信号被屏蔽了,要看等待信号集,sigwaitmask + /* If signal is in wait list and mask list, need unblock it */ //如果信号在等待列表和掩码列表中,需要解除阻止 + if (LOS_ListEmpty(&sigcb->waitList) || + (!LOS_ListEmpty(&sigcb->waitList) && !OsSigIsMember(&sigcb->sigwaitmask, info->si_signo))) { + OsSigAddSet(&sigcb->sigPendFlag, info->si_signo);//将信号加入挂起/待办集 + } + } else {//信号没有被屏蔽的处理 + /* unmasked signal actions */ + OsSigAddSet(&sigcb->sigFlag, info->si_signo);//不屏蔽的信号集 +>>>>>>> remotes/origin/main } if (OsAddSigInfoToTmpList(sigcb, info) == LOS_NOK) { @@ -206,6 +268,7 @@ void OsSigMaskSwitch(LosTaskCB * const rtcb, sigset_t set) sigset_t unmaskset; rtcb->sig.sigprocmask = set; +<<<<<<< HEAD unmaskset = GETUNMASKSET(rtcb->sig.sigprocmask, rtcb->sig.sigPendFlag); if (unmaskset != NULL_SIGNAL_SET) { /* pendlist do */ @@ -214,6 +277,17 @@ void OsSigMaskSwitch(LosTaskCB * const rtcb, sigset_t set) } } +======= + unmaskset = GETUNMASKSET(rtcb->sig.sigprocmask, rtcb->sig.sigPendFlag);//过滤出没有被屏蔽的信号集 + if (unmaskset != NULL_SIGNAL_SET) { + /* pendlist do */ + rtcb->sig.sigFlag |= unmaskset; //加入不屏蔽信号集 + rtcb->sig.sigPendFlag ^= unmaskset;//从挂起/待办集中去掉unmaskset + } +} + + +>>>>>>> remotes/origin/main int OsSigprocMask(int how, const sigset_t_l *setl, sigset_t_l *oldsetl) { LosTaskCB *spcb = NULL; @@ -223,11 +297,19 @@ int OsSigprocMask(int how, const sigset_t_l *setl, sigset_t_l *oldsetl) SCHEDULER_LOCK(intSave); spcb = OsCurrTaskGet(); +<<<<<<< HEAD /* If requested, copy the old mask to user. */ if (oldsetl != NULL) { *(sigset_t *)oldsetl = spcb->sig.sigprocmask; } /* If requested, modify the current signal mask. */ +======= + /* If requested, copy the old mask to user. | 如果需要,请将旧掩码复制给用户*/ + if (oldsetl != NULL) { + *(sigset_t *)oldsetl = spcb->sig.sigprocmask; + } + /* If requested, modify the current signal mask. | 如有要求,修改当前信号屏蔽*/ +>>>>>>> remotes/origin/main if (setl != NULL) { set = *(sigset_t *)setl; /* Okay, determine what we are supposed to do */ @@ -236,35 +318,56 @@ int OsSigprocMask(int how, const sigset_t_l *setl, sigset_t_l *oldsetl) * set pointed to by set as the new sigprocmask. */ case SIG_BLOCK: +<<<<<<< HEAD spcb->sig.sigprocmask |= set; +======= + spcb->sig.sigprocmask |= set;//增加信号屏蔽位 +>>>>>>> remotes/origin/main break; /* Set the intersection of the current set and the * signal set pointed to by set as the new sigprocmask. */ case SIG_UNBLOCK: +<<<<<<< HEAD spcb->sig.sigprocmask &= ~(set); break; /* Set the signal set pointed to by set as the new sigprocmask. */ case SIG_SETMASK: spcb->sig.sigprocmask = set; +======= + spcb->sig.sigprocmask &= ~(set);//解除信号屏蔽位 + break; + /* Set the signal set pointed to by set as the new sigprocmask. */ + case SIG_SETMASK: + spcb->sig.sigprocmask = set;//设置一个新的屏蔽掩码 +>>>>>>> remotes/origin/main break; default: ret = -EINVAL; break; } /* If pending mask not in sigmask, need set sigflag. */ +<<<<<<< HEAD OsSigMaskSwitch(spcb, spcb->sig.sigprocmask); +======= + OsSigMaskSwitch(spcb, spcb->sig.sigprocmask);//更新与屏蔽信号相关的变量 +>>>>>>> remotes/origin/main } SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///让进程的每一个task执行参数函数 +>>>>>>> remotes/origin/main int OsSigProcessForeachChild(LosProcessCB *spcb, ForEachTaskCB handler, void *arg) { int ret; /* Visit the main thread last (if present) */ +<<<<<<< HEAD LosTaskCB *taskCB = NULL; LOS_DL_LIST_FOR_EACH_ENTRY(taskCB, &(spcb->threadSiblingList), LosTaskCB, threadList) { ret = handler(taskCB, arg); @@ -276,6 +379,19 @@ int OsSigProcessForeachChild(LosProcessCB *spcb, ForEachTaskCB handler, void *ar static int SigProcessSignalHandler(LosTaskCB *tcb, void *arg) { struct ProcessSignalInfo *info = (struct ProcessSignalInfo *)arg; +======= + LosTaskCB *taskCB = NULL;//遍历进程的 threadList 链表,里面存放的都是task节点 + LOS_DL_LIST_FOR_EACH_ENTRY(taskCB, &(spcb->threadSiblingList), LosTaskCB, threadList) {//遍历进程的任务列表 + ret = handler(taskCB, arg);//回调参数函数 + OS_RETURN_IF(ret != 0, ret);//这个宏的意思就是只有ret = 0时,啥也不处理.其余就返回 ret + } + return LOS_OK; +} +///信号处理函数,这里就是上面的 handler = SigProcessSignalHandler,见于 OsSigProcessSend +static int SigProcessSignalHandler(LosTaskCB *tcb, void *arg) +{ + struct ProcessSignalInfo *info = (struct ProcessSignalInfo *)arg;//先把参数解出来 +>>>>>>> remotes/origin/main int ret; int isMember; @@ -283,6 +399,7 @@ static int SigProcessSignalHandler(LosTaskCB *tcb, void *arg) return 0; } +<<<<<<< HEAD /* If the default tcb is not set, then set this one as default. */ if (!info->defaultTcb) { info->defaultTcb = tcb; @@ -290,68 +407,126 @@ static int SigProcessSignalHandler(LosTaskCB *tcb, void *arg) isMember = OsSigIsMember(&tcb->sig.sigwaitmask, info->sigInfo->si_signo); if (isMember && (!info->awakenedTcb)) { +======= + /* If the default tcb is not setted, then set this one as default. */ + if (!info->defaultTcb) {//如果没有默认发送方的任务,即默认参数任务. + info->defaultTcb = tcb; + } + + isMember = OsSigIsMember(&tcb->sig.sigwaitmask, info->sigInfo->si_signo);//任务是否在等待这个信号 + if (isMember && (!info->awakenedTcb)) {//是在等待,并尚未向该任务时发送信号时 +>>>>>>> remotes/origin/main /* This means the task is waiting for this signal. Stop looking for it and use this tcb. * The requirement is: if more than one task in this task group is waiting for the signal, * then only one indeterminate task in the group will receive the signal. */ +<<<<<<< HEAD ret = OsTcbDispatch(tcb, info->sigInfo); OS_RETURN_IF(ret < 0, ret); +======= + ret = OsTcbDispatch(tcb, info->sigInfo);//发送信号,注意这是给其他任务发送信号,tcb不是当前任务 + OS_RETURN_IF(ret < 0, ret);//这种写法很有意思 +>>>>>>> remotes/origin/main /* set this tcb as awakenedTcb */ info->awakenedTcb = tcb; OS_RETURN_IF(info->receivedTcb != NULL, SIG_STOP_VISIT); /* Stop search */ } /* Is this signal unblocked on this thread? */ +<<<<<<< HEAD isMember = OsSigIsMember(&tcb->sig.sigprocmask, info->sigInfo->si_signo); if ((!isMember) && (!info->receivedTcb) && (tcb != info->awakenedTcb)) { /* if unblockedTcb of this signal is not set, then set it. */ +======= + isMember = OsSigIsMember(&tcb->sig.sigprocmask, info->sigInfo->si_signo);//任务是否屏蔽了这个信号 + if ((!isMember) && (!info->receivedTcb) && (tcb != info->awakenedTcb)) {//没有屏蔽,有唤醒任务没有接收任务. + /* if unblockedTcb of this signal is not setted, then set it. */ +>>>>>>> remotes/origin/main if (!info->unblockedTcb) { info->unblockedTcb = tcb; } +<<<<<<< HEAD ret = OsTcbDispatch(tcb, info->sigInfo); OS_RETURN_IF(ret < 0, ret); /* set this tcb as receivedTcb */ info->receivedTcb = tcb; +======= + ret = OsTcbDispatch(tcb, info->sigInfo);//向任务发送信号 + OS_RETURN_IF(ret < 0, ret); + /* set this tcb as receivedTcb */ + info->receivedTcb = tcb;//设置这个任务为接收任务 +>>>>>>> remotes/origin/main OS_RETURN_IF(info->awakenedTcb != NULL, SIG_STOP_VISIT); /* Stop search */ } return 0; /* Keep searching */ } +<<<<<<< HEAD static int SigProcessKillSigHandler(LosTaskCB *tcb, void *arg) { struct ProcessSignalInfo *info = (struct ProcessSignalInfo *)arg; +======= +///进程收到 SIGKILL 信号后,通知任务tcb处理. +static int SigProcessKillSigHandler(LosTaskCB *tcb, void *arg) +{ + struct ProcessSignalInfo *info = (struct ProcessSignalInfo *)arg;//转参 +>>>>>>> remotes/origin/main return OsPendingTaskWake(tcb, info->sigInfo->si_signo); } +<<<<<<< HEAD +======= +//处理信号发送 +>>>>>>> remotes/origin/main static void SigProcessLoadTcb(struct ProcessSignalInfo *info, siginfo_t *sigInfo) { LosTaskCB *tcb = NULL; +<<<<<<< HEAD if (info->awakenedTcb == NULL && info->receivedTcb == NULL) { if (info->unblockedTcb) { tcb = info->unblockedTcb; } else if (info->defaultTcb) { +======= + if (info->awakenedTcb == NULL && info->receivedTcb == NULL) {//信号即没有指定接收task 也没有指定被唤醒task + if (info->unblockedTcb) {//如果进程信号信息体中有阻塞task + tcb = info->unblockedTcb;// + } else if (info->defaultTcb) {//如果有默认的发送方task +>>>>>>> remotes/origin/main tcb = info->defaultTcb; } else { return; } /* Deliver the signal to the selected task */ +<<<<<<< HEAD (void)OsTcbDispatch(tcb, sigInfo); } } +======= + (void)OsTcbDispatch(tcb, sigInfo);//向所选任务发送信号 + } +} +///给参数进程发送参数信号 +>>>>>>> remotes/origin/main int OsSigProcessSend(LosProcessCB *spcb, siginfo_t *sigInfo) { int ret; struct ProcessSignalInfo info = { +<<<<<<< HEAD .sigInfo = sigInfo, .defaultTcb = NULL, +======= + .sigInfo = sigInfo, //信号内容 + .defaultTcb = NULL, //以下四个值将在OsSigProcessForeachChild中根据条件完善 +>>>>>>> remotes/origin/main .unblockedTcb = NULL, .awakenedTcb = NULL, .receivedTcb = NULL }; +<<<<<<< HEAD if (info.sigInfo == NULL) { return -EFAULT; @@ -364,20 +539,42 @@ int OsSigProcessSend(LosProcessCB *spcb, siginfo_t *sigInfo) return 0; } else { ret = OsSigProcessForeachChild(spcb, SigProcessSignalHandler, &info); +======= + //总之是要从进程中找个至少一个任务来接受这个信号,优先级 + //awakenedTcb > receivedTcb > unblockedTcb > defaultTcb + if (info.sigInfo == NULL){ + return -EFAULT; + } + /* visit all taskcb and dispatch signal */ //访问所有任务和分发信号 + if (info.sigInfo->si_signo == SIGKILL) {//需要干掉进程时 SIGKILL = 9, #linux kill 9 14 + OsSigAddSet(&spcb->sigShare, info.sigInfo->si_signo);//信号集中增加信号 + (void)OsSigProcessForeachChild(spcb, SigProcessKillSigHandler, &info); + return 0; + } else { + ret = OsSigProcessForeachChild(spcb, SigProcessSignalHandler, &info);//进程通知所有task处理信号 +>>>>>>> remotes/origin/main } if (ret < 0) { return ret; } +<<<<<<< HEAD SigProcessLoadTcb(&info, sigInfo); return 0; } +======= + SigProcessLoadTcb(&info, sigInfo);//确保能给一个任务发送信号 + return 0; +} +///信号集全部清0 +>>>>>>> remotes/origin/main int OsSigEmptySet(sigset_t *set) { *set = NULL_SIGNAL_SET; return 0; } +<<<<<<< HEAD /* Privilege process can't send to kernel and privilege process */ static int OsSignalPermissionToCheck(const LosProcessCB *spcb) { @@ -385,18 +582,33 @@ static int OsSignalPermissionToCheck(const LosProcessCB *spcb) if (gid == OS_KERNEL_PROCESS_GROUP) { return -EPERM; } else if (gid == OS_USER_PRIVILEGE_PROCESS_GROUP) { +======= +/* Privilege process can't send to kernel and privilege process */ //内核进程组和用户特权进程组无法发送 +static int OsSignalPermissionToCheck(const LosProcessCB *spcb) +{ + UINTPTR gid = (UINTPTR)OS_GET_PGROUP_LEADER(spcb->pgroup); + + if (gid == OS_KERNEL_PROCESS_GROUP) {//内核进程组 + return -EPERM; + } else if (gid == OS_USER_PRIVILEGE_PROCESS_GROUP) {//用户特权进程组 +>>>>>>> remotes/origin/main return -EPERM; } return 0; } +<<<<<<< HEAD +======= +///信号分发,发送信号权限/进程组过滤. +>>>>>>> remotes/origin/main STATIC int SendSigPermissionCheck(LosProcessCB *spcb, int permission) { if (spcb == NULL) { return -ESRCH; } +<<<<<<< HEAD if (OsProcessIsUnused(spcb)) { return -ESRCH; } @@ -405,6 +617,16 @@ STATIC int SendSigPermissionCheck(LosProcessCB *spcb, int permission) LosProcessCB *current = OsCurrProcessGet(); /* Kernel process always has kill permission and user process should check permission */ if (OsProcessIsUserMode(current) && !(current->processStatus & OS_PROCESS_FLAG_EXIT)) { +======= + if (OsProcessIsUnused(spcb)) {//进程是否还在使用,不一定是当前进程但必须是个有效进程 + return -ESRCH; + } + +#ifdef LOSCFG_SECURITY_CAPABILITY //启用能力安全模式 + LosProcessCB *current = OsCurrProcessGet();//获取当前进程,检查当前进程是否有发送信号的权限. + /* Kernel process always has kill permission and user process should check permission *///内核进程总是有kill权限,用户进程需要检查权限 + if (OsProcessIsUserMode(current) && !(current->processStatus & OS_PROCESS_FLAG_EXIT)) {//用户进程检查能力范围 +>>>>>>> remotes/origin/main if ((current != spcb) && (!IsCapPermit(CAP_KILL)) && (current->user->userID != spcb->user->userID)) { return -EPERM; } @@ -437,7 +659,11 @@ int OsSendSigToProcess(LosProcessCB *spcb, int sig, int permission) info.si_code = SI_USER; info.si_value.sival_ptr = NULL; +<<<<<<< HEAD return OsSigProcessSend(spcb, &info); +======= + return OsSigProcessSend(spcb, &info);//给参数进程发送信号 +>>>>>>> remotes/origin/main } int OsDispatch(pid_t pid, siginfo_t *info, int permission) @@ -470,14 +696,24 @@ int OsKill(pid_t pid, int sig, int permission) return -EINVAL; } +<<<<<<< HEAD /* Create the siginfo structure */ info.si_signo = sig; info.si_code = SI_USER; +======= + /* Create the siginfo structure */ //创建信号结构体 + info.si_signo = sig; //信号编号 + info.si_code = SI_USER; //来自用户进程信号 +>>>>>>> remotes/origin/main info.si_value.sival_ptr = NULL; if (pid > 0) { /* Send the signal to the specify process */ +<<<<<<< HEAD ret = OsDispatch(pid, &info, permission); +======= + ret = OsDispatch(pid, &info, permission);//发送信号 +>>>>>>> remotes/origin/main } else if (pid == -1) { /* Send SIG to all processes */ ret = OsSendSignalToAllProcess(&info, permission); @@ -489,18 +725,29 @@ int OsKill(pid_t pid, int sig, int permission) } return ret; } +<<<<<<< HEAD +======= +///给发送信号过程加锁 +>>>>>>> remotes/origin/main int OsKillLock(pid_t pid, int sig) { int ret; unsigned int intSave; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsKill(pid, sig, OS_USER_KILL_PERMISSION); SCHEDULER_UNLOCK(intSave); return ret; } +======= + ret = OsKill(pid, sig, OS_USER_KILL_PERMISSION);//用户权限向进程发送信号 + SCHEDULER_UNLOCK(intSave); + return ret; +} +>>>>>>> remotes/origin/main INT32 OsTaskKillUnsafe(UINT32 taskID, INT32 signo) { siginfo_t info; @@ -519,7 +766,11 @@ INT32 OsTaskKillUnsafe(UINT32 taskID, INT32 signo) * dispatch rules. */ return OsTcbDispatch(taskCB, &info); } +<<<<<<< HEAD +======= +///发送信号 +>>>>>>> remotes/origin/main int OsPthreadKill(UINT32 tid, int signo) { int ret; @@ -537,7 +788,11 @@ int OsPthreadKill(UINT32 tid, int signo) SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///向信号集中加入signo信号 +>>>>>>> remotes/origin/main int OsSigAddSet(sigset_t *set, int signo) { /* Verify the signal */ @@ -545,6 +800,7 @@ int OsSigAddSet(sigset_t *set, int signo) return -EINVAL; } else { /* In musl, sig No bits 00000100 present sig No 3, but 1<< 3 = 00001000, so signo needs minus 1 */ +<<<<<<< HEAD signo -= 1; /* Add the signal to the set */ *set |= SIGNO2SET((unsigned int)signo); @@ -552,6 +808,15 @@ int OsSigAddSet(sigset_t *set, int signo) } } +======= + signo -= 1;// 信号范围是 [1 ~ 64 ],而保存变量位的范围是[0 ~ 63] + /* Add the signal to the set */ + *set |= SIGNO2SET((unsigned int)signo);//填充信号集 + return LOS_OK; + } +} +///获取阻塞当前任务的信号集 +>>>>>>> remotes/origin/main int OsSigPending(sigset_t *set) { LosTaskCB *tcb = NULL; @@ -563,7 +828,11 @@ int OsSigPending(sigset_t *set) SCHEDULER_LOCK(intSave); tcb = OsCurrTaskGet(); +<<<<<<< HEAD *set = tcb->sig.sigPendFlag; +======= + *set = tcb->sig.sigPendFlag;//被阻塞的信号集 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return LOS_OK; } @@ -578,7 +847,11 @@ STATIC int FindFirstSetedBit(UINT64 n) for (count = 0; (count < UINT64_BIT_SIZE) && (n ^ 1ULL); n >>= 1, count++) {} return (count < UINT64_BIT_SIZE) ? count : (-1); } +<<<<<<< HEAD +======= +///等待信号时间 +>>>>>>> remotes/origin/main int OsSigTimedWaitNoLock(sigset_t *set, siginfo_t *info, unsigned int timeout) { LosTaskCB *task = NULL; @@ -589,19 +862,32 @@ int OsSigTimedWaitNoLock(sigset_t *set, siginfo_t *info, unsigned int timeout) sigcb = &task->sig; if (sigcb->waitList.pstNext == NULL) { +<<<<<<< HEAD LOS_ListInit(&sigcb->waitList); } /* If pendingflag & set > 0, should clear pending flag */ +======= + LOS_ListInit(&sigcb->waitList);//初始化信号等待链表 + } + /* If pendingflag & set > 0, shound clear pending flag */ +>>>>>>> remotes/origin/main sigset_t clear = sigcb->sigPendFlag & *set; if (clear) { sigcb->sigPendFlag ^= clear; ret = FindFirstSetedBit((UINT64)clear) + 1; OsMoveTmpInfoToUnbInfo(sigcb, ret); } else { +<<<<<<< HEAD OsSigAddSet(set, SIGKILL); OsSigAddSet(set, SIGSTOP); sigcb->sigwaitmask |= *set; +======= + OsSigAddSet(set, SIGKILL);//kill 9 14 必须要处理 + OsSigAddSet(set, SIGSTOP);//终止进程的信号也必须处理 + + sigcb->sigwaitmask |= *set;//按位加到等待集上,也就是说sigwaitmask的信号来了都是要处理的. +>>>>>>> remotes/origin/main OsTaskWaitSetPendMask(OS_TASK_WAIT_SIGNAL, sigcb->sigwaitmask, timeout); ret = task->ops->wait(task, &sigcb->waitList, timeout); if (ret == LOS_ERRNO_TSK_TIMEOUT) { @@ -614,7 +900,11 @@ int OsSigTimedWaitNoLock(sigset_t *set, siginfo_t *info, unsigned int timeout) } return ret; } +<<<<<<< HEAD +======= +///让当前任务等待的信号 +>>>>>>> remotes/origin/main int OsSigTimedWait(sigset_t *set, siginfo_t *info, unsigned int timeout) { int ret; @@ -622,12 +912,20 @@ int OsSigTimedWait(sigset_t *set, siginfo_t *info, unsigned int timeout) SCHEDULER_LOCK(intSave); +<<<<<<< HEAD ret = OsSigTimedWaitNoLock(set, info, timeout); +======= + ret = OsSigTimedWaitNoLock(set, info, timeout);//以不加锁的方式等待 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return ret; } +<<<<<<< HEAD +======= +///通过信号挂起当前任务 +>>>>>>> remotes/origin/main int OsPause(void) { LosTaskCB *spcb = NULL; @@ -637,7 +935,11 @@ int OsPause(void) oldSigprocmask = spcb->sig.sigprocmask; return OsSigSuspend(&oldSigprocmask); } +<<<<<<< HEAD +======= +///用参数set代替进程的原有掩码,并暂停进程执行,直到收到信号再恢复原有掩码并继续执行进程。 +>>>>>>> remotes/origin/main int OsSigSuspend(const sigset_t *set) { unsigned int intSave; @@ -677,6 +979,10 @@ int OsSigSuspend(const sigset_t *set) return -EINTR; } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main int OsSigAction(int sig, const sigaction_t *act, sigaction_t *oact) { UINTPTR addr; @@ -685,6 +991,7 @@ int OsSigAction(int sig, const sigaction_t *act, sigaction_t *oact) if (!GOOD_SIGNO(sig) || sig < 1 || act == NULL) { return -EINVAL; } +<<<<<<< HEAD if (LOS_ArchCopyFromUser(&action, act, sizeof(sigaction_t)) != LOS_OK) { return -EFAULT; } @@ -693,6 +1000,19 @@ int OsSigAction(int sig, const sigaction_t *act, sigaction_t *oact) addr = OsGetSigHandler(); if (addr == 0) { OsSetSigHandler((unsigned long)(UINTPTR)action.sa_handler); +======= + //将数据从用户空间拷贝到内核空间 + if (LOS_ArchCopyFromUser(&action, act, sizeof(sigaction_t)) != LOS_OK) { + return -EFAULT; + } + + if (sig == SIGSYS) {//鸿蒙此处通过错误的系统调用 来安装信号处理函数,有点巧妙. + addr = OsGetSigHandler();//是否已存在信号处理函数 + if (addr == 0) {//进程没有设置信号处理函数时 + OsSetSigHandler((unsigned long)(UINTPTR)action.sa_handler);//设置进程信号处理函数 + //void (*sa_handler)(int); //信号处理函数——普通版 + //void (*sa_sigaction)(int, siginfo_t *, void *);//信号处理函数——高级版 +>>>>>>> remotes/origin/main return LOS_OK; } return -EINVAL; @@ -717,11 +1037,18 @@ VOID OsSigIntUnlock(VOID) (VOID)LOS_AtomicSub((Atomic *)&sigcb->sigIntLock, 1); } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main VOID *OsSaveSignalContext(VOID *sp, VOID *newSp) { UINTPTR sigHandler; UINT32 intSave; +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main LosTaskCB *task = OsCurrTaskGet(); LosProcessCB *process = OsCurrProcessGet(); sig_cb *sigcb = &task->sig; @@ -754,7 +1081,11 @@ VOID *OsSaveSignalContext(VOID *sp, VOID *newSp) OsProcessExitCodeSignalSet(process, signo); sigcb->sigContext = sp; +<<<<<<< HEAD OsInitSignalContext(sp, newSp, sigHandler, signo, sigVal); +======= + OsInitSignalContext(sp, newSp, sigHandler, signo, sigVal);//初始化信号上下文 +>>>>>>> remotes/origin/main /* sig No bits 00000100 present sig No 3, but 1<< 3 = 00001000, so signo needs minus 1 */ sigcb->sigFlag ^= 1ULL << (signo - 1); @@ -767,6 +1098,10 @@ VOID *OsSaveSignalContext(VOID *sp, VOID *newSp) return sp; } +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main VOID *OsRestorSignalContext(VOID *sp) { UINT32 intSave; @@ -785,8 +1120,13 @@ VOID *OsRestorSignalContext(VOID *sp) VOID *saveContext = sigcb->sigContext; sigcb->sigContext = NULL; sigcb->count--; +<<<<<<< HEAD process->sigShare = 0; OsProcessExitCodeSignalClear(process); +======= + process->sigShare = 0; //回到用户态,信号共享清0 + OsProcessExitCodeSignalClear(process);//清空进程退出码 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return saveContext; } diff --git a/src/kernel_liteos_a/kernel/base/mem/common/los_memstat.c b/src/kernel_liteos_a/kernel/base/mem/common/los_memstat.c index 91164b4c..fa219606 100644 --- a/src/kernel_liteos_a/kernel/base/mem/common/los_memstat.c +++ b/src/kernel_liteos_a/kernel/base/mem/common/los_memstat.c @@ -32,9 +32,15 @@ #include "los_memstat_pri.h" #include "los_task_pri.h" +<<<<<<< HEAD LITE_OS_SEC_BSS_MINOR STATIC TskMemUsedInfo g_tskMemUsedInfo[LOSCFG_BASE_CORE_TSK_LIMIT]; +======= +/// 记录每个任务对内存的使用情况 +LITE_OS_SEC_BSS_MINOR STATIC TskMemUsedInfo g_tskMemUsedInfo[LOSCFG_BASE_CORE_TSK_LIMIT]; +/// 计算指定任务对内存使用增加量 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemUsedInc(UINT32 usedSize, UINT32 taskID) { if (taskID >= LOSCFG_BASE_CORE_TSK_LIMIT) { @@ -43,9 +49,15 @@ LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemUsedInc(UINT32 usedSize, UINT32 taskID) if (OS_INT_ACTIVE) { return; } +<<<<<<< HEAD g_tskMemUsedInfo[taskID].memUsed += usedSize; } +======= + g_tskMemUsedInfo[taskID].memUsed += usedSize; ///< 叠加 +} +/// 计算指定任务对内存使用减少量 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemUsedDec(UINT32 usedSize, UINT32 taskID) { if (taskID >= LOSCFG_BASE_CORE_TSK_LIMIT) { @@ -59,9 +71,15 @@ LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemUsedDec(UINT32 usedSize, UINT32 taskID) OsCurrTaskGet()->taskName, g_tskMemUsedInfo[taskID].memUsed, usedSize); return; } +<<<<<<< HEAD g_tskMemUsedInfo[taskID].memUsed -= usedSize; } +======= + g_tskMemUsedInfo[taskID].memUsed -= usedSize; ///< 递减 +} +/// 获取指定任务对内存的使用情况 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsTaskMemUsage(UINT32 taskID) { if (taskID >= LOSCFG_BASE_CORE_TSK_LIMIT) { @@ -70,7 +88,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsTaskMemUsage(UINT32 taskID) return g_tskMemUsedInfo[taskID].memUsed; } +<<<<<<< HEAD +======= +/// 清空任务内存使用记录 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemClear(UINT32 taskID) { if (taskID >= LOSCFG_BASE_CORE_TSK_LIMIT) { @@ -82,8 +104,13 @@ LITE_OS_SEC_TEXT_MINOR VOID OsTaskMemClear(UINT32 taskID) } g_tskMemUsedInfo[taskID].memUsed = 0; } +<<<<<<< HEAD #ifdef LOS_MEM_SLAB +======= +// Slab是一种内存分配器,通过将内存划分不同大小的空间分配给对象使用来进行缓存管理,应用于内核对象的缓存。 +#ifdef LOS_MEM_SLAB // +>>>>>>> remotes/origin/main LITE_OS_SEC_BSS_MINOR STATIC TskSlabUsedInfo g_tskSlabUsedInfo[LOSCFG_BASE_CORE_TSK_LIMIT]; LITE_OS_SEC_TEXT_MINOR VOID OsTaskSlabUsedInc(UINT32 usedSize, UINT32 taskID) diff --git a/src/kernel_liteos_a/kernel/base/mem/membox/los_membox.c b/src/kernel_liteos_a/kernel/base/mem/membox/los_membox.c index 7a4b66de..aed57bd8 100644 --- a/src/kernel_liteos_a/kernel/base/mem/membox/los_membox.c +++ b/src/kernel_liteos_a/kernel/base/mem/membox/los_membox.c @@ -16,6 +16,27 @@ * to endorse or promote products derived from this software without specific prior written * permission. * +<<<<<<< HEAD +======= + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CON/* + * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. + * + * Redistribution and use in source and binary forms, with or without modification, + * are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, this list of + * conditions and the following disclaimer. + * + * 2. Redistributions in binary form must reproduce the above copyright notice, this list + * of conditions and the following disclaimer in the documentation and/or other materials + * provided with the distribution. + * + * 3. Neither the name of the copyright holder nor the names of its contributors may be used + * to endorse or promote products derived from this software without specific prior written + * permission. + * +>>>>>>> remotes/origin/main * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR @@ -33,6 +54,214 @@ #include "los_hwi.h" #include "los_spinlock.h" +<<<<<<< HEAD +======= +#ifdef LOSCFG_AARCH64 +#define OS_MEMBOX_MAGIC 0xa55a5aa5a55a5aa5 +#else +#define OS_MEMBOX_MAGIC 0xa55a5aa5 +#endif +#define OS_MEMBOX_SET_MAGIC(addr) \ + ((LOS_MEMBOX_NODE *)(addr))->pstNext = (LOS_MEMBOX_NODE *)OS_MEMBOX_MAGIC //设置魔法数字 +#define OS_MEMBOX_CHECK_MAGIC(addr) \ + ((((LOS_MEMBOX_NODE *)(addr))->pstNext == (LOS_MEMBOX_NODE *)OS_MEMBOX_MAGIC) ? LOS_OK : LOS_NOK) + +#define OS_MEMBOX_USER_ADDR(addr) \ + ((VOID *)((UINT8 *)(addr) + OS_MEMBOX_NODE_HEAD_SIZE)) +#define OS_MEMBOX_NODE_ADDR(addr) \ + ((LOS_MEMBOX_NODE *)(VOID *)((UINT8 *)(addr) - OS_MEMBOX_NODE_HEAD_SIZE)) //节块 = (节头 + 节体) addr = 节体 +/* spinlock for mem module, only available on SMP mode */ +LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_memboxSpin); +#define MEMBOX_LOCK(state) LOS_SpinLockSave(&g_memboxSpin, &(state)) ///< 获取静态内存池自旋锁 +#define MEMBOX_UNLOCK(state) LOS_SpinUnlockRestore(&g_memboxSpin, (state))///< 释放静态内存池自旋锁 +/// 检查静态内存块 +STATIC INLINE UINT32 OsCheckBoxMem(const LOS_MEMBOX_INFO *boxInfo, const VOID *node) +{ + UINT32 offset; + + if (boxInfo->uwBlkSize == 0) { + return LOS_NOK; + } + + offset = (UINT32)((UINTPTR)node - (UINTPTR)(boxInfo + 1)); + if ((offset % boxInfo->uwBlkSize) != 0) { + return LOS_NOK; + } + + if ((offset / boxInfo->uwBlkSize) >= boxInfo->uwBlkNum) { + return LOS_NOK; + } + + return OS_MEMBOX_CHECK_MAGIC(node);//检查魔法数字是否被修改过了 +} +/// 初始化一个静态内存池,根据入参设定其起始地址、总大小及每个内存块大小 +LITE_OS_SEC_TEXT_INIT UINT32 LOS_MemboxInit(VOID *pool, UINT32 poolSize, UINT32 blkSize) +{ + LOS_MEMBOX_INFO *boxInfo = (LOS_MEMBOX_INFO *)pool;//在内存起始处安置池头 + LOS_MEMBOX_NODE *node = NULL; + UINT32 index; + UINT32 intSave; + + if (pool == NULL) { + return LOS_NOK; + } + + if (blkSize == 0) { + return LOS_NOK; + } + + if (poolSize < sizeof(LOS_MEMBOX_INFO)) { + return LOS_NOK; + } + + MEMBOX_LOCK(intSave); + boxInfo->uwBlkSize = LOS_MEMBOX_ALIGNED(blkSize + OS_MEMBOX_NODE_HEAD_SIZE); //节块总大小(节头+节体) + boxInfo->uwBlkNum = (poolSize - sizeof(LOS_MEMBOX_INFO)) / boxInfo->uwBlkSize;//总节块数量 + boxInfo->uwBlkCnt = 0; //已分配的数量 + if (boxInfo->uwBlkNum == 0) {//只有0块的情况 + MEMBOX_UNLOCK(intSave); + return LOS_NOK; + } + + node = (LOS_MEMBOX_NODE *)(boxInfo + 1);//去除池头,找到第一个节块位置 + + boxInfo->stFreeList.pstNext = node;//池头空闲链表指向第一个节块 + + for (index = 0; index < boxInfo->uwBlkNum - 1; ++index) {//切割节块,挂入空闲链表 + node->pstNext = OS_MEMBOX_NEXT(node, boxInfo->uwBlkSize);//按块大小切割好,统一由pstNext指向 + node = node->pstNext;//node存储了下一个节点的地址信息 + } + + node->pstNext = NULL;//最后一个为null + + MEMBOX_UNLOCK(intSave); + + return LOS_OK; +} +/// 从指定的静态内存池中申请一块静态内存块,整个内核源码只有 OsSwtmrScan中用到了静态内存. +LITE_OS_SEC_TEXT VOID *LOS_MemboxAlloc(VOID *pool) +{ + LOS_MEMBOX_INFO *boxInfo = (LOS_MEMBOX_INFO *)pool; + LOS_MEMBOX_NODE *node = NULL; + LOS_MEMBOX_NODE *nodeTmp = NULL; + UINT32 intSave; + + if (pool == NULL) { + return NULL; + } + + MEMBOX_LOCK(intSave); + node = &(boxInfo->stFreeList);//拿到空闲单链表 + if (node->pstNext != NULL) {//不需要遍历链表,因为这是空闲链表 + nodeTmp = node->pstNext;//先记录要使用的节点 + node->pstNext = nodeTmp->pstNext;//不再空闲了,把节点摘出去了. + OS_MEMBOX_SET_MAGIC(nodeTmp);//为已使用的节块设置魔法数字 + boxInfo->uwBlkCnt++;//已使用块数增加 + } + MEMBOX_UNLOCK(intSave); + + return (nodeTmp == NULL) ? NULL : OS_MEMBOX_USER_ADDR(nodeTmp);//返回可用的虚拟地址 +} +/// 释放指定的一块静态内存块 +LITE_OS_SEC_TEXT UINT32 LOS_MemboxFree(VOID *pool, VOID *box) +{ + LOS_MEMBOX_INFO *boxInfo = (LOS_MEMBOX_INFO *)pool; + UINT32 ret = LOS_NOK; + UINT32 intSave; + + if ((pool == NULL) || (box == NULL)) { + return LOS_NOK; + } + + MEMBOX_LOCK(intSave); + do { + LOS_MEMBOX_NODE *node = OS_MEMBOX_NODE_ADDR(box);//通过节体获取节块首地址 + if (OsCheckBoxMem(boxInfo, node) != LOS_OK) { + break; + } + + node->pstNext = boxInfo->stFreeList.pstNext;//节块指向空闲链表表头 + boxInfo->stFreeList.pstNext = node;//空闲链表表头反指向它,意味节块排到第一,下次申请将首个分配它 + boxInfo->uwBlkCnt--;//已经使用的内存块减一 + ret = LOS_OK; + } while (0);//将被编译时优化 + MEMBOX_UNLOCK(intSave); + + return ret; +} +/// 清零指定静态内存块的内容 +LITE_OS_SEC_TEXT_MINOR VOID LOS_MemboxClr(VOID *pool, VOID *box) +{ + LOS_MEMBOX_INFO *boxInfo = (LOS_MEMBOX_INFO *)pool; + + if ((pool == NULL) || (box == NULL)) { + return; + } + //将魔法数字一并清除了. + (VOID)memset_s(box, (boxInfo->uwBlkSize - OS_MEMBOX_NODE_HEAD_SIZE), 0, + (boxInfo->uwBlkSize - OS_MEMBOX_NODE_HEAD_SIZE)); +} +/// 打印指定静态内存池所有节点信息(打印等级是LOS_INFO_LEVEL),包括内存池起始地址、 +/// 内存块大小、总内存块数量、每个空闲内存块的起始地址、所有内存块的起始地址 +LITE_OS_SEC_TEXT_MINOR VOID LOS_ShowBox(VOID *pool) +{ + UINT32 index; + UINT32 intSave; + LOS_MEMBOX_INFO *boxInfo = (LOS_MEMBOX_INFO *)pool; + LOS_MEMBOX_NODE *node = NULL; + + if (pool == NULL) { + return; + } + MEMBOX_LOCK(intSave); + PRINT_INFO("membox(%p,0x%x,0x%x):\r\n", pool, boxInfo->uwBlkSize, boxInfo->uwBlkNum); + PRINT_INFO("free node list:\r\n"); + + for (node = boxInfo->stFreeList.pstNext, index = 0; node != NULL; + node = node->pstNext, ++index) { + PRINT_INFO("(%u,%p)\r\n", index, node); + } + + PRINT_INFO("all node list:\r\n"); + node = (LOS_MEMBOX_NODE *)(boxInfo + 1); + for (index = 0; index < boxInfo->uwBlkNum; ++index, node = OS_MEMBOX_NEXT(node, boxInfo->uwBlkSize)) { + PRINT_INFO("(%u,%p,%p)\r\n", index, node, node->pstNext); + } + MEMBOX_UNLOCK(intSave); +} +/// 获取指定静态内存池的信息,包括内存池中总内存块数量、已经分配出去的内存块数量、每个内存块的大小 +LITE_OS_SEC_TEXT_MINOR UINT32 LOS_MemboxStatisticsGet(const VOID *boxMem, UINT32 *maxBlk, + UINT32 *blkCnt, UINT32 *blkSize) +{ + if ((boxMem == NULL) || (maxBlk == NULL) || (blkCnt == NULL) || (blkSize == NULL)) { + return LOS_NOK; + } + + *maxBlk = ((OS_MEMBOX_S *)boxMem)->uwBlkNum; + *blkCnt = ((OS_MEMBOX_S *)boxMem)->uwBlkCnt; + *blkSize = ((OS_MEMBOX_S *)boxMem)->uwBlkSize; + + return LOS_OK; +} + +TRIBUTORS + * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, + * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR + * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, + * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, + * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; + * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR + * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF + * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "los_membox.h" +#include "los_hwi.h" +#include "los_spinlock.h" + +>>>>>>> remotes/origin/main #ifdef LOSCFG_AARCH64 #define OS_MEMBOX_MAGIC 0xa55a5aa5a55a5aa5 diff --git a/src/kernel_liteos_a/kernel/base/mem/tlsf/los_memory.c b/src/kernel_liteos_a/kernel/base/mem/tlsf/los_memory.c index 3441b613..ef8e8856 100644 --- a/src/kernel_liteos_a/kernel/base/mem/tlsf/los_memory.c +++ b/src/kernel_liteos_a/kernel/base/mem/tlsf/los_memory.c @@ -43,7 +43,11 @@ #include "los_lms_pri.h" #endif +<<<<<<< HEAD /* Used to cut non-essential functions. */ +======= +/* Used to cut non-essential functions. | 用于削减非必要功能 */ +>>>>>>> remotes/origin/main #define OS_MEM_FREE_BY_TASKID 0 #ifdef LOSCFG_KERNEL_VM #define OS_MEM_EXPAND_ENABLE 1 @@ -56,16 +60,25 @@ /* column num of the output info of mem node */ #define OS_MEM_COLUMN_NUM 8 +<<<<<<< HEAD UINT8 *m_aucSysMem0 = NULL; UINT8 *m_aucSysMem1 = NULL; #ifdef LOSCFG_MEM_MUL_POOL VOID *g_poolHead = NULL; +======= +UINT8 *m_aucSysMem0 = NULL; ///< 异常交互动态内存池地址的起始地址,当不支持异常交互特性时,m_aucSysMem0等于m_aucSysMem1。 +UINT8 *m_aucSysMem1 = NULL; ///< (内核态)系统动态内存池地址的起始地址 + +#ifdef LOSCFG_MEM_MUL_POOL +VOID *g_poolHead = NULL; ///内存池头,由它牵引多个内存池 +>>>>>>> remotes/origin/main #endif /* The following is the macro definition and interface implementation related to the TLSF. */ /* Supposing a Second Level Index: SLI = 3. */ +<<<<<<< HEAD #define OS_MEM_SLI 3 /* Giving 1 free list for each small bucket: 4, 8, 12, up to 124. */ #define OS_MEM_SMALL_BUCKET_COUNT 31 @@ -82,6 +95,24 @@ VOID *g_poolHead = NULL; #define OS_MEM_BITMAP_WORDS ((OS_MEM_FREE_LIST_COUNT >> 5) + 1) #define OS_MEM_BITMAP_MASK 0x1FU +======= +#define OS_MEM_SLI 3 ///< 二级小区间级数, +/* Giving 1 free list for each small bucket: 4, 8, 12, up to 124. */ +#define OS_MEM_SMALL_BUCKET_COUNT 31 ///< 小桶的偏移单位 从 4 ~ 124 ,共32级 +#define OS_MEM_SMALL_BUCKET_MAX_SIZE 128 ///< 小桶的最大数量 +/* Giving OS_MEM_FREE_LIST_NUM free lists for each large bucket. */ +#define OS_MEM_LARGE_BUCKET_COUNT 24 /// 为每个大存储桶空闲列表数量 大桶范围: [2^7, 2^31] ,每个小区间有分为 2^3个小区间 +#define OS_MEM_FREE_LIST_NUM (1 << OS_MEM_SLI) ///< 2^3 = 8 个,即大桶的每个区间又分为8个小区间 +/* OS_MEM_SMALL_BUCKET_MAX_SIZE to the power of 2 is 7. */ +#define OS_MEM_LARGE_START_BUCKET 7 /// 大桶的开始下标 + +/* The count of free list. */ +#define OS_MEM_FREE_LIST_COUNT (OS_MEM_SMALL_BUCKET_COUNT + (OS_MEM_LARGE_BUCKET_COUNT << OS_MEM_SLI)) ///< 总链表的数量 32 + 24 * 8 = 224 +/* The bitmap is used to indicate whether the free list is empty, 1: not empty, 0: empty. */ +#define OS_MEM_BITMAP_WORDS ((OS_MEM_FREE_LIST_COUNT >> 5) + 1) ///< 224 >> 5 + 1 = 7 ,为什么要右移 5 因为 2^5 = 32 是一个32位整型的大小 + ///< 而 32 * 7 = 224 ,也就是说用 int[7]当位图就能表示完 224个链表 ,此处,一定要理解好,因为这是理解 TLSF 算法的关键. +#define OS_MEM_BITMAP_MASK 0x1FU ///< 因为一个int型为 32位, 2^5 = 32,所以此处 0x1FU = 5个1 足以. +>>>>>>> remotes/origin/main /* Used to find the first bit of 1 in bitmap. */ STATIC INLINE UINT16 OsMemFFS(UINT32 bitmap) @@ -101,7 +132,11 @@ STATIC INLINE UINT32 OsMemLog2(UINT32 size) return OsMemFLS(size); } +<<<<<<< HEAD /* Get the first level: f = log2(size). */ +======= +/* Get the first level: f = log2(size). | 获取第一级*/ +>>>>>>> remotes/origin/main STATIC INLINE UINT32 OsMemFlGet(UINT32 size) { if (size < OS_MEM_SMALL_BUCKET_MAX_SIZE) { @@ -110,13 +145,18 @@ STATIC INLINE UINT32 OsMemFlGet(UINT32 size) return OsMemLog2(size); } +<<<<<<< HEAD /* Get the second level: s = (size - 2^f) * 2^SLI / 2^f. */ +======= +/* Get the second level: s = (size - 2^f) * 2^SLI / 2^f. | 获取第二级 */ +>>>>>>> remotes/origin/main STATIC INLINE UINT32 OsMemSlGet(UINT32 size, UINT32 fl) { return (((size << OS_MEM_SLI) >> fl) - OS_MEM_FREE_LIST_NUM); } /* The following is the memory algorithm related macro definition and interface implementation. */ +<<<<<<< HEAD struct OsMemNodeHead { UINT32 magic; @@ -160,6 +200,51 @@ struct OsMemPoolHead { SPIN_LOCK_S spinlock; #ifdef LOSCFG_MEM_MUL_POOL VOID *nextPool; +======= +/// 内存池节点 +struct OsMemNodeHead { + UINT32 magic; ///< 魔法数字 0xABCDDCBA + union {//注意这里的前后指向的是连续的地址节点,用于分割和合并 + struct OsMemNodeHead *prev; /* The prev is used for current node points to the previous node | prev 用于当前节点指向前一个节点*/ + struct OsMemNodeHead *next; /* The next is used for last node points to the expand node | next 用于最后一个节点指向展开节点*/ + } ptr; +#ifdef LOSCFG_MEM_LEAKCHECK //内存泄漏检测 + UINTPTR linkReg[LOS_RECORD_LR_CNT];///< 存放左右节点地址,用于检测 +#endif + UINT32 sizeAndFlag; ///< 节点总大小+标签 +}; +/// 已使用内存池节点 +struct OsMemUsedNodeHead { + struct OsMemNodeHead header;///< 已被使用节点 +#if OS_MEM_FREE_BY_TASKID + UINT32 taskID; ///< 使用节点的任务ID +#endif +}; +/// 内存池空闲节点 +struct OsMemFreeNodeHead { + struct OsMemNodeHead header; ///< 内存池节点 + struct OsMemFreeNodeHead *prev; ///< 前一个空闲前驱节点 + struct OsMemFreeNodeHead *next; ///< 后一个空闲后继节点 +}; +/// 内存池信息 +struct OsMemPoolInfo { + VOID *pool; ///< 指向内存块基地址,仅做记录而已,真正的分配内存跟它没啥关系 + UINT32 totalSize; ///< 总大小,确定了内存池的边界 + UINT32 attr; ///< 属性 default attr: lock, not expand. +#ifdef LOSCFG_MEM_WATERLINE + UINT32 waterLine; /* Maximum usage size in a memory pool | 内存吃水线*/ + UINT32 curUsedSize; /* Current usage size in a memory pool | 当前已使用大小*/ +#endif +}; +/// 内存池头信息 +struct OsMemPoolHead { + struct OsMemPoolInfo info; ///< 记录内存池的信息 + UINT32 freeListBitmap[OS_MEM_BITMAP_WORDS]; ///< 空闲位图 int[7] = 32 * 7 = 224 > 223 + struct OsMemFreeNodeHead *freeList[OS_MEM_FREE_LIST_COUNT];///< 空闲节点链表 32 + 24 * 8 = 224 + SPIN_LOCK_S spinlock; ///< 操作本池的自旋锁,涉及CPU多核竞争,所以必须得是自旋锁 +#ifdef LOSCFG_MEM_MUL_POOL + VOID *nextPool; ///< 指向下一个内存池 OsMemPoolHead 类型 +>>>>>>> remotes/origin/main #endif }; @@ -168,6 +253,7 @@ struct OsMemPoolHead { #define MEM_UNLOCK(pool, state) LOS_SpinUnlockRestore(&(pool)->spinlock, (state)) /* The memory pool support expand. */ +<<<<<<< HEAD #define OS_MEM_POOL_EXPAND_ENABLE 0x01 /* The memory pool support no lock. */ #define OS_MEM_POOL_LOCK_ENABLE 0x02 @@ -178,6 +264,18 @@ struct OsMemPoolHead { #define OS_MEM_NODE_USED_FLAG 0x80000000U #define OS_MEM_NODE_ALIGNED_FLAG 0x40000000U #define OS_MEM_NODE_LAST_FLAG 0x20000000U /* Sentinel Node */ +======= +#define OS_MEM_POOL_EXPAND_ENABLE 0x01 ///< 支持扩展 +/* The memory pool support no lock. */ +#define OS_MEM_POOL_LOCK_ENABLE 0x02 ///< 加锁 + +#define OS_MEM_NODE_MAGIC 0xABCDDCBA ///< 内存节点的魔法数字 +#define OS_MEM_MIN_ALLOC_SIZE (sizeof(struct OsMemFreeNodeHead) - sizeof(struct OsMemUsedNodeHead)) //最小分配空间 +// 必须给指向空闲块的指针留位置 +#define OS_MEM_NODE_USED_FLAG 0x80000000U ///< 已使用标签 +#define OS_MEM_NODE_ALIGNED_FLAG 0x40000000U ///< 对齐标签 +#define OS_MEM_NODE_LAST_FLAG 0x20000000U /* Sentinel Node | 哨兵节点标签*/ +>>>>>>> remotes/origin/main #define OS_MEM_NODE_ALIGNED_AND_USED_FLAG (OS_MEM_NODE_USED_FLAG | OS_MEM_NODE_ALIGNED_FLAG | OS_MEM_NODE_LAST_FLAG) #define OS_MEM_NODE_GET_ALIGNED_FLAG(sizeAndFlag) \ @@ -226,32 +324,56 @@ STATIC INLINE UINT32 OsMemAllocCheck(struct OsMemPoolHead *pool, UINT32 intSave) #if OS_MEM_FREE_BY_TASKID STATIC INLINE VOID OsMemNodeSetTaskID(struct OsMemUsedNodeHead *node) { +<<<<<<< HEAD node->taskID = LOS_CurTaskIDGet(); +======= + node->taskID = LOS_CurTaskIDGet();//将当前任务ID绑定到内存池节点上 +>>>>>>> remotes/origin/main } #endif #ifdef LOSCFG_MEM_WATERLINE STATIC INLINE VOID OsMemWaterUsedRecord(struct OsMemPoolHead *pool, UINT32 size) { +<<<<<<< HEAD pool->info.curUsedSize += size; if (pool->info.curUsedSize > pool->info.waterLine) { pool->info.waterLine = pool->info.curUsedSize; +======= + pool->info.curUsedSize += size; //延长可使用空间 + if (pool->info.curUsedSize > pool->info.waterLine) { + pool->info.waterLine = pool->info.curUsedSize; //警戒线加高 +>>>>>>> remotes/origin/main } } #else STATIC INLINE VOID OsMemWaterUsedRecord(struct OsMemPoolHead *pool, UINT32 size) { +<<<<<<< HEAD (VOID)pool; (VOID)size; +======= + (VOID)pool; + (VOID)size; +>>>>>>> remotes/origin/main } #endif #if OS_MEM_EXPAND_ENABLE +<<<<<<< HEAD STATIC INLINE struct OsMemNodeHead *OsMemLastSentinelNodeGet(const struct OsMemNodeHead *sentinelNode) { struct OsMemNodeHead *node = NULL; VOID *ptr = sentinelNode->ptr.next; UINT32 size = OS_MEM_NODE_GET_SIZE(sentinelNode->sizeAndFlag); +======= +/// 更新哨兵节点内容 +STATIC INLINE struct OsMemNodeHead *OsMemLastSentinelNodeGet(const struct OsMemNodeHead *sentinelNode) +{ + struct OsMemNodeHead *node = NULL; + VOID *ptr = sentinelNode->ptr.next;//返回不连续的内存块 + UINT32 size = OS_MEM_NODE_GET_SIZE(sentinelNode->sizeAndFlag); // 获取大小 +>>>>>>> remotes/origin/main while ((ptr != NULL) && (size != 0)) { node = OS_MEM_END_NODE(ptr, size); @@ -261,7 +383,11 @@ STATIC INLINE struct OsMemNodeHead *OsMemLastSentinelNodeGet(const struct OsMemN return node; } +<<<<<<< HEAD +======= +/// 检查哨兵节点 +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsMemSentinelNodeCheck(struct OsMemNodeHead *sentinelNode) { if (!OS_MEM_NODE_GET_USED_FLAG(sentinelNode->sizeAndFlag)) { @@ -274,7 +400,11 @@ STATIC INLINE BOOL OsMemSentinelNodeCheck(struct OsMemNodeHead *sentinelNode) return TRUE; } +<<<<<<< HEAD +======= +/// 是否为最后一个哨兵节点 +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsMemIsLastSentinelNode(struct OsMemNodeHead *sentinelNode) { if (OsMemSentinelNodeCheck(sentinelNode) == FALSE) { @@ -289,11 +419,19 @@ STATIC INLINE BOOL OsMemIsLastSentinelNode(struct OsMemNodeHead *sentinelNode) return FALSE; } +<<<<<<< HEAD STATIC INLINE VOID OsMemSentinelNodeSet(struct OsMemNodeHead *sentinelNode, VOID *newNode, UINT32 size) { if (sentinelNode->ptr.next != NULL) { sentinelNode = OsMemLastSentinelNodeGet(sentinelNode); +======= +/// 设置哨兵节点内容 +STATIC INLINE VOID OsMemSentinelNodeSet(struct OsMemNodeHead *sentinelNode, VOID *newNode, UINT32 size) +{ + if (sentinelNode->ptr.next != NULL) { //哨兵节点有 逻辑地址不连续的衔接内存块 + sentinelNode = OsMemLastSentinelNodeGet(sentinelNode);//更新哨兵节点内容 +>>>>>>> remotes/origin/main } sentinelNode->sizeAndFlag = size; @@ -329,6 +467,7 @@ STATIC INLINE struct OsMemNodeHead *PreSentinelNodeGet(const VOID *pool, const s return NULL; } +<<<<<<< HEAD UINT32 OsMemLargeNodeFree(const VOID *ptr) { @@ -337,6 +476,16 @@ UINT32 OsMemLargeNodeFree(const VOID *ptr) return LOS_NOK; } LOS_PhysPagesFreeContiguous((VOID *)ptr, page->nPages); +======= +/// 大内存释放 +UINT32 OsMemLargeNodeFree(const VOID *ptr) +{ + LosVmPage *page = OsVmVaddrToPage((VOID *)ptr);//获取物理页 + if ((page == NULL) || (page->nPages == 0)) { + return LOS_NOK; + } + LOS_PhysPagesFreeContiguous((VOID *)ptr, page->nPages);//释放连续的几个物理页 +>>>>>>> remotes/origin/main return LOS_OK; } @@ -375,7 +524,11 @@ STATIC INLINE BOOL TryShrinkPool(const VOID *pool, const struct OsMemNodeHead *n #endif return TRUE; } +<<<<<<< HEAD +======= +/// 内存池扩展实现 +>>>>>>> remotes/origin/main STATIC INLINE INT32 OsMemPoolExpandSub(VOID *pool, UINT32 size, UINT32 intSave) { UINT32 tryCount = MAX_SHRINK_PAGECACHE_TRY; @@ -383,11 +536,19 @@ STATIC INLINE INT32 OsMemPoolExpandSub(VOID *pool, UINT32 size, UINT32 intSave) struct OsMemNodeHead *newNode = NULL; struct OsMemNodeHead *endNode = NULL; +<<<<<<< HEAD size = ROUNDUP(size + OS_MEM_NODE_HEAD_SIZE, PAGE_SIZE); endNode = OS_MEM_END_NODE(pool, poolInfo->info.totalSize); RETRY: newNode = (struct OsMemNodeHead *)LOS_PhysPagesAllocContiguous(size >> PAGE_SHIFT); +======= + size = ROUNDUP(size + OS_MEM_NODE_HEAD_SIZE, PAGE_SIZE);//圆整 + endNode = OS_MEM_END_NODE(pool, poolInfo->info.totalSize);//获取哨兵节点 + +RETRY: + newNode = (struct OsMemNodeHead *)LOS_PhysPagesAllocContiguous(size >> PAGE_SHIFT);//申请新的内存池 | 物理内存 +>>>>>>> remotes/origin/main if (newNode == NULL) { if (tryCount > 0) { tryCount--; @@ -411,6 +572,7 @@ RETRY: size = (resize == 0) ? size : resize; } #endif +<<<<<<< HEAD newNode->sizeAndFlag = (size - OS_MEM_NODE_HEAD_SIZE); newNode->ptr.prev = OS_MEM_END_NODE(newNode, size); OsMemSentinelNodeSet(endNode, newNode, size); @@ -431,6 +593,28 @@ STATIC INLINE INT32 OsMemPoolExpand(VOID *pool, UINT32 allocSize, UINT32 intSave UINT32 expandDefault = MEM_EXPAND_SIZE(LOS_MemPoolSizeGet(pool)); UINT32 expandSize = MAX(expandDefault, allocSize); UINT32 tryCount = 1; +======= + newNode->sizeAndFlag = (size - OS_MEM_NODE_HEAD_SIZE);//设置新节点大小 + newNode->ptr.prev = OS_MEM_END_NODE(newNode, size);//新节点的前节点指向新节点的哨兵节点 + OsMemSentinelNodeSet(endNode, newNode, size);//设置老内存池的哨兵节点信息,其实就是指向新内存块 + OsMemFreeNodeAdd(pool, (struct OsMemFreeNodeHead *)newNode);//将新节点加入空闲链表 + + endNode = OS_MEM_END_NODE(newNode, size);//获取新节点的哨兵节点 + (VOID)memset(endNode, 0, sizeof(*endNode));//清空内存 + endNode->ptr.next = NULL;//新哨兵节点没有后续指向,因为它已成为最后 + endNode->magic = OS_MEM_NODE_MAGIC;//设置新哨兵节的魔法数字 + OsMemSentinelNodeSet(endNode, NULL, 0); //设置新哨兵节点内容 + OsMemWaterUsedRecord(poolInfo, OS_MEM_NODE_HEAD_SIZE);//更新内存池警戒线 + + return 0; +} +/// 扩展内存池 +STATIC INLINE INT32 OsMemPoolExpand(VOID *pool, UINT32 allocSize, UINT32 intSave) +{ + UINT32 expandDefault = MEM_EXPAND_SIZE(LOS_MemPoolSizeGet(pool));//至少要扩展现有内存池的 1/8 大小 + UINT32 expandSize = MAX(expandDefault, allocSize); + UINT32 tryCount = 1;//尝试次数 +>>>>>>> remotes/origin/main UINT32 ret; do { @@ -447,7 +631,11 @@ STATIC INLINE INT32 OsMemPoolExpand(VOID *pool, UINT32 allocSize, UINT32 intSave return -1; } +<<<<<<< HEAD +======= +///< 允许指定内存池扩展 +>>>>>>> remotes/origin/main VOID LOS_MemExpandEnable(VOID *pool) { if (pool == NULL) { @@ -486,7 +674,11 @@ STATIC INLINE VOID OsLmsAllocAlignMark(VOID *ptr, VOID *alignedPtr, UINT32 size) g_lms->simpleMark((UINTPTR)ptr + sizeof(UINT32), (UINTPTR)alignedPtr, LMS_SHADOW_REDZONE_U8); } +<<<<<<< HEAD /* mark remaining as redzone */ +======= + /* mark remining as redzone */ +>>>>>>> remotes/origin/main g_lms->simpleMark(LMS_ADDR_ALIGN((UINTPTR)alignedPtr + size), (UINTPTR)OS_MEM_NEXT_NODE(allocNode), LMS_SHADOW_REDZONE_U8); } @@ -522,8 +714,12 @@ STATIC INLINE VOID OsLmsReallocResizeMark(struct OsMemNodeHead *node, UINT32 res g_lms->simpleMark((UINTPTR)node + resize, (UINTPTR)OS_MEM_NEXT_NODE(node), LMS_SHADOW_REDZONE_U8); } #endif +<<<<<<< HEAD #ifdef LOSCFG_MEM_LEAKCHECK +======= +#ifdef LOSCFG_MEM_LEAKCHECK //内存泄漏检查 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsMemLinkRegisterRecord(struct OsMemNodeHead *node) { LOS_RecordLR(node->linkReg, LOS_RECORD_LR_CNT, LOS_RECORD_LR_CNT, LOS_OMIT_LR_CNT); @@ -549,7 +745,11 @@ STATIC INLINE VOID OsMemUsedNodePrint(struct OsMemNodeHead *node) PRINTK("\n"); } } +<<<<<<< HEAD +======= +/// 打印已使用的节点 +>>>>>>> remotes/origin/main VOID OsMemUsedNodeShow(VOID *pool) { if (pool == NULL) { @@ -626,17 +826,29 @@ STATIC VOID OsMemNodeBacktraceInfo(const struct OsMemNodeHead *tmpNode, STATIC INLINE UINT32 OsMemFreeListIndexGet(UINT32 size) { +<<<<<<< HEAD UINT32 fl = OsMemFlGet(size); +======= + UINT32 fl = OsMemFlGet(size);//获取一级位图 +>>>>>>> remotes/origin/main if (size < OS_MEM_SMALL_BUCKET_MAX_SIZE) { return fl; } +<<<<<<< HEAD UINT32 sl = OsMemSlGet(size, fl); +======= + UINT32 sl = OsMemSlGet(size, fl);//获取二级位图 +>>>>>>> remotes/origin/main return (OS_MEM_SMALL_BUCKET_COUNT + ((fl - OS_MEM_LARGE_START_BUCKET) << OS_MEM_SLI) + sl); } STATIC INLINE struct OsMemFreeNodeHead *OsMemFindCurSuitableBlock(struct OsMemPoolHead *poolHead, +<<<<<<< HEAD UINT32 index, UINT32 size) +======= + UINT32 index, UINT32 size) +>>>>>>> remotes/origin/main { struct OsMemFreeNodeHead *node = NULL; @@ -663,7 +875,11 @@ STATIC INLINE UINT32 OsMemNotEmptyIndexGet(struct OsMemPoolHead *poolHead, UINT3 return OS_MEM_FREE_LIST_COUNT; } +<<<<<<< HEAD +======= +/// 找到下一个合适的块 +>>>>>>> remotes/origin/main STATIC INLINE struct OsMemFreeNodeHead *OsMemFindNextSuitableBlock(VOID *pool, UINT32 size, UINT32 *outIndex) { struct OsMemPoolHead *poolHead = (struct OsMemPoolHead *)pool; @@ -730,13 +946,22 @@ STATIC INLINE VOID OsMemListAdd(struct OsMemPoolHead *pool, UINT32 listIndex, st OsMemSetFreeListBit(pool, listIndex); node->header.magic = OS_MEM_NODE_MAGIC; } +<<<<<<< HEAD +======= +/// 从空闲链表中删除 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsMemListDelete(struct OsMemPoolHead *pool, UINT32 listIndex, struct OsMemFreeNodeHead *node) { if (node == pool->freeList[listIndex]) { pool->freeList[listIndex] = node->next; +<<<<<<< HEAD if (node->next == NULL) { OsMemClearFreeListBit(pool, listIndex); +======= + if (node->next == NULL) {//如果链表空了 + OsMemClearFreeListBit(pool, listIndex);//将位图位 置为 0 +>>>>>>> remotes/origin/main } else { node->next->prev = NULL; } @@ -748,27 +973,47 @@ STATIC INLINE VOID OsMemListDelete(struct OsMemPoolHead *pool, UINT32 listIndex, } node->header.magic = OS_MEM_NODE_MAGIC; } +<<<<<<< HEAD STATIC INLINE VOID OsMemFreeNodeAdd(VOID *pool, struct OsMemFreeNodeHead *node) { UINT32 index = OsMemFreeListIndexGet(node->header.sizeAndFlag); +======= +/// 添加一个空闲节点 +STATIC INLINE VOID OsMemFreeNodeAdd(VOID *pool, struct OsMemFreeNodeHead *node) +{ + UINT32 index = OsMemFreeListIndexGet(node->header.sizeAndFlag);//根据大小定位索引位 +>>>>>>> remotes/origin/main if (index >= OS_MEM_FREE_LIST_COUNT) { LOS_Panic("The index of free lists is error, index = %u\n", index); return; } +<<<<<<< HEAD OsMemListAdd(pool, index, node); } STATIC INLINE VOID OsMemFreeNodeDelete(VOID *pool, struct OsMemFreeNodeHead *node) { UINT32 index = OsMemFreeListIndexGet(node->header.sizeAndFlag); +======= + OsMemListAdd(pool, index, node);//挂入链表 +} +/// 从空闲链表上摘除节点 +STATIC INLINE VOID OsMemFreeNodeDelete(VOID *pool, struct OsMemFreeNodeHead *node) +{ + UINT32 index = OsMemFreeListIndexGet(node->header.sizeAndFlag);//根据大小定位索引位 +>>>>>>> remotes/origin/main if (index >= OS_MEM_FREE_LIST_COUNT) { LOS_Panic("The index of free lists is error, index = %u\n", index); return; } OsMemListDelete(pool, index, node); } +<<<<<<< HEAD +======= +//获取一个空闲的节点 +>>>>>>> remotes/origin/main STATIC INLINE struct OsMemNodeHead *OsMemFreeNodeGet(VOID *pool, UINT32 size) { struct OsMemPoolHead *poolHead = (struct OsMemPoolHead *)pool; @@ -782,11 +1027,16 @@ STATIC INLINE struct OsMemNodeHead *OsMemFreeNodeGet(VOID *pool, UINT32 size) return &firstNode->header; } +<<<<<<< HEAD +======= +/// 合并节点,和前面的节点合并 node 消失 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsMemMergeNode(struct OsMemNodeHead *node) { struct OsMemNodeHead *nextNode = NULL; +<<<<<<< HEAD node->ptr.prev->sizeAndFlag += node->sizeAndFlag; nextNode = (struct OsMemNodeHead *)((UINTPTR)node + node->sizeAndFlag); if (!OS_MEM_NODE_GET_LAST_FLAG(nextNode->sizeAndFlag)) { @@ -794,11 +1044,21 @@ STATIC INLINE VOID OsMemMergeNode(struct OsMemNodeHead *node) } } +======= + node->ptr.prev->sizeAndFlag += node->sizeAndFlag; //前节点长度变长 + nextNode = (struct OsMemNodeHead *)((UINTPTR)node + node->sizeAndFlag); // 下一个节点位置 + if (!OS_MEM_NODE_GET_LAST_FLAG(nextNode->sizeAndFlag)) {//不是哨兵节点 + nextNode->ptr.prev = node->ptr.prev;//后一个节点的前节点变成前前节点 + } +} +/// 切割节点 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsMemSplitNode(VOID *pool, struct OsMemNodeHead *allocNode, UINT32 allocSize) { struct OsMemFreeNodeHead *newFreeNode = NULL; struct OsMemNodeHead *nextNode = NULL; +<<<<<<< HEAD newFreeNode = (struct OsMemFreeNodeHead *)(VOID *)((UINT8 *)allocNode + allocSize); newFreeNode->header.ptr.prev = allocNode; newFreeNode->header.sizeAndFlag = allocNode->sizeAndFlag - allocSize; @@ -824,14 +1084,58 @@ STATIC INLINE VOID *OsMemCreateUsedNode(VOID *addr) #endif #ifdef LOSCFG_KERNEL_LMS +======= + newFreeNode = (struct OsMemFreeNodeHead *)(VOID *)((UINT8 *)allocNode + allocSize);//切割后出现的新空闲节点,在分配节点的右侧 + newFreeNode->header.ptr.prev = allocNode;//新节点指向前节点,说明是从左到右切割 + newFreeNode->header.sizeAndFlag = allocNode->sizeAndFlag - allocSize;//新空闲节点大小 + allocNode->sizeAndFlag = allocSize;//分配节点大小 + nextNode = OS_MEM_NEXT_NODE(&newFreeNode->header);//获取新节点的下一个节点 + if (!OS_MEM_NODE_GET_LAST_FLAG(nextNode->sizeAndFlag)) {//如果下一个节点不是哨兵节点(末尾节点) + nextNode->ptr.prev = &newFreeNode->header;//下一个节点的前节点为新空闲节点 + if (!OS_MEM_NODE_GET_USED_FLAG(nextNode->sizeAndFlag)) {//如果下一个节点也是空闲的 + OsMemFreeNodeDelete(pool, (struct OsMemFreeNodeHead *)nextNode);//删除下一个节点信息 + OsMemMergeNode(nextNode);//下一个节点和新空闲节点 合并成一个新节点 + } + } + + OsMemFreeNodeAdd(pool, newFreeNode);//挂入空闲链表 +} +// +STATIC INLINE VOID *OsMemCreateUsedNode(VOID *addr) +{ + struct OsMemUsedNodeHead *node = (struct OsMemUsedNodeHead *)addr;//直接将地址转成使用节点,说明节点信息也存在内存池中 + //这种用法是非常巧妙的 +#if OS_MEM_FREE_BY_TASKID + OsMemNodeSetTaskID(node);//设置使用内存节点的任务 +#endif + +#ifdef LOSCFG_KERNEL_LMS //检测内存泄漏 +>>>>>>> remotes/origin/main struct OsMemNodeHead *newNode = (struct OsMemNodeHead *)node; if (g_lms != NULL) { g_lms->mallocMark(newNode, OS_MEM_NEXT_NODE(newNode), OS_MEM_NODE_HEAD_SIZE); } #endif +<<<<<<< HEAD return node + 1; } +======= + return node + 1; //@note_good 这个地方挺有意思的,只是将结构体扩展下,留一个 int 位 ,变成了已使用节点,返回的地址正是要分配给应用的地址 +} + +/*! + * @brief OsMemPoolInit 内存池初始化 + * 内存池节点部分包含3种类型节点:未使用空闲内存节点(OsMemFreeNodeHead),已使用内存节点(OsMemUsedNodeHead) 和 尾节点(OsMemNodeHead)。 + * \n 每个内存节点维护一个前序指针,指向内存池中上一个内存节点,还维护内存节点的大小和使用标记。 + * \n 空闲内存节点和已使用内存节点后面的内存区域是数据域 + * @param pool + * @param size + * @return + * + * @see + */ +>>>>>>> remotes/origin/main STATIC UINT32 OsMemPoolInit(VOID *pool, UINT32 size) { struct OsMemPoolHead *poolHead = (struct OsMemPoolHead *)pool; @@ -853,6 +1157,7 @@ STATIC UINT32 OsMemPoolInit(VOID *pool, UINT32 size) #endif LOS_SpinInit(&poolHead->spinlock); +<<<<<<< HEAD poolHead->info.pool = pool; poolHead->info.totalSize = size; poolHead->info.attr = OS_MEM_POOL_LOCK_ENABLE; /* default attr: lock, not expand. */ @@ -877,6 +1182,34 @@ STATIC UINT32 OsMemPoolInit(VOID *pool, UINT32 size) #ifdef LOSCFG_MEM_WATERLINE poolHead->info.curUsedSize = sizeof(struct OsMemPoolHead) + OS_MEM_NODE_HEAD_SIZE; poolHead->info.waterLine = poolHead->info.curUsedSize; +======= + poolHead->info.pool = pool; //内存池的起始地址,但注意真正的内存并不是从此处分配,它只是用来记录这个内存块的开始位置而已. + poolHead->info.totalSize = size;//内存池总大小 + poolHead->info.attr = OS_MEM_POOL_LOCK_ENABLE; /* default attr: lock, not expand. | 默认是上锁,不支持扩展,需扩展得另外设置*/ + + newNode = OS_MEM_FIRST_NODE(pool);//跳到第一个节点位置,即跳过结构体本身位置,真正的分配内存是从newNode开始的. + newNode->sizeAndFlag = (size - sizeof(struct OsMemPoolHead) - OS_MEM_NODE_HEAD_SIZE);//这才是可供分配给外界使用的总内存块大小,即数据域 + //OS_MEM_NODE_HEAD_SIZE 叫当前使用节点,即指 newNode占用的空间 + newNode->ptr.prev = NULL;//开始是空指向 + newNode->magic = OS_MEM_NODE_MAGIC;//魔法数字 用于标识这是一个 OsMemNodeHead 节点, 魔法数字不能被覆盖, + OsMemFreeNodeAdd(pool, (struct OsMemFreeNodeHead *)newNode);//添加一个空闲节点,由此有了首个可供分配的空闲节点 + + /* The last mem node */ + endNode = OS_MEM_END_NODE(pool, size);//确定尾节点位置,尾节点没有数据域 + endNode->magic = OS_MEM_NODE_MAGIC; //填入尾节点的魔法数字 +#if OS_MEM_EXPAND_ENABLE //支持扩展 + endNode->ptr.next = NULL;//尾节点没有后继节点 + OsMemSentinelNodeSet(endNode, NULL, 0);//将尾节点设置为哨兵节点 +#else + endNode->sizeAndFlag = 0;//0代表没有数据域 + endNode->ptr.prev = newNode;//前驱指针指向第一个节点 + OS_MEM_NODE_SET_USED_FLAG(endNode->sizeAndFlag); +#endif +#ifdef LOSCFG_MEM_WATERLINE //吃水线开关 + poolHead->info.curUsedSize = sizeof(struct OsMemPoolHead) + OS_MEM_NODE_HEAD_SIZE;//内存池已使用了这么多空间,这些都是存内存池自身数据的空间, + //但此处是否还要算是 endNode ? @note_thinking + poolHead->info.waterLine = poolHead->info.curUsedSize; //设置吃水线 +>>>>>>> remotes/origin/main #endif #ifdef LOSCFG_KERNEL_LMS if (resize != 0) { @@ -896,13 +1229,21 @@ STATIC VOID OsMemPoolDeinit(const VOID *pool, UINT32 size) #endif (VOID)memset_s(pool, size, 0, sizeof(struct OsMemPoolHead)); } +<<<<<<< HEAD +======= +/// 新增内存池 +>>>>>>> remotes/origin/main STATIC UINT32 OsMemPoolAdd(VOID *pool, UINT32 size) { VOID *nextPool = g_poolHead; VOID *curPool = g_poolHead; UINTPTR poolEnd; +<<<<<<< HEAD while (nextPool != NULL) { +======= + while (nextPool != NULL) {//单链表遍历方式 +>>>>>>> remotes/origin/main poolEnd = (UINTPTR)nextPool + LOS_MemPoolSizeGet(nextPool); if (((pool <= nextPool) && (((UINTPTR)pool + size) > (UINTPTR)nextPool)) || (((UINTPTR)pool < poolEnd) && (((UINTPTR)pool + size) >= poolEnd))) { @@ -916,6 +1257,7 @@ STATIC UINT32 OsMemPoolAdd(VOID *pool, UINT32 size) } if (g_poolHead == NULL) { +<<<<<<< HEAD g_poolHead = pool; } else { ((struct OsMemPoolHead *)curPool)->nextPool = pool; @@ -925,6 +1267,17 @@ STATIC UINT32 OsMemPoolAdd(VOID *pool, UINT32 size) return LOS_OK; } +======= + g_poolHead = pool; //首个内存池 + } else { + ((struct OsMemPoolHead *)curPool)->nextPool = pool; //两池扯上关系 + } + + ((struct OsMemPoolHead *)pool)->nextPool = NULL; //新池下一个无所指 + return LOS_OK; +} +/// 删除内存池 +>>>>>>> remotes/origin/main STATIC UINT32 OsMemPoolDelete(const VOID *pool) { UINT32 ret = LOS_NOK; @@ -955,29 +1308,57 @@ STATIC UINT32 OsMemPoolDelete(const VOID *pool) } #endif +<<<<<<< HEAD +======= +/*! + * @brief LOS_MemInit 初始化一块指定的动态内存池,大小为size + * 初始一个内存池后生成一个内存池控制头、尾节点EndNode,剩余的内存被标记为FreeNode内存节点。 + * @param pool + * @param size + * @return + * @attention EndNode作为内存池末尾的节点,size为0。 + * @see + */ +>>>>>>> remotes/origin/main UINT32 LOS_MemInit(VOID *pool, UINT32 size) { if ((pool == NULL) || (size <= OS_MEM_MIN_POOL_SIZE)) { return OS_ERROR; } +<<<<<<< HEAD size = OS_MEM_ALIGN(size, OS_MEM_ALIGN_SIZE); +======= + size = OS_MEM_ALIGN(size, OS_MEM_ALIGN_SIZE);//4个字节对齐 +>>>>>>> remotes/origin/main if (OsMemPoolInit(pool, size)) { return OS_ERROR; } +<<<<<<< HEAD #ifdef LOSCFG_MEM_MUL_POOL +======= +#ifdef LOSCFG_MEM_MUL_POOL //多内存池开关 +>>>>>>> remotes/origin/main if (OsMemPoolAdd(pool, size)) { (VOID)OsMemPoolDeInit(pool, size); return OS_ERROR; } #endif +<<<<<<< HEAD OsHookCall(LOS_HOOK_TYPE_MEM_INIT, pool, size); +======= + OsHookCall(LOS_HOOK_TYPE_MEM_INIT, pool, size);//打印日志 +>>>>>>> remotes/origin/main return LOS_OK; } #ifdef LOSCFG_MEM_MUL_POOL +<<<<<<< HEAD +======= +/// 删除指定内存池 +>>>>>>> remotes/origin/main UINT32 LOS_MemDeInit(VOID *pool) { struct OsMemPoolHead *tmpPool = (struct OsMemPoolHead *)pool; @@ -997,7 +1378,11 @@ UINT32 LOS_MemDeInit(VOID *pool) OsHookCall(LOS_HOOK_TYPE_MEM_DEINIT, tmpPool); return LOS_OK; } +<<<<<<< HEAD +======= +/// 打印系统中已初始化的所有内存池,包括内存池的起始地址、内存池大小、空闲内存总大小、已使用内存总大小、最大的空闲内存块大小、空闲内存块数量、已使用的内存块数量。 +>>>>>>> remotes/origin/main UINT32 LOS_MemPoolList(VOID) { VOID *nextPool = g_poolHead; @@ -1011,7 +1396,11 @@ UINT32 LOS_MemPoolList(VOID) return index; } #endif +<<<<<<< HEAD +======= +/// 从指定动态内存池中申请size长度的内存 +>>>>>>> remotes/origin/main STATIC INLINE VOID *OsMemAlloc(struct OsMemPoolHead *pool, UINT32 size, UINT32 intSave) { struct OsMemNodeHead *allocNode = NULL; @@ -1024,6 +1413,7 @@ STATIC INLINE VOID *OsMemAlloc(struct OsMemPoolHead *pool, UINT32 size, UINT32 i UINT32 allocSize = OS_MEM_ALIGN(size + OS_MEM_NODE_HEAD_SIZE, OS_MEM_ALIGN_SIZE); #if OS_MEM_EXPAND_ENABLE +<<<<<<< HEAD retry: #endif allocNode = OsMemFreeNodeGet(pool, allocSize); @@ -1033,6 +1423,17 @@ retry: INT32 ret = OsMemPoolExpand(pool, allocSize, intSave); if (ret == 0) { goto retry; +======= +retry: //这种写法也挺赞的 @note_good +#endif + allocNode = OsMemFreeNodeGet(pool, allocSize);//获取空闲节点 + if (allocNode == NULL) {//没有内存了,怎搞? +#if OS_MEM_EXPAND_ENABLE + if (pool->info.attr & OS_MEM_POOL_EXPAND_ENABLE) { + INT32 ret = OsMemPoolExpand(pool, allocSize, intSave);//扩展内存池 + if (ret == 0) { + goto retry;//再来一遍 +>>>>>>> remotes/origin/main } } #endif @@ -1047,6 +1448,7 @@ retry: return NULL; } +<<<<<<< HEAD if ((allocSize + OS_MEM_NODE_HEAD_SIZE + OS_MEM_MIN_ALLOC_SIZE) <= allocNode->sizeAndFlag) { OsMemSplitNode(pool, allocNode, allocSize); } @@ -1063,6 +1465,24 @@ retry: VOID *LOS_MemAlloc(VOID *pool, UINT32 size) { if ((pool == NULL) || (size == 0)) { +======= + if ((allocSize + OS_MEM_NODE_HEAD_SIZE + OS_MEM_MIN_ALLOC_SIZE) <= allocNode->sizeAndFlag) {//所需小于内存池可供分配量 + OsMemSplitNode(pool, allocNode, allocSize);//劈开内存池 + } + + OS_MEM_NODE_SET_USED_FLAG(allocNode->sizeAndFlag);//给节点贴上已使用的标签 + OsMemWaterUsedRecord(pool, OS_MEM_NODE_GET_SIZE(allocNode->sizeAndFlag));//更新吃水线 + +#ifdef LOSCFG_MEM_LEAKCHECK //检测内存泄漏开关 + OsMemLinkRegisterRecord(allocNode); +#endif + return OsMemCreateUsedNode((VOID *)allocNode);//创建已使用节点 +} +/// 从指定内存池中申请size长度的内存,注意这可不是从内核堆空间中申请内存 +VOID *LOS_MemAlloc(VOID *pool, UINT32 size) +{ + if ((pool == NULL) || (size == 0)) {//没提供内存池时 +>>>>>>> remotes/origin/main return (size > 0) ? OsVmBootMemAlloc(size) : NULL; } @@ -1079,6 +1499,7 @@ VOID *LOS_MemAlloc(VOID *pool, UINT32 size) break; } MEM_LOCK(poolHead, intSave); +<<<<<<< HEAD ptr = OsMemAlloc(poolHead, size, intSave); MEM_UNLOCK(poolHead, intSave); } while (0); @@ -1087,6 +1508,16 @@ VOID *LOS_MemAlloc(VOID *pool, UINT32 size) return ptr; } +======= + ptr = OsMemAlloc(poolHead, size, intSave);//真正的分配内存函数,详细查看 鸿蒙内核源码分析(内存池篇) + MEM_UNLOCK(poolHead, intSave); + } while (0); + + OsHookCall(LOS_HOOK_TYPE_MEM_ALLOC, pool, ptr, size);//打印日志,到此一游 + return ptr; +} +/// 从指定内存池中申请size长度的内存且地址按boundary字节对齐的内存 +>>>>>>> remotes/origin/main VOID *LOS_MemAllocAlign(VOID *pool, UINT32 size, UINT32 boundary) { UINT32 gapSize; @@ -1143,10 +1574,17 @@ VOID *LOS_MemAllocAlign(VOID *pool, UINT32 size, UINT32 boundary) ptr = alignedPtr; } while (0); +<<<<<<< HEAD OsHookCall(LOS_HOOK_TYPE_MEM_ALLOCALIGN, pool, ptr, size, boundary); return ptr; } +======= + OsHookCall(LOS_HOOK_TYPE_MEM_ALLOCALIGN, pool, ptr, size, boundary);//打印对齐日志,表示程序曾临幸过此处 + return ptr; +} +/// 内存池有效性检查 +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsMemAddrValidCheck(const struct OsMemPoolHead *pool, const VOID *addr) { UINT32 size; @@ -1160,7 +1598,11 @@ STATIC INLINE BOOL OsMemAddrValidCheck(const struct OsMemPoolHead *pool, const V if (OS_MEM_MIDDLE_ADDR_OPEN_END(pool + 1, addr, (UINTPTR)pool + size)) { return TRUE; } +<<<<<<< HEAD #if OS_MEM_EXPAND_ENABLE +======= +#if OS_MEM_EXPAND_ENABLE //如果支持可扩展 +>>>>>>> remotes/origin/main struct OsMemNodeHead *node = NULL; struct OsMemNodeHead *sentinel = OS_MEM_END_NODE(pool, size); while (OsMemIsLastSentinelNode(sentinel) == FALSE) { @@ -1250,7 +1692,11 @@ STATIC UINT32 OsMemCheckUsedNode(const struct OsMemPoolHead *pool, const struct return LOS_OK; } +<<<<<<< HEAD +======= +/// 释放内存 +>>>>>>> remotes/origin/main STATIC INLINE UINT32 OsMemFree(struct OsMemPoolHead *pool, struct OsMemNodeHead *node) { UINT32 ret = OsMemCheckUsedNode(pool, node); @@ -1260,10 +1706,17 @@ STATIC INLINE UINT32 OsMemFree(struct OsMemPoolHead *pool, struct OsMemNodeHead } #ifdef LOSCFG_MEM_WATERLINE +<<<<<<< HEAD pool->info.curUsedSize -= OS_MEM_NODE_GET_SIZE(node->sizeAndFlag); #endif node->sizeAndFlag = OS_MEM_NODE_GET_SIZE(node->sizeAndFlag); +======= + pool->info.curUsedSize -= OS_MEM_NODE_GET_SIZE(node->sizeAndFlag);//降低水位线 +#endif + + node->sizeAndFlag = OS_MEM_NODE_GET_SIZE(node->sizeAndFlag);//获取大小和标记 +>>>>>>> remotes/origin/main #ifdef LOSCFG_MEM_LEAKCHECK OsMemLinkRegisterRecord(node); #endif @@ -1274,6 +1727,7 @@ STATIC INLINE UINT32 OsMemFree(struct OsMemPoolHead *pool, struct OsMemNodeHead g_lms->check((UINTPTR)node + OS_MEM_NODE_HEAD_SIZE, TRUE); } #endif +<<<<<<< HEAD struct OsMemNodeHead *preNode = node->ptr.prev; /* merage preNode */ if ((preNode != NULL) && !OS_MEM_NODE_GET_USED_FLAG(preNode->sizeAndFlag)) { OsMemFreeNodeDelete(pool, (struct OsMemFreeNodeHead *)preNode); @@ -1285,6 +1739,19 @@ STATIC INLINE UINT32 OsMemFree(struct OsMemPoolHead *pool, struct OsMemNodeHead if ((nextNode != NULL) && !OS_MEM_NODE_GET_USED_FLAG(nextNode->sizeAndFlag)) { OsMemFreeNodeDelete(pool, (struct OsMemFreeNodeHead *)nextNode); OsMemMergeNode(nextNode); +======= + struct OsMemNodeHead *preNode = node->ptr.prev; /* merage preNode | 合并前一个节点*/ + if ((preNode != NULL) && !OS_MEM_NODE_GET_USED_FLAG(preNode->sizeAndFlag)) { + OsMemFreeNodeDelete(pool, (struct OsMemFreeNodeHead *)preNode);//删除前节点的信息 + OsMemMergeNode(node);//向前合并 + node = preNode; + } + + struct OsMemNodeHead *nextNode = OS_MEM_NEXT_NODE(node); /* merage nextNode | 计算后一个节点位置*/ + if ((nextNode != NULL) && !OS_MEM_NODE_GET_USED_FLAG(nextNode->sizeAndFlag)) { + OsMemFreeNodeDelete(pool, (struct OsMemFreeNodeHead *)nextNode);//删除后节点信息 + OsMemMergeNode(nextNode);//合并节点 +>>>>>>> remotes/origin/main } #if OS_MEM_EXPAND_ENABLE @@ -1305,7 +1772,11 @@ STATIC INLINE UINT32 OsMemFree(struct OsMemPoolHead *pool, struct OsMemNodeHead #endif return ret; } +<<<<<<< HEAD +======= +/// 释放从指定动态内存中申请的内存 +>>>>>>> remotes/origin/main UINT32 LOS_MemFree(VOID *pool, VOID *ptr) { UINT32 intSave; @@ -1321,13 +1792,21 @@ UINT32 LOS_MemFree(VOID *pool, VOID *ptr) struct OsMemNodeHead *node = NULL; do { +<<<<<<< HEAD UINT32 gapSize = *(UINT32 *)((UINTPTR)ptr - sizeof(UINT32)); +======= + UINT32 gapSize = *(UINT32 *)((UINTPTR)ptr - sizeof(UINT32));//获取节点大小和标签 即: sizeAndFlag +>>>>>>> remotes/origin/main if (OS_MEM_NODE_GET_ALIGNED_FLAG(gapSize) && OS_MEM_NODE_GET_USED_FLAG(gapSize)) { PRINT_ERR("[%s:%d]gapSize:0x%x error\n", __FUNCTION__, __LINE__, gapSize); break; } +<<<<<<< HEAD node = (struct OsMemNodeHead *)((UINTPTR)ptr - OS_MEM_NODE_HEAD_SIZE); +======= + node = (struct OsMemNodeHead *)((UINTPTR)ptr - OS_MEM_NODE_HEAD_SIZE);//定位到节点开始位置 +>>>>>>> remotes/origin/main if (OS_MEM_NODE_GET_ALIGNED_FLAG(gapSize)) { gapSize = OS_MEM_NODE_GET_ALIGNED_GAPSIZE(gapSize); @@ -1415,7 +1894,11 @@ STATIC INLINE VOID *OsGetRealPtr(const VOID *pool, VOID *ptr) } STATIC INLINE VOID *OsMemRealloc(struct OsMemPoolHead *pool, const VOID *ptr, +<<<<<<< HEAD struct OsMemNodeHead *node, UINT32 size, UINT32 intSave) +======= + struct OsMemNodeHead *node, UINT32 size, UINT32 intSave) +>>>>>>> remotes/origin/main { struct OsMemNodeHead *nextNode = NULL; UINT32 allocSize = OS_MEM_ALIGN(size + OS_MEM_NODE_HEAD_SIZE, OS_MEM_ALIGN_SIZE); @@ -1446,7 +1929,11 @@ STATIC INLINE VOID *OsMemRealloc(struct OsMemPoolHead *pool, const VOID *ptr, } return tmpPtr; } +<<<<<<< HEAD +======= +/// 按size大小重新分配内存块,并将原内存块内容拷贝到新内存块。如果新内存块申请成功,则释放原内存块 +>>>>>>> remotes/origin/main VOID *LOS_MemRealloc(VOID *pool, VOID *ptr, UINT32 size) { if ((pool == NULL) || OS_MEM_NODE_GET_USED_FLAG(size) || OS_MEM_NODE_GET_ALIGNED_FLAG(size)) { @@ -1539,7 +2026,11 @@ UINT32 LOS_MemFreeByTaskID(VOID *pool, UINT32 taskID) return LOS_OK; } #endif +<<<<<<< HEAD +======= +/// 获取指定动态内存池的总大小 +>>>>>>> remotes/origin/main UINT32 LOS_MemPoolSizeGet(const VOID *pool) { UINT32 count = 0; @@ -1548,6 +2039,7 @@ UINT32 LOS_MemPoolSizeGet(const VOID *pool) return LOS_NOK; } +<<<<<<< HEAD count += ((struct OsMemPoolHead *)pool)->info.totalSize; #if OS_MEM_EXPAND_ENABLE @@ -1560,11 +2052,29 @@ UINT32 LOS_MemPoolSizeGet(const VOID *pool) node = OsMemSentinelNodeGet(sentinel); sentinel = OS_MEM_END_NODE(node, size); count += size; +======= + count += ((struct OsMemPoolHead *)pool)->info.totalSize; // 这里的 += 好像没必要吧?, = 就可以了, @note_thinking + +#if OS_MEM_EXPAND_ENABLE //支持扩展 + UINT32 size; + struct OsMemNodeHead *node = NULL; + struct OsMemNodeHead *sentinel = OS_MEM_END_NODE(pool, count);//获取哨兵节点 + + while (OsMemIsLastSentinelNode(sentinel) == FALSE) {//不是最后一个节点 + size = OS_MEM_NODE_GET_SIZE(sentinel->sizeAndFlag);//数据域大小 + node = OsMemSentinelNodeGet(sentinel);//再获取哨兵节点 + sentinel = OS_MEM_END_NODE(node, size);//获取尾节点 + count += size; //内存池大小变大 +>>>>>>> remotes/origin/main } #endif return count; } +<<<<<<< HEAD +======= +/// 获取指定动态内存池的总使用量大小 +>>>>>>> remotes/origin/main UINT32 LOS_MemTotalUsedGet(VOID *pool) { struct OsMemNodeHead *tmpNode = NULL; @@ -1637,7 +2147,11 @@ STATIC UINT32 OsMemAddrValidCheckPrint(const VOID *pool, struct OsMemFreeNodeHea } STATIC UINT32 OsMemIntegrityCheckSub(struct OsMemNodeHead **tmpNode, const VOID *pool, +<<<<<<< HEAD const struct OsMemNodeHead *endNode) +======= + const struct OsMemNodeHead *endNode) +>>>>>>> remotes/origin/main { if (!OS_MEM_MAGIC_VALID(*tmpNode)) { OsMemMagicCheckPrint(tmpNode); @@ -1653,7 +2167,11 @@ STATIC UINT32 OsMemIntegrityCheckSub(struct OsMemNodeHead **tmpNode, const VOID } STATIC UINT32 OsMemFreeListNodeCheck(const struct OsMemPoolHead *pool, +<<<<<<< HEAD const struct OsMemFreeNodeHead *node) +======= + const struct OsMemFreeNodeHead *node) +>>>>>>> remotes/origin/main { if (!OsMemAddrValidCheck(pool, node) || !OsMemAddrValidCheck(pool, node->prev) || @@ -1714,9 +2232,15 @@ OUT: #endif } } +<<<<<<< HEAD STATIC UINT32 OsMemIntegrityCheck(const struct OsMemPoolHead *pool, struct OsMemNodeHead **tmpNode, struct OsMemNodeHead **preNode) +======= +//对指定内存池做完整性检查, +STATIC UINT32 OsMemIntegrityCheck(const struct OsMemPoolHead *pool, struct OsMemNodeHead **tmpNode, + struct OsMemNodeHead **preNode) +>>>>>>> remotes/origin/main { struct OsMemNodeHead *endNode = OS_MEM_END_NODE(pool, pool->info.totalSize); @@ -1839,7 +2363,11 @@ STATIC INLINE UINT32 OsMemAllocCheck(struct OsMemPoolHead *pool, UINT32 intSave) return LOS_OK; } #endif +<<<<<<< HEAD +======= +/// 对指定内存池做完整性检查 +>>>>>>> remotes/origin/main UINT32 LOS_MemIntegrityCheck(const VOID *pool) { if (pool == NULL) { @@ -1864,7 +2392,11 @@ ERROR_OUT: } STATIC INLINE VOID OsMemInfoGet(struct OsMemPoolHead *poolInfo, struct OsMemNodeHead *node, +<<<<<<< HEAD LOS_MEM_POOL_STATUS *poolStatus) +======= + LOS_MEM_POOL_STATUS *poolStatus) +>>>>>>> remotes/origin/main { UINT32 totalUsedSize = 0; UINT32 totalFreeSize = 0; @@ -1893,8 +2425,22 @@ STATIC INLINE VOID OsMemInfoGet(struct OsMemPoolHead *poolInfo, struct OsMemNode poolStatus->freeNodeNum += freeNodeNum; } +<<<<<<< HEAD UINT32 LOS_MemInfoGet(VOID *pool, LOS_MEM_POOL_STATUS *poolStatus) { +======= +/*! + * @brief LOS_MemInfoGet + * 获取指定内存池的内存结构信息,包括空闲内存大小、已使用内存大小、空闲内存块数量、已使用的内存块数量、最大的空闲内存块大小 + * @param pool + * @param poolStatus + * @return + * + * @see + */ +UINT32 LOS_MemInfoGet(VOID *pool, LOS_MEM_POOL_STATUS *poolStatus) +{//内存碎片率计算:同样调用LOS_MemInfoGet接口,可以获取内存池的剩余内存大小和最大空闲内存块大小,然后根据公式(fragment=100-最大空闲内存块大小/剩余内存大小)得出此时的动态内存池碎片率。 +>>>>>>> remotes/origin/main struct OsMemPoolHead *poolInfo = pool; if (poolStatus == NULL) { @@ -1975,7 +2521,11 @@ STATIC VOID OsMemInfoPrint(VOID *pool) status.freeNodeNum); #endif } +<<<<<<< HEAD +======= +/// 打印指定内存池的空闲内存块的大小及数量 +>>>>>>> remotes/origin/main UINT32 LOS_MemFreeNodeShow(VOID *pool) { struct OsMemPoolHead *poolInfo = (struct OsMemPoolHead *)pool; @@ -2022,7 +2572,11 @@ UINT32 LOS_MemFreeNodeShow(VOID *pool) return LOS_OK; } +<<<<<<< HEAD +======= +///内核空间动态内存(堆内存)初始化 , 争取系统动态内存池 +>>>>>>> remotes/origin/main STATUS_T OsKHeapInit(size_t size) { STATUS_T ret; @@ -2041,14 +2595,20 @@ STATUS_T OsKHeapInit(size_t size) return -1; } +<<<<<<< HEAD m_aucSysMem0 = m_aucSysMem1 = ptr; ret = LOS_MemInit(m_aucSysMem0, size); +======= + m_aucSysMem0 = m_aucSysMem1 = ptr;// 指定内核内存池的位置 + ret = LOS_MemInit(m_aucSysMem0, size); //初始化内存池,供内核分配动态内存 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { PRINT_ERR("vmm_kheap_init LOS_MemInit failed!\n"); g_vmBootMemBase -= size; return ret; } #if OS_MEM_EXPAND_ENABLE +<<<<<<< HEAD LOS_MemExpandEnable(OS_SYS_MEM_ADDR); #endif return LOS_OK; @@ -2073,6 +2633,32 @@ BOOL OsMemIsHeapNode(const VOID *ptr) firstNode = OsMemSentinelNodeGet(endNode); endNode = OS_MEM_END_NODE(firstNode, size); if (OS_MEM_MIDDLE_ADDR(firstNode, ptr, endNode)) { +======= + LOS_MemExpandEnable(OS_SYS_MEM_ADDR);//支持扩展系统动态内存 +#endif + return LOS_OK; +} +///< 判断地址是否在堆区 +BOOL OsMemIsHeapNode(const VOID *ptr) +{ + struct OsMemPoolHead *pool = (struct OsMemPoolHead *)m_aucSysMem1;//内核堆区开始地址 + struct OsMemNodeHead *firstNode = OS_MEM_FIRST_NODE(pool);//获取内存池首个节点 + struct OsMemNodeHead *endNode = OS_MEM_END_NODE(pool, pool->info.totalSize);//获取内存池的尾节点 + + if (OS_MEM_MIDDLE_ADDR(firstNode, ptr, endNode)) {//如果在首尾范围内 + return TRUE; + } + +#if OS_MEM_EXPAND_ENABLE//内存池经过扩展后,新旧块的虚拟地址是不连续的,所以需要跳块判断 + UINT32 intSave; + UINT32 size;//详细查看百篇博客系列篇之 鸿蒙内核源码分析(内存池篇) + MEM_LOCK(pool, intSave); //获取自旋锁 + while (OsMemIsLastSentinelNode(endNode) == FALSE) { //哨兵节点是内存池结束的标记 + size = OS_MEM_NODE_GET_SIZE(endNode->sizeAndFlag);//获取节点大小 + firstNode = OsMemSentinelNodeGet(endNode);//获取下一块的开始地址 + endNode = OS_MEM_END_NODE(firstNode, size);//获取下一块的尾节点 + if (OS_MEM_MIDDLE_ADDR(firstNode, ptr, endNode)) {//判断地址是否在该块中 +>>>>>>> remotes/origin/main MEM_UNLOCK(pool, intSave); return TRUE; } @@ -2081,3 +2667,8 @@ BOOL OsMemIsHeapNode(const VOID *ptr) #endif return FALSE; } +<<<<<<< HEAD +======= + + +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/misc/kill_shellcmd.c b/src/kernel_liteos_a/kernel/base/misc/kill_shellcmd.c index e21dca7d..ebefbc76 100644 --- a/src/kernel_liteos_a/kernel/base/misc/kill_shellcmd.c +++ b/src/kernel_liteos_a/kernel/base/misc/kill_shellcmd.c @@ -48,6 +48,26 @@ #include "shell.h" #endif +<<<<<<< HEAD +======= +/********************************************* +命令功能 +命令用于发送特定信号给指定进程。 + +命令格式 +kill [signo | -signo] [pid] + +参数 参数说明 取值范围 +signo 信号ID [1,30] +pid 进程ID [1,MAX_INT] + +须知: signo有效范围为[0,64],建议取值范围为[1,30],其余为保留内容。 + +使用指南 +必须指定发送的信号编号及进程号。 +进程编号取值范围根据系统配置变化,例如系统最大支持pid为256,则取值范围缩小为[1-256]。 +*********************************************/ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsPrintKillUsage(VOID) { PRINTK("\nkill: usage: kill [sigspec] [pid]\n"); diff --git a/src/kernel_liteos_a/kernel/base/misc/los_misc.c b/src/kernel_liteos_a/kernel/base/misc/los_misc.c index 05a083ee..85fa9021 100644 --- a/src/kernel_liteos_a/kernel/base/misc/los_misc.c +++ b/src/kernel_liteos_a/kernel/base/misc/los_misc.c @@ -31,7 +31,14 @@ #include "los_task_pri.h" +<<<<<<< HEAD +======= +// 这个函数接受两个参数,一个是地址(addr),另一个是对齐边界(boundary)。它返回一个对齐后的地址。 +// 这个函数首先检查地址加上边界再减去1是否大于地址,如果是, +// 就返回地址加上边界再减去1然后与~((UINTPTR)(boundary - 1))进行位与运算的结果, +// 否则就返回地址与~((UINTPTR)(boundary - 1))进行位与运算的结果。这实际上是在对地址进行对齐。 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT UINTPTR LOS_Align(UINTPTR addr, UINT32 boundary) { if ((addr + boundary - 1) > addr) { @@ -41,6 +48,7 @@ LITE_OS_SEC_TEXT UINTPTR LOS_Align(UINTPTR addr, UINT32 boundary) } } +<<<<<<< HEAD LITE_OS_SEC_TEXT_MINOR VOID LOS_Msleep(UINT32 msecs) { UINT32 interval; @@ -48,6 +56,16 @@ LITE_OS_SEC_TEXT_MINOR VOID LOS_Msleep(UINT32 msecs) if (msecs == 0) { interval = 0; } else { +======= + +LITE_OS_SEC_TEXT_MINOR VOID LOS_Msleep(UINT32 msecs) +{ + UINT32 interval; + // 这个函数接受一个参数,毫秒数(msecs)。这个函数首先检查毫秒数是否为0,如果是,就将间隔设置为0。 + if (msecs == 0) { + interval = 0; + // } else { 否则,它将毫秒数转换为tick数(可能是操作系统的时间单位),如果转换后的间隔为0,就将间隔设置为1。然后,它调用LOS_TaskDelay函数来延迟指定的间隔。 +>>>>>>> remotes/origin/main interval = LOS_MS2Tick(msecs); if (interval == 0) { interval = 1; diff --git a/src/kernel_liteos_a/kernel/base/misc/los_stackinfo.c b/src/kernel_liteos_a/kernel/base/misc/los_stackinfo.c index 6f96a8a5..08bf636e 100644 --- a/src/kernel_liteos_a/kernel/base/misc/los_stackinfo.c +++ b/src/kernel_liteos_a/kernel/base/misc/los_stackinfo.c @@ -37,14 +37,47 @@ #include "shcmd.h" #include "shell.h" #endif +<<<<<<< HEAD const StackInfo *g_stackInfo = NULL; UINT32 g_stackNum; +======= +/** + * @file los_stackinfo.c + * @brief 栈内容 + * @verbatim + @note_pic OsExcStackInfo 各个CPU栈布局图,其他栈也是一样,CPU各核硬件栈都是紧挨着 + __undef_stack(SMP) + +-------------------+ <--- cpu1 top + | | + | CPU core1 | + | | + +--------------------<--- cpu2 top + | | + | cpu core 2 | + | | + +--------------------<--- cpu3 top + | | + | cpu core 3 | + | | + +--------------------<--- cpu4 top + | | + | cpu core 4 | + | | + +-------------------+ + * @endverbatim + */ + +const StackInfo *g_stackInfo = NULL; ///< CPU所有工作模式的栈信息 +UINT32 g_stackNum; ///< CPU所有工作模式的栈数量 +///获取栈的吃水线 +>>>>>>> remotes/origin/main UINT32 OsStackWaterLineGet(const UINTPTR *stackBottom, const UINTPTR *stackTop, UINT32 *peakUsed) { UINT32 size; const UINTPTR *tmp = NULL; +<<<<<<< HEAD if (*stackTop == OS_STACK_MAGIC_WORD) { tmp = stackTop + 1; while ((tmp < stackBottom) && (*tmp == OS_STACK_INIT)) { @@ -59,6 +92,22 @@ UINT32 OsStackWaterLineGet(const UINTPTR *stackBottom, const UINTPTR *stackTop, } } +======= + if (*stackTop == OS_STACK_MAGIC_WORD) {//栈顶值是否等于 magic 0xCCCCCCCC + tmp = stackTop + 1; + while ((tmp < stackBottom) && (*tmp == OS_STACK_INIT)) {//记录从栈顶到栈低有多少个连续的 0xCACACACA + tmp++; + } + size = (UINT32)((UINTPTR)stackBottom - (UINTPTR)tmp);//剩余多少非0xCACACACA的栈空间 + *peakUsed = (size == 0) ? size : (size + sizeof(CHAR *));//得出高峰用值,还剩多少可用 + return LOS_OK; + } else { + *peakUsed = OS_INVALID_WATERLINE;//栈溢出了 + return LOS_NOK; + } +} +///异常情况下的栈检查,主要就是检查栈顶值有没有被改写 +>>>>>>> remotes/origin/main VOID OsExcStackCheck(VOID) { UINT32 index; @@ -71,7 +120,11 @@ VOID OsExcStackCheck(VOID) for (index = 0; index < g_stackNum; index++) { for (cpuid = 0; cpuid < LOSCFG_KERNEL_CORE_NUM; cpuid++) { stackTop = (UINTPTR *)((UINTPTR)g_stackInfo[index].stackTop + cpuid * g_stackInfo[index].stackSize); +<<<<<<< HEAD if (*stackTop != OS_STACK_MAGIC_WORD) { +======= + if (*stackTop != OS_STACK_MAGIC_WORD) {// 只要栈顶内容不是 0xCCCCCCCCC 就是溢出了. +>>>>>>> remotes/origin/main PRINT_ERR("cpu:%u %s overflow , magic word changed to 0x%x\n", LOSCFG_KERNEL_CORE_NUM - 1 - cpuid, g_stackInfo[index].stackName, *stackTop); } @@ -79,6 +132,10 @@ VOID OsExcStackCheck(VOID) } } +<<<<<<< HEAD +======= +///打印栈的信息 把每个CPU的栈信息打印出来 +>>>>>>> remotes/origin/main VOID OsExcStackInfo(VOID) { UINT32 index; @@ -93,18 +150,27 @@ VOID OsExcStackInfo(VOID) PrintExcInfo("\n stack name cpu id stack addr total size used size\n" " ---------- ------ --------- -------- --------\n"); +<<<<<<< HEAD for (index = 0; index < g_stackNum; index++) { for (cpuid = 0; cpuid < LOSCFG_KERNEL_CORE_NUM; cpuid++) { stackTop = (UINTPTR *)((UINTPTR)g_stackInfo[index].stackTop + cpuid * g_stackInfo[index].stackSize); stack = (UINTPTR *)((UINTPTR)stackTop + g_stackInfo[index].stackSize); (VOID)OsStackWaterLineGet(stack, stackTop, &size); +======= + for (index = 0; index < g_stackNum; index++) { + for (cpuid = 0; cpuid < LOSCFG_KERNEL_CORE_NUM; cpuid++) {//可以看出 各个CPU的栈是紧挨的的 + stackTop = (UINTPTR *)((UINTPTR)g_stackInfo[index].stackTop + cpuid * g_stackInfo[index].stackSize); + stack = (UINTPTR *)((UINTPTR)stackTop + g_stackInfo[index].stackSize); + (VOID)OsStackWaterLineGet(stack, stackTop, &size);//获取吃水线, 鸿蒙用WaterLine 这个词用的很妙 +>>>>>>> remotes/origin/main PrintExcInfo("%11s %-5d %-10p 0x%-8x 0x%-4x\n", g_stackInfo[index].stackName, LOSCFG_KERNEL_CORE_NUM - 1 - cpuid, stackTop, g_stackInfo[index].stackSize, size); } } +<<<<<<< HEAD OsExcStackCheck(); } @@ -120,9 +186,32 @@ VOID OsStackInit(VOID *stacktop, UINT32 stacksize) errno_t ret = memset_s(stacktop, stacksize, (INT32)OS_STACK_INIT, stacksize); if (ret == EOK) { *((UINTPTR *)stacktop) = OS_STACK_MAGIC_WORD; +======= + OsExcStackCheck();//发生异常时栈检查 +} + +///注册栈信息 +VOID OsExcStackInfoReg(const StackInfo *stackInfo, UINT32 stackNum) +{ + g_stackInfo = stackInfo; //全局变量指向g_excStack + g_stackNum = stackNum; +} + +///task栈的初始化,设置固定的值. 0xcccccccc 和 0xcacacaca +VOID OsStackInit(VOID *stacktop, UINT32 stacksize) +{ + /* initialize the task stack, write magic num to stack top */ + errno_t ret = memset_s(stacktop, stacksize, (INT32)OS_STACK_INIT, stacksize);//清一色填 0xCACACACA + if (ret == EOK) { + *((UINTPTR *)stacktop) = OS_STACK_MAGIC_WORD;//0xCCCCCCCCC 中文就是"烫烫烫烫" 这几个字懂点计算机的人都不会陌生了. +>>>>>>> remotes/origin/main } } #ifdef LOSCFG_SHELL_CMD_DEBUG +<<<<<<< HEAD SHELLCMD_ENTRY(stack_shellcmd, CMD_TYPE_EX, "stack", 1, (CmdCallBackFunc)OsExcStackInfo); +======= +SHELLCMD_ENTRY(stack_shellcmd, CMD_TYPE_EX, "stack", 1, (CmdCallBackFunc)OsExcStackInfo);//采用shell命令静态注册方式 +>>>>>>> remotes/origin/main #endif diff --git a/src/kernel_liteos_a/kernel/base/misc/mempt_shellcmd.c b/src/kernel_liteos_a/kernel/base/misc/mempt_shellcmd.c index 0e8ec47b..113cac35 100644 --- a/src/kernel_liteos_a/kernel/base/misc/mempt_shellcmd.c +++ b/src/kernel_liteos_a/kernel/base/misc/mempt_shellcmd.c @@ -193,7 +193,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdUname(INT32 argc, const CHAR *argv[]) if (argc == 1) { if (strcmp(argv[0], "-a") == 0) { +<<<<<<< HEAD PRINTK("%s %d.%d.%d.%d %s %s\n", KERNEL_NAME, KERNEL_MAJOR, KERNEL_MINOR, KERNEL_PATCH, KERNEL_ITRE, \ +======= + PRINTK("%s %d.%d.%d.%d %s %s\n", KERNEL_NAME, KERNEL_MAJOR, KERNEL_MINOR, KERNEL_PATCH, KERNEL_ITRE,\ +>>>>>>> remotes/origin/main __DATE__, __TIME__); return 0; } else if (strcmp(argv[0], "-s") == 0) { diff --git a/src/kernel_liteos_a/kernel/base/misc/swtmr_shellcmd.c b/src/kernel_liteos_a/kernel/base/misc/swtmr_shellcmd.c index 81573ef4..e6814e45 100644 --- a/src/kernel_liteos_a/kernel/base/misc/swtmr_shellcmd.c +++ b/src/kernel_liteos_a/kernel/base/misc/swtmr_shellcmd.c @@ -58,6 +58,7 @@ STATIC VOID OsPrintSwtmrMsg(const SWTMR_CTRL_S *swtmr) (VOID)LOS_SwtmrTimeGet(swtmr->usTimerID, &ticks); PRINTK("%7u%10s%8s%12u%7u%#12x%#12x\n", +<<<<<<< HEAD swtmr->usTimerID % LOSCFG_BASE_CORE_SWTMR_LIMIT, g_shellSwtmrStatus[swtmr->ucState], g_shellSwtmrMode[swtmr->ucMode], @@ -65,13 +66,27 @@ STATIC VOID OsPrintSwtmrMsg(const SWTMR_CTRL_S *swtmr) ticks, swtmr->uwArg, swtmr->pfnHandler); +======= + swtmr->usTimerID % LOSCFG_BASE_CORE_SWTMR_LIMIT, //软件定时器ID。 + g_shellSwtmrStatus[swtmr->ucState], //软件定时器状态,状态可能为:"UnUsed", "Created", "Ticking"。 + g_shellSwtmrMode[swtmr->ucMode], //软件定时器模式。模式可能为:"Once", "Period", "NSD(单次定时器,定时结束后不会自动删除)" + swtmr->uwInterval, //软件定时器使用的Tick数。 + ticks, + swtmr->uwArg, //传入的参数。 + swtmr->pfnHandler); //回调函数的地址。 +>>>>>>> remotes/origin/main } STATIC INLINE VOID OsPrintSwtmrMsgHead(VOID) { PRINTK("\r\nSwTmrID State Mode Interval Count Arg handlerAddr\n"); } +<<<<<<< HEAD +======= +///shell命令之swtmr 命令用于查询系统软件定时器相关信息。 +//参数缺省时,默认显示所有软件定时器的相关信息。 +>>>>>>> remotes/origin/main STATIC UINT32 SwtmrBaseInfoGet(UINT32 timerID) { SWTMR_CTRL_S *swtmr = g_swtmrCBArray; @@ -173,7 +188,11 @@ SWTMR_HELP: PRINTK(" swtmr ID --- Specifies information about a software timer.\n"); return LOS_OK; } +<<<<<<< HEAD SHELLCMD_ENTRY(swtmr_shellcmd, CMD_TYPE_EX, "swtmr", 1, (CmdCallBackFunc)OsShellCmdSwtmrInfoGet); +======= +SHELLCMD_ENTRY(swtmr_shellcmd, CMD_TYPE_EX, "swtmr", 1, (CmdCallBackFunc)OsShellCmdSwtmrInfoGet);//采用shell命令静态注册方式 +>>>>>>> remotes/origin/main #endif /* LOSCFG_SHELL */ diff --git a/src/kernel_liteos_a/kernel/base/misc/sysinfo_shellcmd.c b/src/kernel_liteos_a/kernel/base/misc/sysinfo_shellcmd.c index b7c56f84..70b30ed1 100644 --- a/src/kernel_liteos_a/kernel/base/misc/sysinfo_shellcmd.c +++ b/src/kernel_liteos_a/kernel/base/misc/sysinfo_shellcmd.c @@ -118,7 +118,11 @@ UINT32 OsShellCmdSwtmrCntGet(VOID) LOS_IntRestore(intSave); return swtmrCnt; } +<<<<<<< HEAD +======= +///查看系统资源使用情况 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsShellCmdSystemInfoGet(VOID) { UINT8 isTaskEnable = TRUE; @@ -137,6 +141,7 @@ LITE_OS_SEC_TEXT_MINOR VOID OsShellCmdSystemInfoGet(VOID) #else UINT8 isSwtmrEnable = FALSE; #endif +<<<<<<< HEAD PRINTK("\n Module Used Total Enabled\n"); PRINTK("--------------------------------------------\n"); @@ -158,6 +163,29 @@ LITE_OS_SEC_TEXT_MINOR VOID OsShellCmdSystemInfoGet(VOID) SYSINFO_ENABLED(isSwtmrEnable)); } +======= +//模块名称 当前使用量 最大可用量 模块是否开启 + PRINTK("\n Module Used Total Enabled\n"); + PRINTK("--------------------------------------------\n"); + PRINTK(" Task %-10u%-10d%s\n", + OsShellCmdTaskCntGet(), //有效任务数 + LOSCFG_BASE_CORE_TSK_LIMIT, //任务最大数 128 + SYSINFO_ENABLED(isTaskEnable));//任务是否失效 YES or NO + PRINTK(" Sem %-10u%-10d%s\n", + OsShellCmdSemCntGet(), //信号量的数量 + LOSCFG_BASE_IPC_SEM_LIMIT, //信号量最大数 1024 + SYSINFO_ENABLED(isSemEnable));//信号量是否失效 YES or NO + PRINTK(" Queue %-10u%-10d%s\n", + OsShellCmdQueueCntGet(), //队列的数量 + LOSCFG_BASE_IPC_QUEUE_LIMIT, //队列的最大数 1024 + SYSINFO_ENABLED(isQueueEnable));//队列是否失效 YES or NO + PRINTK(" SwTmr %-10u%-10d%s\n", + OsShellCmdSwtmrCntGet(), //定时器的数量 + LOSCFG_BASE_CORE_SWTMR_LIMIT, //定时器的总数 1024 + SYSINFO_ENABLED(isSwtmrEnable)); //定时器是否失效 YES or NO +} +///systeminfo命令用于显示当前操作系统内资源使用情况,包括任务、信号量、互斥量、队列、定时器等。 +>>>>>>> remotes/origin/main INT32 OsShellCmdSystemInfo(INT32 argc, const CHAR **argv) { if (argc == 0) { diff --git a/src/kernel_liteos_a/kernel/base/misc/vm_shellcmd.c b/src/kernel_liteos_a/kernel/base/misc/vm_shellcmd.c index 445014b3..c55e8610 100644 --- a/src/kernel_liteos_a/kernel/base/misc/vm_shellcmd.c +++ b/src/kernel_liteos_a/kernel/base/misc/vm_shellcmd.c @@ -54,7 +54,11 @@ #define VMM_CMD "vmm" #define OOM_CMD "oom" #define VMM_PMM_CMD "v2p" +<<<<<<< HEAD +======= +//dump内核空间 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OsDumpKernelAspace(VOID) { LosVmSpace *kAspace = LOS_GetKVmSpace(); @@ -104,6 +108,7 @@ LITE_OS_SEC_TEXT_MINOR VOID OsDoDumpVm(pid_t pid) PRINTK("\tThe process [%d] not active\n", pid); } } +<<<<<<< HEAD LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdDumpVm(INT32 argc, const CHAR *argv[]) { @@ -124,6 +129,28 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdDumpVm(INT32 argc, const CHAR *argv[]) OsPrintUsage(); } } else { +======= +///查看进程的虚拟内存使用情况。vmm [-a / -h / --help], vmm [pid] +LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdDumpVm(INT32 argc, const CHAR *argv[]) +{ + if (argc == 0) { //没有参数 使用 # vmm 查看所有进程使用虚拟内存的情况 + OsDumpAllAspace(); + } else if (argc == 1) { + pid_t pid = OsPid(argv[0]); + if (strcmp(argv[0], "-a") == 0) { //# vmm -a 查看所有进程使用虚拟内存的情况 + OsDumpAllAspace(); + } else if (strcmp(argv[0], "-k") == 0) {//# vmm -k 查看内核进程使用虚拟内存的情况 + OsDumpKernelAspace(); + } else if (pid >= 0) { //# vmm 3 查看3号进程使用虚拟内存的情况 + OsDoDumpVm(pid); + } else if (strcmp(argv[0], "-h") == 0 || strcmp(argv[0], "--help") == 0) { //# vmm -h 或者 vmm --help + OsPrintUsage(); + } else { + PRINTK("%s: invalid option: %s\n", VMM_CMD, argv[0]); //格式错误,输出规范格式 + OsPrintUsage(); + } + } else { //多于一个参数 例如 # vmm 3 9 +>>>>>>> remotes/origin/main OsPrintUsage(); } @@ -135,7 +162,11 @@ LITE_OS_SEC_TEXT_MINOR VOID V2PPrintUsage(VOID) PRINTK("pid vaddr(0x1000000~0x3e000000), print physical address of virtual address\n" "-h | --help, print v2p command usage\n"); } +<<<<<<< HEAD +======= +///v2p 虚拟内存对应的物理内存 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdV2P(INT32 argc, const CHAR *argv[]) { UINT32 vaddr; @@ -180,7 +211,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdV2P(INT32 argc, const CHAR *argv[]) return LOS_OK; } +<<<<<<< HEAD +======= +///查看系统内存物理页及pagecache物理页使用情况 , Debug版本才具备的命令 # pmm +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdDumpPmm(VOID) { OsVmPhysDump(); @@ -192,12 +227,22 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdDumpPmm(VOID) LITE_OS_SEC_TEXT_MINOR VOID OomPrintUsage(VOID) { +<<<<<<< HEAD PRINTK("\t-i [interval], set oom check interval (ms)\n" "\t-m [mem byte], set oom low memory threshold (Byte)\n" "\t-r [mem byte], set page cache reclaim memory threshold (Byte)\n" "\t-h | --help, print vmm command usage\n"); } +======= + PRINTK("\t-i [interval], set oom check interval (ms)\n" //设置oom线程任务检查的时间间隔。 + "\t-m [mem byte], set oom low memory threshold (Byte)\n" //设置低内存阈值。 + "\t-r [mem byte], set page cache reclaim memory threshold (Byte)\n" //设置pagecache内存回收阈值。 + "\t-h | --help, print vmm command usage\n"); //使用帮助。 +} +///查看和设置低内存阈值以及pagecache内存回收阈值。参数缺省时,显示oom功能当前配置信息。 +//当系统内存不足时,会打印出内存不足的提示信息。 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdOom(INT32 argc, const CHAR *argv[]) { UINT32 lowMemThreshold; @@ -219,7 +264,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdOom(INT32 argc, const CHAR *argv[]) PRINTK("[oom] low mem threshold %s(byte) invalid.\n", argv[1]); return OS_ERROR; } else { +<<<<<<< HEAD OomSetLowMemThreashold(lowMemThreshold); +======= + OomSetLowMemThreashold(lowMemThreshold);//设置低内存阈值 +>>>>>>> remotes/origin/main } } else if (strcmp(argv[0], "-i") == 0) { checkInterval = strtoul((CHAR *)argv[1], &endPtr, 0); @@ -227,7 +276,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdOom(INT32 argc, const CHAR *argv[]) PRINTK("[oom] check interval %s(us) invalid.\n", argv[1]); return OS_ERROR; } else { +<<<<<<< HEAD OomSetCheckInterval(checkInterval); +======= + OomSetCheckInterval(checkInterval);//设置oom线程任务检查的时间间隔 +>>>>>>> remotes/origin/main } } else if (strcmp(argv[0], "-r") == 0) { reclaimMemThreshold = strtoul((CHAR *)argv[1], &endPtr, 0); @@ -235,7 +288,11 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdOom(INT32 argc, const CHAR *argv[]) PRINTK("[oom] reclaim mem threshold %s(byte) invalid.\n", argv[1]); return OS_ERROR; } else { +<<<<<<< HEAD OomSetReclaimMemThreashold(reclaimMemThreshold); +======= + OomSetReclaimMemThreashold(reclaimMemThreshold);//设置pagecache内存回收阈值 +>>>>>>> remotes/origin/main } } else { PRINTK("%s: invalid option: %s %s\n", OOM_CMD, argv[0], argv[1]); @@ -250,6 +307,7 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OsShellCmdOom(INT32 argc, const CHAR *argv[]) } #ifdef LOSCFG_SHELL_CMD_DEBUG +<<<<<<< HEAD SHELLCMD_ENTRY(oom_shellcmd, CMD_TYPE_SHOW, OOM_CMD, 2, (CmdCallBackFunc)OsShellCmdOom); SHELLCMD_ENTRY(vm_shellcmd, CMD_TYPE_SHOW, VMM_CMD, 1, (CmdCallBackFunc)OsShellCmdDumpVm); SHELLCMD_ENTRY(v2p_shellcmd, CMD_TYPE_SHOW, VMM_PMM_CMD, 1, (CmdCallBackFunc)OsShellCmdV2P); @@ -257,6 +315,15 @@ SHELLCMD_ENTRY(v2p_shellcmd, CMD_TYPE_SHOW, VMM_PMM_CMD, 1, (CmdCallBackFunc)OsS #ifdef LOSCFG_SHELL SHELLCMD_ENTRY(pmm_shellcmd, CMD_TYPE_SHOW, "pmm", 0, (CmdCallBackFunc)OsShellCmdDumpPmm); +======= +SHELLCMD_ENTRY(oom_shellcmd, CMD_TYPE_SHOW, OOM_CMD, 2, (CmdCallBackFunc)OsShellCmdOom);//采用shell命令静态注册方式 +SHELLCMD_ENTRY(vm_shellcmd, CMD_TYPE_SHOW, VMM_CMD, 1, (CmdCallBackFunc)OsShellCmdDumpVm);//采用shell命令静态注册方式 vmm +SHELLCMD_ENTRY(v2p_shellcmd, CMD_TYPE_SHOW, VMM_PMM_CMD, 1, (CmdCallBackFunc)OsShellCmdV2P);//采用shell命令静态注册方式 v2p +#endif + +#ifdef LOSCFG_SHELL +SHELLCMD_ENTRY(pmm_shellcmd, CMD_TYPE_SHOW, "pmm", 0, (CmdCallBackFunc)OsShellCmdDumpPmm);//采用shell命令静态注册方式 +>>>>>>> remotes/origin/main #endif #endif diff --git a/src/kernel_liteos_a/kernel/base/mp/los_mp.c b/src/kernel_liteos_a/kernel/base/mp/los_mp.c index 67bc7012..cfaf4331 100644 --- a/src/kernel_liteos_a/kernel/base/mp/los_mp.c +++ b/src/kernel_liteos_a/kernel/base/mp/los_mp.c @@ -36,13 +36,20 @@ #include "los_swtmr.h" #include "los_task_pri.h" +<<<<<<< HEAD #ifdef LOSCFG_KERNEL_SMP +======= + +#ifdef LOSCFG_KERNEL_SMP +//给参数CPU发送调度信号 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_SMP_CALL LITE_OS_SEC_BSS SPIN_LOCK_INIT(g_mpCallSpin); #define MP_CALL_LOCK(state) LOS_SpinLockSave(&g_mpCallSpin, &(state)) #define MP_CALL_UNLOCK(state) LOS_SpinUnlockRestore(&g_mpCallSpin, (state)) #endif +<<<<<<< HEAD VOID LOS_MpSchedule(UINT32 target) { @@ -51,19 +58,35 @@ VOID LOS_MpSchedule(UINT32 target) HalIrqSendIpi(target, LOS_MP_IPI_SCHEDULE); } +======= +VOID LOS_MpSchedule(UINT32 target)//target每位对应CPU core +{ + UINT32 cpuid = ArchCurrCpuid(); + target &= ~(1U << cpuid);//获取除了自身之外的其他CPU + HalIrqSendIpi(target, LOS_MP_IPI_SCHEDULE);//向目标CPU发送调度信号,核间中断(Inter-Processor Interrupts),IPI +} +///硬中断唤醒处理函数 +>>>>>>> remotes/origin/main VOID OsMpWakeHandler(VOID) { /* generic wakeup ipi, do nothing */ } +<<<<<<< HEAD VOID OsMpScheduleHandler(VOID) { +======= +///硬中断调度处理函数 +VOID OsMpScheduleHandler(VOID) +{//将调度标志设置为与唤醒功能不同,这样就可以在硬中断结束时触发调度程序。 +>>>>>>> remotes/origin/main /* * set schedule flag to differ from wake function, * so that the scheduler can be triggered at the end of irq. */ OsSchedRunqueuePendingSet(); } +<<<<<<< HEAD VOID OsMpHaltHandler(VOID) { @@ -73,6 +96,17 @@ VOID OsMpHaltHandler(VOID) while (1) {} } +======= +///硬中断暂停处理函数 +VOID OsMpHaltHandler(VOID) +{ + (VOID)LOS_IntLock(); + OsPercpuGet()->excFlag = CPU_HALT;//让当前Cpu停止工作 + + while (1) {}//陷入空循环,也就是空闲状态 +} +///MP定时器处理函数, 递归检查所有可用任务 +>>>>>>> remotes/origin/main VOID OsMpCollectTasks(VOID) { LosTaskCB *taskCB = NULL; @@ -80,19 +114,32 @@ VOID OsMpCollectTasks(VOID) UINT32 ret; /* recursive checking all the available task */ +<<<<<<< HEAD for (; taskID <= g_taskMaxNum; taskID++) { +======= + for (; taskID <= g_taskMaxNum; taskID++) { //递归检查所有可用任务 +>>>>>>> remotes/origin/main taskCB = &g_taskCBArray[taskID]; if (OsTaskIsUnused(taskCB) || OsTaskIsRunning(taskCB)) { continue; } +<<<<<<< HEAD /* * though task status is not atomic, this check may success but not accomplish * the deletion; this deletion will be handled until the next run. */ if (taskCB->signal & SIGNAL_KILL) { ret = LOS_TaskDelete(taskID); +======= + /* 虽然任务状态不是原子的,但此检查可能成功,但无法完成删除,此删除将在下次运行之前处理 + * though task status is not atomic, this check may success but not accomplish + * the deletion; this deletion will be handled until the next run. + */ + if (taskCB->signal & SIGNAL_KILL) {//任务收到被干掉信号 + ret = LOS_TaskDelete(taskID);//干掉任务,回归任务池 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { PRINT_WARN("GC collect task failed err:0x%x\n", ret); } @@ -101,6 +148,10 @@ VOID OsMpCollectTasks(VOID) } #ifdef LOSCFG_KERNEL_SMP_CALL +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main VOID OsMpFuncCall(UINT32 target, SMP_FUNC_CALL func, VOID *args) { UINT32 index; @@ -110,6 +161,7 @@ VOID OsMpFuncCall(UINT32 target, SMP_FUNC_CALL func, VOID *args) return; } +<<<<<<< HEAD if (!(target & OS_MP_CPU_ALL)) { return; } @@ -117,6 +169,15 @@ VOID OsMpFuncCall(UINT32 target, SMP_FUNC_CALL func, VOID *args) for (index = 0; index < LOSCFG_KERNEL_CORE_NUM; index++) { if (CPUID_TO_AFFI_MASK(index) & target) { MpCallFunc *mpCallFunc = (MpCallFunc *)LOS_MemAlloc(m_aucSysMem0, sizeof(MpCallFunc)); +======= + if (!(target & OS_MP_CPU_ALL)) {//检查目标CPU是否正确 + return; + } + + for (index = 0; index < LOSCFG_KERNEL_CORE_NUM; index++) {//遍历所有核 + if (CPUID_TO_AFFI_MASK(index) & target) { + MpCallFunc *mpCallFunc = (MpCallFunc *)LOS_MemAlloc(m_aucSysMem0, sizeof(MpCallFunc));//从内核空间 分配回调结构体 +>>>>>>> remotes/origin/main if (mpCallFunc == NULL) { PRINT_ERR("smp func call malloc failed\n"); return; @@ -125,6 +186,7 @@ VOID OsMpFuncCall(UINT32 target, SMP_FUNC_CALL func, VOID *args) mpCallFunc->args = args; MP_CALL_LOCK(intSave); +<<<<<<< HEAD LOS_ListAdd(&g_percpu[index].funcLink, &(mpCallFunc->node)); MP_CALL_UNLOCK(intSave); } @@ -136,10 +198,31 @@ VOID OsMpFuncCallHandler(VOID) { UINT32 intSave; UINT32 cpuid = ArchCurrCpuid(); +======= + LOS_ListAdd(&g_percpu[index].funcLink, &(mpCallFunc->node));//将回调结构体挂入链表尾部 + MP_CALL_UNLOCK(intSave); + } + } + HalIrqSendIpi(target, LOS_MP_IPI_FUNC_CALL);//向目标CPU发起核间中断 +} + +/*! + * @brief OsMpFuncCallHandler + * 回调向当前CPU注册过的函数 + * @return + * + * @see + */ +VOID OsMpFuncCallHandler(VOID) +{ + UINT32 intSave; + UINT32 cpuid = ArchCurrCpuid();//获取当前CPU +>>>>>>> remotes/origin/main LOS_DL_LIST *list = NULL; MpCallFunc *mpCallFunc = NULL; MP_CALL_LOCK(intSave); +<<<<<<< HEAD while (!LOS_ListEmpty(&g_percpu[cpuid].funcLink)) { list = LOS_DL_LIST_FIRST(&g_percpu[cpuid].funcLink); LOS_ListDelete(list); @@ -148,11 +231,22 @@ VOID OsMpFuncCallHandler(VOID) mpCallFunc = LOS_DL_LIST_ENTRY(list, MpCallFunc, node); mpCallFunc->func(mpCallFunc->args); (VOID)LOS_MemFree(m_aucSysMem0, mpCallFunc); +======= + while (!LOS_ListEmpty(&g_percpu[cpuid].funcLink)) {//遍历回调函数链表,知道为空 + list = LOS_DL_LIST_FIRST(&g_percpu[cpuid].funcLink);//获取链表第一个数据 + LOS_ListDelete(list);//将自己从链表上摘除 + MP_CALL_UNLOCK(intSave); + + mpCallFunc = LOS_DL_LIST_ENTRY(list, MpCallFunc, node);//获取回调函数 + mpCallFunc->func(mpCallFunc->args);//获取参数并回调该函数 + (VOID)LOS_MemFree(m_aucSysMem0, mpCallFunc);//释放回调函数内存 +>>>>>>> remotes/origin/main MP_CALL_LOCK(intSave); } MP_CALL_UNLOCK(intSave); } +<<<<<<< HEAD VOID OsMpFuncCallInit(VOID) { @@ -164,20 +258,43 @@ VOID OsMpFuncCallInit(VOID) } #endif /* LOSCFG_KERNEL_SMP_CALL */ +======= +/// CPU层级的回调模块初始化 +VOID OsMpFuncCallInit(VOID) +{ + UINT32 index; + /* init funclink for each core | 为每个CPU核整一个回调函数链表*/ + for (index = 0; index < LOSCFG_KERNEL_CORE_NUM; index++) { + LOS_ListInit(&g_percpu[index].funcLink);//链表初始化 + } +} +#endif /* LOSCFG_KERNEL_SMP_CALL */ +//MP(multiprocessing) 多核处理器初始化 +>>>>>>> remotes/origin/main UINT32 OsMpInit(VOID) { UINT16 swtmrId; +<<<<<<< HEAD (VOID)LOS_SwtmrCreate(OS_MP_GC_PERIOD, LOS_SWTMR_MODE_PERIOD, (SWTMR_PROC_FUNC)OsMpCollectTasks, &swtmrId, 0); (VOID)LOS_SwtmrStart(swtmrId); +======= + (VOID)LOS_SwtmrCreate(OS_MP_GC_PERIOD, LOS_SWTMR_MODE_PERIOD, //创建一个周期性,持续时间为 100个tick的定时器 + (SWTMR_PROC_FUNC)OsMpCollectTasks, &swtmrId, 0);//OsMpCollectTasks为超时回调函数 + (VOID)LOS_SwtmrStart(swtmrId);//开始定时任务 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_SMP_CALL OsMpFuncCallInit(); #endif return LOS_OK; } +<<<<<<< HEAD LOS_MODULE_INIT(OsMpInit, LOS_INIT_LEVEL_KMOD_TASK); +======= +LOS_MODULE_INIT(OsMpInit, LOS_INIT_LEVEL_KMOD_TASK);//多处理器模块初始化 +>>>>>>> remotes/origin/main #endif diff --git a/src/kernel_liteos_a/kernel/base/mp/los_percpu.c b/src/kernel_liteos_a/kernel/base/mp/los_percpu.c index 4d883916..241a8993 100644 --- a/src/kernel_liteos_a/kernel/base/mp/los_percpu.c +++ b/src/kernel_liteos_a/kernel/base/mp/los_percpu.c @@ -33,7 +33,11 @@ #include "los_printf.h" #ifdef LOSCFG_KERNEL_SMP +<<<<<<< HEAD Percpu g_percpu[LOSCFG_KERNEL_CORE_NUM]; +======= +Percpu g_percpu[LOSCFG_KERNEL_CORE_NUM]; ///< CPU池,池大小由CPU核数决定 +>>>>>>> remotes/origin/main VOID OsAllCpuStatusOutput(VOID) { diff --git a/src/kernel_liteos_a/kernel/base/om/los_err.c b/src/kernel_liteos_a/kernel/base/om/los_err.c index 6267a9d1..3ce5b013 100644 --- a/src/kernel_liteos_a/kernel/base/om/los_err.c +++ b/src/kernel_liteos_a/kernel/base/om/los_err.c @@ -31,9 +31,35 @@ #include "los_err.h" +<<<<<<< HEAD LITE_OS_SEC_BSS STATIC LOS_ERRORHANDLE_FUNC g_errHandleHook = NULL; +======= +/** +基本概念 + 错误处理指程序运行错误时,调用错误处理模块的接口函数,上报错误信息,并调用注册的钩子函数 + 进行特定处理,保存现场以便定位问题。通过错误处理,可以控制和提示程序中的非法输入,防止程序崩溃。 + +运作机制 + 错误处理是一种机制,用于处理异常状况。当程序出现错误时,会显示相应的错误码。 + 此外,如果注册了相应的错误处理函数,则会执行这个函数。 + +系统中只有一个错误处理的钩子函数。当多次注册钩子函数时,最后一次注册的钩子函数会覆盖前一次注册的函数 + */ + +LITE_OS_SEC_BSS STATIC LOS_ERRORHANDLE_FUNC g_errHandleHook = NULL;///< 错误接管钩子函数 + +/** + * @brief 调用钩子函数,处理错误 + * @param fileName 存放错误日志的文件名,系统内部调用时,入参为"os_unspecific_file" + * @param lineNo 发生错误的代码行号系统内部调用时,若值为0xa1b2c3f8,表示未传递行号 + * @param errorNo 错误码 + * @param paraLen 入参para的长度系统内部调用时,入参为0 + * @param para 错误标签系统内部调用时,入参为NULL + * @return LITE_OS_SEC_TEXT_INIT + */ +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT UINT32 LOS_ErrHandle(CHAR *fileName, UINT32 lineNo, UINT32 errorNo, UINT32 paraLen, VOID *para) { @@ -43,7 +69,11 @@ LITE_OS_SEC_TEXT_INIT UINT32 LOS_ErrHandle(CHAR *fileName, UINT32 lineNo, UINT32 return LOS_OK; } +<<<<<<< HEAD +======= +///设置钩子函数,处理错误 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_INIT VOID LOS_SetErrHandleHook(LOS_ERRORHANDLE_FUNC fun) { g_errHandleHook = fun; diff --git a/src/kernel_liteos_a/kernel/base/sched/los_idle.c b/src/kernel_liteos_a/kernel/base/sched/los_idle.c index d2538595..da114c76 100644 --- a/src/kernel_liteos_a/kernel/base/sched/los_idle.c +++ b/src/kernel_liteos_a/kernel/base/sched/los_idle.c @@ -44,7 +44,11 @@ STATIC VOID IdleTimeSliceUpdate(SchedRunqueue *rq, LosTaskCB *taskCB, UINT64 cur STATIC INT32 IdleParamCompare(const SchedPolicy *sp1, const SchedPolicy *sp2); STATIC VOID IdlePriorityInheritance(LosTaskCB *owner, const SchedParam *param); STATIC VOID IdlePriorityRestore(LosTaskCB *owner, const LOS_DL_LIST *list, const SchedParam *param); +<<<<<<< HEAD +======= +//空闲调度 +>>>>>>> remotes/origin/main const STATIC SchedOps g_idleOps = { .dequeue = IdleDequeue, .enqueue = IdleEnqueue, diff --git a/src/kernel_liteos_a/kernel/base/sched/los_priority.c b/src/kernel_liteos_a/kernel/base/sched/los_priority.c index 610ba119..8bef4dc4 100644 --- a/src/kernel_liteos_a/kernel/base/sched/los_priority.c +++ b/src/kernel_liteos_a/kernel/base/sched/los_priority.c @@ -1,5 +1,9 @@ /* +<<<<<<< HEAD * Copyright (c) 2022-2023 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2022-2022 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -43,6 +47,11 @@ #define OS_SCHED_READY_MAX 30 #define OS_TIME_SLICE_MIN (INT32)((50 * OS_SYS_NS_PER_US) / OS_NS_PER_CYCLE) /* 50us */ +<<<<<<< HEAD +======= +//基于优先数调度算法 Highest-Priority-First (HPF) + +>>>>>>> remotes/origin/main STATIC HPFRunqueue g_schedHPF; STATIC VOID HPFDequeue(SchedRunqueue *rq, LosTaskCB *taskCB); @@ -63,7 +72,11 @@ STATIC VOID HPFTimeSliceUpdate(SchedRunqueue *rq, LosTaskCB *taskCB, UINT64 curr STATIC INT32 HPFParamCompare(const SchedPolicy *sp1, const SchedPolicy *sp2); STATIC VOID HPFPriorityInheritance(LosTaskCB *owner, const SchedParam *param); STATIC VOID HPFPriorityRestore(LosTaskCB *owner, const LOS_DL_LIST *list, const SchedParam *param); +<<<<<<< HEAD +======= +//优先级调度算法操作 +>>>>>>> remotes/origin/main const STATIC SchedOps g_priorityOps = { .dequeue = HPFDequeue, .enqueue = HPFEnqueue, @@ -243,7 +256,11 @@ STATIC INLINE VOID PriQueInsert(HPFRunqueue *rq, LosTaskCB *taskCB) taskCB->taskStatus &= ~OS_TASK_STATUS_BLOCKED; taskCB->taskStatus |= OS_TASK_STATUS_READY; } +<<<<<<< HEAD +======= +//入就绪队列 +>>>>>>> remotes/origin/main STATIC VOID HPFEnqueue(SchedRunqueue *rq, LosTaskCB *taskCB) { #ifdef LOSCFG_SCHED_HPF_DEBUG @@ -253,14 +270,24 @@ STATIC VOID HPFEnqueue(SchedRunqueue *rq, LosTaskCB *taskCB) #endif PriQueInsert(rq->hpfRunqueue, taskCB); } +<<<<<<< HEAD +======= +//出就绪队列 +>>>>>>> remotes/origin/main STATIC VOID HPFDequeue(SchedRunqueue *rq, LosTaskCB *taskCB) { SchedHPF *sched = (SchedHPF *)&taskCB->sp; +<<<<<<< HEAD if (taskCB->taskStatus & OS_TASK_STATUS_READY) { PriQueDelete(rq->hpfRunqueue, sched->basePrio, &taskCB->pendList, sched->priority); taskCB->taskStatus &= ~OS_TASK_STATUS_READY; +======= + if (taskCB->taskStatus & OS_TASK_STATUS_READY) {//是否有就绪状态 + PriQueDelete(rq->hpfRunqueue, sched->basePrio, &taskCB->pendList, sched->priority); + taskCB->taskStatus &= ~OS_TASK_STATUS_READY;//更新成非就绪状态 +>>>>>>> remotes/origin/main } } @@ -475,7 +502,11 @@ STATIC VOID HPFPriorityInheritance(LosTaskCB *owner, const SchedParam *param) LOS_BitmapSet(&sp->priBitmap, sp->priority); sp->priority = param->priority; } +<<<<<<< HEAD +======= +/// 恢复任务优先级 +>>>>>>> remotes/origin/main STATIC VOID HPFPriorityRestore(LosTaskCB *owner, const LOS_DL_LIST *list, const SchedParam *param) { UINT16 priority; @@ -498,8 +529,13 @@ STATIC VOID HPFPriorityRestore(LosTaskCB *owner, const LOS_DL_LIST *list, const } if ((list != NULL) && !LOS_ListEmpty((LOS_DL_LIST *)list)) { +<<<<<<< HEAD priority = LOS_HighBitGet(sp->priBitmap); LOS_DL_LIST_FOR_EACH_ENTRY(pendedTask, list, LosTaskCB, pendList) { +======= + priority = LOS_HighBitGet(sp->priBitmap);//获取在历史调度中最高优先级 + LOS_DL_LIST_FOR_EACH_ENTRY(pendedTask, list, LosTaskCB, pendList) {//遍历链表 +>>>>>>> remotes/origin/main SchedHPF *pendSp = (SchedHPF *)&pendedTask->sp; if ((pendedTask->ops == owner->ops) && (priority != pendSp->priority)) { LOS_BitmapClr(&sp->priBitmap, pendSp->priority); @@ -537,7 +573,11 @@ VOID HPFProcessDefaultSchedParamGet(SchedParam *param) { param->basePrio = OS_USER_PROCESS_PRIORITY_HIGHEST; } +<<<<<<< HEAD +======= +//HPF 调度策略初始化 +>>>>>>> remotes/origin/main VOID HPFSchedPolicyInit(SchedRunqueue *rq) { if (ArchCurrCpuid() > 0) { diff --git a/src/kernel_liteos_a/kernel/base/sched/los_sortlink.c b/src/kernel_liteos_a/kernel/base/sched/los_sortlink.c index 4f68436f..2eca1624 100644 --- a/src/kernel_liteos_a/kernel/base/sched/los_sortlink.c +++ b/src/kernel_liteos_a/kernel/base/sched/los_sortlink.c @@ -1,6 +1,10 @@ /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. +<<<<<<< HEAD * Copyright (c) 2020-2022 Huawei Device Co., Ltd. All rights reserved. +======= + * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. +>>>>>>> remotes/origin/main * * Redistribution and use in source and binary forms, with or without modification, * are permitted provided that the following conditions are met: @@ -30,7 +34,11 @@ */ #include "los_sortlink_pri.h" +<<<<<<< HEAD +======= +/// 排序链表初始化 +>>>>>>> remotes/origin/main VOID OsSortLinkInit(SortLinkAttribute *sortLinkHeader) { LOS_ListInit(&sortLinkHeader->sortLink); @@ -38,6 +46,7 @@ VOID OsSortLinkInit(SortLinkAttribute *sortLinkHeader) sortLinkHeader->nodeNum = 0; } +<<<<<<< HEAD STATIC INLINE VOID AddNode2SortLink(SortLinkAttribute *sortLinkHeader, SortLinkList *sortList) { LOS_DL_LIST *head = (LOS_DL_LIST *)&sortLinkHeader->sortLink; @@ -63,13 +72,54 @@ STATIC INLINE VOID AddNode2SortLink(SortLinkAttribute *sortLinkHeader, SortLinkL do { listSorted = LOS_DL_LIST_ENTRY(prevNode, SortLinkList, sortLinkNode); if (listSorted->responseTime <= sortList->responseTime) { +======= +/*! + * @brief OsAddNode2SortLink 向链表中插入结点,并按时间顺序排列 + * + * @param sortLinkHeader 被插入的链表 + * @param sortList 要插入的结点 + * @return + * + * @see + */ +STATIC INLINE VOID AddNode2SortLink(SortLinkAttribute *sortLinkHeader, SortLinkList *sortList) +{ + LOS_DL_LIST *head = (LOS_DL_LIST *)&sortLinkHeader->sortLink; //获取双向链表 + + if (LOS_ListEmpty(head)) { //空链表,直接插入 + LOS_ListHeadInsert(head, &sortList->sortLinkNode);//插入结点 + sortLinkHeader->nodeNum++;//CPU的工作量增加了 + return; + } + //链表不为空时,插入分三种情况, responseTime 大于,等于,小于的处理 + SortLinkList *listSorted = LOS_DL_LIST_ENTRY(head->pstNext, SortLinkList, sortLinkNode); + if (listSorted->responseTime > sortList->responseTime) {//如果要插入的节点 responseTime 最小 + LOS_ListAdd(head, &sortList->sortLinkNode);//能跑进来说明是最小的,直接插入到第一的位置 + sortLinkHeader->nodeNum++;//CPU的工作量增加了 + return;//直接返回了 + } else if (listSorted->responseTime == sortList->responseTime) {//相等的情况 + LOS_ListAdd(head->pstNext, &sortList->sortLinkNode);//插到第二的位置 + sortLinkHeader->nodeNum++; + return; + } + //处理大于链表中第一个responseTime的情况,需要遍历链表 + LOS_DL_LIST *prevNode = head->pstPrev;//注意这里用的前一个结点,也就是说前一个结点中的responseTime 是最大的 + do { // @note_good 这里写的有点妙,也是双向链表的魅力所在 + listSorted = LOS_DL_LIST_ENTRY(prevNode, SortLinkList, sortLinkNode);//一个个遍历,先比大的再比小的 + if (listSorted->responseTime <= sortList->responseTime) {//如果时间比你小,就插到后面 +>>>>>>> remotes/origin/main LOS_ListAdd(prevNode, &sortList->sortLinkNode); sortLinkHeader->nodeNum++; break; } +<<<<<<< HEAD prevNode = prevNode->pstPrev; } while (1); +======= + prevNode = prevNode->pstPrev;//再拿上一个更小的responseTime进行比较 + } while (1);//死循环 +>>>>>>> remotes/origin/main } VOID OsAdd2SortLink(SortLinkAttribute *head, SortLinkList *node, UINT64 responseTime, UINT16 idleCpu) diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_dump.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_dump.c index f326ea12..05d64221 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_dump.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_dump.c @@ -159,26 +159,44 @@ UINT32 OsKProcessPmUsage(LosVmSpace *kSpace, UINT32 *actualPm) /* Kernel resident memory, include default heap memory */ memUsed = SYS_MEM_SIZE_DEFAULT - (totalCount << PAGE_SHIFT); +<<<<<<< HEAD spaceList = LOS_GetVmSpaceList(); LosMux *vmSpaceListMux = OsGVmSpaceMuxGet(); (VOID)LOS_MuxAcquire(vmSpaceListMux); LOS_DL_LIST_FOR_EACH_ENTRY(space, spaceList, LosVmSpace, node) { if (space == LOS_GetKVmSpace()) { +======= + spaceList = LOS_GetVmSpaceList();//获取虚拟空间链表,上面挂了所有虚拟空间 + LosMux *vmSpaceListMux = OsGVmSpaceMuxGet(); + (VOID)LOS_MuxAcquire(vmSpaceListMux); + LOS_DL_LIST_FOR_EACH_ENTRY(space, spaceList, LosVmSpace, node) {//遍历链表 + if (space == LOS_GetKVmSpace()) {//内核空间不统计 +>>>>>>> remotes/origin/main continue; } UProcessUsed += OsUProcessPmUsage(space, NULL, NULL); } (VOID)LOS_MuxRelease(vmSpaceListMux); +<<<<<<< HEAD /* Kernel dynamic memory, include extended heap memory */ memUsed += ((usedCount << PAGE_SHIFT) - UProcessUsed); /* Remaining heap memory */ +======= + /* Kernel dynamic memory, include extended heap memory */ //内核动态内存,包括扩展堆内存 + memUsed += ((usedCount << PAGE_SHIFT) - UProcessUsed); + /* Remaining heap memory */ //剩余堆内存 +>>>>>>> remotes/origin/main memUsed -= freeMem; *actualPm = memUsed; return memUsed; } +<<<<<<< HEAD +======= +///shell task 物理内存的使用情况 +>>>>>>> remotes/origin/main UINT32 OsShellCmdProcessPmUsage(LosVmSpace *space, UINT32 *sharePm, UINT32 *actualPm) { if (space == NULL) { @@ -194,7 +212,11 @@ UINT32 OsShellCmdProcessPmUsage(LosVmSpace *space, UINT32 *sharePm, UINT32 *actu } return OsUProcessPmUsage(space, sharePm, actualPm); } +<<<<<<< HEAD +======= +///虚拟空间物理内存的使用情况,参数同时带走共享物理内存sharePm和actualPm单位是字节 +>>>>>>> remotes/origin/main UINT32 OsUProcessPmUsage(LosVmSpace *space, UINT32 *sharePm, UINT32 *actualPm) { LosVmMapRegion *region = NULL; @@ -230,10 +252,17 @@ UINT32 OsUProcessPmUsage(LosVmSpace *space, UINT32 *sharePm, UINT32 *actualPm) continue; } +<<<<<<< HEAD shareRef = LOS_AtomicRead(&page->refCounts); if (shareRef > 1) { if (sharePm != NULL) { *sharePm += PAGE_SIZE; +======= + shareRef = LOS_AtomicRead(&page->refCounts);//ref 大于1 说明page被其他空间也引用了 + if (shareRef > 1) { + if (sharePm != NULL) { + *sharePm += PAGE_SIZE;//一页4K字节 +>>>>>>> remotes/origin/main } pmSize += PAGE_SIZE / shareRef; } else { @@ -250,7 +279,13 @@ UINT32 OsUProcessPmUsage(LosVmSpace *space, UINT32 *sharePm, UINT32 *actualPm) return pmSize; } +<<<<<<< HEAD +======= +/// @brief 通过虚拟空间获取进程实体 +/// @param space +/// @return +>>>>>>> remotes/origin/main LosProcessCB *OsGetPIDByAspace(const LosVmSpace *space) { UINT32 pid; @@ -258,6 +293,7 @@ LosProcessCB *OsGetPIDByAspace(const LosVmSpace *space) LosProcessCB *processCB = NULL; SCHEDULER_LOCK(intSave); +<<<<<<< HEAD for (pid = 0; pid < g_processMaxNum; ++pid) { processCB = g_processCBArray + pid; if (OsProcessIsUnused(processCB)) { @@ -265,6 +301,15 @@ LosProcessCB *OsGetPIDByAspace(const LosVmSpace *space) } if (processCB->vmSpace == space) { +======= + for (pid = 0; pid < g_processMaxNum; ++pid) {//循环进程池,进程池本质是个数组 + processCB = g_processCBArray + pid; + if (OsProcessIsUnused(processCB)) {//进程还没被分配使用 + continue; + } + + if (processCB->vmSpace == space) {//找到了 +>>>>>>> remotes/origin/main SCHEDULER_UNLOCK(intSave); return processCB; } @@ -272,7 +317,15 @@ LosProcessCB *OsGetPIDByAspace(const LosVmSpace *space) SCHEDULER_UNLOCK(intSave); return NULL; } - +<<<<<<< HEAD + +======= +/// @brief 统计虚拟空间中某个线性区的页数 +/// @param space +/// @param region +/// @param pssPages +/// @return +>>>>>>> remotes/origin/main UINT32 OsCountRegionPages(LosVmSpace *space, LosVmMapRegion *region, UINT32 *pssPages) { UINT32 regionPages = 0; @@ -306,7 +359,11 @@ UINT32 OsCountRegionPages(LosVmSpace *space, LosVmMapRegion *region, UINT32 *pss return regionPages; } +<<<<<<< HEAD +======= +///统计虚拟空间的总页数 +>>>>>>> remotes/origin/main UINT32 OsCountAspacePages(LosVmSpace *space) { UINT32 spacePages = 0; @@ -390,27 +447,49 @@ VOID OsDumpRegion2(LosVmSpace *space, LosVmMapRegion *region) region->range.size, flagsStr, regionPages, pssPages); (VOID)LOS_MemFree(m_aucSysMem0, flagsStr); } +<<<<<<< HEAD +======= +///dump 指定虚拟空间的信息 +>>>>>>> remotes/origin/main VOID OsDumpAspace(LosVmSpace *space) { LosVmMapRegion *region = NULL; LosRbNode *pstRbNode = NULL; LosRbNode *pstRbNodeNext = NULL; UINT32 spacePages; +<<<<<<< HEAD LosProcessCB *pcb = OsGetPIDByAspace(space); +======= + LosProcessCB *pcb = OsGetPIDByAspace(space);//通过虚拟空间找到进程实体 +>>>>>>> remotes/origin/main if (pcb == NULL) { return; } +<<<<<<< HEAD spacePages = OsCountAspacePages(space); +======= + //进程ID | 进程虚拟内存控制块地址信息 | 虚拟内存起始地址 | 虚拟内存大小 | 已使用的物理页数量 + spacePages = OsCountAspacePages(space);//获取空间的页数 +>>>>>>> remotes/origin/main PRINTK("\r\n PID aspace name base size pages \n"); PRINTK(" ---- ------ ---- ---- ----- ----\n"); PRINTK(" %-4d %#010x %-10.10s %#010x %#010x %d\n", pcb->processID, space, pcb->processName, space->base, space->size, spacePages); +<<<<<<< HEAD PRINTK("\r\n\t region name base size mmu_flags pages pg/ref\n"); PRINTK("\t ------ ---- ---- ---- --------- ----- -----\n"); RB_SCAN_SAFE(&space->regionRbTree, pstRbNode, pstRbNodeNext) +======= + + //虚拟区间控制块地址信息 | 虚拟区间类型 | 虚拟区间起始地址 | 虚拟区间大小 | 虚拟区间mmu映射属性 | 已使用的物理页数量(包括共享内存部分 | 已使用的物理页数量 + + PRINTK("\r\n\t region name base size mmu_flags pages pg/ref\n"); + PRINTK("\t ------ ---- ---- ---- --------- ----- -----\n"); + RB_SCAN_SAFE(&space->regionRbTree, pstRbNode, pstRbNodeNext)//按region 轮询统计 +>>>>>>> remotes/origin/main region = (LosVmMapRegion *)pstRbNode; if (region != NULL) { OsDumpRegion2(space, region); @@ -421,6 +500,7 @@ VOID OsDumpAspace(LosVmSpace *space) RB_SCAN_SAFE_END(&space->regionRbTree, pstRbNode, pstRbNodeNext) return; } +<<<<<<< HEAD VOID OsDumpAllAspace(VOID) { @@ -429,6 +509,16 @@ VOID OsDumpAllAspace(VOID) LOS_DL_LIST_FOR_EACH_ENTRY(space, aspaceList, LosVmSpace, node) { (VOID)LOS_MuxAcquire(&space->regionMux); OsDumpAspace(space); +======= +///查看所有进程使用虚拟内存的情况 +VOID OsDumpAllAspace(VOID) +{ + LosVmSpace *space = NULL; + LOS_DL_LIST *aspaceList = LOS_GetVmSpaceList();//获取所有空间链表 + LOS_DL_LIST_FOR_EACH_ENTRY(space, aspaceList, LosVmSpace, node) {//循环取出进程虚拟 + (VOID)LOS_MuxAcquire(&space->regionMux); + OsDumpAspace(space);//dump 空间 +>>>>>>> remotes/origin/main (VOID)LOS_MuxRelease(&space->regionMux); } return; @@ -447,11 +537,19 @@ STATUS_T OsRegionOverlapCheck(LosVmSpace *space, LosVmMapRegion *region) (VOID)LOS_MuxRelease(&space->regionMux); return ret; } +<<<<<<< HEAD VOID OsDumpPte(VADDR_T vaddr) { UINT32 l1Index = vaddr >> MMU_DESCRIPTOR_L1_SMALL_SHIFT; LosVmSpace *space = LOS_SpaceGet(vaddr); +======= +///dump 页表项 +VOID OsDumpPte(VADDR_T vaddr) +{ + UINT32 l1Index = vaddr >> MMU_DESCRIPTOR_L1_SMALL_SHIFT; + LosVmSpace *space = LOS_SpaceGet(vaddr);//通过虚拟地址获取空间,内核分三个空间 内核进程空间,内核堆空间,用户进程空间 +>>>>>>> remotes/origin/main UINT32 ttEntry; LosVmPage *page = NULL; PTE_T *l2Table = NULL; @@ -461,6 +559,7 @@ VOID OsDumpPte(VADDR_T vaddr) return; } +<<<<<<< HEAD ttEntry = space->archMmu.virtTtb[l1Index]; if (ttEntry) { l2Table = LOS_PaddrToKVaddr(MMU_DESCRIPTOR_L1_PAGE_TABLE_ADDR(ttEntry)); @@ -469,19 +568,38 @@ VOID OsDumpPte(VADDR_T vaddr) goto ERR; } page = LOS_VmPageGet(l2Table[l2Index] & ~(PAGE_SIZE - 1)); +======= + ttEntry = space->archMmu.virtTtb[l1Index];//找到 L1 页面项 + if (ttEntry) { + l2Table = LOS_PaddrToKVaddr(MMU_DESCRIPTOR_L1_PAGE_TABLE_ADDR(ttEntry));//找到L1页面项对应的 L2表 + l2Index = (vaddr % MMU_DESCRIPTOR_L1_SMALL_SIZE) >> PAGE_SHIFT;//找到L2页面项 + if (l2Table == NULL) { + goto ERR; + } + page = LOS_VmPageGet(l2Table[l2Index] & ~(PAGE_SIZE - 1));//获取物理页框 +>>>>>>> remotes/origin/main if (page == NULL) { goto ERR; } PRINTK("vaddr %p, l1Index %d, ttEntry %p, l2Table %p, l2Index %d, pfn %p count %d\n", +<<<<<<< HEAD vaddr, l1Index, ttEntry, l2Table, l2Index, l2Table[l2Index], LOS_AtomicRead(&page->refCounts)); } else { +======= + vaddr, l1Index, ttEntry, l2Table, l2Index, l2Table[l2Index], LOS_AtomicRead(&page->refCounts));//打印L1 L2 页表项 + } else {//不在L1表 +>>>>>>> remotes/origin/main PRINTK("vaddr %p, l1Index %d, ttEntry %p\n", vaddr, l1Index, ttEntry); } return; ERR: PRINTK("%s, error vaddr: %#x, l2Table: %#x, l2Index: %#x\n", __FUNCTION__, vaddr, l2Table, l2Index); } +<<<<<<< HEAD +======= +///获取段剩余页框数 +>>>>>>> remotes/origin/main UINT32 OsVmPhySegPagesGet(LosVmPhysSeg *seg) { UINT32 intSave; @@ -489,6 +607,7 @@ UINT32 OsVmPhySegPagesGet(LosVmPhysSeg *seg) UINT32 segFreePages = 0; LOS_SpinLockSave(&seg->freeListLock, &intSave); +<<<<<<< HEAD for (flindex = 0; flindex < VM_LIST_ORDER_MAX; flindex++) { segFreePages += ((1 << flindex) * seg->freeList[flindex].listCnt); } @@ -497,6 +616,27 @@ UINT32 OsVmPhySegPagesGet(LosVmPhysSeg *seg) return segFreePages; } +======= + for (flindex = 0; flindex < VM_LIST_ORDER_MAX; flindex++) {//遍历块组 + segFreePages += ((1 << flindex) * seg->freeList[flindex].listCnt);//1 << flindex等于页数, * 节点数 得到组块的总页数. + } + LOS_SpinUnlockRestore(&seg->freeListLock, intSave); + + return segFreePages;//返回剩余未分配的总物理页框 +} +///dump 物理内存 +/*********************************************************** +* phys_seg:物理页控制块地址信息 +* base:第一个物理页地址,即物理页内存起始地址 +* size:物理页内存大小 +* free_pages:空闲物理页数量 +* active anon: pagecache中,活跃的匿名页数量 +* inactive anon: pagecache中,不活跃的匿名页数量 +* active file: pagecache中,活跃的文件页数量 +* inactive file: pagecache中,不活跃的文件页数量 +* pmm pages total:总的物理页数,used:已使用的物理页数,free:空闲的物理页数 +************************************************************/ +>>>>>>> remotes/origin/main VOID OsVmPhysDump(VOID) { LosVmPhysSeg *seg = NULL; @@ -508,7 +648,11 @@ VOID OsVmPhysDump(VOID) UINT32 flindex; UINT32 listCount[VM_LIST_ORDER_MAX] = {0}; +<<<<<<< HEAD for (segIndex = 0; segIndex < g_vmPhysSegNum; segIndex++) { +======= + for (segIndex = 0; segIndex < g_vmPhysSegNum; segIndex++) {//循环取段 +>>>>>>> remotes/origin/main seg = &g_vmPhysSeg[segIndex]; if (seg->size > 0) { segFreePages = OsVmPhySegPagesGet(seg); @@ -538,7 +682,11 @@ VOID OsVmPhysDump(VOID) PRINTK("\n\rpmm pages: total = %u, used = %u, free = %u\n", totalPages, (totalPages - totalFreePages), totalFreePages); } +<<<<<<< HEAD +======= +///获取物理内存的使用信息,两个参数接走数据 +>>>>>>> remotes/origin/main VOID OsVmPhysUsedInfoGet(UINT32 *usedCount, UINT32 *totalCount) { UINT32 index; @@ -551,12 +699,21 @@ VOID OsVmPhysUsedInfoGet(UINT32 *usedCount, UINT32 *totalCount) *usedCount = 0; *totalCount = 0; +<<<<<<< HEAD for (index = 0; index < g_vmPhysSegNum; index++) { physSeg = &g_vmPhysSeg[index]; if (physSeg->size > 0) { *totalCount += physSeg->size >> PAGE_SHIFT; segFreePages = OsVmPhySegPagesGet(physSeg); *usedCount += (*totalCount - segFreePages); +======= + for (index = 0; index < g_vmPhysSegNum; index++) {//循环取段 + physSeg = &g_vmPhysSeg[index]; + if (physSeg->size > 0) { + *totalCount += physSeg->size >> PAGE_SHIFT;//叠加段的总页数 + segFreePages = OsVmPhySegPagesGet(physSeg);//获取段的剩余页数 + *usedCount += (*totalCount - segFreePages);//叠加段的使用页数 +>>>>>>> remotes/origin/main } } } diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_fault.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_fault.c index 86b08199..9118a18f 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_fault.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_fault.c @@ -54,7 +54,11 @@ extern char __exc_table_start[]; extern char __exc_table_end[]; +<<<<<<< HEAD +======= +//线性正确性检查 +>>>>>>> remotes/origin/main STATIC STATUS_T OsVmRegionPermissionCheck(LosVmMapRegion *region, UINT32 flags) { if ((region->regionFlags & VM_MAP_REGION_FLAG_PERM_READ) != VM_MAP_REGION_FLAG_PERM_READ) { @@ -62,14 +66,22 @@ STATIC STATUS_T OsVmRegionPermissionCheck(LosVmMapRegion *region, UINT32 flags) return LOS_NOK; } +<<<<<<< HEAD if ((flags & VM_MAP_PF_FLAG_WRITE) == VM_MAP_PF_FLAG_WRITE) { +======= + if ((flags & VM_MAP_PF_FLAG_WRITE) == VM_MAP_PF_FLAG_WRITE) {//写入许可 +>>>>>>> remotes/origin/main if ((region->regionFlags & VM_MAP_REGION_FLAG_PERM_WRITE) != VM_MAP_REGION_FLAG_PERM_WRITE) { VM_ERR("write permission check failed operation flags %x, region flags %x", flags, region->regionFlags); return LOS_NOK; } } +<<<<<<< HEAD if ((flags & VM_MAP_PF_FLAG_INSTRUCTION) == VM_MAP_PF_FLAG_INSTRUCTION) { +======= + if ((flags & VM_MAP_PF_FLAG_INSTRUCTION) == VM_MAP_PF_FLAG_INSTRUCTION) {//指令 +>>>>>>> remotes/origin/main if ((region->regionFlags & VM_MAP_REGION_FLAG_PERM_EXECUTE) != VM_MAP_REGION_FLAG_PERM_EXECUTE) { VM_ERR("exec permission check failed operation flags %x, region flags %x", flags, region->regionFlags); return LOS_NOK; @@ -97,7 +109,12 @@ STATIC VOID OsFaultTryFixup(ExcContext *frame, VADDR_T excVaddr, STATUS_T *statu } #ifdef LOSCFG_FS_VFS +<<<<<<< HEAD STATIC STATUS_T OsDoReadFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) +======= +//读页时发生缺页的处理 +STATIC STATUS_T OsDoReadFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault)//读缺页 +>>>>>>> remotes/origin/main { status_t ret; PADDR_T paddr; @@ -105,16 +122,25 @@ STATIC STATUS_T OsDoReadFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) VADDR_T vaddr = (VADDR_T)vmPgFault->vaddr; LosVmSpace *space = region->space; +<<<<<<< HEAD ret = LOS_ArchMmuQuery(&space->archMmu, vaddr, NULL, NULL); if (ret == LOS_OK) { return LOS_OK; } if (region->unTypeData.rf.vmFOps == NULL || region->unTypeData.rf.vmFOps->fault == NULL) { +======= + ret = LOS_ArchMmuQuery(&space->archMmu, vaddr, NULL, NULL);//查询是否缺页 + if (ret == LOS_OK) {//注意这里时LOS_OK却返回,都OK了说明查到了物理地址,有页了。 + return LOS_OK;//查到了就说明不缺页的,缺页就是因为虚拟地址没有映射到物理地址嘛 + } + if (region->unTypeData.rf.vmFOps == NULL || region->unTypeData.rf.vmFOps->fault == NULL) {//线性区必须有实现了缺页接口 +>>>>>>> remotes/origin/main VM_ERR("region args invalid, file path: %s", region->unTypeData.rf.vnode->filePath); return LOS_ERRNO_VM_INVALID_ARGS; } (VOID)LOS_MuxAcquire(®ion->unTypeData.rf.vnode->mapping.mux_lock); +<<<<<<< HEAD ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault); if (ret == LOS_OK) { paddr = LOS_PaddrQuery(vmPgFault->pageKVaddr); @@ -125,6 +151,18 @@ STATIC STATUS_T OsDoReadFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) } ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, region->regionFlags & (~VM_MAP_REGION_FLAG_PERM_WRITE)); +======= + ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault);// 函数指针,执行的是g_commVmOps.OsVmmFileFault + if (ret == LOS_OK) { + paddr = LOS_PaddrQuery(vmPgFault->pageKVaddr);//查询物理地址 + page = LOS_VmPageGet(paddr);//获取page + if (page != NULL) { /* just incase of page null */ + LOS_AtomicInc(&page->refCounts);//ref 自增 + OsCleanPageLocked(page); + } + ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, + region->regionFlags & (~VM_MAP_REGION_FLAG_PERM_WRITE));//重新映射为非可写 +>>>>>>> remotes/origin/main if (ret < 0) { VM_ERR("LOS_ArchMmuMap failed"); OsDelMapInfo(region, vmPgFault, false); @@ -140,7 +178,11 @@ STATIC STATUS_T OsDoReadFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) return LOS_ERRNO_VM_NO_MEMORY; } +<<<<<<< HEAD /* unmap a page when cow happened only */ +======= +/* unmap a page when cow happened only *///仅当写时拷贝发生时取消页面映射 +>>>>>>> remotes/origin/main STATIC LosVmPage *OsCowUnmapOrg(LosArchMmu *archMmu, LosVmMapRegion *region, LosVmPgFault *vmf) { UINT32 intSave; @@ -168,7 +210,11 @@ STATIC LosVmPage *OsCowUnmapOrg(LosArchMmu *archMmu, LosVmMapRegion *region, Los return oldPage; } #endif +<<<<<<< HEAD +======= +//在私有线性区写入文件时发生缺页的处理 +>>>>>>> remotes/origin/main status_t OsDoCowFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) { STATUS_T ret; @@ -186,23 +232,40 @@ status_t OsDoCowFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) } space = region->space; +<<<<<<< HEAD ret = LOS_ArchMmuQuery(&space->archMmu, (VADDR_T)vmPgFault->vaddr, &oldPaddr, NULL); if (ret == LOS_OK) { oldPage = OsCowUnmapOrg(&space->archMmu, region, vmPgFault); } newPage = LOS_PhysPageAlloc(); +======= + ret = LOS_ArchMmuQuery(&space->archMmu, (VADDR_T)vmPgFault->vaddr, &oldPaddr, NULL);//查询出老物理地址 + if (ret == LOS_OK) { + oldPage = OsCowUnmapOrg(&space->archMmu, region, vmPgFault);//取消页面映射 + } + + newPage = LOS_PhysPageAlloc();//分配一个新页面 +>>>>>>> remotes/origin/main if (newPage == NULL) { VM_ERR("LOS_PhysPageAlloc failed"); ret = LOS_ERRNO_VM_NO_MEMORY; goto ERR_OUT; } +<<<<<<< HEAD newPaddr = VM_PAGE_TO_PHYS(newPage); kvaddr = OsVmPageToVaddr(newPage); (VOID)LOS_MuxAcquire(®ion->unTypeData.rf.vnode->mapping.mux_lock); ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault); +======= + newPaddr = VM_PAGE_TO_PHYS(newPage);//拿到新的物理地址 + kvaddr = OsVmPageToVaddr(newPage);//拿到新的虚拟地址 + + (VOID)LOS_MuxAcquire(®ion->unTypeData.rf.vnode->mapping.mux_lock); + ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault);// 函数指针 g_commVmOps.OsVmmFileFault +>>>>>>> remotes/origin/main if (ret != LOS_OK) { VM_ERR("call region->vm_ops->fault fail"); (VOID)LOS_MuxRelease(®ion->unTypeData.rf.vnode->mapping.mux_lock); @@ -214,6 +277,7 @@ status_t OsDoCowFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) * we can take it as a normal file cow map. 2.this page has done file cow map, * we can take it as a anonymous cow map. */ +<<<<<<< HEAD if ((oldPaddr == 0) || (LOS_PaddrToKVaddr(oldPaddr) == vmPgFault->pageKVaddr)) { (VOID)memcpy_s(kvaddr, PAGE_SIZE, vmPgFault->pageKVaddr, PAGE_SIZE); LOS_AtomicInc(&newPage->refCounts); @@ -223,11 +287,26 @@ status_t OsDoCowFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) /* use old page free the new one */ if (newPaddr == oldPaddr) { LOS_PhysPageFree(newPage); +======= + if ((oldPaddr == 0) || (LOS_PaddrToKVaddr(oldPaddr) == vmPgFault->pageKVaddr)) {//没有映射或者 已在pagecache有映射 + (VOID)memcpy_s(kvaddr, PAGE_SIZE, vmPgFault->pageKVaddr, PAGE_SIZE);//直接copy到新页 + LOS_AtomicInc(&newPage->refCounts);//引用ref++ + OsCleanPageLocked(LOS_VmPageGet(LOS_PaddrQuery(vmPgFault->pageKVaddr)));//解锁 + } else { + OsPhysSharePageCopy(oldPaddr, &newPaddr, newPage);//调用之前 oldPaddr肯定不等于newPaddr + /* use old page free the new one */ + if (newPaddr == oldPaddr) {//注意这里newPaddr可能已经被改变了,参数传入的是 &newPaddr + LOS_PhysPageFree(newPage);//释放新页,别浪费的内存,内核使用内存是一分钱当十块用. +>>>>>>> remotes/origin/main newPage = NULL; } } +<<<<<<< HEAD ret = LOS_ArchMmuMap(&space->archMmu, (VADDR_T)vmPgFault->vaddr, newPaddr, 1, region->regionFlags); +======= + ret = LOS_ArchMmuMap(&space->archMmu, (VADDR_T)vmPgFault->vaddr, newPaddr, 1, region->regionFlags);//把新物理地址映射给缺页的虚拟地址,这样就不会缺页啦 +>>>>>>> remotes/origin/main if (ret < 0) { VM_ERR("LOS_ArchMmuMap failed"); ret = LOS_ERRNO_VM_NO_MEMORY; @@ -252,7 +331,11 @@ ERR_OUT: return ret; } +<<<<<<< HEAD +======= +///在共享线性区写文件操作发生缺页的情况处理,因为线性区是共享的 +>>>>>>> remotes/origin/main status_t OsDoSharedFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) { STATUS_T ret; @@ -268,10 +351,17 @@ status_t OsDoSharedFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) return LOS_ERRNO_VM_INVALID_ARGS; } +<<<<<<< HEAD ret = LOS_ArchMmuQuery(&space->archMmu, vmPgFault->vaddr, &paddr, NULL); if (ret == LOS_OK) { LOS_ArchMmuUnmap(&space->archMmu, vmPgFault->vaddr, 1); ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, region->regionFlags); +======= + ret = LOS_ArchMmuQuery(&space->archMmu, vmPgFault->vaddr, &paddr, NULL);//查询物理地址 + if (ret == LOS_OK) { + LOS_ArchMmuUnmap(&space->archMmu, vmPgFault->vaddr, 1);//先取消映射 + ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, region->regionFlags);//再重新映射,为啥这么干,是因为regionFlags变了, +>>>>>>> remotes/origin/main if (ret < 0) { VM_ERR("LOS_ArchMmuMap failed. ret=%d", ret); return LOS_ERRNO_VM_NO_MEMORY; @@ -279,16 +369,27 @@ status_t OsDoSharedFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) LOS_SpinLockSave(®ion->unTypeData.rf.vnode->mapping.list_lock, &intSave); fpage = OsFindGetEntry(®ion->unTypeData.rf.vnode->mapping, vmPgFault->pgoff); +<<<<<<< HEAD if (fpage) { OsMarkPageDirty(fpage, region, 0, 0); +======= + if (fpage) {//在页高速缓存(page cache)中找到了 + OsMarkPageDirty(fpage, region, 0, 0);//标记为脏页 +>>>>>>> remotes/origin/main } LOS_SpinUnlockRestore(®ion->unTypeData.rf.vnode->mapping.list_lock, intSave); return LOS_OK; } +<<<<<<< HEAD (VOID)LOS_MuxAcquire(®ion->unTypeData.rf.vnode->mapping.mux_lock); ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault); +======= + //以下是没有映射到物理地址的处理 + (VOID)LOS_MuxAcquire(®ion->unTypeData.rf.vnode->mapping.mux_lock); + ret = region->unTypeData.rf.vmFOps->fault(region, vmPgFault);//函数指针,执行的是g_commVmOps.OsVmmFileFault +>>>>>>> remotes/origin/main if (ret == LOS_OK) { paddr = LOS_PaddrQuery(vmPgFault->pageKVaddr); page = LOS_VmPageGet(paddr); @@ -319,11 +420,16 @@ status_t OsDoSharedFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault) * For COW fault, pagecache is copied to private anonyous pages and the changes on this page * won't write through to the underlying file. For SHARED fault, pagecache is mapping with * region->arch_mmu_flags and the changes on this page will write through to the underlying file +<<<<<<< HEAD */ +======= + */ //操作文件时产生缺页中断 +>>>>>>> remotes/origin/main STATIC STATUS_T OsDoFileFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault, UINT32 flags) { STATUS_T ret; +<<<<<<< HEAD if (flags & VM_MAP_PF_FLAG_WRITE) { if (region->regionFlags & VM_MAP_REGION_FLAG_SHARED) { ret = OsDoSharedFault(region, vmPgFault); @@ -336,6 +442,33 @@ STATIC STATUS_T OsDoFileFault(LosVmMapRegion *region, LosVmPgFault *vmPgFault, U return ret; } +======= + if (flags & VM_MAP_PF_FLAG_WRITE) {//写页的时候产生缺页 + if (region->regionFlags & VM_MAP_REGION_FLAG_SHARED) {//共享线性区 + ret = OsDoSharedFault(region, vmPgFault);//写操作时的共享缺页,最复杂,此页上的更改将写入磁盘文件 + } else {//非共享线性区 + ret = OsDoCowFault(region, vmPgFault);//(写时拷贝技术)写操作时的私有缺页,pagecache被复制到私有的任意一个页面上,并在此页面上进行更改,不会直接写入磁盘文件 + } + } else {//读页的时候产生缺页 + ret = OsDoReadFault(region, vmPgFault);//页面读取操作很简单,只需共享页面缓存(节省内存)并进行读权限映射(region->arch_mmu_flags&(~arch_mmu_FLAG_PERM_WRITE)) + } + return ret; +} +/*************************************************************** +缺页中断处理程序 +通常有两种情况导致 +第一种:由编程错误引起的异常 +第二种:属于进程的地址空间范围但还尚未分配物理页框引起的异常 +***************************************************************/ + +/** + * @brief + * @param vaddr + * @param flags + * @param frame + * @return STATUS_T + */ +>>>>>>> remotes/origin/main STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) { LosVmSpace *space = LOS_SpaceGet(vaddr); @@ -354,9 +487,15 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) return status; } +<<<<<<< HEAD if (((flags & VM_MAP_PF_FLAG_USER) != 0) && (!LOS_IsUserAddress(vaddr))) { VM_ERR("user space not allowed to access invalid address: %#x", vaddr); return LOS_ERRNO_VM_ACCESS_DENIED; +======= + if (((flags & VM_MAP_PF_FLAG_USER) != 0) && (!LOS_IsUserAddress(vaddr))) {//地址保护,用户空间不允许跨界访问 + VM_ERR("user space not allowed to access invalid address: %#x", vaddr); + return LOS_ERRNO_VM_ACCESS_DENIED;//拒绝访问 +>>>>>>> remotes/origin/main } #ifdef LOSCFG_KERNEL_PLIMITS @@ -366,7 +505,11 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) #endif (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD region = LOS_RegionFind(space, vaddr); +======= + region = LOS_RegionFind(space, vaddr);//通过虚拟地址找到所在线性区 +>>>>>>> remotes/origin/main if (region == NULL) { VM_ERR("region not exists, vaddr: %#x", vaddr); status = LOS_ERRNO_VM_NOT_FOUND; @@ -375,11 +518,19 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) status = OsVmRegionPermissionCheck(region, flags); if (status != LOS_OK) { +<<<<<<< HEAD status = LOS_ERRNO_VM_ACCESS_DENIED; goto CHECK_FAILED; } if (OomCheckProcess()) { +======= + status = LOS_ERRNO_VM_ACCESS_DENIED;//拒绝访问 + goto CHECK_FAILED; + } + + if (OomCheckProcess()) {//低内存检查 +>>>>>>> remotes/origin/main /* * under low memory, when user process request memory allocation * it will fail, and result is LOS_NOK and current user process @@ -389,6 +540,7 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) goto CHECK_FAILED; } +<<<<<<< HEAD vaddr = ROUNDDOWN(vaddr, PAGE_SIZE); #ifdef LOSCFG_FS_VFS if (LOS_IsRegionFileValid(region)) { @@ -401,6 +553,20 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) vmPgFault.pageKVaddr = NULL; status = OsDoFileFault(region, &vmPgFault, flags); +======= + vaddr = ROUNDDOWN(vaddr, PAGE_SIZE);//为啥要向下圆整,因为这一页要重新使用,需找到页面基地址 +#ifdef LOSCFG_FS_VFS + if (LOS_IsRegionFileValid(region)) {//是否为文件线性区 + if (region->unTypeData.rf.vnode == NULL) { + goto CHECK_FAILED; + } + vmPgFault.vaddr = vaddr;//虚拟地址 + vmPgFault.pgoff = ((vaddr - region->range.base) >> PAGE_SHIFT) + region->pgOff;//计算出文件读取位置 + vmPgFault.flags = flags; + vmPgFault.pageKVaddr = NULL;//缺失页初始化没有物理地址 + + status = OsDoFileFault(region, &vmPgFault, flags);//缺页处理 +>>>>>>> remotes/origin/main if (status) { VM_ERR("vm fault error, status=%d", status); goto CHECK_FAILED; @@ -408,13 +574,19 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) goto DONE; } #endif +<<<<<<< HEAD newPage = LOS_PhysPageAlloc(); +======= + //请求调页:推迟到不能再推迟为止 + newPage = LOS_PhysPageAlloc();//分配一个新的物理页 +>>>>>>> remotes/origin/main if (newPage == NULL) { status = LOS_ERRNO_VM_NO_MEMORY; goto CHECK_FAILED; } +<<<<<<< HEAD newPaddr = VM_PAGE_TO_PHYS(newPage); (VOID)memset_s(OsVmPageToVaddr(newPage), PAGE_SIZE, 0, PAGE_SIZE); status = LOS_ArchMmuQuery(&space->archMmu, vaddr, &oldPaddr, NULL); @@ -424,11 +596,26 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) /* use old page free the new one */ if (newPaddr == oldPaddr) { LOS_PhysPageFree(newPage); +======= + newPaddr = VM_PAGE_TO_PHYS(newPage);//获取物理地址 + (VOID)memset_s(OsVmPageToVaddr(newPage), PAGE_SIZE, 0, PAGE_SIZE);//获取虚拟地址 清0 + status = LOS_ArchMmuQuery(&space->archMmu, vaddr, &oldPaddr, NULL);//通过虚拟地址查询老物理地址 + if (status >= 0) { + LOS_ArchMmuUnmap(&space->archMmu, vaddr, 1);//解除映射关系 + OsPhysSharePageCopy(oldPaddr, &newPaddr, newPage);//将oldPaddr的数据拷贝到newPage + /* use old page free the new one */ + if (newPaddr == oldPaddr) {//新老物理地址一致 + LOS_PhysPageFree(newPage);//继续使用旧页释放新页 +>>>>>>> remotes/origin/main newPage = NULL; } /* map all of the pages */ +<<<<<<< HEAD status = LOS_ArchMmuMap(&space->archMmu, vaddr, newPaddr, 1, region->regionFlags); +======= + status = LOS_ArchMmuMap(&space->archMmu, vaddr, newPaddr, 1, region->regionFlags);//重新映射新物理地址 +>>>>>>> remotes/origin/main if (status < 0) { VM_ERR("failed to map replacement page, status:%d", status); status = LOS_ERRNO_VM_MAP_FAILED; @@ -439,8 +626,13 @@ STATUS_T OsVmPageFaultHandler(VADDR_T vaddr, UINT32 flags, ExcContext *frame) goto DONE; } else { /* map all of the pages */ +<<<<<<< HEAD LOS_AtomicInc(&newPage->refCounts); status = LOS_ArchMmuMap(&space->archMmu, vaddr, newPaddr, 1, region->regionFlags); +======= + LOS_AtomicInc(&newPage->refCounts);//引用数自增 + status = LOS_ArchMmuMap(&space->archMmu, vaddr, newPaddr, 1, region->regionFlags);//映射新物理地址,如此下次就不会缺页了 +>>>>>>> remotes/origin/main if (status < 0) { VM_ERR("failed to map page, status:%d", status); status = LOS_ERRNO_VM_MAP_FAILED; diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_filemap.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_filemap.c index e95cd826..cc8e97d0 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_filemap.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_filemap.c @@ -66,31 +66,64 @@ VOID ResetPageCacheHitInfo(int *try, int *hit) #define TRACE_TRY_CACHE() #define TRACE_HIT_CACHE() #endif +<<<<<<< HEAD #ifdef LOSCFG_KERNEL_VM +======= +#ifdef LOSCFG_KERNEL_VM + +/** + * @brief + @verbatim + 增加文件页到页高速缓存(page cache) + LosFilePage将一个文件切成了一页一页,因为读文件过程随机seek,所以文件页也不会是连续的, + pgoff记录文件的位置,并确保在cache的文件数据是按顺序排列的. + @endverbatim + * @param page + * @param mapping + * @param pgoff + * @return STATIC + */ +>>>>>>> remotes/origin/main STATIC VOID OsPageCacheAdd(LosFilePage *page, struct page_mapping *mapping, VM_OFFSET_T pgoff) { LosFilePage *fpage = NULL; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) { if (fpage->pgoff > pgoff) { LOS_ListTailInsert(&fpage->node, &page->node); +======= + LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) {//遍历page_list链表 + if (fpage->pgoff > pgoff) {//插入的条件,这样插入保证了按pgoff 从小到大排序 + LOS_ListTailInsert(&fpage->node, &page->node);//等于挂到fpage节点的前面了 +>>>>>>> remotes/origin/main goto done_add; } } +<<<<<<< HEAD LOS_ListTailInsert(&mapping->page_list, &page->node); done_add: mapping->nrpages++; } +======= + LOS_ListTailInsert(&mapping->page_list, &page->node);//将页挂到文件映射的链表上,相当于挂到了最后 + +done_add: + mapping->nrpages++; //文件在缓存中多了一个 文件页 +} +///将页面加到活动文件页LRU链表上 +>>>>>>> remotes/origin/main VOID OsAddToPageacheLru(LosFilePage *page, struct page_mapping *mapping, VM_OFFSET_T pgoff) { OsPageCacheAdd(page, mapping, pgoff); OsLruCacheAdd(page, VM_LRU_ACTIVE_FILE); } +<<<<<<< HEAD VOID OsPageCacheDel(LosFilePage *fpage) { @@ -108,15 +141,44 @@ VOID OsPageCacheDel(LosFilePage *fpage) LOS_MemFree(m_aucSysMem0, fpage); } +======= +///从页高速缓存上删除页 +VOID OsPageCacheDel(LosFilePage *fpage) +{ + /* delete from file cache list */ + LOS_ListDelete(&fpage->node);//将自己从链表上摘除 + fpage->mapping->nrpages--;//文件映射的页总数减少 + + /* unmap and remove map info */ + if (OsIsPageMapped(fpage)) {//是否映射过 + OsUnmapAllLocked(fpage); + } + + LOS_PhysPageFree(fpage->vmPage);//释放物理内存 + + LOS_MemFree(m_aucSysMem0, fpage);//释放文件页结构体内存 +} +/************************************************************************************************** +每个进程都有自己的地址空间, 多个进程可以访问同一个LosFilePage,每个进程使用的虚拟地址都需要单独映射 +所以同一个LosFilePage会映射到多个进程空间.本函数记录页面被哪些进程映射过 +在两个地方会被被空间映射 +1.缺页中断 2.克隆地址空间 +**************************************************************************************************/ +>>>>>>> remotes/origin/main VOID OsAddMapInfo(LosFilePage *page, LosArchMmu *archMmu, VADDR_T vaddr) { LosMapInfo *info = NULL; +<<<<<<< HEAD info = (LosMapInfo *)LOS_MemAlloc(m_aucSysMem0, sizeof(LosMapInfo)); +======= + info = (LosMapInfo *)LOS_MemAlloc(m_aucSysMem0, sizeof(LosMapInfo));//分配一个映射信息 +>>>>>>> remotes/origin/main if (info == NULL) { VM_ERR("OsAddMapInfo alloc memory failed!"); return; } +<<<<<<< HEAD info->page = page; info->archMmu = archMmu; info->vaddr = vaddr; @@ -147,17 +209,57 @@ VOID OsDeletePageCacheLru(LosFilePage *page) OsPageCacheDel(page); } +======= + info->page = page; //文件页 + info->archMmu = archMmu;//进程MMU,完成虚实地址转换 + info->vaddr = vaddr; //虚拟地址 + + LOS_ListAdd(&page->i_mmap, &info->node);//将 LosMapInfo 节点挂入链表 + page->n_maps++;//映射总数++ +} +///通过虚拟地址获取文件页映射信息,archMmu每个进程都有属于自己的mmu +LosMapInfo *OsGetMapInfo(const LosFilePage *page, const LosArchMmu *archMmu, VADDR_T vaddr) +{ + LosMapInfo *info = NULL; + const LOS_DL_LIST *immap = &page->i_mmap;//一个文件页被多个进程映射 + + LOS_DL_LIST_FOR_EACH_ENTRY(info, immap, LosMapInfo, node) {//遍历每个节点 + if ((info->archMmu == archMmu) && (info->vaddr == vaddr) && (info->page == page)) {//全等时返回 + return info; + } + } + return NULL; +} +///删除页高速缓存和LRU,对应 OsAddToPageacheLru +VOID OsDeletePageCacheLru(LosFilePage *page) +{ + /* delete form lru list */ + OsLruCacheDel(page); //将页面从lru列表中删除 + /* delete from cache lits and free pmm if need */ + OsPageCacheDel(page); //从page缓存中删除 +} + +//解除文件页和进程的映射关系 +>>>>>>> remotes/origin/main STATIC VOID OsPageCacheUnmap(LosFilePage *fpage, LosArchMmu *archMmu, VADDR_T vaddr) { UINT32 intSave; LosMapInfo *info = NULL; LOS_SpinLockSave(&fpage->physSeg->lruLock, &intSave); +<<<<<<< HEAD info = OsGetMapInfo(fpage, archMmu, vaddr); if (info == NULL) { VM_ERR("OsPageCacheUnmap get map info failed!"); } else { OsUnmapPageLocked(fpage, info); +======= + info = OsGetMapInfo(fpage, archMmu, vaddr);//获取文件页在进程的映射信息 + if (info == NULL) { + VM_ERR("OsPageCacheUnmap get map info failed!"); + } else { + OsUnmapPageLocked(fpage, info);//解除进程和文件页映射关系 +>>>>>>> remotes/origin/main } if (!(OsIsPageMapped(fpage) && ((fpage->flags & VM_MAP_REGION_FLAG_PERM_EXECUTE) || OsIsPageDirty(fpage->vmPage)))) { @@ -166,7 +268,11 @@ STATIC VOID OsPageCacheUnmap(LosFilePage *fpage, LosArchMmu *archMmu, VADDR_T va LOS_SpinUnlockRestore(&fpage->physSeg->lruLock, intSave); } +<<<<<<< HEAD +======= +///删除文件 +>>>>>>> remotes/origin/main VOID OsVmmFileRemove(LosVmMapRegion *region, LosArchMmu *archMmu, VM_OFFSET_T pgoff) { UINT32 intSave; @@ -179,6 +285,7 @@ VOID OsVmmFileRemove(LosVmMapRegion *region, LosArchMmu *archMmu, VM_OFFSET_T pg LosVmPage *mapPage = NULL; if (!LOS_IsRegionFileValid(region) || (region->unTypeData.rf.vnode == NULL)) { +<<<<<<< HEAD return; } vnode = region->unTypeData.rf.vnode; @@ -186,10 +293,20 @@ VOID OsVmmFileRemove(LosVmMapRegion *region, LosArchMmu *archMmu, VM_OFFSET_T pg vaddr = region->range.base + ((UINT32)(pgoff - region->pgOff) << PAGE_SHIFT); status_t status = LOS_ArchMmuQuery(archMmu, vaddr, &paddr, NULL); +======= + return;//判断是否为文件映射,是否已map + } + vnode = region->unTypeData.rf.vnode; + mapping = &vnode->mapping; + vaddr = region->range.base + ((UINT32)(pgoff - region->pgOff) << PAGE_SHIFT);//得到虚拟地址 + + status_t status = LOS_ArchMmuQuery(archMmu, vaddr, &paddr, NULL);//获取物理地址 +>>>>>>> remotes/origin/main if (status != LOS_OK) { return; } +<<<<<<< HEAD mapPage = LOS_VmPageGet(paddr); /* is page is in cache list */ @@ -204,6 +321,22 @@ VOID OsVmmFileRemove(LosVmMapRegion *region, LosArchMmu *archMmu, VM_OFFSET_T pg OsPageCacheUnmap(fpage, archMmu, vaddr); if (OsIsPageDirty(fpage->vmPage)) { tmpPage = OsDumpDirtyPage(fpage); +======= + mapPage = LOS_VmPageGet(paddr);//获取物理页框 + + /* is page is in cache list */ + LOS_SpinLockSave(&mapping->list_lock, &intSave); + fpage = OsFindGetEntry(mapping, pgoff);//获取fpage + /* no cache or have cache but not map(cow), free it direct */ + if ((fpage == NULL) || (fpage->vmPage != mapPage)) {//没有缓存或有缓存但没有映射(cow),直接释放它 + LOS_PhysPageFree(mapPage);//释放物理页框 + LOS_ArchMmuUnmap(archMmu, vaddr, 1);//取消虚拟地址的映射 + /* this is a page cache map! */ + } else { + OsPageCacheUnmap(fpage, archMmu, vaddr);////取消缓存中的映射 + if (OsIsPageDirty(fpage->vmPage)) {//脏页处理 + tmpPage = OsDumpDirtyPage(fpage);//dump 脏页 +>>>>>>> remotes/origin/main } } LOS_SpinUnlockRestore(&mapping->list_lock, intSave); @@ -213,6 +346,7 @@ VOID OsVmmFileRemove(LosVmMapRegion *region, LosArchMmu *archMmu, VM_OFFSET_T pg } return; } +<<<<<<< HEAD VOID OsMarkPageDirty(LosFilePage *fpage, const LosVmMapRegion *region, INT32 off, INT32 len) { @@ -222,6 +356,17 @@ VOID OsMarkPageDirty(LosFilePage *fpage, const LosVmMapRegion *region, INT32 off fpage->dirtyEnd = len; } else { OsSetPageDirty(fpage->vmPage); +======= +///标记page为脏页 进程修改了高速缓存里的数据时,该页就被内核标记为脏页 +VOID OsMarkPageDirty(LosFilePage *fpage, const LosVmMapRegion *region, INT32 off, INT32 len) +{ + if (region != NULL) { + OsSetPageDirty(fpage->vmPage);//设置为脏页 + fpage->dirtyOff = off;//脏页偏移位置 + fpage->dirtyEnd = len;//脏页结束位置 + } else { + OsSetPageDirty(fpage->vmPage);//设置为脏页 +>>>>>>> remotes/origin/main if ((off + len) > fpage->dirtyEnd) { fpage->dirtyEnd = off + len; } @@ -258,22 +403,37 @@ STATIC UINT32 GetDirtySize(LosFilePage *fpage, struct Vnode *vnode) return PAGE_SIZE; } +<<<<<<< HEAD +======= +///冲洗脏页,回写磁盘 +>>>>>>> remotes/origin/main STATIC INT32 OsFlushDirtyPage(LosFilePage *fpage) { UINT32 ret; size_t len; char *buff = NULL; +<<<<<<< HEAD struct Vnode *vnode = fpage->mapping->host; +======= + struct Vnode *vnode = fpage->mapping->host;/* owner of this mapping */ //此映射属于哪个文件,注意是1:1的关系. +>>>>>>> remotes/origin/main if (vnode == NULL) { VM_ERR("page cache vnode error"); return LOS_NOK; } +<<<<<<< HEAD len = fpage->dirtyEnd - fpage->dirtyOff; len = (len == 0) ? GetDirtySize(fpage, vnode) : len; if (len == 0) { OsCleanPageDirty(fpage->vmPage); +======= + len = fpage->dirtyEnd - fpage->dirtyOff;//计算出脏数据长度 + len = (len == 0) ? GetDirtySize(fpage, vnode) : len; + if (len == 0) {//没有脏数据 + OsCleanPageDirty(fpage->vmPage);//页面取消脏标签 +>>>>>>> remotes/origin/main return LOS_OK; } @@ -290,7 +450,11 @@ STATIC INT32 OsFlushDirtyPage(LosFilePage *fpage) return ret; } +<<<<<<< HEAD +======= +///备份脏页,老脏页撕掉脏页标签 +>>>>>>> remotes/origin/main LosFilePage *OsDumpDirtyPage(LosFilePage *oldFPage) { LosFilePage *newFPage = NULL; @@ -302,11 +466,19 @@ LosFilePage *OsDumpDirtyPage(LosFilePage *oldFPage) } OsCleanPageDirty(oldFPage->vmPage); +<<<<<<< HEAD (VOID)memcpy_s(newFPage, sizeof(LosFilePage), oldFPage, sizeof(LosFilePage)); return newFPage; } +======= + (VOID)memcpy_s(newFPage, sizeof(LosFilePage), oldFPage, sizeof(LosFilePage));//直接内存拷贝 + + return newFPage; +} +///冲洗脏页数据,将脏页数据回写磁盘 +>>>>>>> remotes/origin/main VOID OsDoFlushDirtyPage(LosFilePage *fpage) { if (fpage == NULL) { @@ -328,7 +500,11 @@ STATIC VOID OsReleaseFpage(struct page_mapping *mapping, LosFilePage *fpage) LOS_SpinUnlockRestore(lruLock, lruSave); LOS_SpinUnlockRestore(&mapping->list_lock, intSave); } +<<<<<<< HEAD +======= +///删除映射信息 +>>>>>>> remotes/origin/main VOID OsDelMapInfo(LosVmMapRegion *region, LosVmPgFault *vmf, BOOL cleanDirty) { UINT32 intSave; @@ -349,9 +525,15 @@ VOID OsDelMapInfo(LosVmMapRegion *region, LosVmPgFault *vmf, BOOL cleanDirty) } if (cleanDirty) { +<<<<<<< HEAD OsCleanPageDirty(fpage->vmPage); } info = OsGetMapInfo(fpage, ®ion->space->archMmu, (vaddr_t)vmf->vaddr); +======= + OsCleanPageDirty(fpage->vmPage);//恢复干净页 + } + info = OsGetMapInfo(fpage, ®ion->space->archMmu, (vaddr_t)vmf->vaddr);//通过虚拟地址获取映射信息 +>>>>>>> remotes/origin/main if (info != NULL) { fpage->n_maps--; LOS_ListDelete(&info->node); @@ -362,7 +544,14 @@ VOID OsDelMapInfo(LosVmMapRegion *region, LosVmPgFault *vmf, BOOL cleanDirty) } LOS_SpinUnlockRestore(&mapping->list_lock, intSave); } - +<<<<<<< HEAD + +======= +/*! +文件缺页时的处理,先读入磁盘数据,再重新读页数据 +被 OsDoReadFault(...),OsDoCowFault(...),OsDoSharedFault(...) 等调用 +*/ +>>>>>>> remotes/origin/main INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) { INT32 ret; @@ -374,7 +563,11 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) struct page_mapping *mapping = NULL; LosFilePage *fpage = NULL; +<<<<<<< HEAD if (!LOS_IsRegionFileValid(region) || (region->unTypeData.rf.vnode == NULL) || (vmf == NULL)) { +======= + if (!LOS_IsRegionFileValid(region) || (region->unTypeData.rf.vnode == NULL) || (vmf == NULL)) {//文件是否映射到了内存 +>>>>>>> remotes/origin/main VM_ERR("Input param is NULL"); return LOS_NOK; } @@ -383,6 +576,7 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) /* get or create a new cache node */ LOS_SpinLockSave(&mapping->list_lock, &intSave); +<<<<<<< HEAD fpage = OsFindGetEntry(mapping, vmf->pgoff); TRACE_TRY_CACHE(); if (fpage != NULL) { @@ -390,11 +584,21 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) OsPageRefIncLocked(fpage); } else { fpage = OsPageCacheAlloc(mapping, vmf->pgoff); +======= + fpage = OsFindGetEntry(mapping, vmf->pgoff);//获取文件页 + TRACE_TRY_CACHE(); + if (fpage != NULL) {//找到了,说明该页已经在页高速缓存中 + TRACE_HIT_CACHE(); + OsPageRefIncLocked(fpage); + } else {//真的缺页了,页高速缓存中没找到 + fpage = OsPageCacheAlloc(mapping, vmf->pgoff);//分配一个文件页,将数据初始化好,包括vmpage(物理页框) +>>>>>>> remotes/origin/main if (fpage == NULL) { LOS_SpinUnlockRestore(&mapping->list_lock, intSave); VM_ERR("Failed to alloc a page frame"); return LOS_NOK; } +<<<<<<< HEAD newCache = true; } OsSetPageLocked(fpage->vmPage); @@ -403,6 +607,16 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) /* read file to new page cache */ if (newCache) { +======= + newCache = true;//分配了新文件页 + } + OsSetPageLocked(fpage->vmPage);//对vmpage上锁 + LOS_SpinUnlockRestore(&mapping->list_lock, intSave); + kvaddr = OsVmPageToVaddr(fpage->vmPage);//获取该页框在内核空间的虚拟地址,因为 page cache本身就是在内核空间, + + /* read file to new page cache */ + if (newCache) {//新cache +>>>>>>> remotes/origin/main ret = vnode->vop->ReadPage(vnode, kvaddr, fpage->pgoff << PAGE_SHIFT); if (ret == 0) { VM_ERR("Failed to read from file!"); @@ -410,18 +624,27 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) return LOS_NOK; } LOS_SpinLockSave(&mapping->list_lock, &intSave); +<<<<<<< HEAD OsAddToPageacheLru(fpage, mapping, vmf->pgoff); +======= + OsAddToPageacheLru(fpage, mapping, vmf->pgoff);//将fpage挂入pageCache 和 LruCache +>>>>>>> remotes/origin/main LOS_SpinUnlockRestore(&mapping->list_lock, intSave); } LOS_SpinLockSave(&mapping->list_lock, &intSave); /* cow fault case no need to save mapinfo */ if (!((vmf->flags & VM_MAP_PF_FLAG_WRITE) && !(region->regionFlags & VM_MAP_REGION_FLAG_SHARED))) { +<<<<<<< HEAD OsAddMapInfo(fpage, ®ion->space->archMmu, (vaddr_t)vmf->vaddr); +======= + OsAddMapInfo(fpage, ®ion->space->archMmu, (vaddr_t)vmf->vaddr);//添加<虚拟地址,文件页>的映射关系,如此进程以后就能通过虚拟地址操作文件页了. +>>>>>>> remotes/origin/main fpage->flags = region->regionFlags; } /* share page fault, mark the page dirty */ +<<<<<<< HEAD if ((vmf->flags & VM_MAP_PF_FLAG_WRITE) && (region->regionFlags & VM_MAP_REGION_FLAG_SHARED)) { OsMarkPageDirty(fpage, region, 0, 0); } @@ -431,11 +654,26 @@ INT32 OsVmmFileFault(LosVmMapRegion *region, LosVmPgFault *vmf) return LOS_OK; } +======= + if ((vmf->flags & VM_MAP_PF_FLAG_WRITE) && (region->regionFlags & VM_MAP_REGION_FLAG_SHARED)) {//有过写操作或者为共享线性区 + OsMarkPageDirty(fpage, region, 0, 0);//标记为脏页,要回写磁盘,内核会在适当的时候回写磁盘 + } + + vmf->pageKVaddr = kvaddr;//缺陷页记录文件页的虚拟地址 + LOS_SpinUnlockRestore(&mapping->list_lock, intSave); + return LOS_OK; +} +///文件缓存冲洗,把所有fpage冲洗一边,把脏页洗到dirtyList中,配合OsFileCacheRemove理解 +>>>>>>> remotes/origin/main VOID OsFileCacheFlush(struct page_mapping *mapping) { UINT32 intSave; UINT32 lruLock; +<<<<<<< HEAD LOS_DL_LIST_HEAD(dirtyList); +======= + LOS_DL_LIST_HEAD(dirtyList);//LOS_DL_LIST list = { &(list), &(list) }; +>>>>>>> remotes/origin/main LosFilePage *ftemp = NULL; LosFilePage *fpage = NULL; @@ -443,33 +681,59 @@ VOID OsFileCacheFlush(struct page_mapping *mapping) return; } LOS_SpinLockSave(&mapping->list_lock, &intSave); +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) { LOS_SpinLockSave(&fpage->physSeg->lruLock, &lruLock); if (OsIsPageDirty(fpage->vmPage)) { ftemp = OsDumpDirtyPage(fpage); if (ftemp != NULL) { LOS_ListTailInsert(&dirtyList, &ftemp->node); +======= + LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) {//循环从page_list中取node给fpage + LOS_SpinLockSave(&fpage->physSeg->lruLock, &lruLock); + if (OsIsPageDirty(fpage->vmPage)) {//是否为脏页 + ftemp = OsDumpDirtyPage(fpage);//这里挺妙的,copy出一份新页,老页变成了非脏页继续用 + if (ftemp != NULL) { + LOS_ListTailInsert(&dirtyList, &ftemp->node);//将新页插入脏页List,等待回写磁盘 +>>>>>>> remotes/origin/main } } LOS_SpinUnlockRestore(&fpage->physSeg->lruLock, lruLock); } LOS_SpinUnlockRestore(&mapping->list_lock, intSave); +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, ftemp, &dirtyList, LosFilePage, node) { OsDoFlushDirtyPage(fpage); } } +======= + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, ftemp, &dirtyList, LosFilePage, node) {//仔细看这个宏,关键在 &(item)->member != (list); + OsDoFlushDirtyPage(fpage);//立马洗掉,所以dirtyList可以不是全局变量 + } +} + +/****************************************************************************** + 删除文件缓存,清空文件在page cache的痕迹 + 参数 mapping 可理解为文件在内存的身份证 +******************************************************************************/ +>>>>>>> remotes/origin/main VOID OsFileCacheRemove(struct page_mapping *mapping) { UINT32 intSave; UINT32 lruSave; SPIN_LOCK_S *lruLock = NULL; +<<<<<<< HEAD LOS_DL_LIST_HEAD(dirtyList); +======= + LOS_DL_LIST_HEAD(dirtyList);//定义一个叫dirtyList的双循环链表并初始化,用于挂脏页 +>>>>>>> remotes/origin/main LosFilePage *ftemp = NULL; LosFilePage *fpage = NULL; LosFilePage *fnext = NULL; +<<<<<<< HEAD LOS_SpinLockSave(&mapping->list_lock, &intSave); LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, &mapping->page_list, LosFilePage, node) { lruLock = &fpage->physSeg->lruLock; @@ -501,12 +765,52 @@ LosVmFileOps g_commVmOps = { INT32 OsVfsFileMmap(struct file *filep, LosVmMapRegion *region) { region->unTypeData.rf.vmFOps = &g_commVmOps; +======= + LOS_SpinLockSave(&mapping->list_lock, &intSave);//多进程操作,必须上锁. + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, &mapping->page_list, LosFilePage, node) {//遍历文件在内存中产生的所有文件页(例如1,4,8页)不一定连续,取决于用户的读取顺序 + lruLock = &fpage->physSeg->lruLock; + LOS_SpinLockSave(lruLock, &lruSave);//@note_why 自旋锁有必要从这里开始上锁吗? + if (OsIsPageDirty(fpage->vmPage)) {//数据是脏页吗,脏页就是被修改过数据的页 + ftemp = OsDumpDirtyPage(fpage);//做这个拷贝动作是为了fpage的统一下线,因为数据回写磁盘的速度是很慢的,如果直接在这里处理脏数据 + if (ftemp != NULL) {//会导致函数持有mapping->list_lock自旋锁的时间太长了,影响其他CPU的处理效率 + LOS_ListTailInsert(&dirtyList, &ftemp->node);//将临时脏页挂到记录脏页链表上 + } + } + + OsDeletePageCacheLru(fpage);//删除高速缓存和从置换链表中下线 + LOS_SpinUnlockRestore(lruLock, lruSave); + } + LOS_SpinUnlockRestore(&mapping->list_lock, intSave);//恢复自旋锁,不能让别的CPU等太久 + + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, &dirtyList, LosFilePage, node) {//到这里,再来慢慢的统一处理脏页数据 + OsDoFlushDirtyPage(fpage);//遍历脏页链表,一页一页处理 + } +} +///虚拟内存文件操作实现类 +LosVmFileOps g_commVmOps = {// + .open = NULL, + .close = NULL, + .fault = OsVmmFileFault, //缺页中断处理 + .remove = OsVmmFileRemove,//删除页 +}; +//文件映射 +INT32 OsVfsFileMmap(struct file *filep, LosVmMapRegion *region) +{ + region->unTypeData.rf.vmFOps = &g_commVmOps;//文件操作 +>>>>>>> remotes/origin/main region->unTypeData.rf.vnode = filep->f_vnode; region->unTypeData.rf.f_oflags = filep->f_oflags; return ENOERR; } - +<<<<<<< HEAD + +======= +/*! + 有名映射,可理解为文件映射,跟匿名映射相对应 + 参数filep是广义的文件,在鸿蒙内核,目录/普通文件/字符设备/块设备/网络套接字/管道/链接 都是文件 +*/ +>>>>>>> remotes/origin/main STATUS_T OsNamedMMap(struct file *filep, LosVmMapRegion *region) { struct Vnode *vnode = NULL; @@ -519,10 +823,17 @@ STATUS_T OsNamedMMap(struct file *filep, LosVmMapRegion *region) vnode->useCount++; VnodeDrop(); if (filep->ops != NULL && filep->ops->mmap != NULL) { +<<<<<<< HEAD if (vnode->type == VNODE_TYPE_CHR || vnode->type == VNODE_TYPE_BLK) { LOS_SetRegionTypeDev(region); } else { LOS_SetRegionTypeFile(region); +======= + if (vnode->type == VNODE_TYPE_CHR || vnode->type == VNODE_TYPE_BLK) {//块设备或者字符设备 /dev/.. + LOS_SetRegionTypeDev(region);//设置为设备类型 + } else { + LOS_SetRegionTypeFile(region);//设置为文件类型 +>>>>>>> remotes/origin/main } int ret = filep->ops->mmap(filep, region); if (ret != LOS_OK) { @@ -538,10 +849,18 @@ STATUS_T OsNamedMMap(struct file *filep, LosVmMapRegion *region) return LOS_OK; } +<<<<<<< HEAD +======= +/************************************************************************************************** + 通过位置从文件映射页中找到的指定的文件页 + 举例:mapping->page_list上节点的数据可能只有是文件 1,3,4,6 页的数据,此时来找文件第5页的数据就会没有 +**************************************************************************************************/ +>>>>>>> remotes/origin/main LosFilePage *OsFindGetEntry(struct page_mapping *mapping, VM_OFFSET_T pgoff) { LosFilePage *fpage = NULL; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) { if (fpage->pgoff == pgoff) { return fpage; @@ -549,6 +868,15 @@ LosFilePage *OsFindGetEntry(struct page_mapping *mapping, VM_OFFSET_T pgoff) if (fpage->pgoff > pgoff) { break; +======= + LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) {//遍历文件页 + if (fpage->pgoff == pgoff) {//找到指定的页, + return fpage; + } + + if (fpage->pgoff > pgoff) {//大于之前还没有找到,说明不在链表中,往后的也不用找了, + break;//因为 mapping->page_list节点上的数据都是按 fpage->pgoff 从小到大的顺序排列的. +>>>>>>> remotes/origin/main } } @@ -556,6 +884,14 @@ LosFilePage *OsFindGetEntry(struct page_mapping *mapping, VM_OFFSET_T pgoff) } /* need mutex & change memory to dma zone. */ +<<<<<<< HEAD +======= +/*! +以页高速缓存方式分配一个文件页 LosFilePage + Direct Memory Access(存储器直接访问)指一种高速的数据传输操作,允许在外部设备和存储器之间直接读写数据。 + 整个数据传输操作在一个称为"DMA控制器"的控制下进行的。CPU只需在数据传输开始和结束时做一点处理(开始和结束时候要做中断处理) +*/ +>>>>>>> remotes/origin/main LosFilePage *OsPageCacheAlloc(struct page_mapping *mapping, VM_OFFSET_T pgoff) { VOID *kvaddr = NULL; @@ -563,26 +899,44 @@ LosFilePage *OsPageCacheAlloc(struct page_mapping *mapping, VM_OFFSET_T pgoff) LosVmPage *vmPage = NULL; LosFilePage *fpage = NULL; +<<<<<<< HEAD vmPage = LOS_PhysPageAlloc(); +======= + vmPage = LOS_PhysPageAlloc(); //先分配一个物理页 +>>>>>>> remotes/origin/main if (vmPage == NULL) { VM_ERR("alloc vm page failed"); return NULL; } +<<<<<<< HEAD physSeg = OsVmPhysSegGet(vmPage); kvaddr = OsVmPageToVaddr(vmPage); if ((physSeg == NULL) || (kvaddr == NULL)) { LOS_PhysPageFree(vmPage); +======= + physSeg = OsVmPhysSegGet(vmPage);//通过页获取所在seg + kvaddr = OsVmPageToVaddr(vmPage);//获取内核空间的虚拟地址,具体点进去看函数说明,这里一定要理解透彻! + if ((physSeg == NULL) || (kvaddr == NULL)) { + LOS_PhysPageFree(vmPage); //异常情况要释放vmPage +>>>>>>> remotes/origin/main VM_ERR("alloc vm page failed!"); return NULL; } +<<<<<<< HEAD fpage = (LosFilePage *)LOS_MemAlloc(m_aucSysMem0, sizeof(LosFilePage)); if (fpage == NULL) { LOS_PhysPageFree(vmPage); +======= + fpage = (LosFilePage *)LOS_MemAlloc(m_aucSysMem0, sizeof(LosFilePage));//从内存池中分配一个filePage + if (fpage == NULL) { + LOS_PhysPageFree(vmPage); //异常情况要释放vmPage +>>>>>>> remotes/origin/main VM_ERR("Failed to allocate for page!"); return NULL; } +<<<<<<< HEAD (VOID)memset_s((VOID *)fpage, sizeof(LosFilePage), 0, sizeof(LosFilePage)); LOS_ListInit(&fpage->i_mmap); @@ -596,6 +950,21 @@ LosFilePage *OsPageCacheAlloc(struct page_mapping *mapping, VM_OFFSET_T pgoff) fpage->mapping = mapping; fpage->pgoff = pgoff; (VOID)memset_s(kvaddr, PAGE_SIZE, 0, PAGE_SIZE); +======= + (VOID)memset_s((VOID *)fpage, sizeof(LosFilePage), 0, sizeof(LosFilePage));//调标准库函数 置0 + + LOS_ListInit(&fpage->i_mmap); //初始化映射,链表上挂 MapInfo + LOS_ListInit(&fpage->node); //节点初始化 + LOS_ListInit(&fpage->lru); //LRU初始化 + fpage->n_maps = 0; //映射次数 + fpage->dirtyOff = PAGE_SIZE; //默认页尾部,相当于没有脏数据 + fpage->dirtyEnd = 0; //脏页结束位置 + fpage->physSeg = physSeg; //页框所属段.其中包含了 LRU LIST == + fpage->vmPage = vmPage; //物理页框 + fpage->mapping = mapping; //记录所有文件页映射 + fpage->pgoff = pgoff; //将文件切成一页页,页标 + (VOID)memset_s(kvaddr, PAGE_SIZE, 0, PAGE_SIZE);//页内数据清0 +>>>>>>> remotes/origin/main return fpage; } @@ -609,4 +978,8 @@ INT32 OsVfsFileMmap(struct file *filep, LosVmMapRegion *region) } #endif +<<<<<<< HEAD +#endif +======= #endif +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_iomap.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_iomap.c index a313da14..bb521607 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_iomap.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_iomap.c @@ -1,4 +1,46 @@ /* +<<<<<<< HEAD +======= + 直接内存访问 + 直接内存访问(Direct Memory Access,DMA)是计算机科学中的一种内存访问技术。它允许某些电脑内部的 + 硬件子系统(电脑外设),可以独立地直接读写系统内存,而不需中央处理器(CPU)介入处理 + 在同等程度的处理器负担下,DMA是一种快速的数据传送方式。很多硬件的系统会使用DMA,包含硬盘控制器、 + 绘图显卡、网卡和声卡。 + + DMA是所有现代电脑的重要特色,它允许不同速度的硬件设备来沟通,而不需要依于中央处理器的大量中断负载。 + 否则,中央处理器需要从来源把每一片段的资料复制到寄存器,然后把它们再次写回到新的地方。在这个时间中, + 中央处理器对于其他的工作来说就无法使用。 + + DMA传输常使用在将一个内存区从一个设备复制到另外一个。当中央处理器初始化这个传输动作,传输动作本身是 + 由DMA控制器来实行和完成。典型的例子就是移动一个外部内存的区块到芯片内部更快的内存去。像是这样的操作 + 并没有让处理器工作拖延,使其可以被重新调度去处理其他的工作。DMA传输对于高性能嵌入式系统算法和网络是 + 很重要的。 举个例子,个人电脑的ISA DMA控制器拥有8个DMA通道,其中的7个通道是可以让计算机的中央处理器所利用。 + 每一个DMA通道有一个16位地址寄存器和一个16位计数寄存器。要初始化资料传输时,设备驱动程序一起设置DMA通道的 + 地址和计数寄存器,以及资料传输的方向,读取或写入。然后指示DMA硬件开始这个传输动作。当传输结束的时候, + 设备就会以中断的方式通知中央处理器。 + + "分散-收集"(Scatter-gather)DMA允许在一次单一的DMA处理中传输资料到多个内存区域。相当于把多个简单的DMA要求 + 串在一起。同样,这样做的目的是要减轻中央处理器的多次输出输入中断和资料复制任务。 + DRQ意为DMA要求;DACK意为DMA确认。这些符号一般在有DMA功能的电脑系统硬件概要上可以看到。 + 它们表示了介于中央处理器和DMA控制器之间的电子信号传输线路。 + + 缓存一致性问题 + DMA会导致缓存一致性问题。想像中央处理器带有缓存与外部内存的情况,DMA的运作则是去访问外部内存, + 当中央处理器访问外部内存某个地址的时候,暂时先将新的值写入缓存中,但并未将外部内存的资料更新, + 若在缓存中的资料尚未更新到外部内存前发生了DMA,则DMA过程将会读取到未更新的资料。 + 相同的,如果外部设备写入新的值到外部内存内,则中央处理器若访问缓存时则会访问到尚未更新的资料。 + 这些问题可以用两种方法来解决: + + 缓存同调系统(Cache-coherent system):以硬件方法来完成,当外部设备写入内存时以一个信号来通知 + 缓存控制器某内存地址的值已经过期或是应该更新资料。 + 非同调系统(Non-coherent system):以软件方法来完成,操作系统必须确认缓存读取时,DMA程序已经 + 开始或是禁止DMA发生。 + 第二种的方法会造成DMA的系统负担。 + */ + + +/* +>>>>>>> remotes/origin/main * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. * @@ -36,7 +78,11 @@ #include "los_vm_map.h" #include "los_memory.h" +<<<<<<< HEAD +======= +///分配DMA空间 +>>>>>>> remotes/origin/main VOID *LOS_DmaMemAlloc(DMA_ADDR_T *dmaAddr, size_t size, size_t align, enum DmaMemType type) { VOID *kVaddr = NULL; @@ -53,24 +99,42 @@ VOID *LOS_DmaMemAlloc(DMA_ADDR_T *dmaAddr, size_t size, size_t align, enum DmaMe #ifdef LOSCFG_KERNEL_VM kVaddr = LOS_KernelMallocAlign(size, align); #else +<<<<<<< HEAD kVaddr = LOS_MemAllocAlign(OS_SYS_MEM_ADDR, size, align); #endif if (kVaddr == NULL) { VM_ERR("failed, size = %u, align = %u", size, align); +======= + kVaddr = LOS_MemAllocAlign(OS_SYS_MEM_ADDR, size, align);//不走内存池方式, 直接申请物理页 +#endif + if (kVaddr == NULL) { + VM_ERR("failed, size = %u, align = %u", size, align);//从内存池中申请 +>>>>>>> remotes/origin/main return NULL; } if (dmaAddr != NULL) { +<<<<<<< HEAD *dmaAddr = (DMA_ADDR_T)LOS_PaddrQuery(kVaddr); } if (type == DMA_NOCACHE) { +======= + *dmaAddr = (DMA_ADDR_T)LOS_PaddrQuery(kVaddr);//查询物理地址, DMA直接将数据灌到物理地址 + } + + if (type == DMA_NOCACHE) {//无缓存模式 , 计算新的虚拟地址 +>>>>>>> remotes/origin/main kVaddr = (VOID *)VMM_TO_UNCACHED_ADDR((UINTPTR)kVaddr); } return kVaddr; } +<<<<<<< HEAD +======= +///释放DMA指针 +>>>>>>> remotes/origin/main VOID LOS_DmaMemFree(VOID *vaddr) { UINTPTR addr; @@ -79,6 +143,7 @@ VOID LOS_DmaMemFree(VOID *vaddr) return; } addr = (UINTPTR)vaddr; +<<<<<<< HEAD if ((addr >= UNCACHED_VMM_BASE) && (addr < UNCACHED_VMM_BASE + UNCACHED_VMM_SIZE)) { addr = UNCACHED_TO_VMM_ADDR(addr); @@ -86,6 +151,15 @@ VOID LOS_DmaMemFree(VOID *vaddr) LOS_KernelFree((VOID *)addr); #else LOS_MemFree(OS_SYS_MEM_ADDR, (VOID *)addr); +======= + // 未缓存区 + if ((addr >= UNCACHED_VMM_BASE) && (addr < UNCACHED_VMM_BASE + UNCACHED_VMM_SIZE)) { + addr = UNCACHED_TO_VMM_ADDR(addr);//转换成未缓存区地址 +#ifdef LOSCFG_KERNEL_VM + LOS_KernelFree((VOID *)addr); +#else + LOS_MemFree(OS_SYS_MEM_ADDR, (VOID *)addr);//内存池方式释放 +>>>>>>> remotes/origin/main #endif } else if ((addr >= KERNEL_VMM_BASE) && (addr < KERNEL_VMM_BASE + KERNEL_VMM_SIZE)) { #ifdef LOSCFG_KERNEL_VM diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_map.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_map.c index 5d4f62b3..21b8809e 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_map.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_map.c @@ -1,4 +1,47 @@ /* +<<<<<<< HEAD +======= +基本概念 + 虚拟内存管理是计算机系统管理内存的一种技术。每个进程都有连续的虚拟地址空间,虚拟地址空间的大小由CPU的位数决定, + 32位的硬件平台可以提供的最大的寻址空间为0-4GiB。整个4GiB空间分成两部分,LiteOS-A内核占据3GiB的高地址空间, + 1GiB的低地址空间留给进程使用。各个进程空间的虚拟地址空间是独立的,代码、数据互不影响。 + + 系统将虚拟内存分割为称为虚拟页的内存块,大小一般为4KiB或64KiB,LiteOS-A内核默认的页的大小是4KiB, + 根据需要可以对MMU(Memory Management Units)进行配置。虚拟内存管理操作的最小单位就是一个页, + LiteOS-A内核中一个虚拟地址区间region包含地址连续的多个虚拟页,也可只有一个页。同样,物理内存也会按照页大小进行分割, + 分割后的每个内存块称为页帧。虚拟地址空间划分:内核态占高地址3GiB(0x40000000 ~ 0xFFFFFFFF), + 用户态占低地址1GiB(0x01000000 ~ 0x3F000000),具体见下表,详细可以查看或配置los_vm_zone.h。 + +内核态地址规划: + Zone名称 描述 属性 + ---------------------------------------------------------------------------- + DMA zone 供IO设备的DMA使用。 Uncache + + Normal zone 加载内核代码段、数据段、堆和栈的地址区间。 Cache + + high mem zone可以分配连续的虚拟内存,但其所映射的物理内存不一定连续。Cache + +用户态地址规划: + Zone名称 描述 属性 + ---------------------------------------------------------------------------- + 代码段 用户态代码段地址区间。 Cache + 堆 用户态堆地址区间。 Cache + 栈 用户态栈地址区间。 Cache + 共享库 用于加载用户态共享库的地址区间,包括mmap所映射的区间。 Cache + +运行机制 + 虚拟内存管理中,虚拟地址空间是连续的,但是其映射的物理内存并不一定是连续的,如下图所示。 + 可执行程序加载运行,CPU访问虚拟地址空间的代码或数据时存在两种情况: + + 1. CPU访问的虚拟地址所在的页,如V0,已经与具体的物理页P0做映射,CPU通过找到进程对应的页表条目(详见虚实映射), + 根据页表条目中的物理地址信息访问物理内存中的内容并返回。 + 2. CPU访问的虚拟地址所在的页,如V2,没有与具体的物理页做映射,系统会触发缺页异常,系统申请一个物理页, + 并把相应的信息拷贝到物理页中,并且把物理页的起始地址更新到页表条目中。此时CPU重新执行访问虚拟内存的指令 + 便能够访问到具体的代码或数据。 + */ + +/* +>>>>>>> remotes/origin/main * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. * @@ -51,16 +94,30 @@ #ifdef LOSCFG_KERNEL_VM +<<<<<<< HEAD #define VM_MAP_WASTE_MEM_LEVEL (PAGE_SIZE >> 2) LosMux g_vmSpaceListMux; LOS_DL_LIST_HEAD(g_vmSpaceList); LosVmSpace g_kVmSpace; LosVmSpace g_vMallocSpace; +======= +#define VM_MAP_WASTE_MEM_LEVEL (PAGE_SIZE >> 2) ///< 浪费内存大小(1K) +LosMux g_vmSpaceListMux; ///< 用于锁g_vmSpaceList的互斥量 +LOS_DL_LIST_HEAD(g_vmSpaceList); ///< 初始化全局虚拟空间节点,所有虚拟空间都挂到此节点上. +LosVmSpace g_kVmSpace; ///< 内核非分配空间,用于内核运行栈,代码区,数据区 +LosVmSpace g_vMallocSpace; ///< 内核分配空间,用于内核分配内存 + +/************************************************************ +* 获取进程空间系列接口 +************************************************************/ +/// 获取当前进程空间结构体指针 +>>>>>>> remotes/origin/main LosVmSpace *LOS_CurrSpaceGet(VOID) { return OsCurrProcessGet()->vmSpace; } +<<<<<<< HEAD LosVmSpace *LOS_SpaceGet(VADDR_T vaddr) { @@ -70,38 +127,76 @@ LosVmSpace *LOS_SpaceGet(VADDR_T vaddr) return LOS_CurrSpaceGet(); } else if (LOS_IsVmallocAddress(vaddr)) { return LOS_GetVmallocSpace(); +======= +/// 获取虚拟地址对应的进程空间结构体指针 +LosVmSpace *LOS_SpaceGet(VADDR_T vaddr) +{ + if (LOS_IsKernelAddress(vaddr)) { //是否为内核空间 + return LOS_GetKVmSpace(); //获取内核空间 + } else if (LOS_IsUserAddress(vaddr)) {//是否为用户空间 + return LOS_CurrSpaceGet(); + } else if (LOS_IsVmallocAddress(vaddr)) {//是否为内核分配空间 + return LOS_GetVmallocSpace();//获取内核分配空间 +>>>>>>> remotes/origin/main } else { return NULL; } } +<<<<<<< HEAD +======= +///内核空间只有g_kVmSpace一个,所有的内核进程都共用一个内核空间 +>>>>>>> remotes/origin/main LosVmSpace *LOS_GetKVmSpace(VOID) { return &g_kVmSpace; } +<<<<<<< HEAD +======= +///获取进程空间链表指针 g_vmSpaceList中挂的是进程空间 g_kVmSpace, g_vMallocSpace,所有用户进程的空间(独有一个进程空间) +>>>>>>> remotes/origin/main LOS_DL_LIST *LOS_GetVmSpaceList(VOID) { return &g_vmSpaceList; } +<<<<<<< HEAD +======= +///获取内核堆空间的全局变量 +>>>>>>> remotes/origin/main LosVmSpace *LOS_GetVmallocSpace(VOID) { return &g_vMallocSpace; } +<<<<<<< HEAD +======= +/************************************************************ +* 虚拟地址区间region相关的操作 +************************************************************/ +///释放挂在红黑树上节点,等于释放了线性区 +>>>>>>> remotes/origin/main ULONG_T OsRegionRbFreeFn(LosRbNode *pstNode) { LOS_MemFree(m_aucSysMem0, pstNode); return LOS_OK; } +<<<<<<< HEAD +======= +///通过红黑树节点找到对应的线性区 +>>>>>>> remotes/origin/main VOID *OsRegionRbGetKeyFn(LosRbNode *pstNode) { LosVmMapRegion *region = (LosVmMapRegion *)LOS_DL_LIST_ENTRY(pstNode, LosVmMapRegion, rbNode); return (VOID *)®ion->range; } +<<<<<<< HEAD +======= +///比较两个红黑树节点 +>>>>>>> remotes/origin/main ULONG_T OsRegionRbCmpKeyFn(const VOID *pNodeKeyA, const VOID *pNodeKeyB) { LosVmMapRange rangeA = *(LosVmMapRange *)pNodeKeyA; @@ -111,6 +206,7 @@ ULONG_T OsRegionRbCmpKeyFn(const VOID *pNodeKeyA, const VOID *pNodeKeyB) UINT32 startB = rangeB.base; UINT32 endB = rangeB.base + rangeB.size - 1; +<<<<<<< HEAD if (startA > endB) { return RB_BIGGER; } else if (startA >= startB) { @@ -127,21 +223,57 @@ ULONG_T OsRegionRbCmpKeyFn(const VOID *pNodeKeyA, const VOID *pNodeKeyB) } } else if (endA < startB) { return RB_SMALLER; +======= + if (startA > endB) {// A基地址大于B的结束地址 + return RB_BIGGER; //说明线性区A更大,在右边 + } else if (startA >= startB) { + if (endA <= endB) { + return RB_EQUAL; //相等,说明 A在B中 + } else { + return RB_BIGGER; //说明 A的结束地址更大 + } + } else if (startA <= startB) { //A基地址小于等于B的基地址 + if (endA >= endB) { + return RB_EQUAL; //相等 说明 B在A中 + } else { + return RB_SMALLER;//说明A的结束地址更小 + } + } else if (endA < startB) {//A结束地址小于B的开始地址 + return RB_SMALLER;//说明A在 +>>>>>>> remotes/origin/main } return RB_EQUAL; } +<<<<<<< HEAD STATIC BOOL OsVmSpaceInitCommon(LosVmSpace *vmSpace, VADDR_T *virtTtb) { LOS_RbInitTree(&vmSpace->regionRbTree, OsRegionRbCmpKeyFn, OsRegionRbFreeFn, OsRegionRbGetKeyFn); status_t retval = LOS_MuxInit(&vmSpace->regionMux, NULL); +======= +/*! + * @brief OsVmSpaceInitCommon 初始化进程虚拟空间,必须提供L1表的虚拟内存地址 + * + * @param virtTtb L1表的地址,TTB表地址 + * @param vmSpace + * @return + * + * @see + */ +STATIC BOOL OsVmSpaceInitCommon(LosVmSpace *vmSpace, VADDR_T *virtTtb) +{ + LOS_RbInitTree(&vmSpace->regionRbTree, OsRegionRbCmpKeyFn, OsRegionRbFreeFn, OsRegionRbGetKeyFn);//初始化虚拟存储空间-以红黑树组织方式 + + status_t retval = LOS_MuxInit(&vmSpace->regionMux, NULL);//初始化互斥量 +>>>>>>> remotes/origin/main if (retval != LOS_OK) { VM_ERR("Create mutex for vm space failed, status: %d", retval); return FALSE; } (VOID)LOS_MuxAcquire(&g_vmSpaceListMux); +<<<<<<< HEAD LOS_ListAdd(&g_vmSpaceList, &vmSpace->node); (VOID)LOS_MuxRelease(&g_vmSpaceListMux); @@ -151,10 +283,22 @@ STATIC BOOL OsVmSpaceInitCommon(LosVmSpace *vmSpace, VADDR_T *virtTtb) VOID OsVmMapInit(VOID) { status_t retval = LOS_MuxInit(&g_vmSpaceListMux, NULL); +======= + LOS_ListAdd(&g_vmSpaceList, &vmSpace->node);//将虚拟空间挂入全局虚拟空间双循环链表上 + (VOID)LOS_MuxRelease(&g_vmSpaceListMux); + + return OsArchMmuInit(&vmSpace->archMmu, virtTtb);//对mmu初始化 +} +///@note_thinking 这个函数名称和内容不太搭 +VOID OsVmMapInit(VOID) +{ + status_t retval = LOS_MuxInit(&g_vmSpaceListMux, NULL);//初始化虚拟空间的互斥量 +>>>>>>> remotes/origin/main if (retval != LOS_OK) { VM_ERR("Create mutex for g_vmSpaceList failed, status: %d", retval); } } +<<<<<<< HEAD BOOL OsKernVmSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb) { @@ -162,10 +306,33 @@ BOOL OsKernVmSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb) vmSpace->size = KERNEL_ASPACE_SIZE; vmSpace->mapBase = KERNEL_VMM_BASE; vmSpace->mapSize = KERNEL_VMM_SIZE; +======= +///初始化内核虚拟空间 +BOOL OsKernVmSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb)//内核空间页表是编译时放在bbs段指定的,共用 L1表 +{ + vmSpace->base = KERNEL_ASPACE_BASE;//内核空间基地址, 线性区将分配在此范围 + vmSpace->size = KERNEL_ASPACE_SIZE;//内核空间大小 + vmSpace->mapBase = KERNEL_VMM_BASE;//内核空间映射区基地址 + vmSpace->mapSize = KERNEL_VMM_SIZE;//内核空间映射区大小 +#ifdef LOSCFG_DRIVERS_TZDRIVER + vmSpace->codeStart = 0; //代码区开始地址 + vmSpace->codeEnd = 0; //代码区结束地址 +#endif + return OsVmSpaceInitCommon(vmSpace, virtTtb);//virtTtb 用于初始化 mmu +} +///初始化内核堆空间 +BOOL OsVMallocSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb)//内核动态空间的页表是动态申请得来,共用 L1表 +{ + vmSpace->base = VMALLOC_START; //内核堆空间基地址 + vmSpace->size = VMALLOC_SIZE; //内核堆空间大小 + vmSpace->mapBase = VMALLOC_START; //内核堆空间映射基地址 + vmSpace->mapSize = VMALLOC_SIZE; //内核堆空间映射区大小 +>>>>>>> remotes/origin/main #ifdef LOSCFG_DRIVERS_TZDRIVER vmSpace->codeStart = 0; vmSpace->codeEnd = 0; #endif +<<<<<<< HEAD return OsVmSpaceInitCommon(vmSpace, virtTtb); } @@ -198,17 +365,53 @@ BOOL OsUserVmSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb) vmSpace->heapBase = USER_HEAP_BASE; vmSpace->heapNow = USER_HEAP_BASE; vmSpace->heap = NULL; +======= + return OsVmSpaceInitCommon(vmSpace, virtTtb);//创建MMU,为后续的虚实映射做好初始化的工作 +} +///内核虚拟空间初始化 +VOID OsKSpaceInit(VOID) +{ + OsVmMapInit();//初始化后续操作 g_vmSpaceList 的互斥锁 + OsKernVmSpaceInit(&g_kVmSpace, OsGFirstTableGet()); //初始化内核进程虚拟空间 + OsVMallocSpaceInit(&g_vMallocSpace, OsGFirstTableGet());//初始化内核动态分配空间 +} +/*! + * @brief OsUserVmSpaceInit 用户空间的TTB表是动态申请得来,每个进程有属于自己的L1,L2表 + * 初始化用户进程虚拟空间,主要划分数据区,堆区,映射区和创建mmu + * @param virtTtb + * @param vmSpace + * @return + * + * @see + */ +BOOL OsUserVmSpaceInit(LosVmSpace *vmSpace, VADDR_T *virtTtb) +{ + vmSpace->base = USER_ASPACE_BASE;//用户空间基地址 + vmSpace->size = USER_ASPACE_SIZE;//用户空间大小 + vmSpace->mapBase = USER_MAP_BASE;//用户空间映射基地址 + vmSpace->mapSize = USER_MAP_SIZE;//用户空间映射大小 + vmSpace->heapBase = USER_HEAP_BASE;//用户堆区开始地址,只有用户进程需要设置这里,动态内存的开始地址 + vmSpace->heapNow = USER_HEAP_BASE;//堆区最新指向地址,用户堆空间大小可通过系统调用 do_brk()扩展 + vmSpace->heap = NULL; //最近分配的一个堆线性区 +>>>>>>> remotes/origin/main #ifdef LOSCFG_DRIVERS_TZDRIVER vmSpace->codeStart = 0; vmSpace->codeEnd = 0; #endif +<<<<<<< HEAD return OsVmSpaceInitCommon(vmSpace, virtTtb); } +======= + return OsVmSpaceInitCommon(vmSpace, virtTtb);//创建MMU,为后续的虚实映射做好初始化的工作 +} +/// 创建用户进程空间 +>>>>>>> remotes/origin/main LosVmSpace *OsCreateUserVmSpace(VOID) { BOOL retVal = FALSE; +<<<<<<< HEAD LosVmSpace *space = LOS_MemAlloc(m_aucSysMem0, sizeof(LosVmSpace)); if (space == NULL) { return NULL; @@ -223,17 +426,41 @@ LosVmSpace *OsCreateUserVmSpace(VOID) (VOID)memset_s(ttb, PAGE_SIZE, 0, PAGE_SIZE); retVal = OsUserVmSpaceInit(space, ttb); LosVmPage *vmPage = OsVmVaddrToPage(ttb); +======= + LosVmSpace *space = LOS_MemAlloc(m_aucSysMem0, sizeof(LosVmSpace));//在内核空间申请用户进程空间 + if (space == NULL) { + return NULL; + } + //此处为何直接申请物理页帧存放用户进程的页表,大概是因为所有页表都被存放在内核空间(g_kVmSpace)而非内核分配空间(g_vMallocSpace) + VADDR_T *ttb = LOS_PhysPagesAllocContiguous(1);//分配一个物理页用于存放虚实映射关系表, 即:L1表 + if (ttb == NULL) {//若连页表都没有,剩下的也别玩了. + (VOID)LOS_MemFree(m_aucSysMem0, space); + return NULL; + } + + (VOID)memset_s(ttb, PAGE_SIZE, 0, PAGE_SIZE);//4K空间置0 + retVal = OsUserVmSpaceInit(space, ttb);//初始化用户空间,mmu + LosVmPage *vmPage = OsVmVaddrToPage(ttb);//找到所在物理页框 +>>>>>>> remotes/origin/main if ((retVal == FALSE) || (vmPage == NULL)) { (VOID)LOS_MemFree(m_aucSysMem0, space); LOS_PhysPagesFreeContiguous(ttb, 1); return NULL; } +<<<<<<< HEAD LOS_ListAdd(&space->archMmu.ptList, &(vmPage->node)); +======= + LOS_ListAdd(&space->archMmu.ptList, &(vmPage->node));//页表链表,先挂上L1,后续还会挂上 N个L2表 +>>>>>>> remotes/origin/main return space; } +<<<<<<< HEAD STATIC BOOL OsVmSpaceParamCheck(const LosVmSpace *vmSpace) +======= +STATIC BOOL OsVmSpaceParamCheck(const LosVmSpace *vmSpace)//这么简单也要写个函数? +>>>>>>> remotes/origin/main { if (vmSpace == NULL) { return FALSE; @@ -241,6 +468,10 @@ STATIC BOOL OsVmSpaceParamCheck(const LosVmSpace *vmSpace) return TRUE; } +<<<<<<< HEAD +======= +//虚拟内存空间克隆,被用于fork进程 +>>>>>>> remotes/origin/main STATUS_T LOS_VmSpaceClone(UINT32 cloneFlags, LosVmSpace *oldVmSpace, LosVmSpace *newVmSpace) { LosRbNode *pstRbNode = NULL; @@ -255,6 +486,7 @@ STATUS_T LOS_VmSpaceClone(UINT32 cloneFlags, LosVmSpace *oldVmSpace, LosVmSpace return LOS_ERRNO_VM_INVALID_ARGS; } +<<<<<<< HEAD if ((OsIsVmRegionEmpty(oldVmSpace) == TRUE) || (oldVmSpace == &g_kVmSpace)) { return LOS_ERRNO_VM_INVALID_ARGS; } @@ -265,13 +497,29 @@ STATUS_T LOS_VmSpaceClone(UINT32 cloneFlags, LosVmSpace *oldVmSpace, LosVmSpace newVmSpace->heapNow = oldVmSpace->heapNow; (VOID)LOS_MuxAcquire(&oldVmSpace->regionMux); RB_SCAN_SAFE(&oldVmSpace->regionRbTree, pstRbNode, pstRbNodeNext) +======= + if ((OsIsVmRegionEmpty(oldVmSpace) == TRUE) || (oldVmSpace == &g_kVmSpace)) {//不允许clone内核空间,内核空间是独一无二的. + return LOS_ERRNO_VM_INVALID_ARGS; + } + //空间克隆的主体实现是:线性区重新一个个分配物理内存,重新映射. + /* search the region list */ + newVmSpace->mapBase = oldVmSpace->mapBase; //复制映射区基址 + newVmSpace->heapBase = oldVmSpace->heapBase; //复制堆区基址 + newVmSpace->heapNow = oldVmSpace->heapNow; //复制堆区当前使用到哪了 + (VOID)LOS_MuxAcquire(&oldVmSpace->regionMux); + RB_SCAN_SAFE(&oldVmSpace->regionRbTree, pstRbNode, pstRbNodeNext)//红黑树循环开始 +>>>>>>> remotes/origin/main LosVmMapRegion *oldRegion = (LosVmMapRegion *)pstRbNode; #if defined(LOSCFG_KERNEL_SHM) && defined(LOSCFG_IPC_CONTAINER) if ((oldRegion->regionFlags & VM_MAP_REGION_FLAG_SHM) && (cloneFlags & CLONE_NEWIPC)) { continue; } #endif +<<<<<<< HEAD LosVmMapRegion *newRegion = OsVmRegionDup(newVmSpace, oldRegion, oldRegion->range.base, oldRegion->range.size); +======= + LosVmMapRegion *newRegion = OsVmRegionDup(newVmSpace, oldRegion, oldRegion->range.base, oldRegion->range.size);//复制线性区 +>>>>>>> remotes/origin/main if (newRegion == NULL) { VM_ERR("dup new region failed"); ret = LOS_ERRNO_VM_NO_MEMORY; @@ -279,6 +527,7 @@ STATUS_T LOS_VmSpaceClone(UINT32 cloneFlags, LosVmSpace *oldVmSpace, LosVmSpace } #ifdef LOSCFG_KERNEL_SHM +<<<<<<< HEAD if (oldRegion->regionFlags & VM_MAP_REGION_FLAG_SHM) { OsShmFork(newVmSpace, oldRegion, newRegion); continue; @@ -308,21 +557,64 @@ STATUS_T LOS_VmSpaceClone(UINT32 cloneFlags, LosVmSpace *oldVmSpace, LosVmSpace #ifdef LOSCFG_FS_VFS if (LOS_IsRegionFileValid(oldRegion)) { +======= + if (oldRegion->regionFlags & VM_MAP_REGION_FLAG_SHM) {//如果老线性区是共享内存 + OsShmFork(newVmSpace, oldRegion, newRegion);//fork共享线性区,如此新虚拟空间也能用那个线性区 + continue;//不往下走了,因为共享内存不需要重新映射,下面无非就是需要MMU映射虚拟地址<-->物理地址 + } +#endif + + if (oldRegion == oldVmSpace->heap) {//如果这个线性区是堆区 + newVmSpace->heap = newRegion;//那么新的线性区也是新虚拟空间的堆区 + } + + numPages = newRegion->range.size >> PAGE_SHIFT;//计算线性区页数 + for (i = 0; i < numPages; i++) {//一页一页进行重新映射 + vaddr = newRegion->range.base + (i << PAGE_SHIFT); + if (LOS_ArchMmuQuery(&oldVmSpace->archMmu, vaddr, &paddr, &flags) != LOS_OK) {//先查物理地址 + continue; + } + + page = LOS_VmPageGet(paddr);//通过物理页获取物理内存的页框 + if (page != NULL) { + LOS_AtomicInc(&page->refCounts);//refCounts 自增 + } + if (flags & VM_MAP_REGION_FLAG_PERM_WRITE) {//可写入区标签 + LOS_ArchMmuUnmap(&oldVmSpace->archMmu, vaddr, 1);//先删除老空间映射 + LOS_ArchMmuMap(&oldVmSpace->archMmu, vaddr, paddr, 1, flags & ~VM_MAP_REGION_FLAG_PERM_WRITE);//老空间重新映射 + } + LOS_ArchMmuMap(&newVmSpace->archMmu, vaddr, paddr, 1, flags & ~VM_MAP_REGION_FLAG_PERM_WRITE);//映射新空间 + +#ifdef LOSCFG_FS_VFS //文件系统开关 + if (LOS_IsRegionFileValid(oldRegion)) {//是都是一个文件映射线性区 +>>>>>>> remotes/origin/main LosFilePage *fpage = NULL; LOS_SpinLockSave(&oldRegion->unTypeData.rf.vnode->mapping.list_lock, &intSave); fpage = OsFindGetEntry(&oldRegion->unTypeData.rf.vnode->mapping, newRegion->pgOff + i); if ((fpage != NULL) && (fpage->vmPage == page)) { /* cow page no need map */ +<<<<<<< HEAD OsAddMapInfo(fpage, &newVmSpace->archMmu, vaddr); +======= + OsAddMapInfo(fpage, &newVmSpace->archMmu, vaddr);//添加文件页映射,记录页面被进程映射过 +>>>>>>> remotes/origin/main } LOS_SpinUnlockRestore(&oldRegion->unTypeData.rf.vnode->mapping.list_lock, intSave); } #endif } +<<<<<<< HEAD RB_SCAN_SAFE_END(&oldVmSpace->regionRbTree, pstRbNode, pstRbNodeNext) (VOID)LOS_MuxRelease(&oldVmSpace->regionMux); return ret; } +======= + RB_SCAN_SAFE_END(&oldVmSpace->regionRbTree, pstRbNode, pstRbNodeNext)//红黑树循环结束 + (VOID)LOS_MuxRelease(&oldVmSpace->regionMux); + return ret; +} +///通过虚拟(线性)地址查找所属线性区,红黑树 +>>>>>>> remotes/origin/main LosVmMapRegion *OsFindRegion(LosRbTree *regionRbTree, VADDR_T vaddr, size_t len) { LosVmMapRegion *regionRst = NULL; @@ -336,18 +628,30 @@ LosVmMapRegion *OsFindRegion(LosRbTree *regionRbTree, VADDR_T vaddr, size_t len) } return regionRst; } +<<<<<<< HEAD +======= +/// 查找线性区 根据起始地址在进程空间内查找是否存在 +>>>>>>> remotes/origin/main LosVmMapRegion *LOS_RegionFind(LosVmSpace *vmSpace, VADDR_T addr) { LosVmMapRegion *region = NULL; +<<<<<<< HEAD (VOID)LOS_MuxAcquire(&vmSpace->regionMux); +======= + (VOID)LOS_MuxAcquire(&vmSpace->regionMux);//因进程空间是隔离的,所以此处只会涉及到任务(线程)之间的竞争,故使用互斥锁,而自旋锁则用于CPU核间的竞争 +>>>>>>> remotes/origin/main region = OsFindRegion(&vmSpace->regionRbTree, addr, 1); (VOID)LOS_MuxRelease(&vmSpace->regionMux); return region; } +<<<<<<< HEAD +======= +/// 查找线性区 根据地址区间在进程空间内查找是否存在 +>>>>>>> remotes/origin/main LosVmMapRegion *LOS_RegionRangeFind(LosVmSpace *vmSpace, VADDR_T addr, size_t len) { LosVmMapRegion *region = NULL; @@ -358,14 +662,22 @@ LosVmMapRegion *LOS_RegionRangeFind(LosVmSpace *vmSpace, VADDR_T addr, size_t le return region; } +<<<<<<< HEAD +======= +/// 分配指定长度的线性区 +>>>>>>> remotes/origin/main VADDR_T OsAllocRange(LosVmSpace *vmSpace, size_t len) { LosVmMapRegion *curRegion = NULL; LosRbNode *pstRbNode = NULL; LosRbNode *pstRbNodeTmp = NULL; LosRbTree *regionRbTree = &vmSpace->regionRbTree; +<<<<<<< HEAD VADDR_T curEnd = vmSpace->mapBase; +======= + VADDR_T curEnd = vmSpace->mapBase;//获取映射区基地址 +>>>>>>> remotes/origin/main VADDR_T nextStart; curRegion = LOS_RegionFind(vmSpace, vmSpace->mapBase); @@ -384,7 +696,11 @@ VADDR_T OsAllocRange(LosVmSpace *vmSpace, size_t len) curEnd = curRegion->range.base + curRegion->range.size; } RB_MID_SCAN_END(regionRbTree, pstRbNode) +<<<<<<< HEAD } else { +======= + } else {//红黑树扫描排序,从小到大 +>>>>>>> remotes/origin/main /* rbtree scan is sorted, from small to big */ RB_SCAN_SAFE(regionRbTree, pstRbNode, pstRbNodeTmp) curRegion = (LosVmMapRegion *)pstRbNode; @@ -407,34 +723,58 @@ VADDR_T OsAllocRange(LosVmSpace *vmSpace, size_t len) return 0; } +<<<<<<< HEAD +======= +/// 分配指定开始地址和长度的线性区 +>>>>>>> remotes/origin/main VADDR_T OsAllocSpecificRange(LosVmSpace *vmSpace, VADDR_T vaddr, size_t len, UINT32 regionFlags) { STATUS_T status; +<<<<<<< HEAD if (LOS_IsRangeInSpace(vmSpace, vaddr, len) == FALSE) { +======= + if (LOS_IsRangeInSpace(vmSpace, vaddr, len) == FALSE) {//虚拟地址是否在进程空间范围内 +>>>>>>> remotes/origin/main return 0; } if ((LOS_RegionFind(vmSpace, vaddr) != NULL) || (LOS_RegionFind(vmSpace, vaddr + len - 1) != NULL) || +<<<<<<< HEAD (LOS_RegionRangeFind(vmSpace, vaddr, len - 1) != NULL)) { if ((regionFlags & VM_MAP_REGION_FLAG_FIXED_NOREPLACE) != 0) { return 0; } else if ((regionFlags & VM_MAP_REGION_FLAG_FIXED) != 0) { status = LOS_UnMMap(vaddr, len); +======= + (LOS_RegionRangeFind(vmSpace, vaddr, len - 1) != NULL)) {//没找到的情况 + if ((regionFlags & VM_MAP_REGION_FLAG_FIXED_NOREPLACE) != 0) { + return 0; + } else if ((regionFlags & VM_MAP_REGION_FLAG_FIXED) != 0) {//线性区未填满,则解除这部分空间的映射 + status = LOS_UnMMap(vaddr, len);//解除映射 +>>>>>>> remotes/origin/main if (status != LOS_OK) { VM_ERR("unmap specific range va: %#x, len: %#x failed, status: %d", vaddr, len, status); return 0; } } else { +<<<<<<< HEAD return OsAllocRange(vmSpace, len); +======= + return OsAllocRange(vmSpace, len);//默认分配一个 +>>>>>>> remotes/origin/main } } return vaddr; } +<<<<<<< HEAD +======= +///映射类型为文件的线性区是否有效 +>>>>>>> remotes/origin/main BOOL LOS_IsRegionFileValid(LosVmMapRegion *region) { if ((region != NULL) && (LOS_IsRegionTypeFile(region)) && @@ -443,7 +783,11 @@ BOOL LOS_IsRegionFileValid(LosVmMapRegion *region) } return FALSE; } +<<<<<<< HEAD +======= +///向红黑树中插入线性区 +>>>>>>> remotes/origin/main BOOL OsInsertRegion(LosRbTree *regionRbTree, LosVmMapRegion *region) { if (LOS_RbAddNode(regionRbTree, (LosRbNode *)region) == FALSE) { @@ -453,14 +797,22 @@ BOOL OsInsertRegion(LosRbTree *regionRbTree, LosVmMapRegion *region) } return TRUE; } +<<<<<<< HEAD LosVmMapRegion *OsCreateRegion(VADDR_T vaddr, size_t len, UINT32 regionFlags, unsigned long offset) { LosVmMapRegion *region = LOS_MemAlloc(m_aucSysMem0, sizeof(LosVmMapRegion)); +======= +///创建一个线性区 +LosVmMapRegion *OsCreateRegion(VADDR_T vaddr, size_t len, UINT32 regionFlags, unsigned long offset) +{ + LosVmMapRegion *region = LOS_MemAlloc(m_aucSysMem0, sizeof(LosVmMapRegion));//只是分配一个线性区结构体 +>>>>>>> remotes/origin/main if (region == NULL) { VM_ERR("memory allocate for LosVmMapRegion failed"); return region; } +<<<<<<< HEAD (void)memset_s(region, sizeof(LosVmMapRegion), 0, sizeof(LosVmMapRegion)); region->range.base = vaddr; @@ -473,12 +825,27 @@ LosVmMapRegion *OsCreateRegion(VADDR_T vaddr, size_t len, UINT32 regionFlags, un return region; } +======= + //创建线性区的本质就是在画饼,见如下操作: + (void)memset_s(region, sizeof(LosVmMapRegion), 0, sizeof(LosVmMapRegion)); + region->range.base = vaddr; //虚拟地址作为线性区的基地址 + region->range.size = len; //线性区大小,这是线性区构思最巧妙的地方,只要不过分,蓝图随便画。 + region->pgOff = offset; //页标 + region->regionFlags = regionFlags;//标识,可读/可写/可执行 + region->regionType = VM_MAP_REGION_TYPE_NONE;//未映射 + region->forkFlags = 0; // + region->shmid = -1; //默认线性区为不共享,无共享资源ID + return region; +} +///通过虚拟地址查询映射的物理地址 +>>>>>>> remotes/origin/main PADDR_T LOS_PaddrQuery(VOID *vaddr) { PADDR_T paddr = 0; STATUS_T status; LosVmSpace *space = NULL; LosArchMmu *archMmu = NULL; +<<<<<<< HEAD if (LOS_IsKernelAddress((VADDR_T)(UINTPTR)vaddr)) { archMmu = &g_kVmSpace.archMmu; @@ -486,13 +853,26 @@ PADDR_T LOS_PaddrQuery(VOID *vaddr) space = OsCurrProcessGet()->vmSpace; archMmu = &space->archMmu; } else if (LOS_IsVmallocAddress((VADDR_T)(UINTPTR)vaddr)) { +======= + //先取出对应空间的mmu + if (LOS_IsKernelAddress((VADDR_T)(UINTPTR)vaddr)) {//是否内核空间地址 + archMmu = &g_kVmSpace.archMmu; + } else if (LOS_IsUserAddress((VADDR_T)(UINTPTR)vaddr)) {//是否为用户空间地址 + space = OsCurrProcessGet()->vmSpace; + archMmu = &space->archMmu; + } else if (LOS_IsVmallocAddress((VADDR_T)(UINTPTR)vaddr)) {//是否为分配空间地址,堆区地址 +>>>>>>> remotes/origin/main archMmu = &g_vMallocSpace.archMmu; } else { VM_ERR("vaddr is beyond range"); return 0; } +<<<<<<< HEAD status = LOS_ArchMmuQuery(archMmu, (VADDR_T)(UINTPTR)vaddr, &paddr, 0); +======= + status = LOS_ArchMmuQuery(archMmu, (VADDR_T)(UINTPTR)vaddr, &paddr, 0);//查询物理地址 +>>>>>>> remotes/origin/main if (status == LOS_OK) { return paddr; } else { @@ -500,6 +880,13 @@ PADDR_T LOS_PaddrQuery(VOID *vaddr) } } +<<<<<<< HEAD +======= +/*! + * 这里不是真的分配物理内存,而是逻辑上画一个连续的区域,标记这个区域可以拿用,表示内存已经归你了。 + 但真正的物理内存的占用会延迟到使用的时候才由缺页中断调入内存 +*/ +>>>>>>> remotes/origin/main LosVmMapRegion *LOS_RegionAlloc(LosVmSpace *vmSpace, VADDR_T vaddr, size_t len, UINT32 regionFlags, VM_OFFSET_T pgoff) { VADDR_T rstVaddr; @@ -510,37 +897,70 @@ LosVmMapRegion *LOS_RegionAlloc(LosVmSpace *vmSpace, VADDR_T vaddr, size_t len, * this is the most portable method of creating a new mapping. If addr is not NULL, * then the kernel takes it as where to place the mapping; */ +<<<<<<< HEAD (VOID)LOS_MuxAcquire(&vmSpace->regionMux); if (vaddr == 0) { rstVaddr = OsAllocRange(vmSpace, len); } else { /* if it is already mmapped here, we unmmap it */ rstVaddr = OsAllocSpecificRange(vmSpace, vaddr, len, regionFlags); +======= + (VOID)LOS_MuxAcquire(&vmSpace->regionMux);//获得互斥锁 + if (vaddr == 0) {//如果地址是0,根据线性区管理的实际情况,自动创建虚拟地址, 这是创建新映射的最便捷的方法。 + rstVaddr = OsAllocRange(vmSpace, len); + } else { + /* if it is already mmapped here, we unmmap it | 如果已经被映射了, 则解除映射关系*/ + rstVaddr = OsAllocSpecificRange(vmSpace, vaddr, len, regionFlags);//创建包含指定虚拟地址的线性区, rstVaddr != vaddr || rstVaddr == vaddr +>>>>>>> remotes/origin/main if (rstVaddr == 0) { VM_ERR("alloc specific range va: %#x, len: %#x failed", vaddr, len); goto OUT; } } +<<<<<<< HEAD if (rstVaddr == 0) { goto OUT; } newRegion = OsCreateRegion(rstVaddr, len, regionFlags, pgoff); +======= + if (rstVaddr == 0) {//没有可供映射的虚拟地址 + goto OUT; + } + + newRegion = OsCreateRegion(rstVaddr, len, regionFlags, pgoff);//创建一个线性区,指定线性区的开始地址rstVaddr ... +>>>>>>> remotes/origin/main if (newRegion == NULL) { goto OUT; } newRegion->space = vmSpace; +<<<<<<< HEAD isInsertSucceed = OsInsertRegion(&vmSpace->regionRbTree, newRegion); if (isInsertSucceed == FALSE) { (VOID)LOS_MemFree(m_aucSysMem0, newRegion); +======= + isInsertSucceed = OsInsertRegion(&vmSpace->regionRbTree, newRegion);//插入红黑树和双循环链表中管理 + if (isInsertSucceed == FALSE) {//插入失败 + (VOID)LOS_MemFree(m_aucSysMem0, newRegion);//从内存池中释放 +>>>>>>> remotes/origin/main newRegion = NULL; } OUT: +<<<<<<< HEAD (VOID)LOS_MuxRelease(&vmSpace->regionMux); return newRegion; } +======= + (VOID)LOS_MuxRelease(&vmSpace->regionMux);//释放互斥锁 + return newRegion; +} +/*! + * 删除匿名页,匿名页就是内存映射页 + * 1.解除映射关系 2.释放物理内存 +*/ +>>>>>>> remotes/origin/main STATIC VOID OsAnonPagesRemove(LosArchMmu *archMmu, VADDR_T vaddr, UINT32 count) { status_t status; @@ -552,20 +972,36 @@ STATIC VOID OsAnonPagesRemove(LosArchMmu *archMmu, VADDR_T vaddr, UINT32 count) return; } +<<<<<<< HEAD while (count > 0) { count--; status = LOS_ArchMmuQuery(archMmu, vaddr, &paddr, NULL); if (status != LOS_OK) { +======= + while (count > 0) {//一页页操作 + count--; + status = LOS_ArchMmuQuery(archMmu, vaddr, &paddr, NULL);//通过虚拟地址拿到物理地址 + if (status != LOS_OK) {//失败,拿下一页的物理地址 +>>>>>>> remotes/origin/main vaddr += PAGE_SIZE; continue; } +<<<<<<< HEAD LOS_ArchMmuUnmap(archMmu, vaddr, 1); page = LOS_VmPageGet(paddr); if (page != NULL) { if (!OsIsPageShared(page)) { LOS_PhysPageFree(page); +======= + LOS_ArchMmuUnmap(archMmu, vaddr, 1);//解除一页的映射 + + page = LOS_VmPageGet(paddr);//通过物理地址获取所在物理页框的起始地址 + if (page != NULL) {//获取成功 + if (!OsIsPageShared(page)) {//不是共享页,共享页会有专门的共享标签,共享本质是有无多个进程对该页的引用 + LOS_PhysPageFree(page);//释放物理页框 +>>>>>>> remotes/origin/main } } vaddr += PAGE_SIZE; @@ -609,7 +1045,11 @@ STATIC VOID OsFilePagesRemove(LosVmSpace *space, LosVmMapRegion *region) } } #endif +<<<<<<< HEAD +======= +/// 释放进程空间指定线性区 +>>>>>>> remotes/origin/main STATUS_T LOS_RegionFree(LosVmSpace *space, LosVmMapRegion *region) { if ((space == NULL) || (region == NULL)) { @@ -619,14 +1059,21 @@ STATUS_T LOS_RegionFree(LosVmSpace *space, LosVmMapRegion *region) (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD #ifdef LOSCFG_FS_VFS if (LOS_IsRegionFileValid(region)) { OsFilePagesRemove(space, region); +======= +#ifdef LOSCFG_FS_VFS //文件开关 + if (LOS_IsRegionFileValid(region)) {//是否为文件线性区 + OsFilePagesRemove(space, region);//删除文件页 +>>>>>>> remotes/origin/main VnodeHold(); region->unTypeData.rf.vnode->useCount--; VnodeDrop(); } else #endif +<<<<<<< HEAD #ifdef LOSCFG_KERNEL_SHM if (OsIsShmRegion(region)) { @@ -648,6 +1095,28 @@ STATUS_T LOS_RegionFree(LosVmSpace *space, LosVmMapRegion *region) return LOS_OK; } +======= +#ifdef LOSCFG_KERNEL_SHM //共享内存开关 + if (OsIsShmRegion(region)) { //是否为共享内存线性区 + OsShmRegionFree(space, region);//释放共享线性区 + } else if (LOS_IsRegionTypeDev(region)) { +#else + if (LOS_IsRegionTypeDev(region)) {//如果是设备线性区 +#endif + OsDevPagesRemove(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT);//删除映射设备 + } else { + OsAnonPagesRemove(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT);//删除匿名映射 + } + + /* remove it from space */ + LOS_RbDelNode(&space->regionRbTree, ®ion->rbNode);//从红黑树中删除线性区 + /* free it */ + LOS_MemFree(m_aucSysMem0, region);//释放线性区结构体占用的内存 + (VOID)LOS_MuxRelease(&space->regionMux); + return LOS_OK; +} +/// 复制线性区 +>>>>>>> remotes/origin/main LosVmMapRegion *OsVmRegionDup(LosVmSpace *space, LosVmMapRegion *oldRegion, VADDR_T vaddr, size_t size) { LosVmMapRegion *newRegion = NULL; @@ -655,16 +1124,26 @@ LosVmMapRegion *OsVmRegionDup(LosVmSpace *space, LosVmMapRegion *oldRegion, VADD (VOID)LOS_MuxAcquire(&space->regionMux); regionFlags = oldRegion->regionFlags; +<<<<<<< HEAD if (vaddr == 0) { regionFlags &= ~(VM_MAP_REGION_FLAG_FIXED | VM_MAP_REGION_FLAG_FIXED_NOREPLACE); } else { regionFlags |= VM_MAP_REGION_FLAG_FIXED; } newRegion = LOS_RegionAlloc(space, vaddr, size, regionFlags, oldRegion->pgOff); +======= + if (vaddr == 0) {//不指定地址 + regionFlags &= ~(VM_MAP_REGION_FLAG_FIXED | VM_MAP_REGION_FLAG_FIXED_NOREPLACE); //撕掉两个标签 + } else { + regionFlags |= VM_MAP_REGION_FLAG_FIXED; //贴上填满线性区标签 + } + newRegion = LOS_RegionAlloc(space, vaddr, size, regionFlags, oldRegion->pgOff); //分配一个线性区 +>>>>>>> remotes/origin/main if (newRegion == NULL) { VM_ERR("LOS_RegionAlloc failed"); goto REGIONDUPOUT; } +<<<<<<< HEAD newRegion->regionType = oldRegion->regionType; #ifdef LOSCFG_KERNEL_SHM @@ -680,6 +1159,23 @@ LosVmMapRegion *OsVmRegionDup(LosVmSpace *space, LosVmMapRegion *oldRegion, VADD newRegion->unTypeData.rf.f_oflags = oldRegion->unTypeData.rf.f_oflags; VnodeHold(); newRegion->unTypeData.rf.vnode->useCount++; +======= + newRegion->regionType = oldRegion->regionType;//复制线性区类型(文件,设备,匿名) + +#ifdef LOSCFG_KERNEL_SHM + if (OsIsShmRegion(oldRegion)) {//如果是共享内存 + newRegion->shmid = oldRegion->shmid;//复制共享ID + } +#endif + +#ifdef LOSCFG_FS_VFS //文件开关 + if (LOS_IsRegionTypeFile(oldRegion)) {//如果是文件线性区 + newRegion->unTypeData.rf.vmFOps = oldRegion->unTypeData.rf.vmFOps; //文件操作接口 + newRegion->unTypeData.rf.vnode = oldRegion->unTypeData.rf.vnode; //文件索引节点 + newRegion->unTypeData.rf.f_oflags = oldRegion->unTypeData.rf.f_oflags;//读写标签 + VnodeHold(); + newRegion->unTypeData.rf.vnode->useCount++;//索引节点使用数增加 +>>>>>>> remotes/origin/main VnodeDrop(); } #endif @@ -688,14 +1184,24 @@ REGIONDUPOUT: (VOID)LOS_MuxRelease(&space->regionMux); return newRegion; } +<<<<<<< HEAD +======= +/// 劈开线性区 +>>>>>>> remotes/origin/main STATIC LosVmMapRegion *OsVmRegionSplit(LosVmMapRegion *oldRegion, VADDR_T newRegionStart) { LosVmMapRegion *newRegion = NULL; LosVmSpace *space = oldRegion->space; +<<<<<<< HEAD size_t size = LOS_RegionSize(newRegionStart, LOS_RegionEndAddr(oldRegion)); oldRegion->range.size = LOS_RegionSize(oldRegion->range.base, newRegionStart - 1); +======= + size_t size = LOS_RegionSize(newRegionStart, LOS_RegionEndAddr(oldRegion));//获取线性区大小 + + oldRegion->range.size = LOS_RegionSize(oldRegion->range.base, newRegionStart - 1);//获取旧线性区大小 +>>>>>>> remotes/origin/main if (oldRegion->range.size == 0) { LOS_RbDelNode(&space->regionRbTree, &oldRegion->rbNode); } @@ -710,14 +1216,22 @@ STATIC LosVmMapRegion *OsVmRegionSplit(LosVmMapRegion *oldRegion, VADDR_T newReg #endif return newRegion; } +<<<<<<< HEAD +======= +///对线性区进行调整 +>>>>>>> remotes/origin/main STATUS_T OsVmRegionAdjust(LosVmSpace *space, VADDR_T newRegionStart, size_t size) { LosVmMapRegion *region = NULL; VADDR_T nextRegionBase = newRegionStart + size; LosVmMapRegion *newRegion = NULL; +<<<<<<< HEAD region = LOS_RegionFind(space, newRegionStart); +======= + region = LOS_RegionFind(space, newRegionStart);//先找到线性区 +>>>>>>> remotes/origin/main if ((region != NULL) && (newRegionStart > region->range.base)) { newRegion = OsVmRegionSplit(region, newRegionStart); if (newRegion == NULL) { @@ -737,7 +1251,11 @@ STATUS_T OsVmRegionAdjust(LosVmSpace *space, VADDR_T newRegionStart, size_t size return LOS_OK; } +<<<<<<< HEAD +======= +///删除线性区 +>>>>>>> remotes/origin/main STATUS_T OsRegionsRemove(LosVmSpace *space, VADDR_T regionBase, size_t size) { STATUS_T status; @@ -748,12 +1266,20 @@ STATUS_T OsRegionsRemove(LosVmSpace *space, VADDR_T regionBase, size_t size) (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD status = OsVmRegionAdjust(space, regionBase, size); +======= + status = OsVmRegionAdjust(space, regionBase, size);//线性区调整 +>>>>>>> remotes/origin/main if (status != LOS_OK) { goto ERR_REGION_SPLIT; } +<<<<<<< HEAD RB_SCAN_SAFE(&space->regionRbTree, pstRbNodeTemp, pstRbNodeNext) +======= + RB_SCAN_SAFE(&space->regionRbTree, pstRbNodeTemp, pstRbNodeNext)//扫描虚拟空间内的线性区 +>>>>>>> remotes/origin/main regionTemp = (LosVmMapRegion *)pstRbNodeTemp; if (regionTemp->range.base > regionEnd) { break; @@ -772,7 +1298,11 @@ ERR_REGION_SPLIT: (VOID)LOS_MuxRelease(&space->regionMux); return status; } +<<<<<<< HEAD +======= +///根据指定参数范围[addr,addr+len] 释放用户空间中堆区所占用的物理内存 +>>>>>>> remotes/origin/main INT32 OsUserHeapFree(LosVmSpace *vmSpace, VADDR_T addr, size_t len) { LosVmMapRegion *vmRegion = NULL; @@ -781,15 +1311,24 @@ INT32 OsUserHeapFree(LosVmSpace *vmSpace, VADDR_T addr, size_t len) VADDR_T vaddr; STATUS_T ret; +<<<<<<< HEAD if (vmSpace == LOS_GetKVmSpace() || vmSpace->heap == NULL) { return -1; } vmRegion = LOS_RegionFind(vmSpace, addr); +======= + if (vmSpace == LOS_GetKVmSpace() || vmSpace->heap == NULL) {//虚拟空间堆区必须在非内核空间 + return -1; + } + + vmRegion = LOS_RegionFind(vmSpace, addr);//通过参数虚拟地址红黑树找到线性区,线性区范围内包含了参数虚拟地址 +>>>>>>> remotes/origin/main if (vmRegion == NULL) { return -1; } +<<<<<<< HEAD if (vmRegion == vmSpace->heap) { vaddr = addr; while (len > 0) { @@ -800,6 +1339,18 @@ INT32 OsUserHeapFree(LosVmSpace *vmSpace, VADDR_T addr, size_t len) } vmPage = LOS_VmPageGet(paddr); LOS_PhysPageFree(vmPage); +======= + if (vmRegion == vmSpace->heap) {//地址所在的线性区为堆区 + vaddr = addr; + while (len > 0) {//参数0 代表不获取 flags 信息 + if (LOS_ArchMmuQuery(&vmSpace->archMmu, vaddr, &paddr, 0) == LOS_OK) {//通过虚拟地址查到物理地址 + ret = LOS_ArchMmuUnmap(&vmSpace->archMmu, vaddr, 1);//解除映射关系以页为单位,这里解除1页 + if (ret <= 0) { + VM_ERR("unmap failed, ret = %d", ret); + } + vmPage = LOS_VmPageGet(paddr);//获取物理页面信息 + LOS_PhysPageFree(vmPage);//释放页 +>>>>>>> remotes/origin/main } vaddr += PAGE_SIZE; len -= PAGE_SIZE; @@ -809,7 +1360,11 @@ INT32 OsUserHeapFree(LosVmSpace *vmSpace, VADDR_T addr, size_t len) return -1; } +<<<<<<< HEAD +======= +///线性区是否支持扩展 +>>>>>>> remotes/origin/main STATUS_T OsIsRegionCanExpand(LosVmSpace *space, LosVmMapRegion *region, size_t size) { LosVmMapRegion *nextRegion = NULL; @@ -826,13 +1381,21 @@ STATUS_T OsIsRegionCanExpand(LosVmSpace *space, LosVmMapRegion *region, size_t s return LOS_NOK; } +<<<<<<< HEAD +======= +///解除一定范围的虚拟地址的映射关系 +>>>>>>> remotes/origin/main STATUS_T OsUnMMap(LosVmSpace *space, VADDR_T addr, size_t size) { size = LOS_Align(size, PAGE_SIZE); addr = LOS_Align(addr, PAGE_SIZE); (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD STATUS_T status = OsRegionsRemove(space, addr, size); +======= + STATUS_T status = OsRegionsRemove(space, addr, size);//删除线性区 +>>>>>>> remotes/origin/main if (status != LOS_OK) { status = -EINVAL; VM_ERR("region_split failed"); @@ -843,19 +1406,29 @@ ERR_REGION_SPLIT: (VOID)LOS_MuxRelease(&space->regionMux); return status; } +<<<<<<< HEAD +======= +/// 释放所有线性区 +>>>>>>> remotes/origin/main STATIC VOID OsVmSpaceAllRegionFree(LosVmSpace *space) { LosRbNode *pstRbNode = NULL; LosRbNode *pstRbNodeNext = NULL; /* free all of the regions */ +<<<<<<< HEAD RB_SCAN_SAFE(&space->regionRbTree, pstRbNode, pstRbNodeNext) LosVmMapRegion *region = (LosVmMapRegion *)pstRbNode; +======= + RB_SCAN_SAFE(&space->regionRbTree, pstRbNode, pstRbNodeNext) //遍历红黑树 + LosVmMapRegion *region = (LosVmMapRegion *)pstRbNode;//拿到线性区 +>>>>>>> remotes/origin/main if (region->range.size == 0) { VM_ERR("space free, region: %#x flags: %#x, base:%#x, size: %#x", region, region->regionFlags, region->range.base, region->range.size); } +<<<<<<< HEAD STATUS_T ret = LOS_RegionFree(space, region); if (ret != LOS_OK) { VM_ERR("free region error, space %p, region %p", space, region); @@ -865,6 +1438,17 @@ STATIC VOID OsVmSpaceAllRegionFree(LosVmSpace *space) return; } +======= + STATUS_T ret = LOS_RegionFree(space, region);//释放线性区 + if (ret != LOS_OK) { + VM_ERR("free region error, space %p, region %p", space, region); + } + RB_SCAN_SAFE_END(&space->regionRbTree, pstRbNode, pstRbNodeNext)//要好好研究下这几个宏,有点意思 + + return; +} +/// 释放虚拟空间 +>>>>>>> remotes/origin/main STATUS_T OsVmSpaceRegionFree(LosVmSpace *space) { if (space == NULL) { @@ -882,22 +1466,34 @@ STATUS_T OsVmSpaceRegionFree(LosVmSpace *space) return LOS_OK; } +<<<<<<< HEAD +======= +///释放虚拟空间,注意内核空间不能被释放掉,永驻内存 +>>>>>>> remotes/origin/main STATUS_T LOS_VmSpaceFree(LosVmSpace *space) { if (space == NULL) { return LOS_ERRNO_VM_INVALID_ARGS; } +<<<<<<< HEAD if (space == &g_kVmSpace) { +======= + if (space == &g_kVmSpace) {//不能释放内核虚拟空间,内核空间常驻内存 +>>>>>>> remotes/origin/main VM_ERR("try to free kernel aspace, not allowed"); return LOS_OK; } /* pop it out of the global aspace list */ (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD LOS_ListDelete(&space->node); +======= + LOS_ListDelete(&space->node);//从g_vmSpaceList链表里删除,g_vmSpaceList记录了所有空间节点。 +>>>>>>> remotes/origin/main OsVmSpaceAllRegionFree(space); @@ -920,7 +1516,11 @@ STATUS_T LOS_VmSpaceFree(LosVmSpace *space) LOS_MemFree(m_aucSysMem0, space); return LOS_OK; } +<<<<<<< HEAD +======= +///虚拟地址和size是否在空间 +>>>>>>> remotes/origin/main BOOL LOS_IsRangeInSpace(const LosVmSpace *space, VADDR_T vaddr, size_t size) { /* is the starting address within the address space */ @@ -940,7 +1540,11 @@ BOOL LOS_IsRangeInSpace(const LosVmSpace *space, VADDR_T vaddr, size_t size) } return TRUE; } +<<<<<<< HEAD +======= +/// 在进程空间中预留一块内存空间 +>>>>>>> remotes/origin/main STATUS_T LOS_VmSpaceReserve(LosVmSpace *space, size_t size, VADDR_T vaddr) { UINT32 regionFlags = 0; @@ -961,7 +1565,11 @@ STATUS_T LOS_VmSpaceReserve(LosVmSpace *space, size_t size, VADDR_T vaddr) return region ? LOS_OK : LOS_ERRNO_VM_NO_MEMORY; } +<<<<<<< HEAD +======= +///实现从虚拟地址到物理地址的映射,将指定长度的物理地址区间与虚拟地址区间做映射,需提前申请物理地址区间 +>>>>>>> remotes/origin/main STATUS_T LOS_VaddrToPaddrMmap(LosVmSpace *space, VADDR_T vaddr, PADDR_T paddr, size_t len, UINT32 flags) { STATUS_T ret; @@ -976,19 +1584,34 @@ STATUS_T LOS_VaddrToPaddrMmap(LosVmSpace *space, VADDR_T vaddr, PADDR_T paddr, s } if (space == NULL) { +<<<<<<< HEAD space = OsCurrProcessGet()->vmSpace; } region = LOS_RegionFind(space, vaddr); if (region != NULL) { +======= + space = OsCurrProcessGet()->vmSpace;//获取当前进程的空间 + } + + region = LOS_RegionFind(space, vaddr);//通过虚拟地址查找线性区 + if (region != NULL) {//已经被映射过了,失败返回 +>>>>>>> remotes/origin/main VM_ERR("vaddr : 0x%x already used!", vaddr); return LOS_ERRNO_VM_BUSY; } +<<<<<<< HEAD region = LOS_RegionAlloc(space, vaddr, len, flags, 0); if (region == NULL) { VM_ERR("failed"); return LOS_ERRNO_VM_NO_MEMORY; +======= + region = LOS_RegionAlloc(space, vaddr, len, flags, 0);//通过虚拟地址 创建一个region + if (region == NULL) { + VM_ERR("failed"); + return LOS_ERRNO_VM_NO_MEMORY;//内存不够 +>>>>>>> remotes/origin/main } while (len > 0) { @@ -998,9 +1621,15 @@ STATUS_T LOS_VaddrToPaddrMmap(LosVmSpace *space, VADDR_T vaddr, PADDR_T paddr, s VM_ERR("Page is NULL"); return LOS_ERRNO_VM_NOT_VALID; } +<<<<<<< HEAD LOS_AtomicInc(&vmPage->refCounts); ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, region->regionFlags); +======= + LOS_AtomicInc(&vmPage->refCounts);//ref自增 + + ret = LOS_ArchMmuMap(&space->archMmu, vaddr, paddr, 1, region->regionFlags);//mmu map +>>>>>>> remotes/origin/main if (ret <= 0) { VM_ERR("LOS_ArchMmuMap failed: %d", ret); LOS_RegionFree(space, region); @@ -1014,9 +1643,16 @@ STATUS_T LOS_VaddrToPaddrMmap(LosVmSpace *space, VADDR_T vaddr, PADDR_T paddr, s return LOS_OK; } +<<<<<<< HEAD VOID *LOS_VMalloc(size_t size) { LosVmSpace *space = &g_vMallocSpace; +======= +//对外接口|申请内核堆空间内存 +VOID *LOS_VMalloc(size_t size) +{ + LosVmSpace *space = &g_vMallocSpace;//从内核动态空间申请 +>>>>>>> remotes/origin/main LosVmMapRegion *region = NULL; size_t sizeCount; size_t count; @@ -1025,6 +1661,7 @@ VOID *LOS_VMalloc(size_t size) PADDR_T pa; STATUS_T ret; +<<<<<<< HEAD size = LOS_Align(size, PAGE_SIZE); if ((size == 0) || (size > space->size)) { return NULL; @@ -1035,18 +1672,36 @@ VOID *LOS_VMalloc(size_t size) (VOID)LOS_MuxAcquire(&space->regionMux); count = LOS_PhysPagesAlloc(sizeCount, &pageList); +======= + size = LOS_Align(size, PAGE_SIZE);// + if ((size == 0) || (size > space->size)) { + return NULL; + } + sizeCount = size >> PAGE_SHIFT;//按页申请所以需右移12位 + + LOS_DL_LIST_HEAD(pageList); + (VOID)LOS_MuxAcquire(&space->regionMux);//获得互斥锁 + + count = LOS_PhysPagesAlloc(sizeCount, &pageList);//一页一页申请,并从pageList尾部插入 +>>>>>>> remotes/origin/main if (count < sizeCount) { VM_ERR("failed to allocate enough pages (ask %zu, got %zu)", sizeCount, count); goto ERROR; } +<<<<<<< HEAD /* allocate a region and put it in the aspace list */ region = LOS_RegionAlloc(space, 0, size, VM_MAP_REGION_FLAG_PERM_READ | VM_MAP_REGION_FLAG_PERM_WRITE, 0); +======= + /* allocate a region and put it in the aspace list *///分配一个可读写的线性区,并挂在space + region = LOS_RegionAlloc(space, 0, size, VM_MAP_REGION_FLAG_PERM_READ | VM_MAP_REGION_FLAG_PERM_WRITE, 0);//注意第二个参数是 vaddr = 0 !!! +>>>>>>> remotes/origin/main if (region == NULL) { VM_ERR("alloc region failed, size = %x", size); goto ERROR; } +<<<<<<< HEAD va = region->range.base; while ((vmPage = LOS_ListRemoveHeadType(&pageList, LosVmPage, node))) { pa = vmPage->physAddr; @@ -1067,6 +1722,28 @@ ERROR: return NULL; } +======= + va = region->range.base;//va 该区范围基地址为虚拟地址的开始位置,理解va怎么来的是理解线性地址的关键! + while ((vmPage = LOS_ListRemoveHeadType(&pageList, LosVmPage, node))) {//从pageList循环拿page + pa = vmPage->physAddr;//获取page物理地址,因上面是通过LOS_PhysPagesAlloc分配 + LOS_AtomicInc(&vmPage->refCounts);//refCounts 自增 + ret = LOS_ArchMmuMap(&space->archMmu, va, pa, 1, region->regionFlags);//一页一页的map + if (ret != 1) { + VM_ERR("LOS_ArchMmuMap failed!, err;%d", ret); + } + va += PAGE_SIZE;//一页映射完成,进入下一页 + }//va 注意 region的虚拟地址页是连续的,但物理页可以不连续! 很重要!!! + + (VOID)LOS_MuxRelease(&space->regionMux);//释放互斥锁 + return (VOID *)(UINTPTR)region->range.base;//返回虚拟基地址供应用使用 + +ERROR: + (VOID)LOS_PhysPagesFree(&pageList);//释放物理内存页 + (VOID)LOS_MuxRelease(&space->regionMux);//释放互斥锁 + return NULL; +} +///对外接口|释放内核堆空间内存 +>>>>>>> remotes/origin/main VOID LOS_VFree(const VOID *addr) { LosVmSpace *space = &g_vMallocSpace; @@ -1080,13 +1757,21 @@ VOID LOS_VFree(const VOID *addr) (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD region = LOS_RegionFind(space, (VADDR_T)(UINTPTR)addr); +======= + region = LOS_RegionFind(space, (VADDR_T)(UINTPTR)addr);//先找到线性区 +>>>>>>> remotes/origin/main if (region == NULL) { VM_ERR("find region failed"); goto DONE; } +<<<<<<< HEAD ret = LOS_RegionFree(space, region); +======= + ret = LOS_RegionFree(space, region);//释放线性区 +>>>>>>> remotes/origin/main if (ret) { VM_ERR("free region failed, ret = %d", ret); } @@ -1097,10 +1782,16 @@ DONE: LosMux *OsGVmSpaceMuxGet(VOID) { +<<<<<<< HEAD return &g_vmSpaceListMux; } STATIC INLINE BOOL OsMemLargeAlloc(UINT32 size) +======= + return &g_vmSpaceListMux; +} +STATIC INLINE BOOL OsMemLargeAlloc(UINT32 size)//是不是分配浪费大于1K的内存 +>>>>>>> remotes/origin/main { if (g_kHeapInited == FALSE) { return FALSE; @@ -1122,6 +1813,7 @@ PADDR_T LOS_PaddrQuery(VOID *vaddr) return (PADDR_T)VMM_TO_DMA_ADDR((VADDR_T)vaddr); } #endif +<<<<<<< HEAD VOID *LOS_KernelMalloc(UINT32 size) { @@ -1134,16 +1826,38 @@ VOID *LOS_KernelMalloc(UINT32 size) #endif { ptr = LOS_MemAlloc(OS_SYS_MEM_ADDR, size); +======= +///内核空间内存分配,申请小于16KiB的内存则通过堆内存池获取,否则申请多个连续物理页 +VOID *LOS_KernelMalloc(UINT32 size) +{ + VOID *ptr = NULL; + //从本函数可知,内核空间的分配有两种方式 +#ifdef LOSCFG_KERNEL_VM + if (OsMemLargeAlloc(size)) {//是不是分配浪费小于1K的内存 + ptr = LOS_PhysPagesAllocContiguous(ROUNDUP(size, PAGE_SIZE) >> PAGE_SHIFT);//分配连续的物理内存页 + } else +#endif + { + ptr = LOS_MemAlloc(OS_SYS_MEM_ADDR, size);//从内存池分配 +>>>>>>> remotes/origin/main } return ptr; } +<<<<<<< HEAD +======= +/// 申请具有对齐属性的内存,申请规则:申请小于16KiB的内存则通过堆内存池获取,否则申请多个连续物理页 +>>>>>>> remotes/origin/main VOID *LOS_KernelMallocAlign(UINT32 size, UINT32 boundary) { VOID *ptr = NULL; +<<<<<<< HEAD #ifdef LOSCFG_KERNEL_VM +======= +#ifdef LOSCFG_KERNEL_VM +>>>>>>> remotes/origin/main if (OsMemLargeAlloc(size) && IS_ALIGNED(PAGE_SIZE, boundary)) { ptr = LOS_PhysPagesAllocContiguous(ROUNDUP(size, PAGE_SIZE) >> PAGE_SHIFT); } else @@ -1154,7 +1868,11 @@ VOID *LOS_KernelMallocAlign(UINT32 size, UINT32 boundary) return ptr; } +<<<<<<< HEAD +======= +/// 重新分配内核内存空间 +>>>>>>> remotes/origin/main VOID *LOS_KernelRealloc(VOID *ptr, UINT32 size) { VOID *tmpPtr = NULL; @@ -1162,6 +1880,10 @@ VOID *LOS_KernelRealloc(VOID *ptr, UINT32 size) #ifdef LOSCFG_KERNEL_VM LosVmPage *page = NULL; errno_t ret; +<<<<<<< HEAD +======= + +>>>>>>> remotes/origin/main if (ptr == NULL) { tmpPtr = LOS_KernelMalloc(size); } else { @@ -1198,7 +1920,11 @@ VOID LOS_KernelFree(VOID *ptr) { #ifdef LOSCFG_KERNEL_VM UINT32 ret; +<<<<<<< HEAD if (OsMemIsHeapNode(ptr) == FALSE) { +======= + if (OsMemIsHeapNode(ptr) == FALSE) {//判断地址是否在堆区 +>>>>>>> remotes/origin/main ret = OsMemLargeNodeFree(ptr); if (ret != LOS_OK) { VM_ERR("KernelFree %p failed", ptr); @@ -1207,6 +1933,12 @@ VOID LOS_KernelFree(VOID *ptr) } else #endif { +<<<<<<< HEAD (VOID)LOS_MemFree(OS_SYS_MEM_ADDR, ptr); } } +======= + (VOID)LOS_MemFree(OS_SYS_MEM_ADDR, ptr);//从内存池中释放 + } +} +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_page.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_page.c index 650fd0b0..17eba645 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_page.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_page.c @@ -40,6 +40,7 @@ #ifdef LOSCFG_KERNEL_VM +<<<<<<< HEAD LosVmPage *g_vmPageArray = NULL; size_t g_vmPageArraySize; @@ -51,15 +52,35 @@ STATIC VOID OsVmPageInit(LosVmPage *page, paddr_t pa, UINT8 segID) page->physAddr = pa; page->segID = segID; page->order = VM_LIST_ORDER_MAX; +======= +LosVmPage *g_vmPageArray = NULL;//物理页框数组 +size_t g_vmPageArraySize;//物理页框大小 +//页框初始化 +STATIC VOID OsVmPageInit(LosVmPage *page, paddr_t pa, UINT8 segID) +{ + LOS_ListInit(&page->node);//页节点初始化 + page->flags = FILE_PAGE_FREE;//页标签,初始为空闲页 + LOS_AtomicSet(&page->refCounts, 0);//引用次数0 + page->physAddr = pa;//物理地址 + page->segID = segID;//物理地址使用段管理,段ID + page->order = VM_LIST_ORDER_MAX;//初始化值,不属于任何块组 +>>>>>>> remotes/origin/main page->nPages = 0; #ifdef LOSCFG_PAGE_TABLE_FINE_LOCK LOS_SpinInit(&page->lock); #endif } +<<<<<<< HEAD STATIC INLINE VOID OsVmPageOrderListInit(LosVmPage *page, size_t nPages) { OsVmPhysPagesFreeContiguous(page, nPages); +======= +///伙伴算法初始化 +STATIC INLINE VOID OsVmPageOrderListInit(LosVmPage *page, size_t nPages) +{//@note_why 此时所有页面 page->order = VM_LIST_ORDER_MAX,能挂入伙伴算法的链表吗? + OsVmPhysPagesFreeContiguous(page, nPages);//释放连续的物理页框 +>>>>>>> remotes/origin/main } #define VMPAGEINIT(page, pa, segID) do { \ @@ -68,6 +89,13 @@ STATIC INLINE VOID OsVmPageOrderListInit(LosVmPage *page, size_t nPages) (pa) += PAGE_SIZE; \ } while (0) +<<<<<<< HEAD +======= +/*! + 完成对物理内存整体初始化,本函数一定运行在实模式下 + 1.申请大块内存g_vmPageArray存放LosVmPage,按4K一页划分物理内存存放在数组中. +*/ +>>>>>>> remotes/origin/main VOID OsVmPageStartup(VOID) { struct VmPhysSeg *seg = NULL; @@ -76,7 +104,11 @@ VOID OsVmPageStartup(VOID) UINT32 nPage; INT32 segID; +<<<<<<< HEAD OsVmPhysAreaSizeAdjust(ROUNDUP((g_vmBootMemBase - KERNEL_ASPACE_BASE), PAGE_SIZE)); +======= + OsVmPhysAreaSizeAdjust(ROUNDUP((g_vmBootMemBase - KERNEL_ASPACE_BASE), PAGE_SIZE));//校正 g_physArea size +>>>>>>> remotes/origin/main /* * Pages getting from OsVmPhysPageNumGet() interface here contain the memory @@ -85,6 +117,7 @@ VOID OsVmPageStartup(VOID) */ UINT32 pageNum = OsVmPhysPageNumGet(); nPage = pageNum * PAGE_SIZE / (sizeof(LosVmPage) + PAGE_SIZE); +<<<<<<< HEAD g_vmPageArraySize = nPage * sizeof(LosVmPage); g_vmPageArray = (LosVmPage *)OsVmBootMemAlloc(g_vmPageArraySize); @@ -92,13 +125,28 @@ VOID OsVmPageStartup(VOID) OsVmPhysSegAdd(); OsVmPhysInit(); +======= + g_vmPageArraySize = nPage * sizeof(LosVmPage);//页表总大小 + g_vmPageArray = (LosVmPage *)OsVmBootMemAlloc(g_vmPageArraySize);//实模式下申请内存,此时还没有初始化MMU + + OsVmPhysAreaSizeAdjust(ROUNDUP(g_vmPageArraySize, PAGE_SIZE)); + + OsVmPhysSegAdd();// 完成对段的初始化 + OsVmPhysInit();// 加入空闲链表和设置置换算法,LRU(最近最久未使用)算法 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_PLIMITS OsMemLimitSetLimit(pageNum * PAGE_SIZE); #endif +<<<<<<< HEAD for (segID = 0; segID < g_vmPhysSegNum; segID++) { seg = &g_vmPhysSeg[segID]; nPage = seg->size >> PAGE_SHIFT; +======= + for (segID = 0; segID < g_vmPhysSegNum; segID++) {//遍历物理段,将段切成一页一页 + seg = &g_vmPhysSeg[segID]; + nPage = seg->size >> PAGE_SHIFT;//本段总页数 +>>>>>>> remotes/origin/main UINT32 count = nPage >> 3; /* 3: 2 ^ 3, nPage / 8, cycle count */ UINT32 left = nPage & 0x7; /* 0x7: nPage % 8, left page */ @@ -116,17 +164,29 @@ VOID OsVmPageStartup(VOID) for (; left > 0; left--) { VMPAGEINIT(page, pa, segID); } +<<<<<<< HEAD OsVmPageOrderListInit(seg->pageBase, nPage); } } +======= + OsVmPageOrderListInit(seg->pageBase, nPage);//伙伴算法初始化,将所有页加入空闲链表供分配 + } +} +///通过物理地址获取页框 +>>>>>>> remotes/origin/main LosVmPage *LOS_VmPageGet(PADDR_T paddr) { INT32 segID; LosVmPage *page = NULL; +<<<<<<< HEAD for (segID = 0; segID < g_vmPhysSegNum; segID++) { page = OsVmPhysToPage(paddr, segID); +======= + for (segID = 0; segID < g_vmPhysSegNum; segID++) {//物理内存采用段页管理 + page = OsVmPhysToPage(paddr, segID);//通过物理地址和段ID找出物理页框 +>>>>>>> remotes/origin/main if (page != NULL) { break; } diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_phys.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_phys.c index 2c581456..c36052ea 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_phys.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_phys.c @@ -1,3 +1,32 @@ +<<<<<<< HEAD +======= +/*! +基本概念 + 物理内存是计算机上最重要的资源之一,指的是实际的内存设备提供的、可以通过CPU总线直接进行寻址的内存空间, + 其主要作用是为操作系统及程序提供临时存储空间。LiteOS-A内核管理物理内存是通过分页实现的,除了内核堆占用的一部分内存外, + 其余可用内存均以4KiB为单位划分成页帧,内存分配和内存回收便是以页帧为单位进行操作。内核采用伙伴算法管理空闲页面, + 可以降低一定的内存碎片率,提高内存分配和释放的效率,但是一个很小的块往往也会阻塞一个大块的合并,导致不能分配较大的内存块。 +运行机制 + LiteOS-A内核的物理内存使用分布视图,主要由内核镜像、内核堆及物理页组成。内核堆部分见堆内存管理一节。 + ----------------------------------------------------- + + kernel.bin | heap | page frames + (内核镜像) | (内核堆) | (物理页框) + ----------------------------------------------------- + 伙伴算法把所有空闲页帧分成9个内存块组,每组中内存块包含2的幂次方个页帧,例如:第0组的内存块包含2的0次方个页帧, + 即1个页帧;第8组的内存块包含2的8次方个页帧,即256个页帧。相同大小的内存块挂在同一个链表上进行管理。 + +申请内存 + 系统申请20KiB内存,按一页帧4K算,即5个页帧时,9个内存块组中索引为2的链表挂着一块大小为8个页帧的内存块满足要求,分配出20KiB内存后还剩余12KiB内存, + 即3个页帧,将3个页帧分成2的幂次方之和,即0跟1,尝试查找伙伴进行合并。2个页帧的内存块没有伙伴则直接插到索引为1的链表上, + 继续查找1个页帧的内存块是否有伙伴,索引为0的链表上此时有1个,如果两个内存块地址连续则进行合并,并将内存块挂到索引为0的链表上,否则不做处理。 +释放内存 + 系统释放12KiB内存,即3个页帧,将3个页帧分成2的幂次方之和,即2跟1,尝试查找伙伴进行合并,索引为1的链表上有1个内存块, + 若地址连续则合并,并将合并后的内存块挂到索引为2的链表上,索引为0的链表上此时也有1个,如果地址连续则进行合并, + 并将合并后的内存块挂到索引为1的链表上,此时继续判断是否有伙伴,重复上述操作。 + */ + +>>>>>>> remotes/origin/main /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. * Copyright (c) 2020-2021 Huawei Device Co., Ltd. All rights reserved. @@ -41,6 +70,7 @@ #define ONE_PAGE 1 +<<<<<<< HEAD /* Physical memory area array */ STATIC struct VmPhysArea g_physArea[] = { { @@ -52,15 +82,33 @@ STATIC struct VmPhysArea g_physArea[] = { struct VmPhysSeg g_vmPhysSeg[VM_PHYS_SEG_MAX]; INT32 g_vmPhysSegNum = 0; +======= +/* Physical memory area array | 物理内存区数组 */ +STATIC struct VmPhysArea g_physArea[] = {///< 这里只有一个区域,即只生成一个段 + { + .start = SYS_MEM_BASE, //整个物理内存基地址,#define SYS_MEM_BASE DDR_MEM_ADDR , 0x80000000 + .size = SYS_MEM_SIZE_DEFAULT,//整个物理内存总大小 0x07f00000 + }, +}; + +struct VmPhysSeg g_vmPhysSeg[VM_PHYS_SEG_MAX]; ///< 最大32段 +INT32 g_vmPhysSegNum = 0; ///< 段数 +/// 获取段数组,全局变量,变量放在 .bbs 区 +>>>>>>> remotes/origin/main LosVmPhysSeg *OsGVmPhysSegGet(void) { return g_vmPhysSeg; } +<<<<<<< HEAD +======= +/// 初始化Lru置换链表 +>>>>>>> remotes/origin/main STATIC VOID OsVmPhysLruInit(struct VmPhysSeg *seg) { INT32 i; UINT32 intSave; +<<<<<<< HEAD LOS_SpinInit(&seg->lruLock); LOS_SpinLockSave(&seg->lruLock, &intSave); @@ -71,6 +119,18 @@ STATIC VOID OsVmPhysLruInit(struct VmPhysSeg *seg) LOS_SpinUnlockRestore(&seg->lruLock, intSave); } +======= + LOS_SpinInit(&seg->lruLock);//初始化自旋锁,自旋锁用于CPU多核同步 + + LOS_SpinLockSave(&seg->lruLock, &intSave); + for (i = 0; i < VM_NR_LRU_LISTS; i++) { //五个双循环链表 + seg->lruSize[i] = 0; //记录链表节点数 + LOS_ListInit(&seg->lruList[i]); //初始化LRU链表 + } + LOS_SpinUnlockRestore(&seg->lruLock, intSave); +} +/// 创建物理段,由区划分转成段管理 +>>>>>>> remotes/origin/main STATIC INT32 OsVmPhysSegCreate(paddr_t start, size_t size) { struct VmPhysSeg *seg = NULL; @@ -79,8 +139,13 @@ STATIC INT32 OsVmPhysSegCreate(paddr_t start, size_t size) return -1; } +<<<<<<< HEAD seg = &g_vmPhysSeg[g_vmPhysSegNum++]; for (; (seg > g_vmPhysSeg) && ((seg - 1)->start > (start + size)); seg--) { +======= + seg = &g_vmPhysSeg[g_vmPhysSegNum++];//拿到一段数据 + for (; (seg > g_vmPhysSeg) && ((seg - 1)->start > (start + size)); seg--) {//定位到合适的段 +>>>>>>> remotes/origin/main *seg = *(seg - 1); } seg->start = start; @@ -88,21 +153,35 @@ STATIC INT32 OsVmPhysSegCreate(paddr_t start, size_t size) return 0; } +<<<<<<< HEAD +======= +/// 添加物理段 +>>>>>>> remotes/origin/main VOID OsVmPhysSegAdd(VOID) { INT32 i, ret; LOS_ASSERT(g_vmPhysSegNum < VM_PHYS_SEG_MAX); +<<<<<<< HEAD for (i = 0; i < (sizeof(g_physArea) / sizeof(g_physArea[0])); i++) { ret = OsVmPhysSegCreate(g_physArea[i].start, g_physArea[i].size); +======= + + for (i = 0; i < (sizeof(g_physArea) / sizeof(g_physArea[0])); i++) {//遍历g_physArea数组 + ret = OsVmPhysSegCreate(g_physArea[i].start, g_physArea[i].size);//由区划分转成段管理 +>>>>>>> remotes/origin/main if (ret != 0) { VM_ERR("create phys seg failed"); } } } +<<<<<<< HEAD +======= +/// 段区域大小调整 +>>>>>>> remotes/origin/main VOID OsVmPhysAreaSizeAdjust(size_t size) { /* @@ -113,24 +192,38 @@ VOID OsVmPhysAreaSizeAdjust(size_t size) g_physArea[0].size -= size; } +<<<<<<< HEAD +======= +/// 获得物理内存的总页数 +>>>>>>> remotes/origin/main UINT32 OsVmPhysPageNumGet(VOID) { UINT32 nPages = 0; INT32 i; for (i = 0; i < (sizeof(g_physArea) / sizeof(g_physArea[0])); i++) { +<<<<<<< HEAD nPages += g_physArea[i].size >> PAGE_SHIFT; } return nPages; } +======= + nPages += g_physArea[i].size >> PAGE_SHIFT;//右移12位,相当于除以4K, 计算出总页数 + } + + return nPages;//返回所有物理内存总页数 +} +/// 初始化空闲链表,分配物理页框使用伙伴算法 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsVmPhysFreeListInit(struct VmPhysSeg *seg) { int i; UINT32 intSave; struct VmFreeList *list = NULL; +<<<<<<< HEAD LOS_SpinInit(&seg->freeListLock); LOS_SpinLockSave(&seg->freeListLock, &intSave); @@ -142,6 +235,19 @@ STATIC INLINE VOID OsVmPhysFreeListInit(struct VmPhysSeg *seg) LOS_SpinUnlockRestore(&seg->freeListLock, intSave); } +======= + LOS_SpinInit(&seg->freeListLock);//初始化用于分配的自旋锁 + + LOS_SpinLockSave(&seg->freeListLock, &intSave); + for (i = 0; i < VM_LIST_ORDER_MAX; i++) {//遍历伙伴算法空闲块组链表 + list = &seg->freeList[i]; //一个个来 + LOS_ListInit(&list->node); //LosVmPage.node将挂到list->node上 + list->listCnt = 0; //链表上的数量默认0 + } + LOS_SpinUnlockRestore(&seg->freeListLock, intSave); +} +/// 物理段初始化 +>>>>>>> remotes/origin/main VOID OsVmPhysInit(VOID) { struct VmPhysSeg *seg = NULL; @@ -150,6 +256,7 @@ VOID OsVmPhysInit(VOID) for (i = 0; i < g_vmPhysSegNum; i++) { seg = &g_vmPhysSeg[i]; +<<<<<<< HEAD seg->pageBase = &g_vmPageArray[nPages]; nPages += seg->size >> PAGE_SHIFT; OsVmPhysFreeListInit(seg); @@ -157,6 +264,15 @@ VOID OsVmPhysInit(VOID) } } +======= + seg->pageBase = &g_vmPageArray[nPages];//记录本段首页物理页框地址 + nPages += seg->size >> PAGE_SHIFT;//偏移12位,按4K一页,算出本段总页数 + OsVmPhysFreeListInit(seg); //初始化空闲链表,分配页框使用伙伴算法 + OsVmPhysLruInit(seg); //初始化LRU置换链表 + } +} +/// 将页框挂入空闲链表,分配物理页框从空闲链表里拿 +>>>>>>> remotes/origin/main STATIC VOID OsVmPhysFreeListAddUnsafe(LosVmPage *page, UINT8 order) { struct VmPhysSeg *seg = NULL; @@ -173,12 +289,17 @@ STATIC VOID OsVmPhysFreeListAddUnsafe(LosVmPage *page, UINT8 order) LOS_ListTailInsert(&list->node, &page->node); list->listCnt++; } +<<<<<<< HEAD +======= +///将物理页框从空闲链表上摘除,见于物理页框被分配的情况 +>>>>>>> remotes/origin/main STATIC VOID OsVmPhysFreeListDelUnsafe(LosVmPage *page) { struct VmPhysSeg *seg = NULL; struct VmFreeList *list = NULL; +<<<<<<< HEAD if ((page->segID >= VM_PHYS_SEG_MAX) || (page->order >= VM_LIST_ORDER_MAX)) { LOS_Panic("The page segment id(%u) or order(%u) is invalid\n", page->segID, page->order); } @@ -190,11 +311,33 @@ STATIC VOID OsVmPhysFreeListDelUnsafe(LosVmPage *page) page->order = VM_LIST_ORDER_MAX; } +======= + if ((page->segID >= VM_PHYS_SEG_MAX) || (page->order >= VM_LIST_ORDER_MAX)) {//等于VM_LIST_ORDER_MAX也不行,说明伙伴算法最大支持 2^8的分配 + LOS_Panic("The page segment id(%u) or order(%u) is invalid\n", page->segID, page->order); + } + + seg = &g_vmPhysSeg[page->segID]; //找到物理页框对应的段 + list = &seg->freeList[page->order]; //根据伙伴算法组序号找到空闲链表 + list->listCnt--; //链表节点总数减一 + LOS_ListDelete(&page->node); //将自己从链表上摘除 + page->order = VM_LIST_ORDER_MAX; //告诉系统物理页框已不在空闲链表上, 用于OsVmPhysPagesSpiltUnsafe的断言 +} + + +/** + * @brief 本函数很像卖猪肉的,拿一大块肉剁,先把多余的放回到小块肉堆里去. + * @param page + * @param oldOrder 原本要买 2^2肉 + * @param newOrder 却找到个 2^8肉块 + * @return STATIC + */ +>>>>>>> remotes/origin/main STATIC VOID OsVmPhysPagesSpiltUnsafe(LosVmPage *page, UINT8 oldOrder, UINT8 newOrder) { UINT32 order; LosVmPage *buddyPage = NULL; +<<<<<<< HEAD for (order = newOrder; order > oldOrder;) { order--; buddyPage = &page[VM_ORDER_TO_PAGES(order)]; @@ -203,6 +346,16 @@ STATIC VOID OsVmPhysPagesSpiltUnsafe(LosVmPage *page, UINT8 oldOrder, UINT8 newO } } +======= + for (order = newOrder; order > oldOrder;) {//把肉剁碎的过程,把多余的肉块切成2^7,2^6...标准块, + order--;//越切越小,逐一挂到对应的空闲链表上 + buddyPage = &page[VM_ORDER_TO_PAGES(order)];//@note_good 先把多余的肉割出来,这句代码很赞!因为LosVmPage本身是在一个大数组上,page[nPages]可直接定位 + LOS_ASSERT(buddyPage->order == VM_LIST_ORDER_MAX);//没挂到伙伴算法对应组块空闲链表上的物理页框的order必须是VM_LIST_ORDER_MAX + OsVmPhysFreeListAddUnsafe(buddyPage, order);//将劈开的节点挂到对应序号的链表上,buddyPage->order = order + } +} +///通过物理地址获取所属参数段的物理页框 +>>>>>>> remotes/origin/main LosVmPage *OsVmPhysToPage(paddr_t pa, UINT8 segID) { struct VmPhysSeg *seg = NULL; @@ -216,8 +369,13 @@ LosVmPage *OsVmPhysToPage(paddr_t pa, UINT8 segID) return NULL; } +<<<<<<< HEAD offset = pa - seg->start; return (seg->pageBase + (offset >> PAGE_SHIFT)); +======= + offset = pa - seg->start;//得到物理地址的偏移量 + return (seg->pageBase + (offset >> PAGE_SHIFT));//得到对应的物理页框 +>>>>>>> remotes/origin/main } LosVmPage *OsVmPaddrToPage(paddr_t paddr) @@ -233,6 +391,7 @@ LosVmPage *OsVmPaddrToPage(paddr_t paddr) } return NULL; } +<<<<<<< HEAD VOID *OsVmPageToVaddr(LosVmPage *page) { @@ -252,12 +411,43 @@ LosVmPage *OsVmVaddrToPage(VOID *ptr) seg = &g_vmPhysSeg[segID]; if ((pa >= seg->start) && (pa < (seg->start + seg->size))) { return seg->pageBase + ((pa - seg->start) >> PAGE_SHIFT); +======= +/*! + * @brief 通过page获取内核空间的虚拟地址 参考OsArchMmuInit + \n #define SYS_MEM_BASE DDR_MEM_ADDR /* physical memory base 物理地址的起始地址 * / + \n 本函数非常重要,通过一个物理地址找到内核虚拟地址 + \n 内核静态映射:提升虚实转化效率,段映射减少页表项 + * @param page + * @return VOID* + */ +VOID *OsVmPageToVaddr(LosVmPage *page)// +{ + VADDR_T vaddr; + vaddr = KERNEL_ASPACE_BASE + page->physAddr - SYS_MEM_BASE;//表示申请的物理地址在物理空间的偏移量等于映射的虚拟地址在内核空间的偏移量 + return (VOID *)(UINTPTR)vaddr;//不需要存储映射关系,这简直就是神来之笔,拍案叫绝。@note_good 详见 鸿蒙内核源码分析(页表管理篇) +} +///通过虚拟地址找映射的物理页框 +LosVmPage *OsVmVaddrToPage(VOID *ptr) +{ + struct VmPhysSeg *seg = NULL; + PADDR_T pa = LOS_PaddrQuery(ptr);//通过空间的虚拟地址查询物理地址 + UINT32 segID; + + for (segID = 0; segID < g_vmPhysSegNum; segID++) {//遍历所有段 + seg = &g_vmPhysSeg[segID]; + if ((pa >= seg->start) && (pa < (seg->start + seg->size))) {//找到物理地址所在的段 + return seg->pageBase + ((pa - seg->start) >> PAGE_SHIFT);//段基地址+页偏移索引 得到虚拟地址经映射所在物理页框 +>>>>>>> remotes/origin/main } } return NULL; } +<<<<<<< HEAD +======= +/// 回收一定范围内的页框 +>>>>>>> remotes/origin/main STATIC INLINE VOID OsVmRecycleExtraPages(LosVmPage *page, size_t startPage, size_t endPage) { if (startPage >= endPage) { @@ -266,7 +456,11 @@ STATIC INLINE VOID OsVmRecycleExtraPages(LosVmPage *page, size_t startPage, size OsVmPhysPagesFreeContiguous(page, endPage - startPage); } +<<<<<<< HEAD +======= +/// 大块的物理内存分配 +>>>>>>> remotes/origin/main STATIC LosVmPage *OsVmPhysLargeAlloc(struct VmPhysSeg *seg, size_t nPages) { struct VmFreeList *list = NULL; @@ -276,11 +470,19 @@ STATIC LosVmPage *OsVmPhysLargeAlloc(struct VmPhysSeg *seg, size_t nPages) PADDR_T paEnd; size_t size = nPages << PAGE_SHIFT; +<<<<<<< HEAD list = &seg->freeList[VM_LIST_ORDER_MAX - 1]; LOS_DL_LIST_FOR_EACH_ENTRY(page, &list->node, LosVmPage, node) { paStart = page->physAddr; paEnd = paStart + size; if (paEnd > (seg->start + seg->size)) { +======= + list = &seg->freeList[VM_LIST_ORDER_MAX - 1];//先找伙伴算法中内存块最大的开撸 + LOS_DL_LIST_FOR_EACH_ENTRY(page, &list->node, LosVmPage, node) {//遍历链表 + paStart = page->physAddr; + paEnd = paStart + size; + if (paEnd > (seg->start + seg->size)) {//匹配物理地址范围 +>>>>>>> remotes/origin/main continue; } @@ -302,7 +504,11 @@ STATIC LosVmPage *OsVmPhysLargeAlloc(struct VmPhysSeg *seg, size_t nPages) return NULL; } +<<<<<<< HEAD +======= +/// 申请物理页并挂在对应的链表上 +>>>>>>> remotes/origin/main STATIC LosVmPage *OsVmPhysPagesAlloc(struct VmPhysSeg *seg, size_t nPages) { struct VmFreeList *list = NULL; @@ -312,6 +518,7 @@ STATIC LosVmPage *OsVmPhysPagesAlloc(struct VmPhysSeg *seg, size_t nPages) UINT32 newOrder; order = OsVmPagesToOrder(nPages); +<<<<<<< HEAD if (order < VM_LIST_ORDER_MAX) { for (newOrder = order; newOrder < VM_LIST_ORDER_MAX; newOrder++) { list = &seg->freeList[newOrder]; @@ -319,6 +526,15 @@ STATIC LosVmPage *OsVmPhysPagesAlloc(struct VmPhysSeg *seg, size_t nPages) continue; } page = LOS_DL_LIST_ENTRY(LOS_DL_LIST_FIRST(&list->node), LosVmPage, node); +======= + if (order < VM_LIST_ORDER_MAX) {//按正常的伙伴算法分配 + for (newOrder = order; newOrder < VM_LIST_ORDER_MAX; newOrder++) {//从小往大了撸 + list = &seg->freeList[newOrder]; + if (LOS_ListEmpty(&list->node)) {//这条链路上没有可分配的物理页框 + continue;//继续往大的找 + } + page = LOS_DL_LIST_ENTRY(LOS_DL_LIST_FIRST(&list->node), LosVmPage, node);//找到了直接返回第一个节点 +>>>>>>> remotes/origin/main goto DONE; } } else { @@ -339,7 +555,11 @@ DONE: return page; } +<<<<<<< HEAD +======= +/// 释放物理页框,所谓释放物理页就是把页挂到空闲链表中 +>>>>>>> remotes/origin/main VOID OsVmPhysPagesFree(LosVmPage *page, UINT8 order) { paddr_t pa; @@ -349,6 +569,7 @@ VOID OsVmPhysPagesFree(LosVmPage *page, UINT8 order) return; } +<<<<<<< HEAD if (order < VM_LIST_ORDER_MAX - 1) { pa = VM_PAGE_TO_PHYS(page); do { @@ -358,21 +579,39 @@ VOID OsVmPhysPagesFree(LosVmPage *page, UINT8 order) break; } OsVmPhysFreeListDelUnsafe(buddyPage); +======= + if (order < VM_LIST_ORDER_MAX - 1) {//order[0,7] + pa = VM_PAGE_TO_PHYS(page);//获取物理地址 + do {//按位异或 + pa ^= VM_ORDER_TO_PHYS(order);//@note_good 注意这里是高位和低位的 ^= ,也就是说跳到order块组物理地址处,此处处理甚妙! + buddyPage = OsVmPhysToPage(pa, page->segID);//通过物理地址拿到页框 + if ((buddyPage == NULL) || (buddyPage->order != order)) {//页框所在组块必须要对应 + break; + } + OsVmPhysFreeListDelUnsafe(buddyPage);//注意buddypage是连续的物理页框 例如order=2时,2^2=4页就是一个块组 |_|_|_|_| +>>>>>>> remotes/origin/main order++; pa &= ~(VM_ORDER_TO_PHYS(order) - 1); page = OsVmPhysToPage(pa, page->segID); } while (order < VM_LIST_ORDER_MAX - 1); } +<<<<<<< HEAD OsVmPhysFreeListAddUnsafe(page, order); } +======= + OsVmPhysFreeListAddUnsafe(page, order);//伙伴算法 空闲节点增加 +} +///连续的释放物理页框, 如果8页连在一块是一起释放的 +>>>>>>> remotes/origin/main VOID OsVmPhysPagesFreeContiguous(LosVmPage *page, size_t nPages) { paddr_t pa; UINT32 order; size_t n; +<<<<<<< HEAD while (TRUE) { pa = VM_PAGE_TO_PHYS(page); order = VM_PHYS_TO_ORDER(pa); @@ -394,6 +633,37 @@ VOID OsVmPhysPagesFreeContiguous(LosVmPage *page, size_t nPages) } } +======= + while (TRUE) {//死循环 + pa = VM_PAGE_TO_PHYS(page);//获取页面物理地址 + order = VM_PHYS_TO_ORDER(pa);//通过物理地址找到伙伴算法的级别 + n = VM_ORDER_TO_PAGES(order);//通过级别找到物理页块 (1< nPages) {//nPages只剩下小于2^order时,退出循环 + break; + } + OsVmPhysPagesFree(page, order);//释放伙伴算法对应块组 + nPages -= n;//总页数减少 + page += n;//释放的页数增多 + } + //举例剩下 7个页框时,依次用 2^2 2^1 2^0 方式释放 + while (nPages > 0) { + order = LOS_HighBitGet(nPages);//从高到低块组释放 + n = VM_ORDER_TO_PAGES(order);//2^order次方 + OsVmPhysPagesFree(page, order);//释放块组 + nPages -= n; + page += n;//相当于page[n] + } +} + +/*! + * @brief OsVmPhysPagesGet 获取一定数量的页框 LosVmPage实体是放在全局大数组中的, + * LosVmPage->nPages 标记了分配页数 + * @param nPages + * @return + * + * @see + */ +>>>>>>> remotes/origin/main STATIC LosVmPage *OsVmPhysPagesGet(size_t nPages) { UINT32 intSave; @@ -404,11 +674,19 @@ STATIC LosVmPage *OsVmPhysPagesGet(size_t nPages) for (segID = 0; segID < g_vmPhysSegNum; segID++) { seg = &g_vmPhysSeg[segID]; LOS_SpinLockSave(&seg->freeListLock, &intSave); +<<<<<<< HEAD page = OsVmPhysPagesAlloc(seg, nPages); if (page != NULL) { /* the first page of continuous physical addresses holds refCounts */ LOS_AtomicSet(&page->refCounts, 0); page->nPages = nPages; +======= + page = OsVmPhysPagesAlloc(seg, nPages);//分配指定页数的物理页,nPages需小于伙伴算法一次能分配的最大页数 + if (page != NULL) {//分配成功 + /* */ + LOS_AtomicSet(&page->refCounts, 0);//设置引用次数为0 + page->nPages = nPages;//页数 +>>>>>>> remotes/origin/main LOS_SpinUnlockRestore(&seg->freeListLock, intSave); return page; } @@ -416,7 +694,11 @@ STATIC LosVmPage *OsVmPhysPagesGet(size_t nPages) } return NULL; } +<<<<<<< HEAD +======= +///分配连续的物理页 +>>>>>>> remotes/origin/main VOID *LOS_PhysPagesAllocContiguous(size_t nPages) { LosVmPage *page = NULL; @@ -424,15 +706,26 @@ VOID *LOS_PhysPagesAllocContiguous(size_t nPages) if (nPages == 0) { return NULL; } +<<<<<<< HEAD page = OsVmPhysPagesGet(nPages); +======= + //鸿蒙 nPages 不能大于 2^8 次方,即256个页,1M内存,仅限于内核态,用户态不限制分配大小. + page = OsVmPhysPagesGet(nPages);//通过伙伴算法获取物理上连续的页 +>>>>>>> remotes/origin/main if (page == NULL) { return NULL; } +<<<<<<< HEAD return OsVmPageToVaddr(page); } +======= + return OsVmPageToVaddr(page);//通过物理页找虚拟地址 +} +/// 释放指定页数地址连续的物理内存 +>>>>>>> remotes/origin/main VOID LOS_PhysPagesFreeContiguous(VOID *ptr, size_t nPages) { UINT32 intSave; @@ -443,17 +736,29 @@ VOID LOS_PhysPagesFreeContiguous(VOID *ptr, size_t nPages) return; } +<<<<<<< HEAD page = OsVmVaddrToPage(ptr); +======= + page = OsVmVaddrToPage(ptr);//通过虚拟地址找到页框 +>>>>>>> remotes/origin/main if (page == NULL) { VM_ERR("vm page of ptr(%#x) is null", ptr); return; } +<<<<<<< HEAD page->nPages = 0; +======= + page->nPages = 0;//被分配的页数置为0,表示不被分配 +>>>>>>> remotes/origin/main seg = &g_vmPhysSeg[page->segID]; LOS_SpinLockSave(&seg->freeListLock, &intSave); +<<<<<<< HEAD OsVmPhysPagesFreeContiguous(page, nPages); +======= + OsVmPhysPagesFreeContiguous(page, nPages);//具体释放实现 +>>>>>>> remotes/origin/main LOS_SpinUnlockRestore(&seg->freeListLock, intSave); #ifdef LOSCFG_KERNEL_PLIMITS @@ -468,7 +773,11 @@ PADDR_T OsKVaddrToPaddr(VADDR_T kvaddr) } return (kvaddr - KERNEL_ASPACE_BASE + SYS_MEM_BASE); } +<<<<<<< HEAD +======= +/// 通过物理地址获取内核虚拟地址 +>>>>>>> remotes/origin/main VADDR_T *LOS_PaddrToKVaddr(PADDR_T paddr) { struct VmPhysSeg *seg = NULL; @@ -484,10 +793,17 @@ VADDR_T *LOS_PaddrToKVaddr(PADDR_T paddr) return (VADDR_T *)(UINTPTR)(paddr - SYS_MEM_BASE + KERNEL_ASPACE_BASE); } } +<<<<<<< HEAD return (VADDR_T *)(UINTPTR)(paddr - SYS_MEM_BASE + KERNEL_ASPACE_BASE); } +======= + //内核 + return (VADDR_T *)(UINTPTR)(paddr - SYS_MEM_BASE + KERNEL_ASPACE_BASE);// +} +///释放一个物理页框 +>>>>>>> remotes/origin/main VOID LOS_PhysPageFree(LosVmPage *page) { UINT32 intSave; @@ -497,12 +813,21 @@ VOID LOS_PhysPageFree(LosVmPage *page) return; } +<<<<<<< HEAD if (LOS_AtomicDecRet(&page->refCounts) <= 0) { seg = &g_vmPhysSeg[page->segID]; LOS_SpinLockSave(&seg->freeListLock, &intSave); OsVmPhysPagesFreeContiguous(page, ONE_PAGE); LOS_AtomicSet(&page->refCounts, 0); +======= + if (LOS_AtomicDecRet(&page->refCounts) <= 0) {//减少引用数后不能小于0 + seg = &g_vmPhysSeg[page->segID]; + LOS_SpinLockSave(&seg->freeListLock, &intSave); + + OsVmPhysPagesFreeContiguous(page, ONE_PAGE);//释放一页 + LOS_AtomicSet(&page->refCounts, 0);//只要物理内存被释放了,引用数就必须得重置为 0 +>>>>>>> remotes/origin/main LOS_SpinUnlockRestore(&seg->freeListLock, intSave); } @@ -510,12 +835,31 @@ VOID LOS_PhysPageFree(LosVmPage *page) OsMemLimitMemFree(PAGE_SIZE); #endif } +<<<<<<< HEAD LosVmPage *LOS_PhysPageAlloc(VOID) { return OsVmPhysPagesGet(ONE_PAGE); } +======= +/// 申请一个物理页 +LosVmPage *LOS_PhysPageAlloc(VOID) +{ + return OsVmPhysPagesGet(ONE_PAGE);//分配一页物理页 +} + +/*! + * @brief LOS_PhysPagesAlloc 分配nPages页个物理页框,并将页框挂入list + \n 返回已分配的页面大小,不负责一定能分配到nPages的页框 + * + * @param list + * @param nPages + * @return + * + * @see + */ +>>>>>>> remotes/origin/main size_t LOS_PhysPagesAlloc(size_t nPages, LOS_DL_LIST *list) { LosVmPage *page = NULL; @@ -526,17 +870,29 @@ size_t LOS_PhysPagesAlloc(size_t nPages, LOS_DL_LIST *list) } while (nPages--) { +<<<<<<< HEAD page = OsVmPhysPagesGet(ONE_PAGE); if (page == NULL) { break; } LOS_ListTailInsert(list, &page->node); +======= + page = OsVmPhysPagesGet(ONE_PAGE);//一页一页分配,由伙伴算法分配 + if (page == NULL) { + break; + } + LOS_ListTailInsert(list, &page->node);//从参数链表list尾部挂入新页面结点 +>>>>>>> remotes/origin/main count++; } return count; } +<<<<<<< HEAD +======= +///拷贝共享页面 +>>>>>>> remotes/origin/main VOID OsPhysSharePageCopy(PADDR_T oldPaddr, PADDR_T *newPaddr, LosVmPage *newPage) { UINT32 intSave; @@ -550,12 +906,17 @@ VOID OsPhysSharePageCopy(PADDR_T oldPaddr, PADDR_T *newPaddr, LosVmPage *newPage return; } +<<<<<<< HEAD oldPage = LOS_VmPageGet(oldPaddr); +======= + oldPage = LOS_VmPageGet(oldPaddr);//由物理地址得到页框 +>>>>>>> remotes/origin/main if (oldPage == NULL) { VM_ERR("invalid oldPaddr %p", oldPaddr); return; } +<<<<<<< HEAD seg = &g_vmPhysSeg[oldPage->segID]; LOS_SpinLockSave(&seg->freeListLock, &intSave); if (LOS_AtomicRead(&oldPage->refCounts) == 1) { @@ -573,20 +934,49 @@ VOID OsPhysSharePageCopy(PADDR_T oldPaddr, PADDR_T *newPaddr, LosVmPage *newPage LOS_AtomicInc(&newPage->refCounts); LOS_AtomicDec(&oldPage->refCounts); +======= + seg = &g_vmPhysSeg[oldPage->segID];//拿到物理段 + LOS_SpinLockSave(&seg->freeListLock, &intSave); + if (LOS_AtomicRead(&oldPage->refCounts) == 1) {//页面引用次数仅一次,说明只有一个进程在操作 + *newPaddr = oldPaddr;//新老指向同一块物理地址 + } else {//是个共享内存 + newMem = LOS_PaddrToKVaddr(*newPaddr); //新页虚拟地址 + oldMem = LOS_PaddrToKVaddr(oldPaddr); //老页虚拟地址 + if ((newMem == NULL) || (oldMem == NULL)) { + LOS_SpinUnlockRestore(&seg->freeListLock, intSave); + return; + }//请记住,在保护模式下,物理地址只能用于计算,操作(包括拷贝)需要虚拟地址! + if (memcpy_s(newMem, PAGE_SIZE, oldMem, PAGE_SIZE) != EOK) {//老页内容复制给新页,需操作虚拟地址,拷贝一页数据 + VM_ERR("memcpy_s failed"); + } + + LOS_AtomicInc(&newPage->refCounts);//新页引用次数以原子方式自动减量 + LOS_AtomicDec(&oldPage->refCounts);//老页引用次数以原子方式自动减量 +>>>>>>> remotes/origin/main } LOS_SpinUnlockRestore(&seg->freeListLock, intSave); return; } +<<<<<<< HEAD +======= +///获取物理页框所在段 +>>>>>>> remotes/origin/main struct VmPhysSeg *OsVmPhysSegGet(LosVmPage *page) { if ((page == NULL) || (page->segID >= VM_PHYS_SEG_MAX)) { return NULL; } +<<<<<<< HEAD return (OsGVmPhysSegGet() + page->segID); } +======= + return (OsGVmPhysSegGet() + page->segID);//等用于OsGVmPhysSegGet()[page->segID] +} +///获取参数nPages对应的块组,例如 7 -> 2^3 返回 3 +>>>>>>> remotes/origin/main UINT32 OsVmPagesToOrder(size_t nPages) { UINT32 order; @@ -595,7 +985,11 @@ UINT32 OsVmPagesToOrder(size_t nPages) return order; } +<<<<<<< HEAD +======= +///释放双链表中的所有节点内存,本质是回归到伙伴orderlist中 +>>>>>>> remotes/origin/main size_t LOS_PhysPagesFree(LOS_DL_LIST *list) { UINT32 intSave; @@ -608,6 +1002,7 @@ size_t LOS_PhysPagesFree(LOS_DL_LIST *list) return 0; } +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(page, nPage, list, LosVmPage, node) { LOS_ListDelete(&page->node); if (LOS_AtomicDecRet(&page->refCounts) <= 0) { @@ -618,6 +1013,18 @@ size_t LOS_PhysPagesFree(LOS_DL_LIST *list) LOS_SpinUnlockRestore(&seg->freeListLock, intSave); } count++; +======= + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(page, nPage, list, LosVmPage, node) {//宏循环 + LOS_ListDelete(&page->node);//先把自己摘出去 + if (LOS_AtomicDecRet(&page->refCounts) <= 0) {//无引用 + seg = &g_vmPhysSeg[page->segID];//获取物理段 + LOS_SpinLockSave(&seg->freeListLock, &intSave);//锁住freeList + OsVmPhysPagesFreeContiguous(page, ONE_PAGE);//连续释放,注意这里的ONE_PAGE其实有误导,让人以为是释放4K,其实是指连续的物理页框,如果3页连在一块是一起释放的. + LOS_AtomicSet(&page->refCounts, 0);//引用重置为0 + LOS_SpinUnlockRestore(&seg->freeListLock, intSave);//恢复锁 + } + count++;//继续取下一个node +>>>>>>> remotes/origin/main } return count; @@ -632,4 +1039,7 @@ VADDR_T *LOS_PaddrToKVaddr(PADDR_T paddr) return (VADDR_T *)DMA_TO_VMM_ADDR(paddr); } #endif +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_scan.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_scan.c index ab467818..7c79155d 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_scan.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_scan.c @@ -37,6 +37,13 @@ #ifdef LOSCFG_KERNEL_VM /* unmap a lru page by map record info caller need lru lock */ +<<<<<<< HEAD +======= +/************************************************************************************************** + 解除文件页和进程(mmu)的映射关系 + 参数info记录了进程的MMU +**************************************************************************************************/ +>>>>>>> remotes/origin/main VOID OsUnmapPageLocked(LosFilePage *page, LosMapInfo *info) { if (page == NULL || info == NULL) { @@ -47,21 +54,33 @@ VOID OsUnmapPageLocked(LosFilePage *page, LosMapInfo *info) LOS_ListDelete(&info->node); LOS_AtomicDec(&page->vmPage->refCounts); LOS_ArchMmuUnmap(info->archMmu, info->vaddr, 1); +<<<<<<< HEAD LOS_MemFree(m_aucSysMem0, info); } +======= + LOS_MemFree(m_aucSysMem0, info);//释放虚拟 +} +///解除文件页在所有进程的映射 +>>>>>>> remotes/origin/main VOID OsUnmapAllLocked(LosFilePage *page) { LosMapInfo *info = NULL; LosMapInfo *next = NULL; LOS_DL_LIST *immap = &page->i_mmap; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(info, next, immap, LosMapInfo, node) { +======= + + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(info, next, immap, LosMapInfo, node) {//遍历 immap->info 链表 +>>>>>>> remotes/origin/main OsUnmapPageLocked(page, info); } } /* add a new lru node to lru list, lruType can be file or anon */ +<<<<<<< HEAD VOID OsLruCacheAdd(LosFilePage *fpage, enum OsLruList lruType) { UINT32 intSave; @@ -129,6 +148,75 @@ STATIC INLINE VOID OsMoveToInactiveHead(LosFilePage *fpage) LosVmPhysSeg *physSeg = fpage->physSeg; LOS_ListDelete(&fpage->lru); LOS_ListTailInsert(&physSeg->lruList[VM_LRU_INACTIVE_FILE], &fpage->lru); +======= +VOID OsLruCacheAdd(LosFilePage *fpage, enum OsLruList lruType)//在lru列表中添加一个新的lru节点,lruType可以是文件或匿名 +{ + UINT32 intSave; + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + LosVmPage *page = fpage->vmPage; //得到物理页面 + + LOS_SpinLockSave(&physSeg->lruLock, &intSave);//自旋锁:最多只能被一个内核持有,CPU内核 互斥锁 + OsSetPageActive(page); //设置页面为活动页 + OsCleanPageReferenced(page);//清除页面被引用位 + physSeg->lruSize[lruType]++; //lruType页总size++ + LOS_ListTailInsert(&physSeg->lruList[lruType], &fpage->lru);//加入lruType页双循环链表中 + + LOS_SpinUnlockRestore(&physSeg->lruLock, intSave);//解锁 +} + +/* dellete a lru node, caller need hold lru_lock */ +VOID OsLruCacheDel(LosFilePage *fpage)//删除lru节点,调用者需要拿到lru锁 +{ + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + int type = OsIsPageActive(fpage->vmPage) ? VM_LRU_ACTIVE_FILE : VM_LRU_INACTIVE_FILE;//得到页面LRU类型 + + physSeg->lruSize[type]--; //type页总size-- + LOS_ListDelete(&fpage->lru);//将自己从lru链表中摘出来 +} +///非活动文件页低于活动文件页吗 +BOOL OsInactiveListIsLow(LosVmPhysSeg *physSeg) +{ + return (physSeg->lruSize[VM_LRU_ACTIVE_FILE] > + physSeg->lruSize[VM_LRU_INACTIVE_FILE]) ? TRUE : FALSE;//直接对比size,效率杠杠的 +} + +/* move a page from inactive list to active list head */ +STATIC INLINE VOID OsMoveToActiveList(LosFilePage *fpage)//将页面从非活动列表移动到活动列表 +{ + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + + physSeg->lruSize[VM_LRU_ACTIVE_FILE]++; //活动页总size++ + physSeg->lruSize[VM_LRU_INACTIVE_FILE]--; //不活动页总size-- + LOS_ListDelete(&fpage->lru); //将自己从lru链表中摘出来 + LOS_ListTailInsert(&physSeg->lruList[VM_LRU_ACTIVE_FILE], &fpage->lru);//加入活动页双循环链表中 +} + +/* move a page from active list to inactive list head */ +STATIC INLINE VOID OsMoveToInactiveList(LosFilePage *fpage)//将页面从活动列表移动到非活动列表 +{ + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + + physSeg->lruSize[VM_LRU_ACTIVE_FILE]--; //活动页总size-- + physSeg->lruSize[VM_LRU_INACTIVE_FILE]++; //不活动页总size++ + LOS_ListDelete(&fpage->lru); //将自己从lru链表中摘出来 + LOS_ListTailInsert(&physSeg->lruList[VM_LRU_INACTIVE_FILE], &fpage->lru);//加入不活动页双循环链表中 +} + +/* move a page to the most active pos in lru list(active head) *///将页面移至lru列表中最活跃的位置 +STATIC INLINE VOID OsMoveToActiveHead(LosFilePage *fpage) +{ + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + LOS_ListDelete(&fpage->lru); //将自己从lru链表中摘出来 + LOS_ListTailInsert(&physSeg->lruList[VM_LRU_ACTIVE_FILE], &fpage->lru);//加入活动页双循环链表中 +} + +/* move a page to the most active pos in lru list(inactive head) */ +STATIC INLINE VOID OsMoveToInactiveHead(LosFilePage *fpage)//鸿蒙会从inactive链表的尾部开始进行回收,跟linux一样 +{ + LosVmPhysSeg *physSeg = fpage->physSeg; //得到页面对应段 + LOS_ListDelete(&fpage->lru); //将自己从lru链表中摘出来 + LOS_ListTailInsert(&physSeg->lruList[VM_LRU_INACTIVE_FILE], &fpage->lru);//加入不活动页双循环链表中 +>>>>>>> remotes/origin/main } /* page referced add: (call by page cache get) @@ -138,7 +226,11 @@ ref:0, act:0 --> ref:1, act:0 ref:1, act:0 --> ref:0, act:1 ref:0, act:1 --> ref:1, act:1 */ +<<<<<<< HEAD VOID OsPageRefIncLocked(LosFilePage *fpage) +======= +VOID OsPageRefIncLocked(LosFilePage *fpage)// ref ,act 标签转换功能 +>>>>>>> remotes/origin/main { BOOL isOrgActive; UINT32 intSave; @@ -148,6 +240,7 @@ VOID OsPageRefIncLocked(LosFilePage *fpage) return; } +<<<<<<< HEAD LOS_SpinLockSave(&fpage->physSeg->lruLock, &intSave); page = fpage->vmPage; @@ -158,6 +251,18 @@ VOID OsPageRefIncLocked(LosFilePage *fpage) OsSetPageActive(page); } else if (!OsIsPageReferenced(page)) { OsSetPageReferenced(page); +======= + LOS_SpinLockSave(&fpage->physSeg->lruLock, &intSave);//要处理lruList,先拿锁 + + page = fpage->vmPage;//拿到物理页框 + isOrgActive = OsIsPageActive(page);//页面是否在活动 + + if (OsIsPageReferenced(page) && !OsIsPageActive(page)) {//身兼 不活动和引用标签 + OsCleanPageReferenced(page);//撕掉引用标签 ref:1, act:0 --> ref:0, act:1 + OsSetPageActive(page); //贴上活动标签 + } else if (!OsIsPageReferenced(page)) { + OsSetPageReferenced(page);//ref:0, act:0 --> ref:1, act:0 +>>>>>>> remotes/origin/main } if (!isOrgActive && OsIsPageActive(page)) { @@ -175,14 +280,22 @@ VOID OsPageRefIncLocked(LosFilePage *fpage) LOS_SpinUnlockRestore(&fpage->physSeg->lruLock, intSave); } +<<<<<<< HEAD /* page referced dec: (call by shrinker) +======= +/* page referced dec: (call by thrinker) +>>>>>>> remotes/origin/main ----------inactive----------|----------active------------ [ref:0,act:0], [ref:1,act:0]|[ref:0,act:1], [ref:1,act:1] ref:1, act:1 --> ref:0, act:1 ref:0, act:1 --> ref:1, act:0 ref:1, act:0 --> ref:0, act:0 */ +<<<<<<< HEAD VOID OsPageRefDecNoLock(LosFilePage *fpage) +======= +VOID OsPageRefDecNoLock(LosFilePage *fpage) // ref ,act 标签转换功能 +>>>>>>> remotes/origin/main { BOOL isOrgActive; LosVmPage *page = NULL; @@ -194,7 +307,11 @@ VOID OsPageRefDecNoLock(LosFilePage *fpage) page = fpage->vmPage; isOrgActive = OsIsPageActive(page); +<<<<<<< HEAD if (!OsIsPageReferenced(page) && OsIsPageActive(page)) { +======= + if (!OsIsPageReferenced(page) && OsIsPageActive(page)) {//[ref:0,act:1]的情况 +>>>>>>> remotes/origin/main OsCleanPageActive(page); OsSetPageReferenced(page); } else if (OsIsPageReferenced(page)) { @@ -205,13 +322,18 @@ VOID OsPageRefDecNoLock(LosFilePage *fpage) OsMoveToInactiveList(fpage); } } +<<<<<<< HEAD +======= +///缩小活动页链表 +>>>>>>> remotes/origin/main VOID OsShrinkActiveList(LosVmPhysSeg *physSeg, int nScan) { LosFilePage *fpage = NULL; LosFilePage *fnext = NULL; LOS_DL_LIST *activeFile = &physSeg->lruList[VM_LRU_ACTIVE_FILE]; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, activeFile, LosFilePage, lru) { if (LOS_SpinTrylock(&fpage->mapping->list_lock) != LOS_OK) { continue; @@ -231,13 +353,38 @@ VOID OsShrinkActiveList(LosVmPhysSeg *physSeg, int nScan) OsPageRefDecNoLock(fpage); LOS_SpinUnlock(&fpage->mapping->list_lock); +======= + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, activeFile, LosFilePage, lru) {//一页一页处理 + if (LOS_SpinTrylock(&fpage->mapping->list_lock) != LOS_OK) {//尝试获取文件页所在的page_mapping锁 + continue;//接着处理下一文件页 + } + + /* happend when caller hold cache lock and try reclaim this page *///调用方持有缓存锁并尝试回收此页时发生 + if (OsIsPageLocked(fpage->vmPage)) {//页面是否被锁 + LOS_SpinUnlock(&fpage->mapping->list_lock);//失败时,一定要释放page_mapping锁. + continue;//接着处理下一文件页 + } + + if (OsIsPageMapped(fpage) && (fpage->flags & VM_MAP_REGION_FLAG_PERM_EXECUTE)) {//文件页是否被映射而且是个可执行文件 ? + LOS_SpinUnlock(&fpage->mapping->list_lock);//是时,一定要释放page_mapping锁. + continue;//接着处理下一文件页 + } + //找了可以收缩的文件页 + OsPageRefDecNoLock(fpage); //将页面移到未活动文件链表 + + LOS_SpinUnlock(&fpage->mapping->list_lock); //释放page_mapping锁. +>>>>>>> remotes/origin/main if (--nScan <= 0) { break; } } } +<<<<<<< HEAD +======= +///缩小未活动页链表 +>>>>>>> remotes/origin/main int OsShrinkInactiveList(LosVmPhysSeg *physSeg, int nScan, LOS_DL_LIST *list) { UINT32 nrReclaimed = 0; @@ -248,6 +395,7 @@ int OsShrinkInactiveList(LosVmPhysSeg *physSeg, int nScan, LOS_DL_LIST *list) LosFilePage *ftemp = NULL; LOS_DL_LIST *inactive_file = &physSeg->lruList[VM_LRU_INACTIVE_FILE]; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, inactive_file, LosFilePage, lru) { flock = &fpage->mapping->list_lock; @@ -278,6 +426,38 @@ int OsShrinkInactiveList(LosVmPhysSeg *physSeg, int nScan, LOS_DL_LIST *list) nrReclaimed++; if (--nScan <= 0) { +======= + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, inactive_file, LosFilePage, lru) {//遍历链表一页一页处理 + flock = &fpage->mapping->list_lock; + + if (LOS_SpinTrylock(flock) != LOS_OK) {//尝试获取文件页所在的page_mapping锁 + continue;//接着处理下一文件页 + } + + page = fpage->vmPage;//获取物理页框 + if (OsIsPageLocked(page)) {//页面是否被锁 + LOS_SpinUnlock(flock); + continue;//接着处理下一文件页 + } + + if (OsIsPageMapped(fpage) && (OsIsPageDirty(page) || (fpage->flags & VM_MAP_REGION_FLAG_PERM_EXECUTE))) { + LOS_SpinUnlock(flock);//文件页是否被映射而且是个脏页获取是个可执行文件 ? + continue;//接着处理下一文件页 + } + + if (OsIsPageDirty(page)) {//是脏页 + ftemp = OsDumpDirtyPage(fpage);//备份脏页 + if (ftemp != NULL) {//备份成功了 + LOS_ListTailInsert(list, &ftemp->node);//将脏页挂到参数链表上带走 + } + } + + OsDeletePageCacheLru(fpage);//将文件页从LRU和pagecache上摘除 + LOS_SpinUnlock(flock); + nrReclaimed++;//成功回收了一页 + + if (--nScan <= 0) {//继续回收 +>>>>>>> remotes/origin/main break; } } @@ -286,25 +466,38 @@ int OsShrinkInactiveList(LosVmPhysSeg *physSeg, int nScan, LOS_DL_LIST *list) } #ifdef LOSCFG_FS_VFS +<<<<<<< HEAD int OsTryShrinkMemory(size_t nPage) +======= +int OsTryShrinkMemory(size_t nPage)//尝试收缩文件页 +>>>>>>> remotes/origin/main { UINT32 intSave; size_t totalPages; size_t nReclaimed = 0; LosVmPhysSeg *physSeg = NULL; UINT32 index; +<<<<<<< HEAD LOS_DL_LIST_HEAD(dirtyList); +======= + LOS_DL_LIST_HEAD(dirtyList);//初始化脏页链表,上面将挂所有脏页用于同步到磁盘后回收 +>>>>>>> remotes/origin/main LosFilePage *fpage = NULL; LosFilePage *fnext = NULL; if (nPage == 0) { +<<<<<<< HEAD nPage = VM_FILEMAP_MIN_SCAN; +======= + nPage = VM_FILEMAP_MIN_SCAN;// +>>>>>>> remotes/origin/main } if (nPage > VM_FILEMAP_MAX_SCAN) { nPage = VM_FILEMAP_MAX_SCAN; } +<<<<<<< HEAD for (index = 0; index < g_vmPhysSegNum; index++) { physSeg = &g_vmPhysSeg[index]; LOS_SpinLockSave(&physSeg->lruLock, &intSave); @@ -328,6 +521,31 @@ int OsTryShrinkMemory(size_t nPage) LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, &dirtyList, LosFilePage, node) { OsDoFlushDirtyPage(fpage); +======= + for (index = 0; index < g_vmPhysSegNum; index++) {//遍历整个物理段组 + physSeg = &g_vmPhysSeg[index];//一段段来 + LOS_SpinLockSave(&physSeg->lruLock, &intSave); + totalPages = physSeg->lruSize[VM_LRU_ACTIVE_FILE] + physSeg->lruSize[VM_LRU_INACTIVE_FILE];//统计所有文件页 + if (totalPages < VM_FILEMAP_MIN_SCAN) {//文件页占用内存不多的情况下,怎么处理? + LOS_SpinUnlockRestore(&physSeg->lruLock, intSave); + continue;//放过这一段,找下一段 + } + + if (OsInactiveListIsLow(physSeg)) { + OsShrinkActiveList(physSeg, (nPage < VM_FILEMAP_MIN_SCAN) ? VM_FILEMAP_MIN_SCAN : nPage);//缩小活动页 + } + + nReclaimed += OsShrinkInactiveList(physSeg, nPage, &dirtyList);//缩小未活动页,带出脏页链表 + LOS_SpinUnlockRestore(&physSeg->lruLock, intSave); + + if (nReclaimed >= nPage) {//够了,够了,达到目的了. + break;//退出收缩 + } + } + + LOS_DL_LIST_FOR_EACH_ENTRY_SAFE(fpage, fnext, &dirtyList, LosFilePage, node) {//遍历处理脏页数据 + OsDoFlushDirtyPage(fpage);//冲洗脏页数据,将脏页数据回写磁盘 +>>>>>>> remotes/origin/main } return nReclaimed; @@ -340,4 +558,8 @@ int OsTryShrinkMemory(size_t nPage) #endif #endif +<<<<<<< HEAD +#endif +======= #endif +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/los_vm_syscall.c b/src/kernel_liteos_a/kernel/base/vm/los_vm_syscall.c index 6aa23993..b0ea89e0 100644 --- a/src/kernel_liteos_a/kernel/base/vm/los_vm_syscall.c +++ b/src/kernel_liteos_a/kernel/base/vm/los_vm_syscall.c @@ -65,7 +65,11 @@ STATUS_T OsCheckMMapParams(VADDR_T *vaddr, unsigned long flags, size_t len, unsi return -EINVAL; } +<<<<<<< HEAD if ((flags & MAP_SUPPORT_MASK) == 0) { +======= + if ((flags & MAP_SUPPORT_MASK) == 0) {//映射权限限制 +>>>>>>> remotes/origin/main return -EINVAL; } if (((flags & MAP_SHARED_PRIVATE) == 0) || ((flags & MAP_SHARED_PRIVATE) == MAP_SHARED_PRIVATE)) { @@ -95,20 +99,66 @@ STATUS_T OsNamedMmapingPermCheck(struct file *filep, unsigned long flags, unsign return LOS_OK; } +<<<<<<< HEAD +======= +///匿名映射 +>>>>>>> remotes/origin/main STATUS_T OsAnonMMap(LosVmMapRegion *region) { LOS_SetRegionTypeAnon(region); return LOS_OK; } +<<<<<<< HEAD +======= +/** + mmap基础概念: + 一种内存映射文件的方法,即将一个文件或者其它对象映射到进程的地址空间,实现文件磁盘地址和进程虚拟地址空间中一段虚拟地址的一一对映关系. + 实现这样的映射关系后,进程就可以采用指针的方式读写操作这一段内存,而系统会自动回写脏页面到对应的文件磁盘上, + 即完成了对文件的操作而不必再调用read,write等系统调用函数。相反,内核空间对这段区域的修改也直接反映用户空间, + 从而可以实现不同进程间的文件共享。 + + https://www.cnblogs.com/huxiao-tee/p/4660352.html + http://abcdxyzk.github.io/blog/2015/09/11/kernel-mm-mmap/ + + 参数 描述 + addr 指向欲映射的内存起始地址,通常设为 NULL,代表让系统自动选定地址,映射成功后返回该地址。 + length 代表将文件中多大的部分映射到内存。 + prot 用于设置内存段的访问权限,有如下权限: + PROT_EXEC 映射区域可被执行 + PROT_READ 映射区域可被读取 + PROT_WRITE 映射区域可被写入 + PROT_NONE 映射区域不能存取 + + flags 控制程序对内存段的改变所造成的影响,有如下属性: + MAP_FIXED 如果参数start所指的地址无法成功建立映射时,则放弃映射,不对地址做修正。通常不鼓励用此旗标。 + MAP_SHARED 对映射区域的写入数据会复制回文件内,而且允许其他映射该文件的进程共享。 + MAP_PRIVATE 对映射区域的写入操作会产生一个映射文件的复制,即私人的“写入时复制”(copy on write)对此区域作的任何修改都不会写回原来的文件内容。 + MAP_ANONYMOUS建立匿名映射。此时会忽略参数fd,不涉及文件,而且映射区域无法和其他进程共享。 + MAP_DENYWRITE只允许对映射区域的写入操作,其他对文件直接写入的操作将会被拒绝。 + MAP_LOCKED 将映射区域锁定住,这表示该区域不会被置换(swap)。 + + fd: 要映射到内存中的文件描述符。如果使用匿名内存映射时,即flags中设置了MAP_ANONYMOUS,fd设为-1。 + 有些系统不支持匿名内存映射,则可以使用fopen打开/dev/zero文件,然后对该文件进行映射,可以同样达到匿名内存映射的效果。 + + offset 文件映射的偏移量,通常设置为0,代表从文件最前方开始对应,offset必须是PAGE_SIZE的整数倍。 + 成功返回:虚拟内存地址,这地址是页对齐。 + 失败返回:(void *)-1。 + */ +>>>>>>> remotes/origin/main VADDR_T LOS_MMap(VADDR_T vaddr, size_t len, unsigned prot, unsigned long flags, int fd, unsigned long pgoff) { STATUS_T status; VADDR_T resultVaddr; UINT32 regionFlags; +<<<<<<< HEAD LosVmMapRegion *newRegion = NULL; struct file *filep = NULL; +======= + LosVmMapRegion *newRegion = NULL;//应用的内存分配对应到内核就是分配一个线性区 + struct file *filep = NULL;// inode : file = 1:N ,一对多关系,一个inode可以被多个进程打开,返回不同的file但都指向同一个inode +>>>>>>> remotes/origin/main LosVmSpace *vmSpace = OsCurrProcessGet()->vmSpace; len = ROUNDUP(len, PAGE_SIZE); @@ -116,9 +166,15 @@ VADDR_T LOS_MMap(VADDR_T vaddr, size_t len, unsigned prot, unsigned long flags, if (checkRst != LOS_OK) { return checkRst; } +<<<<<<< HEAD if (LOS_IsNamedMapping(flags)) { status = fs_getfilep(fd, &filep); +======= + + if (LOS_IsNamedMapping(flags)) {//是否文件映射 + status = fs_getfilep(fd, &filep);//获取文件描述符和状态 +>>>>>>> remotes/origin/main if (status < 0) { return -EBADF; } @@ -131,6 +187,7 @@ VADDR_T LOS_MMap(VADDR_T vaddr, size_t len, unsigned prot, unsigned long flags, (VOID)LOS_MuxAcquire(&vmSpace->regionMux); /* user mode calls mmap to release heap physical memory without releasing heap virtual space */ +<<<<<<< HEAD status = OsUserHeapFree(vmSpace, vaddr, len); if (status == LOS_OK) { resultVaddr = vaddr; @@ -155,6 +212,32 @@ VADDR_T LOS_MMap(VADDR_T vaddr, size_t len, unsigned prot, unsigned long flags, if (status != LOS_OK) { LOS_RbDelNode(&vmSpace->regionRbTree, &newRegion->rbNode); LOS_RegionFree(vmSpace, newRegion); +======= + status = OsUserHeapFree(vmSpace, vaddr, len);//用户模式释放堆物理内存而不释放堆虚拟空间 + if (status == LOS_OK) {//OsUserHeapFree 干两件事 1.解除映射关系 2.释放物理页 + resultVaddr = vaddr; + goto MMAP_DONE; + } + //地址不在堆区 + regionFlags = OsCvtProtFlagsToRegionFlags(prot, flags);//将参数flag转换Region的flag + newRegion = LOS_RegionAlloc(vmSpace, vaddr, len, regionFlags, pgoff);//分配一个线性区 + if (newRegion == NULL) { + resultVaddr = (VADDR_T)-ENOMEM;//ENOMEM:内存溢出 + goto MMAP_DONE; + } + newRegion->regionFlags |= VM_MAP_REGION_FLAG_MMAP; + resultVaddr = newRegion->range.base;//线性区基地址为分配的地址 + + if (LOS_IsNamedMapping(flags)) { + status = OsNamedMMap(filep, newRegion);//文件映射 + } else { + status = OsAnonMMap(newRegion);//匿名映射 + } + + if (status != LOS_OK) { + LOS_RbDelNode(&vmSpace->regionRbTree, &newRegion->rbNode);//从红黑树和双循环链表中删除 + LOS_RegionFree(vmSpace, newRegion);//释放 +>>>>>>> remotes/origin/main resultVaddr = (VADDR_T)-ENOMEM; goto MMAP_DONE; } @@ -163,7 +246,11 @@ MMAP_DONE: (VOID)LOS_MuxRelease(&vmSpace->regionMux); return resultVaddr; } +<<<<<<< HEAD +======= +///解除映射关系 +>>>>>>> remotes/origin/main STATUS_T LOS_UnMMap(VADDR_T addr, size_t size) { if ((addr <= 0) || (size == 0)) { @@ -172,7 +259,10 @@ STATUS_T LOS_UnMMap(VADDR_T addr, size_t size) return OsUnMMap(OsCurrProcessGet()->vmSpace, addr, size); } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main STATIC INLINE BOOL OsProtMprotectPermCheck(unsigned long prot, LosVmMapRegion *region) { UINT32 protFlags = 0; @@ -186,11 +276,16 @@ STATIC INLINE BOOL OsProtMprotectPermCheck(unsigned long prot, LosVmMapRegion *r return ((protFlags & permFlags) == protFlags); } +<<<<<<< HEAD +======= +/// 收缩堆区 +>>>>>>> remotes/origin/main VOID *OsShrinkHeap(VOID *addr, LosVmSpace *space) { VADDR_T newBrk, oldBrk; +<<<<<<< HEAD newBrk = LOS_Align((VADDR_T)(UINTPTR)addr, PAGE_SIZE); oldBrk = LOS_Align(space->heapNow, PAGE_SIZE); if (LOS_UnMMap(newBrk, (oldBrk - newBrk)) < 0) { @@ -200,6 +295,25 @@ VOID *OsShrinkHeap(VOID *addr, LosVmSpace *space) return addr; } +======= + newBrk = LOS_Align((VADDR_T)(UINTPTR)addr, PAGE_SIZE);//新堆顶 + oldBrk = LOS_Align(space->heapNow, PAGE_SIZE);//旧堆顶 + if (LOS_UnMMap(newBrk, (oldBrk - newBrk)) < 0) {//解除相差区的映射 + return (void *)(UINTPTR)space->heapNow;//解除失败就持续现有的 + } + space->heapNow = (VADDR_T)(UINTPTR)addr;//返回新堆顶 + return addr; +} + +/** + 用户进程向内核申请空间,进一步说用于扩展用户堆栈空间,或者回收用户堆栈空间 + 扩展当前进程的堆空间 + 一个进程所有的线性区都在进程指定的线性地址范围内, + 线性区之间是不会有地址的重叠的,开始都是连续的,随着进程的运行出现了释放再分配的情况 + 由此出现了断断续续的线性区,内核回收线性区时会检测是否和周边的线性区可合并成一个更大 + 的线性区用于分配。 + */ +>>>>>>> remotes/origin/main VOID *LOS_DoBrk(VOID *addr) { LosVmSpace *space = OsCurrProcessGet()->vmSpace; @@ -209,6 +323,7 @@ VOID *LOS_DoBrk(VOID *addr) VOID *alignAddr = NULL; VOID *shrinkAddr = NULL; +<<<<<<< HEAD if (addr == NULL) { return (void *)(UINTPTR)space->heapNow; } @@ -225,25 +340,55 @@ VOID *LOS_DoBrk(VOID *addr) (VOID)LOS_MuxAcquire(&space->regionMux); if (addr < (VOID *)(UINTPTR)space->heapNow) { shrinkAddr = OsShrinkHeap(addr, space); +======= + if (addr == NULL) {//参数地址未传情况 + return (void *)(UINTPTR)space->heapNow;//以现有指向地址为基础进行扩展 + } + + if ((UINTPTR)addr < (UINTPTR)space->heapBase) {//heapBase是堆区的开始地址,所以参数地址不能低于它 + return (VOID *)-ENOMEM; + } + + size = (UINTPTR)addr - (UINTPTR)space->heapBase;//算出大小 + size = ROUNDUP(size, PAGE_SIZE); //圆整size + alignAddr = (CHAR *)(UINTPTR)(space->heapBase) + size;//得到新的线性区的结束地址 + PRINT_INFO("brk addr %p , size 0x%x, alignAddr %p, align %d\n", addr, size, alignAddr, PAGE_SIZE); + + (VOID)LOS_MuxAcquire(&space->regionMux); + if (addr < (VOID *)(UINTPTR)space->heapNow) {//如果地址小于堆区现地址 + shrinkAddr = OsShrinkHeap(addr, space);//收缩堆区 +>>>>>>> remotes/origin/main (VOID)LOS_MuxRelease(&space->regionMux); return shrinkAddr; } +<<<<<<< HEAD if ((UINTPTR)alignAddr >= space->mapBase) { VM_ERR("Process heap memory space is insufficient"); +======= + if ((UINTPTR)alignAddr >= space->mapBase) {//参数地址 大于映射区地址 + VM_ERR("Process heap memory space is insufficient");//进程堆空间不足 +>>>>>>> remotes/origin/main ret = (VOID *)-ENOMEM; goto REGION_ALLOC_FAILED; } +<<<<<<< HEAD if (space->heapBase == space->heapNow) { region = LOS_RegionAlloc(space, space->heapBase, size, VM_MAP_REGION_FLAG_PERM_READ | VM_MAP_REGION_FLAG_PERM_WRITE | +======= + if (space->heapBase == space->heapNow) {//往往是第一次调用本函数才会出现,因为初始化时 heapBase = heapNow + region = LOS_RegionAlloc(space, space->heapBase, size,//分配一个可读/可写/可使用的线性区,只需分配一次 + VM_MAP_REGION_FLAG_PERM_READ | VM_MAP_REGION_FLAG_PERM_WRITE |//线性区的大小由range.size决定 +>>>>>>> remotes/origin/main VM_MAP_REGION_FLAG_FIXED | VM_MAP_REGION_FLAG_PERM_USER, 0); if (region == NULL) { ret = (VOID *)-ENOMEM; VM_ERR("LOS_RegionAlloc failed"); goto REGION_ALLOC_FAILED; } +<<<<<<< HEAD region->regionFlags |= VM_MAP_REGION_FLAG_HEAP; space->heap = region; } @@ -251,18 +396,36 @@ VOID *LOS_DoBrk(VOID *addr) space->heapNow = (VADDR_T)(UINTPTR)alignAddr; space->heap->range.size = size; ret = (VOID *)(UINTPTR)space->heapNow; +======= + region->regionFlags |= VM_MAP_REGION_FLAG_HEAP;//贴上线性区类型为堆区的标签,注意一个线性区可以有多种标签 + space->heap = region;//指定线性区为堆区 + } + + space->heapNow = (VADDR_T)(UINTPTR)alignAddr;//更新堆区顶部位置 + space->heap->range.size = size; //更新堆区大小,经此操作线性区变大或缩小了 + ret = (VOID *)(UINTPTR)space->heapNow;//返回堆顶 +>>>>>>> remotes/origin/main REGION_ALLOC_FAILED: (VOID)LOS_MuxRelease(&space->regionMux); return ret; } +<<<<<<< HEAD +======= +/// 继承老线性区的标签 +>>>>>>> remotes/origin/main STATIC UINT32 OsInheritOldRegionName(UINT32 oldRegionFlags) { UINT32 vmFlags = 0; +<<<<<<< HEAD if (oldRegionFlags & VM_MAP_REGION_FLAG_HEAP) { vmFlags |= VM_MAP_REGION_FLAG_HEAP; +======= + if (oldRegionFlags & VM_MAP_REGION_FLAG_HEAP) { //如果是从大堆区中申请的 + vmFlags |= VM_MAP_REGION_FLAG_HEAP; //线性区则贴上堆区标签 +>>>>>>> remotes/origin/main } else if (oldRegionFlags & VM_MAP_REGION_FLAG_STACK) { vmFlags |= VM_MAP_REGION_FLAG_STACK; } else if (oldRegionFlags & VM_MAP_REGION_FLAG_TEXT) { @@ -277,7 +440,11 @@ STATIC UINT32 OsInheritOldRegionName(UINT32 oldRegionFlags) return vmFlags; } +<<<<<<< HEAD +======= +///修改内存段的访问权限 +>>>>>>> remotes/origin/main INT32 LOS_DoMprotect(VADDR_T vaddr, size_t len, unsigned long prot) { LosVmSpace *space = OsCurrProcessGet()->vmSpace; @@ -287,7 +454,11 @@ INT32 LOS_DoMprotect(VADDR_T vaddr, size_t len, unsigned long prot) int ret; (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD region = LOS_RegionFind(space, vaddr); +======= + region = LOS_RegionFind(space, vaddr);//通过虚拟地址找到线性区 +>>>>>>> remotes/origin/main if (!IS_ALIGNED(vaddr, PAGE_SIZE) || (region == NULL) || (vaddr > vaddr + len)) { ret = -EINVAL; goto OUT_MPROTECT; @@ -297,19 +468,30 @@ INT32 LOS_DoMprotect(VADDR_T vaddr, size_t len, unsigned long prot) ret = -EINVAL; goto OUT_MPROTECT; } +<<<<<<< HEAD +======= + //如果是堆区或VDSO区,说明区内容是不能修改的 +>>>>>>> remotes/origin/main if ((region->regionFlags & VM_MAP_REGION_FLAG_VDSO) || (region->regionFlags & VM_MAP_REGION_FLAG_HEAP)) { ret = -EPERM; goto OUT_MPROTECT; } +<<<<<<< HEAD +======= + //如果是共享文件,说明内容也不能修改 +>>>>>>> remotes/origin/main if (LOS_IsRegionTypeFile(region) && (region->regionFlags & VM_MAP_REGION_FLAG_SHARED)) { if (!OsProtMprotectPermCheck(prot, region)) { ret = -EACCES; goto OUT_MPROTECT; } } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main len = LOS_Align(len, PAGE_SIZE); /* can't operation cross region */ if ((region->range.base + region->range.size) < (vaddr + len)) { @@ -318,11 +500,19 @@ INT32 LOS_DoMprotect(VADDR_T vaddr, size_t len, unsigned long prot) } /* if only move some part of region, we need to split first */ +<<<<<<< HEAD if (region->range.size > len) { OsVmRegionAdjust(space, vaddr, len); } vmFlags = OsCvtProtFlagsToRegionFlags(prot, 0); +======= + if (region->range.size > len) {//如果只修改部分区域,我们需要先拆分区 + OsVmRegionAdjust(space, vaddr, len);//调整下线性区范围 + } + + vmFlags = OsCvtProtFlagsToRegionFlags(prot, 0);//转换FLAGS +>>>>>>> remotes/origin/main vmFlags |= (region->regionFlags & VM_MAP_REGION_FLAG_SHARED) ? VM_MAP_REGION_FLAG_SHARED : 0; vmFlags |= OsInheritOldRegionName(region->regionFlags); region = LOS_RegionFind(space, vaddr); @@ -332,7 +522,11 @@ INT32 LOS_DoMprotect(VADDR_T vaddr, size_t len, unsigned long prot) } region->regionFlags = vmFlags; count = len >> PAGE_SHIFT; +<<<<<<< HEAD ret = LOS_ArchMmuChangeProt(&space->archMmu, vaddr, count, region->regionFlags); +======= + ret = LOS_ArchMmuChangeProt(&space->archMmu, vaddr, count, region->regionFlags);//修改访问权限实体函数 +>>>>>>> remotes/origin/main if (ret) { ret = -ENOMEM; goto OUT_MPROTECT; @@ -387,7 +581,11 @@ STATUS_T OsMremapCheck(VADDR_T addr, size_t oldLen, VADDR_T newAddr, size_t newL } } +<<<<<<< HEAD /* avoid new region overlapping with the old one */ +======= + /* avoid new region overlaping with the old one */ +>>>>>>> remotes/origin/main if (flags & MREMAP_FIXED) { if (((region->range.base + region->range.size) > newAddr) && (region->range.base < (newAddr + newLen))) { @@ -401,7 +599,11 @@ STATUS_T OsMremapCheck(VADDR_T addr, size_t oldLen, VADDR_T newAddr, size_t newL return LOS_OK; } +<<<<<<< HEAD +======= +///重新映射虚拟内存地址。 +>>>>>>> remotes/origin/main VADDR_T LOS_DoMremap(VADDR_T oldAddress, size_t oldSize, size_t newSize, int flags, VADDR_T newAddr) { LosVmMapRegion *regionOld = NULL; @@ -496,7 +698,11 @@ OUT_MREMAP: (VOID)LOS_MuxRelease(&space->regionMux); return ret; } +<<<<<<< HEAD +======= +///输出内存线性区 +>>>>>>> remotes/origin/main VOID LOS_DumpMemRegion(VADDR_T vaddr) { LosVmSpace *space = NULL; @@ -506,6 +712,7 @@ VOID LOS_DumpMemRegion(VADDR_T vaddr) return; } +<<<<<<< HEAD if (LOS_IsRangeInSpace(space, ROUNDDOWN(vaddr, MB), MB) == FALSE) { return; } @@ -515,3 +722,13 @@ VOID LOS_DumpMemRegion(VADDR_T vaddr) } #endif +======= + if (LOS_IsRangeInSpace(space, ROUNDDOWN(vaddr, MB), MB) == FALSE) {//是否在空间范围内 + return; + } + + OsDumpPte(vaddr);//dump L1 L2 + OsDumpAspace(space);//dump 空间 +} +#endif +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/oom.c b/src/kernel_liteos_a/kernel/base/vm/oom.c index 72996781..192b3276 100644 --- a/src/kernel_liteos_a/kernel/base/vm/oom.c +++ b/src/kernel_liteos_a/kernel/base/vm/oom.c @@ -47,8 +47,13 @@ #ifdef LOSCFG_KERNEL_VM +<<<<<<< HEAD LITE_OS_SEC_BSS OomCB *g_oomCB = NULL; static SPIN_LOCK_INIT(g_oomSpinLock); +======= +LITE_OS_SEC_BSS OomCB *g_oomCB = NULL; //全局内存溢出控制块 +static SPIN_LOCK_INIT(g_oomSpinLock);//内存溢出自旋锁 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR STATIC UINT32 OomScoreProcess(LosProcessCB *candidateProcess) { @@ -57,20 +62,32 @@ LITE_OS_SEC_TEXT_MINOR STATIC UINT32 OomScoreProcess(LosProcessCB *candidateProc #ifndef LOSCFG_KERNEL_SMP (VOID)LOS_MuxAcquire(&candidateProcess->vmSpace->regionMux); #endif +<<<<<<< HEAD /* we only consider actual physical memory here. */ +======= + /* we only consider actual physical memory here. */ //只考虑实际的物理内存 +>>>>>>> remotes/origin/main OsUProcessPmUsage(candidateProcess->vmSpace, NULL, &actualPm); #ifndef LOSCFG_KERNEL_SMP (VOID)LOS_MuxRelease(&candidateProcess->vmSpace->regionMux); #endif return actualPm; } +<<<<<<< HEAD +======= +///用于设置 g_oomCB->processVictimCB 回调函数 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR STATIC UINT32 OomKillProcess(UINTPTR param) { /* we will not kill process, and do nothing here */ return LOS_OK; } +<<<<<<< HEAD +======= +///强制收缩内存 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR STATIC UINT32 OomForceShrinkMemory(VOID) { UINT32 i; @@ -80,13 +97,21 @@ LITE_OS_SEC_TEXT_MINOR STATIC UINT32 OomForceShrinkMemory(VOID) * TryShrinkMemory maybe reclaim 0 pages in the first time from active list * to inactive list, and in the second time reclaim memory from inactive list. */ +<<<<<<< HEAD +======= + //TryShrinkMemory可能会在第一时间从活动列表中回收0页到非活动列表,并在第二次从非活动列表中回收内存。 +>>>>>>> remotes/origin/main for (i = 0; i < MAX_SHRINK_PAGECACHE_TRY; i++) { reclaimMemPages += OsTryShrinkMemory(0); } return reclaimMemPages; } +<<<<<<< HEAD +======= +///内存不足时回收页高速缓存 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR STATIC BOOL OomReclaimPageCache(VOID) { UINT32 totalPm = 0; @@ -96,14 +121,21 @@ LITE_OS_SEC_TEXT_MINOR STATIC BOOL OomReclaimPageCache(VOID) UINT32 i; for (i = 0; i < MAX_SHRINK_PAGECACHE_TRY; i++) { +<<<<<<< HEAD OsVmPhysUsedInfoGet(&usedPm, &totalPm); isReclaimMemory = ((totalPm - usedPm) << PAGE_SHIFT) < g_oomCB->reclaimMemThreshold; if (isReclaimMemory) { +======= + OsVmPhysUsedInfoGet(&usedPm, &totalPm);//获取总的和已经使用的物理内存数量 + isReclaimMemory = ((totalPm - usedPm) << PAGE_SHIFT) < g_oomCB->reclaimMemThreshold;//检查是否过了回收门槛 + if (isReclaimMemory) {//要回收了 +>>>>>>> remotes/origin/main /* * we do force memory reclaim from page cache here. * if we get memory, we will reclaim pagecache memory again. * if there is no memory to reclaim, we will return. */ +<<<<<<< HEAD reclaimMemPages = OomForceShrinkMemory(); if (reclaimMemPages > 0) { continue; @@ -120,19 +152,47 @@ LITE_OS_SEC_TEXT_MINOR STATIC BOOL OomReclaimPageCache(VOID) * return is kill process or not. */ LITE_OS_SEC_TEXT_MINOR BOOL OomCheckProcess(VOID) +======= + //在这里强制从页缓存中回收内存, + reclaimMemPages = OomForceShrinkMemory();//强制回收内存 + if (reclaimMemPages > 0) {//如果得到内存,将再次回收pagecache内存 + continue; + } + } + break;//实在没有内存可回收 + } + + return isReclaimMemory;//返回回收的数量 +} + +/* + * check is low memory or not, if low memory, try to kill process. + * return is kill process or not. + */ +LITE_OS_SEC_TEXT_MINOR BOOL OomCheckProcess(VOID)//检查内存是否不足,如果内存不足,请尝试终止进程,返回是否kill进程 +>>>>>>> remotes/origin/main { UINT32 totalPm; UINT32 usedPm; BOOL isLowMemory = FALSE; /* +<<<<<<< HEAD * spinlock the current core schedule, make sure oom process atomic * spinlock other place entering OomCheckProcess, make sure oom process mutex +======= + * spinlock the current core schedule, make sure oom process atomic //旋转锁定当前核心计划,确保oom进程原子化 + * spinlock other place entering OomCheckProcess, make sure oom process mutex //旋转锁定其他进入OomCheckProcess的地方,确保oom进程互斥 +>>>>>>> remotes/origin/main */ LOS_SpinLock(&g_oomSpinLock); /* first we will check if we need to reclaim pagecache memory */ +<<<<<<< HEAD if (OomReclaimPageCache() == FALSE) { +======= + if (OomReclaimPageCache() == FALSE) {// +>>>>>>> remotes/origin/main LOS_SpinUnlock(&g_oomSpinLock); goto NO_VICTIM_PROCESS; } @@ -140,9 +200,13 @@ LITE_OS_SEC_TEXT_MINOR BOOL OomCheckProcess(VOID) /* get free bytes */ OsVmPhysUsedInfoGet(&usedPm, &totalPm); isLowMemory = ((totalPm - usedPm) << PAGE_SHIFT) < g_oomCB->lowMemThreshold; +<<<<<<< HEAD LOS_SpinUnlock(&g_oomSpinLock); +======= + LOS_SpinUnlock(&g_oomSpinLock); +>>>>>>> remotes/origin/main if (isLowMemory) { PRINTK("[oom] OS is in low memory state\n" "total physical memory: %#x(byte), used: %#x(byte)," @@ -155,6 +219,7 @@ NO_VICTIM_PROCESS: return isLowMemory; } +<<<<<<< HEAD #ifdef LOSCFG_ENABLE_OOM_LOOP_TASK STATIC VOID OomWriteEvent(VOID) { @@ -163,6 +228,16 @@ STATIC VOID OomWriteEvent(VOID) #endif LITE_OS_SEC_TEXT_MINOR VOID OomInfodump(VOID) +======= +#ifdef LOSCFG_ENABLE_OOM_LOOP_TASK //内存溢出监测任务开关 +STATIC VOID OomWriteEvent(VOID) // OomTaskInit中创建的定时器回调 +{ + OsWriteResourceEvent(OS_RESOURCE_EVENT_OOM);//广播内存溢出事件 +} +#endif +//打印内存不足时的信息 +LITE_OS_SEC_TEXT_MINOR VOID OomInfodump(VOID) //打印内存溢出信息 +>>>>>>> remotes/origin/main { PRINTK("[oom] oom loop task status: %s\n" " oom low memory threshold: %#x(byte)\n" @@ -172,7 +247,11 @@ LITE_OS_SEC_TEXT_MINOR VOID OomInfodump(VOID) g_oomCB->lowMemThreshold, g_oomCB->reclaimMemThreshold, g_oomCB->checkInterval); } +<<<<<<< HEAD +======= +///设置低内存门槛 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OomSetLowMemThreashold(UINT32 lowMemThreshold) { if ((lowMemThreshold > OOM_DEFAULT_LOW_MEM_THRESHOLD_MAX)) { @@ -186,7 +265,11 @@ LITE_OS_SEC_TEXT_MINOR VOID OomSetLowMemThreashold(UINT32 lowMemThreshold) g_oomCB->lowMemThreshold); } } +<<<<<<< HEAD +======= +///设置回收内存的门槛 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OomSetReclaimMemThreashold(UINT32 reclaimMemThreshold) { UINT32 totalPm = 0; @@ -204,7 +287,11 @@ LITE_OS_SEC_TEXT_MINOR VOID OomSetReclaimMemThreashold(UINT32 reclaimMemThreshol g_oomCB->reclaimMemThreshold); } } +<<<<<<< HEAD +======= +///设置监控间隔 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR VOID OomSetCheckInterval(UINT32 checkInterval) { if ((checkInterval >= OOM_CHECK_MIN) && (checkInterval <= OOM_CHECK_MAX)) { @@ -216,7 +303,11 @@ LITE_OS_SEC_TEXT_MINOR VOID OomSetCheckInterval(UINT32 checkInterval) g_oomCB->checkInterval, OOM_CHECK_MIN, OOM_CHECK_MAX); } } +<<<<<<< HEAD +======= +///内存不足监控任务初始化, OOM 通过开一个软件定时器来检查内存的使用情况 +>>>>>>> remotes/origin/main LITE_OS_SEC_TEXT_MINOR UINT32 OomTaskInit(VOID) { g_oomCB = (OomCB *)LOS_MemAlloc(m_aucSysMem0, sizeof(OomCB)); @@ -225,6 +316,7 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OomTaskInit(VOID) return LOS_NOK; } +<<<<<<< HEAD g_oomCB->lowMemThreshold = OOM_DEFAULT_LOW_MEM_THRESHOLD; g_oomCB->reclaimMemThreshold = OOM_DEFAULT_RECLAIM_MEM_THRESHOLD; g_oomCB->checkInterval = OOM_DEFAULT_CHECK_INTERVAL; @@ -236,17 +328,40 @@ LITE_OS_SEC_TEXT_MINOR UINT32 OomTaskInit(VOID) g_oomCB->enabled = TRUE; UINT32 ret = LOS_SwtmrCreate(g_oomCB->checkInterval, LOS_SWTMR_MODE_PERIOD, (SWTMR_PROC_FUNC)OomWriteEvent, &g_oomCB->swtmrID, (UINTPTR)g_oomCB); +======= + g_oomCB->lowMemThreshold = OOM_DEFAULT_LOW_MEM_THRESHOLD; //运行任务的门槛 + g_oomCB->reclaimMemThreshold = OOM_DEFAULT_RECLAIM_MEM_THRESHOLD; //回收内存的门槛 + g_oomCB->checkInterval = OOM_DEFAULT_CHECK_INTERVAL; //检测时间间隔 1S + g_oomCB->processVictimCB = (OomFn)OomKillProcess; //出问题时对进程的处理函数 + g_oomCB->scoreCB = (OomFn)OomScoreProcess; //统计进程占用的物理内存 + g_oomCB->enabled = FALSE; //是否启用监控 + +#ifdef LOSCFG_ENABLE_OOM_LOOP_TASK //内存溢出检测开关 + g_oomCB->enabled = TRUE; + UINT32 ret = LOS_SwtmrCreate(g_oomCB->checkInterval, LOS_SWTMR_MODE_PERIOD, (SWTMR_PROC_FUNC)OomWriteEvent, + &g_oomCB->swtmrID, (UINTPTR)g_oomCB);//创建检测定时器 +>>>>>>> remotes/origin/main if (ret != LOS_OK) { return ret; } +<<<<<<< HEAD return LOS_SwtmrStart(g_oomCB->swtmrID); +======= + return LOS_SwtmrStart(g_oomCB->swtmrID);//启动定时器 +>>>>>>> remotes/origin/main #else return LOS_OK; #endif } +<<<<<<< HEAD LOS_MODULE_INIT(OomTaskInit, LOS_INIT_LEVEL_KMOD_TASK); #endif +======= +LOS_MODULE_INIT(OomTaskInit, LOS_INIT_LEVEL_KMOD_TASK);//初始化内存监控模块 + +#endif +>>>>>>> remotes/origin/main diff --git a/src/kernel_liteos_a/kernel/base/vm/shm.c b/src/kernel_liteos_a/kernel/base/vm/shm.c index 637c6167..6582d99f 100644 --- a/src/kernel_liteos_a/kernel/base/vm/shm.c +++ b/src/kernel_liteos_a/kernel/base/vm/shm.c @@ -1,3 +1,20 @@ +<<<<<<< HEAD +======= +/*! + 什么是共享内存 + 顾名思义,共享内存就是允许两个不相关的进程访问同一个物理内存。共享内存是在两个正在运行的进程之间 + 共享和传递数据的一种非常有效的方式。不同进程之间共享的内存通常安排为同一段物理内存。进程可以将同 + 一段共享内存连接到它们自己的地址空间中,所有进程都可以访问共享内存中的地址,就好像它们是由用C语言 + 函数malloc()分配的内存一样。而如果某个进程向共享内存写入数据,所做的改动将立即影响到可以访问同一段 + 共享内存的任何其他进程。 + + 特别提醒:共享内存并未提供同步机制,也就是说,在第一个进程结束对共享内存的写操作之前,并无自动机制 + 可以阻止第二个进程开始对它进行读取。所以我们通常需要用其他的机制来同步对共享内存的访问 + + 共享线性区可以由任意的进程创建,每个使用共享线性区都必须经过映射. + */ + +>>>>>>> remotes/origin/main /* * Copyright (c) 2013-2019 Huawei Technologies Co., Ltd. All rights reserved. * Copyright (c) 2020-2023 Huawei Device Co., Ltd. All rights reserved. @@ -55,9 +72,15 @@ #ifdef LOSCFG_KERNEL_SHM +<<<<<<< HEAD #define SHM_SEG_FREE 0x2000 #define SHM_SEG_USED 0x4000 #define SHM_SEG_REMOVE 0x8000 +======= +#define SHM_SEG_FREE 0x2000 //空闲未使用 +#define SHM_SEG_USED 0x4000 //已使用 +#define SHM_SEG_REMOVE 0x8000 //删除 +>>>>>>> remotes/origin/main #ifndef SHM_M #define SHM_M 010000 @@ -66,14 +89,21 @@ #ifndef SHM_X #define SHM_X 0100 #endif +<<<<<<< HEAD #ifndef ACCESSPERMS #define ACCESSPERMS (S_IRWXU | S_IRWXG | S_IRWXO) #endif +======= +#ifndef ACCESSPERMS +#define ACCESSPERMS (S_IRWXU | S_IRWXG | S_IRWXO)//文件权限值意思就是 用户,用户组,其他可读可写. +#endif //代表含义U:user G:group O:other +>>>>>>> remotes/origin/main #define SHM_S_IRUGO (S_IRUSR | S_IRGRP | S_IROTH) #define SHM_S_IWUGO (S_IWUSR | S_IWGRP | S_IWOTH) #define SHM_S_IXUGO (S_IXUSR | S_IXGRP | S_IXOTH) +<<<<<<< HEAD #define SHM_GROUPE_TO_USER 3 #define SHM_OTHER_TO_USER 6 @@ -81,6 +111,12 @@ #ifndef LOSCFG_IPC_CONTAINER STATIC LosMux g_sysvShmMux; +======= +#define SHM_GROUPE_TO_USER 3 +#define SHM_OTHER_TO_USER 6 +#ifndef LOSCFG_IPC_CONTAINER +STATIC LosMux g_sysvShmMux; +>>>>>>> remotes/origin/main /* private data */ STATIC struct shminfo g_shmInfo; STATIC struct shmIDSource *g_shmSegs = NULL; @@ -96,6 +132,63 @@ STATIC UINT32 g_shmUsedPageCount; #define SYSV_SHM_LOCK() (VOID)LOS_MuxLock(&IPC_SHM_SYS_VSHM_MUTEX, LOS_WAIT_FOREVER) #define SYSV_SHM_UNLOCK() (VOID)LOS_MuxUnlock(&IPC_SHM_SYS_VSHM_MUTEX) +<<<<<<< HEAD +======= +#if 0 // @note_#if0 + +//内核为每一个IPC对象保存一个ipc_perm结构体,该结构说明了IPC对象的权限和所有者 +struct ipc_perm { + key_t __ipc_perm_key; //调用shmget()时给出的关键字 + uid_t uid; //共享内存所有者的有效用户ID + gid_t gid; //共享内存所有者所属组的有效组ID + uid_t cuid; //共享内存创建 者的有效用户ID + gid_t cgid; //共享内存创建者所属组的有效组ID + mode_t mode; //权限 + SHM_DEST / SHM_LOCKED /SHM_HUGETLB 标志位 + int __ipc_perm_seq; //序列号 + long __pad1; //保留扩展用 + long __pad2; +}; +//每个共享内存段在内核中维护着一个内部结构shmid_ds +struct shmid_ds { + struct ipc_perm shm_perm;///< 操作许可,里面包含共享内存的用户ID、组ID等信息 + size_t shm_segsz; ///< 共享内存段的大小,单位为字节 + time_t shm_atime; ///< 最后一个进程访问共享内存的时间 + time_t shm_dtime; ///< 最后一个进程离开共享内存的时间 + time_t shm_ctime; ///< 创建时间 + pid_t shm_cpid; ///< 创建共享内存的进程ID + pid_t shm_lpid; ///< 最后操作共享内存的进程ID + unsigned long shm_nattch; ///< 当前使用该共享内存段的进程数量 + unsigned long __pad1; //保留扩展用 + unsigned long __pad2; +}; +// 共享内存模块设置信息 +struct shminfo { + unsigned long shmmax, shmmin, shmmni, shmseg, shmall, __unused[4]; +}; +struct shmIDSource {//共享内存描述符 + struct shmid_ds ds; //是内核为每一个共享内存段维护的数据结构 + UINT32 status; //状态 SHM_SEG_FREE ... + LOS_DL_LIST node; //节点,挂VmPage +#ifdef LOSCFG_SHELL + CHAR ownerName[OS_PCB_NAME_LEN]; +#endif +}; + +/* private data */ +STATIC struct shminfo g_shmInfo = { //描述共享内存范围的全局变量 + .shmmax = SHM_MAX,//共享内存单个上限 4096页 即 16M + .shmmin = SHM_MIN,//共享内存单个下限 1页 即:4K + .shmmni = SHM_MNI,//共享内存总数 默认192 + .shmseg = SHM_SEG,//每个用户进程可以使用的最多的共享内存段的数目 128 + .shmall = SHM_ALL,//系统范围内共享内存的总页数, 4096页 +}; + +STATIC struct shmIDSource *g_shmSegs = NULL; +STATIC UINT32 g_shmUsedPageCount; +#endif + +//共享内存初始化 +>>>>>>> remotes/origin/main struct shmIDSource *OsShmCBInit(LosMux *sysvShmMux, struct shminfo *shmInfo, UINT32 *shmUsedPageCount) { UINT32 ret; @@ -104,7 +197,10 @@ struct shmIDSource *OsShmCBInit(LosMux *sysvShmMux, struct shminfo *shmInfo, UIN if ((sysvShmMux == NULL) || (shmInfo == NULL) || (shmUsedPageCount == NULL)) { return NULL; } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main ret = LOS_MuxInit(sysvShmMux, NULL); if (ret != LOS_OK) { goto ERROR; @@ -115,7 +211,10 @@ struct shmIDSource *OsShmCBInit(LosMux *sysvShmMux, struct shminfo *shmInfo, UIN shmInfo->shmmni = SHM_MNI; shmInfo->shmseg = SHM_SEG; shmInfo->shmall = SHM_ALL; +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main struct shmIDSource *shmSegs = LOS_MemAlloc((VOID *)OS_SYS_MEM_ADDR, sizeof(struct shmIDSource) * shmInfo->shmmni); if (shmSegs == NULL) { (VOID)LOS_MuxDestroy(sysvShmMux); @@ -125,9 +224,15 @@ struct shmIDSource *OsShmCBInit(LosMux *sysvShmMux, struct shminfo *shmInfo, UIN 0, (sizeof(struct shmIDSource) * shmInfo->shmmni)); for (i = 0; i < shmInfo->shmmni; i++) { +<<<<<<< HEAD shmSegs[i].status = SHM_SEG_FREE; shmSegs[i].ds.shm_perm.seq = i + 1; LOS_ListInit(&shmSegs[i].node); +======= + shmSegs[i].status = SHM_SEG_FREE;//节点初始状态为空闲 + shmSegs[i].ds.shm_perm.seq = i + 1;//struct ipc_perm shm_perm;系统为每一个IPC对象保存一个ipc_perm结构体,结构说明了IPC对象的权限和所有者 + LOS_ListInit(&shmSegs[i].node);//初始化节点 +>>>>>>> remotes/origin/main } *shmUsedPageCount = 0; @@ -137,7 +242,10 @@ ERROR: VM_ERR("ShmInit fail\n"); return NULL; } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main UINT32 ShmInit(VOID) { #ifndef LOSCFG_IPC_CONTAINER @@ -149,8 +257,14 @@ UINT32 ShmInit(VOID) return LOS_OK; } +<<<<<<< HEAD LOS_MODULE_INIT(ShmInit, LOS_INIT_LEVEL_VM_COMPLETE); +======= + +LOS_MODULE_INIT(ShmInit, LOS_INIT_LEVEL_VM_COMPLETE);//共享内存模块初始化 +//共享内存反初始化 +>>>>>>> remotes/origin/main UINT32 ShmDeinit(VOID) { UINT32 ret; @@ -165,7 +279,11 @@ UINT32 ShmDeinit(VOID) return 0; } +<<<<<<< HEAD +======= +///给共享段中所有物理页框贴上共享标签 +>>>>>>> remotes/origin/main STATIC inline VOID ShmSetSharedFlag(struct shmIDSource *seg) { LosVmPage *page = NULL; @@ -174,7 +292,11 @@ STATIC inline VOID ShmSetSharedFlag(struct shmIDSource *seg) OsSetPageShared(page); } } +<<<<<<< HEAD +======= +///给共享段中所有物理页框撕掉共享标签 +>>>>>>> remotes/origin/main STATIC inline VOID ShmClearSharedFlag(struct shmIDSource *seg) { LosVmPage *page = NULL; @@ -183,7 +305,11 @@ STATIC inline VOID ShmClearSharedFlag(struct shmIDSource *seg) OsCleanPageShared(page); } } +<<<<<<< HEAD +======= +///seg下所有共享页引用减少 +>>>>>>> remotes/origin/main STATIC VOID ShmPagesRefDec(struct shmIDSource *seg) { LosVmPage *page = NULL; @@ -193,6 +319,18 @@ STATIC VOID ShmPagesRefDec(struct shmIDSource *seg) } } +<<<<<<< HEAD +======= +/** + * @brief 为共享段分配物理内存 + 例如:参数size = 4097, LOS_Align(size, PAGE_SIZE) = 8192 + 分配页数 size >> PAGE_SHIFT = 2页 + * @param key + * @param size + * @param shmflg + * @return STATIC + */ +>>>>>>> remotes/origin/main STATIC INT32 ShmAllocSegCheck(key_t key, size_t *size, INT32 *segNum) { INT32 i; @@ -201,7 +339,11 @@ STATIC INT32 ShmAllocSegCheck(key_t key, size_t *size, INT32 *segNum) return -EINVAL; } +<<<<<<< HEAD *size = LOS_Align(*size, PAGE_SIZE); +======= + *size = LOS_Align(*size, PAGE_SIZE);//必须对齐 +>>>>>>> remotes/origin/main if ((IPC_SHM_USED_PAGE_COUNT + (*size >> PAGE_SHIFT)) > IPC_SHM_INFO.shmall) { return -ENOMEM; } @@ -211,11 +353,18 @@ STATIC INT32 ShmAllocSegCheck(key_t key, size_t *size, INT32 *segNum) return -ENOMEM; } #endif +<<<<<<< HEAD for (i = 0; i < IPC_SHM_INFO.shmmni; i++) { if (IPC_SHM_SEGS[i].status & SHM_SEG_FREE) { IPC_SHM_SEGS[i].status &= ~SHM_SEG_FREE; *segNum = i; +======= + for (i = 0; i < IPC_SHM_INFO.shmmni; i++) {//试图找到一个空闲段与参数key绑定 + if (IPC_SHM_SEGS[i].status & SHM_SEG_FREE) {//找到空闲段 + IPC_SHM_SEGS[i].status &= ~SHM_SEG_FREE;//变成非空闲状态 + *segNum = i;//标号 +>>>>>>> remotes/origin/main break; } } @@ -236,17 +385,26 @@ STATIC INT32 ShmAllocSeg(key_t key, size_t size, INT32 shmflg) if (ret < 0) { return ret; } +<<<<<<< HEAD seg = &IPC_SHM_SEGS[segNum]; count = LOS_PhysPagesAlloc(size >> PAGE_SHIFT, &seg->node); if (count != (size >> PAGE_SHIFT)) { (VOID)LOS_PhysPagesFree(&seg->node); seg->status = SHM_SEG_FREE; +======= + seg = &IPC_SHM_SEGS[segNum]; + count = LOS_PhysPagesAlloc(size >> PAGE_SHIFT, &seg->node);//分配共享页面,函数内部把node都挂好了. + if (count != (size >> PAGE_SHIFT)) {//当未分配到足够的内存时,处理方式是:不稀罕给那么点,舍弃! + (VOID)LOS_PhysPagesFree(&seg->node);//释放节点上的物理页框 + seg->status = SHM_SEG_FREE;//共享段变回空闲状态 +>>>>>>> remotes/origin/main #ifdef LOSCFG_KERNEL_IPC_PLIMIT OsIPCLimitShmFree(size); #endif return -ENOMEM; } +<<<<<<< HEAD ShmSetSharedFlag(seg); IPC_SHM_USED_PAGE_COUNT += size >> PAGE_SHIFT; @@ -265,20 +423,49 @@ STATIC INT32 ShmAllocSeg(key_t key, size_t size, INT32 shmflg) seg->ds.shm_atime = 0; seg->ds.shm_dtime = 0; seg->ds.shm_ctime = time(NULL); +======= + ShmSetSharedFlag(seg);//将node的每个页面设置为共享页 + IPC_SHM_USED_PAGE_COUNT += size >> PAGE_SHIFT; + + seg->status |= SHM_SEG_USED; //共享段贴上已在使用的标签 + seg->ds.shm_perm.mode = (UINT32)shmflg & ACCESSPERMS; + seg->ds.shm_perm.key = key;//保存参数key,如此 key 和 共享ID绑定在一块 + seg->ds.shm_segsz = size; //共享段的大小 + seg->ds.shm_perm.cuid = LOS_GetUserID(); //设置用户ID + seg->ds.shm_perm.uid = LOS_GetUserID(); //设置用户ID + seg->ds.shm_perm.cgid = LOS_GetGroupID(); //设置组ID + seg->ds.shm_perm.gid = LOS_GetGroupID(); //设置组ID + seg->ds.shm_lpid = 0; //最后一个操作的进程 + seg->ds.shm_nattch = 0; //绑定进程的数量 + seg->ds.shm_cpid = LOS_GetCurrProcessID(); //获取进程ID + seg->ds.shm_atime = 0; //访问时间 + seg->ds.shm_dtime = 0; //detach 分离时间 共享内存使用完之后,需要将它从进程地址空间中分离出来;将共享内存分离并不是删除它,只是使该共享内存对当前的进程不再可用 + seg->ds.shm_ctime = time(NULL);//创建时间 +>>>>>>> remotes/origin/main #ifdef LOSCFG_SHELL (VOID)memcpy_s(seg->ownerName, OS_PCB_NAME_LEN, OsCurrProcessGet()->processName, OS_PCB_NAME_LEN); #endif return segNum; } +<<<<<<< HEAD +======= +///释放seg->node 所占物理页框,seg本身重置 +>>>>>>> remotes/origin/main STATIC INLINE VOID ShmFreeSeg(struct shmIDSource *seg, UINT32 *shmUsedPageCount) { UINT32 count; +<<<<<<< HEAD ShmClearSharedFlag(seg); count = LOS_PhysPagesFree(&seg->node); if (count != (seg->ds.shm_segsz >> PAGE_SHIFT)) { +======= + ShmClearSharedFlag(seg);//先撕掉 seg->node 中vmpage的共享标签 + count = LOS_PhysPagesFree(&seg->node);//再挨个删除物理页框 + if (count != (seg->ds.shm_segsz >> PAGE_SHIFT)) {//异常,必须要一样 +>>>>>>> remotes/origin/main VM_ERR("free physical pages failed, count = %d, size = %d", count, seg->ds.shm_segsz >> PAGE_SHIFT); return; } @@ -288,31 +475,54 @@ STATIC INLINE VOID ShmFreeSeg(struct shmIDSource *seg, UINT32 *shmUsedPageCount) if (shmUsedPageCount != NULL) { (*shmUsedPageCount) -= seg->ds.shm_segsz >> PAGE_SHIFT; } +<<<<<<< HEAD seg->status = SHM_SEG_FREE; LOS_ListInit(&seg->node); } +======= + seg->status = SHM_SEG_FREE;//seg恢复自由之身 + LOS_ListInit(&seg->node);//重置node +} +///通过key查找 shmId +>>>>>>> remotes/origin/main STATIC INT32 ShmFindSegByKey(key_t key) { INT32 i; struct shmIDSource *seg = NULL; +<<<<<<< HEAD for (i = 0; i < IPC_SHM_INFO.shmmni; i++) { seg = &IPC_SHM_SEGS[i]; if ((seg->status & SHM_SEG_USED) && (seg->ds.shm_perm.key == key)) { +======= + for (i = 0; i < IPC_SHM_INFO.shmmni; i++) {//遍历共享段池,找到与key绑定的共享ID + seg = &IPC_SHM_SEGS[i]; + if ((seg->status & SHM_SEG_USED) && + (seg->ds.shm_perm.key == key)) {//满足两个条件,找到后返回 +>>>>>>> remotes/origin/main return i; } } return -1; } +<<<<<<< HEAD STATIC INT32 ShmSegValidCheck(INT32 segNum, size_t size, INT32 shmFlg) { struct shmIDSource *seg = &IPC_SHM_SEGS[segNum]; if (size > seg->ds.shm_segsz) { +======= +///共享内存段有效性检查 +STATIC INT32 ShmSegValidCheck(INT32 segNum, size_t size, INT32 shmFlg) +{ + struct shmIDSource *seg = &IPC_SHM_SEGS[segNum];//拿到shmID + + if (size > seg->ds.shm_segsz) {//段长 +>>>>>>> remotes/origin/main return -EINVAL; } @@ -323,7 +533,11 @@ STATIC INT32 ShmSegValidCheck(INT32 segNum, size_t size, INT32 shmFlg) return segNum; } +<<<<<<< HEAD +======= +///通过ID找到共享内存资源 +>>>>>>> remotes/origin/main STATIC struct shmIDSource *ShmFindSeg(int shmid) { struct shmIDSource *seg = NULL; @@ -341,7 +555,11 @@ STATIC struct shmIDSource *ShmFindSeg(int shmid) return seg; } +<<<<<<< HEAD +======= +///共享内存映射 +>>>>>>> remotes/origin/main STATIC VOID ShmVmmMapping(LosVmSpace *space, LOS_DL_LIST *pageList, VADDR_T vaddr, UINT32 regionFlags) { LosVmPage *vmPage = NULL; @@ -349,29 +567,45 @@ STATIC VOID ShmVmmMapping(LosVmSpace *space, LOS_DL_LIST *pageList, VADDR_T vadd PADDR_T pa; STATUS_T ret; +<<<<<<< HEAD LOS_DL_LIST_FOR_EACH_ENTRY(vmPage, pageList, LosVmPage, node) { pa = VM_PAGE_TO_PHYS(vmPage); LOS_AtomicInc(&vmPage->refCounts); ret = LOS_ArchMmuMap(&space->archMmu, va, pa, 1, regionFlags); +======= + LOS_DL_LIST_FOR_EACH_ENTRY(vmPage, pageList, LosVmPage, node) {//遍历一页一页映射 + pa = VM_PAGE_TO_PHYS(vmPage);//拿到物理地址 + LOS_AtomicInc(&vmPage->refCounts);//自增 + ret = LOS_ArchMmuMap(&space->archMmu, va, pa, 1, regionFlags);//虚实映射 +>>>>>>> remotes/origin/main if (ret != 1) { VM_ERR("LOS_ArchMmuMap failed, ret = %d", ret); } va += PAGE_SIZE; } } +<<<<<<< HEAD +======= +///fork 一个共享线性区 +>>>>>>> remotes/origin/main VOID OsShmFork(LosVmSpace *space, LosVmMapRegion *oldRegion, LosVmMapRegion *newRegion) { struct shmIDSource *seg = NULL; SYSV_SHM_LOCK(); +<<<<<<< HEAD seg = ShmFindSeg(oldRegion->shmid); +======= + seg = ShmFindSeg(oldRegion->shmid);//通过老区ID获取对应的共享资源ID结构体 +>>>>>>> remotes/origin/main if (seg == NULL) { SYSV_SHM_UNLOCK(); VM_ERR("shm fork failed!"); return; } +<<<<<<< HEAD newRegion->shmid = oldRegion->shmid; newRegion->forkFlags = oldRegion->forkFlags; ShmVmmMapping(space, &seg->node, newRegion->range.base, newRegion->regionFlags); @@ -379,34 +613,63 @@ VOID OsShmFork(LosVmSpace *space, LosVmMapRegion *oldRegion, LosVmMapRegion *new SYSV_SHM_UNLOCK(); } +======= + newRegion->shmid = oldRegion->shmid;//一样的共享区ID + newRegion->forkFlags = oldRegion->forkFlags;//forkFlags也一样了 + ShmVmmMapping(space, &seg->node, newRegion->range.base, newRegion->regionFlags);//新线性区与共享内存进行映射 + seg->ds.shm_nattch++;//附在共享线性区上的进程数++ + SYSV_SHM_UNLOCK(); +} +///释放共享线性区 +>>>>>>> remotes/origin/main VOID OsShmRegionFree(LosVmSpace *space, LosVmMapRegion *region) { struct shmIDSource *seg = NULL; SYSV_SHM_LOCK(); +<<<<<<< HEAD seg = ShmFindSeg(region->shmid); +======= + seg = ShmFindSeg(region->shmid);//通过线性区ID获取对应的共享资源ID结构体 +>>>>>>> remotes/origin/main if (seg == NULL) { SYSV_SHM_UNLOCK(); return; } +<<<<<<< HEAD LOS_ArchMmuUnmap(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT); ShmPagesRefDec(seg); seg->ds.shm_nattch--; if (seg->ds.shm_nattch <= 0 && (seg->status & SHM_SEG_REMOVE)) { ShmFreeSeg(seg, &IPC_SHM_USED_PAGE_COUNT); +======= + LOS_ArchMmuUnmap(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT);//解除线性区的映射 + ShmPagesRefDec(seg);//ref -- + seg->ds.shm_nattch--;//附在共享线性区上的进程数-- + if (seg->ds.shm_nattch <= 0 && (seg->status & SHM_SEG_REMOVE)) { + ShmFreeSeg(seg, &IPC_SHM_USED_PAGE_COUNT);//就释放掉物理内存!注意是:物理内存 +>>>>>>> remotes/origin/main } else { seg->ds.shm_dtime = time(NULL); seg->ds.shm_lpid = LOS_GetCurrProcessID(); /* may not be the space's PID. */ } SYSV_SHM_UNLOCK(); } +<<<<<<< HEAD +======= +///是否为共享线性区,是否有标签? +>>>>>>> remotes/origin/main BOOL OsIsShmRegion(LosVmMapRegion *region) { return (region->regionFlags & VM_MAP_REGION_FLAG_SHM) ? TRUE : FALSE; } +<<<<<<< HEAD +======= +///获取共享内存池中已被使用的段数量 +>>>>>>> remotes/origin/main STATIC INT32 ShmSegUsedCount(VOID) { INT32 i; @@ -415,16 +678,27 @@ STATIC INT32 ShmSegUsedCount(VOID) for (i = 0; i < IPC_SHM_INFO.shmmni; i++) { seg = &IPC_SHM_SEGS[i]; +<<<<<<< HEAD if (seg->status & SHM_SEG_USED) { +======= + if (seg->status & SHM_SEG_USED) {//找到一个 +>>>>>>> remotes/origin/main count++; } } return count; } +<<<<<<< HEAD STATIC INT32 ShmPermCheck(struct shmIDSource *seg, mode_t mode) { INT32 uid = LOS_GetUserID(); +======= +///对共享内存段权限检查 +STATIC INT32 ShmPermCheck(struct shmIDSource *seg, mode_t mode) +{ + INT32 uid = LOS_GetUserID();//当前进程的用户ID +>>>>>>> remotes/origin/main UINT32 tmpMode = 0; mode_t privMode = seg->ds.shm_perm.mode; mode_t accMode; @@ -466,6 +740,25 @@ STATIC INT32 ShmPermCheck(struct shmIDSource *seg, mode_t mode) } } +<<<<<<< HEAD +======= +/*! + * @brief ShmGet + * 得到一个共享内存标识符或创建一个共享内存对象 + * @param key 建立新共享内存对象 标识符是IPC对象的内部名。为使多个合作进程能够在同一IPC对象上汇聚,需要提供一个外部命名方案。 + 为此,每个IPC对象都与一个键(key)相关联,这个键作为该对象的外部名,无论何时创建IPC结构(通过msgget、semget、shmget创建), + 都应给IPC指定一个键, key_t由ftok创建,ftok当然在本工程里找不到,所以要写这么多. + * @param shmflg IPC_CREAT IPC_EXCL + IPC_CREAT: 在创建新的IPC时,如果key参数是IPC_PRIVATE或者和当前某种类型的IPC结构无关,则需要指明flag参数的IPC_CREAT标志位, + 则用来创建一个新的IPC结构。(如果IPC结构已存在,并且指定了IPC_CREAT,则IPC_CREAT什么都不做,函数也不出错) + IPC_EXCL: 此参数一般与IPC_CREAT配合使用来创建一个新的IPC结构。如果创建的IPC结构已存在函数就出错返回, + 返回EEXIST(这与open函数指定O_CREAT和O_EXCL标志原理相同) + * @param size 新建的共享内存大小,以字节为单位 + * @return + * + * @see + */ +>>>>>>> remotes/origin/main INT32 ShmGet(key_t key, size_t size, INT32 shmflg) { INT32 ret; @@ -476,6 +769,7 @@ INT32 ShmGet(key_t key, size_t size, INT32 shmflg) if (key == IPC_PRIVATE) { ret = ShmAllocSeg(key, size, shmflg); } else { +<<<<<<< HEAD ret = ShmFindSegByKey(key); if (ret < 0) { if (((UINT32)shmflg & IPC_CREAT) == 0) { @@ -483,6 +777,15 @@ INT32 ShmGet(key_t key, size_t size, INT32 shmflg) goto ERROR; } else { ret = ShmAllocSeg(key, size, shmflg); +======= + ret = ShmFindSegByKey(key);//通过key查找资源ID + if (ret < 0) { + if (((UINT32)shmflg & IPC_CREAT) == 0) {// + ret = -ENOENT; + goto ERROR; + } else { + ret = ShmAllocSeg(key, size, shmflg);//分配一个共享内存 +>>>>>>> remotes/origin/main } } else { shmid = ret; @@ -491,7 +794,11 @@ INT32 ShmGet(key_t key, size_t size, INT32 shmflg) ret = -EEXIST; goto ERROR; } +<<<<<<< HEAD ret = ShmPermCheck(ShmFindSeg(shmid), (UINT32)shmflg & ACCESSPERMS); +======= + ret = ShmPermCheck(ShmFindSeg(shmid), (UINT32)shmflg & ACCESSPERMS);//对共享内存权限检查 +>>>>>>> remotes/origin/main if (ret != 0) { ret = -ret; goto ERROR; @@ -526,13 +833,21 @@ INT32 ShmatParamCheck(const VOID *shmaddr, INT32 shmflg) return 0; } +<<<<<<< HEAD +======= +///分配一个共享线性区并映射好 +>>>>>>> remotes/origin/main LosVmMapRegion *ShmatVmmAlloc(struct shmIDSource *seg, const VOID *shmaddr, INT32 shmflg, UINT32 prot) { LosVmSpace *space = OsCurrProcessGet()->vmSpace; LosVmMapRegion *region = NULL; +<<<<<<< HEAD UINT32 flags = MAP_ANONYMOUS | MAP_SHARED; +======= + UINT32 flags = MAP_ANONYMOUS | MAP_SHARED;//本线性区为共享+匿名标签 +>>>>>>> remotes/origin/main UINT32 mapFlags = flags | MAP_FIXED; VADDR_T vaddr; UINT32 regionFlags; @@ -543,29 +858,48 @@ LosVmMapRegion *ShmatVmmAlloc(struct shmIDSource *seg, const VOID *shmaddr, } regionFlags = OsCvtProtFlagsToRegionFlags(prot, flags); (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD if (shmaddr == NULL) { region = LOS_RegionAlloc(space, 0, seg->ds.shm_segsz, regionFlags, 0); } else { +======= + if (shmaddr == NULL) {//未指定了共享内存连接到当前进程中的地址位置 + region = LOS_RegionAlloc(space, 0, seg->ds.shm_segsz, regionFlags, 0);//分配线性区 + } else {//指定时,就需要先找地址所在的线性区 +>>>>>>> remotes/origin/main if ((UINT32)shmflg & SHM_RND) { vaddr = ROUNDDOWN((VADDR_T)(UINTPTR)shmaddr, SHMLBA); } else { vaddr = (VADDR_T)(UINTPTR)shmaddr; +<<<<<<< HEAD } +======= + }//找到线性区并重新映射,当指定地址时需贴上重新映射的标签 +>>>>>>> remotes/origin/main if (!((UINT32)shmflg & SHM_REMAP) && (LOS_RegionFind(space, vaddr) || LOS_RegionFind(space, vaddr + seg->ds.shm_segsz - 1) || LOS_RegionRangeFind(space, vaddr, seg->ds.shm_segsz - 1))) { ret = EINVAL; goto ERROR; } +<<<<<<< HEAD vaddr = (VADDR_T)LOS_MMap(vaddr, seg->ds.shm_segsz, prot, mapFlags, -1, 0); region = LOS_RegionFind(space, vaddr); +======= + vaddr = (VADDR_T)LOS_MMap(vaddr, seg->ds.shm_segsz, prot, mapFlags, -1, 0);//做好映射 + region = LOS_RegionFind(space, vaddr);//重新查找线性区,用于返回. +>>>>>>> remotes/origin/main } if (region == NULL) { ret = ENOMEM; goto ERROR; } +<<<<<<< HEAD ShmVmmMapping(space, &seg->node, region->range.base, regionFlags); +======= + ShmVmmMapping(space, &seg->node, region->range.base, regionFlags);//共享内存映射 +>>>>>>> remotes/origin/main (VOID)LOS_MuxRelease(&space->regionMux); return region; ERROR: @@ -574,6 +908,20 @@ ERROR: return NULL; } +<<<<<<< HEAD +======= +/*! + * @brief ShmAt + * 用来启动对该共享内存的访问,并把共享内存连接到当前进程的地址空间。 + * @param shm_flg 是一组标志位,通常为0。 + * @param shmaddr 指定共享内存连接到当前进程中的地址位置,通常为空,表示让系统来选择共享内存的地址。 + * @param shmid 是shmget()函数返回的共享内存标识符 + * @return + * 如果shmat成功执行,那么内核将使与该共享存储相关的shmid_ds结构中的shm_nattch计数器值加1 + shmid 就是个索引,就跟进程和线程的ID一样 g_shmSegs[shmid] shmid > 192个 + * @see + */ +>>>>>>> remotes/origin/main VOID *ShmAt(INT32 shmid, const VOID *shmaddr, INT32 shmflg) { INT32 ret; @@ -582,13 +930,21 @@ VOID *ShmAt(INT32 shmid, const VOID *shmaddr, INT32 shmflg) struct shmIDSource *seg = NULL; LosVmMapRegion *r = NULL; +<<<<<<< HEAD ret = ShmatParamCheck(shmaddr, shmflg); +======= + ret = ShmatParamCheck(shmaddr, shmflg);//参数检查 +>>>>>>> remotes/origin/main if (ret != 0) { set_errno(ret); return (VOID *)-1; } +<<<<<<< HEAD if ((UINT32)shmflg & SHM_EXEC) { +======= + if ((UINT32)shmflg & SHM_EXEC) {//flag 转换 +>>>>>>> remotes/origin/main prot |= PROT_EXEC; acc_mode |= SHM_S_IXUGO; } else if (((UINT32)shmflg & SHM_RDONLY) == 0) { @@ -597,7 +953,11 @@ VOID *ShmAt(INT32 shmid, const VOID *shmaddr, INT32 shmflg) } SYSV_SHM_LOCK(); +<<<<<<< HEAD seg = ShmFindSeg(shmid); +======= + seg = ShmFindSeg(shmid);//找到段 +>>>>>>> remotes/origin/main if (seg == NULL) { SYSV_SHM_UNLOCK(); return (VOID *)-1; @@ -608,18 +968,30 @@ VOID *ShmAt(INT32 shmid, const VOID *shmaddr, INT32 shmflg) goto ERROR; } +<<<<<<< HEAD seg->ds.shm_nattch++; r = ShmatVmmAlloc(seg, shmaddr, shmflg, prot); +======= + seg->ds.shm_nattch++;//ds上记录有一个进程绑定上来 + r = ShmatVmmAlloc(seg, shmaddr, shmflg, prot);//在当前进程空间分配一个线性区并映射到共享内存 +>>>>>>> remotes/origin/main if (r == NULL) { seg->ds.shm_nattch--; SYSV_SHM_UNLOCK(); return (VOID *)-1; } +<<<<<<< HEAD r->shmid = shmid; r->regionFlags |= VM_MAP_REGION_FLAG_SHM; seg->ds.shm_atime = time(NULL); seg->ds.shm_lpid = LOS_GetCurrProcessID(); +======= + r->shmid = shmid;//把ID给线性区的shmid + r->regionFlags |= VM_MAP_REGION_FLAG_SHM;//这是一个共享线性区 + seg->ds.shm_atime = time(NULL);//访问时间 + seg->ds.shm_lpid = LOS_GetCurrProcessID();//进程ID +>>>>>>> remotes/origin/main SYSV_SHM_UNLOCK(); return (VOID *)(UINTPTR)r->range.base; @@ -630,6 +1002,22 @@ ERROR: return (VOID *)-1; } +<<<<<<< HEAD +======= +/*! + * @brief ShmCtl + * 此函数可以对shmid指定的共享存储进行多种操作(删除、取信息、加锁、解锁等) + * @param buf 是一个结构指针,它指向共享内存模式和访问权限的结构。 + * @param cmd command是要采取的操作,它可以取下面的三个值 : + IPC_STAT:把shmid_ds结构中的数据设置为共享内存的当前关联值,即用共享内存的当前关联值覆盖shmid_ds的值。 + IPC_SET:如果进程有足够的权限,就把共享内存的当前关联值设置为shmid_ds结构中给出的值 + IPC_RMID:删除共享内存段 + * @param shmid 是shmget()函数返回的共享内存标识符 + * @return + * + * @see + */ +>>>>>>> remotes/origin/main INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) { struct shmIDSource *seg = NULL; @@ -642,7 +1030,11 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) SYSV_SHM_LOCK(); if ((cmd != IPC_INFO) && (cmd != SHM_INFO)) { +<<<<<<< HEAD seg = ShmFindSeg(shmid); +======= + seg = ShmFindSeg(shmid);//通过索引ID找到seg +>>>>>>> remotes/origin/main if (seg == NULL) { SYSV_SHM_UNLOCK(); return -1; @@ -656,13 +1048,21 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) switch (cmd) { case IPC_STAT: +<<<<<<< HEAD case SHM_STAT: +======= + case SHM_STAT://取段结构 +>>>>>>> remotes/origin/main ret = ShmPermCheck(seg, SHM_S_IRUGO); if (ret != 0) { goto ERROR; } +<<<<<<< HEAD ret = LOS_ArchCopyToUser(buf, &seg->ds, sizeof(struct shmid_ds)); +======= + ret = LOS_ArchCopyToUser(buf, &seg->ds, sizeof(struct shmid_ds));//把内核空间的共享页数据拷贝到用户空间 +>>>>>>> remotes/origin/main if (ret != 0) { ret = EFAULT; goto ERROR; @@ -671,13 +1071,21 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) ret = (unsigned int)((unsigned int)seg->ds.shm_perm.seq << 16) | (unsigned int)((unsigned int)shmid & 0xffff); /* 16: use the seq as the upper 16 bits */ } break; +<<<<<<< HEAD case IPC_SET: +======= + case IPC_SET://重置共享段 +>>>>>>> remotes/origin/main ret = ShmPermCheck(seg, SHM_M); if (ret != 0) { ret = EPERM; goto ERROR; } +<<<<<<< HEAD +======= + //从用户空间拷贝数据到内核空间 +>>>>>>> remotes/origin/main ret = LOS_ArchCopyFromUser(&shm_perm, &buf->shm_perm, sizeof(struct ipc_perm)); if (ret != 0) { ret = EFAULT; @@ -686,14 +1094,22 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) seg->ds.shm_perm.uid = shm_perm.uid; seg->ds.shm_perm.gid = shm_perm.gid; seg->ds.shm_perm.mode = (seg->ds.shm_perm.mode & ~ACCESSPERMS) | +<<<<<<< HEAD (shm_perm.mode & ACCESSPERMS); +======= + (shm_perm.mode & ACCESSPERMS);//可访问 +>>>>>>> remotes/origin/main seg->ds.shm_ctime = time(NULL); #ifdef LOSCFG_SHELL (VOID)memcpy_s(seg->ownerName, OS_PCB_NAME_LEN, OS_PCB_FROM_PID(shm_perm.uid)->processName, OS_PCB_NAME_LEN); #endif break; +<<<<<<< HEAD case IPC_RMID: +======= + case IPC_RMID://删除共享段 +>>>>>>> remotes/origin/main ret = ShmPermCheck(seg, SHM_M); if (ret != 0) { ret = EPERM; @@ -701,11 +1117,19 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) } seg->status |= SHM_SEG_REMOVE; +<<<<<<< HEAD if (seg->ds.shm_nattch <= 0) { ShmFreeSeg(seg, &IPC_SHM_USED_PAGE_COUNT); } break; case IPC_INFO: +======= + if (seg->ds.shm_nattch <= 0) {//没有任何进程在使用了 + ShmFreeSeg(seg, &IPC_SHM_USED_PAGE_COUNT); + } + break; + case IPC_INFO://把内核空间的共享页数据拷贝到用户空间 +>>>>>>> remotes/origin/main ret = LOS_ArchCopyToUser(buf, &IPC_SHM_INFO, sizeof(struct shminfo)); if (ret != 0) { ret = EFAULT; @@ -719,8 +1143,13 @@ INT32 ShmCtl(INT32 shmid, INT32 cmd, struct shmid_ds *buf) shmInfo.shm_tot = 0; shmInfo.swap_attempts = 0; shmInfo.swap_successes = 0; +<<<<<<< HEAD shmInfo.used_ids = ShmSegUsedCount(); ret = LOS_ArchCopyToUser(buf, &shmInfo, sizeof(struct shm_info)); +======= + shmInfo.used_ids = ShmSegUsedCount();//在使用的seg数 + ret = LOS_ArchCopyToUser(buf, &shmInfo, sizeof(struct shm_info));//把内核空间的共享页数据拷贝到用户空间 +>>>>>>> remotes/origin/main if (ret != 0) { ret = EFAULT; goto ERROR; @@ -743,33 +1172,63 @@ ERROR: return -1; } +<<<<<<< HEAD INT32 ShmDt(const VOID *shmaddr) { LosVmSpace *space = OsCurrProcessGet()->vmSpace; +======= +/** + * @brief 当对共享存储的操作已经结束时,则调用shmdt与该存储段分离 + 如果shmat成功执行,那么内核将使与该共享存储相关的shmid_ds结构中的shm_nattch计数器值减1 + * @attention 注意:这并不从系统中删除共享存储的标识符以及其相关的数据结构。共享存储的仍然存在, + 直至某个进程带IPC_RMID命令的调用shmctl特地删除共享存储为止 + * @param shmaddr + * @return INT32 + */ +INT32 ShmDt(const VOID *shmaddr) +{ + LosVmSpace *space = OsCurrProcessGet()->vmSpace;//获取进程空间 +>>>>>>> remotes/origin/main struct shmIDSource *seg = NULL; LosVmMapRegion *region = NULL; INT32 shmid; INT32 ret; +<<<<<<< HEAD if (IS_PAGE_ALIGNED(shmaddr) == 0) { +======= + if (IS_PAGE_ALIGNED(shmaddr) == 0) {//地址是否对齐 +>>>>>>> remotes/origin/main ret = EINVAL; goto ERROR; } (VOID)LOS_MuxAcquire(&space->regionMux); +<<<<<<< HEAD region = LOS_RegionFind(space, (VADDR_T)(UINTPTR)shmaddr); +======= + region = LOS_RegionFind(space, (VADDR_T)(UINTPTR)shmaddr);//找到线性区 +>>>>>>> remotes/origin/main if (region == NULL) { ret = EINVAL; goto ERROR_WITH_LOCK; } +<<<<<<< HEAD shmid = region->shmid; if (region->range.base != (VADDR_T)(UINTPTR)shmaddr) { ret = EINVAL; +======= + shmid = region->shmid;//线性区共享ID + + if (region->range.base != (VADDR_T)(UINTPTR)shmaddr) {//这是用户空间和内核空间的一次解绑 + ret = EINVAL; //shmaddr 必须要等于region->range.base +>>>>>>> remotes/origin/main goto ERROR_WITH_LOCK; } /* remove it from aspace */ +<<<<<<< HEAD LOS_RbDelNode(&space->regionRbTree, ®ion->rbNode); LOS_ArchMmuUnmap(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT); (VOID)LOS_MuxRelease(&space->regionMux); @@ -778,12 +1237,23 @@ INT32 ShmDt(const VOID *shmaddr) SYSV_SHM_LOCK(); seg = ShmFindSeg(shmid); +======= + LOS_RbDelNode(&space->regionRbTree, ®ion->rbNode);//从红黑树和链表中摘除节点 + LOS_ArchMmuUnmap(&space->archMmu, region->range.base, region->range.size >> PAGE_SHIFT);//解除线性区的映射 + (VOID)LOS_MuxRelease(&space->regionMux); + /* free it */ + free(region);//释放线性区所占内存池中的内存 + + SYSV_SHM_LOCK(); + seg = ShmFindSeg(shmid);//找到seg,线性区和共享段的关系是 1:N 的关系,其他空间的线性区也会绑在共享段上 +>>>>>>> remotes/origin/main if (seg == NULL) { ret = EINVAL; SYSV_SHM_UNLOCK(); goto ERROR; } +<<<<<<< HEAD ShmPagesRefDec(seg); seg->ds.shm_nattch--; if ((seg->ds.shm_nattch <= 0) && @@ -792,6 +1262,16 @@ INT32 ShmDt(const VOID *shmaddr) } else { seg->ds.shm_dtime = time(NULL); seg->ds.shm_lpid = LOS_GetCurrProcessID(); +======= + ShmPagesRefDec(seg);//页面引用数 -- + seg->ds.shm_nattch--;//使用共享内存的进程数少了一个 + if ((seg->ds.shm_nattch <= 0) && //无任何进程使用共享内存 + (seg->status & SHM_SEG_REMOVE)) {//状态为删除时需要释放物理页内存了,否则其他进程还要继续使用共享内存 + ShmFreeSeg(seg, &IPC_SHM_USED_PAGE_COUNT);//释放seg 页框链表中的页框内存,再重置seg状态 + } else { + seg->ds.shm_dtime = time(NULL);//记录分离的时间 + seg->ds.shm_lpid = LOS_GetCurrProcessID();//记录操作进程ID +>>>>>>> remotes/origin/main } SYSV_SHM_UNLOCK(); @@ -847,7 +1327,10 @@ STATIC VOID OsShmInfoCmd(VOID) } SYSV_SHM_UNLOCK(); } +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main STATIC VOID OsShmDeleteCmd(INT32 shmid) { struct shmIDSource *seg = NULL; @@ -876,7 +1359,11 @@ STATIC VOID OsShmCmdUsage(VOID) "\t-r [shmid], Recycle the specified shared memory about shmid\n" "\t-h | --help, print shm command usage\n"); } +<<<<<<< HEAD +======= +///共享内存 +>>>>>>> remotes/origin/main UINT32 OsShellCmdShm(INT32 argc, const CHAR *argv[]) { INT32 shmid; @@ -911,4 +1398,7 @@ DONE: SHELLCMD_ENTRY(shm_shellcmd, CMD_TYPE_SHOW, "shm", 2, (CmdCallBackFunc)OsShellCmdShm); #endif #endif +<<<<<<< HEAD +======= +>>>>>>> remotes/origin/main