mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2025-08-16 12:34:01 +00:00

When running mm selftest to verify mm patches, 'compaction_test' case
failed on an x86 server with 1TB memory. And the root cause is that it
has too much free memory than what the test supports.
The test case tries to allocate 100000 huge pages, which is about 200 GB
for that x86 server, and when it succeeds, it expects it's large than 1/3
of 80% of the free memory in system. This logic only works for platform
with 750 GB ( 200 / (1/3) / 80% ) or less free memory, and may raise false
alarm for others.
Fix it by changing the fixed page number to self-adjustable number
according to the real number of free memory.
Link: https://lkml.kernel.org/r/20250423103645.2758-1-feng.tang@linux.alibaba.com
Fixes: bd67d5c15c
("Test compaction of mlocked memory")
Signed-off-by: Feng Tang <feng.tang@linux.alibaba.com>
Acked-by: Dev Jain <dev.jain@arm.com>
Reviewed-by: Baolin Wang <baolin.wang@linux.alibaba.com>
Tested-by: Baolin Wang <baolin.wang@inux.alibaba.com>
Cc: Shuah Khan <shuah@kernel.org>
Cc: Sri Jayaramappa <sjayaram@akamai.com>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
277 lines
6.6 KiB
C
277 lines
6.6 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
*
|
|
* A test for the patch "Allow compaction of unevictable pages".
|
|
* With this patch we should be able to allocate at least 1/4
|
|
* of RAM in huge pages. Without the patch much less is
|
|
* allocated.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/resource.h>
|
|
#include <fcntl.h>
|
|
#include <errno.h>
|
|
#include <unistd.h>
|
|
#include <string.h>
|
|
|
|
#include "../kselftest.h"
|
|
|
|
#define MAP_SIZE_MB 100
|
|
#define MAP_SIZE (MAP_SIZE_MB * 1024 * 1024)
|
|
|
|
struct map_list {
|
|
void *map;
|
|
struct map_list *next;
|
|
};
|
|
|
|
int read_memory_info(unsigned long *memfree, unsigned long *hugepagesize)
|
|
{
|
|
char buffer[256] = {0};
|
|
char *cmd = "cat /proc/meminfo | grep -i memfree | grep -o '[0-9]*'";
|
|
FILE *cmdfile = popen(cmd, "r");
|
|
|
|
if (!(fgets(buffer, sizeof(buffer), cmdfile))) {
|
|
ksft_print_msg("Failed to read meminfo: %s\n", strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
pclose(cmdfile);
|
|
|
|
*memfree = atoll(buffer);
|
|
cmd = "cat /proc/meminfo | grep -i hugepagesize | grep -o '[0-9]*'";
|
|
cmdfile = popen(cmd, "r");
|
|
|
|
if (!(fgets(buffer, sizeof(buffer), cmdfile))) {
|
|
ksft_print_msg("Failed to read meminfo: %s\n", strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
pclose(cmdfile);
|
|
*hugepagesize = atoll(buffer);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int prereq(void)
|
|
{
|
|
char allowed;
|
|
int fd;
|
|
|
|
fd = open("/proc/sys/vm/compact_unevictable_allowed",
|
|
O_RDONLY | O_NONBLOCK);
|
|
if (fd < 0) {
|
|
ksft_print_msg("Failed to open /proc/sys/vm/compact_unevictable_allowed: %s\n",
|
|
strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
if (read(fd, &allowed, sizeof(char)) != sizeof(char)) {
|
|
ksft_print_msg("Failed to read from /proc/sys/vm/compact_unevictable_allowed: %s\n",
|
|
strerror(errno));
|
|
close(fd);
|
|
return -1;
|
|
}
|
|
|
|
close(fd);
|
|
if (allowed == '1')
|
|
return 0;
|
|
|
|
ksft_print_msg("Compaction isn't allowed\n");
|
|
return -1;
|
|
}
|
|
|
|
int check_compaction(unsigned long mem_free, unsigned long hugepage_size,
|
|
unsigned long initial_nr_hugepages)
|
|
{
|
|
unsigned long nr_hugepages_ul;
|
|
int fd, ret = -1;
|
|
int compaction_index = 0;
|
|
char nr_hugepages[20] = {0};
|
|
char init_nr_hugepages[24] = {0};
|
|
char target_nr_hugepages[24] = {0};
|
|
int slen;
|
|
|
|
snprintf(init_nr_hugepages, sizeof(init_nr_hugepages),
|
|
"%lu", initial_nr_hugepages);
|
|
|
|
/* We want to test with 80% of available memory. Else, OOM killer comes
|
|
in to play */
|
|
mem_free = mem_free * 0.8;
|
|
|
|
fd = open("/proc/sys/vm/nr_hugepages", O_RDWR | O_NONBLOCK);
|
|
if (fd < 0) {
|
|
ksft_print_msg("Failed to open /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
ret = -1;
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Request huge pages for about half of the free memory. The Kernel
|
|
* will allocate as much as it can, and we expect it will get at least 1/3
|
|
*/
|
|
nr_hugepages_ul = mem_free / hugepage_size / 2;
|
|
snprintf(target_nr_hugepages, sizeof(target_nr_hugepages),
|
|
"%lu", nr_hugepages_ul);
|
|
|
|
slen = strlen(target_nr_hugepages);
|
|
if (write(fd, target_nr_hugepages, slen) != slen) {
|
|
ksft_print_msg("Failed to write %lu to /proc/sys/vm/nr_hugepages: %s\n",
|
|
nr_hugepages_ul, strerror(errno));
|
|
goto close_fd;
|
|
}
|
|
|
|
lseek(fd, 0, SEEK_SET);
|
|
|
|
if (read(fd, nr_hugepages, sizeof(nr_hugepages)) <= 0) {
|
|
ksft_print_msg("Failed to re-read from /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
goto close_fd;
|
|
}
|
|
|
|
/* We should have been able to request at least 1/3 rd of the memory in
|
|
huge pages */
|
|
nr_hugepages_ul = strtoul(nr_hugepages, NULL, 10);
|
|
if (!nr_hugepages_ul) {
|
|
ksft_print_msg("ERROR: No memory is available as huge pages\n");
|
|
goto close_fd;
|
|
}
|
|
compaction_index = mem_free/(nr_hugepages_ul * hugepage_size);
|
|
|
|
lseek(fd, 0, SEEK_SET);
|
|
|
|
if (write(fd, init_nr_hugepages, strlen(init_nr_hugepages))
|
|
!= strlen(init_nr_hugepages)) {
|
|
ksft_print_msg("Failed to write value to /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
goto close_fd;
|
|
}
|
|
|
|
ksft_print_msg("Number of huge pages allocated = %lu\n",
|
|
nr_hugepages_ul);
|
|
|
|
if (compaction_index > 3) {
|
|
ksft_print_msg("ERROR: Less than 1/%d of memory is available\n"
|
|
"as huge pages\n", compaction_index);
|
|
goto close_fd;
|
|
}
|
|
|
|
ret = 0;
|
|
|
|
close_fd:
|
|
close(fd);
|
|
out:
|
|
ksft_test_result(ret == 0, "check_compaction\n");
|
|
return ret;
|
|
}
|
|
|
|
int set_zero_hugepages(unsigned long *initial_nr_hugepages)
|
|
{
|
|
int fd, ret = -1;
|
|
char nr_hugepages[20] = {0};
|
|
|
|
fd = open("/proc/sys/vm/nr_hugepages", O_RDWR | O_NONBLOCK);
|
|
if (fd < 0) {
|
|
ksft_print_msg("Failed to open /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
goto out;
|
|
}
|
|
if (read(fd, nr_hugepages, sizeof(nr_hugepages)) <= 0) {
|
|
ksft_print_msg("Failed to read from /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
goto close_fd;
|
|
}
|
|
|
|
lseek(fd, 0, SEEK_SET);
|
|
|
|
/* Start with the initial condition of 0 huge pages */
|
|
if (write(fd, "0", sizeof(char)) != sizeof(char)) {
|
|
ksft_print_msg("Failed to write 0 to /proc/sys/vm/nr_hugepages: %s\n",
|
|
strerror(errno));
|
|
goto close_fd;
|
|
}
|
|
|
|
*initial_nr_hugepages = strtoul(nr_hugepages, NULL, 10);
|
|
ret = 0;
|
|
|
|
close_fd:
|
|
close(fd);
|
|
|
|
out:
|
|
return ret;
|
|
}
|
|
|
|
int main(int argc, char **argv)
|
|
{
|
|
struct rlimit lim;
|
|
struct map_list *list = NULL, *entry;
|
|
size_t page_size, i;
|
|
void *map = NULL;
|
|
unsigned long mem_free = 0;
|
|
unsigned long hugepage_size = 0;
|
|
long mem_fragmentable_MB = 0;
|
|
unsigned long initial_nr_hugepages;
|
|
|
|
ksft_print_header();
|
|
|
|
if (prereq() || geteuid())
|
|
ksft_exit_skip("Prerequisites unsatisfied\n");
|
|
|
|
ksft_set_plan(1);
|
|
|
|
/* Start the test without hugepages reducing mem_free */
|
|
if (set_zero_hugepages(&initial_nr_hugepages))
|
|
ksft_exit_fail();
|
|
|
|
lim.rlim_cur = RLIM_INFINITY;
|
|
lim.rlim_max = RLIM_INFINITY;
|
|
if (setrlimit(RLIMIT_MEMLOCK, &lim))
|
|
ksft_exit_fail_msg("Failed to set rlimit: %s\n", strerror(errno));
|
|
|
|
page_size = getpagesize();
|
|
|
|
if (read_memory_info(&mem_free, &hugepage_size) != 0)
|
|
ksft_exit_fail_msg("Failed to get meminfo\n");
|
|
|
|
mem_fragmentable_MB = mem_free * 0.8 / 1024;
|
|
|
|
while (mem_fragmentable_MB > 0) {
|
|
map = mmap(NULL, MAP_SIZE, PROT_READ | PROT_WRITE,
|
|
MAP_ANONYMOUS | MAP_PRIVATE | MAP_LOCKED, -1, 0);
|
|
if (map == MAP_FAILED)
|
|
break;
|
|
|
|
entry = malloc(sizeof(struct map_list));
|
|
if (!entry) {
|
|
munmap(map, MAP_SIZE);
|
|
break;
|
|
}
|
|
entry->map = map;
|
|
entry->next = list;
|
|
list = entry;
|
|
|
|
/* Write something (in this case the address of the map) to
|
|
* ensure that KSM can't merge the mapped pages
|
|
*/
|
|
for (i = 0; i < MAP_SIZE; i += page_size)
|
|
*(unsigned long *)(map + i) = (unsigned long)map + i;
|
|
|
|
mem_fragmentable_MB -= MAP_SIZE_MB;
|
|
}
|
|
|
|
for (entry = list; entry != NULL; entry = entry->next) {
|
|
munmap(entry->map, MAP_SIZE);
|
|
if (!entry->next)
|
|
break;
|
|
entry = entry->next;
|
|
}
|
|
|
|
if (check_compaction(mem_free, hugepage_size,
|
|
initial_nr_hugepages) == 0)
|
|
ksft_exit_pass();
|
|
|
|
ksft_exit_fail();
|
|
}
|