aboutsummaryrefslogtreecommitdiff
path: root/utils/lvm2/patches/003-no-mallinfo.patch
blob: 8f7f25d1d35c1d92188771e7492b06b3dc72e887 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
--- a/lib/mm/memlock.c
+++ b/lib/mm/memlock.c
@@ -183,12 +183,15 @@ static void _allocate_memory(void)
          *  memory on free(), this is good enough for our purposes.
          */
 	while (missing > 0) {
+#ifdef __GLIBC__
 		struct mallinfo inf = mallinfo();
 		hblks = inf.hblks;
+#endif
 
 		if ((areas[area] = malloc(_size_malloc_tmp)))
 			_touch_memory(areas[area], _size_malloc_tmp);
 
+#ifdef __GLIBC__
 		inf = mallinfo();
 
 		if (hblks < inf.hblks) {
@@ -198,9 +201,12 @@ static void _allocate_memory(void)
 			free(areas[area]);
 			_size_malloc_tmp /= 2;
 		} else {
+#endif
 			++ area;
 			missing -= _size_malloc_tmp;
+#ifdef __GLIBC__
 		}
+#endif
 
 		if (area == max_areas && missing > 0) {
 			/* Too bad. Warn the user and proceed, as things are
@@ -521,8 +527,13 @@ static void _lock_mem(struct cmd_context
 	 * will not block memory locked thread
 	 * Note: assuming _memlock_count_daemon is updated before _memlock_count
 	 */
+#ifdef __GLIBC__
 	_use_mlockall = _memlock_count_daemon ? 1 :
 		find_config_tree_bool(cmd, activation_use_mlockall_CFG, NULL);
+#else
+	/* always use mlockall on musl */
+	_use_mlockall = 1;
+#endif
 
 	if (!_use_mlockall) {
 		if (!*_procselfmaps &&