mirror of
https://github.com/Motorhead1991/qemu.git
synced 2025-08-30 21:42:06 -06:00
accel/tcg: Move user-related declarations out of 'exec/cpu-all.h' (2/4)
Move declarations related to page protection under user emulation from "exec/cpu-all.h" to "user/page-protection.h". Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org> Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Message-Id: <20241212185341.2857-13-philmd@linaro.org>
This commit is contained in:
parent
970ae60e9b
commit
62ef949bbc
4 changed files with 61 additions and 56 deletions
|
@ -57,6 +57,7 @@
|
||||||
|
|
||||||
#include "qemu-bsd.h"
|
#include "qemu-bsd.h"
|
||||||
#include "exec/page-protection.h"
|
#include "exec/page-protection.h"
|
||||||
|
#include "user/page-protection.h"
|
||||||
|
|
||||||
extern struct bsd_shm_regions bsd_shm_regions[];
|
extern struct bsd_shm_regions bsd_shm_regions[];
|
||||||
extern abi_ulong target_brk;
|
extern abi_ulong target_brk;
|
||||||
|
|
|
@ -130,21 +130,6 @@ extern const TargetPageBits target_page;
|
||||||
|
|
||||||
int page_get_flags(target_ulong address);
|
int page_get_flags(target_ulong address);
|
||||||
|
|
||||||
/**
|
|
||||||
* page_set_flags:
|
|
||||||
* @start: first byte of range
|
|
||||||
* @last: last byte of range
|
|
||||||
* @flags: flags to set
|
|
||||||
* Context: holding mmap lock
|
|
||||||
*
|
|
||||||
* Modify the flags of a page and invalidate the code if necessary.
|
|
||||||
* The flag PAGE_WRITE_ORG is positioned automatically depending
|
|
||||||
* on PAGE_WRITE. The mmap_lock should already be held.
|
|
||||||
*/
|
|
||||||
void page_set_flags(target_ulong start, target_ulong last, int flags);
|
|
||||||
|
|
||||||
void page_reset_target_data(target_ulong start, target_ulong last);
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* page_check_range
|
* page_check_range
|
||||||
* @start: first byte of range
|
* @start: first byte of range
|
||||||
|
@ -157,46 +142,6 @@ void page_reset_target_data(target_ulong start, target_ulong last);
|
||||||
*/
|
*/
|
||||||
bool page_check_range(target_ulong start, target_ulong last, int flags);
|
bool page_check_range(target_ulong start, target_ulong last, int flags);
|
||||||
|
|
||||||
/**
|
|
||||||
* page_check_range_empty:
|
|
||||||
* @start: first byte of range
|
|
||||||
* @last: last byte of range
|
|
||||||
* Context: holding mmap lock
|
|
||||||
*
|
|
||||||
* Return true if the entire range [@start, @last] is unmapped.
|
|
||||||
* The memory lock must be held so that the caller will can ensure
|
|
||||||
* the result stays true until a new mapping can be installed.
|
|
||||||
*/
|
|
||||||
bool page_check_range_empty(target_ulong start, target_ulong last);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* page_find_range_empty
|
|
||||||
* @min: first byte of search range
|
|
||||||
* @max: last byte of search range
|
|
||||||
* @len: size of the hole required
|
|
||||||
* @align: alignment of the hole required (power of 2)
|
|
||||||
*
|
|
||||||
* If there is a range [x, x+@len) within [@min, @max] such that
|
|
||||||
* x % @align == 0, then return x. Otherwise return -1.
|
|
||||||
* The memory lock must be held, as the caller will want to ensure
|
|
||||||
* the returned range stays empty until a new mapping can be installed.
|
|
||||||
*/
|
|
||||||
target_ulong page_find_range_empty(target_ulong min, target_ulong max,
|
|
||||||
target_ulong len, target_ulong align);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* page_get_target_data(address)
|
|
||||||
* @address: guest virtual address
|
|
||||||
*
|
|
||||||
* Return TARGET_PAGE_DATA_SIZE bytes of out-of-band data to associate
|
|
||||||
* with the guest page at @address, allocating it if necessary. The
|
|
||||||
* caller should already have verified that the address is valid.
|
|
||||||
*
|
|
||||||
* The memory will be freed when the guest page is deallocated,
|
|
||||||
* e.g. with the munmap system call.
|
|
||||||
*/
|
|
||||||
void *page_get_target_data(target_ulong address)
|
|
||||||
__attribute__((returns_nonnull));
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
CPUArchState *cpu_copy(CPUArchState *env);
|
CPUArchState *cpu_copy(CPUArchState *env);
|
||||||
|
|
|
@ -18,6 +18,63 @@
|
||||||
|
|
||||||
void page_protect(tb_page_addr_t page_addr);
|
void page_protect(tb_page_addr_t page_addr);
|
||||||
int page_unprotect(tb_page_addr_t address, uintptr_t pc);
|
int page_unprotect(tb_page_addr_t address, uintptr_t pc);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* page_set_flags:
|
||||||
|
* @start: first byte of range
|
||||||
|
* @last: last byte of range
|
||||||
|
* @flags: flags to set
|
||||||
|
* Context: holding mmap lock
|
||||||
|
*
|
||||||
|
* Modify the flags of a page and invalidate the code if necessary.
|
||||||
|
* The flag PAGE_WRITE_ORG is positioned automatically depending
|
||||||
|
* on PAGE_WRITE. The mmap_lock should already be held.
|
||||||
|
*/
|
||||||
|
void page_set_flags(target_ulong start, target_ulong last, int flags);
|
||||||
|
|
||||||
|
void page_reset_target_data(target_ulong start, target_ulong last);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* page_check_range_empty:
|
||||||
|
* @start: first byte of range
|
||||||
|
* @last: last byte of range
|
||||||
|
* Context: holding mmap lock
|
||||||
|
*
|
||||||
|
* Return true if the entire range [@start, @last] is unmapped.
|
||||||
|
* The memory lock must be held so that the caller will can ensure
|
||||||
|
* the result stays true until a new mapping can be installed.
|
||||||
|
*/
|
||||||
|
bool page_check_range_empty(target_ulong start, target_ulong last);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* page_find_range_empty
|
||||||
|
* @min: first byte of search range
|
||||||
|
* @max: last byte of search range
|
||||||
|
* @len: size of the hole required
|
||||||
|
* @align: alignment of the hole required (power of 2)
|
||||||
|
*
|
||||||
|
* If there is a range [x, x+@len) within [@min, @max] such that
|
||||||
|
* x % @align == 0, then return x. Otherwise return -1.
|
||||||
|
* The memory lock must be held, as the caller will want to ensure
|
||||||
|
* the returned range stays empty until a new mapping can be installed.
|
||||||
|
*/
|
||||||
|
target_ulong page_find_range_empty(target_ulong min, target_ulong max,
|
||||||
|
target_ulong len, target_ulong align);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* page_get_target_data(address)
|
||||||
|
* @address: guest virtual address
|
||||||
|
*
|
||||||
|
* Return TARGET_PAGE_DATA_SIZE bytes of out-of-band data to associate
|
||||||
|
* with the guest page at @address, allocating it if necessary. The
|
||||||
|
* caller should already have verified that the address is valid.
|
||||||
|
*
|
||||||
|
* The memory will be freed when the guest page is deallocated,
|
||||||
|
* e.g. with the munmap system call.
|
||||||
|
*/
|
||||||
|
__attribute__((returns_nonnull))
|
||||||
|
void *page_get_target_data(target_ulong address);
|
||||||
|
|
||||||
typedef int (*walk_memory_regions_fn)(void *, target_ulong,
|
typedef int (*walk_memory_regions_fn)(void *, target_ulong,
|
||||||
target_ulong, unsigned long);
|
target_ulong, unsigned long);
|
||||||
|
|
||||||
|
|
|
@ -23,7 +23,9 @@
|
||||||
#include "internals.h"
|
#include "internals.h"
|
||||||
#include "exec/exec-all.h"
|
#include "exec/exec-all.h"
|
||||||
#include "exec/page-protection.h"
|
#include "exec/page-protection.h"
|
||||||
#ifndef CONFIG_USER_ONLY
|
#ifdef CONFIG_USER_ONLY
|
||||||
|
#include "user/page-protection.h"
|
||||||
|
#else
|
||||||
#include "exec/ram_addr.h"
|
#include "exec/ram_addr.h"
|
||||||
#endif
|
#endif
|
||||||
#include "exec/cpu_ldst.h"
|
#include "exec/cpu_ldst.h"
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue