mirror of
https://gitee.com/bianbu-linux/linux-6.6
synced 2025-04-24 14:07:52 -04:00
* Support for the Svpbmt extension, which allows memory attributes to be encoded in pages. * Support for the Allwinner D1's implementation of page-based memory attributes. * Support for running rv32 binaries on rv64 systems, via the compat subsystem. * Support for kexec_file(). * Support for the new generic ticket-based spinlocks, which allows us to also move to qrwlock. These should have already gone in through the asm-geneic tree as well. * A handful of cleanups and fixes, include some larger ones around atomics and XIP. -----BEGIN PGP SIGNATURE----- iQJHBAABCAAxFiEEKzw3R0RoQ7JKlDp6LhMZ81+7GIkFAmKWOx8THHBhbG1lckBk YWJiZWx0LmNvbQAKCRAuExnzX7sYieAiEADAUdP7ctoaSQwk5skd/fdA3b4KJuKn 1Zjl+Br32WP0DlbirYBYWRUQZnCCsvABbTiwSJMcG7NBpU5pyQ5XDtB3OA5kJswO Fdp8Nd53//+GK1M5zdEM9OdgvT9fbfTZ3qTu8bKsROOQhGwnYL+Csc9KjFRqEmzN oQii0jlb3n5PM4FL3GsbV4uMn9zzkP9mnVAPQktcock2EKFEK/Fy3uNYMQiO2KPi n8O6bIDaeRdQ6SurzWOuOkt0cro0tEF85ilzT04mynQsOU0el5oGqCxnOhNH3VWg ndqPT6Yafw12hZOtbKJeP+nF8IIR6aJLP3jOtRwEVgcfbXYAw4QwbAV8kQZISefN ipn8JGY7GX9Y9TYU692OUGkcmAb3/dxb6c0WihBdvJ0M6YyLD5X+YKHNuG2onLgK ss43C5Mxsu629rsjdu/PV91B1+pve3rG9siVmF+g4eo0x9rjMq6/JB0Kal/8SLI1 Je5T55d5ujV1a2XxhZLQOSD5owrK7J1M9owb0bloTnr9nVwFTWDrfEQEU82o3kP+ Xm+FfXktnz9ai55NjkMbbEur5D++dKJhBavwCTnBcTrJmMtEH0R45GTK9ZehP+WC rNVrRXjIsS18wsTfJxnkZeFQA38as6VBKTzvwHvOgzTrrZU1/xk3lpkouYtAO6BG gKacHshVilmUuA== =Loi6 -----END PGP SIGNATURE----- Merge tag 'riscv-for-linus-5.19-mw0' of git://git.kernel.org/pub/scm/linux/kernel/git/riscv/linux Pull RISC-V updates from Palmer Dabbelt: - Support for the Svpbmt extension, which allows memory attributes to be encoded in pages - Support for the Allwinner D1's implementation of page-based memory attributes - Support for running rv32 binaries on rv64 systems, via the compat subsystem - Support for kexec_file() - Support for the new generic ticket-based spinlocks, which allows us to also move to qrwlock. These should have already gone in through the asm-geneic tree as well - A handful of cleanups and fixes, include some larger ones around atomics and XIP * tag 'riscv-for-linus-5.19-mw0' of git://git.kernel.org/pub/scm/linux/kernel/git/riscv/linux: (51 commits) RISC-V: Prepare dropping week attribute from arch_kexec_apply_relocations[_add] riscv: compat: Using seperated vdso_maps for compat_vdso_info RISC-V: Fix the XIP build RISC-V: Split out the XIP fixups into their own file RISC-V: ignore xipImage RISC-V: Avoid empty create_*_mapping definitions riscv: Don't output a bogus mmu-type on a no MMU kernel riscv: atomic: Add custom conditional atomic operation implementation riscv: atomic: Optimize dec_if_positive functions riscv: atomic: Cleanup unnecessary definition RISC-V: Load purgatory in kexec_file RISC-V: Add purgatory RISC-V: Support for kexec_file on panic RISC-V: Add kexec_file support RISC-V: use memcpy for kexec_file mode kexec_file: Fix kexec_file.c build error for riscv platform riscv: compat: Add COMPAT Kbuild skeletal support riscv: compat: ptrace: Add compat_arch_ptrace implement riscv: compat: signal: Add rt_frame implementation riscv: add memory-type errata for T-Head ...
373 lines
8.3 KiB
C
373 lines
8.3 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Copyright (C) 2012 Regents of the University of California
|
|
*/
|
|
|
|
#ifndef _ASM_RISCV_PGTABLE_64_H
|
|
#define _ASM_RISCV_PGTABLE_64_H
|
|
|
|
#include <linux/bits.h>
|
|
#include <linux/const.h>
|
|
#include <asm/errata_list.h>
|
|
|
|
extern bool pgtable_l4_enabled;
|
|
extern bool pgtable_l5_enabled;
|
|
|
|
#define PGDIR_SHIFT_L3 30
|
|
#define PGDIR_SHIFT_L4 39
|
|
#define PGDIR_SHIFT_L5 48
|
|
#define PGDIR_SIZE_L3 (_AC(1, UL) << PGDIR_SHIFT_L3)
|
|
|
|
#define PGDIR_SHIFT (pgtable_l5_enabled ? PGDIR_SHIFT_L5 : \
|
|
(pgtable_l4_enabled ? PGDIR_SHIFT_L4 : PGDIR_SHIFT_L3))
|
|
/* Size of region mapped by a page global directory */
|
|
#define PGDIR_SIZE (_AC(1, UL) << PGDIR_SHIFT)
|
|
#define PGDIR_MASK (~(PGDIR_SIZE - 1))
|
|
|
|
/* p4d is folded into pgd in case of 4-level page table */
|
|
#define P4D_SHIFT 39
|
|
#define P4D_SIZE (_AC(1, UL) << P4D_SHIFT)
|
|
#define P4D_MASK (~(P4D_SIZE - 1))
|
|
|
|
/* pud is folded into pgd in case of 3-level page table */
|
|
#define PUD_SHIFT 30
|
|
#define PUD_SIZE (_AC(1, UL) << PUD_SHIFT)
|
|
#define PUD_MASK (~(PUD_SIZE - 1))
|
|
|
|
#define PMD_SHIFT 21
|
|
/* Size of region mapped by a page middle directory */
|
|
#define PMD_SIZE (_AC(1, UL) << PMD_SHIFT)
|
|
#define PMD_MASK (~(PMD_SIZE - 1))
|
|
|
|
/* Page 4th Directory entry */
|
|
typedef struct {
|
|
unsigned long p4d;
|
|
} p4d_t;
|
|
|
|
#define p4d_val(x) ((x).p4d)
|
|
#define __p4d(x) ((p4d_t) { (x) })
|
|
#define PTRS_PER_P4D (PAGE_SIZE / sizeof(p4d_t))
|
|
|
|
/* Page Upper Directory entry */
|
|
typedef struct {
|
|
unsigned long pud;
|
|
} pud_t;
|
|
|
|
#define pud_val(x) ((x).pud)
|
|
#define __pud(x) ((pud_t) { (x) })
|
|
#define PTRS_PER_PUD (PAGE_SIZE / sizeof(pud_t))
|
|
|
|
/* Page Middle Directory entry */
|
|
typedef struct {
|
|
unsigned long pmd;
|
|
} pmd_t;
|
|
|
|
#define pmd_val(x) ((x).pmd)
|
|
#define __pmd(x) ((pmd_t) { (x) })
|
|
|
|
#define PTRS_PER_PMD (PAGE_SIZE / sizeof(pmd_t))
|
|
|
|
/*
|
|
* rv64 PTE format:
|
|
* | 63 | 62 61 | 60 54 | 53 10 | 9 8 | 7 | 6 | 5 | 4 | 3 | 2 | 1 | 0
|
|
* N MT RSV PFN reserved for SW D A G U X W R V
|
|
*/
|
|
#define _PAGE_PFN_MASK GENMASK(53, 10)
|
|
|
|
/*
|
|
* [62:61] Svpbmt Memory Type definitions:
|
|
*
|
|
* 00 - PMA Normal Cacheable, No change to implied PMA memory type
|
|
* 01 - NC Non-cacheable, idempotent, weakly-ordered Main Memory
|
|
* 10 - IO Non-cacheable, non-idempotent, strongly-ordered I/O memory
|
|
* 11 - Rsvd Reserved for future standard use
|
|
*/
|
|
#define _PAGE_NOCACHE_SVPBMT (1UL << 61)
|
|
#define _PAGE_IO_SVPBMT (1UL << 62)
|
|
#define _PAGE_MTMASK_SVPBMT (_PAGE_NOCACHE_SVPBMT | _PAGE_IO_SVPBMT)
|
|
|
|
/*
|
|
* [63:59] T-Head Memory Type definitions:
|
|
*
|
|
* 00000 - NC Weakly-ordered, Non-cacheable, Non-bufferable, Non-shareable, Non-trustable
|
|
* 01110 - PMA Weakly-ordered, Cacheable, Bufferable, Shareable, Non-trustable
|
|
* 10000 - IO Strongly-ordered, Non-cacheable, Non-bufferable, Non-shareable, Non-trustable
|
|
*/
|
|
#define _PAGE_PMA_THEAD ((1UL << 62) | (1UL << 61) | (1UL << 60))
|
|
#define _PAGE_NOCACHE_THEAD 0UL
|
|
#define _PAGE_IO_THEAD (1UL << 63)
|
|
#define _PAGE_MTMASK_THEAD (_PAGE_PMA_THEAD | _PAGE_IO_THEAD | (1UL << 59))
|
|
|
|
static inline u64 riscv_page_mtmask(void)
|
|
{
|
|
u64 val;
|
|
|
|
ALT_SVPBMT(val, _PAGE_MTMASK);
|
|
return val;
|
|
}
|
|
|
|
static inline u64 riscv_page_nocache(void)
|
|
{
|
|
u64 val;
|
|
|
|
ALT_SVPBMT(val, _PAGE_NOCACHE);
|
|
return val;
|
|
}
|
|
|
|
static inline u64 riscv_page_io(void)
|
|
{
|
|
u64 val;
|
|
|
|
ALT_SVPBMT(val, _PAGE_IO);
|
|
return val;
|
|
}
|
|
|
|
#define _PAGE_NOCACHE riscv_page_nocache()
|
|
#define _PAGE_IO riscv_page_io()
|
|
#define _PAGE_MTMASK riscv_page_mtmask()
|
|
|
|
/* Set of bits to preserve across pte_modify() */
|
|
#define _PAGE_CHG_MASK (~(unsigned long)(_PAGE_PRESENT | _PAGE_READ | \
|
|
_PAGE_WRITE | _PAGE_EXEC | \
|
|
_PAGE_USER | _PAGE_GLOBAL | \
|
|
_PAGE_MTMASK))
|
|
|
|
static inline int pud_present(pud_t pud)
|
|
{
|
|
return (pud_val(pud) & _PAGE_PRESENT);
|
|
}
|
|
|
|
static inline int pud_none(pud_t pud)
|
|
{
|
|
return (pud_val(pud) == 0);
|
|
}
|
|
|
|
static inline int pud_bad(pud_t pud)
|
|
{
|
|
return !pud_present(pud);
|
|
}
|
|
|
|
#define pud_leaf pud_leaf
|
|
static inline int pud_leaf(pud_t pud)
|
|
{
|
|
return pud_present(pud) && (pud_val(pud) & _PAGE_LEAF);
|
|
}
|
|
|
|
static inline int pud_user(pud_t pud)
|
|
{
|
|
return pud_val(pud) & _PAGE_USER;
|
|
}
|
|
|
|
static inline void set_pud(pud_t *pudp, pud_t pud)
|
|
{
|
|
*pudp = pud;
|
|
}
|
|
|
|
static inline void pud_clear(pud_t *pudp)
|
|
{
|
|
set_pud(pudp, __pud(0));
|
|
}
|
|
|
|
static inline pud_t pfn_pud(unsigned long pfn, pgprot_t prot)
|
|
{
|
|
return __pud((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
|
|
}
|
|
|
|
static inline unsigned long _pud_pfn(pud_t pud)
|
|
{
|
|
return pud_val(pud) >> _PAGE_PFN_SHIFT;
|
|
}
|
|
|
|
static inline pmd_t *pud_pgtable(pud_t pud)
|
|
{
|
|
return (pmd_t *)pfn_to_virt(__page_val_to_pfn(pud_val(pud)));
|
|
}
|
|
|
|
static inline struct page *pud_page(pud_t pud)
|
|
{
|
|
return pfn_to_page(__page_val_to_pfn(pud_val(pud)));
|
|
}
|
|
|
|
#define mm_p4d_folded mm_p4d_folded
|
|
static inline bool mm_p4d_folded(struct mm_struct *mm)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
#define mm_pud_folded mm_pud_folded
|
|
static inline bool mm_pud_folded(struct mm_struct *mm)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
#define pmd_index(addr) (((addr) >> PMD_SHIFT) & (PTRS_PER_PMD - 1))
|
|
|
|
static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot)
|
|
{
|
|
unsigned long prot_val = pgprot_val(prot);
|
|
|
|
ALT_THEAD_PMA(prot_val);
|
|
|
|
return __pmd((pfn << _PAGE_PFN_SHIFT) | prot_val);
|
|
}
|
|
|
|
static inline unsigned long _pmd_pfn(pmd_t pmd)
|
|
{
|
|
return __page_val_to_pfn(pmd_val(pmd));
|
|
}
|
|
|
|
#define mk_pmd(page, prot) pfn_pmd(page_to_pfn(page), prot)
|
|
|
|
#define pmd_ERROR(e) \
|
|
pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
|
|
|
|
#define pud_ERROR(e) \
|
|
pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
|
|
|
|
#define p4d_ERROR(e) \
|
|
pr_err("%s:%d: bad p4d %016lx.\n", __FILE__, __LINE__, p4d_val(e))
|
|
|
|
static inline void set_p4d(p4d_t *p4dp, p4d_t p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
*p4dp = p4d;
|
|
else
|
|
set_pud((pud_t *)p4dp, (pud_t){ p4d_val(p4d) });
|
|
}
|
|
|
|
static inline int p4d_none(p4d_t p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return (p4d_val(p4d) == 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int p4d_present(p4d_t p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return (p4d_val(p4d) & _PAGE_PRESENT);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static inline int p4d_bad(p4d_t p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return !p4d_present(p4d);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline void p4d_clear(p4d_t *p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
set_p4d(p4d, __p4d(0));
|
|
}
|
|
|
|
static inline p4d_t pfn_p4d(unsigned long pfn, pgprot_t prot)
|
|
{
|
|
return __p4d((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
|
|
}
|
|
|
|
static inline unsigned long _p4d_pfn(p4d_t p4d)
|
|
{
|
|
return p4d_val(p4d) >> _PAGE_PFN_SHIFT;
|
|
}
|
|
|
|
static inline pud_t *p4d_pgtable(p4d_t p4d)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return (pud_t *)pfn_to_virt(p4d_val(p4d) >> _PAGE_PFN_SHIFT);
|
|
|
|
return (pud_t *)pud_pgtable((pud_t) { p4d_val(p4d) });
|
|
}
|
|
#define p4d_page_vaddr(p4d) ((unsigned long)p4d_pgtable(p4d))
|
|
|
|
static inline struct page *p4d_page(p4d_t p4d)
|
|
{
|
|
return pfn_to_page(p4d_val(p4d) >> _PAGE_PFN_SHIFT);
|
|
}
|
|
|
|
#define pud_index(addr) (((addr) >> PUD_SHIFT) & (PTRS_PER_PUD - 1))
|
|
|
|
#define pud_offset pud_offset
|
|
static inline pud_t *pud_offset(p4d_t *p4d, unsigned long address)
|
|
{
|
|
if (pgtable_l4_enabled)
|
|
return p4d_pgtable(*p4d) + pud_index(address);
|
|
|
|
return (pud_t *)p4d;
|
|
}
|
|
|
|
static inline void set_pgd(pgd_t *pgdp, pgd_t pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
*pgdp = pgd;
|
|
else
|
|
set_p4d((p4d_t *)pgdp, (p4d_t){ pgd_val(pgd) });
|
|
}
|
|
|
|
static inline int pgd_none(pgd_t pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return (pgd_val(pgd) == 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int pgd_present(pgd_t pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return (pgd_val(pgd) & _PAGE_PRESENT);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static inline int pgd_bad(pgd_t pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return !pgd_present(pgd);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline void pgd_clear(pgd_t *pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
set_pgd(pgd, __pgd(0));
|
|
}
|
|
|
|
static inline p4d_t *pgd_pgtable(pgd_t pgd)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return (p4d_t *)pfn_to_virt(pgd_val(pgd) >> _PAGE_PFN_SHIFT);
|
|
|
|
return (p4d_t *)p4d_pgtable((p4d_t) { pgd_val(pgd) });
|
|
}
|
|
#define pgd_page_vaddr(pgd) ((unsigned long)pgd_pgtable(pgd))
|
|
|
|
static inline struct page *pgd_page(pgd_t pgd)
|
|
{
|
|
return pfn_to_page(pgd_val(pgd) >> _PAGE_PFN_SHIFT);
|
|
}
|
|
#define pgd_page(pgd) pgd_page(pgd)
|
|
|
|
#define p4d_index(addr) (((addr) >> P4D_SHIFT) & (PTRS_PER_P4D - 1))
|
|
|
|
#define p4d_offset p4d_offset
|
|
static inline p4d_t *p4d_offset(pgd_t *pgd, unsigned long address)
|
|
{
|
|
if (pgtable_l5_enabled)
|
|
return pgd_pgtable(*pgd) + p4d_index(address);
|
|
|
|
return (p4d_t *)pgd;
|
|
}
|
|
|
|
#endif /* _ASM_RISCV_PGTABLE_64_H */
|