Viewing file: processor.h (5.35 KB) -rw-r--r-- Select action/file-type: (+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
/* SPDX-License-Identifier: GPL-2.0 */ /* * Copyright (C) 2020-2022 Loongson Technology Corporation Limited */ #ifndef _ASM_PROCESSOR_H #define _ASM_PROCESSOR_H
#include <linux/atomic.h> #include <linux/cpumask.h> #include <linux/sizes.h>
#include <asm/cpu.h> #include <asm/cpu-info.h> #include <asm/hw_breakpoint.h> #include <asm/loongarch.h> #include <asm/vdso/processor.h> #include <uapi/asm/ptrace.h> #include <uapi/asm/sigcontext.h>
#ifdef CONFIG_32BIT
#define TASK_SIZE 0x80000000UL #define TASK_SIZE_MIN TASK_SIZE #define STACK_TOP_MAX TASK_SIZE
#define TASK_IS_32BIT_ADDR 1
#endif
#ifdef CONFIG_64BIT
#define TASK_SIZE32 0x100000000UL #define TASK_SIZE64 (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits))
#define TASK_SIZE (test_thread_flag(TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64) #define TASK_SIZE_MIN TASK_SIZE32 #define STACK_TOP_MAX TASK_SIZE64
#define TASK_SIZE_OF(tsk) \ (test_tsk_thread_flag(tsk, TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64)
#define TASK_IS_32BIT_ADDR test_thread_flag(TIF_32BIT_ADDR)
#endif
#define VDSO_RANDOMIZE_SIZE (TASK_IS_32BIT_ADDR ? SZ_1M : SZ_64M)
unsigned long stack_top(void); #define STACK_TOP stack_top()
/* * This decides where the kernel will search for a free chunk of vm * space during mmap's. */ #define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
#define FPU_REG_WIDTH 256 #define FPU_ALIGN __attribute__((aligned(32)))
union fpureg { __u32 val32[FPU_REG_WIDTH / 32]; __u64 val64[FPU_REG_WIDTH / 64]; };
#define FPR_IDX(width, idx) (idx)
#define BUILD_FPR_ACCESS(width) \ static inline u##width get_fpr##width(union fpureg *fpr, unsigned idx) \ { \ return fpr->val##width[FPR_IDX(width, idx)]; \ } \ \ static inline void set_fpr##width(union fpureg *fpr, unsigned int idx, \ u##width val) \ { \ fpr->val##width[FPR_IDX(width, idx)] = val; \ }
BUILD_FPR_ACCESS(32) BUILD_FPR_ACCESS(64)
struct loongarch_fpu { uint64_t fcc; /* 8x8 */ uint32_t fcsr; uint32_t ftop; union fpureg fpr[NUM_FPU_REGS]; };
struct loongarch_lbt { /* Scratch registers */ unsigned long scr0; unsigned long scr1; unsigned long scr2; unsigned long scr3; /* Eflags register */ unsigned long eflags; };
#define INIT_CPUMASK { \ {0,} \ }
#define ARCH_MIN_TASKALIGN 32
struct loongarch_vdso_info;
/* * If you change thread_struct remember to change the #defines below too! */ struct thread_struct { /* Main processor registers. */ unsigned long reg01, reg03, reg22; /* ra sp fp */ unsigned long reg23, reg24, reg25, reg26; /* s0-s3 */ unsigned long reg27, reg28, reg29, reg30, reg31; /* s4-s8 */
/* __schedule() return address / call frame address */ unsigned long sched_ra; unsigned long sched_cfa;
/* CSR registers */ unsigned long csr_prmd; unsigned long csr_crmd; unsigned long csr_euen; unsigned long csr_ecfg; unsigned long csr_badvaddr; /* Last user fault */
/* Other stuff associated with the thread. */ unsigned long trap_nr; unsigned long error_code; unsigned long single_step; /* Used by PTRACE_SINGLESTEP */ struct loongarch_vdso_info *vdso;
/* * FPU & vector registers, must be at the last of inherited * context because they are conditionally copied at fork(). */ struct loongarch_fpu fpu FPU_ALIGN; struct loongarch_lbt lbt; /* Also conditionally copied */
/* Hardware breakpoints pinned to this task. */ struct perf_event *hbp_break[LOONGARCH_MAX_BRP]; struct perf_event *hbp_watch[LOONGARCH_MAX_WRP]; };
#define thread_saved_ra(tsk) (tsk->thread.sched_ra) #define thread_saved_fp(tsk) (tsk->thread.sched_cfa)
#define INIT_THREAD { \ /* \ * Main processor registers \ */ \ .reg01 = 0, \ .reg03 = 0, \ .reg22 = 0, \ .reg23 = 0, \ .reg24 = 0, \ .reg25 = 0, \ .reg26 = 0, \ .reg27 = 0, \ .reg28 = 0, \ .reg29 = 0, \ .reg30 = 0, \ .reg31 = 0, \ .sched_ra = 0, \ .sched_cfa = 0, \ .csr_crmd = 0, \ .csr_prmd = 0, \ .csr_euen = 0, \ .csr_ecfg = 0, \ .csr_badvaddr = 0, \ /* \ * Other stuff associated with the process \ */ \ .trap_nr = 0, \ .error_code = 0, \ /* \ * FPU & vector registers \ */ \ .fpu = { \ .fcc = 0, \ .fcsr = 0, \ .ftop = 0, \ .fpr = {{{0,},},}, \ }, \ .hbp_break = {0}, \ .hbp_watch = {0}, \ }
struct task_struct;
enum idle_boot_override {IDLE_NO_OVERRIDE = 0, IDLE_HALT, IDLE_NOMWAIT, IDLE_POLL};
extern unsigned long boot_option_idle_override; /* * Do necessary setup to start up a newly executed thread. */ extern void start_thread(struct pt_regs *regs, unsigned long pc, unsigned long sp);
unsigned long __get_wchan(struct task_struct *p);
#define __KSTK_TOS(tsk) ((unsigned long)task_stack_page(tsk) + \ THREAD_SIZE - sizeof(struct pt_regs)) #define task_pt_regs(tsk) ((struct pt_regs *)__KSTK_TOS(tsk)) #define KSTK_EIP(tsk) (task_pt_regs(tsk)->csr_era) #define KSTK_ESP(tsk) (task_pt_regs(tsk)->regs[3]) #define KSTK_EUEN(tsk) (task_pt_regs(tsk)->csr_euen) #define KSTK_ECFG(tsk) (task_pt_regs(tsk)->csr_ecfg)
#define return_address() ({__asm__ __volatile__("":::"$1"); __builtin_return_address(0);})
#ifdef CONFIG_CPU_HAS_PREFETCH
#define ARCH_HAS_PREFETCH #define prefetch(x) __builtin_prefetch((x), 0, 1)
#define ARCH_HAS_PREFETCHW #define prefetchw(x) __builtin_prefetch((x), 1, 1)
#endif
#endif /* _ASM_PROCESSOR_H */
|