!C99Shell v. 2.0 [PHP 7 Update] [25.02.2019]!

Software: nginx/1.23.4. PHP/5.6.40-65+ubuntu20.04.1+deb.sury.org+1 

uname -a: Linux foro-restaurado-2 5.15.0-1040-oracle #46-Ubuntu SMP Fri Jul 14 21:47:21 UTC 2023
aarch64
 

uid=33(www-data) gid=33(www-data) groups=33(www-data) 

Safe-mode: OFF (not secure)

/usr/src/linux-oracle-headers-5.15.0-1040/arch/openrisc/include/asm/   drwxr-xr-x
Free 83.24 GB of 96.73 GB (86.06%)
Home    Back    Forward    UPDIR    Refresh    Search    Buffer    Encoder    Tools    Proc.    FTP brute    Sec.    SQL    PHP-code    Update    Feedback    Self remove    Logout    


Viewing file:     atomic.h (3.52 KB)      -rw-r--r--
Select action/file-type:
(+) | (+) | (+) | Code (+) | Session (+) | (+) | SDB (+) | (+) | (+) | (+) | (+) | (+) |
/*
 * Copyright (C) 2014 Stefan Kristiansson <[email protected]>
 *
 * This file is licensed under the terms of the GNU General Public License
 * version 2.  This program is licensed "as is" without any warranty of any
 * kind, whether express or implied.
 */

#ifndef __ASM_OPENRISC_ATOMIC_H
#define __ASM_OPENRISC_ATOMIC_H

#include <linux/types.h>

/* Atomically perform op with v->counter and i */
#define ATOMIC_OP(op)                            \
static inline void arch_atomic_##op(int i, atomic_t *v)            \
{                                    \
    int tmp;                            \
                                    \
    __asm__ __volatile__(                        \
        "1:    l.lwa    %0,0(%1)    \n"            \
        "    l." #op " %0,%0,%2    \n"            \
        "    l.swa    0(%1),%0    \n"            \
        "    l.bnf    1b        \n"            \
        "     l.nop            \n"            \
        : "=&r"(tmp)                        \
        : "r"(&v->counter), "r"(i)                \
        : "cc", "memory");                    \
}

/* Atomically perform op with v->counter and i, return the result */
#define ATOMIC_OP_RETURN(op)                        \
static inline int arch_atomic_##op##_return(int i, atomic_t *v)        \
{                                    \
    int tmp;                            \
                                    \
    __asm__ __volatile__(                        \
        "1:    l.lwa    %0,0(%1)    \n"            \
        "    l." #op " %0,%0,%2    \n"            \
        "    l.swa    0(%1),%0    \n"            \
        "    l.bnf    1b        \n"            \
        "     l.nop            \n"            \
        : "=&r"(tmp)                        \
        : "r"(&v->counter), "r"(i)                \
        : "cc", "memory");                    \
                                    \
    return tmp;                            \
}

/* Atomically perform op with v->counter and i, return orig v->counter */
#define ATOMIC_FETCH_OP(op)                        \
static inline int arch_atomic_fetch_##op(int i, atomic_t *v)        \
{                                    \
    int tmp, old;                            \
                                    \
    __asm__ __volatile__(                        \
        "1:    l.lwa    %0,0(%2)    \n"            \
        "    l." #op " %1,%0,%3    \n"            \
        "    l.swa    0(%2),%1    \n"            \
        "    l.bnf    1b        \n"            \
        "     l.nop            \n"            \
        : "=&r"(old), "=&r"(tmp)                \
        : "r"(&v->counter), "r"(i)                \
        : "cc", "memory");                    \
                                    \
    return old;                            \
}

ATOMIC_OP_RETURN(add)
ATOMIC_OP_RETURN(sub)

ATOMIC_FETCH_OP(add)
ATOMIC_FETCH_OP(sub)
ATOMIC_FETCH_OP(and)
ATOMIC_FETCH_OP(or)
ATOMIC_FETCH_OP(xor)

ATOMIC_OP(add)
ATOMIC_OP(sub)
ATOMIC_OP(and)
ATOMIC_OP(or)
ATOMIC_OP(xor)

#undef ATOMIC_FETCH_OP
#undef ATOMIC_OP_RETURN
#undef ATOMIC_OP

#define arch_atomic_add_return    arch_atomic_add_return
#define arch_atomic_sub_return    arch_atomic_sub_return
#define arch_atomic_fetch_add    arch_atomic_fetch_add
#define arch_atomic_fetch_sub    arch_atomic_fetch_sub
#define arch_atomic_fetch_and    arch_atomic_fetch_and
#define arch_atomic_fetch_or    arch_atomic_fetch_or
#define arch_atomic_fetch_xor    arch_atomic_fetch_xor
#define arch_atomic_add        arch_atomic_add
#define arch_atomic_sub        arch_atomic_sub
#define arch_atomic_and        arch_atomic_and
#define arch_atomic_or        arch_atomic_or
#define arch_atomic_xor        arch_atomic_xor

/*
 * Atomically add a to v->counter as long as v is not already u.
 * Returns the original value at v->counter.
 *
 * This is often used through atomic_inc_not_zero()
 */
static inline int arch_atomic_fetch_add_unless(atomic_t *v, int a, int u)
{
    int old, tmp;

    __asm__ __volatile__(
        "1:    l.lwa %0, 0(%2)        \n"
        "    l.sfeq %0, %4        \n"
        "    l.bf 2f            \n"
        "     l.add %1, %0, %3    \n"
        "    l.swa 0(%2), %1        \n"
        "    l.bnf 1b        \n"
        "     l.nop            \n"
        "2:                \n"
        : "=&r"(old), "=&r" (tmp)
        : "r"(&v->counter), "r"(a), "r"(u)
        : "cc", "memory");

    return old;
}
#define arch_atomic_fetch_add_unless    arch_atomic_fetch_add_unless

#define arch_atomic_read(v)        READ_ONCE((v)->counter)
#define arch_atomic_set(v,i)        WRITE_ONCE((v)->counter, (i))

#include <asm/cmpxchg.h>

#define arch_atomic_xchg(ptr, v)        (arch_xchg(&(ptr)->counter, (v)))
#define arch_atomic_cmpxchg(v, old, new)    (arch_cmpxchg(&((v)->counter), (old), (new)))

#endif /* __ASM_OPENRISC_ATOMIC_H */

:: Command execute ::

Enter:
 
Select:
 

:: Search ::
  - regexp 

:: Upload ::
 
[ Read-Only ]

:: Make Dir ::
 
[ Read-Only ]
:: Make File ::
 
[ Read-Only ]

:: Go Dir ::
 
:: Go File ::
 

--[ c99shell v. 2.0 [PHP 7 Update] [25.02.2019] maintained by HackingTool | HackingTool | Generation time: 0.0059 ]--