mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2024-12-29 12:29:35 +03:00
55d7afa4ad
Both Clang's Static Analyzer and Synopsys' Coverity would ignore assertions. Following Clang's advice, we annotate our assertions: https://clang-analyzer.llvm.org/annotations.html#custom_assertions This makes both Clang's Static Analyzer and Coverity properly identify assertions. This change reduced Clang's reported defects from 246 to 180. It also reduced the false positives reported by Coverityi by 10, while enabling Coverity to find 9 more defects that previously were false negatives. A couple examples of this would be CID-1524417 and CID-1524423. After submitting a build to coverity with the modified assertions, CID-1524417 disappeared while the report for CID-1524423 no longer claimed that the assertion tripped. Coincidentally, it turns out that it is possible to more accurately annotate our headers than the Coverity modelling file permits in the case of format strings. Since we can do that and this patch annotates headers whenever `__coverity_panic__()` would have been used in the model file, we drop all models that use `__coverity_panic__()` from the model file. Upon seeing the success in eliminating false positives involving assertions, it occurred to me that we could also modify our headers to eliminate coverity's false positives involving byte swaps. We now have coverity specific byteswap macros, that do nothing, to disable Coverity's false positives when we do byte swaps. This allowed us to also drop the byteswap definitions from the model file. Lastly, a model file update has been done beyond the mentioned deletions: * The definitions of `umem_alloc_aligned()`, `umem_alloc()` andi `umem_zalloc()` were originally implemented in a way that was intended to inform coverity that when KM_SLEEP has been passed these functions, they do not return NULL. A small error in how this was done was found, so we correct it. * Definitions for umem_cache_alloc() and umem_cache_free() have been added. In practice, no false positives were avoided by making these changes, but in the interest of correctness from future coverity builds, we make them anyway. Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Reviewed-by: Ryan Moeller <ryan@iXsystems.com> Signed-off-by: Richard Yao <richard.yao@alumni.stonybrook.edu> Closes #13902
215 lines
6.3 KiB
C
215 lines
6.3 KiB
C
/*
|
|
* Copyright (C) 2007-2010 Lawrence Livermore National Security, LLC.
|
|
* Copyright (C) 2007 The Regents of the University of California.
|
|
* Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
|
|
* Written by Brian Behlendorf <behlendorf1@llnl.gov>.
|
|
* UCRL-CODE-235197
|
|
*
|
|
* This file is part of the SPL, Solaris Porting Layer.
|
|
*
|
|
* The SPL is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2 of the License, or (at your
|
|
* option) any later version.
|
|
*
|
|
* The SPL is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with the SPL. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#ifndef _SPL_KMEM_H
|
|
#define _SPL_KMEM_H
|
|
|
|
#include <sys/debug.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/vmalloc.h>
|
|
|
|
extern int kmem_debugging(void);
|
|
extern char *kmem_vasprintf(const char *fmt, va_list ap)
|
|
__attribute__((format(printf, 1, 0)));
|
|
extern char *kmem_asprintf(const char *fmt, ...)
|
|
__attribute__((format(printf, 1, 2)));
|
|
extern char *kmem_strdup(const char *str);
|
|
extern void kmem_strfree(char *str);
|
|
|
|
/*
|
|
* Memory allocation interfaces
|
|
*/
|
|
#define KM_SLEEP 0x0000 /* can block for memory; success guaranteed */
|
|
#define KM_NOSLEEP 0x0001 /* cannot block for memory; may fail */
|
|
#define KM_PUSHPAGE 0x0004 /* can block for memory; may use reserve */
|
|
#define KM_ZERO 0x1000 /* zero the allocation */
|
|
#define KM_VMEM 0x2000 /* caller is vmem_* wrapper */
|
|
|
|
#define KM_PUBLIC_MASK (KM_SLEEP | KM_NOSLEEP | KM_PUSHPAGE)
|
|
|
|
static int spl_fstrans_check(void);
|
|
void *spl_kvmalloc(size_t size, gfp_t flags);
|
|
|
|
/*
|
|
* Convert a KM_* flags mask to its Linux GFP_* counterpart. The conversion
|
|
* function is context aware which means that KM_SLEEP allocations can be
|
|
* safely used in syncing contexts which have set PF_FSTRANS.
|
|
*/
|
|
static inline gfp_t
|
|
kmem_flags_convert(int flags)
|
|
{
|
|
gfp_t lflags = __GFP_NOWARN | __GFP_COMP;
|
|
|
|
if (flags & KM_NOSLEEP) {
|
|
lflags |= GFP_ATOMIC | __GFP_NORETRY;
|
|
} else {
|
|
lflags |= GFP_KERNEL;
|
|
if (spl_fstrans_check())
|
|
lflags &= ~(__GFP_IO|__GFP_FS);
|
|
}
|
|
|
|
if (flags & KM_PUSHPAGE)
|
|
lflags |= __GFP_HIGH;
|
|
|
|
if (flags & KM_ZERO)
|
|
lflags |= __GFP_ZERO;
|
|
|
|
return (lflags);
|
|
}
|
|
|
|
typedef struct {
|
|
struct task_struct *fstrans_thread;
|
|
unsigned int saved_flags;
|
|
} fstrans_cookie_t;
|
|
|
|
/*
|
|
* Introduced in Linux 3.9, however this cannot be solely relied on before
|
|
* Linux 3.18 as it doesn't turn off __GFP_FS as it should.
|
|
*/
|
|
#ifdef PF_MEMALLOC_NOIO
|
|
#define __SPL_PF_MEMALLOC_NOIO (PF_MEMALLOC_NOIO)
|
|
#else
|
|
#define __SPL_PF_MEMALLOC_NOIO (0)
|
|
#endif
|
|
|
|
/*
|
|
* PF_FSTRANS is removed from Linux 4.12
|
|
*/
|
|
#ifdef PF_FSTRANS
|
|
#define __SPL_PF_FSTRANS (PF_FSTRANS)
|
|
#else
|
|
#define __SPL_PF_FSTRANS (0)
|
|
#endif
|
|
|
|
#define SPL_FSTRANS (__SPL_PF_FSTRANS|__SPL_PF_MEMALLOC_NOIO)
|
|
|
|
static inline fstrans_cookie_t
|
|
spl_fstrans_mark(void)
|
|
{
|
|
fstrans_cookie_t cookie;
|
|
|
|
BUILD_BUG_ON(SPL_FSTRANS == 0);
|
|
|
|
cookie.fstrans_thread = current;
|
|
cookie.saved_flags = current->flags & SPL_FSTRANS;
|
|
current->flags |= SPL_FSTRANS;
|
|
|
|
return (cookie);
|
|
}
|
|
|
|
static inline void
|
|
spl_fstrans_unmark(fstrans_cookie_t cookie)
|
|
{
|
|
ASSERT3P(cookie.fstrans_thread, ==, current);
|
|
ASSERT((current->flags & SPL_FSTRANS) == SPL_FSTRANS);
|
|
|
|
current->flags &= ~SPL_FSTRANS;
|
|
current->flags |= cookie.saved_flags;
|
|
}
|
|
|
|
static inline int
|
|
spl_fstrans_check(void)
|
|
{
|
|
return (current->flags & SPL_FSTRANS);
|
|
}
|
|
|
|
/*
|
|
* specifically used to check PF_FSTRANS flag, cannot be relied on for
|
|
* checking spl_fstrans_mark().
|
|
*/
|
|
static inline int
|
|
__spl_pf_fstrans_check(void)
|
|
{
|
|
return (current->flags & __SPL_PF_FSTRANS);
|
|
}
|
|
|
|
/*
|
|
* Kernel compatibility for GFP flags
|
|
*/
|
|
/* < 4.13 */
|
|
#ifndef __GFP_RETRY_MAYFAIL
|
|
#define __GFP_RETRY_MAYFAIL __GFP_REPEAT
|
|
#endif
|
|
/* < 4.4 */
|
|
#ifndef __GFP_RECLAIM
|
|
#define __GFP_RECLAIM __GFP_WAIT
|
|
#endif
|
|
|
|
#ifdef HAVE_ATOMIC64_T
|
|
#define kmem_alloc_used_add(size) atomic64_add(size, &kmem_alloc_used)
|
|
#define kmem_alloc_used_sub(size) atomic64_sub(size, &kmem_alloc_used)
|
|
#define kmem_alloc_used_read() atomic64_read(&kmem_alloc_used)
|
|
#define kmem_alloc_used_set(size) atomic64_set(&kmem_alloc_used, size)
|
|
extern atomic64_t kmem_alloc_used;
|
|
extern unsigned long long kmem_alloc_max;
|
|
#else /* HAVE_ATOMIC64_T */
|
|
#define kmem_alloc_used_add(size) atomic_add(size, &kmem_alloc_used)
|
|
#define kmem_alloc_used_sub(size) atomic_sub(size, &kmem_alloc_used)
|
|
#define kmem_alloc_used_read() atomic_read(&kmem_alloc_used)
|
|
#define kmem_alloc_used_set(size) atomic_set(&kmem_alloc_used, size)
|
|
extern atomic_t kmem_alloc_used;
|
|
extern unsigned long long kmem_alloc_max;
|
|
#endif /* HAVE_ATOMIC64_T */
|
|
|
|
extern unsigned int spl_kmem_alloc_warn;
|
|
extern unsigned int spl_kmem_alloc_max;
|
|
|
|
#define kmem_alloc(sz, fl) spl_kmem_alloc((sz), (fl), __func__, __LINE__)
|
|
#define kmem_zalloc(sz, fl) spl_kmem_zalloc((sz), (fl), __func__, __LINE__)
|
|
#define kmem_free(ptr, sz) spl_kmem_free((ptr), (sz))
|
|
#define kmem_cache_reap_active spl_kmem_cache_reap_active
|
|
|
|
extern void *spl_kmem_alloc(size_t sz, int fl, const char *func, int line)
|
|
__attribute__((alloc_size(1)));
|
|
extern void *spl_kmem_zalloc(size_t sz, int fl, const char *func, int line)
|
|
__attribute__((alloc_size(1)));
|
|
extern void spl_kmem_free(const void *ptr, size_t sz);
|
|
|
|
/*
|
|
* 5.8 API change, pgprot_t argument removed.
|
|
*/
|
|
#ifdef HAVE_VMALLOC_PAGE_KERNEL
|
|
#define spl_vmalloc(size, flags) __vmalloc(size, flags, PAGE_KERNEL)
|
|
#else
|
|
#define spl_vmalloc(size, flags) __vmalloc(size, flags)
|
|
#endif
|
|
|
|
/*
|
|
* The following functions are only available for internal use.
|
|
*/
|
|
extern void *spl_kmem_alloc_impl(size_t size, int flags, int node);
|
|
extern void *spl_kmem_alloc_debug(size_t size, int flags, int node);
|
|
extern void *spl_kmem_alloc_track(size_t size, int flags,
|
|
const char *func, int line, int node);
|
|
extern void spl_kmem_free_impl(const void *buf, size_t size);
|
|
extern void spl_kmem_free_debug(const void *buf, size_t size);
|
|
extern void spl_kmem_free_track(const void *buf, size_t size);
|
|
|
|
extern int spl_kmem_init(void);
|
|
extern void spl_kmem_fini(void);
|
|
extern int spl_kmem_cache_reap_active(void);
|
|
|
|
#endif /* _SPL_KMEM_H */
|