mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2024-12-30 21:09:38 +03:00
2e7f664f04
I recently gained the ability to run Clang's static analyzer on the linux kernel modules via a few hacks. This extended coverage to code that was previously missed since Clang's static analyzer only looked at code that we built in userspace. Running it against the Linux kernel modules built from my local branch produced a total of 72 reports against my local branch. Of those, 50 were reports of logic errors and 22 were reports of dead code. Since we already had cleaned up all of the previous dead code reports, I felt it would be a good next step to clean up these dead code reports. Clang did a further breakdown of the dead code reports into: Dead assignment 15 Dead increment 2 Dead nested assignment 5 The benefit of cleaning these up, especially in the case of dead nested assignment, is that they can expose places where our error handling is incorrect. A number of them were fairly straight forward. However several were not: In vdev_disk_physio_completion(), not only were we not using the return value from the static function vdev_disk_dio_put(), but nothing used it, so I changed it to return void and removed the existing (void) cast in the other area where we call it in addition to no longer storing it to a stack value. In FSE_createDTable(), the function is dead code. Its helper function FSE_freeDTable() is also dead code, as are the CPP definitions in `module/zstd/include/zstd_compat_wrapper.h`. We just delete it all. In zfs_zevent_wait(), we have an optimization opportunity. cv_wait_sig() returns 0 if there are waiting signals and 1 if there are none. The Linux SPL version literally returns `signal_pending(current) ? 0 : 1)` and FreeBSD implements the same semantics, we can just do `!cv_wait_sig()` in place of `signal_pending(current)` to avoid unnecessarily calling it again. zfs_setattr() on FreeBSD version did not have error handling issue because the code was removed entirely from FreeBSD version. The error is from updating the attribute directory's files. After some thought, I decided to propapage errors on it to userspace. In zfs_secpolicy_tmp_snapshot(), we ignore a lack of permission from the first check in favor of checking three other permissions. I assume this is intentional. In zfs_create_fs(), the return value of zap_update() was not checked despite setting an important version number. I see no backward compatibility reason to permit failures, so we add an assertion to catch failures. Interestingly, Linux is still using ASSERT(error == 0) from OpenSolaris while FreeBSD has switched to the improved ASSERT0(error) from illumos, although illumos has yet to adopt it here. ASSERT(error == 0) was used on Linux while ASSERT0(error) was used on FreeBSD since the entire file needs conversion and that should be the subject of another patch. dnode_move()'s issue was caused by us not having implemented POINTER_IS_VALID() on Linux. We have a stub in `include/os/linux/spl/sys/kmem_cache.h` for it, when it really should be in `include/os/linux/spl/sys/kmem.h` to be consistent with Illumos/OpenSolaris. FreeBSD put both `POINTER_IS_VALID()` and `POINTER_INVALIDATE()` in `include/os/freebsd/spl/sys/kmem.h`, so we copy what it did. Whenever a report was in platform-specific code, I checked the FreeBSD version to see if it also applied to FreeBSD, but it was only relevant a few times. Lastly, the patch that enabled Clang's static analyzer to be run on the Linux kernel modules needs more work before it can be put into a PR. I plan to do that in the future as part of the on-going static analysis work that I am doing. Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Richard Yao <richard.yao@alumni.stonybrook.edu> Closes #14380
213 lines
8.5 KiB
C
213 lines
8.5 KiB
C
/*
|
|
* Copyright (C) 2007-2010 Lawrence Livermore National Security, LLC.
|
|
* Copyright (C) 2007 The Regents of the University of California.
|
|
* Produced at Lawrence Livermore National Laboratory (cf, DISCLAIMER).
|
|
* Written by Brian Behlendorf <behlendorf1@llnl.gov>.
|
|
* UCRL-CODE-235197
|
|
*
|
|
* This file is part of the SPL, Solaris Porting Layer.
|
|
*
|
|
* The SPL is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; either version 2 of the License, or (at your
|
|
* option) any later version.
|
|
*
|
|
* The SPL is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with the SPL. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#ifndef _SPL_KMEM_CACHE_H
|
|
#define _SPL_KMEM_CACHE_H
|
|
|
|
#include <sys/taskq.h>
|
|
|
|
/*
|
|
* Slab allocation interfaces. The SPL slab differs from the standard
|
|
* Linux SLAB or SLUB primarily in that each cache may be backed by slabs
|
|
* allocated from the physical or virtual memory address space. The virtual
|
|
* slabs allow for good behavior when allocation large objects of identical
|
|
* size. This slab implementation also supports both constructors and
|
|
* destructors which the Linux slab does not.
|
|
*/
|
|
typedef enum kmc_bit {
|
|
KMC_BIT_NODEBUG = 1, /* Default behavior */
|
|
KMC_BIT_KVMEM = 7, /* Use kvmalloc linux allocator */
|
|
KMC_BIT_SLAB = 8, /* Use Linux slab cache */
|
|
KMC_BIT_DEADLOCKED = 14, /* Deadlock detected */
|
|
KMC_BIT_GROWING = 15, /* Growing in progress */
|
|
KMC_BIT_REAPING = 16, /* Reaping in progress */
|
|
KMC_BIT_DESTROY = 17, /* Destroy in progress */
|
|
KMC_BIT_TOTAL = 18, /* Proc handler helper bit */
|
|
KMC_BIT_ALLOC = 19, /* Proc handler helper bit */
|
|
KMC_BIT_MAX = 20, /* Proc handler helper bit */
|
|
} kmc_bit_t;
|
|
|
|
/* kmem move callback return values */
|
|
typedef enum kmem_cbrc {
|
|
KMEM_CBRC_YES = 0, /* Object moved */
|
|
KMEM_CBRC_NO = 1, /* Object not moved */
|
|
KMEM_CBRC_LATER = 2, /* Object not moved, try again later */
|
|
KMEM_CBRC_DONT_NEED = 3, /* Neither object is needed */
|
|
KMEM_CBRC_DONT_KNOW = 4, /* Object unknown */
|
|
} kmem_cbrc_t;
|
|
|
|
#define KMC_NODEBUG (1 << KMC_BIT_NODEBUG)
|
|
#define KMC_KVMEM (1 << KMC_BIT_KVMEM)
|
|
#define KMC_SLAB (1 << KMC_BIT_SLAB)
|
|
#define KMC_DEADLOCKED (1 << KMC_BIT_DEADLOCKED)
|
|
#define KMC_GROWING (1 << KMC_BIT_GROWING)
|
|
#define KMC_REAPING (1 << KMC_BIT_REAPING)
|
|
#define KMC_DESTROY (1 << KMC_BIT_DESTROY)
|
|
#define KMC_TOTAL (1 << KMC_BIT_TOTAL)
|
|
#define KMC_ALLOC (1 << KMC_BIT_ALLOC)
|
|
#define KMC_MAX (1 << KMC_BIT_MAX)
|
|
|
|
#define KMC_REAP_CHUNK INT_MAX
|
|
#define KMC_DEFAULT_SEEKS 1
|
|
|
|
#define KMC_RECLAIM_ONCE 0x1 /* Force a single shrinker pass */
|
|
|
|
extern struct list_head spl_kmem_cache_list;
|
|
extern struct rw_semaphore spl_kmem_cache_sem;
|
|
|
|
#define SKM_MAGIC 0x2e2e2e2e
|
|
#define SKO_MAGIC 0x20202020
|
|
#define SKS_MAGIC 0x22222222
|
|
#define SKC_MAGIC 0x2c2c2c2c
|
|
|
|
#define SPL_KMEM_CACHE_OBJ_PER_SLAB 8 /* Target objects per slab */
|
|
#define SPL_KMEM_CACHE_ALIGN 8 /* Default object alignment */
|
|
#ifdef _LP64
|
|
#define SPL_KMEM_CACHE_MAX_SIZE 32 /* Max slab size in MB */
|
|
#else
|
|
#define SPL_KMEM_CACHE_MAX_SIZE 4 /* Max slab size in MB */
|
|
#endif
|
|
|
|
#define SPL_MAX_ORDER (MAX_ORDER - 3)
|
|
#define SPL_MAX_ORDER_NR_PAGES (1 << (SPL_MAX_ORDER - 1))
|
|
|
|
#ifdef CONFIG_SLUB
|
|
#define SPL_MAX_KMEM_CACHE_ORDER PAGE_ALLOC_COSTLY_ORDER
|
|
#define SPL_MAX_KMEM_ORDER_NR_PAGES (1 << (SPL_MAX_KMEM_CACHE_ORDER - 1))
|
|
#else
|
|
#define SPL_MAX_KMEM_ORDER_NR_PAGES (KMALLOC_MAX_SIZE >> PAGE_SHIFT)
|
|
#endif
|
|
|
|
typedef int (*spl_kmem_ctor_t)(void *, void *, int);
|
|
typedef void (*spl_kmem_dtor_t)(void *, void *);
|
|
|
|
typedef struct spl_kmem_magazine {
|
|
uint32_t skm_magic; /* Sanity magic */
|
|
uint32_t skm_avail; /* Available objects */
|
|
uint32_t skm_size; /* Magazine size */
|
|
uint32_t skm_refill; /* Batch refill size */
|
|
struct spl_kmem_cache *skm_cache; /* Owned by cache */
|
|
unsigned int skm_cpu; /* Owned by cpu */
|
|
void *skm_objs[0]; /* Object pointers */
|
|
} spl_kmem_magazine_t;
|
|
|
|
typedef struct spl_kmem_obj {
|
|
uint32_t sko_magic; /* Sanity magic */
|
|
void *sko_addr; /* Buffer address */
|
|
struct spl_kmem_slab *sko_slab; /* Owned by slab */
|
|
struct list_head sko_list; /* Free object list linkage */
|
|
} spl_kmem_obj_t;
|
|
|
|
typedef struct spl_kmem_slab {
|
|
uint32_t sks_magic; /* Sanity magic */
|
|
uint32_t sks_objs; /* Objects per slab */
|
|
struct spl_kmem_cache *sks_cache; /* Owned by cache */
|
|
struct list_head sks_list; /* Slab list linkage */
|
|
struct list_head sks_free_list; /* Free object list */
|
|
unsigned long sks_age; /* Last modify jiffie */
|
|
uint32_t sks_ref; /* Ref count used objects */
|
|
} spl_kmem_slab_t;
|
|
|
|
typedef struct spl_kmem_alloc {
|
|
struct spl_kmem_cache *ska_cache; /* Owned by cache */
|
|
int ska_flags; /* Allocation flags */
|
|
taskq_ent_t ska_tqe; /* Task queue entry */
|
|
} spl_kmem_alloc_t;
|
|
|
|
typedef struct spl_kmem_emergency {
|
|
struct rb_node ske_node; /* Emergency tree linkage */
|
|
unsigned long ske_obj; /* Buffer address */
|
|
} spl_kmem_emergency_t;
|
|
|
|
typedef struct spl_kmem_cache {
|
|
uint32_t skc_magic; /* Sanity magic */
|
|
uint32_t skc_name_size; /* Name length */
|
|
char *skc_name; /* Name string */
|
|
spl_kmem_magazine_t **skc_mag; /* Per-CPU warm cache */
|
|
uint32_t skc_mag_size; /* Magazine size */
|
|
uint32_t skc_mag_refill; /* Magazine refill count */
|
|
spl_kmem_ctor_t skc_ctor; /* Constructor */
|
|
spl_kmem_dtor_t skc_dtor; /* Destructor */
|
|
void *skc_private; /* Private data */
|
|
void *skc_vmp; /* Unused */
|
|
struct kmem_cache *skc_linux_cache; /* Linux slab cache if used */
|
|
unsigned long skc_flags; /* Flags */
|
|
uint32_t skc_obj_size; /* Object size */
|
|
uint32_t skc_obj_align; /* Object alignment */
|
|
uint32_t skc_slab_objs; /* Objects per slab */
|
|
uint32_t skc_slab_size; /* Slab size */
|
|
atomic_t skc_ref; /* Ref count callers */
|
|
taskqid_t skc_taskqid; /* Slab reclaim task */
|
|
struct list_head skc_list; /* List of caches linkage */
|
|
struct list_head skc_complete_list; /* Completely alloc'ed */
|
|
struct list_head skc_partial_list; /* Partially alloc'ed */
|
|
struct rb_root skc_emergency_tree; /* Min sized objects */
|
|
spinlock_t skc_lock; /* Cache lock */
|
|
spl_wait_queue_head_t skc_waitq; /* Allocation waiters */
|
|
uint64_t skc_slab_fail; /* Slab alloc failures */
|
|
uint64_t skc_slab_create; /* Slab creates */
|
|
uint64_t skc_slab_destroy; /* Slab destroys */
|
|
uint64_t skc_slab_total; /* Slab total current */
|
|
uint64_t skc_slab_alloc; /* Slab alloc current */
|
|
uint64_t skc_slab_max; /* Slab max historic */
|
|
uint64_t skc_obj_total; /* Obj total current */
|
|
uint64_t skc_obj_alloc; /* Obj alloc current */
|
|
struct percpu_counter skc_linux_alloc; /* Linux-backed Obj alloc */
|
|
uint64_t skc_obj_max; /* Obj max historic */
|
|
uint64_t skc_obj_deadlock; /* Obj emergency deadlocks */
|
|
uint64_t skc_obj_emergency; /* Obj emergency current */
|
|
uint64_t skc_obj_emergency_max; /* Obj emergency max */
|
|
} spl_kmem_cache_t;
|
|
#define kmem_cache_t spl_kmem_cache_t
|
|
|
|
extern spl_kmem_cache_t *spl_kmem_cache_create(const char *name, size_t size,
|
|
size_t align, spl_kmem_ctor_t ctor, spl_kmem_dtor_t dtor,
|
|
void *reclaim, void *priv, void *vmp, int flags);
|
|
extern void spl_kmem_cache_set_move(spl_kmem_cache_t *,
|
|
kmem_cbrc_t (*)(void *, void *, size_t, void *));
|
|
extern void spl_kmem_cache_destroy(spl_kmem_cache_t *skc);
|
|
extern void *spl_kmem_cache_alloc(spl_kmem_cache_t *skc, int flags);
|
|
extern void spl_kmem_cache_free(spl_kmem_cache_t *skc, void *obj);
|
|
extern void spl_kmem_cache_set_allocflags(spl_kmem_cache_t *skc, gfp_t flags);
|
|
extern void spl_kmem_cache_reap_now(spl_kmem_cache_t *skc);
|
|
extern void spl_kmem_reap(void);
|
|
extern uint64_t spl_kmem_cache_inuse(kmem_cache_t *cache);
|
|
extern uint64_t spl_kmem_cache_entry_size(kmem_cache_t *cache);
|
|
|
|
#define kmem_cache_create(name, size, align, ctor, dtor, rclm, priv, vmp, fl) \
|
|
spl_kmem_cache_create(name, size, align, ctor, dtor, rclm, priv, vmp, fl)
|
|
#define kmem_cache_set_move(skc, move) spl_kmem_cache_set_move(skc, move)
|
|
#define kmem_cache_destroy(skc) spl_kmem_cache_destroy(skc)
|
|
#define kmem_cache_alloc(skc, flags) spl_kmem_cache_alloc(skc, flags)
|
|
#define kmem_cache_free(skc, obj) spl_kmem_cache_free(skc, obj)
|
|
#define kmem_cache_reap_now(skc) spl_kmem_cache_reap_now(skc)
|
|
#define kmem_reap() spl_kmem_reap()
|
|
|
|
/*
|
|
* The following functions are only available for internal use.
|
|
*/
|
|
extern int spl_kmem_cache_init(void);
|
|
extern void spl_kmem_cache_fini(void);
|
|
|
|
#endif /* _SPL_KMEM_CACHE_H */
|