bf251437e9
Many changes were necessary this time around: * QAPI was changed to avoid redundant has_* variables, see commit 44ea9d9be3 ("qapi: Start to elide redundant has_FOO in generated C") for details. This affected many QMP commands added by Proxmox too. * Pending querying for migration got split into two functions, one to estimate, one for exact value, see commit c8df4a7aef ("migration: Split save_live_pending() into state_pending_*") for details. Relevant for savevm-async and PBS dirty bitmap. * Some block (driver) functions got converted to coroutines, so the Proxmox block drivers needed to be adapted. * Alloc track auto-detaching during PBS live restore got broken by AioContext-related changes resulting in a deadlock. The current, hacky method was replaced by a simpler one. Stefan apparently ran into a problem with that when he wrote the driver, but there were improvements in the stream job code since then and I didn't manage to reproduce the issue. It's a separate patch "alloc-track: fix deadlock during drop" for now, you can find the details there. * Async snapshot-related changes: - The pending querying got adapted to the above-mentioned split and a patch is added to optimize it/make it more similar to what upstream code does. - Added initialization of the compression counters (for future-proofing). - It's necessary the hold the BQL (big QEMU lock = iothread mutex) during the setup phase, because block layer functions are used there and not doing so leads to racy, hard-to-debug crashes or hangs. It's necessary to change some upstream code too for this, a version of the patch "migration: for snapshots, hold the BQL during setup callbacks" is intended to be upstreamed. - Need to take the bdrv graph read lock before flushing. * hmp_info_balloon was moved to a different file. * Needed to include a new headers from time to time to still get the correct functions. Signed-off-by: Fiona Ebner <f.ebner@proxmox.com>
51 lines
1.7 KiB
Diff
51 lines
1.7 KiB
Diff
From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001
|
|
From: Wolfgang Bumiller <w.bumiller@proxmox.com>
|
|
Date: Mon, 6 Apr 2020 12:16:35 +0200
|
|
Subject: [PATCH] PVE: [Config] glusterfs: no default logfile if daemonized
|
|
|
|
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
|
|
---
|
|
block/gluster.c | 15 +++++++++++----
|
|
1 file changed, 11 insertions(+), 4 deletions(-)
|
|
|
|
diff --git a/block/gluster.c b/block/gluster.c
|
|
index 185a83e5e5..f11a40aa9e 100644
|
|
--- a/block/gluster.c
|
|
+++ b/block/gluster.c
|
|
@@ -43,7 +43,7 @@
|
|
#define GLUSTER_DEBUG_DEFAULT 4
|
|
#define GLUSTER_DEBUG_MAX 9
|
|
#define GLUSTER_OPT_LOGFILE "logfile"
|
|
-#define GLUSTER_LOGFILE_DEFAULT "-" /* handled in libgfapi as /dev/stderr */
|
|
+#define GLUSTER_LOGFILE_DEFAULT NULL
|
|
/*
|
|
* Several versions of GlusterFS (3.12? -> 6.0.1) fail when the transfer size
|
|
* is greater or equal to 1024 MiB, so we are limiting the transfer size to 512
|
|
@@ -425,6 +425,7 @@ static struct glfs *qemu_gluster_glfs_init(BlockdevOptionsGluster *gconf,
|
|
int old_errno;
|
|
SocketAddressList *server;
|
|
unsigned long long port;
|
|
+ const char *logfile;
|
|
|
|
glfs = glfs_find_preopened(gconf->volume);
|
|
if (glfs) {
|
|
@@ -467,9 +468,15 @@ static struct glfs *qemu_gluster_glfs_init(BlockdevOptionsGluster *gconf,
|
|
}
|
|
}
|
|
|
|
- ret = glfs_set_logging(glfs, gconf->logfile, gconf->debug);
|
|
- if (ret < 0) {
|
|
- goto out;
|
|
+ logfile = gconf->logfile;
|
|
+ if (!logfile && !is_daemonized()) {
|
|
+ logfile = "-";
|
|
+ }
|
|
+ if (logfile) {
|
|
+ ret = glfs_set_logging(glfs, logfile, gconf->debug);
|
|
+ if (ret < 0) {
|
|
+ goto out;
|
|
+ }
|
|
}
|
|
|
|
ret = glfs_init(glfs);
|