From b0bdd900dbf589ed55c2f88b99b495351ce826d2 Mon Sep 17 00:00:00 2001 From: CentOS Sources Date: Tue, 9 Aug 2022 04:11:22 +0000 Subject: [PATCH] Auto sync2gitlab import of libtirpc-1.1.4-8.el8.src.rpm --- libtirpc-1.1.4-multithr-cleanup.patch | 1230 +++++++++++++++++++++++++ libtirpc-1.1.4-v2proto-mech.patch | 88 ++ libtirpc.spec | 10 +- 3 files changed, 1327 insertions(+), 1 deletion(-) create mode 100644 libtirpc-1.1.4-multithr-cleanup.patch create mode 100644 libtirpc-1.1.4-v2proto-mech.patch diff --git a/libtirpc-1.1.4-multithr-cleanup.patch b/libtirpc-1.1.4-multithr-cleanup.patch new file mode 100644 index 0000000..d4fc2d6 --- /dev/null +++ b/libtirpc-1.1.4-multithr-cleanup.patch @@ -0,0 +1,1230 @@ +diff --git a/src/bindresvport.c b/src/bindresvport.c +index ef9b345..5c0ddcf 100644 +--- a/src/bindresvport.c ++++ b/src/bindresvport.c +@@ -164,10 +164,11 @@ bindresvport_sa(sd, sa) + int endport = ENDPORT; + int i; + ++ mutex_lock(&port_lock); ++ + if (!blacklist_read) + load_blacklist(); + +- mutex_lock(&port_lock); + nports = ENDPORT - startport + 1; + + if (sa == NULL) { +diff --git a/src/clnt_dg.c b/src/clnt_dg.c +index d8890b5..166af63 100644 +--- a/src/clnt_dg.c ++++ b/src/clnt_dg.c +@@ -53,6 +53,7 @@ + #include + #include + #include "rpc_com.h" ++#include "clnt_fd_locks.h" + + #ifdef IP_RECVERR + #include +@@ -81,24 +82,29 @@ static void clnt_dg_destroy(CLIENT *); + * This machinery implements per-fd locks for MT-safety. It is not + * sufficient to do per-CLIENT handle locks for MT-safety because a + * user may create more than one CLIENT handle with the same fd behind +- * it. Therfore, we allocate an array of flags (dg_fd_locks), protected +- * by the clnt_fd_lock mutex, and an array (dg_cv) of condition variables +- * similarly protected. Dg_fd_lock[fd] == 1 => a call is activte on some +- * CLIENT handle created for that fd. ++ * it. ++ * ++ * We keep track of a list of per-fd locks, protected by the clnt_fd_lock ++ * mutex. Each per-fd lock consists of a predicate indicating whether is ++ * active or not: fd_lock->active == TRUE => a call is active on some ++ * CLIENT handle created for that fd. Each fd predicate is guarded by a ++ * condition variable so that the global mutex can be unlocked while ++ * waiting for the predicate to change. ++ * + * The current implementation holds locks across the entire RPC and reply, + * including retransmissions. Yes, this is silly, and as soon as this + * code is proven to work, this should be the first thing fixed. One step + * at a time. + */ +-static int *dg_fd_locks; ++static fd_locks_t *dg_fd_locks; + extern mutex_t clnt_fd_lock; +-static cond_t *dg_cv; +-#define release_fd_lock(fd, mask) { \ ++#define release_fd_lock(fd_lock, mask) { \ + mutex_lock(&clnt_fd_lock); \ +- dg_fd_locks[fd] = 0; \ +- mutex_unlock(&clnt_fd_lock); \ ++ fd_lock->active = FALSE; \ ++ fd_lock->pending--; \ + thr_sigsetmask(SIG_SETMASK, &(mask), NULL); \ +- cond_signal(&dg_cv[fd]); \ ++ cond_signal(&fd_lock->cv); \ ++ mutex_unlock(&clnt_fd_lock); \ + } + + static const char mem_err_clnt_dg[] = "clnt_dg_create: out of memory"; +@@ -110,6 +116,7 @@ static const char mem_err_clnt_dg[] = "clnt_dg_create: out of memory"; + */ + struct cu_data { + int cu_fd; /* connections fd */ ++ fd_lock_t *cu_fd_lock; + bool_t cu_closeit; /* opened by library */ + struct sockaddr_storage cu_raddr; /* remote address */ + int cu_rlen; +@@ -158,46 +165,22 @@ clnt_dg_create(fd, svcaddr, program, version, sendsz, recvsz) + sigset_t newmask; + struct __rpc_sockinfo si; + int one = 1; ++ fd_lock_t *fd_lock; + + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- if (dg_fd_locks == (int *) NULL) { +- size_t cv_allocsz, fd_allocsz; +- unsigned int dtbsize = __rpc_dtbsize(); +- +- if ( (size_t) dtbsize > SIZE_MAX/sizeof(cond_t)) { ++ if (dg_fd_locks == (fd_locks_t *) NULL) { ++ dg_fd_locks = fd_locks_init(); ++ if (dg_fd_locks == (fd_locks_t *) NULL) { + mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- errno = EOVERFLOW; + goto err1; + } +- +- fd_allocsz = dtbsize * sizeof (int); +- dg_fd_locks = (int *) mem_alloc(fd_allocsz); +- if (dg_fd_locks == (int *) NULL) { +- mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- errno = ENOMEM; +- goto err1; +- } else +- memset(dg_fd_locks, '\0', fd_allocsz); +- +- cv_allocsz = dtbsize * sizeof (cond_t); +- dg_cv = (cond_t *) mem_alloc(cv_allocsz); +- if (dg_cv == (cond_t *) NULL) { +- mem_free(dg_fd_locks, fd_allocsz); +- dg_fd_locks = (int *) NULL; +- mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- errno = ENOMEM; +- goto err1; +- } else { +- int i; +- +- for (i = 0; i < dtbsize; i++) +- cond_init(&dg_cv[i], 0, (void *) 0); +- } ++ } ++ fd_lock = fd_lock_create(fd, dg_fd_locks); ++ if (fd_lock == (fd_lock_t *) NULL) { ++ mutex_unlock(&clnt_fd_lock); ++ goto err1; + } + + mutex_unlock(&clnt_fd_lock); +@@ -277,6 +260,7 @@ clnt_dg_create(fd, svcaddr, program, version, sendsz, recvsz) + */ + cu->cu_closeit = FALSE; + cu->cu_fd = fd; ++ cu->cu_fd_lock = fd_lock; + cl->cl_ops = clnt_dg_ops(); + cl->cl_private = (caddr_t)(void *)cu; + cl->cl_auth = authnone_create(); +@@ -322,17 +306,16 @@ clnt_dg_call(cl, proc, xargs, argsp, xresults, resultsp, utimeout) + sigset_t newmask; + socklen_t salen; + ssize_t recvlen = 0; +- int rpc_lock_value; + u_int32_t xid, inval, outval; + + outlen = 0; + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (dg_fd_locks[cu->cu_fd]) +- cond_wait(&dg_cv[cu->cu_fd], &clnt_fd_lock); +- rpc_lock_value = 1; +- dg_fd_locks[cu->cu_fd] = rpc_lock_value; ++ cu->cu_fd_lock->pending++; ++ while (cu->cu_fd_lock->active) ++ cond_wait(&cu->cu_fd_lock->cv, &clnt_fd_lock); ++ cu->cu_fd_lock->active = TRUE; + mutex_unlock(&clnt_fd_lock); + if (cu->cu_total.tv_usec == -1) { + timeout = utimeout; /* use supplied timeout */ +@@ -481,7 +464,7 @@ get_reply: + e = (struct sock_extended_err *) CMSG_DATA(cmsg); + cu->cu_error.re_errno = e->ee_errno; + mem_free(cbuf, (outlen + 256)); +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (cu->cu_error.re_status = RPC_CANTRECV); + } + mem_free(cbuf, (outlen + 256)); +@@ -561,7 +544,7 @@ get_reply: + + } + out: +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (cu->cu_error.re_status); + } + +@@ -590,13 +573,15 @@ clnt_dg_freeres(cl, xdr_res, res_ptr) + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (dg_fd_locks[cu->cu_fd]) +- cond_wait(&dg_cv[cu->cu_fd], &clnt_fd_lock); ++ cu->cu_fd_lock->pending++; ++ while (cu->cu_fd_lock->active) ++ cond_wait(&cu->cu_fd_lock->cv, &clnt_fd_lock); + xdrs->x_op = XDR_FREE; + dummy = (*xdr_res)(xdrs, res_ptr); +- mutex_unlock(&clnt_fd_lock); ++ cu->cu_fd_lock->pending--; + thr_sigsetmask(SIG_SETMASK, &mask, NULL); +- cond_signal(&dg_cv[cu->cu_fd]); ++ cond_signal(&cu->cu_fd_lock->cv); ++ mutex_unlock(&clnt_fd_lock); + return (dummy); + } + +@@ -617,36 +602,35 @@ clnt_dg_control(cl, request, info) + struct netbuf *addr; + sigset_t mask; + sigset_t newmask; +- int rpc_lock_value; + + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (dg_fd_locks[cu->cu_fd]) +- cond_wait(&dg_cv[cu->cu_fd], &clnt_fd_lock); +- rpc_lock_value = 1; +- dg_fd_locks[cu->cu_fd] = rpc_lock_value; ++ cu->cu_fd_lock->pending++; ++ while (cu->cu_fd_lock->active) ++ cond_wait(&cu->cu_fd_lock->cv, &clnt_fd_lock); ++ cu->cu_fd_lock->active = TRUE; + mutex_unlock(&clnt_fd_lock); + switch (request) { + case CLSET_FD_CLOSE: + cu->cu_closeit = TRUE; +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (TRUE); + case CLSET_FD_NCLOSE: + cu->cu_closeit = FALSE; +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (TRUE); + } + + /* for other requests which use info */ + if (info == NULL) { +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (FALSE); + } + switch (request) { + case CLSET_TIMEOUT: + if (time_not_ok((struct timeval *)info)) { +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (FALSE); + } + cu->cu_total = *(struct timeval *)info; +@@ -660,7 +644,7 @@ clnt_dg_control(cl, request, info) + break; + case CLSET_RETRY_TIMEOUT: + if (time_not_ok((struct timeval *)info)) { +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (FALSE); + } + cu->cu_wait = *(struct timeval *)info; +@@ -680,7 +664,7 @@ clnt_dg_control(cl, request, info) + case CLSET_SVC_ADDR: /* set to new address */ + addr = (struct netbuf *)info; + if (addr->len < sizeof cu->cu_raddr) { +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (FALSE); + } + (void) memcpy(&cu->cu_raddr, addr->buf, addr->len); +@@ -743,10 +727,10 @@ clnt_dg_control(cl, request, info) + cu->cu_connect = *(int *)info; + break; + default: +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (FALSE); + } +- release_fd_lock(cu->cu_fd, mask); ++ release_fd_lock(cu->cu_fd_lock, mask); + return (TRUE); + } + +@@ -756,14 +740,21 @@ clnt_dg_destroy(cl) + { + struct cu_data *cu = (struct cu_data *)cl->cl_private; + int cu_fd = cu->cu_fd; ++ fd_lock_t *cu_fd_lock = cu->cu_fd_lock; + sigset_t mask; + sigset_t newmask; + + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (dg_fd_locks[cu_fd]) +- cond_wait(&dg_cv[cu_fd], &clnt_fd_lock); ++ /* wait until all pending operations on client are completed. */ ++ while (cu_fd_lock->pending > 0) { ++ /* If a blocked operation can be awakened, then do it. */ ++ if (cu_fd_lock->active == FALSE) ++ cond_signal(&cu_fd_lock->cv); ++ /* keep waiting... */ ++ cond_wait(&cu_fd_lock->cv, &clnt_fd_lock); ++ } + if (cu->cu_closeit) + (void)close(cu_fd); + XDR_DESTROY(&(cu->cu_outxdrs)); +@@ -773,9 +764,10 @@ clnt_dg_destroy(cl) + if (cl->cl_tp && cl->cl_tp[0]) + mem_free(cl->cl_tp, strlen(cl->cl_tp) +1); + mem_free(cl, sizeof (CLIENT)); ++ cond_signal(&cu_fd_lock->cv); ++ fd_lock_destroy(cu_fd, cu_fd_lock, dg_fd_locks); + mutex_unlock(&clnt_fd_lock); + thr_sigsetmask(SIG_SETMASK, &mask, NULL); +- cond_signal(&dg_cv[cu_fd]); + } + + static struct clnt_ops * +diff --git a/src/clnt_fd_locks.h b/src/clnt_fd_locks.h +new file mode 100644 +index 0000000..6ba62cb +--- /dev/null ++++ b/src/clnt_fd_locks.h +@@ -0,0 +1,209 @@ ++/* ++ * debug.h -- debugging routines for libtirpc ++ * ++ * Copyright (c) 2020 SUSE LINUX GmbH, Nuernberg, Germany. ++ * ++ * Redistribution and use in source and binary forms, with or without ++ * modification, are permitted provided that the following conditions are met: ++ * - Redistributions of source code must retain the above copyright notice, ++ * this list of conditions and the following disclaimer. ++ * - Redistributions in binary form must reproduce the above copyright notice, ++ * this list of conditions and the following disclaimer in the documentation ++ * and/or other materials provided with the distribution. ++ * - Neither the name of Sun Microsystems, Inc. nor the names of its ++ * contributors may be used to endorse or promote products derived ++ * from this software without specific prior written permission. ++ * ++ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" ++ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE ++ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ++ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE ++ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR ++ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF ++ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS ++ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN ++ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ++ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE ++ * POSSIBILITY OF SUCH DAMAGE. ++ */ ++ ++#ifndef _CLNT_FD_LOCKS_H ++#define _CLNT_FD_LOCKS_H ++ ++#include ++#include ++#include ++#include ++ ++ ++/* ++ * This utility manages a list of per-fd locks for the clients. ++ * ++ * If MAX_FDLOCKS_PREALLOC is defined, a number of pre-fd locks will be ++ * pre-allocated. This number is the minimum of MAX_FDLOCKS_PREALLOC or ++ * the process soft limit of allowed fds. ++ */ ++#ifdef MAX_FDLOCKS_PREALLOC ++static unsigned int fd_locks_prealloc = 0; ++#endif ++ ++/* per-fd lock */ ++struct fd_lock_t { ++ bool_t active; ++ int pending; /* Number of pending operations on fd */ ++ cond_t cv; ++}; ++typedef struct fd_lock_t fd_lock_t; ++ ++ ++/* internal type to store per-fd locks in a list */ ++struct fd_lock_item_t { ++ /* fd_lock_t first so we can cast to fd_lock_item_t */ ++ fd_lock_t fd_lock; ++ int fd; ++ unsigned int refs; ++ TAILQ_ENTRY(fd_lock_item_t) link; ++}; ++typedef struct fd_lock_item_t fd_lock_item_t; ++#define to_fd_lock_item(fdlock_t_ptr) ((fd_lock_item_t*) fdlock_t_ptr) ++ ++ ++/* internal list of per-fd locks */ ++typedef TAILQ_HEAD(,fd_lock_item_t) fd_lock_list_t; ++ ++ ++#ifdef MAX_FDLOCKS_PREALLOC ++ ++/* With pre-allocation, keep track of both an array and a list */ ++struct fd_locks_t { ++ fd_lock_list_t fd_lock_list; ++ fd_lock_t *fd_lock_array; ++}; ++typedef struct fd_locks_t fd_locks_t; ++#define to_fd_lock_list(fd_locks_t_ptr) (&fd_locks_t_ptr->fd_lock_list) ++ ++#else ++ ++/* With no pre-allocation, just keep track of a list */ ++typedef fd_lock_list_t fd_locks_t; ++#define to_fd_lock_list(fd_locks_t_ptr) ((fd_lock_list_t *) fd_locks_t_ptr) ++ ++#endif ++ ++ ++/* allocate fd locks */ ++static inline ++fd_locks_t* fd_locks_init() { ++ fd_locks_t *fd_locks; ++ ++ fd_locks = (fd_locks_t *) mem_alloc(sizeof(fd_locks_t)); ++ if (fd_locks == (fd_locks_t *) NULL) { ++ errno = ENOMEM; ++ return (NULL); ++ } ++ TAILQ_INIT(to_fd_lock_list(fd_locks)); ++ ++#ifdef MAX_FDLOCKS_PREALLOC ++ size_t fd_lock_arraysz; ++ ++ if (fd_locks_prealloc == 0) { ++ unsigned int dtbsize = __rpc_dtbsize(); ++ if (0 < dtbsize && dtbsize < MAX_FDLOCKS_PREALLOC) ++ fd_locks_prealloc = dtbsize; ++ else ++ fd_locks_prealloc = MAX_FDLOCKS_PREALLOC; ++ } ++ ++ if ( (size_t) fd_locks_prealloc > SIZE_MAX/sizeof(fd_lock_t)) { ++ mem_free(fd_locks, sizeof (*fd_locks)); ++ errno = EOVERFLOW; ++ return (NULL); ++ } ++ ++ fd_lock_arraysz = fd_locks_prealloc * sizeof (fd_lock_t); ++ fd_locks->fd_lock_array = (fd_lock_t *) mem_alloc(fd_lock_arraysz); ++ if (fd_locks->fd_lock_array == (fd_lock_t *) NULL) { ++ mem_free(fd_locks, sizeof (*fd_locks)); ++ errno = ENOMEM; ++ return (NULL); ++ } ++ else { ++ int i; ++ ++ for (i = 0; i < fd_locks_prealloc; i++) { ++ fd_locks->fd_lock_array[i].active = FALSE; ++ cond_init(&fd_locks->fd_lock_array[i].cv, 0, (void *) 0); ++ } ++ } ++#endif ++ ++ return fd_locks; ++} ++ ++/* de-allocate fd locks */ ++static inline ++void fd_locks_destroy(fd_locks_t *fd_locks) { ++#ifdef MAX_FDLOCKS_PREALLOC ++ fd_lock_t *array = fd_locks->fd_lock_array; ++ mem_free(array, fd_locks_prealloc * sizeof (fd_lock_t)); ++#endif ++ fd_lock_item_t *item; ++ fd_lock_list_t *list = to_fd_lock_list(fd_locks); ++ ++ TAILQ_FOREACH(item, list, link) { ++ cond_destroy(&item->fd_lock.cv); ++ mem_free(item, sizeof (*item)); ++ } ++ mem_free(fd_locks, sizeof (*fd_locks)); ++} ++ ++/* allocate per-fd lock */ ++static inline ++fd_lock_t* fd_lock_create(int fd, fd_locks_t *fd_locks) { ++#ifdef MAX_FDLOCKS_PREALLOC ++ if (fd < fd_locks_prealloc) { ++ return &fd_locks->fd_lock_array[fd]; ++ } ++#endif ++ fd_lock_item_t *item; ++ fd_lock_list_t *list = to_fd_lock_list(fd_locks); ++ ++ for (item = TAILQ_FIRST(list); ++ item != (fd_lock_item_t *) NULL && item->fd != fd; ++ item = TAILQ_NEXT(item, link)); ++ ++ if (item == (fd_lock_item_t *) NULL) { ++ item = (fd_lock_item_t *) mem_alloc(sizeof(fd_lock_item_t)); ++ if (item == (fd_lock_item_t *) NULL) { ++ errno = ENOMEM; ++ return (NULL); ++ } ++ item->fd = fd; ++ item->refs = 1; ++ item->fd_lock.active = FALSE; ++ item->fd_lock.pending = 0; ++ cond_init(&item->fd_lock.cv, 0, (void *) 0); ++ TAILQ_INSERT_HEAD(list, item, link); ++ } else { ++ item->refs++; ++ } ++ return &item->fd_lock; ++} ++ ++/* de-allocate per-fd lock */ ++static inline ++void fd_lock_destroy(int fd, fd_lock_t *fd_lock, fd_locks_t *fd_locks) { ++#ifdef MAX_FDLOCKS_PREALLOC ++ if (fd < fd_locks_prealloc) ++ return; ++#endif ++ fd_lock_item_t* item = to_fd_lock_item(fd_lock); ++ item->refs--; ++ if (item->refs <= 0) { ++ TAILQ_REMOVE(to_fd_lock_list(fd_locks), item, link); ++ cond_destroy(&item->fd_lock.cv); ++ mem_free(item, sizeof (*item)); ++ } ++} ++ ++#endif /* _CLNT_FD_LOCKS_H */ +diff --git a/src/clnt_vc.c b/src/clnt_vc.c +index 3ba55de..7fe3016 100644 +--- a/src/clnt_vc.c ++++ b/src/clnt_vc.c +@@ -67,6 +67,7 @@ + + #include + #include "rpc_com.h" ++#include "clnt_fd_locks.h" + + #ifdef HAVE_RPCSEC_GSS + #include +@@ -114,6 +115,7 @@ static int write_vc(void *, void *, int); + + struct ct_data { + int ct_fd; /* connection's fd */ ++ fd_lock_t *ct_fd_lock; + bool_t ct_closeit; /* close it on destroy */ + struct timeval ct_wait; /* wait interval in milliseconds */ + bool_t ct_waitset; /* wait set by clnt_control? */ +@@ -128,27 +130,33 @@ struct ct_data { + }; + + /* +- * This machinery implements per-fd locks for MT-safety. It is not +- * sufficient to do per-CLIENT handle locks for MT-safety because a +- * user may create more than one CLIENT handle with the same fd behind +- * it. Therfore, we allocate an array of flags (vc_fd_locks), protected +- * by the clnt_fd_lock mutex, and an array (vc_cv) of condition variables +- * similarly protected. Vc_fd_lock[fd] == 1 => a call is active on some +- * CLIENT handle created for that fd. +- * The current implementation holds locks across the entire RPC and reply. +- * Yes, this is silly, and as soon as this code is proven to work, this +- * should be the first thing fixed. One step at a time. ++ * This machinery implements per-fd locks for MT-safety. It is not ++ * sufficient to do per-CLIENT handle locks for MT-safety because a ++ * user may create more than one CLIENT handle with the same fd behind ++ * it. ++ * ++ * We keep track of a list of per-fd locks, protected by the clnt_fd_lock ++ * mutex. Each per-fd lock consists of a predicate indicating whether is ++ * active or not: fd_lock->active == TRUE => a call is active on some ++ * CLIENT handle created for that fd. Each fd predicate is guarded by a ++ * condition variable so that the global mutex can be unlocked while ++ * waiting for the predicate to change. ++ * ++ * The current implementation holds locks across the entire RPC and reply, ++ * including retransmissions. Yes, this is silly, and as soon as this ++ * code is proven to work, this should be the first thing fixed. One step ++ * at a time. + */ +-static int *vc_fd_locks; ++static fd_locks_t *vc_fd_locks; + extern pthread_mutex_t disrupt_lock; + extern mutex_t clnt_fd_lock; +-static cond_t *vc_cv; +-#define release_fd_lock(fd, mask) { \ ++#define release_fd_lock(fd_lock, mask) { \ + mutex_lock(&clnt_fd_lock); \ +- vc_fd_locks[fd] = 0; \ +- mutex_unlock(&clnt_fd_lock); \ ++ fd_lock->active = FALSE; \ ++ fd_lock->pending--; \ + thr_sigsetmask(SIG_SETMASK, &(mask), (sigset_t *) NULL); \ +- cond_signal(&vc_cv[fd]); \ ++ cond_signal(&fd_lock->cv); \ ++ mutex_unlock(&clnt_fd_lock); \ + } + + static const char clnt_vc_errstr[] = "%s : %s"; +@@ -185,6 +193,7 @@ clnt_vc_create(fd, raddr, prog, vers, sendsz, recvsz) + struct sockaddr_storage ss; + socklen_t slen; + struct __rpc_sockinfo si; ++ fd_lock_t *fd_lock; + + mutex_lock(&disrupt_lock); + if (disrupt == 0) +@@ -205,49 +214,26 @@ clnt_vc_create(fd, raddr, prog, vers, sendsz, recvsz) + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- if (vc_fd_locks == (int *) NULL) { +- size_t cv_allocsz, fd_allocsz; +- unsigned int dtbsize = __rpc_dtbsize(); +- struct rpc_createerr *ce = &get_rpc_createerr(); +- +- if ( (size_t) dtbsize > SIZE_MAX/sizeof(cond_t)) { ++ if (vc_fd_locks == (fd_locks_t *) NULL) { ++ vc_fd_locks = fd_locks_init(); ++ if (vc_fd_locks == (fd_locks_t *) NULL) { ++ struct rpc_createerr *ce; + mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- ce->cf_stat = RPC_SYSTEMERROR; +- ce->cf_error.re_errno = EOVERFLOW; +- goto err; +- } +- +- fd_allocsz = dtbsize * sizeof (int); +- vc_fd_locks = (int *) mem_alloc(fd_allocsz); +- if (vc_fd_locks == (int *) NULL) { +- mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- ce->cf_stat = RPC_SYSTEMERROR; +- ce->cf_error.re_errno = ENOMEM; +- goto err; +- } else +- memset(vc_fd_locks, '\0', fd_allocsz); +- +- assert(vc_cv == (cond_t *) NULL); +- cv_allocsz = dtbsize * sizeof (cond_t); +- vc_cv = (cond_t *) mem_alloc(cv_allocsz); +- if (vc_cv == (cond_t *) NULL) { +- mem_free(vc_fd_locks, fd_allocsz); +- vc_fd_locks = (int *) NULL; +- mutex_unlock(&clnt_fd_lock); +- thr_sigsetmask(SIG_SETMASK, &(mask), NULL); ++ ce = &get_rpc_createerr(); + ce->cf_stat = RPC_SYSTEMERROR; +- ce->cf_error.re_errno = ENOMEM; ++ ce->cf_error.re_errno = errno; + goto err; +- } else { +- int i; +- +- for (i = 0; i < dtbsize; i++) +- cond_init(&vc_cv[i], 0, (void *) 0); + } +- } else +- assert(vc_cv != (cond_t *) NULL); ++ } ++ fd_lock = fd_lock_create(fd, vc_fd_locks); ++ if (fd_lock == (fd_lock_t *) NULL) { ++ struct rpc_createerr *ce; ++ mutex_unlock(&clnt_fd_lock); ++ ce = &get_rpc_createerr(); ++ ce->cf_stat = RPC_SYSTEMERROR; ++ ce->cf_error.re_errno = errno; ++ goto err; ++ } + + /* + * Do not hold mutex during connect +@@ -283,6 +269,7 @@ clnt_vc_create(fd, raddr, prog, vers, sendsz, recvsz) + * Set up private data struct + */ + ct->ct_fd = fd; ++ ct->ct_fd_lock = fd_lock; + ct->ct_wait.tv_usec = 0; + ct->ct_waitset = FALSE; + ct->ct_addr.buf = malloc(raddr->maxlen); +@@ -362,17 +349,16 @@ clnt_vc_call(cl, proc, xdr_args, args_ptr, xdr_results, results_ptr, timeout) + bool_t shipnow; + int refreshes = 2; + sigset_t mask, newmask; +- int rpc_lock_value; + + assert(cl != NULL); + + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (vc_fd_locks[ct->ct_fd]) +- cond_wait(&vc_cv[ct->ct_fd], &clnt_fd_lock); +- rpc_lock_value = 1; +- vc_fd_locks[ct->ct_fd] = rpc_lock_value; ++ ct->ct_fd_lock->pending++; ++ while (ct->ct_fd_lock->active) ++ cond_wait(&ct->ct_fd_lock->cv, &clnt_fd_lock); ++ ct->ct_fd_lock->active = TRUE; + mutex_unlock(&clnt_fd_lock); + if (!ct->ct_waitset) { + /* If time is not within limits, we ignore it. */ +@@ -401,22 +387,22 @@ call_again: + if (ct->ct_error.re_status == RPC_SUCCESS) + ct->ct_error.re_status = RPC_CANTENCODEARGS; + (void)xdrrec_endofrecord(xdrs, TRUE); +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (ct->ct_error.re_status); + } + if (! xdrrec_endofrecord(xdrs, shipnow)) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (ct->ct_error.re_status = RPC_CANTSEND); + } + if (! shipnow) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (RPC_SUCCESS); + } + /* + * Hack to provide rpc-based message passing + */ + if (timeout.tv_sec == 0 && timeout.tv_usec == 0) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return(ct->ct_error.re_status = RPC_TIMEDOUT); + } + +@@ -430,14 +416,14 @@ call_again: + reply_msg.acpted_rply.ar_results.where = NULL; + reply_msg.acpted_rply.ar_results.proc = (xdrproc_t)xdr_void; + if (! xdrrec_skiprecord(xdrs)) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (ct->ct_error.re_status); + } + /* now decode and validate the response header */ + if (! xdr_replymsg(xdrs, &reply_msg)) { + if (ct->ct_error.re_status == RPC_SUCCESS) + continue; +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (ct->ct_error.re_status); + } + if (reply_msg.rm_xid == x_id) +@@ -470,7 +456,7 @@ call_again: + if (refreshes-- && AUTH_REFRESH(cl->cl_auth, &reply_msg)) + goto call_again; + } /* end of unsuccessful completion */ +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (ct->ct_error.re_status); + } + +@@ -508,13 +494,15 @@ clnt_vc_freeres(cl, xdr_res, res_ptr) + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (vc_fd_locks[ct->ct_fd]) +- cond_wait(&vc_cv[ct->ct_fd], &clnt_fd_lock); ++ ct->ct_fd_lock->pending++; ++ while (ct->ct_fd_lock->active) ++ cond_wait(&ct->ct_fd_lock->cv, &clnt_fd_lock); + xdrs->x_op = XDR_FREE; + dummy = (*xdr_res)(xdrs, res_ptr); +- mutex_unlock(&clnt_fd_lock); ++ ct->ct_fd_lock->pending--; + thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- cond_signal(&vc_cv[ct->ct_fd]); ++ cond_signal(&ct->ct_fd_lock->cv); ++ mutex_unlock(&clnt_fd_lock); + + return dummy; + } +@@ -536,7 +524,6 @@ clnt_vc_control(cl, request, info) + void *infop = info; + sigset_t mask; + sigset_t newmask; +- int rpc_lock_value; + u_int32_t tmp; + u_int32_t ltmp; + +@@ -547,20 +534,20 @@ clnt_vc_control(cl, request, info) + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (vc_fd_locks[ct->ct_fd]) +- cond_wait(&vc_cv[ct->ct_fd], &clnt_fd_lock); +- rpc_lock_value = 1; +- vc_fd_locks[ct->ct_fd] = rpc_lock_value; ++ ct->ct_fd_lock->pending++; ++ while (ct->ct_fd_lock->active) ++ cond_wait(&ct->ct_fd_lock->cv, &clnt_fd_lock); ++ ct->ct_fd_lock->active = TRUE; + mutex_unlock(&clnt_fd_lock); + + switch (request) { + case CLSET_FD_CLOSE: + ct->ct_closeit = TRUE; +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (TRUE); + case CLSET_FD_NCLOSE: + ct->ct_closeit = FALSE; +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (TRUE); + default: + break; +@@ -568,13 +555,13 @@ clnt_vc_control(cl, request, info) + + /* for other requests which use info */ + if (info == NULL) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (FALSE); + } + switch (request) { + case CLSET_TIMEOUT: + if (time_not_ok((struct timeval *)info)) { +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (FALSE); + } + ct->ct_wait = *(struct timeval *)infop; +@@ -594,7 +581,7 @@ clnt_vc_control(cl, request, info) + *(struct netbuf *)info = ct->ct_addr; + break; + case CLSET_SVC_ADDR: /* set to new address */ +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (FALSE); + case CLGET_XID: + /* +@@ -648,10 +635,10 @@ clnt_vc_control(cl, request, info) + break; + + default: +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (FALSE); + } +- release_fd_lock(ct->ct_fd, mask); ++ release_fd_lock(ct->ct_fd_lock, mask); + return (TRUE); + } + +@@ -660,20 +647,24 @@ static void + clnt_vc_destroy(cl) + CLIENT *cl; + { ++ assert(cl != NULL); + struct ct_data *ct = (struct ct_data *) cl->cl_private; + int ct_fd = ct->ct_fd; ++ fd_lock_t *ct_fd_lock = ct->ct_fd_lock; + sigset_t mask; + sigset_t newmask; + +- assert(cl != NULL); +- +- ct = (struct ct_data *) cl->cl_private; +- + sigfillset(&newmask); + thr_sigsetmask(SIG_SETMASK, &newmask, &mask); + mutex_lock(&clnt_fd_lock); +- while (vc_fd_locks[ct_fd]) +- cond_wait(&vc_cv[ct_fd], &clnt_fd_lock); ++ /* wait until all pending operations on client are completed. */ ++ while (ct_fd_lock->pending > 0) { ++ /* If a blocked operation can be awakened, then do it. */ ++ if (ct_fd_lock->active == FALSE) ++ cond_signal(&ct_fd_lock->cv); ++ /* keep waiting... */ ++ cond_wait(&ct_fd_lock->cv, &clnt_fd_lock); ++ } + if (ct->ct_closeit && ct->ct_fd != -1) { + (void)close(ct->ct_fd); + } +@@ -686,9 +677,10 @@ clnt_vc_destroy(cl) + if (cl->cl_tp && cl->cl_tp[0]) + mem_free(cl->cl_tp, strlen(cl->cl_tp) +1); + mem_free(cl, sizeof(CLIENT)); ++ cond_signal(&ct_fd_lock->cv); ++ fd_lock_destroy(ct_fd, ct_fd_lock, vc_fd_locks); + mutex_unlock(&clnt_fd_lock); + thr_sigsetmask(SIG_SETMASK, &(mask), NULL); +- cond_signal(&vc_cv[ct_fd]); + } + + /* +diff --git a/src/getpublickey.c b/src/getpublickey.c +index 8cf4dc2..be37a24 100644 +--- a/src/getpublickey.c ++++ b/src/getpublickey.c +@@ -74,7 +74,7 @@ __getpublickey_real(netname, publickey) + return (0); + } + *p = '\0'; +- (void) strncpy(publickey, lookup, HEXKEYBYTES); ++ memcpy(publickey, lookup, HEXKEYBYTES); + publickey[HEXKEYBYTES] = '\0'; + return (1); + } +diff --git a/src/mt_misc.c b/src/mt_misc.c +index 5a49b78..3a2bc51 100644 +--- a/src/mt_misc.c ++++ b/src/mt_misc.c +@@ -13,7 +13,7 @@ pthread_rwlock_t svc_lock = PTHREAD_RWLOCK_INITIALIZER; + pthread_rwlock_t svc_fd_lock = PTHREAD_RWLOCK_INITIALIZER; + + /* protects the RPCBIND address cache */ +-pthread_rwlock_t rpcbaddr_cache_lock = PTHREAD_RWLOCK_INITIALIZER; ++pthread_mutex_t rpcbaddr_cache_lock = PTHREAD_MUTEX_INITIALIZER; + + /* protects authdes cache (svcauth_des.c) */ + pthread_mutex_t authdes_lock = PTHREAD_MUTEX_INITIALIZER; +diff --git a/src/rpcb_clnt.c b/src/rpcb_clnt.c +index a5923cb..65ff43e 100644 +--- a/src/rpcb_clnt.c ++++ b/src/rpcb_clnt.c +@@ -85,7 +85,7 @@ static int cachesize; + + extern int __rpc_lowvers; + +-static struct address_cache *check_cache(const char *, const char *); ++static struct address_cache *copy_of_cached(const char *, char *); + static void delete_cache(struct netbuf *); + static void add_cache(const char *, const char *, struct netbuf *, char *); + static CLIENT *getclnthandle(const char *, const struct netconfig *, char **); +@@ -94,6 +94,82 @@ static CLIENT *local_rpcb(void); + static struct netbuf *got_entry(rpcb_entry_list_ptr, const struct netconfig *); + #endif + ++/* ++ * Destroys a cached address entry structure. ++ * ++ */ ++static void ++destroy_addr(addr) ++ struct address_cache *addr; ++{ ++ if (addr == NULL) ++ return; ++ if(addr->ac_host != NULL) ++ free(addr->ac_host); ++ if(addr->ac_netid != NULL) ++ free(addr->ac_netid); ++ if(addr->ac_uaddr != NULL) ++ free(addr->ac_uaddr); ++ if(addr->ac_taddr != NULL) { ++ if(addr->ac_taddr->buf != NULL) ++ free(addr->ac_taddr->buf); ++ } ++ free(addr); ++} ++ ++/* ++ * Creates an unlinked copy of an address cache entry. If the argument is NULL ++ * or the new entry cannot be allocated then NULL is returned. ++ */ ++static struct address_cache * ++copy_addr(addr) ++ const struct address_cache *addr; ++{ ++ struct address_cache *copy; ++ ++ if (addr == NULL) ++ return (NULL); ++ ++ copy = calloc(1, sizeof(*addr)); ++ if (copy == NULL) ++ return (NULL); ++ ++ if (addr->ac_host != NULL) { ++ copy->ac_host = strdup(addr->ac_host); ++ if (copy->ac_host == NULL) ++ goto err; ++ } ++ if (addr->ac_netid != NULL) { ++ copy->ac_netid = strdup(addr->ac_netid); ++ if (copy->ac_netid == NULL) ++ goto err; ++ } ++ if (addr->ac_uaddr != NULL) { ++ copy->ac_uaddr = strdup(addr->ac_uaddr); ++ if (copy->ac_uaddr == NULL) ++ goto err; ++ } ++ ++ if (addr->ac_taddr == NULL) ++ return (copy); ++ ++ copy->ac_taddr = calloc(1, sizeof(*addr->ac_taddr)); ++ if (copy->ac_taddr == NULL) ++ goto err; ++ ++ memcpy(copy->ac_taddr, addr->ac_taddr, sizeof(*addr->ac_taddr)); ++ copy->ac_taddr->buf = malloc(addr->ac_taddr->len); ++ if (copy->ac_taddr->buf == NULL) ++ goto err; ++ ++ memcpy(copy->ac_taddr->buf, addr->ac_taddr->buf, addr->ac_taddr->len); ++ return (copy); ++ ++err: ++ destroy_addr(copy); ++ return (NULL); ++} ++ + /* + * This routine adjusts the timeout used for calls to the remote rpcbind. + * Also, this routine can be used to set the use of portmapper version 2 +@@ -125,17 +201,18 @@ __rpc_control(request, info) + } + + /* +- * It might seem that a reader/writer lock would be more reasonable here. +- * However because getclnthandle(), the only user of the cache functions, +- * may do a delete_cache() operation if a check_cache() fails to return an +- * address useful to clnt_tli_create(), we may as well use a mutex. +- */ +-/* +- * As it turns out, if the cache lock is *not* a reader/writer lock, we will +- * block all clnt_create's if we are trying to connect to a host that's down, +- * since the lock will be held all during that time. ++ * Protect against concurrent access to the address cache and modifications ++ * (esp. deletions) of cache entries. ++ * ++ * Previously a bidirectional R/W lock was used. However, R/W locking is ++ * dangerous as it allows concurrent modification (e.g. deletion with write ++ * lock) at the same time as the deleted element is accessed via check_cache() ++ * and a read lock). We absolutely need a single mutex for all access to ++ * prevent cache corruption. If the mutexing is restricted to only the ++ * relevant code sections, deadlocking should be avoided even with recursed ++ * client creation. + */ +-extern rwlock_t rpcbaddr_cache_lock; ++extern pthread_mutex_t rpcbaddr_cache_lock; + + /* + * The routines check_cache(), add_cache(), delete_cache() manage the +@@ -143,49 +220,52 @@ extern rwlock_t rpcbaddr_cache_lock; + */ + + static struct address_cache * +-check_cache(host, netid) +- const char *host, *netid; ++copy_of_cached(host, netid) ++ const char *host; ++ char *netid; + { +- struct address_cache *cptr; +- +- /* READ LOCK HELD ON ENTRY: rpcbaddr_cache_lock */ ++ struct address_cache *cptr, *copy = NULL; + ++ mutex_lock(&rpcbaddr_cache_lock); + for (cptr = front; cptr != NULL; cptr = cptr->ac_next) { + if (!strcmp(cptr->ac_host, host) && + !strcmp(cptr->ac_netid, netid)) { + LIBTIRPC_DEBUG(3, ("check_cache: Found cache entry for %s: %s\n", + host, netid)); +- return (cptr); ++ copy = copy_addr(cptr); ++ break; + } + } +- return ((struct address_cache *) NULL); ++ mutex_unlock(&rpcbaddr_cache_lock); ++ return copy; + } + + static void + delete_cache(addr) + struct netbuf *addr; + { +- struct address_cache *cptr, *prevptr = NULL; ++ struct address_cache *cptr = NULL, *prevptr = NULL; ++ ++ /* LOCK HELD ON ENTRY: rpcbaddr_cache_lock */ ++ mutex_lock(&rpcbaddr_cache_lock); + +- /* WRITE LOCK HELD ON ENTRY: rpcbaddr_cache_lock */ + for (cptr = front; cptr != NULL; cptr = cptr->ac_next) { + if (!memcmp(cptr->ac_taddr->buf, addr->buf, addr->len)) { +- free(cptr->ac_host); +- free(cptr->ac_netid); +- free(cptr->ac_taddr->buf); +- free(cptr->ac_taddr); ++ /* Unlink from cache. We'll destroy it after releasing the mutex. */ + if (cptr->ac_uaddr) + free(cptr->ac_uaddr); + if (prevptr) + prevptr->ac_next = cptr->ac_next; + else + front = cptr->ac_next; +- free(cptr); + cachesize--; + break; + } + prevptr = cptr; + } ++ ++ mutex_unlock(&rpcbaddr_cache_lock); ++ destroy_addr(cptr); + } + + static void +@@ -217,7 +297,7 @@ add_cache(host, netid, taddr, uaddr) + + /* VARIABLES PROTECTED BY rpcbaddr_cache_lock: cptr */ + +- rwlock_wrlock(&rpcbaddr_cache_lock); ++ mutex_lock(&rpcbaddr_cache_lock); + if (cachesize < CACHESIZE) { + ad_cache->ac_next = front; + front = ad_cache; +@@ -250,7 +330,7 @@ add_cache(host, netid, taddr, uaddr) + } + free(cptr); + } +- rwlock_unlock(&rpcbaddr_cache_lock); ++ mutex_unlock(&rpcbaddr_cache_lock); + return; + + out_free: +@@ -261,6 +341,7 @@ out_free: + free(ad_cache); + } + ++ + /* + * This routine will return a client handle that is connected to the + * rpcbind. If targaddr is non-NULL, the "universal address" of the +@@ -275,11 +356,9 @@ getclnthandle(host, nconf, targaddr) + char **targaddr; + { + CLIENT *client; +- struct netbuf *addr, taddr; +- struct netbuf addr_to_delete; ++ struct netbuf taddr; + struct __rpc_sockinfo si; + struct addrinfo hints, *res, *tres; +- struct address_cache *ad_cache; + char *tmpaddr; + + if (nconf == NULL) { +@@ -294,47 +373,35 @@ getclnthandle(host, nconf, targaddr) + return NULL; + } + +-/* VARIABLES PROTECTED BY rpcbaddr_cache_lock: ad_cache */ ++ + + /* Get the address of the rpcbind. Check cache first */ + client = NULL; + if (targaddr) + *targaddr = NULL; +- addr_to_delete.len = 0; +- rwlock_rdlock(&rpcbaddr_cache_lock); +- ad_cache = NULL; +- +- if (host != NULL) +- ad_cache = check_cache(host, nconf->nc_netid); +- if (ad_cache != NULL) { +- addr = ad_cache->ac_taddr; +- client = clnt_tli_create(RPC_ANYFD, nconf, addr, +- (rpcprog_t)RPCBPROG, (rpcvers_t)RPCBVERS4, 0, 0); +- if (client != NULL) { +- if (targaddr && ad_cache->ac_uaddr) +- *targaddr = strdup(ad_cache->ac_uaddr); +- rwlock_unlock(&rpcbaddr_cache_lock); +- return (client); +- } +- addr_to_delete.len = addr->len; +- addr_to_delete.buf = (char *)malloc(addr->len); +- if (addr_to_delete.buf == NULL) { +- addr_to_delete.len = 0; +- } else { +- memcpy(addr_to_delete.buf, addr->buf, addr->len); ++ ++ if (host != NULL) { ++ struct address_cache *ad_cache; ++ ++ /* Get an MT-safe copy of the cached address (if any) */ ++ ad_cache = copy_of_cached(host, nconf->nc_netid); ++ if (ad_cache != NULL) { ++ client = clnt_tli_create(RPC_ANYFD, nconf, ad_cache->ac_taddr, ++ (rpcprog_t)RPCBPROG, (rpcvers_t)RPCBVERS4, 0, 0); ++ if (client != NULL) { ++ if (targaddr && ad_cache->ac_uaddr) { ++ *targaddr = ad_cache->ac_uaddr; ++ ad_cache->ac_uaddr = NULL; /* De-reference before destruction */ ++ } ++ destroy_addr(ad_cache); ++ return (client); ++ } ++ ++ delete_cache(ad_cache->ac_taddr); ++ destroy_addr(ad_cache); + } + } +- rwlock_unlock(&rpcbaddr_cache_lock); +- if (addr_to_delete.len != 0) { +- /* +- * Assume this may be due to cache data being +- * outdated +- */ +- rwlock_wrlock(&rpcbaddr_cache_lock); +- delete_cache(&addr_to_delete); +- rwlock_unlock(&rpcbaddr_cache_lock); +- free(addr_to_delete.buf); +- } ++ + if (!__rpc_nconf2sockinfo(nconf, &si)) { + rpc_createerr.cf_stat = RPC_UNKNOWNPROTO; + assert(client == NULL); +diff --git a/tirpc/reentrant.h b/tirpc/reentrant.h +index 5f5c96e..5bb581a 100644 +--- a/tirpc/reentrant.h ++++ b/tirpc/reentrant.h +@@ -57,6 +57,7 @@ + #define mutex_unlock(m) pthread_mutex_unlock(m) + + #define cond_init(c, a, p) pthread_cond_init(c, a) ++#define cond_destroy(c) pthread_cond_destroy(c) + #define cond_signal(m) pthread_cond_signal(m) + #define cond_broadcast(m) pthread_cond_broadcast(m) + #define cond_wait(c, m) pthread_cond_wait(c, m) diff --git a/libtirpc-1.1.4-v2proto-mech.patch b/libtirpc-1.1.4-v2proto-mech.patch new file mode 100644 index 0000000..024cf86 --- /dev/null +++ b/libtirpc-1.1.4-v2proto-mech.patch @@ -0,0 +1,88 @@ +diff -up libtirpc-1.1.4/man/rpcbind.3t.orig libtirpc-1.1.4/man/rpcbind.3t +--- libtirpc-1.1.4/man/rpcbind.3t.orig 2018-08-27 10:06:49.000000000 -0400 ++++ libtirpc-1.1.4/man/rpcbind.3t 2022-08-02 11:21:30.134642780 -0400 +@@ -187,6 +187,8 @@ in + .El + .Sh AVAILABILITY + These functions are part of libtirpc. ++.Sh ENVIRONMENT ++If RPCB_V2FIRST is defined, rpcbind protocol version tryout algorithm changes from v4,v2,v3 to v2,v4,v3. + .Sh SEE ALSO + .Xr rpc_clnt_calls 3 , + .Xr rpc_svc_calls 3 , +diff -up libtirpc-1.1.4/src/rpcb_clnt.c.orig libtirpc-1.1.4/src/rpcb_clnt.c +--- libtirpc-1.1.4/src/rpcb_clnt.c.orig 2022-08-02 11:20:42.795833195 -0400 ++++ libtirpc-1.1.4/src/rpcb_clnt.c 2022-08-02 11:21:30.135642797 -0400 +@@ -818,7 +818,8 @@ error: + * The algorithm used: If the transports is TCP or UDP, it first tries + * version 4 (srv4), then 3 and then fall back to version 2 (portmap). + * With this algorithm, we get performance as well as a plan for +- * obsoleting version 2. ++ * obsoleting version 2. This behaviour is reverted to old algorithm ++ * if RPCB_V2FIRST environment var is defined + * + * For all other transports, the algorithm remains as 4 and then 3. + * +@@ -839,6 +840,10 @@ __rpcb_findaddr_timed(program, version, + #ifdef NOTUSED + static bool_t check_rpcbind = TRUE; + #endif ++ ++#ifdef PORTMAP ++ static bool_t portmap_first = FALSE; ++#endif + CLIENT *client = NULL; + RPCB parms; + enum clnt_stat clnt_st; +@@ -895,8 +900,18 @@ __rpcb_findaddr_timed(program, version, + parms.r_addr = (char *) &nullstring[0]; + } + +- /* First try from start_vers(4) and then version 3 (RPCBVERS) */ ++ /* First try from start_vers(4) and then version 3 (RPCBVERS), except ++ * if env. var RPCB_V2FIRST is defined */ ++ ++#ifdef PORTMAP ++ if (getenv(V2FIRST)) { ++ portmap_first = TRUE; ++ LIBTIRPC_DEBUG(3, ("__rpcb_findaddr_timed: trying v2-port first\n")); ++ goto portmap; ++ } ++#endif + ++rpcbind: + CLNT_CONTROL(client, CLSET_RETRY_TIMEOUT, (char *) &rpcbrmttime); + for (vers = start_vers; vers >= RPCBVERS; vers--) { + /* Set the version */ +@@ -944,10 +959,17 @@ __rpcb_findaddr_timed(program, version, + } + + #ifdef PORTMAP /* Try version 2 for TCP or UDP */ ++ if (portmap_first) ++ goto error; /* we tried all versions if reached here */ ++portmap: + if (strcmp(nconf->nc_protofmly, NC_INET) == 0) { + address = __try_protocol_version_2(program, version, nconf, host, tp); +- if (address == NULL) +- goto error; ++ if (address == NULL) { ++ if (portmap_first) ++ goto rpcbind; ++ else ++ goto error; ++ } + } + #endif /* PORTMAP */ + +diff -up libtirpc-1.1.4/tirpc/rpc/pmap_prot.h.orig libtirpc-1.1.4/tirpc/rpc/pmap_prot.h +--- libtirpc-1.1.4/tirpc/rpc/pmap_prot.h.orig 2018-08-27 10:06:49.000000000 -0400 ++++ libtirpc-1.1.4/tirpc/rpc/pmap_prot.h 2022-08-02 11:21:30.135642797 -0400 +@@ -84,6 +84,8 @@ + #define PMAPPROC_DUMP ((u_long)4) + #define PMAPPROC_CALLIT ((u_long)5) + ++#define V2FIRST "RPCB_V2FIRST" ++ + struct pmap { + long unsigned pm_prog; + long unsigned pm_vers; diff --git a/libtirpc.spec b/libtirpc.spec index 8c687fe..a259868 100644 --- a/libtirpc.spec +++ b/libtirpc.spec @@ -2,7 +2,7 @@ Name: libtirpc Version: 1.1.4 -Release: 7%{?dist} +Release: 8%{?dist} Summary: Transport Independent RPC Library Group: System Environment/Libraries License: SISSL and BSD @@ -42,6 +42,10 @@ Patch006: libtirpc-1.1.4-dos-fix.patch # # bz 2042196 Patch007: libtirpc-1.1.4-dgcall-free.patch +# bz 2107650 +Patch008: libtirpc-1.1.4-v2proto-mech.patch +# bz 2112116 +Patch009: libtirpc-1.1.4-multithr-cleanup.patch BuildRequires: automake, autoconf, libtool, pkgconfig BuildRequires: krb5-devel @@ -162,6 +166,10 @@ mv %{buildroot}%{_mandir}/man3 %{buildroot}%{_mandir}/man3t %{_mandir}/*/* %changelog +* Wed Aug 3 2022 Steve Dickson 1.1.4-8 +- rpcb_clnt.c add mechanism to try v2 protocol first (bz 2107650) +- Multithreaded cleanup (bz 2112116) + * Tue May 31 2022 Steve Dickson 1.1.4-7 - clnt_dg_call: Fix use-after-free accessing the error number (bz 2042196)