summaryrefslogtreecommitdiffstats
path: root/sys/dev/raidframe/rf_memchunk.c
diff options
context:
space:
mode:
authorscottl <scottl@FreeBSD.org>2002-10-20 08:17:39 +0000
committerscottl <scottl@FreeBSD.org>2002-10-20 08:17:39 +0000
commit710948de69ddeae56bda663219319f6d859aea1f (patch)
tree71c65823ba2e8591de708d5cb2e990a75135ee11 /sys/dev/raidframe/rf_memchunk.c
parent63bd46464d6d4587c20c1ca62fb6a6e3be132db9 (diff)
downloadFreeBSD-src-710948de69ddeae56bda663219319f6d859aea1f.zip
FreeBSD-src-710948de69ddeae56bda663219319f6d859aea1f.tar.gz
After much delay and anticipation, welcome RAIDFrame into the FreeBSD
world. This should be considered highly experimental. Approved-by: re
Diffstat (limited to 'sys/dev/raidframe/rf_memchunk.c')
-rw-r--r--sys/dev/raidframe/rf_memchunk.c211
1 files changed, 211 insertions, 0 deletions
diff --git a/sys/dev/raidframe/rf_memchunk.c b/sys/dev/raidframe/rf_memchunk.c
new file mode 100644
index 0000000..b4aae57
--- /dev/null
+++ b/sys/dev/raidframe/rf_memchunk.c
@@ -0,0 +1,211 @@
+/* $FreeBSD$ */
+/* $NetBSD: rf_memchunk.c,v 1.4 1999/08/13 03:41:56 oster Exp $ */
+/*
+ * Copyright (c) 1995 Carnegie-Mellon University.
+ * All rights reserved.
+ *
+ * Author: Mark Holland
+ *
+ * Permission to use, copy, modify and distribute this software and
+ * its documentation is hereby granted, provided that both the copyright
+ * notice and this permission notice appear in all copies of the
+ * software, derivative works or modified versions, and any portions
+ * thereof, and that both notices appear in supporting documentation.
+ *
+ * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
+ * CONDITION. CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
+ * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
+ *
+ * Carnegie Mellon requests users of this software to return to
+ *
+ * Software Distribution Coordinator or Software.Distribution@CS.CMU.EDU
+ * School of Computer Science
+ * Carnegie Mellon University
+ * Pittsburgh PA 15213-3890
+ *
+ * any improvements or extensions that they make and grant Carnegie the
+ * rights to redistribute these changes.
+ */
+
+/*********************************************************************************
+ * rf_memchunk.c
+ *
+ * experimental code. I've found that the malloc and free calls in the DAG
+ * creation code are very expensive. Since for any given workload the DAGs
+ * created for different accesses are likely to be similar to each other, the
+ * amount of memory used for any given DAG data structure is likely to be one
+ * of a small number of values. For example, in UNIX, all reads and writes will
+ * be less than 8k and will not span stripe unit boundaries. Thus in the absence
+ * of failure, the only DAGs that will ever get created are single-node reads
+ * and single-stripe-unit atomic read-modify-writes. So, I'm very likely to
+ * be continually asking for chunks of memory equal to the sizes of these two
+ * DAGs.
+ *
+ * This leads to the idea of holding on to these chunks of memory when the DAG is
+ * freed and then, when a new DAG is created, trying to find such a chunk before
+ * calling malloc.
+ *
+ * the "chunk list" is a list of lists. Each header node contains a size value
+ * and a pointer to a list of chunk descriptors, each of which holds a pointer
+ * to a chunk of memory of the indicated size.
+ *
+ * There is currently no way to purge memory out of the chunk list. My
+ * initial thought on this is to have a low-priority thread that wakes up every
+ * 1 or 2 seconds, purges all the chunks with low reuse counts, and sets all
+ * the reuse counts to zero.
+ *
+ * This whole idea may be bad, since malloc may be able to do this more efficiently.
+ * It's worth a try, though, and it can be turned off by setting useMemChunks to 0.
+ *
+ ********************************************************************************/
+
+#include <dev/raidframe/rf_types.h>
+#include <dev/raidframe/rf_threadstuff.h>
+#include <dev/raidframe/rf_debugMem.h>
+#include <dev/raidframe/rf_memchunk.h>
+#include <dev/raidframe/rf_general.h>
+#include <dev/raidframe/rf_options.h>
+#include <dev/raidframe/rf_shutdown.h>
+
+typedef struct RF_ChunkHdr_s RF_ChunkHdr_t;
+struct RF_ChunkHdr_s {
+ int size;
+ RF_ChunkDesc_t *list;
+ RF_ChunkHdr_t *next;
+};
+
+static RF_ChunkHdr_t *chunklist, *chunk_hdr_free_list;
+static RF_ChunkDesc_t *chunk_desc_free_list;
+RF_DECLARE_STATIC_MUTEX(chunkmutex)
+ static void rf_ShutdownMemChunk(void *);
+ static RF_ChunkDesc_t *NewMemChunk(int, char *);
+
+
+ static void rf_ShutdownMemChunk(ignored)
+ void *ignored;
+{
+ RF_ChunkDesc_t *pt, *p;
+ RF_ChunkHdr_t *hdr, *ht;
+
+ if (rf_memChunkDebug)
+ printf("Chunklist:\n");
+ for (hdr = chunklist; hdr;) {
+ for (p = hdr->list; p;) {
+ if (rf_memChunkDebug)
+ printf("Size %d reuse count %d\n", p->size, p->reuse_count);
+ pt = p;
+ p = p->next;
+ RF_Free(pt->buf, pt->size);
+ RF_Free(pt, sizeof(*pt));
+ }
+ ht = hdr;
+ hdr = hdr->next;
+ RF_Free(ht, sizeof(*ht));
+ }
+
+ rf_mutex_destroy(&chunkmutex);
+}
+
+int
+rf_ConfigureMemChunk(listp)
+ RF_ShutdownList_t **listp;
+{
+ int rc;
+
+ chunklist = NULL;
+ chunk_hdr_free_list = NULL;
+ chunk_desc_free_list = NULL;
+ rc = rf_mutex_init(&chunkmutex, __FUNCTION__);
+ if (rc) {
+ RF_ERRORMSG3("Unable to init mutex file %s line %d rc=%d\n", __FILE__,
+ __LINE__, rc);
+ }
+ rc = rf_ShutdownCreate(listp, rf_ShutdownMemChunk, NULL);
+ if (rc) {
+ RF_ERRORMSG3("Unable to add to shutdown list file %s line %d rc=%d\n", __FILE__,
+ __LINE__, rc);
+ rf_mutex_destroy(&chunkmutex);
+ }
+ return (rc);
+}
+/* called to get a chunk descriptor for a newly-allocated chunk of memory
+ * MUTEX MUST BE LOCKED
+ *
+ * free list is not currently used
+ */
+static RF_ChunkDesc_t *
+NewMemChunk(size, buf)
+ int size;
+ char *buf;
+{
+ RF_ChunkDesc_t *p;
+
+ if (chunk_desc_free_list) {
+ p = chunk_desc_free_list;
+ chunk_desc_free_list = p->next;
+ } else
+ RF_Malloc(p, sizeof(RF_ChunkDesc_t), (RF_ChunkDesc_t *));
+ p->size = size;
+ p->buf = buf;
+ p->next = NULL;
+ p->reuse_count = 0;
+ return (p);
+}
+/* looks for a chunk of memory of acceptable size. If none, allocates one and returns
+ * a chunk descriptor for it, but does not install anything in the list. This is done
+ * when the chunk is released.
+ */
+RF_ChunkDesc_t *
+rf_GetMemChunk(size)
+ int size;
+{
+ RF_ChunkHdr_t *hdr = chunklist;
+ RF_ChunkDesc_t *p = NULL;
+ char *buf;
+
+ RF_LOCK_MUTEX(chunkmutex);
+ for (hdr = chunklist; hdr; hdr = hdr->next)
+ if (hdr->size >= size) {
+ p = hdr->list;
+ if (p) {
+ hdr->list = p->next;
+ p->next = NULL;
+ p->reuse_count++;
+ }
+ break;
+ }
+ if (!p) {
+ RF_Malloc(buf, size, (char *));
+ p = NewMemChunk(size, buf);
+ }
+ RF_UNLOCK_MUTEX(chunkmutex);
+ (void) bzero(p->buf, size);
+ return (p);
+}
+
+void
+rf_ReleaseMemChunk(chunk)
+ RF_ChunkDesc_t *chunk;
+{
+ RF_ChunkHdr_t *hdr, *ht = NULL, *new;
+
+ RF_LOCK_MUTEX(chunkmutex);
+ for (hdr = chunklist; hdr && hdr->size < chunk->size; ht = hdr, hdr = hdr->next);
+ if (hdr && hdr->size == chunk->size) {
+ chunk->next = hdr->list;
+ hdr->list = chunk;
+ } else {
+ RF_Malloc(new, sizeof(RF_ChunkHdr_t), (RF_ChunkHdr_t *));
+ new->size = chunk->size;
+ new->list = chunk;
+ chunk->next = NULL;
+ if (ht) {
+ new->next = ht->next;
+ ht->next = new;
+ } else {
+ new->next = hdr;
+ chunklist = new;
+ }
+ }
+ RF_UNLOCK_MUTEX(chunkmutex);
+}
OpenPOWER on IntegriCloud