1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264
|
/* BLURB lgpl
Coda File System
Release 5
Copyright (c) 1987-1999 Carnegie Mellon University
Additional copyrights listed below
This code is distributed "AS IS" without warranty of any kind under
the terms of the GNU Library General Public Licence Version 2, as
shown in the file LICENSE. The technical and financial contributors to
Coda are listed in the file CREDITS.
Additional copyrights
none currently
#*/
#include <stdlib.h>
#include <stdio.h>
#include <unistd.h>
#include <stdlib.h>
#include "rds_private.h"
/************** NOTE: ***************/
/* we create our own transactions in the following routines, even
* though there is a tid in the interface. This might result in unreferenced
* objects if the thread which malloc'ed the object decides to abort. These
* routines have no way of knowing if such an event happens. Also, if the user
* free's then aborts, could have pointers into the free list objects.
*
* Interface has been changed so that if a user feels confident that his
* transactions are serialized, (the above problem won't arise), he can pass
* in a non-null tid ptr and it will be used. So if you aren't sure, make the
* tidptr is zero!
*/
/* There's another bug too: if tid1 frees an object, then tid2 mallocs it,
* then tid1 aborts, the object is referenced twice. To avoid this and the
* above two problems, rather than actually freeing an object during a
* transaction, keep a record of our intension to do so and only do it when
* the transaction commits. If it aborts, we don't want to free the object
* so we can just throw away the intension record. Thus we can keep that record
* in VM. The record only needs to indicate which transaction and which object.
* By keeping a list per transaction, each entry is simply an address.
* So rds_fake_free(addr, list) adds one's intension of freeing the object.
* rds_do_free(list, mode) actually frees the objects on the list, using
* mode as the end_transaction mode.
*/
int
rds_free(addr, tid, err)
char *addr;
rvm_tid_t *tid;
int *err;
{
free_block_t *bp = BLOCK_HDR(addr); /* find pointer to block header */
rvm_tid_t *atid;
rvm_return_t rvmret;
/* Make sure the heap has been initialized */
if (!HEAP_INIT) {
(*err) = EHEAP_INIT;
return -1;
}
RDS_LOG("rdstrace: Error!!! rds_free called\n");
/* Make sure that the pointer is word aligned */
if ((bp == NULL) || ((unsigned long)bp % sizeof(void *)) != 0) {
(*err) = EBAD_ARGS;
return -1;
}
/* Verify that the guards are intact */
if (bp->type == FREE_GUARD) { /* object was already freed */
(*err) = EFREED_TWICE;
return -1;
}
/* improper guard --> memory corruption */
if ((bp->type != ALLOC_GUARD) || (*BLOCK_END(bp) != END_GUARD)) {
(*err) = ECORRUPT;
return -1;
}
if (tid == NULL) { /* Use input tid if non-null */
atid = rvm_malloc_tid();
rvmret = rvm_begin_transaction(atid, restore);
if (rvmret != RVM_SUCCESS) {
(*err) = (int) rvmret;
rvm_free_tid(atid);
return -1;
}
} else
atid = tid;
*err = SUCCESS; /* Initialize the error value */
CRITICAL({
/* Update statistics */
rvmret = rvm_set_range(atid, &RDS_STATS, sizeof(rds_stats_t));
if (rvmret != RVM_SUCCESS) {
(*err) = (int)rvmret;
} else {
/* Set the lowguard to reflect that the block has been allocated. */
rvmret = rvm_set_range(atid, &(bp->type), sizeof(guard_t));
if (rvmret != RVM_SUCCESS) {
if (tid == NULL) {
rvm_abort_transaction(atid);
rvm_free_tid(atid);
}
(*err) = (int)rvmret;
} else {
bp->type = FREE_GUARD;
RDS_STATS.free++;
RDS_STATS.freebytes += bp->size * RDS_CHUNK_SIZE;
RDS_STATS.mallocbytes -= bp->size * RDS_CHUNK_SIZE;
/* try to merge with trailing free blocks */
merge_with_next_free(bp, atid, err);
/* Add the block to the approprite free list. */
if (*err == SUCCESS)
put_block(bp, atid, err); /* Error is picked up below... */
}
}
if ((*err != SUCCESS) && (tid == NULL)) {
rvm_abort_transaction(atid);
rvm_free_tid(atid);
} else if (tid == NULL) {
rvmret = rvm_end_transaction(atid, no_flush);
rvm_free_tid(atid);
if (rvmret != RVM_SUCCESS) {
(*err) = (int)rvmret;
}
}
});
return 0;
}
/* Assume only one thread can have an open tid at a time. Since we're only
* modifying per-tid structures, we don't have any concurency.
*/
int rds_fake_free(addr, list)
char *addr;
intentionList_t *list;
{
char **temp;
free_block_t *bp = BLOCK_HDR(addr); /* find pointer to block header */
/* Make sure the heap has been initialized */
if (!HEAP_INIT) {
return EHEAP_INIT;
}
/* Freeing a NULL ptr? */
if (!addr)
return SUCCESS;
/* Make sure that the pointer is word aligned */
if (((unsigned long)bp % sizeof(void *)) != 0)
return EBAD_ARGS;
/* Verify that the guards are intact */
if (bp->type == FREE_GUARD) /* object was already freed */
return EFREED_TWICE;
/* improper guard --> memory corruption */
if ((bp->type != ALLOC_GUARD) || (*BLOCK_END(bp) != END_GUARD))
return ECORRUPT;
/* If no intention list has been built for this tid yet, build one. */
if (list->table == NULL) {
list->size = STARTSIZE;
list->table = (char **)malloc(list->size);
list->count = 0;
} else if ((list->count * 4) == list->size) {
list->size *= 2;
temp = (char **)malloc(list->size);
BCOPY(list->table, temp, list->count * sizeof(char *));
free(list->table);
list->table = temp;
}
(list->table)[list->count++] = (char *)addr;
return(SUCCESS);
}
int rds_do_free(list, mode)
intentionList_t *list;
rvm_mode_t mode;
{
int i, err;
rvm_tid_t *tid = rvm_malloc_tid();
rvm_return_t rvmret;
rvmret = rvm_begin_transaction(tid, restore);
if (rvmret != RVM_SUCCESS) {
rvm_free_tid(tid);
return (int) rvmret;
}
RDS_LOG("rdstrace: start do_free\n");
err = SUCCESS; /* Initialize the error value */
CRITICAL({
/* Only need to set the range once. To be safe, doing in critical...*/
rvmret = rvm_set_range(tid, &RDS_STATS, sizeof(rds_stats_t));
if (rvmret != RVM_SUCCESS) {
err = (int)rvmret;
} else
for (i = 0; i < list->count; i++) {
/* find pointer to block header */
free_block_t *bp = BLOCK_HDR((list->table)[i]);
/* Set the lowguard to reflect that the block has been allocated. */
assert(bp->type == ALLOC_GUARD);
rvmret = rvm_set_range(tid, &(bp->type), sizeof(guard_t));
if (rvmret != RVM_SUCCESS) {
err = (int)rvmret;
break;
}
bp->type = FREE_GUARD;
/* Update statistics */
RDS_STATS.free++;
RDS_STATS.freebytes += bp->size * RDS_CHUNK_SIZE;
RDS_STATS.mallocbytes -= bp->size * RDS_CHUNK_SIZE;
RDS_LOG("rdstrace: addr %p size %lx\n",
USER_BLOCK(bp) , bp->size * RDS_CHUNK_SIZE);
/* try to merge with trailing free blocks */
merge_with_next_free(bp, tid, &err);
if (err != SUCCESS)
break;
/* Add the block to the approprite free list. */
put_block(bp, tid, &err);
if (err != SUCCESS)
break;
}
RDS_LOG("rdstrace: end do_free\n");
if (err != SUCCESS) {
rvm_abort_transaction(tid);
} else {
rvmret = rvm_end_transaction(tid, mode);
}
});
rvm_free_tid(tid);
free(list->table);
list->table = NULL; /* Just to be safe */
if (err != SUCCESS) return err;
if (rvmret != RVM_SUCCESS) return (int) rvmret;
return (err != SUCCESS) ? err : ((rvmret != RVM_SUCCESS) ? (int)rvmret : 0);
}
|