* Sebastian Hammer, Adam Dickmeiss
*
* $Log: physical.c,v $
- * Revision 1.6 1995-09-04 12:33:47 adam
+ * Revision 1.13 1996-10-29 13:56:57 adam
+ * Include of zebrautl.h instead of alexutil.h.
+ *
+ * Revision 1.12 1996/03/20 16:17:11 quinn
+ * Bug hunting
+ *
+ * Revision 1.11 1996/03/20 13:29:17 quinn
+ * Bug-fix
+ *
+ * Revision 1.10 1996/03/19 19:22:44 quinn
+ * Fixed update-bug
+ *
+ * Revision 1.9 1996/02/06 10:19:57 quinn
+ * Attempt at fixing bug. Not all blocks were read before they were unlinked
+ * prior to a remap operation.
+ *
+ * Revision 1.8 1996/01/29 09:47:11 quinn
+ * Fixed mean little bug in the read-table code.
+ *
+ * Revision 1.7 1995/12/06 14:48:27 quinn
+ * Fixed some strange bugs.
+ *
+ * Revision 1.6 1995/09/04 12:33:47 adam
* Various cleanup. YAZ util used instead.
*
* Revision 1.5 1994/09/28 11:29:33 quinn
#include <assert.h>
#include <stdio.h>
+#include <log.h>
#include <isam.h>
static int is_freestore_alloc(ISAM is, int type)
is_mbuf *buf;
assert(block->state == IS_MBSTATE_UNREAD);
- block->data = buf = xmalloc_mbuf(IS_MBUF_TYPE_LARGE);
+ block->data = buf = xmalloc_mbuf(IS_MBUF_TYPE_LARGE);
toread = tab->is->types[tab->pos_type].blocksize;
if (toread > is_mbuf_size[buf->type])
{
/* extract header info */
buf->offset = 0;
memcpy(&block->num_records, buf->data, sizeof(block->num_records));
+ assert(block->num_records > 0);
buf->offset += sizeof(block->num_records);
memcpy(&block->nextpos, buf->data + buf->offset,
sizeof(block->nextpos));
sizeof(tab->num_records));
buf->offset +=sizeof(tab->num_records);
}
+ logf(LOG_DEBUG, "R: Block #%d: num %d nextpos %d total %d",
+ block->diskpos, block->num_records, block->nextpos,
+ block == tab->data ? tab->num_records : -1);
buf->num = (toread - buf->offset) / is_keysize(tab->is);
if (buf->num >= block->num_records)
{
block->state = IS_MBSTATE_CLEAN;
}
else
- block->bread = buf->num * is_keysize(tab->is);
+ block->bread = buf->offset + buf->num * is_keysize(tab->is);
return 0;
}
dread += toread;
block->bread += toread * is_keysize(tab->is);
}
+ block->state = IS_MBSTATE_CLEAN;
}
logf (LOG_DEBUG, "R: Block #%d contains %d records.", block->diskpos, block->num_records);
return 0;
sizeof(tab->num_records));
sum += sizeof(tab->num_records);
}
+ logf (LOG_DEBUG, "W: Block #%d contains %d records.", p->diskpos,
+ p->num_records);
+ assert(p->num_records > 0);
for (b = p->data; b; b = b->next)
{
- memcpy(type->dbuf + sum, b->data + b->offset, v = b->num *
- is_keysize(tab->is));
+ logf(LOG_DEBUG, " buf: offset %d, keys %d, type %d, ref %d",
+ b->offset, b->num, b->type, b->refcount);
+ if ((v = b->num * is_keysize(tab->is)) > 0)
+ memcpy(type->dbuf + sum, b->data + b->offset, v);
+
sum += v;
assert(sum <= type->blocksize);
}
logf (LOG_FATAL, "Failed to write block.");
exit(1);
}
- logf (LOG_DEBUG, "W: Block #%d contains %d records.", p->diskpos, p->num_records);
}
}
is_mblock *p;
for (p = tab->data; p; p = p->next)
+ {
if (p->diskpos >= 0)
{
is_freestore_free(tab->is, tab->pos_type, p->diskpos);
p->diskpos = -1;
}
+ }
}
static is_mbuf *mbuf_takehead(is_mbuf **mb, int *num, int keysize)
}
else
{
- tab->data = tab->data->next;
- tab->data->state = IS_MBSTATE_DIRTY;
- next = tab->data;
+ next = tab->data->next;
+ if (next)
+ {
+ if (next->state < IS_MBSTATE_CLEAN)
+ {
+ if (is_p_read_full(tab, next) < 0)
+ {
+ logf(LOG_FATAL, "Error during re-alignment");
+ abort();
+ }
+ if (next->nextpos && !next->next)
+ {
+ next->next = xmalloc_mblock();
+ next->next->diskpos = next->nextpos;
+ next->next->state = IS_MBSTATE_UNREAD;
+ next->next->data = 0;
+ }
+ }
+ next->state = IS_MBSTATE_DIRTY; /* force re-process */
+ tab->data = next;
+ }
}
if (mblock->diskpos >= 0)
is_freestore_free(tab->is, tab->pos_type, mblock->diskpos);