diff options
author | cvsadm <cvsadm> | 2005-01-21 00:44:34 +0000 |
---|---|---|
committer | cvsadm <cvsadm> | 2005-01-21 00:44:34 +0000 |
commit | b2093e3016027d6b5cf06b3f91f30769bfc099e2 (patch) | |
tree | cf58939393a9032182c4fbc4441164a9456e82f8 /ldap/servers/slapd/back-ldbm/import-threads.c | |
download | ds-b2093e3016027d6b5cf06b3f91f30769bfc099e2.tar.gz ds-b2093e3016027d6b5cf06b3f91f30769bfc099e2.tar.xz ds-b2093e3016027d6b5cf06b3f91f30769bfc099e2.zip |
Moving NSCP Directory Server from DirectoryBranch to TRUNK, initial drop. (foxworth)ldapserver7x
Diffstat (limited to 'ldap/servers/slapd/back-ldbm/import-threads.c')
-rw-r--r-- | ldap/servers/slapd/back-ldbm/import-threads.c | 1992 |
1 files changed, 1992 insertions, 0 deletions
diff --git a/ldap/servers/slapd/back-ldbm/import-threads.c b/ldap/servers/slapd/back-ldbm/import-threads.c new file mode 100644 index 00000000..413eaca6 --- /dev/null +++ b/ldap/servers/slapd/back-ldbm/import-threads.c @@ -0,0 +1,1992 @@ +/** BEGIN COPYRIGHT BLOCK + * Copyright 2001 Sun Microsystems, Inc. + * Portions copyright 1999, 2001-2003 Netscape Communications Corporation. + * All rights reserved. + * END COPYRIGHT BLOCK **/ +/* + * the threads that make up an import: + * producer (1) + * foreman (1) + * worker (N: 1 for each index) + * + * a wire import (aka "fast replica" import) won't have a producer thread. + */ + +#include "back-ldbm.h" +#include "vlv_srch.h" +#include "import.h" +#ifdef XP_WIN32 +#define STDIN_FILENO 0 +#endif + + +static struct backentry *import_make_backentry(Slapi_Entry *e, ID id) +{ + struct backentry *ep = backentry_alloc(); + + if (NULL != ep) { + ep->ep_entry = e; + ep->ep_id = id; + } + return ep; +} + +static void import_decref_entry(struct backentry *ep) +{ + PR_AtomicDecrement(&(ep->ep_refcnt)); + PR_ASSERT(ep->ep_refcnt >= 0); +} + +/* generate uniqueid if requested */ +static void import_generate_uniqueid(ImportJob *job, Slapi_Entry *e) +{ + const char *uniqueid = slapi_entry_get_uniqueid(e); + int rc; + + if (!uniqueid && (job->uuid_gen_type != SLAPI_UNIQUEID_GENERATE_NONE)) { + char *newuniqueid; + + /* generate id based on dn */ + if (job->uuid_gen_type == SLAPI_UNIQUEID_GENERATE_NAME_BASED) { + char *dn = slapi_entry_get_dn(e); + + rc = slapi_uniqueIDGenerateFromNameString(&newuniqueid, + job->uuid_namespace, dn, strlen(dn)); + } else { + /* time based */ + rc = slapi_uniqueIDGenerateString(&newuniqueid); + } + + if (rc == UID_SUCCESS) { + slapi_entry_set_uniqueid (e, newuniqueid); + } else { + char ebuf[BUFSIZ]; + LDAPDebug( LDAP_DEBUG_ANY, + "import_generate_uniqueid: failed to generate " + "uniqueid for %s; error=%d.\n", + escape_string(slapi_entry_get_dn_const(e), ebuf), rc, 0 ); + } + } +} + + +/********** BETTER LDIF PARSER **********/ + + +/* like the function in libldif, except this one doesn't need to use + * FILE (which breaks on various platforms for >4G files or large numbers + * of open files) + */ +#define LDIF_BUFFER_SIZE 8192 + +typedef struct { + char *b; /* buffer */ + size_t size; /* how full the buffer is */ + size_t offset; /* where the current entry starts */ +} ldif_context; + +static void import_init_ldif(ldif_context *c) +{ + c->size = c->offset = 0; + c->b = NULL; +} + +static void import_free_ldif(ldif_context *c) +{ + if (c->b) + FREE(c->b); + import_init_ldif(c); +} + +static char *import_get_entry(ldif_context *c, int fd, int *lineno) +{ + int ret; + int done = 0, got_lf = 0; + size_t bufSize = 0, bufOffset = 0, i; + char *buf = NULL; + + while (!done) { + + /* If there's no data in the buffer, get some */ + if ((c->size == 0) || (c->offset == c->size)) { + /* Do we even have a buffer ? */ + if (! c->b) { + c->b = slapi_ch_malloc(LDIF_BUFFER_SIZE); + if (! c->b) + return NULL; + } + ret = read(fd, c->b, LDIF_BUFFER_SIZE); + if (ret < 0) { + /* Must be error */ + goto error; + } else if (ret == 0) { + /* eof */ + if (buf) { + /* last entry */ + buf[bufOffset] = 0; + return buf; + } + return NULL; + } else { + /* read completed OK */ + c->size = ret; + c->offset = 0; + } + } + + /* skip blank lines at start of entry */ + if (bufOffset == 0) { + size_t n; + char *p; + + for (n = c->offset, p = c->b + n; n < c->size; n++, p++) { + if (!(*p == '\r' || *p == '\n' || *p == ' '|| *p == '\t')) + break; + } + c->offset = n; + if (c->offset == c->size) continue; + } + + i = c->offset; + while (!done && (i < c->size)) { + /* scan forward in the buffer, looking for the end of the entry */ + while ((i < c->size) && (c->b[i] != '\n')) + i++; + + if ((i < c->size) && (c->b[i] == '\n')) { + if (got_lf && ((i == 0) || ((i == 1) && (c->b[0] == '\r')))) { + /* saw an lf at the end of the last buffer */ + i++, (*lineno)++; + done = 1; + got_lf = 0; + break; + } + got_lf = 0; + (*lineno)++; + /* is this the end? (need another linefeed) */ + if (++i < c->size) { + if (c->b[i] == '\n') { + /* gotcha! */ + i++, (*lineno)++; + done = 1; + } else if (c->b[i] == '\r') { + if (++i < c->size) { + if (c->b[i] == '\n') { + /* gotcha! (nt) */ + i++, (*lineno)++; + done = 1; + } + } else { + got_lf = 1; + } + } + } else { + /* lf at the very end of the buffer */ + got_lf = 1; + } + } + } + + /* copy what we did so far into the output buffer */ + /* (first, make sure the output buffer is large enough) */ + if (bufSize - bufOffset < i - c->offset + 1) { + char *newbuf = NULL; + size_t newsize = (buf ? bufSize*2 : LDIF_BUFFER_SIZE); + + newbuf = slapi_ch_malloc(newsize); + if (! newbuf) + goto error; + /* copy over the old data (if there was any) */ + if (buf) { + memmove(newbuf, buf, bufOffset); + slapi_ch_free((void **)&buf); + } + buf = newbuf; + bufSize = newsize; + } + memmove(buf + bufOffset, c->b + c->offset, i - c->offset); + bufOffset += (i - c->offset); + c->offset = i; + } + + /* add terminating NUL char */ + buf[bufOffset] = 0; + return buf; + +error: + if (buf) + slapi_ch_free((void **)&buf); + return NULL; +} + + +/********** THREADS **********/ + +/* + * Description: + * 1) return the ldif version # + * 2) replace "version: 1" with "#ersion: 1" + * to pretend like a comment for the str2entry + */ +static int +import_get_version(char *str) +{ + char *s; + char *type; + char *valuecharptr; + char *mystr, *ms; + int offset; + int valuelen; + int my_version = 0; + int retmalloc = 0; + + if ((s = strstr(str, "version:")) == NULL) + return 0; + + offset = s - str; + mystr = ms = slapi_ch_strdup(str); + while ( (s = ldif_getline( &ms )) != NULL ) { + char *errmsg = NULL; + if ( (retmalloc = ldif_parse_line( s, &type, &valuecharptr, &valuelen, &errmsg )) >= 0 ) { + if (!strcasecmp(type, "version")) { + my_version = atoi(valuecharptr); + *(str + offset) = '#'; + /* the memory below was not allocated by the slapi_ch_ functions */ + if (errmsg) slapi_ch_free((void **) &errmsg); + if (retmalloc) slapi_ch_free((void **) &valuecharptr); + break; + } + } else if ( errmsg != NULL ) { + LDAPDebug( LDAP_DEBUG_PARSE, "%s", errmsg, 0, 0 ); + } + /* the memory below was not allocated by the slapi_ch_ functions */ + if (errmsg) slapi_ch_free((void **) &errmsg); + if (retmalloc) slapi_ch_free((void **) &valuecharptr); + } + + slapi_ch_free((void **)&mystr); + return my_version; +} + +/* producer thread: + * read through the given file list, parsing entries (str2entry), assigning + * them IDs and queueing them on the entry FIFO. other threads will do + * the indexing. + */ +void import_producer(void *param) +{ + ImportWorkerInfo *info = (ImportWorkerInfo *)param; + ImportJob *job = info->job; + ID id = job->first_ID, id_filestart = id; + Slapi_Entry *e = NULL; + struct backentry *ep = NULL, *old_ep = NULL; + ldbm_instance *inst = job->inst; + PRIntervalTime sleeptime; + char *estr = NULL; + int str2entry_flags = + SLAPI_STR2ENTRY_TOMBSTONE_CHECK | + SLAPI_STR2ENTRY_REMOVEDUPVALS | + SLAPI_STR2ENTRY_EXPAND_OBJECTCLASSES | + SLAPI_STR2ENTRY_ADDRDNVALS | + SLAPI_STR2ENTRY_NOT_WELL_FORMED_LDIF; + int finished = 0; + int detected_eof = 0; + int fd, curr_file, curr_lineno; + char *curr_filename = NULL; + int idx; + ldif_context c; + int my_version = 0; + size_t newesize = 0; + + PR_ASSERT(info != NULL); + PR_ASSERT(inst != NULL); + + if ( job->flags & FLAG_ABORT ) { + goto error; + } + + sleeptime = PR_MillisecondsToInterval(import_sleep_time); + + /* pause until we're told to run */ + while ((info->command == PAUSE) && !(job->flags & FLAG_ABORT)) { + info->state = WAITING; + DS_Sleep(sleeptime); + } + info->state = RUNNING; + import_init_ldif(&c); + + /* jumpstart by opening the first file */ + curr_file = 0; + fd = -1; + detected_eof = finished = 0; + + /* we loop around reading the input files and processing each entry + * as we read it. + */ + while (! finished) { + Slapi_Attr *attr = NULL; + int flags = 0; + int prev_lineno = 0; + int lines_in_entry = 0; + + if (job->flags & FLAG_ABORT) { + goto error; + } + + /* move on to next file? */ + if (detected_eof) { + /* check if the file can still be read, whine if so... */ + if (read(fd, (void *)&idx, 1) > 0) { + import_log_notice(job, "WARNING: Unexpected end of file found " + "at line %d of file \"%s\"", curr_lineno, + curr_filename); + } + + if (fd == STDIN_FILENO) { + import_log_notice(job, "Finished scanning file stdin (%lu " + "entries)", (u_long)(id-id_filestart)); + } else { + import_log_notice(job, "Finished scanning file \"%s\" (%lu " + "entries)", curr_filename, (u_long)(id-id_filestart)); + } + close(fd); + fd = -1; + detected_eof = 0; + id_filestart = id; + curr_file++; + if (job->task) { + job->task->task_progress++; + slapi_task_status_changed(job->task); + } + if (job->input_filenames[curr_file] == NULL) { + /* done! */ + finished = 1; + break; + } + } + + /* separate from above, because this is also triggered when we + * start (to open the first file) + */ + if (fd < 0) { + curr_lineno = 0; + curr_filename = job->input_filenames[curr_file]; + if (strcmp(curr_filename, "-") == 0) { + fd = STDIN_FILENO; + } else { + int o_flag = O_RDONLY; +#ifdef XP_WIN32 + /* 613041 Somehow the windows low level io lose "\n" + at a very particular situation using O_TEXT mode read. + I think it is a windows bug for O_TEXT mode read. + Use O_BINARY instead, which honestly returns chars + without any translation. + */ + o_flag |= O_BINARY; +#endif + fd = dblayer_open_huge_file(curr_filename, o_flag, 0); + } + if (fd < 0) { + import_log_notice(job, "Could not open LDIF file \"%s\"", + curr_filename); + goto error; + } + if (fd == STDIN_FILENO) { + import_log_notice(job, "Processing file stdin"); + } else { + import_log_notice(job, "Processing file \"%s\"", curr_filename); + } + } + if (job->flags & FLAG_ABORT) { + goto error; + } + + + while ((info->command == PAUSE) && !(job->flags & FLAG_ABORT)){ + info->state = WAITING; + DS_Sleep(sleeptime); + } + info->state = RUNNING; + + prev_lineno = curr_lineno; + estr = import_get_entry(&c, fd, &curr_lineno); + + lines_in_entry = curr_lineno - prev_lineno; + if (!estr) { + /* error reading entry, or end of file */ + detected_eof = 1; + continue; + } + + if (0 == my_version && strstr(estr, "version:")) { + my_version = import_get_version(estr); + str2entry_flags |= SLAPI_STR2ENTRY_INCLUDE_VERSION_STR; + } + + /* If there are more than so many lines in the entry, we tell + * str2entry to optimize for a large entry. + */ + if (lines_in_entry > STR2ENTRY_ATTRIBUTE_PRESENCE_CHECK_THRESHOLD) { + flags = str2entry_flags | SLAPI_STR2ENTRY_BIGENTRY; + } else { + flags = str2entry_flags; + } + e = slapi_str2entry(estr, flags); + FREE(estr); + if (! e) { + if (!(str2entry_flags & SLAPI_STR2ENTRY_INCLUDE_VERSION_STR)) + import_log_notice(job, "WARNING: skipping bad LDIF entry " + "ending line %d of file \"%s\"", curr_lineno, + curr_filename); + continue; + } + if (0 == my_version) { + /* after the first entry version string won't be given */ + my_version = -1; + } + + if (! import_entry_belongs_here(e, inst->inst_be)) { + /* silently skip */ + if (e) { + job->not_here_skipped++; + slapi_entry_free(e); + } + continue; + } + + if (slapi_entry_schema_check(NULL, e) != 0) { + char ebuf[BUFSIZ]; + import_log_notice(job, "WARNING: skipping entry \"%s\" which " + "violates schema, ending line %d of file " + "\"%s\"", escape_string(slapi_entry_get_dn(e), ebuf), + curr_lineno, curr_filename); + if (e) + slapi_entry_free(e); + job->skipped++; + continue; + } + + /* generate uniqueid if necessary */ + import_generate_uniqueid(job, e); + + ep = import_make_backentry(e, id); + if (!ep) + goto error; + + /* check for include/exclude subtree lists */ + if (! ldbm_back_ok_to_dump(backentry_get_ndn(ep), + job->include_subtrees, + job->exclude_subtrees)) { + backentry_free(&ep); + continue; + } + + /* not sure what this does, but it looked like it could be + * simplified. if it's broken, it's my fault. -robey + */ + if (slapi_entry_attr_find(ep->ep_entry, "userpassword", &attr) == 0) { + Slapi_Value **va = attr_get_present_values(attr); + + pw_encodevals( (Slapi_Value **)va ); /* jcm - cast away const */ + } + + if (job->flags & FLAG_ABORT) { + goto error; + } + + + /* Now we have this new entry, all decoded + * Next thing we need to do is: + * (1) see if the appropriate fifo location contains an + * entry which had been processed by the indexers. + * If so, proceed. + * If not, spin waiting for it to become free. + * (2) free the old entry and store the new one there. + * (3) Update the job progress indicators so the indexers + * can use the new entry. + */ + idx = id % job->fifo.size; + old_ep = job->fifo.item[idx].entry; + if (old_ep) { + /* for the slot to be recycled, it needs to be already absorbed + * by the foreman (id >= ready_ID), and all the workers need to + * be finished with it (refcount = 0). + */ + while (((old_ep->ep_refcnt > 0) || + (old_ep->ep_id >= job->ready_ID)) + && (info->command != ABORT) && !(job->flags & FLAG_ABORT)) { + info->state = WAITING; + DS_Sleep(sleeptime); + } + if (job->flags & FLAG_ABORT){ + goto error; + } + info->state = RUNNING; + PR_ASSERT(old_ep == job->fifo.item[idx].entry); + job->fifo.item[idx].entry = NULL; + if (job->fifo.c_bsize > job->fifo.item[idx].esize) + job->fifo.c_bsize -= job->fifo.item[idx].esize; + else + job->fifo.c_bsize = 0; + backentry_free(&old_ep); + } + + newesize = (slapi_entry_size(ep->ep_entry) + sizeof(struct backentry)); + if (newesize > job->fifo.bsize) { /* entry too big */ + char ebuf[BUFSIZ]; + import_log_notice(job, "WARNING: skipping entry \"%s\" " + "ending line %d of file \"%s\"", + escape_string(slapi_entry_get_dn(e), ebuf), + curr_lineno, curr_filename); + import_log_notice(job, "REASON: entry too large (%d bytes) for " + "the buffer size (%d bytes)", newesize, job->fifo.bsize); + backentry_free(&ep); + job->skipped++; + continue; + } + /* Now check if fifo has enough space for the new entry */ + if ((job->fifo.c_bsize + newesize) > job->fifo.bsize) { + import_wait_for_space_in_fifo( job, newesize ); + } + + /* We have enough space */ + job->fifo.item[idx].filename = curr_filename; + job->fifo.item[idx].line = curr_lineno; + job->fifo.item[idx].entry = ep; + job->fifo.item[idx].bad = 0; + job->fifo.item[idx].esize = newesize; + + /* Add the entry size to total fifo size */ + job->fifo.c_bsize += ep->ep_entry? job->fifo.item[idx].esize : 0; + + /* Update the job to show our progress */ + job->lead_ID = id; + if ((id - info->first_ID) <= job->fifo.size) { + job->trailing_ID = info->first_ID; + } else { + job->trailing_ID = id - job->fifo.size; + } + + /* Update our progress meter too */ + info->last_ID_processed = id; + id++; + if (job->flags & FLAG_ABORT){ + goto error; + } + if (info->command == STOP) { + if (fd >= 0) + close(fd); + finished = 1; + } + } + + import_free_ldif(&c); + info->state = FINISHED; + return; + +error: + info->state = ABORTED; +} + +#if defined(UPGRADEDB) +/* producer thread for re-indexing: + * read id2entry, parsing entries (str2entry) (needed???), assigning + * them IDs (again, needed???) and queueing them on the entry FIFO. + * other threads will do the indexing -- same as in import. + */ +void index_producer(void *param) +{ + ImportWorkerInfo *info = (ImportWorkerInfo *)param; + ImportJob *job = info->job; + ID id = job->first_ID; + Slapi_Entry *e = NULL; + struct backentry *ep = NULL, *old_ep = NULL; + ldbm_instance *inst = job->inst; + PRIntervalTime sleeptime; + int finished = 0; + int idx; + + /* vars for Berkeley DB */ + DB_ENV *env = NULL; + DB *db = NULL; + DBC *dbc = NULL; + DBT key = {0}; + DBT data = {0}; + int db_rval = -1; + backend *be = inst->inst_be; + int isfirst = 1; + int curr_entry = 0; + size_t newesize = 0; + + PR_ASSERT(info != NULL); + PR_ASSERT(inst != NULL); + PR_ASSERT(be != NULL); + + if ( job->flags & FLAG_ABORT ) + goto error; + + sleeptime = PR_MillisecondsToInterval(import_sleep_time); + + /* pause until we're told to run */ + while ((info->command == PAUSE) && !(job->flags & FLAG_ABORT)) { + info->state = WAITING; + DS_Sleep(sleeptime); + } + info->state = RUNNING; + + /* open id2entry with dedicated db env and db handler */ + if ( dblayer_get_aux_id2entry( be, &db, &env ) != 0 || db == NULL || + env == NULL) { + LDAPDebug( LDAP_DEBUG_ANY, "Could not open id2entry\n", 0, 0, 0 ); + goto error; + } + + /* get a cursor to we can walk over the table */ + db_rval = db->cursor(db, NULL, &dbc, 0); + if ( 0 != db_rval ) { + LDAPDebug( LDAP_DEBUG_ANY, + "Failed to get cursor for reindexing\n", 0, 0, 0 ); + dblayer_release_id2entry(be, db); + goto error; + } + + /* we loop around reading the input files and processing each entry + * as we read it. + */ + finished = 0; + while (!finished) { + Slapi_Attr *attr = NULL; + ID temp_id; + + if (job->flags & FLAG_ABORT) { + goto error; + } + while ((info->command == PAUSE) && !(job->flags & FLAG_ABORT)){ + info->state = WAITING; + DS_Sleep(sleeptime); + } + info->state = RUNNING; + + key.flags = DB_DBT_MALLOC; + data.flags = DB_DBT_MALLOC; + if (isfirst) + { + db_rval = dbc->c_get(dbc, &key, &data, DB_FIRST); + isfirst = 0; + } + else + { + db_rval = dbc->c_get(dbc, &key, &data, DB_NEXT); + } + + if (0 != db_rval) { + if (DB_NOTFOUND != db_rval) { + LDAPDebug(LDAP_DEBUG_ANY, "%s: Failed to read database, " + "errno=%d (%s)\n", inst->inst_name, db_rval, + dblayer_strerror(db_rval)); + if (job->task) { + slapi_task_log_notice(job->task, + "%s: Failed to read database, err %d (%s)", + inst->inst_name, db_rval, + dblayer_strerror(db_rval)); + } + } + break; + } + curr_entry++; + temp_id = id_stored_to_internal((char *)key.data); + free(key.data); + + /* call post-entry plugin */ + plugin_call_entryfetch_plugins((char **) &data.dptr, &data.dsize); + e = slapi_str2entry(data.data, 0); + if ( NULL == e ) { + if (job->task) { + slapi_task_log_notice(job->task, + "%s: WARNING: skipping badly formatted entry (id %lu)", + inst->inst_name, (u_long)temp_id); + } + LDAPDebug(LDAP_DEBUG_ANY, + "%s: WARNING: skipping badly formatted entry (id %lu)\n", + inst->inst_name, (u_long)temp_id, 0); + continue; + } + free(data.data); + + /* generate uniqueid if necessary */ + import_generate_uniqueid(job, e); + + ep = import_make_backentry(e, temp_id); + if (!ep) + goto error; + + /* not sure what this does, but it looked like it could be + * simplified. if it's broken, it's my fault. -robey + */ + if (slapi_entry_attr_find(ep->ep_entry, "userpassword", &attr) == 0) { + Slapi_Value **va = attr_get_present_values(attr); + + pw_encodevals( (Slapi_Value **)va ); /* jcm - cast away const */ + } + + if (job->flags & FLAG_ABORT) + goto error; + + /* Now we have this new entry, all decoded + * Next thing we need to do is: + * (1) see if the appropriate fifo location contains an + * entry which had been processed by the indexers. + * If so, proceed. + * If not, spin waiting for it to become free. + * (2) free the old entry and store the new one there. + * (3) Update the job progress indicators so the indexers + * can use the new entry. + */ + idx = id % job->fifo.size; + old_ep = job->fifo.item[idx].entry; + if (old_ep) { + /* for the slot to be recycled, it needs to be already absorbed + * by the foreman (id >= ready_ID), and all the workers need to + * be finished with it (refcount = 0). + */ + while (((old_ep->ep_refcnt > 0) || + (old_ep->ep_id >= job->ready_ID)) + && (info->command != ABORT) && !(job->flags & FLAG_ABORT)) { + info->state = WAITING; + DS_Sleep(sleeptime); + } + if (job->flags & FLAG_ABORT) + goto error; + + info->state = RUNNING; + PR_ASSERT(old_ep == job->fifo.item[idx].entry); + job->fifo.item[idx].entry = NULL; + if (job->fifo.c_bsize > job->fifo.item[idx].esize) + job->fifo.c_bsize -= job->fifo.item[idx].esize; + else + job->fifo.c_bsize = 0; + backentry_free(&old_ep); + } + + newesize = (slapi_entry_size(ep->ep_entry) + sizeof(struct backentry)); + if (newesize > job->fifo.bsize) { /* entry too big */ + char ebuf[BUFSIZ]; + import_log_notice(job, "WARNING: skipping entry \"%s\"", + escape_string(slapi_entry_get_dn(e), ebuf)); + import_log_notice(job, "REASON: entry too large (%d bytes) for " + "the buffer size (%d bytes)", newesize, job->fifo.bsize); + backentry_free(&ep); + job->skipped++; + continue; + } + /* Now check if fifo has enough space for the new entry */ + if ((job->fifo.c_bsize + newesize) > job->fifo.bsize) { + import_wait_for_space_in_fifo( job, newesize ); + } + + /* We have enough space */ + job->fifo.item[idx].filename = ID2ENTRY LDBM_FILENAME_SUFFIX; + job->fifo.item[idx].line = curr_entry; + job->fifo.item[idx].entry = ep; + job->fifo.item[idx].bad = 0; + job->fifo.item[idx].esize = newesize; + + /* Add the entry size to total fifo size */ + job->fifo.c_bsize += ep->ep_entry? job->fifo.item[idx].esize : 0; + + /* Update the job to show our progress */ + job->lead_ID = id; + if ((id - info->first_ID) <= job->fifo.size) { + job->trailing_ID = info->first_ID; + } else { + job->trailing_ID = id - job->fifo.size; + } + + /* Update our progress meter too */ + info->last_ID_processed = id; + id++; + if (job->flags & FLAG_ABORT) + goto error; + if (info->command == STOP) + { + finished = 1; + } + } + + dbc->c_close(dbc); + dblayer_release_aux_id2entry( be, db, env ); + info->state = FINISHED; + return; + +error: + dbc->c_close(dbc); + dblayer_release_aux_id2entry( be, db, env ); + info->state = ABORTED; +} +#endif + +static void +import_wait_for_space_in_fifo(ImportJob *job, size_t new_esize) +{ + struct backentry *temp_ep = NULL; + size_t i; + int slot_found; + PRIntervalTime sleeptime; + + sleeptime = PR_MillisecondsToInterval(import_sleep_time); + + /* Now check if fifo has enough space for the new entry */ + while ((job->fifo.c_bsize + new_esize) > job->fifo.bsize) { + for ( i = 0, slot_found = 0 ; i < job->fifo.size ; i++ ) { + temp_ep = job->fifo.item[i].entry; + if (temp_ep) { + if (temp_ep->ep_refcnt == 0 && temp_ep->ep_id < job->ready_ID) { + job->fifo.item[i].entry = NULL; + if (job->fifo.c_bsize > job->fifo.item[i].esize) + job->fifo.c_bsize -= job->fifo.item[i].esize; + else + job->fifo.c_bsize = 0; + backentry_free(&temp_ep); + slot_found = 1; + } + } + } + if ( slot_found == 0 ) + DS_Sleep(sleeptime); + } +} + +/* helper function for the foreman: */ +static int foreman_do_parentid(ImportJob *job, struct backentry *entry, + struct attrinfo *parentid_ai) +{ + backend *be = job->inst->inst_be; + Slapi_Value **svals = NULL; + Slapi_Attr *attr = NULL; + int idl_disposition = 0; + int ret = 0; + + if (slapi_entry_attr_find(entry->ep_entry, "parentid", &attr) == 0) { + svals = attr_get_present_values(attr); + ret = index_addordel_values_ext_sv(be, "parentid", svals, NULL, entry->ep_id, + BE_INDEX_ADD, NULL, &idl_disposition, NULL); + if (idl_disposition != IDL_INSERT_NORMAL) { + char *attr_value = slapi_value_get_berval(svals[0])->bv_val; + ID parent_id = atol(attr_value); + + if (idl_disposition == IDL_INSERT_NOW_ALLIDS) { + import_subcount_mother_init(job->mothers, parent_id, + idl_get_allidslimit(parentid_ai)+1); + } else if (idl_disposition == IDL_INSERT_ALLIDS) { + import_subcount_mother_count(job->mothers, parent_id); + } + } + if (ret != 0) { + import_log_notice(job, "ERROR: Can't update parentid index " + "(error %d)", ret); + return ret; + } + } + + return 0; +} + +/* helper function for the foreman: */ +static int foreman_do_entrydn(ImportJob *job, FifoItem *fi) +{ + backend *be = job->inst->inst_be; + struct berval bv; + int err = 0, ret = 0; + IDList *IDL; + + /* insert into the entrydn index */ + bv.bv_val = (void*)backentry_get_ndn(fi->entry); /* jcm - Had to cast away const */ + bv.bv_len = strlen(backentry_get_ndn(fi->entry)); + + /* We need to check here whether the DN is already present in + * the entrydn index. If it is then the input ldif + * contained a duplicate entry, which it isn't allowed to */ + /* Due to popular demand, we only warn on this, given the + * tendency for customers to want to import dirty data */ + /* So, we do an index read first */ + err = 0; + IDL = index_read(be, "entrydn", indextype_EQUALITY, &bv, NULL, &err); + + /* Did this work ? */ + if (NULL != IDL) { + /* IMPOSTER ! Get thee hence... */ + import_log_notice(job, "WARNING: Skipping duplicate entry " + "\"%s\" found at line %d of file \"%s\"", + slapi_entry_get_dn(fi->entry->ep_entry), + fi->line, fi->filename); + idl_free(IDL); + /* skip this one */ + fi->bad = 1; + job->skipped++; + return -1; /* skip to next entry */ + } + if ((ret = index_addordel_string(be, "entrydn", + bv.bv_val, + fi->entry->ep_id, + BE_INDEX_ADD, NULL)) != 0) { + import_log_notice(job, "Error writing entrydn index " + "(error %d: %s)", + ret, dblayer_strerror(ret)); + return ret; + } + return 0; +} + +/* foreman thread: + * i go through the FIFO just like the other worker threads, but i'm + * responsible for the interrelated indexes: entrydn, id2entry, and the + * operational attributes (plus the parentid index). + */ +void import_foreman(void *param) +{ + ImportWorkerInfo *info = (ImportWorkerInfo *)param; + ImportJob *job = info->job; + ldbm_instance *inst = job->inst; + backend *be = inst->inst_be; + PRIntervalTime sleeptime; + int finished = 0; + ID id = info->first_ID; + int ret = 0; + struct attrinfo *parentid_ai; + Slapi_PBlock *pb = slapi_pblock_new(); + + PR_ASSERT(info != NULL); + PR_ASSERT(inst != NULL); + + if (job->flags & FLAG_ABORT) { + goto error; + } + + /* the pblock is used only by add_op_attrs */ + slapi_pblock_set(pb, SLAPI_BACKEND, be); + sleeptime = PR_MillisecondsToInterval(import_sleep_time); + info->state = RUNNING; + + ainfo_get(be, "parentid", &parentid_ai); + + while (! finished) { + FifoItem *fi = NULL; + int parent_status = 0; + + + if (job->flags & FLAG_ABORT) { + goto error; + } + + while ( ((info->command == PAUSE) || (id > job->lead_ID)) && + (info->command != STOP) && (info->command != ABORT) && !(job->flags & FLAG_ABORT)) { + /* Check to see if we've been told to stop */ + info->state = WAITING; + DS_Sleep(sleeptime); + } + if (info->command == STOP) { + finished = 1; + continue; + } + if (job->flags & FLAG_ABORT) { + goto error; + } + + info->state = RUNNING; + + /* Read that entry from the cache */ + fi = import_fifo_fetch(job, id, 0); + if (! fi) { + import_log_notice(job, "ERROR: foreman fifo error"); + goto error; + } + + /* first, fill in any operational attributes */ + /* add_op_attrs wants a pblock for some reason. */ + if (add_op_attrs(pb, inst->inst_li, fi->entry, &parent_status) != 0) { + import_log_notice(job, "ERROR: Could not add op attrs to " + "entry ending at line %d of file \"%s\"", + fi->line, fi->filename); + goto error; + } + + if (! slapi_entry_flag_is_set(fi->entry->ep_entry, + SLAPI_ENTRY_FLAG_TOMBSTONE)) { + /* + * Only check for a parent and add to the entry2dn index if + * the entry is not a tombstone. + */ + if (job->flags & FLAG_ABORT) { + goto error; + } + + if (parent_status == IMPORT_ADD_OP_ATTRS_NO_PARENT) { + /* If this entry is a suffix entry, this is not a problem */ + /* However, if it is not, this is an error---it means that + * someone tried to import an entry before importing its parent + * we reject the entry but carry on since we've not stored + * anything related to this entry. + */ + if (! slapi_be_issuffix(inst->inst_be, backentry_get_sdn(fi->entry))) { + import_log_notice(job, "WARNING: Skipping entry \"%s\" " + "which has no parent, ending at line %d " + "of file \"%s\"", + slapi_entry_get_dn(fi->entry->ep_entry), + fi->line, fi->filename); + /* skip this one */ + fi->bad = 1; + job->skipped++; + goto cont; /* below */ + } + } + if (job->flags & FLAG_ABORT) { + goto error; + } + + + /* insert into the entrydn index */ + ret = foreman_do_entrydn(job, fi); + if (ret == -1) + goto cont; /* skip entry */ + if (ret != 0) + goto error; + } + + if (job->flags & FLAG_ABORT) { + goto error; + } + +#if defined (UPGRADEDB) + if (!(job->flags & FLAG_REINDEXING))/* reindex reads data from id2entry */ +#endif + { + /* insert into the id2entry index + * (that isn't really an index -- it's the storehouse of the entries + * themselves.) + */ + if ((ret = id2entry_add_ext(be, fi->entry, NULL, job->encrypt)) != 0) { + /* DB_RUNRECOVERY usually occurs if disk fills */ + if (LDBM_OS_ERR_IS_DISKFULL(ret)) { + import_log_notice(job, "ERROR: OUT OF SPACE ON DISK or FILE TOO LARGE -- " + "Could not store the entry ending at line " + "%d of file \"%s\"", + fi->line, fi->filename); + } else if (ret == DB_RUNRECOVERY) { + import_log_notice(job, "FATAL ERROR: (LARGEFILE SUPPORT NOT ENABLED? OUT OF SPACE ON DISK?) -- " + "Could not store the entry ending at line " + "%d of file \"%s\"", + fi->line, fi->filename); + } else { + import_log_notice(job, "ERROR: Could not store the entry " + "ending at line %d of file \"%s\" -- " + "error %d", fi->line, fi->filename, ret); + } + goto error; + } + } + + if (job->flags & FLAG_ABORT) { + goto error; + } + + if (! slapi_entry_flag_is_set(fi->entry->ep_entry, + SLAPI_ENTRY_FLAG_TOMBSTONE)) { + /* parentid index + * (we have to do this here, because the parentID is dependent on + * looking up by entrydn.) + * Only add to the parent index if the entry is not a tombstone. + */ + ret = foreman_do_parentid(job, fi->entry, parentid_ai); + if (ret != 0) + goto error; + + /* Lastly, before we're finished with the entry, pass it to the + vlv code to see whether it's within the scope a VLV index. */ + vlv_grok_new_import_entry(fi->entry, be); + } + if (job->flags & FLAG_ABORT) { + goto error; + } + + + /* Remove the entry from the cache (caused by id2entry_add) */ +#if defined (UPGRADEDB) + if (!(job->flags & FLAG_REINDEXING))/* reindex reads data from id2entry */ +#endif + cache_remove(&inst->inst_cache, fi->entry); + fi->entry->ep_refcnt = job->number_indexers; + + cont: + if (job->flags & FLAG_ABORT) { + goto error; + } + + job->ready_ID = id; + info->last_ID_processed = id; + id++; + + if (job->flags & FLAG_ABORT){ + goto error; + } + } + + slapi_pblock_destroy(pb); + info->state = FINISHED; + return; + +error: + slapi_pblock_destroy(pb); + info->state = ABORTED; +} + + +/* worker thread: + * given an attribute, this worker plows through the entry FIFO, building + * up the attribute index. + */ +void import_worker(void *param) +{ + ImportWorkerInfo *info = (ImportWorkerInfo *)param; + ImportJob *job = info->job; + ldbm_instance *inst = job->inst; + backend *be = inst->inst_be; + PRIntervalTime sleeptime; + int finished = 0; + ID id = info->first_ID; + int ret = 0; + int idl_disposition = 0; + struct vlvIndex* vlv_index = NULL; + void *substring_key_buffer = NULL; + FifoItem *fi; + int is_objectclass_attribute; + int is_nsuniqueid_attribute; + void *attrlist_cursor; + + PR_ASSERT(NULL != info); + PR_ASSERT(NULL != inst); + + if (job->flags & FLAG_ABORT) { + goto error; + } + + if (INDEX_VLV == info->index_info->ai->ai_indexmask) { + vlv_index = vlv_find_indexname(info->index_info->name, be); + if (NULL == vlv_index) { + goto error; + } + } + + /* + * If the entry is a Tombstone, then we only add it to the nsuniqeid index + * and the idlist for (objectclass=tombstone). These two flags are just + * handy for working out what to do in this case. + */ + is_objectclass_attribute = + (strcasecmp(info->index_info->name, "objectclass") == 0); + is_nsuniqueid_attribute = + (strcasecmp(info->index_info->name, SLAPI_ATTR_UNIQUEID) == 0); + + if (1 != idl_get_idl_new()) { + /* Is there substring indexing going on here ? */ + if ( (INDEX_SUB & info->index_info->ai->ai_indexmask) && + (info->index_buffer_size > 0) ) { + /* Then make a key buffer thing */ + ret = index_buffer_init(info->index_buffer_size, 0, + &substring_key_buffer); + if (0 != ret) { + import_log_notice(job, "IMPORT FAIL 1 (error %d)", ret); + } + } + } + + sleeptime = PR_MillisecondsToInterval(import_sleep_time); + info->state = RUNNING; + info->last_ID_processed = id-1; + + while (! finished) { + struct backentry *ep = NULL; + Slapi_Value **svals = NULL; + Slapi_Attr *attr = NULL; + + if (job->flags & FLAG_ABORT) { + goto error; + } + + /* entry can be NULL if it turned out to be bogus */ + while (!finished && !ep) { + /* This worker thread must wait if the command flag is "PAUSE" or + * the entry corresponds to the current entry treated by the foreman + * thread, and the state is neither STOP nor ABORT + */ + while (((info->command == PAUSE) || (id > job->ready_ID)) && + (info->command != STOP) && (info->command != ABORT) && !(job->flags & FLAG_ABORT)) { + /* Check to see if we've been told to stop */ + info->state = WAITING; + DS_Sleep(sleeptime); + } + + if (info->command == STOP) { + finished = 1; + continue; + } + if (job->flags & FLAG_ABORT) { + goto error; + } + + info->state = RUNNING; + + /* Read that entry from the cache */ + fi = import_fifo_fetch(job, id, 1); + ep = fi ? fi->entry : NULL; + if (!ep) { + /* skipping an entry that turned out to be bad */ + info->last_ID_processed = id; + id++; + } + } + if (finished) + continue; + + if (! slapi_entry_flag_is_set(fi->entry->ep_entry, + SLAPI_ENTRY_FLAG_TOMBSTONE)) { + /* This is not a tombstone entry. */ + /* Is this a VLV index ? */ + + if (job->flags & FLAG_ABORT) { + goto error; + } + + if (INDEX_VLV == info->index_info->ai->ai_indexmask) { + /* Yes, call VLV code -- needs pblock to find backend */ + Slapi_PBlock *pb = slapi_pblock_new(); + + PR_ASSERT(NULL != vlv_index); + slapi_pblock_set(pb, SLAPI_BACKEND, be); + vlv_update_index(vlv_index, NULL, inst->inst_li, pb, NULL, ep); + slapi_pblock_destroy(pb); + } else { + /* No, process regular index */ + /* Look for the attribute we're indexing and its subtypes */ + /* For each attr write to the index */ + attrlist_cursor = NULL; + while ((attr = attrlist_find_ex(ep->ep_entry->e_attrs, + info->index_info->name, + NULL, + NULL, + &attrlist_cursor)) != NULL) { + + if (job->flags & FLAG_ABORT) { + goto error; + } + if(valueset_isempty(&(attr->a_present_values))) continue; + svals = attr_get_present_values(attr); + ret = index_addordel_values_ext_sv(be, info->index_info->name, + svals, NULL, ep->ep_id, BE_INDEX_ADD | (job->encrypt ? 0 : BE_INDEX_DONT_ENCRYPT), NULL, &idl_disposition, + substring_key_buffer); + + if (0 != ret) { + /* Something went wrong, eg disk filled up */ + goto error; + } + } + } + } else { + /* This is a Tombstone entry... we only add it to the nsuniqeid + * index and the idlist for (objectclass=nstombstone). + */ + if (job->flags & FLAG_ABORT) { + goto error; + } + if (is_nsuniqueid_attribute) { + ret = index_addordel_string(be, SLAPI_ATTR_UNIQUEID, + slapi_entry_get_uniqueid(ep->ep_entry), ep->ep_id, + BE_INDEX_ADD, NULL); + if (0 != ret) { + /* Something went wrong, eg disk filled up */ + goto error; + } + } + if (is_objectclass_attribute) { + ret = index_addordel_string(be, SLAPI_ATTR_OBJECTCLASS, + SLAPI_ATTR_VALUE_TOMBSTONE, ep->ep_id, BE_INDEX_ADD, NULL); + if (0 != ret) { + /* Something went wrong, eg disk filled up */ + goto error; + } + } + } + import_decref_entry(ep); + info->last_ID_processed = id; + id++; + + if (job->flags & FLAG_ABORT) { + goto error; + } + } + + if (job->flags & FLAG_ABORT) { + goto error; + } + + + /* If we were buffering index keys, now flush them */ + if (substring_key_buffer) { + ret = index_buffer_flush(substring_key_buffer, + inst->inst_be, NULL, + info->index_info->ai); + if (0 != ret) { + goto error; + } + index_buffer_terminate(substring_key_buffer); + } + info->state = FINISHED; + return; + +error: + if (ret == DB_RUNRECOVERY) { + LDAPDebug(LDAP_DEBUG_ANY,"cannot import; database recovery needed\n", + 0,0,0); + } else if (ret == DB_LOCK_DEADLOCK) { + /* can this occur? */ + } + + info->state = ABORTED; +} + + + +/* + * import entries to a backend, over the wire -- entries will arrive + * asynchronously, so this method has no "producer" thread. instead, the + * front-end drops new entries in as they arrive. + * + * this is sometimes called "fast replica initialization". + * + * some of this code is duplicated from ldif2ldbm, but i don't think we + * can avoid it. + */ +static int bulk_import_start(Slapi_PBlock *pb) +{ + struct ldbminfo *li = NULL; + ImportJob *job = NULL; + backend *be = NULL; + PRThread *thread = NULL; + int ret = 0; + + job = CALLOC(ImportJob); + if (job == NULL) { + LDAPDebug(LDAP_DEBUG_ANY, "not enough memory to do import job\n", + 0, 0, 0); + return -1; + } + + slapi_pblock_get(pb, SLAPI_BACKEND, &be); + PR_ASSERT(be != NULL); + li = (struct ldbminfo *)(be->be_database->plg_private); + job->inst = (ldbm_instance *)be->be_instance_info; + + /* check if an import/restore is already ongoing... */ + PR_Lock(job->inst->inst_config_mutex); + if (job->inst->inst_flags & INST_FLAG_BUSY) { + PR_Unlock(job->inst->inst_config_mutex); + LDAPDebug(LDAP_DEBUG_ANY, "ldbm: '%s' is already in the middle of " + "another task and cannot be disturbed.\n", + job->inst->inst_name, 0, 0); + FREE(job); + return SLAPI_BI_ERR_BUSY; + } + job->inst->inst_flags |= INST_FLAG_BUSY; + PR_Unlock(job->inst->inst_config_mutex); + + /* take backend offline */ + slapi_mtn_be_disable(be); + + /* get uniqueid info */ + slapi_pblock_get(pb, SLAPI_LDIF2DB_GENERATE_UNIQUEID, &job->uuid_gen_type); + if (job->uuid_gen_type == SLAPI_UNIQUEID_GENERATE_NAME_BASED) { + char *namespaceid; + + slapi_pblock_get(pb, SLAPI_LDIF2DB_NAMESPACEID, &namespaceid); + job->uuid_namespace = slapi_ch_strdup(namespaceid); + } + + job->flags = 0; /* don't use files */ + job->flags |= FLAG_INDEX_ATTRS; + job->flags |= FLAG_ONLINE; + job->starting_ID = 1; + job->first_ID = 1; + + job->mothers = CALLOC(import_subcount_stuff); + /* how much space should we allocate to index buffering? */ + job->job_index_buffer_size = import_get_index_buffer_size(); + if (job->job_index_buffer_size == 0) { + /* 10% of the allocated cache size + one meg */ + job->job_index_buffer_size = (job->inst->inst_li->li_dbcachesize/10) + + (1024*1024); + } + import_subcount_stuff_init(job->mothers); + job->wire_lock = PR_NewLock(); + job->wire_cv = PR_NewCondVar(job->wire_lock); + + /* COPIED from ldif2ldbm.c : */ + + /* shutdown this instance of the db */ + cache_clear(&job->inst->inst_cache); + dblayer_instance_close(be); + + /* Delete old database files */ + dblayer_delete_instance_dir(be); + /* it's okay to fail -- it might already be gone */ + + /* dblayer_instance_start will init the id2entry index. */ + /* it also (finally) fills in inst_dir_name */ + ret = dblayer_instance_start(be, DBLAYER_IMPORT_MODE); + if (ret != 0) + goto fail; + + /* END OF COPIED SECTION */ + + PR_Lock(job->wire_lock); + vlv_init(job->inst); + + /* create thread for import_main, so we can return */ + thread = PR_CreateThread(PR_USER_THREAD, import_main, (void *)job, + PR_PRIORITY_NORMAL, PR_GLOBAL_THREAD, + PR_JOINABLE_THREAD, + SLAPD_DEFAULT_THREAD_STACKSIZE); + if (thread == NULL) { + PRErrorCode prerr = PR_GetError(); + LDAPDebug(LDAP_DEBUG_ANY, "unable to spawn import thread, " + SLAPI_COMPONENT_NAME_NSPR " error %d (%s)\n", + prerr, slapd_pr_strerror(prerr), 0); + PR_Unlock(job->wire_lock); + ret = -2; + goto fail; + } + + job->main_thread = thread; + slapi_set_object_extension(li->li_bulk_import_object, pb->pb_conn, + li->li_bulk_import_handle, job); + + /* wait for the import_main to signal that it's ready for entries */ + /* (don't want to send the success code back to the LDAP client until + * we're ready for the adds to start rolling in) + */ + PR_WaitCondVar(job->wire_cv, PR_INTERVAL_NO_TIMEOUT); + PR_Unlock(job->wire_lock); + + return 0; + +fail: + PR_Lock(job->inst->inst_config_mutex); + job->inst->inst_flags &= ~INST_FLAG_BUSY; + PR_Unlock(job->inst->inst_config_mutex); + import_free_job(job); + FREE(job); + return ret; +} + +/* returns 0 on success, or < 0 on error + * + * on error, the import process is aborted -- so if this returns an error, + * don't try to queue any more entries or you'll be sorry. + * + * flag_block in used to know if this thread should block when + * the fifo is full or return an error LDAP_BUSY + * Typically, import done on from the GUI or the command line will + * block while online import as used by the replication total update + * will not block + */ +static int bulk_import_queue(ImportJob *job, Slapi_Entry *entry, int flag_block) +{ + struct backentry *ep = NULL, *old_ep = NULL; + int idx; + ID id = job->lead_ID + 1; + Slapi_Attr *attr = NULL; + size_t newesize = 0; + + PR_Lock(job->wire_lock); + + /* generate uniqueid if necessary */ + import_generate_uniqueid(job, entry); + + /* make into backentry */ + ep = import_make_backentry(entry, id); + if (!ep) { + import_abort_all(job, 1); + PR_Unlock(job->wire_lock); + return -1; + } + + /* encode the password */ + if (slapi_entry_attr_find(ep->ep_entry, "userpassword", &attr) == 0) { + Slapi_Value **va = attr_get_present_values(attr); + + pw_encodevals( (Slapi_Value **)va ); /* jcm - had to cast away const */ + } + + /* Now we have this new entry, all decoded + * Next thing we need to do is: + * (1) see if the appropriate fifo location contains an + * entry which had been processed by the indexers. + * If so, proceed. + * If not, spin waiting for it to become free. + * (2) free the old entry and store the new one there. + * (3) Update the job progress indicators so the indexers + * can use the new entry. + */ + idx = id % job->fifo.size; + old_ep = job->fifo.item[idx].entry; + if (old_ep) { + while ((old_ep->ep_refcnt > 0) && !(job->flags & FLAG_ABORT)) + { + if (flag_block) + DS_Sleep(PR_MillisecondsToInterval(import_sleep_time)); + else + { + PR_Unlock(job->wire_lock); + return LDAP_BUSY; + } + } + + /* the producer could be running thru the fifo while + * everyone else is cycling to a new pass... + * double-check that this entry is < ready_ID + */ + while ((old_ep->ep_id >= job->ready_ID) && !(job->flags & FLAG_ABORT)) + { + if (flag_block) + DS_Sleep(PR_MillisecondsToInterval(import_sleep_time)); + else + { + PR_Unlock(job->wire_lock); + return LDAP_BUSY; + } + } + + if (job->flags & FLAG_ABORT) { + PR_Unlock(job->wire_lock); + return -2; + } + + PR_ASSERT(old_ep == job->fifo.item[idx].entry); + job->fifo.item[idx].entry = NULL; + if (job->fifo.c_bsize > job->fifo.item[idx].esize) + job->fifo.c_bsize -= job->fifo.item[idx].esize; + else + job->fifo.c_bsize = 0; + backentry_free(&old_ep); + } + + newesize = (slapi_entry_size(ep->ep_entry) + sizeof(struct backentry)); + if (newesize > job->fifo.bsize) { /* entry too big */ + char ebuf[BUFSIZ]; + import_log_notice(job, "WARNING: skipping entry \"%s\"", + escape_string(slapi_entry_get_dn(ep->ep_entry), ebuf)); + import_log_notice(job, "REASON: entry too large (%d bytes) for " + "the buffer size (%d bytes)", newesize, job->fifo.bsize); + backentry_free(&ep); + PR_Unlock(job->wire_lock); + return -1; + } + /* Now check if fifo has enough space for the new entry */ + if ((job->fifo.c_bsize + newesize) > job->fifo.bsize) { + import_wait_for_space_in_fifo( job, newesize ); + } + + /* We have enough space */ + job->fifo.item[idx].filename = "(bulk import)"; + job->fifo.item[idx].line = 0; + job->fifo.item[idx].entry = ep; + job->fifo.item[idx].bad = 0; + job->fifo.item[idx].esize = newesize; + + /* Add the entry size to total fifo size */ + job->fifo.c_bsize += ep->ep_entry? job->fifo.item[idx].esize : 0; + + /* Update the job to show our progress */ + job->lead_ID = id; + if ((id - job->starting_ID) <= job->fifo.size) { + job->trailing_ID = job->starting_ID; + } else { + job->trailing_ID = id - job->fifo.size; + } + + PR_Unlock(job->wire_lock); + return 0; +} + +void *factory_constructor(void *object, void *parent) +{ + return NULL; +} + +void factory_destructor(void *extension, void *object, void *parent) +{ + ImportJob *job = (ImportJob *)extension; + PRThread *thread; + + if (extension == NULL) + return; + + /* connection was destroyed while we were still storing the extension -- + * this is bad news and means we have a bulk import that needs to be + * aborted! + */ + thread = job->main_thread; + LDAPDebug(LDAP_DEBUG_ANY, "ERROR bulk import abandoned\n", + 0, 0, 0); + import_abort_all(job, 1); + /* wait for import_main to finish... */ + PR_JoinThread(thread); + /* extension object is free'd by import_main */ + return; +} + +/* plugin entry function for replica init */ +int ldbm_back_wire_import(Slapi_PBlock *pb) +{ + struct ldbminfo *li; + backend *be = NULL; + ImportJob *job; + PRThread *thread; + int state; + + slapi_pblock_get(pb, SLAPI_BACKEND, &be); + PR_ASSERT(be != NULL); + li = (struct ldbminfo *)(be->be_database->plg_private); + slapi_pblock_get(pb, SLAPI_BULK_IMPORT_STATE, &state); + if (state == SLAPI_BI_STATE_START) { + /* starting a new import */ + return bulk_import_start(pb); + } + + PR_ASSERT(pb->pb_conn != NULL); + if (pb->pb_conn != NULL) { + job = (ImportJob *)slapi_get_object_extension(li->li_bulk_import_object, pb->pb_conn, li->li_bulk_import_handle); + } + + if ((job == NULL) || (pb->pb_conn == NULL)) { + /* import might be aborting */ + return -1; + } + + if (state == SLAPI_BI_STATE_ADD) { + /* continuing previous import */ + if (! import_entry_belongs_here(pb->pb_import_entry, + job->inst->inst_be)) { + /* silently skip */ + return 0; + } + return bulk_import_queue(job, pb->pb_import_entry, + job->flags & FLAG_USE_FILES); + } + + thread = job->main_thread; + + if (state == SLAPI_BI_STATE_DONE) { + /* finished with an import */ + job->flags |= FLAG_PRODUCER_DONE; + /* "job" struct may vanish at any moment after we set the DONE + * flag, so keep a copy of the thread id in 'thread' for safekeeping. + */ + /* wait for import_main to finish... */ + PR_JoinThread(thread); + slapi_set_object_extension(li->li_bulk_import_object, pb->pb_conn, + li->li_bulk_import_handle, NULL); + return 0; + } + + /* ??? unknown state */ + LDAPDebug(LDAP_DEBUG_ANY, + "ERROR: ldbm_back_wire_import: unknown state %d\n", + state, 0, 0); + return -1; +} + +/* + * backup index configuration + * this function is called from dblayer_backup (ldbm2archive) + * [547427] index config must not change between backup and restore + */ +#define DSE_INDEX "dse_index.ldif" +#define DSE_INSTANCE "dse_instance.ldif" +#define DSE_INDEX_FILTER "(objectclass=nsIndex)" +#define DSE_INSTANCE_FILTER "(objectclass=nsBackendInstance)" +static int +dse_conf_backup_core(struct ldbminfo *li, char *dest_dir, char *file_name, char *filter) +{ + Slapi_PBlock *srch_pb = NULL; + Slapi_Entry **entries = NULL; + Slapi_Entry **ep = NULL; + Slapi_Attr *attr = NULL; + char *attr_name; + char *filename = NULL; + PRFileDesc *prfd = NULL; + int rval = 0; + int dlen = 0; + PRInt32 prrval; + char tmpbuf[BUFSIZ]; + char *tp = NULL; + + dlen = strlen(dest_dir); + if (0 == dlen) + { + filename = file_name; + } + else + { + filename = (char *)slapi_ch_malloc(strlen(file_name) + dlen + 2); + sprintf(filename, "%s/%s", dest_dir, file_name); + } + LDAPDebug(LDAP_DEBUG_TRACE, "dse_conf_backup(%s): backup file %s\n", + filter, filename, 0); + + /* Open the file to write */ + if ((prfd = PR_Open(filename, PR_RDWR | PR_CREATE_FILE | PR_TRUNCATE, + SLAPD_DEFAULT_FILE_MODE)) == NULL) + { + LDAPDebug(LDAP_DEBUG_ANY, + "dse_conf_backup(%s): open %s failed: (%s)\n", + filter, filename, slapd_pr_strerror(PR_GetError())); + rval = -1; + goto out; + } + + srch_pb = slapi_pblock_new(); + slapi_search_internal_set_pb(srch_pb, li->li_plugin->plg_dn, + LDAP_SCOPE_SUBTREE, filter, NULL, 0, NULL, NULL, li->li_identity, 0); + slapi_search_internal_pb(srch_pb); + slapi_pblock_get(srch_pb, SLAPI_PLUGIN_INTOP_SEARCH_ENTRIES, &entries); + for (ep = entries; ep != NULL && *ep != NULL; ep++) + { + size_t l = strlen(slapi_entry_get_dn_const(*ep)) + 5 /* "dn: \n" */; + LDAPDebug(LDAP_DEBUG_TRACE, "\ndn: %s\n", + slapi_entry_get_dn_const(*ep), 0, 0); + + if (l <= BUFSIZ) + tp = tmpbuf; + else + tp = (char *)slapi_ch_malloc(l); /* should be very rare ... */ + sprintf(tp, "dn: %s\n", slapi_entry_get_dn_const(*ep)); + prrval = PR_Write(prfd, tp, l); + if ((size_t)prrval != l) + { + LDAPDebug(LDAP_DEBUG_ANY, + "dse_conf_backup(%s): write %s failed: %d (%s)\n", + filter, PR_GetError(), slapd_pr_strerror(PR_GetError())); + rval = -1; + if (l > BUFSIZ) + slapi_ch_free_string(&tp); + goto out; + } + if (l > BUFSIZ) + slapi_ch_free_string(&tp); + + for (slapi_entry_first_attr(*ep, &attr); attr; + slapi_entry_next_attr(*ep, attr, &attr)) + { + int i; + Slapi_Value *sval = NULL; + const struct berval *attr_val; + int attr_name_len; + + slapi_attr_get_type(attr, &attr_name); + /* numsubordinates should not be backed up */ + if (!strcasecmp("numsubordinates", attr_name)) + continue; + attr_name_len = strlen(attr_name); + for (i = slapi_attr_first_value(attr, &sval); i != -1; + i = slapi_attr_next_value(attr, i, &sval)) + { + attr_val = slapi_value_get_berval(sval); + l = strlen(attr_val->bv_val) + attr_name_len + 3; /* : \n" */ + LDAPDebug(LDAP_DEBUG_TRACE, "%s: %s\n", attr_name, + attr_val->bv_val, 0); + if (l <= BUFSIZ) + tp = tmpbuf; + else + tp = (char *)slapi_ch_malloc(l); + sprintf(tp, "%s: %s\n", attr_name, attr_val->bv_val); + prrval = PR_Write(prfd, tp, l); + if ((size_t)prrval != l) + { + LDAPDebug(LDAP_DEBUG_ANY, + "dse_conf_backup(%s): write %s failed: %d (%s)\n", + filter, PR_GetError(), slapd_pr_strerror(PR_GetError())); + rval = -1; + if (l > BUFSIZ) + slapi_ch_free_string(&tp); + goto out; + } + if (l > BUFSIZ) + slapi_ch_free_string(&tp); + } + } + if (ep+1 != NULL && *(ep+1) != NULL) + { + prrval = PR_Write(prfd, "\n", 1); + if ((int)prrval != 1) + { + LDAPDebug(LDAP_DEBUG_ANY, + "dse_conf_backup(%s): write %s failed: %d (%s)\n", + filter, PR_GetError(), slapd_pr_strerror(PR_GetError())); + rval = -1; + goto out; + } + } + } + +out: + slapi_free_search_results_internal(srch_pb); + if (srch_pb) + { + slapi_pblock_destroy(srch_pb); + } + + if (0 != dlen) + { + slapi_ch_free_string(&filename); + } + + if (prfd) + { + prrval = PR_Close(prfd); + if (PR_SUCCESS != prrval) + { + LDAPDebug( LDAP_DEBUG_ANY, + "Fatal Error---Failed to back up dse indexes %d (%s)\n", + PR_GetError(), slapd_pr_strerror(PR_GetError()), 0); + rval = -1; + } + } + + return rval; +} + +int +dse_conf_backup(struct ldbminfo *li, char *dest_dir) +{ + int rval = 0; + rval = dse_conf_backup_core(li, dest_dir, DSE_INSTANCE, DSE_INSTANCE_FILTER); + rval += dse_conf_backup_core(li, dest_dir, DSE_INDEX, DSE_INDEX_FILTER); + return rval; +} + +/* + * read the backed up index configuration + * adjust them if the current configuration is different from it. + * this function is called from dblayer_restore (archive2ldbm) + * these functions are placed here to borrow import_get_entry + * [547427] index config must not change between backup and restore + */ +int +dse_conf_verify_core(struct ldbminfo *li, char *src_dir, char *file_name, char *filter, char *log_str) +{ + char *filename = NULL; + int rval = 0; + ldif_context c; + int fd = -1; + int curr_lineno = 0; + int finished = 0; + int backup_entry_len = 256; + Slapi_Entry **backup_entries = NULL; + Slapi_Entry **bep = NULL; + Slapi_Entry **curr_entries = NULL; + Slapi_PBlock srch_pb; + + filename = (char *)slapi_ch_malloc(strlen(file_name) + strlen(src_dir) + 2); + sprintf(filename, "%s/%s", src_dir, file_name); + + if (PR_SUCCESS != PR_Access(filename, PR_ACCESS_READ_OK)) + { + LDAPDebug(LDAP_DEBUG_ANY, + "Warning: config backup file %s not found in backup\n", + file_name, 0, 0); + rval = 0; + goto out; + } + + fd = dblayer_open_huge_file(filename, O_RDONLY, 0); + if (fd < 0) + { + LDAPDebug(LDAP_DEBUG_ANY, + "Warning: can't open config backup file: %s\n", filename, 0, 0); + rval = -1; + goto out; + } + + import_init_ldif(&c); + bep = backup_entries = (Slapi_Entry **)slapi_ch_calloc(1, + backup_entry_len * sizeof(Slapi_Entry *)); + + while (!finished) + { + char *estr = NULL; + Slapi_Entry *e = NULL; + estr = import_get_entry(&c, fd, &curr_lineno); + + if (!estr) + break; + + e = slapi_str2entry(estr, 0); + slapi_ch_free_string(&estr); + if (!e) { + LDAPDebug(LDAP_DEBUG_ANY, "WARNING: skipping bad LDIF entry " + "ending line %d of file \"%s\"", curr_lineno, filename, 0); + continue; + } + if (bep - backup_entries >= backup_entry_len) + { + backup_entries = (Slapi_Entry **)slapi_ch_realloc((char *)backup_entries, + 2 * backup_entry_len * sizeof(Slapi_Entry *)); + bep = backup_entries + backup_entry_len; + backup_entry_len *= 2; + } + *bep = e; + bep++; + } + // 623986: terminate the list if we reallocated backup_entries + if (backup_entry_len > 256) + *bep = NULL; + + pblock_init(&srch_pb); + slapi_search_internal_set_pb(&srch_pb, li->li_plugin->plg_dn, + LDAP_SCOPE_SUBTREE, filter, NULL, 0, NULL, NULL, li->li_identity, 0); + slapi_search_internal_pb(&srch_pb); + slapi_pblock_get(&srch_pb, SLAPI_PLUGIN_INTOP_SEARCH_ENTRIES, &curr_entries); + + if (0 != slapi_entries_diff(backup_entries, curr_entries, 1 /* test_all */, + log_str, 1 /* force_update */, li->li_identity)) + { + LDAPDebug(LDAP_DEBUG_ANY, "WARNING!!: current %s is " + "different from backed up configuration; " + "The backup is restored.\n", log_str, 0, 0); + } + + slapi_free_search_results_internal(&srch_pb); + pblock_done(&srch_pb); + import_free_ldif(&c); +out: + for (bep = backup_entries; bep && *bep; bep++) + slapi_entry_free(*bep); + slapi_ch_free((void **)&backup_entries); + + slapi_ch_free_string(&filename); + + if (fd > 0) + close(fd); + + return rval; +} + +int +dse_conf_verify(struct ldbminfo *li, char *src_dir) +{ + int rval; + rval = dse_conf_verify_core(li, src_dir, DSE_INSTANCE, DSE_INSTANCE_FILTER, + "Instance Config"); + rval += dse_conf_verify_core(li, src_dir, DSE_INDEX, DSE_INDEX_FILTER, + "Index Config"); + return rval; +} |