Ralph,
Thanks for posting your issue, although I'm sorry to hear you're having such problems.
I suspect the problem is with your interface (ATA, SATA, FireWire, ...). The problem isn't the archive, it's that QRecall can't reliably read data from the device(s) and those devices are randomly dropping offline.
Let's take your issues one at a time:
I seem to have an archive that is beyond repair.
Technically, that's virtually impossible.
QRecall archives are carefully crafted so that no one incident of lost data will corrupt the entire archive. An archive is, essentially, a huge set of individual data records. The repair process really does nothing more than read the entire record set from start to finish, and notes which data records are valid and intact. It then creates a new index that stitches the valid records back together into a functioning archive and erases everything else. For an archive to be damaged "beyond repair" it would require that almost
every single data record in the entire archive was either damaged or unreadable. Unless your hard drive was driven over with a truck, that's seems unlikely.
Several weeks ago the failures got worse, with the drive regularly unmounting itself during a backup.
I think that this is the core problem, but I can't tell you what's causing it.
When I tried to open the archive qrecall would tell me it needed to be reindexed
That's simply because the archive had already been damaged, and had never been successfully repaired (for one reason or another). QRecall can't browse an unindexed archive.
and when I tried to repair it the drive would unmount itself part way through.
QRecall will not cause your drive to unmount itself. But a drive that does unmount itself will certainly tank whatever QRecall action is in progress.
Well, that first statement isn't entirely true. QRecall reads data from files using the same OS functions that every other application you own uses. One significant difference is that QRecall reads and write
a whole lot of data. Applications that read a few megabytes aren't likely to encounter a problem, but read a few gigabytes and you'll eventually run into something. So it's possible for QRecall to exacerbate an existing problem, but it can't cause it.
Looking at your log files, that's exactly what it appears is happening here. The last report you attempted to perform shows a pattern of data read failures, timeouts, and volumes spontaneously unmounting.
2011-08-22 18:05:12.708 -0700 ------- Repair bu-archive.quanta
QRecall starts the repair process at 18:08 by reading the primary archive file.
Everything seems to run fine for several hours, until ...
2011-08-22 18:20:08.259 -0700 Caution Data problems found
2011-08-22 18:20:08.259 -0700 Warning Invalid data
2011-08-22 18:20:08.260 -0700 Details 32784 bytes at 12817714648
2011-08-22 18:20:08.756 -0700 Details 32784 bytes at 12822828952
2011-08-22 18:20:09.532 -0700 Details 24880 bytes at 12839047576
2011-08-22 18:20:13.020 -0700 Details 32784 bytes at 12935966552
2011-08-22 18:20:14.038 -0700 Details 32784 bytes at 12957223328
a burst of data errors are encountered. QRecall is either getting corrupted data during the read process, or the data in these records was previously scrambled by some other problem.
Twenty seconds after this string of data errors, the volume unmounts itself:
2011-08-22 18:46:53.872 -0700 #debug# DiskDisappeared /dev/disk2s3
2011-08-22 18:46:53.875 -0700 #debug# removed volume 'spare'
2011-08-22 18:47:03.897 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 18:47:03.901 -0700 #debug# unmounted volume /Volumes/spare
A few minutes later, the following is logged:
2011-08-22 19:17:06.069 -0700 Caution listener threw exception in message progressUpdate:
2011-08-22 19:17:06.069 -0700 Details NSDistantObject (0x1023012a0) is invalid (no connection)
2011-08-22 19:17:06.069 -0700 #debug# NSObjectInaccessibleException exception
These messages tell me that the communications link between the repair process and the user interface timed out. This usually means that one of processes terminated abnormally, but it can also happen if one of the processes stop communicating for more than a couple of minutes. That tells me that the repair process was likely "hung" waiting for the drive to return data.
I suspect that the drive and/or interface is having problems reading some data. It either loses connection with the drive (causing hung processes) or panics and unmounts the volume, or both. Two more bits of information in the log infer this. First, is the string of volume mount and unmount events logged:
2011-08-22 17:28:03.347 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:29:12.919 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:29:27.452 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:29:27.617 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:29:27.725 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:32:24.371 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:32:50.309 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:32:50.380 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:33:18.777 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:34:06.202 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:34:19.468 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:34:27.942 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:34:32.833 -0700 #debug# volume /Volumes/FantomHD unmount notification
2011-08-22 17:34:32.907 -0700 #debug# volume /Volumes/MBPclone unmount notification
2011-08-22 17:38:58.680 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:39:17.792 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:39:33.070 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:39:33.236 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:39:33.350 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:40:28.764 -0700 #debug# volume /Volumes/rstrauch mount notification
2011-08-22 17:41:24.718 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:43:07.119 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:43:55.041 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:44:09.609 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:46:20.231 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 17:47:14.190 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 17:48:40.586 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 18:01:55.406 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 18:47:03.897 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-22 20:42:27.107 -0700 #debug# volume /Volumes/spare mount notification
2011-08-22 20:43:50.591 -0700 #debug# volume /Volumes/spare unmount notification
2011-08-23 01:36:29.097 -0700 #debug# volume /Volumes/spare1 mount notification
2011-08-23 08:53:02.072 -0700 #debug# volume /Volumes/spare1 unmount notification
2011-08-23 10:05:30.332 -0700 #debug# volume /Volumes/Files mount notification
2011-08-23 10:38:14.210 -0700 #debug# volume /Volumes/spare1 mount notification
While many of these are obviously normal activity, particularly since you said you were moving volumes around, some are clearly not. Of particular note are the mount events that follow unmount events by a fraction of a second. This is not normal and indicates that the volume spontaneously unmounted itself and then immediately remounted.
Your log is also full of -35 (Volume not found) and -43 (Directory not found) errors, indicating the the archive's volume was not mounted at the time the action attempted to do something.
Other files on the drive copied properly, but part way through the archive copy the source drive dismounted and left and incomplete file on the destination -- just as it had done during repair attempts.
This tells me that the Finder is encountering exactly the same problem that the QRecall repair action did: it was reading the file just fine, then it encountered data read errors, then the drive unmounted. This points to a problem with the drive, the controllers, or the I/O subsystem.
If it becomes possible to recover the data that's on the drive(s), I suspect that the bulk of the archive is fine, could be repaired, and eventually put back into service.