- Fixed pricing on recovery (You know what you are paying - no nasty surprises).
- Quick recovery turnaround at no extra cost. (Our average recovery time is 2 days).
- Memory card chip reading services (1st in the UK to offer this service).
- Raid recoding service (Specialist service for our business customers who have suffered a failed server rebuild).
- Our offices are 100% UK based and we never outsource any recovery work.
- Strict Non-disclosure privacy and security is 100% guaranteed.
Case Studies
Case Studies
We have helped a wide range of clients over the last 25 years and reasons as to the whys and wherefores of the loss of their data. To that end we have a broad knowledge of problems relating to data recovery as our case studies show.
Case Study 1 — iMac G4 Not Booting (Question-Mark Folder)
Issue (Client Impact)
An accountant’s iMac G4 failed to boot. Instead of reaching macOS, the machine displayed the classic flashing folder with a question mark (and intermittent “sad face”)—an indicator that the Mac could not locate a valid system folder or a readable boot device. The workstation contained 52 client datasets (Excel workbooks and Sage accounting files) required urgently.
Initial Triage & Diagnostics
-
Intake & Write-block: Drive removed from the iMac and connected via a forensic write-blocker to prevent inadvertent changes.
-
Drive identity: 3.5″ PATA/IDE HDD (typical for iMac G4 generation). No stable enumerate on first power-up; motor spun, but no ID over ATA.
-
Electronics check: Measured rail resistance and inrush current; 5 V rail abnormal. Visual inspection showed heat-stressed motor driver and a scorched TVS footprint.
-
Firmware access: Service-port access attempted; device would not respond—consistent with PCB/controller failure.
-
Conclusion: Electronic failure of the PCB; mechanical subsystem (spindle/heads) appeared sound (no abnormal noises, normal spin-up profile).
Recovery Actions
-
Donor PCB selection & unique ROM/adaptives transfer
-
Sourced a donor PCB with matching board revision and microcode from in-house inventory.
-
Read the original board’s ROM/NVRAM (contains adaptive calibration data unique to the head-disk assembly).
-
Transferred ROM to donor PCB (hot-air rework) so that the donor electronics matched the platter/head adaptives.
-
-
Controlled power-up & firmware sanity checks
-
After PCB swap, drive enumerated reliably in Device-Safe mode.
-
Verified SMART without allowing automatic background processes.
-
Locked background scan, disabled idle-time write/cache features.
-
-
Hardware-assisted imaging (PC-3000)
-
Built a head map and confirmed no weak head behavior.
-
Performed multi-pass imaging (soft→hard) with tuned timeouts and LBA windowing to avoid repeated retries.
-
Imaged 100% of user LBAs to a sterile target.
-
-
Filesystem & data extraction
-
Detected an HFS+ (Journaled) volume.
-
Replayed the journal, checked catalog/extent B-trees, repaired minor directory issues.
-
Extracted Excel and Sage data; verified Sage company datasets open properly and Excel hashes match expected totals.
-
-
Delivery
-
As the client needed immediate access, we provisioned a secure download (encrypted link). Chain-of-custody and hash manifests provided.
-
Outcome
-
100% data recovered.
-
Client resumed operations the same day using the secure download set.
Note: The “question-mark folder” on older Macs commonly maps to storage not detected / unbootable rather than OS corruption alone. Electronic failures on vintage PATA drives are frequent; success hinges on retaining the original ROM/adaptives when replacing the PCB.
Case Study 2 — Failed Rebuild on a Buffalo LinkStation (RAID 5)
Issue (Client Impact)
A Buffalo LinkStation that had served for years as an office file server suddenly disappeared from the network share list. The web UI only showed “Users” → “Local Users”; shares and volumes were missing. Client later disclosed they attempted a rebuild which failed at ~2%. The array held active departmental documents.
Array & Filesystem Topology (as Received)
-
Enclosure: Buffalo LinkStation (4-bay)
-
RAID level: RAID 5, 4 × 750 GB SATA HDDs
-
OS stack (typical): Linux mdadm layer with XFS or ext3 filesystem on top (confirmed during analysis)
Diagnostics
-
SMART screening: Two members exhibited reallocated/pending sector escalation and high UDMA CRC event counts.
-
On-disk metadata: mdadm superblocks indicated previous degradations and a recent rebuild attempt. Parity and data were out of sync (classic “write-hole” effect compounded by the failed rebuild).
-
Risk: Continuing controller-level rebuilds would likely amplify corruption by writing new parity over stale data.
Recovery Actions
-
Per-disk hardware imaging
-
Used PC-3000/Atola to image each disk independently.
-
Soft→hard imaging strategy with head-map selection and skip/late-fill for unreadable regions.
-
Captured full images; flagged bad LBA ranges for later parity reasoning.
-
-
Geometry discovery & validation
-
Extracted mdadm superblocks to confirm RAID 5 geometry (block/stripe size, rotation pattern, start offset).
-
Validated with stripe-alignment tests against known filesystem signatures (superblocks, inode structures).
-
-
Virtual array reconstruction (no writes to originals)
-
Reassembled array in software using the disk images, not the hardware controller.
-
Where one member had unreadable sectors, computed the missing blocks from the parity of the remaining members (when possible).
-
-
Filesystem repair
-
Mounted the reconstructed volume read-only; detected XFS metadata inconsistencies likely introduced during the failed rebuild.
-
Executed a metadata repair workflow (log replay, directory/inode btree checks), then exported user data to a sterile target.
-
-
Delivery & validation
-
Supplied a 4 TB external HDD with the restored directory tree.
-
Provided a report with hash manifests and a list of previously unreadable LBAs (all covered by parity reconstruction, no user-file loss).
-
Outcome
-
Full logical recovery of the share set.
-
Client re-published the shares from a new NAS, using our copy as the seed.
Note: On mdadm-based NAS, if a disk drops and another is marginal, controller-level rebuilds can quickly convert a “recoverable” state into a parity/data inconsistency. Best practice is to image first, then reconstruct virtually.
Case Study 3 — HP Server: Post-Rebuild Boot Failure on RAID 5 (8+2)
Issue (Client Impact)
A factory file server (HP rack server) hosted ~2 TB of live data for ~100 staff on a RAID 5 array (8 data drives + 2 hot spares on a Smart Array controller). The server froze, was power-cycled, and reported one failed drive. The hot spare did not auto-promote, so IT manually initiated a rebuild to one hot spare. The rebuild reached 100%, but the server would not boot afterwards. Their MSP attempted recovery and referred the case to us.
As-Received Condition
-
10 drives total (8 active members, 2 hot spares) delivered loose.
-
Controller logs unavailable; several disks showed recent power-on resets and re-allocations.
-
By the time the set reached our lab, two additional drives displayed degraded behavior—effectively 3 impaired members in the original 8-drive RAID 5.
In RAID 5, any period with >1 unavailable member causes irrecoverable stripes unless the failing members can still yield partial readable sectors during lab imaging.
Diagnostics
-
SMART & surface: Three drives presented distinct failure modes:
-
Disk A: escalating pending sectors and slow seeks (weak heads).
-
Disk B: read channel timeouts (probable preamp/head issue).
-
Disk C: controller/firmware timeouts under load.
-
-
On-disk metadata: Smart Array headers indicated recent configuration changes matching the forced rebuild.
-
Filesystem on top: NTFS on a single logical volume (validated later).
Recovery Actions
-
Stabilisation & per-disk imaging
-
Isolated each member to a dedicated imager.
-
Head-map imaging on Disk A (skipping failing head ranges, back-filling later).
-
Adaptive timeouts & power-cycle strategy on Disk B to capture intermittent bands.
-
Vendor-specific quiesce on Disk C to limit background firmware tasks interfering with reads.
-
Result: High-coverage images for all eight members; bad-block maps retained.
-
-
Targeted mechanical/electronics service (on the worst offenders)
-
Performed head-stack replacements on two drives exhibiting clear read-channel/preamp faults (using matched donors; unique ROM/adaptive data preserved).
-
One PCB driver issue corrected via donor board + ROM transfer to stabilise firmware access.
-
Post-service imaging improved coverage enough to mathematically reconstruct previously missing stripes.
-
-
Array geometry reconstruction
-
Derived stripe size and parity rotation from controller metadata and content analysis.
-
Normalised capacity where one image reported a truncated LBA span (HPA/DCO artifact).
-
Built a virtual RAID 5 from the eight images; computed missing blocks in stripes where only a single member’s data was absent.
-
Where two members lacked the same stripe region, we merged best-of sectors from those partially imaged drives to reduce the net “unknown” to ≤1 block per stripe, allowing parity to resolve.
-
-
Filesystem repair & extraction
-
Mounted reconstructed NTFS read-only; replayed $LogFile.
-
Repaired $MFT/$MFTMirr mismatches; validated ACLs and key shares.
-
Exported the full 2 TB dataset to new storage; generated hash manifests for acceptance testing.
-
-
Delivery & advisory
-
Clean export provided with a concise incident/recovery report (timeline, drive serials, imaging coverage, irrecoverable sector list = none impacting user data).
-
Recommended: controller firmware update, removal of marginal drives, and pre-production resiliency test before going live.
-
Outcome
-
Complete logical recovery with no material file loss.
-
Client restored services on replacement hardware using our dataset.
Note: It’s common for a stressed array to experience secondary failures during or after a rebuild. The correct lab approach is per-disk imaging first, then virtual reconstruction—never rebuild on the originals.
Closing (Service Notes)
-
For urgent cases, we operate a critical path workflow that prioritises stabilisation → imaging → virtual rebuild → extraction with engineer-to-engineer communication.
-
For shipments: place the drive(s)/NAS in anti-static bags, bubble-wrap, and a padded box, include your contact details and incident description. Avoid further power-ons.
