Abstract
The costs of data loss and unavailability can be large, so businesses use many data protection techniques such as remote mirroring, snapshots, and backups to guard against failures. Choosing an appropriate combination of techniques is difficult because there are numerous approaches for protecting data and allocating resources. Storage system architects typically use ad hoc techniques, often resulting in overengineered expensive solutions or underprovisioned inadequate ones. In contrast, this paper presents a principled automated approach for designing dependable storage solutions for multiple applications in shared environments. Our contributions include search heuristics for intelligent exploration of the large design space and modeling techniques for capturing interactions between applications during recovery. Using realistic storage system requirements, we show that our design tool produces designs that cost up to two times less in initial outlays and expected data penalties than the designs produced by an emulated human design process. Additionally, we compare our design tool to a random search heuristic and a genetic algorithm metaheuristic, and show that our approach consistently produces better designs for the cases we have studied. Finally, we study the sensitivity of our design tool to several input parameters.
Original language | English (US) |
---|---|
Article number | 4586394 |
Pages (from-to) | 366-380 |
Number of pages | 15 |
Journal | IEEE Transactions on Dependable and Secure Computing |
Volume | 7 |
Issue number | 4 |
DOIs | |
State | Published - 2010 |
Keywords
- Backup procedures
- Data protection techniques
- Maintenance
- Reliability
- Secondary storage
- Testing
- and Fault-Tolerance
- design space exploration
- discrete-event simulation
- genetic algorithm
- search heuristic
- storage system design
ASJC Scopus subject areas
- Electrical and Electronic Engineering