fbpx
Wikipedia

Backup

In information technology, a backup, or data backup is a copy of computer data taken and stored elsewhere so that it may be used to restore the original after a data loss event. The verb form, referring to the process of doing so, is "back up", whereas the noun and adjective form is "backup".[1] Backups can be used to recover data after its loss from data deletion or corruption, or to recover data from an earlier time.[2] Backups provide a simple form of disaster recovery; however not all backup systems are able to reconstitute a computer system or other complex configuration such as a computer cluster, active directory server, or database server.[3]

A backup system contains at least one copy of all data considered worth saving. The data storage requirements can be large. An information repository model may be used to provide structure to this storage. There are different types of data storage devices used for copying backups of data that is already in secondary storage onto archive files.[note 1][4] There are also different ways these devices can be arranged to provide geographic dispersion, data security, and portability.

Data is selected, extracted, and manipulated for storage. The process can include methods for dealing with live data, including open files, as well as compression, encryption, and de-duplication. Additional techniques apply to enterprise client-server backup. Backup schemes may include dry runs that validate the reliability of the data being backed up. There are limitations[5] and human factors involved in any backup scheme.

Storage edit

A backup strategy requires an information repository, "a secondary storage space for data"[6] that aggregates backups of data "sources". The repository could be as simple as a list of all backup media (DVDs, etc.) and the dates produced, or could include a computerized index, catalog, or relational database.

The backup data needs to be stored, requiring a backup rotation scheme,[4] which is a system of backing up data to computer media that limits the number of backups of different dates retained separately, by appropriate re-use of the data storage media by overwriting of backups no longer needed. The scheme determines how and when each piece of removable storage is used for a backup operation and how long it is retained once it has backup data stored on it. The 3-2-1 rule can aid in the backup process. It states that there should be at least 3 copies of the data, stored on 2 different types of storage media, and one copy should be kept offsite, in a remote location (this can include cloud storage). 2 or more different media should be used to eliminate data loss due to similar reasons (for example, optical discs may tolerate being underwater while LTO tapes may not, and SSDs cannot fail due to head crashes or damaged spindle motors since they do not have any moving parts, unlike hard drives). An offsite copy protects against fire, theft of physical media (such as tapes or discs) and natural disasters like floods and earthquakes. Physically protected hard drives are an alternative to an offsite copy, but they have limitations like only being able to resist fire for a limited period of time, so an offsite copy still remains as the ideal choice.

Backup methods edit

Unstructured edit

An unstructured repository may simply be a stack of tapes, DVD-Rs or external HDDs with minimal information about what was backed up and when. This method is the easiest to implement, but unlikely to achieve a high level of recoverability as it lacks automation.

Full only/System imaging edit

A repository using this backup method contains complete source data copies taken at one or more specific points in time. Copying system images, this method is frequently used by computer technicians to record known good configurations. However, imaging[7] is generally more useful as a way of deploying a standard configuration to many systems rather than as a tool for making ongoing backups of diverse systems.

Incremental edit

An incremental backup stores data changed since a reference point in time. Duplicate copies of unchanged data are not copied. Typically a full backup of all files is made once or at infrequent intervals, serving as the reference point for an incremental repository. Subsequently, a number of incremental backups are made after successive time periods. Restores begin with the last full backup and then apply the incrementals.[8] Some backup systems[9] can create a synthetic full backup from a series of incrementals, thus providing the equivalent of frequently doing a full backup. When done to modify a single archive file, this speeds restores of recent versions of files.

Near-CDP edit

Continuous Data Protection (CDP) refers to a backup that instantly saves a copy of every change made to the data. This allows restoration of data to any point in time and is the most comprehensive and advanced data protection.[10] Near-CDP backup applications—often marketed as "CDP"—automatically take incremental backups at a specific interval, for example every 15 minutes, one hour, or 24 hours. They can therefore only allow restores to an interval boundary.[10] Near-CDP backup applications use journaling and are typically based on periodic "snapshots",[11] read-only copies of the data frozen at a particular point in time.

Near-CDP (except for Apple Time Machine)[12] intent-logs every change on the host system,[13] often by saving byte or block-level differences rather than file-level differences. This backup method differs from simple disk mirroring in that it enables a roll-back of the log and thus a restoration of old images of data. Intent-logging allows precautions for the consistency of live data, protecting self-consistent files but requiring applications "be quiesced and made ready for backup."

Near-CDP is more practicable for ordinary personal backup applications, as opposed to true CDP, which must be run in conjunction with a virtual machine[14][15] or equivalent[16] and is therefore generally used in enterprise client-server backups.

Software may create copies of individual files such as written documents, multimedia projects, or user preferences, to prevent failed write events caused by power outages, operating system crashes, or exhausted disk space, from causing data loss. A common implementation is an appended ".bak" extension to the file name.

Reverse incremental edit

A Reverse incremental backup method stores a recent archive file "mirror" of the source data and a series of differences between the "mirror" in its current state and its previous states. A reverse incremental backup method starts with a non-image full backup. After the full backup is performed, the system periodically synchronizes the full backup with the live copy, while storing the data necessary to reconstruct older versions. This can either be done using hard links—as Apple Time Machine does, or using binary diffs.

Differential edit

A differential backup saves only the data that has changed since the last full backup. This means a maximum of two backups from the repository are used to restore the data. However, as time from the last full backup (and thus the accumulated changes in data) increases, so does the time to perform the differential backup. Restoring an entire system requires starting from the most recent full backup and then applying just the last differential backup.

A differential backup copies files that have been created or changed since the last full backup, regardless of whether any other differential backups have been made since, whereas an incremental backup copies files that have been created or changed since the most recent backup of any type (full or incremental). Changes in files may be detected through a more recent date/time of last modification file attribute, and/or changes in file size. Other variations of incremental backup include multi-level incrementals and block-level incrementals that compare parts of files instead of just entire files.

Storage media edit

 
From left to right, a DVD disc in plastic cover, a USB flash drive and an external hard drive

Regardless of the repository model that is used, the data has to be copied onto an archive file data storage medium. The medium used is also referred to as the type of backup destination.

Magnetic tape edit

Magnetic tape was for a long time the most commonly used medium for bulk data storage, backup, archiving, and interchange. It was previously a less expensive option, but this is no longer the case for smaller amounts of data.[17] Tape is a sequential access medium, so the rate of continuously writing or reading data can be very fast. While tape media itself has a low cost per space, tape drives are typically dozens of times as expensive as hard disk drives and optical drives.

Many tape formats have been proprietary or specific to certain markets like mainframes or a particular brand of personal computer. By 2014 LTO had become the primary tape technology.[18] The other remaining viable "super" format is the IBM 3592 (also referred to as the TS11xx series). The Oracle StorageTek T10000 was discontinued in 2016.[19]

Hard disk edit

The use of hard disk storage has increased over time as it has become progressively cheaper. Hard disks are usually easy to use, widely available, and can be accessed quickly.[18] However, hard disk backups are close-tolerance mechanical devices and may be more easily damaged than tapes, especially while being transported.[20] In the mid-2000s, several drive manufacturers began to produce portable drives employing ramp loading and accelerometer technology (sometimes termed a "shock sensor"),[21][22] and by 2010 the industry average in drop tests for drives with that technology showed drives remaining intact and working after a 36-inch non-operating drop onto industrial carpeting.[23] Some manufacturers also offer 'ruggedized' portable hard drives, which include a shock-absorbing case around the hard disk, and claim a range of higher drop specifications.[23][24][25] Over a period of years the stability of hard disk backups is shorter than that of tape backups.[19][26][20]

External hard disks can be connected via local interfaces like SCSI, USB, FireWire, or eSATA, or via longer-distance technologies like Ethernet, iSCSI, or Fibre Channel. Some disk-based backup systems, via Virtual Tape Libraries or otherwise, support data deduplication, which can reduce the amount of disk storage capacity consumed by daily and weekly backup data.[27][28][29]

Optical storage edit

 
Optical discs are not vulnerable to water, making them likely to survive a flood disaster.

Optical storage uses lasers to store and retrieve data. Recordable CDs, DVDs, and Blu-ray Discs are commonly used with personal computers and are generally cheap. The capacities and speeds of these discs have typically been lower than hard disks or tapes. Advances in optical media may shrink that gap in the future.[30][31]

Potential future data losses caused by gradual media degradation can be predicted by measuring the rate of correctable minor data errors, of which consecutively too many increase the risk of uncorrectable sectors. Support for error scanning varies among optical drive vendors.[32]

Many optical disc formats are WORM type, which makes them useful for archival purposes since the data cannot be changed. Moreover, optical discs are not vulnerable to head crashes, magnetism, imminent water ingress or power surges; and, a fault of the drive typically just halts the spinning.

Optical media is modular; the storage controller is not tied to media itself like with hard drives or flash storage (→flash memory controller), allowing it to be removed and accessed through a different drive. However, recordable media may degrade earlier under long-term exposure to light.[33]

Some optical storage systems allow for cataloged data backups without human contact with the discs, allowing for longer data integrity. A French study in 2008 indicated that the lifespan of typically-sold CD-Rs was 2–10 years,[34] but one manufacturer later estimated the longevity of its CD-Rs with a gold-sputtered layer to be as high as 100 years.[35] Sony's proprietary Optical Disc Archive[18] can in 2016 reach a read rate of 250 MB/s.[36]

Solid-state drive edit

Solid-state drives (SSDs) use integrated circuit assemblies to store data. Flash memory, thumb drives, USB flash drives, CompactFlash, SmartMedia, Memory Sticks, and Secure Digital card devices are relatively expensive for their low capacity, but convenient for backing up relatively low data volumes. A solid-state drive does not contain any movable parts, making it less susceptible to physical damage, and can have huge throughput of around 500 Mbit/s up to 6 Gbit/s. Available SSDs have become more capacious and cheaper.[37][24] Flash memory backups are stable for fewer years than hard disk backups.[19]

Remote backup service edit

Remote backup services or cloud backups involve service providers storing data offsite. This has been used to protect against events such as fires, floods, or earthquakes which could destroy locally stored backups.[38] Cloud-based backup (through services like or similar to Google Drive, and Microsoft OneDrive) provides a layer of data protection.[20] However, the users must trust the provider to maintain the privacy and integrity of their data, with confidentiality enhanced by the use of encryption. Because speed and availability are limited by a user's online connection,[20] users with large amounts of data may need to use cloud seeding and large-scale recovery.

Management edit

Various methods can be used to manage backup media, striking a balance between accessibility, security and cost. These media management methods are not mutually exclusive and are frequently combined to meet the user's needs. Using on-line disks for staging data before it is sent to a near-line tape library is a common example.[39][40]

Online edit

Online backup storage is typically the most accessible type of data storage, and can begin a restore in milliseconds. An internal hard disk or a disk array (maybe connected to SAN) is an example of an online backup. This type of storage is convenient and speedy, but is vulnerable to being deleted or overwritten, either by accident, by malevolent action, or in the wake of a data-deleting virus payload.

Near-line edit

Nearline storage is typically less accessible and less expensive than online storage, but still useful for backup data storage. A mechanical device is usually used to move media units from storage into a drive where the data can be read or written. Generally it has safety properties similar to on-line storage. An example is a tape library with restore times ranging from seconds to a few minutes.

Off-line edit

Off-line storage requires some direct action to provide access to the storage media: for example, inserting a tape into a tape drive or plugging in a cable. Because the data is not accessible via any computer except during limited periods in which they are written or read back, they are largely immune to on-line backup failure modes. Access time varies depending on whether the media are on-site or off-site.

Off-site data protection edit

Backup media may be sent to an off-site vault to protect against a disaster or other site-specific problem. The vault can be as simple as a system administrator's home office or as sophisticated as a disaster-hardened, temperature-controlled, high-security bunker with facilities for backup media storage. A data replica can be off-site but also on-line (e.g., an off-site RAID mirror). Such a replica has fairly limited value as a backup.

Backup site edit

A backup site or disaster recovery center is used to store data that can enable computer systems and networks to be restored and properly configured in the event of a disaster. Some organisations have their own data recovery centres, while others contract this out to a third-party. Due to high costs, backing up is rarely considered the preferred method of moving data to a DR site. A more typical way would be remote disk mirroring, which keeps the DR data as up to date as possible.

Selection and extraction of data edit

A backup operation starts with selecting and extracting coherent units of data. Most data on modern computer systems is stored in discrete units, known as files. These files are organized into filesystems. Deciding what to back up at any given time involves tradeoffs. By backing up too much redundant data, the information repository will fill up too quickly. Backing up an insufficient amount of data can eventually lead to the loss of critical information.[41]

Files edit

  • Copying files: Making copies of files is the simplest and most common way to perform a backup. A means to perform this basic function is included in all backup software and all operating systems.
  • Partial file copying: A backup may include only the blocks or bytes within a file that have changed in a given period of time. This can substantially reduce needed storage space, but requires higher sophistication to reconstruct files in a restore situation. Some implementations require integration with the source file system.
  • Deleted files: To prevent the unintentional restoration of files that have been intentionally deleted, a record of the deletion must be kept.
  • Versioning of files: Most backup applications, other than those that do only full only/System imaging, also back up files that have been modified since the last backup. "That way, you can retrieve many different versions of a given file, and if you delete it on your hard disk, you can still find it in your [information repository] archive."[4]

Filesystems edit

  • Filesystem dump: A copy of the whole filesystem in block-level can be made. This is also known as a "raw partition backup" and is related to disk imaging. The process usually involves unmounting the filesystem and running a program like dd (Unix).[42] Because the disk is read sequentially and with large buffers, this type of backup can be faster than reading every file normally, especially when the filesystem contains many small files, is highly fragmented, or is nearly full. But because this method also reads the free disk blocks that contain no useful data, this method can also be slower than conventional reading, especially when the filesystem is nearly empty. Some filesystems, such as XFS, provide a "dump" utility that reads the disk sequentially for high performance while skipping unused sections. The corresponding restore utility can selectively restore individual files or the entire volume at the operator's choice.[43]
  • Identification of changes: Some filesystems have an archive bit for each file that says it was recently changed. Some backup software looks at the date of the file and compares it with the last backup to determine whether the file was changed.
  • Versioning file system: A versioning filesystem tracks all changes to a file. The NILFS versioning filesystem for Linux is an example.[44]

Live data edit

Files that are actively being updated present a challenge to back up. One way to back up live data is to temporarily quiesce them (e.g., close all files), take a "snapshot", and then resume live operations. At this point the snapshot can be backed up through normal methods.[45] A snapshot is an instantaneous function of some filesystems that presents a copy of the filesystem as if it were frozen at a specific point in time, often by a copy-on-write mechanism. Snapshotting a file while it is being changed results in a corrupted file that is unusable. This is also the case across interrelated files, as may be found in a conventional database or in applications such as Microsoft Exchange Server.[11] The term fuzzy backup can be used to describe a backup of live data that looks like it ran correctly, but does not represent the state of the data at a single point in time.[46]

Backup options for data files that cannot be or are not quiesced include:[47]

  • Open file backup: Many backup software applications undertake to back up open files in an internally consistent state.[48] Some applications simply check whether open files are in use and try again later.[45] Other applications exclude open files that are updated very frequently.[49] Some low-availability interactive applications can be backed up via natural/induced pausing.
  • Interrelated database files backup: Some interrelated database file systems offer a means to generate a "hot backup"[50] of the database while it is online and usable. This may include a snapshot of the data files plus a snapshotted log of changes made while the backup is running. Upon a restore, the changes in the log files are applied to bring the copy of the database up to the point in time at which the initial backup ended.[51] Other low-availability interactive applications can be backed up via coordinated snapshots. However, genuinely-high-availability interactive applications can be only be backed up via Continuous Data Protection.

Metadata edit

Not all information stored on the computer is stored in files. Accurately recovering a complete system from scratch requires keeping track of this non-file data too.[52]

  • System description: System specifications are needed to procure an exact replacement after a disaster.
  • Boot sector: The boot sector can sometimes be recreated more easily than saving it. It usually isn't a normal file and the system won't boot without it.
  • Partition layout: The layout of the original disk, as well as partition tables and filesystem settings, is needed to properly recreate the original system.
  • File metadata: Each file's permissions, owner, group, ACLs, and any other metadata need to be backed up for a restore to properly recreate the original environment.
  • System metadata: Different operating systems have different ways of storing configuration information. Microsoft Windows keeps a registry of system information that is more difficult to restore than a typical file.

Manipulation of data and dataset optimization edit

It is frequently useful or required to manipulate the data being backed up to optimize the backup process. These manipulations can improve backup speed, restore speed, data security, media usage and/or reduced bandwidth requirements.

Automated data grooming edit

Out-of-date data can be automatically deleted, but for personal backup applications—as opposed to enterprise client-server backup applications where automated data "grooming" can be customized—the deletion[note 2][53][54] can at most[55] be globally delayed or be disabled.[56]

Compression edit

Various schemes can be employed to shrink the size of the source data to be stored so that it uses less storage space. Compression is frequently a built-in feature of tape drive hardware.[57]

Deduplication edit

Redundancy due to backing up similarly configured workstations can be reduced, thus storing just one copy. This technique can be applied at the file or raw block level. This potentially large reduction[57] is called deduplication. It can occur on a server before any data moves to backup media, sometimes referred to as source/client side deduplication. This approach also reduces bandwidth required to send backup data to its target media. The process can also occur at the target storage device, sometimes referred to as inline or back-end deduplication.

Duplication edit

Sometimes backups are duplicated to a second set of storage media. This can be done to rearrange the archive files to optimize restore speed, or to have a second copy at a different location or on a different storage medium—as in the disk-to-disk-to-tape capability of Enterprise client-server backup.

Encryption edit

High-capacity removable storage media such as backup tapes present a data security risk if they are lost or stolen.[58] Encrypting the data on these media can mitigate this problem, however encryption is a CPU intensive process that can slow down backup speeds, and the security of the encrypted backups is only as effective as the security of the key management policy.[57]

Multiplexing edit

When there are many more computers to be backed up than there are destination storage devices, the ability to use a single storage device with several simultaneous backups can be useful.[59] However cramming the scheduled backup window via "multiplexed backup" is only used for tape destinations.[59]

Refactoring edit

The process of rearranging the sets of backups in an archive file is known as refactoring. For example, if a backup system uses a single tape each day to store the incremental backups for all the protected computers, restoring one of the computers could require many tapes. Refactoring could be used to consolidate all the backups for a single computer onto a single tape, creating a "synthetic full backup". This is especially useful for backup systems that do incrementals forever style backups.

Staging edit

Sometimes backups are copied to a staging disk before being copied to tape.[59] This process is sometimes referred to as D2D2T, an acronym for Disk-to-disk-to-tape. It can be useful if there is a problem matching the speed of the final destination device with the source device, as is frequently faced in network-based backup systems. It can also serve as a centralized location for applying other data manipulation techniques.

Objectives edit

  • Recovery point objective (RPO): The point in time that the restarted infrastructure will reflect, expressed as "the maximum targeted period in which data (transactions) might be lost from an IT service due to a major incident". Essentially, this is the roll-back that will be experienced as a result of the recovery. The most desirable RPO would be the point just prior to the data loss event. Making a more recent recovery point achievable requires increasing the frequency of synchronization between the source data and the backup repository.[60]
  • Recovery time objective (RTO): The amount of time elapsed between disaster and restoration of business functions.[61]
  • Data security: In addition to preserving access to data for its owners, data must be restricted from unauthorized access. Backups must be performed in a manner that does not compromise the original owner's undertaking. This can be achieved with data encryption and proper media handling policies.[62]
  • Data retention period: Regulations and policy can lead to situations where backups are expected to be retained for a particular period, but not any further. Retaining backups after this period can lead to unwanted liability and sub-optimal use of storage media.[62]
  • Checksum or hash function validation: Applications that back up to tape archive files need this option to verify that the data was accurately copied.[63]
  • Backup process monitoring: Enterprise client-server backup applications need a user interface that allows administrators to monitor the backup process, and proves compliance to regulatory bodies outside the organization; for example, an insurance company in the USA might be required under HIPAA to demonstrate that its client data meet records retention requirements.[64]
  • User-initiated backups and restores: To avoid or recover from minor disasters, such as inadvertently deleting or overwriting the "good" versions of one or more files, the computer user—rather than an administrator—may initiate backups and restores (from not necessarily the most-recent backup) of files or folders.

See also edit

About backup

Related topics

Notes edit

  1. ^ In contrast to everyday use of the term "archive", the data stored in an "archive file" is not necessarily old or of historical interest.
  2. ^ Some backup applications—notably rsync and CrashPlan—term removing backup data "pruning" instead of "grooming".

References edit

  1. ^ "back•up". The American Heritage Dictionary of the English Language. Houghton Mifflin Harcourt. 2018. Retrieved 9 May 2018.
  2. ^ S. Nelson (2011). "Chapter 1: Introduction to Backup and Recovery". Pro Data Backup and Recovery. Apress. pp. 1–16. ISBN 978-1-4302-2663-5. Retrieved 8 May 2018.
  3. ^ Cougias, D.J.; Heiberger, E.L.; Koop, K. (2003). "Chapter 1: What's a Disaster Without a Recovery?". The Backup Book: Disaster Recovery from Desktop to Data Center. Network Frontiers. pp. 1–14. ISBN 0-9729039-0-9.
  4. ^ a b c Joe Kissell (2007). (PDF) (Version 2.0 ed.). Ithaca, NY: TidBITS Electronic Publishing. pp. 18–20 ("The Archive", meaning information repository, including versioning), 24 (client-server), 82–83 (archive file), 112–114 (Off-site storage backup rotation scheme), 126–141 (old Retrospect terminology and GUI—still used in Windows variant), 165 (client-server), 128 (subvolume—later renamed Favorite Folder in Macintosh variant). ISBN 978-0-9759503-0-2. Archived from the original (PDF) on 1 December 2020. Retrieved 17 May 2019.
  5. ^ Terry Sullivan (11 January 2018). "A Beginner's Guide to Backing Up Photos". The New York Times. a hard drive ... an established company ... declared bankruptcy ... where many ... had ...
  6. ^ McMahon, Mary (1 April 2019). "What Is an Information Repository?". wiseGEEK. Conjecture Corporation. Retrieved 8 May 2019. In the sense of an approach to data management, an information repository is a secondary storage space for data.
  7. ^ "Five key questions to ask about your backup solution". sysgen.ca. 23 March 2014. Does your company have a low tolerance to longer "data access outages" and/or would you like to minimize the time your company may be without its data?. from the original on 4 March 2016. Retrieved 23 September 2015.
  8. ^ . Tech-FAQ. Independent Media. 13 June 2005. Archived from the original on 21 June 2016. Retrieved 10 March 2006.
  9. ^ Pond, James (31 August 2013). . Apple OSX and Time Machine Tips. baligu.com. File System Event Store, Hard Links. Archived from the original on 21 June 2019. Retrieved 19 May 2019.
  10. ^ a b Behzad Behtash (6 May 2010). "Why Continuous Data Protection's Getting More Practical". Disaster recovery/business continuity. InformationWeek. Retrieved 12 November 2011. A true CDP approach should capture all data writes, thus continuously backing up data and eliminating backup windows.... CDP is the gold standard—the most comprehensive and advanced data protection. But "near CDP" technologies can deliver enough protection for many companies with less complexity and cost. For example, snapshots can provide a reasonable near-CDP-level of protection for file shares, letting users directly access data on the file share at regular intervals--say, every half-hour or 15 minutes. That's certainly a higher level of protection than tape-based or disk-based nightly backups and may be all you need.
  11. ^ a b "Continuous data protection (CDP) explained: True CDP vs near-CDP". ComputerWeekly.com. TechTarget. July 2010. Retrieved 22 June 2019. ... copies data from a source to a target. True CDP does this every time a change is made, while so-called near-CDP does this at pre-set time intervals. Near-CDP is effectively the same as snapshotting....True CDP systems record every write and copy them to the target where all changes are stored in a log. [new paragraph] By contrast, near-CDP/snapshot systems copy files in a straightforward manner but require applications to be quiesced and made ready for backup, either via the application's backup mode or using, for example, Microsoft's Volume Shadow Copy Services (VSS).
  12. ^ Pond, James (31 August 2013). "How Time Machine Works its Magic". Apple OSX and Time Machine Tips. Baligu.com (as mirrored after James Pond died in 2013). Retrieved 10 July 2019. The File System Event Store is a hidden log that OSX keeps on each HFS+ formatted disk/partition of changes made to the data on it. It doesn't list every file that's changed, but each directory (folder) that's had anything changed inside it.
  13. ^ de Guise, P. (2009). Enterprise Systems Backup and Recovery: A Corporate Insurance Policy. CRC Press. pp. 285–287. ISBN 978-1-4200-7639-4.
  14. ^ Wu, Victor (4 March 2017). "EMC RecoverPoint for Virtual Machine Overview". Victor Virtual. WuChiKin. Retrieved 22 June 2019. The splitter splits out the Write IOs to the VMDK/RDM of a VM and sends a copy to the production VMDK and also to the RecoverPoint for VMs cluster.
  15. ^ "Zerto or Veeam?". RES-Q Services. March 2017. Retrieved 7 July 2019. Zerto doesn't use snapshot technology like Veeam. Instead, Zerto deploys small virtual machines on its physical hosts. These Zerto VMs capture the data as it is written to the host and then send a copy of that data to the replication site.....However, Veeam has the advantage of being able to more efficiently capture and store data for long-term retention needs. There is also a significant pricing difference, with Veeam being cheaper than Zerto.
  16. ^ "Agent Related". CloudEndure.com. 2019. What does the CloudEndure Agent do?. Retrieved 3 July 2019. The CloudEndure Agent performs an initial block-level read of the content of any volume attached to the server and replicates it to the Replication Server. The Agent then acts as an OS-level read filter to capture writes and synchronizes any block level modifications to the CloudEndure Replication Server, ensuring near-zero RPO.
  17. ^ Gardner, Steve (9 December 2004). . Engenio. Peaceful coexistence. Archived from the original on 7 February 2005. Retrieved 26 May 2019.
  18. ^ a b c "Digital Data Storage Outlook 2017" (PDF). Spectra. Spectra Logic. 2017. p. 7(Solid-State), 10(Magnetic Disk), 14(Tape), 17(Optical). Retrieved 11 July 2018.
  19. ^ a b c Tom Coughlin (29 June 2014). "Keeping Data for a Long Time". Forbes. para. Magnetic Tapes(popular formats, storage life), para. Hard Disk Drives(active archive), para. First consider flash memory in archiving(... may not have good media archive life). Retrieved 19 April 2018.
  20. ^ a b c d Jacobi, John L. (29 February 2016). "Hard-core data preservation: The best media and methods for archiving your data". PC World. sec. External Hard Drives(on the shelf, magnetic properties, mechanical stresses, vulnerable to shocks), Tape, Online storage. Retrieved 19 April 2018.
  21. ^ "Ramp Load/Unload Technology in Hard Disk Drives" (PDF). HGST. Western Digital. November 2007. p. 3(sec. Enhanced Shock Tolerance). Retrieved 29 June 2018.
  22. ^ "Toshiba Portable Hard Drive (Canvio® 3.0)". Toshiba Data Dynamics Singapore. Toshiba Data Dynamics Pte Ltd. 2018. sec. Overview(Internal shock sensor and ramp loading technology). Retrieved 16 June 2018.
  23. ^ a b "Iomega Drop Guard ™ Technology" (PDF). Hard Drive Storage Solutions. Iomega Corp. 20 September 2010. pp. 2(What is Drop Shock Technology?, What is Drop Guard Technology? (... features special internal cushioning .... 40% above the industry average)), 3(*NOTE). Retrieved 12 July 2018.
  24. ^ a b John Burek (15 May 2018). "The Best Rugged Hard Drives and SSDs". PC Magazine. Ziff Davis. What Exactly Makes a Drive Rugged?(When a drive is encased ... you're mostly at the mercy of the drive vendor to tell you the rated maximum drop distance for the drive). Retrieved 4 August 2018.
  25. ^ Justin Krajeski; Kimber Streams (20 March 2017). . The New York Times. Archived from the original on 31 March 2017. Retrieved 4 August 2018.
  26. ^ "Best Long-Term Data Archive Solutions". Iron Mountain. Iron Mountain Inc. 2018. sec. More Reliable(average mean time between failure ... rates, best practice for migrating data). Retrieved 19 April 2018.
  27. ^ Kissell, Joe (2011). Take Control of Backing Up Your Mac. Ithaca NY: TidBITS Publishing Inc. p. 41(Deduplication). ISBN 978-1-61542-394-1. Retrieved 17 September 2019.
  28. ^ "Symantec Shows Backup Exec a Little Dedupe Love; Lays out Source Side Deduplication Roadmap – DCIG". DCIG. 7 July 2009. from the original on 4 March 2016. Retrieved 26 February 2016.
  29. ^ "Veritas NetBackup™ Deduplication Guide". Veritas. Veritas Technologies LLC. 2016. Retrieved 26 July 2018.
  30. ^ S. Wan; Q. Cao; C. Xie (2014). "Optical storage: An emerging option in long-term digital preservation". Frontiers of Optoelectronics. 7 (4): 486–492. doi:10.1007/s12200-014-0442-2. S2CID 60816607.
  31. ^ Q. Zhang; Z. Xia; Y.-B. Cheng; M. Gu (2018). "High-capacity optical long data memory based on enhanced Young's modulus in nanoplasmonic hybrid glass composites". Nature Communications. 9 (1): 1183. Bibcode:2018NatCo...9.1183Z. doi:10.1038/s41467-018-03589-y. PMC 5864957. PMID 29568055.
  32. ^ Bärwaldt, Erik (2014). "Full Control » Linux Magazine". Linux Magazine.
  33. ^ "5. Conditions That Affect CDs and DVDs • CLIR". CLIR.
  34. ^ Gérard Poirier; Foued Berahou (3 March 2008). "Journal de 20 Heures". Institut national de l'audiovisuel. approximately minute 30 of the TV news broadcast. Retrieved 3 March 2008.
  35. ^ . Delkin Devices. Delkin Devices Inc. Archived from the original on 27 September 2013.
  36. ^ "Optical Disc Archive Generation 2" (PDF). Optical Disc Archive. Sony. April 2016. p. 12(World’s First 8-Channel Optical Drive Unit). Retrieved 15 August 2019.
  37. ^ R. Micheloni; P. Olivo (2017). "Solid-State Drives (SSDs)". Proceedings of the IEEE. 105 (9): 1586–88. doi:10.1109/JPROC.2017.2727228.
  38. ^ "Remote Backup". EMC Glossary. Dell, Inc. Retrieved 8 May 2018. Effective remote backup requires that production data be regularly backed up to a location far enough away from the primary location so that both locations would not be affected by the same disruptive event.
  39. ^ Stackpole, B.; Hanrion, P. (2007). Software Deployment, Updating, and Patching. CRC Press. pp. 164–165. ISBN 978-1-4200-1329-0. Retrieved 8 May 2018.
  40. ^ Gnanasundaram, S.; Shrivastava, A., eds. (2012). Information Storage and Management: Storing, Managing, and Protecting Digital Information in Classic, Virtualized, and Cloud Environments. John Wiley and Sons. p. 255. ISBN 978-1-118-23696-3. Retrieved 8 May 2018.
  41. ^ Lee (25 January 2017). "What to backup – a critical look at your data". Irontree Blog. Irontree Internet Services CC. Retrieved 8 May 2018.
  42. ^ Preston, W.C. (2007). Backup & Recovery: Inexpensive Backup Solutions for Open Systems. O'Reilly Media, Inc. pp. 111–114. ISBN 978-0-596-55504-7. Retrieved 8 May 2018.
  43. ^ Preston, W.C. (1999). Unix Backup & Recovery. O'Reilly Media, Inc. pp. 73–91. ISBN 978-1-56592-642-4. Retrieved 8 May 2018.
  44. ^ "NILFS Home". NILFS Continuous Snapshotting System. NILFS Community. 2019. Retrieved 22 August 2019.
  45. ^ a b Cougias, D.J.; Heiberger, E.L.; Koop, K. (2003). "Chapter 11: Open file backup for databases". The Backup Book: Disaster Recovery from Desktop to Data Center. Network Frontiers. pp. 356–360. ISBN 0-9729039-0-9.
  46. ^ Liotine, M. (2003). Mission-critical Network Planning. Artech House. p. 244. ISBN 978-1-58053-559-5. Retrieved 8 May 2018.
  47. ^ de Guise, P. (2009). Enterprise Systems Backup and Recovery: A Corporate Insurance Policy. CRC Press. pp. 50–54. ISBN 978-1-4200-7639-4.
  48. ^ "Open File Backup Software for Windows". Handy Backup. Novosoft LLC. 8 November 2018. Retrieved 29 November 2018.
  49. ^ Reitshamer, Stefan (5 July 2017). "Troubleshooting backing up open/locked files on Windows". Arq Blog. Haystack Software. Stefan Reitshamer is the principal developer of Arq. Retrieved 29 November 2018.
  50. ^ Boss, Nina (10 December 1997). . www.wisc.edu. University of Wisconsin. Archived from the original on 2 March 2007. Retrieved 1 December 2018.
  51. ^ "What is ARCHIVE-LOG and NO-ARCHIVE-LOG mode in Oracle and the advantages & disadvantages of these modes?". Arcserve Backup. Arcserve. 27 September 2018. Retrieved 29 November 2018.
  52. ^ Grešovnik, Igor (April 2016). . Archived from the original on 25 April 2016. Retrieved 21 April 2016.
  53. ^ Tridgell, Andrew; Mackerras, Paul; Davison, Wayne. "rsync(1) - Linux man page". linux.die.net.
  54. ^ "Archive maintenance". Code42 Support. 2023.
  55. ^ Pond, James (2 June 2012). "12. Should I delete old backups? If so, How?". Time Machine. baligu.com. Green box, Gray box. Retrieved 21 June 2019.
  56. ^ Kissell, Joe (12 March 2019). "The Best Online Cloud Backup Service". wirecutter. The New York Times. Next, there’s file retention. Retrieved 21 June 2019.
  57. ^ a b c D. Cherry (2015). Securing SQL Server: Protecting Your Database from Attackers. Syngress. pp. 306–308. ISBN 978-0-12-801375-5. Retrieved 8 May 2018.
  58. ^ Backups tapes a backdoor for identity thieves 5 April 2016 at the Wayback Machine (28 April 2004). Retrieved 10 March 2007
  59. ^ a b c Preston, W.C. (2007). Backup & Recovery: Inexpensive Backup Solutions for Open Systems. O'Reilly Media, Inc. pp. 219–220. ISBN 978-0-596-55504-7. Retrieved 8 May 2018.
  60. ^ "Recovery Point Objective (Definition)". ARL Risky Thinking. Albion Research Ltd. 2007. Retrieved 4 August 2019.
  61. ^ "Recovery Time Objective (Definition)". ARL Risky Thinking. Albion Research Ltd. 2007. Retrieved 4 August 2019.
  62. ^ a b Little, D.B. (2003). "Chapter 2: Business Requirements of Backup Systems". Implementing Backup and Recovery: The Readiness Guide for the Enterprise. John Wiley and Sons. pp. 17–30. ISBN 978-0-471-48081-5. Retrieved 8 May 2018.
  63. ^ "How do the "verify" and "write checksums to media" processes work and why are they necessary?". Veritas Support. Veritas.com. 15 October 2015. Write checksums to media. Retrieved 16 September 2019.
  64. ^ HIPAA Advisory 11 April 2007 at the Wayback Machine. Retrieved 10 March 2007

External links edit

  •   The dictionary definition of backup at Wiktionary
  •   Media related to Backup at Wikimedia Commons

backup, this, article, about, duplicate, data, computer, systems, data, recovery, other, uses, disambiguation, information, technology, backup, data, backup, copy, computer, data, taken, stored, elsewhere, that, used, restore, original, after, data, loss, even. This article is about duplicate data in computer systems for data recovery For other uses see Backup disambiguation In information technology a backup or data backup is a copy of computer data taken and stored elsewhere so that it may be used to restore the original after a data loss event The verb form referring to the process of doing so is back up whereas the noun and adjective form is backup 1 Backups can be used to recover data after its loss from data deletion or corruption or to recover data from an earlier time 2 Backups provide a simple form of disaster recovery however not all backup systems are able to reconstitute a computer system or other complex configuration such as a computer cluster active directory server or database server 3 A backup system contains at least one copy of all data considered worth saving The data storage requirements can be large An information repository model may be used to provide structure to this storage There are different types of data storage devices used for copying backups of data that is already in secondary storage onto archive files note 1 4 There are also different ways these devices can be arranged to provide geographic dispersion data security and portability Data is selected extracted and manipulated for storage The process can include methods for dealing with live data including open files as well as compression encryption and de duplication Additional techniques apply to enterprise client server backup Backup schemes may include dry runs that validate the reliability of the data being backed up There are limitations 5 and human factors involved in any backup scheme Contents 1 Storage 1 1 Backup methods 1 1 1 Unstructured 1 1 2 Full only System imaging 1 1 3 Incremental 1 1 4 Near CDP 1 1 5 Reverse incremental 1 1 6 Differential 1 2 Storage media 1 2 1 Magnetic tape 1 2 2 Hard disk 1 2 3 Optical storage 1 2 4 Solid state drive 1 2 5 Remote backup service 1 3 Management 1 3 1 Online 1 3 2 Near line 1 3 3 Off line 1 3 4 Off site data protection 1 3 5 Backup site 2 Selection and extraction of data 2 1 Files 2 2 Filesystems 2 3 Live data 2 4 Metadata 3 Manipulation of data and dataset optimization 3 1 Automated data grooming 3 2 Compression 3 3 Deduplication 3 4 Duplication 3 5 Encryption 3 6 Multiplexing 3 7 Refactoring 3 8 Staging 3 9 Objectives 4 See also 5 Notes 6 References 7 External linksStorage editA backup strategy requires an information repository a secondary storage space for data 6 that aggregates backups of data sources The repository could be as simple as a list of all backup media DVDs etc and the dates produced or could include a computerized index catalog or relational database The backup data needs to be stored requiring a backup rotation scheme 4 which is a system of backing up data to computer media that limits the number of backups of different dates retained separately by appropriate re use of the data storage media by overwriting of backups no longer needed The scheme determines how and when each piece of removable storage is used for a backup operation and how long it is retained once it has backup data stored on it The 3 2 1 rule can aid in the backup process It states that there should be at least 3 copies of the data stored on 2 different types of storage media and one copy should be kept offsite in a remote location this can include cloud storage 2 or more different media should be used to eliminate data loss due to similar reasons for example optical discs may tolerate being underwater while LTO tapes may not and SSDs cannot fail due to head crashes or damaged spindle motors since they do not have any moving parts unlike hard drives An offsite copy protects against fire theft of physical media such as tapes or discs and natural disasters like floods and earthquakes Physically protected hard drives are an alternative to an offsite copy but they have limitations like only being able to resist fire for a limited period of time so an offsite copy still remains as the ideal choice Backup methods edit Unstructured edit An unstructured repository may simply be a stack of tapes DVD Rs or external HDDs with minimal information about what was backed up and when This method is the easiest to implement but unlikely to achieve a high level of recoverability as it lacks automation Full only System imaging edit A repository using this backup method contains complete source data copies taken at one or more specific points in time Copying system images this method is frequently used by computer technicians to record known good configurations However imaging 7 is generally more useful as a way of deploying a standard configuration to many systems rather than as a tool for making ongoing backups of diverse systems Incremental edit An incremental backup stores data changed since a reference point in time Duplicate copies of unchanged data are not copied Typically a full backup of all files is made once or at infrequent intervals serving as the reference point for an incremental repository Subsequently a number of incremental backups are made after successive time periods Restores begin with the last full backup and then apply the incrementals 8 Some backup systems 9 can create a synthetic full backup from a series of incrementals thus providing the equivalent of frequently doing a full backup When done to modify a single archive file this speeds restores of recent versions of files Near CDP edit Continuous Data Protection CDP refers to a backup that instantly saves a copy of every change made to the data This allows restoration of data to any point in time and is the most comprehensive and advanced data protection 10 Near CDP backup applications often marketed as CDP automatically take incremental backups at a specific interval for example every 15 minutes one hour or 24 hours They can therefore only allow restores to an interval boundary 10 Near CDP backup applications use journaling and are typically based on periodic snapshots 11 read only copies of the data frozen at a particular point in time Near CDP except for Apple Time Machine 12 intent logs every change on the host system 13 often by saving byte or block level differences rather than file level differences This backup method differs from simple disk mirroring in that it enables a roll back of the log and thus a restoration of old images of data Intent logging allows precautions for the consistency of live data protecting self consistent files but requiring applications be quiesced and made ready for backup Near CDP is more practicable for ordinary personal backup applications as opposed to true CDP which must be run in conjunction with a virtual machine 14 15 or equivalent 16 and is therefore generally used in enterprise client server backups Software may create copies of individual files such as written documents multimedia projects or user preferences to prevent failed write events caused by power outages operating system crashes or exhausted disk space from causing data loss A common implementation is an appended bak extension to the file name Reverse incremental edit A Reverse incremental backup method stores a recent archive file mirror of the source data and a series of differences between the mirror in its current state and its previous states A reverse incremental backup method starts with a non image full backup After the full backup is performed the system periodically synchronizes the full backup with the live copy while storing the data necessary to reconstruct older versions This can either be done using hard links as Apple Time Machine does or using binary diffs Differential edit A differential backup saves only the data that has changed since the last full backup This means a maximum of two backups from the repository are used to restore the data However as time from the last full backup and thus the accumulated changes in data increases so does the time to perform the differential backup Restoring an entire system requires starting from the most recent full backup and then applying just the last differential backup A differential backup copies files that have been created or changed since the last full backup regardless of whether any other differential backups have been made since whereas an incremental backup copies files that have been created or changed since the most recent backup of any type full or incremental Changes in files may be detected through a more recent date time of last modification file attribute and or changes in file size Other variations of incremental backup include multi level incrementals and block level incrementals that compare parts of files instead of just entire files Storage media edit nbsp From left to right a DVD disc in plastic cover a USB flash drive and an external hard drive Regardless of the repository model that is used the data has to be copied onto an archive file data storage medium The medium used is also referred to as the type of backup destination Magnetic tape edit Magnetic tape was for a long time the most commonly used medium for bulk data storage backup archiving and interchange It was previously a less expensive option but this is no longer the case for smaller amounts of data 17 Tape is a sequential access medium so the rate of continuously writing or reading data can be very fast While tape media itself has a low cost per space tape drives are typically dozens of times as expensive as hard disk drives and optical drives Many tape formats have been proprietary or specific to certain markets like mainframes or a particular brand of personal computer By 2014 LTO had become the primary tape technology 18 The other remaining viable super format is the IBM 3592 also referred to as the TS11xx series The Oracle StorageTek T10000 was discontinued in 2016 19 Hard disk edit The use of hard disk storage has increased over time as it has become progressively cheaper Hard disks are usually easy to use widely available and can be accessed quickly 18 However hard disk backups are close tolerance mechanical devices and may be more easily damaged than tapes especially while being transported 20 In the mid 2000s several drive manufacturers began to produce portable drives employing ramp loading and accelerometer technology sometimes termed a shock sensor 21 22 and by 2010 the industry average in drop tests for drives with that technology showed drives remaining intact and working after a 36 inch non operating drop onto industrial carpeting 23 Some manufacturers also offer ruggedized portable hard drives which include a shock absorbing case around the hard disk and claim a range of higher drop specifications 23 24 25 Over a period of years the stability of hard disk backups is shorter than that of tape backups 19 26 20 External hard disks can be connected via local interfaces like SCSI USB FireWire or eSATA or via longer distance technologies like Ethernet iSCSI or Fibre Channel Some disk based backup systems via Virtual Tape Libraries or otherwise support data deduplication which can reduce the amount of disk storage capacity consumed by daily and weekly backup data 27 28 29 Optical storage edit nbsp Optical discs are not vulnerable to water making them likely to survive a flood disaster See also Optical media preservation Optical storage uses lasers to store and retrieve data Recordable CDs DVDs and Blu ray Discs are commonly used with personal computers and are generally cheap The capacities and speeds of these discs have typically been lower than hard disks or tapes Advances in optical media may shrink that gap in the future 30 31 Potential future data losses caused by gradual media degradation can be predicted by measuring the rate of correctable minor data errors of which consecutively too many increase the risk of uncorrectable sectors Support for error scanning varies among optical drive vendors 32 Many optical disc formats are WORM type which makes them useful for archival purposes since the data cannot be changed Moreover optical discs are not vulnerable to head crashes magnetism imminent water ingress or power surges and a fault of the drive typically just halts the spinning Optical media is modular the storage controller is not tied to media itself like with hard drives or flash storage flash memory controller allowing it to be removed and accessed through a different drive However recordable media may degrade earlier under long term exposure to light 33 Some optical storage systems allow for cataloged data backups without human contact with the discs allowing for longer data integrity A French study in 2008 indicated that the lifespan of typically sold CD Rs was 2 10 years 34 but one manufacturer later estimated the longevity of its CD Rs with a gold sputtered layer to be as high as 100 years 35 Sony s proprietary Optical Disc Archive 18 can in 2016 reach a read rate of 250 MB s 36 Solid state drive edit Solid state drives SSDs use integrated circuit assemblies to store data Flash memory thumb drives USB flash drives CompactFlash SmartMedia Memory Sticks and Secure Digital card devices are relatively expensive for their low capacity but convenient for backing up relatively low data volumes A solid state drive does not contain any movable parts making it less susceptible to physical damage and can have huge throughput of around 500 Mbit s up to 6 Gbit s Available SSDs have become more capacious and cheaper 37 24 Flash memory backups are stable for fewer years than hard disk backups 19 Remote backup service edit Remote backup services or cloud backups involve service providers storing data offsite This has been used to protect against events such as fires floods or earthquakes which could destroy locally stored backups 38 Cloud based backup through services like or similar to Google Drive and Microsoft OneDrive provides a layer of data protection 20 However the users must trust the provider to maintain the privacy and integrity of their data with confidentiality enhanced by the use of encryption Because speed and availability are limited by a user s online connection 20 users with large amounts of data may need to use cloud seeding and large scale recovery Management edit Various methods can be used to manage backup media striking a balance between accessibility security and cost These media management methods are not mutually exclusive and are frequently combined to meet the user s needs Using on line disks for staging data before it is sent to a near line tape library is a common example 39 40 Online edit Online backup storage is typically the most accessible type of data storage and can begin a restore in milliseconds An internal hard disk or a disk array maybe connected to SAN is an example of an online backup This type of storage is convenient and speedy but is vulnerable to being deleted or overwritten either by accident by malevolent action or in the wake of a data deleting virus payload Near line edit Nearline storage is typically less accessible and less expensive than online storage but still useful for backup data storage A mechanical device is usually used to move media units from storage into a drive where the data can be read or written Generally it has safety properties similar to on line storage An example is a tape library with restore times ranging from seconds to a few minutes Off line edit Off line storage requires some direct action to provide access to the storage media for example inserting a tape into a tape drive or plugging in a cable Because the data is not accessible via any computer except during limited periods in which they are written or read back they are largely immune to on line backup failure modes Access time varies depending on whether the media are on site or off site Off site data protection edit Backup media may be sent to an off site vault to protect against a disaster or other site specific problem The vault can be as simple as a system administrator s home office or as sophisticated as a disaster hardened temperature controlled high security bunker with facilities for backup media storage A data replica can be off site but also on line e g an off site RAID mirror Such a replica has fairly limited value as a backup Backup site edit A backup site or disaster recovery center is used to store data that can enable computer systems and networks to be restored and properly configured in the event of a disaster Some organisations have their own data recovery centres while others contract this out to a third party Due to high costs backing up is rarely considered the preferred method of moving data to a DR site A more typical way would be remote disk mirroring which keeps the DR data as up to date as possible Selection and extraction of data editA backup operation starts with selecting and extracting coherent units of data Most data on modern computer systems is stored in discrete units known as files These files are organized into filesystems Deciding what to back up at any given time involves tradeoffs By backing up too much redundant data the information repository will fill up too quickly Backing up an insufficient amount of data can eventually lead to the loss of critical information 41 Files edit Copying files Making copies of files is the simplest and most common way to perform a backup A means to perform this basic function is included in all backup software and all operating systems Partial file copying A backup may include only the blocks or bytes within a file that have changed in a given period of time This can substantially reduce needed storage space but requires higher sophistication to reconstruct files in a restore situation Some implementations require integration with the source file system Deleted files To prevent the unintentional restoration of files that have been intentionally deleted a record of the deletion must be kept Versioning of files Most backup applications other than those that do only full only System imaging also back up files that have been modified since the last backup That way you can retrieve many different versions of a given file and if you delete it on your hard disk you can still find it in your information repository archive 4 Filesystems edit Filesystem dump A copy of the whole filesystem in block level can be made This is also known as a raw partition backup and is related to disk imaging The process usually involves unmounting the filesystem and running a program like dd Unix 42 Because the disk is read sequentially and with large buffers this type of backup can be faster than reading every file normally especially when the filesystem contains many small files is highly fragmented or is nearly full But because this method also reads the free disk blocks that contain no useful data this method can also be slower than conventional reading especially when the filesystem is nearly empty Some filesystems such as XFS provide a dump utility that reads the disk sequentially for high performance while skipping unused sections The corresponding restore utility can selectively restore individual files or the entire volume at the operator s choice 43 Identification of changes Some filesystems have an archive bit for each file that says it was recently changed Some backup software looks at the date of the file and compares it with the last backup to determine whether the file was changed Versioning file system A versioning filesystem tracks all changes to a file The NILFS versioning filesystem for Linux is an example 44 Live data edit Files that are actively being updated present a challenge to back up One way to back up live data is to temporarily quiesce them e g close all files take a snapshot and then resume live operations At this point the snapshot can be backed up through normal methods 45 A snapshot is an instantaneous function of some filesystems that presents a copy of the filesystem as if it were frozen at a specific point in time often by a copy on write mechanism Snapshotting a file while it is being changed results in a corrupted file that is unusable This is also the case across interrelated files as may be found in a conventional database or in applications such as Microsoft Exchange Server 11 The term fuzzy backup can be used to describe a backup of live data that looks like it ran correctly but does not represent the state of the data at a single point in time 46 Backup options for data files that cannot be or are not quiesced include 47 Open file backup Many backup software applications undertake to back up open files in an internally consistent state 48 Some applications simply check whether open files are in use and try again later 45 Other applications exclude open files that are updated very frequently 49 Some low availability interactive applications can be backed up via natural induced pausing Interrelated database files backup Some interrelated database file systems offer a means to generate a hot backup 50 of the database while it is online and usable This may include a snapshot of the data files plus a snapshotted log of changes made while the backup is running Upon a restore the changes in the log files are applied to bring the copy of the database up to the point in time at which the initial backup ended 51 Other low availability interactive applications can be backed up via coordinated snapshots However genuinely high availability interactive applications can be only be backed up via Continuous Data Protection Metadata edit Not all information stored on the computer is stored in files Accurately recovering a complete system from scratch requires keeping track of this non file data too 52 System description System specifications are needed to procure an exact replacement after a disaster Boot sector The boot sector can sometimes be recreated more easily than saving it It usually isn t a normal file and the system won t boot without it Partition layout The layout of the original disk as well as partition tables and filesystem settings is needed to properly recreate the original system File metadata Each file s permissions owner group ACLs and any other metadata need to be backed up for a restore to properly recreate the original environment System metadata Different operating systems have different ways of storing configuration information Microsoft Windows keeps a registry of system information that is more difficult to restore than a typical file Manipulation of data and dataset optimization editIt is frequently useful or required to manipulate the data being backed up to optimize the backup process These manipulations can improve backup speed restore speed data security media usage and or reduced bandwidth requirements Automated data grooming edit Out of date data can be automatically deleted but for personal backup applications as opposed to enterprise client server backup applications where automated data grooming can be customized the deletion note 2 53 54 can at most 55 be globally delayed or be disabled 56 Compression edit Various schemes can be employed to shrink the size of the source data to be stored so that it uses less storage space Compression is frequently a built in feature of tape drive hardware 57 Deduplication edit Redundancy due to backing up similarly configured workstations can be reduced thus storing just one copy This technique can be applied at the file or raw block level This potentially large reduction 57 is called deduplication It can occur on a server before any data moves to backup media sometimes referred to as source client side deduplication This approach also reduces bandwidth required to send backup data to its target media The process can also occur at the target storage device sometimes referred to as inline or back end deduplication Duplication edit Sometimes backups are duplicated to a second set of storage media This can be done to rearrange the archive files to optimize restore speed or to have a second copy at a different location or on a different storage medium as in the disk to disk to tape capability of Enterprise client server backup Encryption edit High capacity removable storage media such as backup tapes present a data security risk if they are lost or stolen 58 Encrypting the data on these media can mitigate this problem however encryption is a CPU intensive process that can slow down backup speeds and the security of the encrypted backups is only as effective as the security of the key management policy 57 Multiplexing edit When there are many more computers to be backed up than there are destination storage devices the ability to use a single storage device with several simultaneous backups can be useful 59 However cramming the scheduled backup window via multiplexed backup is only used for tape destinations 59 Refactoring edit The process of rearranging the sets of backups in an archive file is known as refactoring For example if a backup system uses a single tape each day to store the incremental backups for all the protected computers restoring one of the computers could require many tapes Refactoring could be used to consolidate all the backups for a single computer onto a single tape creating a synthetic full backup This is especially useful for backup systems that do incrementals forever style backups Staging edit Sometimes backups are copied to a staging disk before being copied to tape 59 This process is sometimes referred to as D2D2T an acronym for Disk to disk to tape It can be useful if there is a problem matching the speed of the final destination device with the source device as is frequently faced in network based backup systems It can also serve as a centralized location for applying other data manipulation techniques Objectives edit Recovery point objective RPO The point in time that the restarted infrastructure will reflect expressed as the maximum targeted period in which data transactions might be lost from an IT service due to a major incident Essentially this is the roll back that will be experienced as a result of the recovery The most desirable RPO would be the point just prior to the data loss event Making a more recent recovery point achievable requires increasing the frequency of synchronization between the source data and the backup repository 60 Recovery time objective RTO The amount of time elapsed between disaster and restoration of business functions 61 Data security In addition to preserving access to data for its owners data must be restricted from unauthorized access Backups must be performed in a manner that does not compromise the original owner s undertaking This can be achieved with data encryption and proper media handling policies 62 Data retention period Regulations and policy can lead to situations where backups are expected to be retained for a particular period but not any further Retaining backups after this period can lead to unwanted liability and sub optimal use of storage media 62 Checksum or hash function validation Applications that back up to tape archive files need this option to verify that the data was accurately copied 63 Backup process monitoring Enterprise client server backup applications need a user interface that allows administrators to monitor the backup process and proves compliance to regulatory bodies outside the organization for example an insurance company in the USA might be required under HIPAA to demonstrate that its client data meet records retention requirements 64 User initiated backups and restores To avoid or recover from minor disasters such as inadvertently deleting or overwriting the good versions of one or more files the computer user rather than an administrator may initiate backups and restores from not necessarily the most recent backup of files or folders See also editAbout backup Backup software and services List of backup software Comparison of online backup services Glossary of backup terms Virtual backup appliance Related topics Data consistency Data degradation Data portability Data proliferation Database dump Digital preservation Disaster recovery and business continuity auditing World Backup DayNotes edit In contrast to everyday use of the term archive the data stored in an archive file is not necessarily old or of historical interest Some backup applications notably rsync and CrashPlan term removing backup data pruning instead of grooming References edit back up The American Heritage Dictionary of the English Language Houghton Mifflin Harcourt 2018 Retrieved 9 May 2018 S Nelson 2011 Chapter 1 Introduction to Backup and Recovery Pro Data Backup and Recovery Apress pp 1 16 ISBN 978 1 4302 2663 5 Retrieved 8 May 2018 Cougias D J Heiberger E L Koop K 2003 Chapter 1 What s a Disaster Without a Recovery The Backup Book Disaster Recovery from Desktop to Data Center Network Frontiers pp 1 14 ISBN 0 9729039 0 9 a b c Joe Kissell 2007 Take Control of Mac OS X Backups PDF Version 2 0 ed Ithaca NY TidBITS Electronic Publishing pp 18 20 The Archive meaning information repository including versioning 24 client server 82 83 archive file 112 114 Off site storage backup rotation scheme 126 141 old Retrospect terminology and GUI still used in Windows variant 165 client server 128 subvolume later renamed Favorite Folder in Macintosh variant ISBN 978 0 9759503 0 2 Archived from the original PDF on 1 December 2020 Retrieved 17 May 2019 Terry Sullivan 11 January 2018 A Beginner s Guide to Backing Up Photos The New York Times a hard drive an established company declared bankruptcy where many had McMahon Mary 1 April 2019 What Is an Information Repository wiseGEEK Conjecture Corporation Retrieved 8 May 2019 In the sense of an approach to data management an information repository is a secondary storage space for data Five key questions to ask about your backup solution sysgen ca 23 March 2014 Does your company have a low tolerance to longer data access outages and or would you like to minimize the time your company may be without its data Archived from the original on 4 March 2016 Retrieved 23 September 2015 Incremental Backup Tech FAQ Independent Media 13 June 2005 Archived from the original on 21 June 2016 Retrieved 10 March 2006 Pond James 31 August 2013 How Time Machine Works its Magic Apple OSX and Time Machine Tips baligu com File System Event Store Hard Links Archived from the original on 21 June 2019 Retrieved 19 May 2019 a b Behzad Behtash 6 May 2010 Why Continuous Data Protection s Getting More Practical Disaster recovery business continuity InformationWeek Retrieved 12 November 2011 A true CDP approach should capture all data writes thus continuously backing up data and eliminating backup windows CDP is the gold standard the most comprehensive and advanced data protection But near CDP technologies can deliver enough protection for many companies with less complexity and cost For example snapshots can provide a reasonable near CDP level of protection for file shares letting users directly access data on the file share at regular intervals say every half hour or 15 minutes That s certainly a higher level of protection than tape based or disk based nightly backups and may be all you need a b Continuous data protection CDP explained True CDP vs near CDP ComputerWeekly com TechTarget July 2010 Retrieved 22 June 2019 copies data from a source to a target True CDP does this every time a change is made while so called near CDP does this at pre set time intervals Near CDP is effectively the same as snapshotting True CDP systems record every write and copy them to the target where all changes are stored in a log new paragraph By contrast near CDP snapshot systems copy files in a straightforward manner but require applications to be quiesced and made ready for backup either via the application s backup mode or using for example Microsoft s Volume Shadow Copy Services VSS Pond James 31 August 2013 How Time Machine Works its Magic Apple OSX and Time Machine Tips Baligu com as mirrored after James Pond died in 2013 Retrieved 10 July 2019 The File System Event Store is a hidden log that OSX keeps on each HFS formatted disk partition of changes made to the data on it It doesn t list every file that s changed but each directory folder that s had anything changed inside it de Guise P 2009 Enterprise Systems Backup and Recovery A Corporate Insurance Policy CRC Press pp 285 287 ISBN 978 1 4200 7639 4 Wu Victor 4 March 2017 EMC RecoverPoint for Virtual Machine Overview Victor Virtual WuChiKin Retrieved 22 June 2019 The splitter splits out the Write IOs to the VMDK RDM of a VM and sends a copy to the production VMDK and also to the RecoverPoint for VMs cluster Zerto or Veeam RES Q Services March 2017 Retrieved 7 July 2019 Zerto doesn t use snapshot technology like Veeam Instead Zerto deploys small virtual machines on its physical hosts These Zerto VMs capture the data as it is written to the host and then send a copy of that data to the replication site However Veeam has the advantage of being able to more efficiently capture and store data for long term retention needs There is also a significant pricing difference with Veeam being cheaper than Zerto Agent Related CloudEndure com 2019 What does the CloudEndure Agent do Retrieved 3 July 2019 The CloudEndure Agent performs an initial block level read of the content of any volume attached to the server and replicates it to the Replication Server The Agent then acts as an OS level read filter to capture writes and synchronizes any block level modifications to the CloudEndure Replication Server ensuring near zero RPO Gardner Steve 9 December 2004 Disk to Disk Backup versus Tape War or Truce Engenio Peaceful coexistence Archived from the original on 7 February 2005 Retrieved 26 May 2019 a b c Digital Data Storage Outlook 2017 PDF Spectra Spectra Logic 2017 p 7 Solid State 10 Magnetic Disk 14 Tape 17 Optical Retrieved 11 July 2018 a b c Tom Coughlin 29 June 2014 Keeping Data for a Long Time Forbes para Magnetic Tapes popular formats storage life para Hard Disk Drives active archive para First consider flash memory in archiving may not have good media archive life Retrieved 19 April 2018 a b c d Jacobi John L 29 February 2016 Hard core data preservation The best media and methods for archiving your data PC World sec External Hard Drives on the shelf magnetic properties mechanical stresses vulnerable to shocks Tape Online storage Retrieved 19 April 2018 Ramp Load Unload Technology in Hard Disk Drives PDF HGST Western Digital November 2007 p 3 sec Enhanced Shock Tolerance Retrieved 29 June 2018 Toshiba Portable Hard Drive Canvio 3 0 Toshiba Data Dynamics Singapore Toshiba Data Dynamics Pte Ltd 2018 sec Overview Internal shock sensor and ramp loading technology Retrieved 16 June 2018 a b Iomega Drop Guard Technology PDF Hard Drive Storage Solutions Iomega Corp 20 September 2010 pp 2 What is Drop Shock Technology What is Drop Guard Technology features special internal cushioning 40 above the industry average 3 NOTE Retrieved 12 July 2018 a b John Burek 15 May 2018 The Best Rugged Hard Drives and SSDs PC Magazine Ziff Davis What Exactly Makes a Drive Rugged When a drive is encased you re mostly at the mercy of the drive vendor to tell you the rated maximum drop distance for the drive Retrieved 4 August 2018 Justin Krajeski Kimber Streams 20 March 2017 The Best Portable Hard Drive The New York Times Archived from the original on 31 March 2017 Retrieved 4 August 2018 Best Long Term Data Archive Solutions Iron Mountain Iron Mountain Inc 2018 sec More Reliable average mean time between failure rates best practice for migrating data Retrieved 19 April 2018 Kissell Joe 2011 Take Control of Backing Up Your Mac Ithaca NY TidBITS Publishing Inc p 41 Deduplication ISBN 978 1 61542 394 1 Retrieved 17 September 2019 Symantec Shows Backup Exec a Little Dedupe Love Lays out Source Side Deduplication Roadmap DCIG DCIG 7 July 2009 Archived from the original on 4 March 2016 Retrieved 26 February 2016 Veritas NetBackup Deduplication Guide Veritas Veritas Technologies LLC 2016 Retrieved 26 July 2018 S Wan Q Cao C Xie 2014 Optical storage An emerging option in long term digital preservation Frontiers of Optoelectronics 7 4 486 492 doi 10 1007 s12200 014 0442 2 S2CID 60816607 Q Zhang Z Xia Y B Cheng M Gu 2018 High capacity optical long data memory based on enhanced Young s modulus in nanoplasmonic hybrid glass composites Nature Communications 9 1 1183 Bibcode 2018NatCo 9 1183Z doi 10 1038 s41467 018 03589 y PMC 5864957 PMID 29568055 Barwaldt Erik 2014 Full Control Linux Magazine Linux Magazine 5 Conditions That Affect CDs and DVDs CLIR CLIR Gerard Poirier Foued Berahou 3 March 2008 Journal de 20 Heures Institut national de l audiovisuel approximately minute 30 of the TV news broadcast Retrieved 3 March 2008 Archival Gold CD R 300 Year Disc Binder of 10 Discs with Scratch Armor Surface Delkin Devices Delkin Devices Inc Archived from the original on 27 September 2013 Optical Disc Archive Generation 2 PDF Optical Disc Archive Sony April 2016 p 12 World s First 8 Channel Optical Drive Unit Retrieved 15 August 2019 R Micheloni P Olivo 2017 Solid State Drives SSDs Proceedings of the IEEE 105 9 1586 88 doi 10 1109 JPROC 2017 2727228 Remote Backup EMC Glossary Dell Inc Retrieved 8 May 2018 Effective remote backup requires that production data be regularly backed up to a location far enough away from the primary location so that both locations would not be affected by the same disruptive event Stackpole B Hanrion P 2007 Software Deployment Updating and Patching CRC Press pp 164 165 ISBN 978 1 4200 1329 0 Retrieved 8 May 2018 Gnanasundaram S Shrivastava A eds 2012 Information Storage and Management Storing Managing and Protecting Digital Information in Classic Virtualized and Cloud Environments John Wiley and Sons p 255 ISBN 978 1 118 23696 3 Retrieved 8 May 2018 Lee 25 January 2017 What to backup a critical look at your data Irontree Blog Irontree Internet Services CC Retrieved 8 May 2018 Preston W C 2007 Backup amp Recovery Inexpensive Backup Solutions for Open Systems O Reilly Media Inc pp 111 114 ISBN 978 0 596 55504 7 Retrieved 8 May 2018 Preston W C 1999 Unix Backup amp Recovery O Reilly Media Inc pp 73 91 ISBN 978 1 56592 642 4 Retrieved 8 May 2018 NILFS Home NILFS Continuous Snapshotting System NILFS Community 2019 Retrieved 22 August 2019 a b Cougias D J Heiberger E L Koop K 2003 Chapter 11 Open file backup for databases The Backup Book Disaster Recovery from Desktop to Data Center Network Frontiers pp 356 360 ISBN 0 9729039 0 9 Liotine M 2003 Mission critical Network Planning Artech House p 244 ISBN 978 1 58053 559 5 Retrieved 8 May 2018 de Guise P 2009 Enterprise Systems Backup and Recovery A Corporate Insurance Policy CRC Press pp 50 54 ISBN 978 1 4200 7639 4 Open File Backup Software for Windows Handy Backup Novosoft LLC 8 November 2018 Retrieved 29 November 2018 Reitshamer Stefan 5 July 2017 Troubleshooting backing up open locked files on Windows Arq Blog Haystack Software Stefan Reitshamer is the principal developer of Arq Retrieved 29 November 2018 Boss Nina 10 December 1997 Oracle Tips Session 3 Oracle Backups www wisc edu University of Wisconsin Archived from the original on 2 March 2007 Retrieved 1 December 2018 What is ARCHIVE LOG and NO ARCHIVE LOG mode in Oracle and the advantages amp disadvantages of these modes Arcserve Backup Arcserve 27 September 2018 Retrieved 29 November 2018 Gresovnik Igor April 2016 Preparation of Bootable Media and Images Archived from the original on 25 April 2016 Retrieved 21 April 2016 Tridgell Andrew Mackerras Paul Davison Wayne rsync 1 Linux man page linux die net Archive maintenance Code42 Support 2023 Pond James 2 June 2012 12 Should I delete old backups If so How Time Machine baligu com Green box Gray box Retrieved 21 June 2019 Kissell Joe 12 March 2019 The Best Online Cloud Backup Service wirecutter The New York Times Next there s file retention Retrieved 21 June 2019 a b c D Cherry 2015 Securing SQL Server Protecting Your Database from Attackers Syngress pp 306 308 ISBN 978 0 12 801375 5 Retrieved 8 May 2018 Backups tapes a backdoor for identity thieves Archived 5 April 2016 at the Wayback Machine 28 April 2004 Retrieved 10 March 2007 a b c Preston W C 2007 Backup amp Recovery Inexpensive Backup Solutions for Open Systems O Reilly Media Inc pp 219 220 ISBN 978 0 596 55504 7 Retrieved 8 May 2018 Recovery Point Objective Definition ARL Risky Thinking Albion Research Ltd 2007 Retrieved 4 August 2019 Recovery Time Objective Definition ARL Risky Thinking Albion Research Ltd 2007 Retrieved 4 August 2019 a b Little D B 2003 Chapter 2 Business Requirements of Backup Systems Implementing Backup and Recovery The Readiness Guide for the Enterprise John Wiley and Sons pp 17 30 ISBN 978 0 471 48081 5 Retrieved 8 May 2018 How do the verify and write checksums to media processes work and why are they necessary Veritas Support Veritas com 15 October 2015 Write checksums to media Retrieved 16 September 2019 HIPAA Advisory Archived 11 April 2007 at the Wayback Machine Retrieved 10 March 2007External links edit nbsp The dictionary definition of backup at Wiktionary nbsp Media related to Backup at Wikimedia Commons Retrieved from https en wikipedia org w index php title Backup amp oldid 1220710580, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.