Regular backups are important to prevent data loss from hardware failures, software issues, user errors, and other problems. There are several backup methods like full, differential, and incremental backups that back up data in different ways. Data can be backed up locally to external hard drives or networks or backed up offsite to cloud services. Proper file formats, documentation, security practices, and testing backups help ensure data is preserved and accessible over time.
The document discusses asset management and digital workflow best practices for photographers. It covers organizing digital files through proper naming conventions, file storage using external hard drives and cloud services, and establishing backup and archiving systems. The Universal Photographic Digital Imaging Guidelines (UPDIG) recommend a workflow process of planning shoots, capturing images, and ingesting files to ensure quality reproduction and efficiency. Proper asset management is important for organizing a growing portfolio and facilitating production.
This document provides guidelines for planning and maintaining digital media storage. It discusses various storage options such as online, near-line, and offline storage. Online storage allows immediate access but is more expensive, while offline storage has limited accessibility but focuses on long-term preservation. The document also outlines best practices for storage environments including temperature, power, security and backups to protect digital records for long-term access. Proper planning, labeling, inventorying and care of storage media and facilities is necessary to preserve digital records over time.
The document discusses file storage and backup systems. It defines files and folders, describes common file storage devices like hard drives, optical disks, solid state drives and memory sticks. It emphasizes the importance of backing up files regularly using either manual backups, backup applications, automatic backup systems or cloud backup services to protect against data loss from hardware failure. Simple backup strategies and resources for learning more are provided.
Backing up data is important to protect against data loss or corruption. There are two main types of backups: incremental backups that only back up changed files, and image backups that make a complete copy of all files. Backup options include external drives, cloud storage, or network-attached storage. It is important to back up files regularly and store backups separately from the computer to ensure security.
Selection & formulation of Research Problem
Literature survey
Development of working hypotheses
Research design
Sampling strategy or sample design
Pilot (quick & dirty) study
Data collection
Processing & analysis of data
Testing hypotheses
Interpretation & generalisation
Preparation of the report
This document discusses data integrity and how to preserve it. It defines data integrity as the completeness, accuracy, and consistency of data throughout its lifecycle. Data integrity ensures the validity and accuracy of data and is different from data security which protects data. Threats to data integrity include human error, transfer errors, security issues, attacks, and hardware problems. Tools are recommended to check for duplicate files and remove them to maintain integrity. Backups are also important to prevent data loss. The document outlines features of data integrity like accessibility and reliability of data. It provides a checklist for input and data validation, duplicate removal, and access controls to preserve integrity.
Power point presentation on backup and recovery.
A good presentation cover all topics.
For any other type of ppt's or pdf's to be created on demand contact -dhawalm8@gmail.com
mob. no-7023419969
This document provides practical steps organizations can take to preserve digital information without a digital archive. It discusses preserving the bitstream by ensuring backup and storage of files on multiple media types and locations. It also covers preserving accessibility by addressing software and format obsolescence through normalization to long-term preservation formats. Preserving context involves classifying records and capturing metadata to maintain understanding of records over time. The key recommendations are to survey digital holdings, ensure backup systems are effective, monitor file formats and choose preservation formats, and work towards systems to capture record context.
This document provides an overview of practical steps organizations can take to preserve digital information without a full digital archive program. It discusses preserving the bitstream by ensuring proper storage on multiple media types and locations with backups. It also covers preserving accessibility by addressing issues of software and format obsolescence through normalization to long-term preservation formats. Finally, it addresses preserving context by implementing recordkeeping practices like classification systems and metadata capture to maintain authenticity and integrity of records over time. The key recommendations are to survey digital holdings, conduct risk assessments, choose preservation formats, and work towards systems to manage records and their context.
This presentation covers a number of best practices for managing research data. The main topics include: file naming and organization conventions, data documentation, and data storage and backups.
Proact backup and storage vs ransomware 2021Karel Kannel
油
This document discusses recommendations for protecting against ransomware attacks based on statements from the UK's National Cyber Security Centre. It recommends taking a holistic approach involving multiple layers of protection for primary storage, backup systems, and backup data. Specific recommendations include implementing access controls, regular backups, monitoring for anomalies, isolating backup systems and data, and having independent copies of data stored on different media in different locations. Solutions discussed that can help implement these recommendations include Veritas NetBackup, Rubrik, ExaGrid, and NetApp.
University of Bath Research Data Management training for researchersJez Cope
油
際際滷s from a workshop on Research Data Management for research staff and students at the University of Bath.
Part of the Research360 project (http://blogs.bath.ac.uk/research360).
Authors: Cathy Pink and Jez Cope, University of Bath
This document summarizes Peter Chan's presentation on accessioning born-digital materials. The presentation covered literature reviews on best practices, putting accessioning in context within Stanford's workflow, and a demonstration of their forensic workflow. The workflow involves surveying collections, creating accession records, photographing media, virus checking, creating disk images, generating summaries, and transferring data to secure storage. Questions from attendees were also taken and a tour of the forensic lab was included.
This document summarizes the UW Desktop Encryption Project. The project aims to research encryption tools to protect restricted data on lost or stolen devices. It will recommend a product for pilot testing and evaluate its full disk and file/folder encryption. Challenges include supporting different platforms, key management, and gaining user acceptance. The project selected SafeBoot due to its features and will pilot it through June before recommending a solution to sponsors.
Data Management for Undergraduate Researchers (updated - 02/2016)Rebekah Cummings
油
This document summarizes a seminar on data management for undergraduate researchers. It discusses what data is, why it needs to be managed, and key aspects of effective data management including data organization, metadata, storage and archiving. Specific topics covered include creating data management plans, file naming conventions, structuring folders, describing data through codebooks and documentation, backup strategies, and long-term archival options. The goal is to help researchers organize and document their data so it can be understood and preserved over time.
Data backup involves copying files and data to external or online storage so they are preserved if the original files are lost or damaged. Reasons for data loss include hardware failures, viruses, file corruption, and disasters. The main purpose of data backup is to avoid data loss of important financial, customer, and company information that would be difficult to replace. Backup options include external drives, internal drives, department servers, online backup sites, and cloud storage services.
This document summarizes various methods for protecting data security. It discusses procedures like using passwords, biometric identification, encryption, and access hierarchies to restrict data access. It also covers consequences of not protecting data like loss of trade secrets, privacy violations, loss of reputation, income loss, and potential legal prosecution. The document recommends regular backups stored offsite, using virus scanners and firewalls, and properly disposing of or destroying old storage devices.
Android forensics an Custom Recovery ImageMohamed Khaled
油
Mobile Forensic Process
Different Mobile Forensic Scenario
Acquisition Guide
Challenges of Android Forensics
How to Circumvent the Pass Code
Types Of Analyses(Logical analysis)
Types Of Analyses(Physical analysis)
Android Partition Layout
Custom Recovery Modifications
How Data are Stored In Android
Example of Useful Data extracted from Android Image
Webinar: Are You Treating Unstructured Data as a Second Class Citizen?Storage Switzerland
油
This document discusses the importance of properly protecting unstructured data and compares current practices for protecting structured data versus unstructured data. It notes that unstructured data sets are often the largest in terms of capacity but are typically only backed up once per night. This leaves them vulnerable, especially to ransomware attacks. The document advocates for using a purpose-built solution for unstructured data protection that uses snapshot-like data capture for granularity and content search abilities while leveraging object storage or public clouds for long-term retention and instant restore capabilities. It promotes the Aparavi File Protect & Insight solution.
Trigonometry studies relationships involving lengths and angles of triangles. It emerged from applications of geometry to astronomy during the 3rd century BC. Basic trigonometric identities include tan x sin x + cos x = sec x and cos^2 20+cos^2 70/sin^2 31+sin^2 59+sin^2 64+cos64.sin26 = 2. Applications involve the line of sight from an observer's eye to an airplane overhead, defining the angle of elevation, and the pilot's downward line of sight, defining the angle of depression.
17-MOD 6 Conducting Security Audits & MOD 7 Information Security Audit Prepar...abhichowdary16
油
This document discusses information security audits and their key features. It describes the different types of security audits and phases of an information security audit. It outlines the audit process, including defining the security perimeter, describing system components, determining threats, and using appropriate tools. It also discusses auditor roles and skills, as well as elements that characterize a good security audit like clearly defined objectives and an experienced independent audit team.
More Related Content
Similar to 7-Backups of security Devices-03-06-2023.ppt (20)
Power point presentation on backup and recovery.
A good presentation cover all topics.
For any other type of ppt's or pdf's to be created on demand contact -dhawalm8@gmail.com
mob. no-7023419969
This document provides practical steps organizations can take to preserve digital information without a digital archive. It discusses preserving the bitstream by ensuring backup and storage of files on multiple media types and locations. It also covers preserving accessibility by addressing software and format obsolescence through normalization to long-term preservation formats. Preserving context involves classifying records and capturing metadata to maintain understanding of records over time. The key recommendations are to survey digital holdings, ensure backup systems are effective, monitor file formats and choose preservation formats, and work towards systems to capture record context.
This document provides an overview of practical steps organizations can take to preserve digital information without a full digital archive program. It discusses preserving the bitstream by ensuring proper storage on multiple media types and locations with backups. It also covers preserving accessibility by addressing issues of software and format obsolescence through normalization to long-term preservation formats. Finally, it addresses preserving context by implementing recordkeeping practices like classification systems and metadata capture to maintain authenticity and integrity of records over time. The key recommendations are to survey digital holdings, conduct risk assessments, choose preservation formats, and work towards systems to manage records and their context.
This presentation covers a number of best practices for managing research data. The main topics include: file naming and organization conventions, data documentation, and data storage and backups.
Proact backup and storage vs ransomware 2021Karel Kannel
油
This document discusses recommendations for protecting against ransomware attacks based on statements from the UK's National Cyber Security Centre. It recommends taking a holistic approach involving multiple layers of protection for primary storage, backup systems, and backup data. Specific recommendations include implementing access controls, regular backups, monitoring for anomalies, isolating backup systems and data, and having independent copies of data stored on different media in different locations. Solutions discussed that can help implement these recommendations include Veritas NetBackup, Rubrik, ExaGrid, and NetApp.
University of Bath Research Data Management training for researchersJez Cope
油
際際滷s from a workshop on Research Data Management for research staff and students at the University of Bath.
Part of the Research360 project (http://blogs.bath.ac.uk/research360).
Authors: Cathy Pink and Jez Cope, University of Bath
This document summarizes Peter Chan's presentation on accessioning born-digital materials. The presentation covered literature reviews on best practices, putting accessioning in context within Stanford's workflow, and a demonstration of their forensic workflow. The workflow involves surveying collections, creating accession records, photographing media, virus checking, creating disk images, generating summaries, and transferring data to secure storage. Questions from attendees were also taken and a tour of the forensic lab was included.
This document summarizes the UW Desktop Encryption Project. The project aims to research encryption tools to protect restricted data on lost or stolen devices. It will recommend a product for pilot testing and evaluate its full disk and file/folder encryption. Challenges include supporting different platforms, key management, and gaining user acceptance. The project selected SafeBoot due to its features and will pilot it through June before recommending a solution to sponsors.
Data Management for Undergraduate Researchers (updated - 02/2016)Rebekah Cummings
油
This document summarizes a seminar on data management for undergraduate researchers. It discusses what data is, why it needs to be managed, and key aspects of effective data management including data organization, metadata, storage and archiving. Specific topics covered include creating data management plans, file naming conventions, structuring folders, describing data through codebooks and documentation, backup strategies, and long-term archival options. The goal is to help researchers organize and document their data so it can be understood and preserved over time.
Data backup involves copying files and data to external or online storage so they are preserved if the original files are lost or damaged. Reasons for data loss include hardware failures, viruses, file corruption, and disasters. The main purpose of data backup is to avoid data loss of important financial, customer, and company information that would be difficult to replace. Backup options include external drives, internal drives, department servers, online backup sites, and cloud storage services.
This document summarizes various methods for protecting data security. It discusses procedures like using passwords, biometric identification, encryption, and access hierarchies to restrict data access. It also covers consequences of not protecting data like loss of trade secrets, privacy violations, loss of reputation, income loss, and potential legal prosecution. The document recommends regular backups stored offsite, using virus scanners and firewalls, and properly disposing of or destroying old storage devices.
Android forensics an Custom Recovery ImageMohamed Khaled
油
Mobile Forensic Process
Different Mobile Forensic Scenario
Acquisition Guide
Challenges of Android Forensics
How to Circumvent the Pass Code
Types Of Analyses(Logical analysis)
Types Of Analyses(Physical analysis)
Android Partition Layout
Custom Recovery Modifications
How Data are Stored In Android
Example of Useful Data extracted from Android Image
Webinar: Are You Treating Unstructured Data as a Second Class Citizen?Storage Switzerland
油
This document discusses the importance of properly protecting unstructured data and compares current practices for protecting structured data versus unstructured data. It notes that unstructured data sets are often the largest in terms of capacity but are typically only backed up once per night. This leaves them vulnerable, especially to ransomware attacks. The document advocates for using a purpose-built solution for unstructured data protection that uses snapshot-like data capture for granularity and content search abilities while leveraging object storage or public clouds for long-term retention and instant restore capabilities. It promotes the Aparavi File Protect & Insight solution.
Trigonometry studies relationships involving lengths and angles of triangles. It emerged from applications of geometry to astronomy during the 3rd century BC. Basic trigonometric identities include tan x sin x + cos x = sec x and cos^2 20+cos^2 70/sin^2 31+sin^2 59+sin^2 64+cos64.sin26 = 2. Applications involve the line of sight from an observer's eye to an airplane overhead, defining the angle of elevation, and the pilot's downward line of sight, defining the angle of depression.
17-MOD 6 Conducting Security Audits & MOD 7 Information Security Audit Prepar...abhichowdary16
油
This document discusses information security audits and their key features. It describes the different types of security audits and phases of an information security audit. It outlines the audit process, including defining the security perimeter, describing system components, determining threats, and using appropriate tools. It also discusses auditor roles and skills, as well as elements that characterize a good security audit like clearly defined objectives and an experienced independent audit team.
20-MOD 8 Self and Work Management-20-07-2023.pptabhichowdary16
油
Establishing good working relationships with colleagues has many benefits, including making tasks easier to complete, increasing productivity, and getting ideas and feedback. It is important to understand one's own scope of work and responsibilities clearly to plan work better, build trust, and reduce conflicts. Organizations have policies and procedures to help with regulatory compliance, optimize performance, and reduce errors. Maintaining confidentiality of information is important for legal and business reasons, and organizations have policies for securely storing and disposing of confidential documents and data.
Class IX Maths Construction Module 1 .pptxabhichowdary16
油
This document provides an overview of the topics that will be covered in Chapter 11 of the Class IX Mathematics textbook, which focuses on constructions. The chapter will introduce basic constructions and how to construct triangles. It will conclude with a summary of the key points covered.
1. The document discusses properties and congruence criteria of triangles. It defines triangles as having three sides, three angles, and three vertices.
2. There are five criteria for determining if two triangles are congruent: side-angle-side (SAS), angle-side-angle (ASA), angle-angle-side (AAS), side-side-side (SSS), and right-angle-hypotenuse-side (RHS).
3. Additional properties discussed include: angles opposite equal sides are equal; sides opposite equal angles are equal; the sum of any two sides is greater than the third side.
1. The document discusses properties and congruence criteria of triangles. It defines triangles as having three sides, three angles, and three vertices.
2. There are five criteria for determining if two triangles are congruent: side-angle-side (SAS), angle-side-angle (ASA), angle-angle-side (AAS), side-side-side (SSS), and right-angle-hypotenuse-side (RHS).
3. Additional properties discussed include: angles opposite equal sides are equal; sides opposite equal angles are equal; the sum of any two sides is greater than the third side.
11-Incident Response, Risk Management Sample Question and Answer-24-06-2023.pptabhichowdary16
油
This document outlines the steps of an incident response process including identification, recording, initial response, communication, containment, response strategy formulation, classification, investigation, and recovery. It discusses strategies for each step such as gathering information, validating incidents, determining appropriate response personnel, containment techniques, and formulating strategies based on business impact and recovery efforts. Common security incidents and appropriate reporting procedures are also addressed.
Air pollution is contamination of the indoor or outdoor environment by any ch...dhanashree78
油
Air pollution is contamination of the indoor or outdoor environment by any chemical, physical or biological agent that modifies the natural characteristics of the atmosphere.
Household combustion devices, motor vehicles, industrial facilities and forest fires are common sources of air pollution. Pollutants of major public health concern include particulate matter, carbon monoxide, ozone, nitrogen dioxide and sulfur dioxide. Outdoor and indoor air pollution cause respiratory and other diseases and are important sources of morbidity and mortality.
WHO data show that almost all of the global population (99%) breathe air that exceeds WHO guideline limits and contains high levels of pollutants, with low- and middle-income countries suffering from the highest exposures.
Air quality is closely linked to the earths climate and ecosystems globally. Many of the drivers of air pollution (i.e. combustion of fossil fuels) are also sources of greenhouse gas emissions. Policies to reduce air pollution, therefore, offer a win-win strategy for both climate and health, lowering the burden of disease attributable to air pollution, as well as contributing to the near- and long-term mitigation of climate change.
Welcome to the March 2025 issue of WIPAC Monthly the magazine brought to you by the LinkedIn Group WIPAC Monthly.
In this month's edition, on top of the month's news from the water industry we cover subjects from the intelligent use of wastewater networks, the use of machine learning in water quality as well as how, we as an industry, need to develop the skills base in developing areas such as Machine Learning and Artificial Intelligence.
Enjoy the latest edition
Integration of Additive Manufacturing (AM) with IoT : A Smart Manufacturing A...ASHISHDESAI85
油
Combining 3D printing with Internet of Things (IoT) enables the creation of smart, connected, and customizable objects that can monitor, control, and optimize their performance, potentially revolutionizing various industries. oT-enabled 3D printers can use sensors to monitor the quality of prints during the printing process. If any defects or deviations from the desired specifications are detected, the printer can adjust its parameters in real time to ensure that the final product meets the required standards.
Preface: The ReGenX Generator innovation operates with a US Patented Frequency Dependent Load Current Delay which delays the creation and storage of created Electromagnetic Field Energy around the exterior of the generator coil. The result is the created and Time Delayed Electromagnetic Field Energy performs any magnitude of Positive Electro-Mechanical Work at infinite efficiency on the generator's Rotating Magnetic Field, increasing its Kinetic Energy and increasing the Kinetic Energy of an EV or ICE Vehicle to any magnitude without requiring any Externally Supplied Input Energy. In Electricity Generation applications the ReGenX Generator innovation now allows all electricity to be generated at infinite efficiency requiring zero Input Energy, zero Input Energy Cost, while producing zero Greenhouse Gas Emissions, zero Air Pollution and zero Nuclear Waste during the Electricity Generation Phase. In Electric Motor operation the ReGen-X Quantum Motor now allows any magnitude of Work to be performed with zero Electric Input Energy.
Demonstration Protocol: The demonstration protocol involves three prototypes;
1. Protytpe #1, demonstrates the ReGenX Generator's Load Current Time Delay when compared to the instantaneous Load Current Sine Wave for a Conventional Generator Coil.
2. In the Conventional Faraday Generator operation the created Electromagnetic Field Energy performs Negative Work at infinite efficiency and it reduces the Kinetic Energy of the system.
3. The Magnitude of the Negative Work / System Kinetic Energy Reduction (in Joules) is equal to the Magnitude of the created Electromagnetic Field Energy (also in Joules).
4. When the Conventional Faraday Generator is placed On-Load, Negative Work is performed and the speed of the system decreases according to Lenz's Law of Induction.
5. In order to maintain the System Speed and the Electric Power magnitude to the Loads, additional Input Power must be supplied to the Prime Mover and additional Mechanical Input Power must be supplied to the Generator's Drive Shaft.
6. For example, if 100 Watts of Electric Power is delivered to the Load by the Faraday Generator, an additional >100 Watts of Mechanical Input Power must be supplied to the Generator's Drive Shaft by the Prime Mover.
7. If 1 MW of Electric Power is delivered to the Load by the Faraday Generator, an additional >1 MW Watts of Mechanical Input Power must be supplied to the Generator's Drive Shaft by the Prime Mover.
8. Generally speaking the ratio is 2 Watts of Mechanical Input Power to every 1 Watt of Electric Output Power generated.
9. The increase in Drive Shaft Mechanical Input Power is provided by the Prime Mover and the Input Energy Source which powers the Prime Mover.
10. In the Heins ReGenX Generator operation the created and Time Delayed Electromagnetic Field Energy performs Positive Work at infinite efficiency and it increases the Kinetic Energy of the system.
Indian Soil Classification System in Geotechnical EngineeringRajani Vyawahare
油
This PowerPoint presentation provides a comprehensive overview of the Indian Soil Classification System, widely used in geotechnical engineering for identifying and categorizing soils based on their properties. It covers essential aspects such as particle size distribution, sieve analysis, and Atterberg consistency limits, which play a crucial role in determining soil behavior for construction and foundation design. The presentation explains the classification of soil based on particle size, including gravel, sand, silt, and clay, and details the sieve analysis experiment used to determine grain size distribution. Additionally, it explores the Atterberg consistency limits, such as the liquid limit, plastic limit, and shrinkage limit, along with a plasticity chart to assess soil plasticity and its impact on engineering applications. Furthermore, it discusses the Indian Standard Soil Classification (IS 1498:1970) and its significance in construction, along with a comparison to the Unified Soil Classification System (USCS). With detailed explanations, graphs, charts, and practical applications, this presentation serves as a valuable resource for students, civil engineers, and researchers in the field of geotechnical engineering.
2. Backup
Why Backup?
If you are not backing up your files
regularly, you deserve to lose them.
Average user experiences loss once a year
3. Backup
What Can Cause Data Loss?
Incorrect software use
Input data incorrectly
Software may harm data
Hard disk malfunctions
Accidentally delete files
Virus infection
5. Full Backup
Complete type of backup operation
make a copy of all the data including files, folders,
settings, applications
On - storage devices like hard drive, SSD, HDD, etc.
files and folders will be backed up again entirely in any
subsequent backup operations redundant copies of data
Needs much disk space
10. Local Backup
backup where capacity medium is kept close within reach
or in structure with source
second inward hard drive, connected outer hard drive,
CD/DVD-More, Network Attached storage (NAS)
11. Offsite Backup
Backup - set in an alternate topographical area from
source
Cloud Backup
FTP Backup
12. Hardware and Software
Desktops and laptops Computer equipment can be damaged, lost, stolen.
Removable media Media degrades over time. Software programs to read
the media change and can become obsolete. Hardware changes over time,
and is not always backward compatible.
Cloud storage
Cloud providers go out of business
Data formats change (what you upload may not be useable
when you download it)
Accidents happen. Data is corrupted, or stolen.
13. File Formats
Think about the ability to use and re-use data in the future. Both for you,
and for others.
Accessibility of future data because of technology changes - proactively
plan for hardware and software obsolescence.
Think about who needs access to your data. Are you collaborating with
someone within the University, or outside of it?
Conducting funded research - be aware of any data storage and data
sharing requirements.
Think about data security.
14. Best Practices for File Formats
Formats most likely to be accessible in the future are:
non-proprietary
open
documented standard commonly used by a discipline-specific
research community
standard representation (ASCII, Unicode)
unencrypted and uncompressed
Remember that comprehensive documentation (metadata) is
essential to accurate use, and reuse, of all data.
15. Data Security and Access Control
Network security
Keep confidential or highly sensitive data off
computers or servers connected to the internet
Physical security
Access to buildings and rooms
Computer systems & files
Use strong passwords on files and systems
Virus protection (updated continuously and running!)
Encryption
16. Data Backups
Reduces the risk of damage or loss
Use multiple locations (here-near-far or 3-2-1)
-Keep at least 3 copies of data
- Store 2 backup copies on different storage media.
- Store 1 backup copy offsite.
Create a backup schedule and put someone dependable in
charge
Use reliable backup medium
Test your backup system (test file recovery, data consistency, data
accuracy)
Editor's Notes
#13: Your desktop or laptop computer should store only your current working copy of your data. Your master copy should be stored in a safe environment, and it should be copied in another location as well. Data redundancy is very important.
Your laptop can break down, the hard drive can crash, your computer can be infected with a virus, your laptop is stolen. What happens if it is the ONLY copy of your data?
Removable media are vulnerable to temperature and humidity fluctuations, poor handling, air, moisture, and light conditions. The program you used to generate or analyze the data could be upgraded by the manufacturer, and it might no longer work with your data. The software has changed, and no longer runs on your computer.
Removable media should be migrated to new media every 3-5 years. Hard drives should be migrated every 5 years. The consistency of the recorded data should be checked yearly.
Cloud storage providers are providing a service for a fee. They are a business, and as such they can go bankrupt, be bought out, or be legislated out of existence. Cloud storage is simply storage. What you put in is what you get out. Dont make the mistake of thinking of them as a long-term storage solution, like an archive. Have another option available, because if they go out of business, you may have a very limited amount of time to move your data to a new provider. You will not be able to move it directly from one service to another.
#14: Plan for the future at the start of any data collection activity. You may need your data in one format to analyze it, and another to visualize it. You may want to share it with a collaborator who needs it in a specific format. Perhaps the software tool you need to use requires a format that only it has. Being aware of these requirements will make the later stages of your work easier and more productive.
Think about how software and hardware change over time. Proprietary software often has a short shelf life. Microsoft has changed Word quite a few times over the years. If you try to open an old 1989 copy of a Word file with the 2010 version, you will get nothing but gibberish. (The first Word program was released in 1983. It was in MS-DOS). Later versions are not always backwards compatible. The same holds true for many proprietary software products. Hardware also changes over time. Operating systems change, hard drive file formats change. The mechanical bits of computers change and break down. The data may be still on an older hard drive, but if the software that reads it is no longer available, or the bearings that spin the disc are frozen, you will never be able to access that data.
Access for your collaborators is something you need to be aware of. If they are at UVa, they are probably using the same systems as you, or the data can be migrated to a format you can both access and use. But if they are in Europe, or Asia, they may not have access to the same equipment or software. This is especially important if you are working with highly sensitive data, such as personal identifiers or HIPPA regulated data. Do your collaborators have access to your data storage environments?
Many funders, especially US government ones, require data sharing as a part of your funding. They may specify specific formats for data sharing and data storage.
Data security is an important consideration. Can you move the data easily? Consider factors such as size, security level, format. What are the requirements imposed by the funder, the institution, or collaborators?
#15: You will probably need to keep your working data in a proprietary format until your data analysis is finished. It is always a good idea to have your primary data file backed up in both the proprietary format you are using for analysis, and a non-proprietary one. Data redundancy is a simple but effective deterrent to unimagined software and hardware failures.
For long-term data storage (archiving) and sharing, you want to select a file format which is non-proprietary for maximum reuse potential. Open formats are best. Convert your data from proprietary formats to open or discipline-defined standards. The following list comes from the UK Data Archive, the preeminent authority on digital research data in the social sciences and humanities. [http://www.data-archive.ac.uk/create-manage/format/formats-table]
Quantitative tabular data: SPSS portable format (.por), comma-separated values (.csv), tab-delimited files (.tab), text files (.txt)
Geospatial data: ESRI Shapefile (.shp, .shx, .prj, .sbx, .sbn), TIFF (.tif, .tfw), Cad data (.dwg)
Qualitative data: XML, rich text format (.rtf), plain text data (ASCII .txt)
Digital Image data: TIFF (.tif) uncompressed
Digital Audio data: Free Lossless Audio Codec (FLAC) (.flac)
Digital Video data: MPEG-4 (.mp4), motion JPEG 2000 (.mj2)
Documentation and scripts: Rich text format (.rtf), PDF/A or PDF (.pdf), HTML (.htm), Open Document Text (.odt)
Dont forget the metadata! Comprehensive and complete documentation is essential so that all future users can understand the data.
#16: The PI, or an assigned team member, should be responsible for maintaining the master copy.
Restrict write access to specific members of the team. Document and update your guidelines (include responsibilities) so if team members change, new members will know the requirements.
Record changes with your working data with Version control.
Network security: Keep highly sensitive, confidential or IP-valuable data off internet-connected servers (or behind firewalls).
Physical security: Who has access to your office and to the data files? Be careful of allowing repairs by an outside company.
Computer systems and files: Keep virus protection up to date, and active. Be sure your computer has a login password. Do not send personal or confidential data via e-mail or FTP. Encrypt your data prior to sending.
If you have IRB requirements, be sure to follow them. You may need to address data storage when acquiring consent and confidentially agreements for data users. Many IRBs require 2 levels of security when working with sensitive or confidential data. This can be a password-secured computer in a locked office, or password-protected data files on a password-protected computer. Encryption is another option to ensure the security of data.
#17: Data redundancy is very important. Regular back-ups protect against disasters and accidental or malicious data loss. Types of problems:
Hardware failures
Software problems
Virus infection
Hacking
Power failures
Human errors
Ideally, your backup should be stored in a different zone: you want to be able to recover your data if something bad happens. If you keep your backup on the same device/media as your primary, or master, then in the case of a disaster you will lose all of it.
3-2-1 Rule: Keep 3 copies of your data. 2 different media (for example - hard drive and optical media). 1 copy stored offsite, ideally in a different threat zone.
Here-near-far: Keep 3 copies of your data. 1 here (your working copy), 1 near (your primary backup), 1 far (your primary data, kept in a different threat zone)
What is a threat zone? A different geographical location from the one you are working in. If a natural disaster occurred in Virginia, and all of your data files are here, then you will probably lose everything. Put your primary, or primary backup, on a media that is secure, and send it to a friend or family member in another state for safe keeping. Can the cloud be considered a different zone? Yes, but only if the servers arent in the same zone you are in.
Backup your data frequently. Copy that backup frequently. Dont touch the primary unless you need it. Ideally, create a schedule for backing up your data, and follow it. Have someone be responsible for the backup schedule. Follow-up.
Periodically test your backed up data for accuracy and consistency. Your schedule should include this. Your schedule should also include the dates when data files are created, so you know when you should migrate to fresh media.
Disaster recovery: If the worst happens, recover your data from the primary backup. You may have to start your analysis again, but it is better than having to collect the data again. If you have been saving and backing up your data on a regular schedule, you should lose no more than your work between backups. This is a great reason to do those backups weekly.