Skip to content

langchr86/mini-nas

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

76 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

mini-nas

Ansible playbook and vagrantfile to deploy simple NAS OS based on ubuntu with advanced btrfs features.

Introduction

This is an example of a simple NAS installation targeted on private networks where private data should be centralized and backups from client machines stored on.

This example shows the capabilities of the used technologies and how to setup a simple but powerfull setup. Even though this setup is not complete and should be considered incomplete for real-world usages.

The whole point is to save all the important data in a secure place that is resilient to hardware, software and user failures. This covers e.g. the following use cases:

  • Harddisk reaches end of live and generates write/read errors or dies instantly.
  • Natural bit-flips (bitrot) occur in some file (where the corresponding application cannot handle this false data).
  • User deletes or overwrites data that is still needed.

To not need to implement mechanisms against such data loss scenarios on multiple PCs/Servers we collect all the data of all such clients on one machine which provides us with the full blown feature set against data loss.

Used technologies

  • advanced filesystem: btrfs
    • atomic snapshots (created by 3rd party script: btrbk)
    • checksum over each file (protects against bitrot)
    • RAID0 or RAID1 (do not use 5/6)
  • widespread fileservice: smb
    • used in windows, mac and linux environments
    • samba implements this protocol
    • provide snapshots to users via shadow-copy
  • collecting backups: rsync
    • mature and simple tool to copy data between hosts
    • synchronize a destination folder (backup) with the state of the source (client)
    • fast because of multiple algorithms/mechanisms to transfer only changed memory blocks
  • monitoring: S.M.A.R.T.
    • hardware feature of all kind of disks
    • provides statistic data (e.g. age, power on hours, read error rate, etc.)
    • some self made scripts to get a fast overview of the current state of the filesystem/hardware

Getting started

This example project consists of mainly two parts:

  • vagrantfile: definition of a virtual machine with all needed hardware parts to simulate a NAS
  • ansible playbook: setup definition which installs all services/tools/configs used to provide the described features.

First you need to install vagrant and virtualbox. The Vagrantfile is not compatible with other VM provides because of virtualbox specific configuration.

# windows (using Chocolatey)
choco install -y git.install vagrant virtualbox

# linux
sudo apt install git vagrant virtualbox

Then you need to clone this project somewhere on your disk:

git clone https://github.com/langchr86/mini-nas

And startup vagrant to setup the virtual machine and install all the tools using ansible. Ensure you have a console with admin privileges.

cd mini-nas
vagrant up --provision

After that step you have a full working NAS running in a VM.

Customization

If you want to customize the setup you can change the following files:

  • Vagrantfile
  • ansible/playbook.yml

If you only changed the playbook you mostly only need to re-run ansible with:

vagrant provision

If you changed some VM settings or more intrusive ansible things you will need to destroy the VM and start from scratch:

vagrant halt
vagrant destroy -f
vagrant up --provision

Usage

Connect samba share

Now your virtual NAS is running and it is time to access the SMB share. This can be done by opening an explorer and access the URL: \\mini-nas. If this does not work you need to use the IP of the VM. Evaluate it with connecting to the VM with vagrant ssh and then get the IP address with ip a.

The share is password protected. Use one of the defined users in the playbook (or your own defined). All users are set by default with the exact same password as the user name. If you change the password of the samba user later the ansible role will not change it.

Access old snapshots of data

The samba client allows to access so called shadow-copy versions of your files. These can be accessed e.g. in windows directly by using the context-menu: Restore previous versions. This is possible on each file or subfolder existing in your share. At least if there is a snapshot with a previous version of this file/folder. Snapshots are created with the btrbk utility by creating read-only snapshots of a btrfs subvolume. All the created snapshots are directly accessable under /mnt/pool-main/snapshots. The snapshots are automatically cleaned-up, each time the tool runs. When the tool is executed is controlled by: timer_OnCalendar: "*-*-* *:00,30:00". In this case it is run each 30 minuntes. This means we get a new snapshot each 30 minutes. The cleanup is controlled by:

snapshot_preserve_min: "1h"
snapshot_preserve: "24h 14d 5w 3m"

In this case we keep all snapshots not longer then 1 hour. Note that cleanup will only happen at full hours. This means intermediate snapshots may exist at most 1 hour and 59 minutes. After that we keep 24 hourly, 14 daily, 5 weekly and 3 monthly snapshots.

Data redundancy

We use RAID1 configuration for data and metadata in the btrfs volume. This means not only the management data of the filesystem but also the user data on the volume is organized that each data block is contained 2 times on individual physical hard disk.

This means that each block exists on both used disks in our setup. But RAID mechanisms do not provide some kind of backup but only high availability. So if one drive dies we can still access and work with our data. But if the user does delete a file which is needed later the data is erased on both disks.

To secure users against such mistakes we have the previous described snapshot mechanism which let us access a defined state of the whole samba share at a given time. So the primary mechanism to fight data loss are the snapshots.

RAID1 is optional but gives us more security and in addition helps when bitrot happens. In such a case the filesystem detects errors in data blocks by using checksums. If we have a RAID1 configuration the filesystem corrects the defective block by using the sibling on the other disk.

Disaster backups

A third layer of security can be added by adding a third disk which uses a different filesystem and collects regularly a complete state of the btrfs volume. This is implemented by a simple rsync synchronization. The different filesystem is used to be secure to systematic failures that could happen because of a bug in the btrfs implementation. The disaster copy is mounted under /mnt/backups/share-main/ and can also be accessed by a read-only samba share.

Client backups

As described in the introduction we want to leverage the introduced mechanism for other hosts too. This is implemented by backing-up client data to the share-main on the NAS. Like that we get all the snapshots and other mechanisms for these backups too.

Therefore the rsync-daemon is running on the NAS to which clients can send their backups with rsync. A simple and portable client application for this is qtdsync. This is a simple Java-GUI with a MinGW compiled version of ssh and rsync.

To setup a backup job we have to use these settings:

  • Configure the destination as rsync and use mini-nas/clang as host.
  • Use the qtdsync user and the corresponding password for authentication.
  • As path you can choose the host name. In this example lang-ct2014.
  • All the needed subfolders need to be defined in the rsync-daemon role too.

With the scheduling you can control how many times your data is synchronized to the NAS. To access old/lost data from your client hosts simply search them in the share-main. Each user can only access its own backups.

Monitoring

btrfs

To see an overall status of a btrfs volume we can use:

sudo btrfs-status /mnt/pool-main/

This is also executed automatically once per day by the systemd timer btrfs-status-pool-main and the results are stored in /mnt/pool-main/subvolumes/share-main/btrfs-status-pool-main.log.

To see all space used by snapshots of one volume we can use:

sudo btrfs-snapshot-quotas /mnt/pool-main/

S.M.A.R.T. and general disk status

It is important to not only monitor the health of the filesystem but also of the disks. Most modern disks provide the so called 'S.M.A.R.T.' interface to read out statistics that can help to detect when hardware failures will happen.

To read out those statistics for all drives we can use the script:

sudo disk-status

This gives us the current disk temperatures and a short health report at the top. After that the extensive statistics report of all disks follows.

In addition we can run short or long self-tests on all disks with:

sudo smart-short-test
# or
sudo smart-long-test

The status script and the short disk self-test will also automatically run daily and log the statistics into the configured disk_monitoring_status_log_path.

Maintenance

To keep your btrfs volume clean and ensure data integrity you should execute some housekeeping approx. once per year:

Scrubbing

This checks all checksums and corrects possible errors by using redundant copy in RAID1 configuration. If no RAID1 an error is printed. This is the main mechanism against bitrot. This process needs a long time and is really I/O heavy.

# start and control scrub process.
sudo btrfs scrub start /mnt/pool-main
sudo btrfs scrub status /mnt/pool-main
sudo btrfs scrub cancel /mnt/pool-main
sudo btrfs scrub resume /mnt/pool-main

# check status (keep all those open in tmux panes)
watch -n 5 sudo btrfs scrub status /mnt
watch -n 5 sudo btrfs scrub status -R -d /mnt
watch -n 5 sudo cat /var/lib/btrfs/scrub.status.<UUID>
systemctl -f

Defragmentation

Helps to optimize free space by moving data of mostly empty blocks into others. Therefore frees blocks and allows to store big files in less blocks.

# try to rebalance blocks (data / metadata) with less then 80% usage
sudo btrfs balance start -dusage=80 /mnt/pool-main
sudo btrfs balance start -musage=80 /mnt/pool-main

Replace failed disk

See: Replacing failed devices

Real hardware

If you want build up a real NAS with all these features this should be really simple. Install an ubuntu distribution (e.g. ubuntu-server or armbian) on your hardware, install newest ansible, clone this repository, customize some things in the playbook (mainly the harddisk paths) and let ansible do its magic:

sudo apt-get install software-properties-common
sudo apt-add-repository ppa:ansible/ansible
sudo apt-get update
sudo apt-get install git ansible -y
# or
sudo apt-get install python3-pip
sudo pip3 install ansible

git clone https://github.com/langchr86/mini-nas
cd mini-nas

nano ansible/playbook.yml
# enable disk-monitoring role
# make own customization

./run-local.sh

Known issues / future features

  • Playbook is not idempotent because the samba role does create its config file via template mechanism and the btrbk role does manipulate this.

Presentation

This example was part of a presentation that shows the danger of personal data loss and a simple and inexpensive way to create a save place for all private data in a private network.

Slides: Data Safety and Backups

References

Some parts of my personal setup are published in other github projects:

License

MIT

Versioning

There exists no version numbers, releases, tags or branches. The master should be considered the current stable release. All other existing branches are feature/development branches and are considered unstable.

Author Information

Christian Lang lang.chr86@gmail.com

About

Ansible playbook and vagrantfile to deploy simple NAS OS based on ubuntu with advanced btrfs features

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published