This is part 3 in a series of posts reviewing the DiskStation DS1621xs+.
In part 1, I unboxed the Synology and configured storage. In part 2, I set up Cloud Sync. Now I’m going to use the Active Backup for G Suite tool to do a full backup of a G Suite.
I have a domain that’s hosted by Google’s G Suite. Synology offer an on-premise backup solution for G Suite - Active Backup for G Suite.
The first time you open Active Backup for G Suite, you’ll be prompted to activate the package by entering your Synology account.
Enter your Synology account email address and password.
And you’re good to go.
We’re now ready to start creating a backup task. First up, enter your domain, email address and service key. I wasn’t sure what a ‘Service Key’ was, but you’ll notice there’s a hyperlink right there in the description that takes you a step by step tutorial on configuring G Suite and creating the service key.
I followed the tutorial through and created the service key JSON file that I then selected for this step.
Next enter a task name and confirm the destination folder. Unlike previous configurations, the only choice is to pick the top-level directory. The task will create a subdirectory under this location for you.
You can then select services to be automatically added should new G Suite accounts be created in the future.
You can choose whether to run backup continuously, or on a regular schedule. You also can set whether to keep all versions of files, or just a specific number of days.
Finally, a summary of all the settings before you finally create the task.
And then your task is created! Pretty easy really.
If you configured it to run on a schedule, then you can then choose to start your first backup immediately (otherwise I assume it will start at the specified time).
You can monitor the backup progress from the Task List page.
Here’s an example having the Active Backup for G Suite run on a nightly schedule. The first couple of times were quite large, but the most recent backup was quite small, which makes sense if it’s just the changes since the previous backup. Looks like I possibly had a network outage the previous night which interrupted that backup, but everything is green now.
Now backups are great, but they’re essentially useless unless you can restore from them.
You can explore the backups using the Active Backup for G Suite Portal.
It is possible to allow multiple users to have access to the portal, but in my case I’m just using the admin account. You can select which Google account’s services you’re inspecting and (as per the screenshot) you can switch between the different services (eg. Drive, Mail, Contacts and Calendar).
The timeline across the bottom allows you to choose the point in time to restore from.
To restore an item, you simply select it, and then choose the ‘Restore’ button. There’s also the option to restore the entire service (eg. mailbox).
You get a chance to confirm the restore action.
And then watch the restore progress.
Checking the inbox of the relevant account, I could then see the email ‘magically’ restored!
Handily it also adds a label on the email so you can easily identify it was restored.
In summary, I think this is a really useful feature that I plan to keep using. I can also imagine it would be valuable for organisations that need to retain data to comply with legal requirements.
Now that the Synology is up and running with disks configured we can start installing some extra features. First on my list to take a look at is Cloud Sync.
I think this could be useful for a small business that wants to have a second backup of content that employees might have in OneDrive or DropBox, or as a way of aggregating that content locally. The aggregation idea could also make sense if for some reason you had a mix of services in use too. The list of supported services is impressive (and to be honest I haven’t even heard of some of these):
- Alibaba Cloud OS
- Azure storage
- Backblaze B2
- Baidu Cloud
- Google Cloud Storage
- Google Drive
- Google Shared Drive
- hicloud S3
- JD Cloud OSS
- Microsoft OneDrive
- Microsoft OneDrive for Business
- Microsoft SharePoint
- OpenStack Swift
- S3 storage
- SFR NAS Backup
- Tencent Cloud COS
- Yandex Disk
I’ve already added a OneDrive account. You can see it shows a nice summary that everything is up to date and even shows how your usage for that specific account is going. In this case my OneDrive has 1 TB capacity and I’m currently using 22.66 GB.
Adding a service is as easy as clicking on the “+” button from the Cloud Sync app.
Then I select the specific provider I want to add. I’m going to add a Dropbox account.
A new browser window opens which allows me to enter my credentials and authorise my Synology device to connect to my Dropbox account.
I’m now given an opportunity to specify where on the Synology I want to store the synchronised files. I clicked on the folder icon in the “Local Path” field
I can select an existing folder or create a new subfolder. I tried to create a new folder:
But I got this error message. I’m not sure why this is the case as my account is a full administrator on the Synology server.
I was able to work around the problem by going to the File Station app and successfully creating the folder there. Once it was created then I could select it.
I can also specify what kind of synchronisation to perform. The default is bidirectional, but might also want to pick another option like ‘Download remote changes only’ which would make sense if you’re planning to use the Synology as a secondary backup for your cloud files.
Finally a chance to review all the settings before you hit Apply to finish setting up the new sync task.
This is a really simple, straightforward process that I was able to repeat for multiple different cloud providers.
I’ve done a few hardware reviews on my blog, but one is a little different in that Synology have sent me one of their units to review. Specifically their DiskStation DS1621xs+, which is aimed at small and medium-sized businesses.
This is not the first Synology I’ve used - we actually had one at my last office as a local file server (I believe that was a DS220j), but it had been set up before my time. So I’ll be starting from scratch, and as much as possible I’ll be attempting to wear the ‘small to medium business IT/support person’ hat to consider how well it will fit their needs.
There’s a lot to cover, so I’m going to break the review down into a number of parts that I’ll be publishing over the next few weeks:
- Installing and configuring (this post)
- Cloud Sync
- Active Backup for G Suite
- Virtual Machine Manager and Docker
- Other extras
And before I go any further I should also point out some great posts that my fellow Adelaide-based Microsoft MVP Adam Fowler has done on Synology and their software. In particular I’ll call out his review of Active Backup for Microsoft 365
But let’s begin! In the box was a quick installation guide, power cable, two network cables, drive tray keys, some screws (for mounting 2.5” drives), the DiskStation unit and two complementary SSDs.
To facilitate the evaluation I purchased two Western Digital Red NAS 4TB (WD40EFRX) hard disks (which conveniently happen to be listed on the Synology Products Compatibility List). I also decided to add a couple of older spare drives I had sitting around (1TB and 1.5TB). I wasn’t sure whether one or both of these might be faulty - I was interested to see what the Synology would make of them.
The Quick Installation Guide has pictorial instructions showing how to install hard disks into the 6 drive bays. I noticed it didn’t mention SSDs (and I couldn’t see anywhere obvious for them to go). I discovered that there’s a whole lot of documentation online as well, including the Hardware Installation Guide. Chapter 2 of that guide includes a section “Add an M.2 NVMe SSD to Synology NAS”, which to summarise has you removing all the drive bays so that you can then slide in the SSDs onto what turns out to be the side of the case.
With those installed it was time to add the hard drives. That was quite straight forward, again following the instructions.
The DS1621xs+ comes with three network ports - one 10 Gigabit Ethernet and two Gigabit Ethernet ports. It’s nice to see they’ve catered both for higher capacity networks as well as allowing for some network redundancy. There’s also 3 x USB3 ports (1 on the front, 2 on the back) for connecting external drives, or if you’re serious about expansion there’s eSATA ports into which you can connect Synology Expansion Units.
All of this allows for up to 96 TB (6 bays with 16TB drives) or up to an impressive 256TB with expansion units.
With the first two drives installed, it’s time to start everything up. Power and network cable in, and press the power button on the front of the unit.
My first reaction was ‘wow, that’s loud!’. The fans kicked in and reminded me a bit of one of those rack-mounted servers that would normally live in a dedicated server room. (I was relieved to find that later on in the installation process the fans became a lot quieter and remained so.)
Rather than continue reading the instructions at this point, instead I looked up the devices on my network and found that ‘diskstation’ was now registered. Hitting that with my browser brought up this welcome screen:
I guess I’d assumed that the DiskStation would be all ready to go, but it turns out the first part of the process is to actually download and install DiskStation Manager (DSM) - the Synology’s operating system. An advantage of this is that you’re always going to be running the latest version at the time of configuration.
At this point I’d just added the two WD 4TB drives, so it was good to have confirmation that any data on them would be removed:
The 10 minute estimate was pretty accurate. It then restarted, and I got this screen:
That will teach me not to read the instructions! Checking the network and I found ‘synologynas’ was now the name. Changing to that URL and I was back in business.
Next a prompt to customise the server name and supply a username and password, followed by the option to set up QuickConnect (and associate it with a Synology account). I wasn’t quite clear what ‘QuickConnect’ was. A link to further information would have been useful at this step. I found more details in the Synology Support Center. If you skip this step you can set it up later.
And with that, the main DSM desktop is displayed!
A number of times I found myself wanting more information. I should point out that most applications have a ‘?’ icon in the title bar that does link to context-sensitive help. What seemed to missing is a similar feature when various wizards operate.
I then had a security advisor screen show that even with the latest version of DSM installed, two of the bundled packages had additional updates pending.
Those packages were then updated.
At this point if you try to do much more you’ll quickly discover that the next thing that must be addressed is storage. Our disks actually need to be configured into a storage pool. Opening Storage Manager shows that while everything is ‘healthy’, there aren’t actually any volumes or storage pools. The two drives are listed as ‘unused’.
What is a storage pool? Good question. There’s a bit more of an explanation over in the online help. A storage pool has one or more “volumes”. You need to create the pool first, and then you can create volumes within it.
Creating a new storage pool starts with the Storage Pool Creation Wizard. The first thing to decide is which type of storage pool to create: “Better performance” or “Higher flexibility”. Well both of those sound desirable to me. A link to more info in the wizard would have been helpful, but I did manage to find more details under Create a New Storage Pool in the online help.
My reading of that online help is if you’re just planning to have a single volume in your pool, then the “Better performance” option is right. If you want to have multiple volumes in the pool, then choose “Higher flexibility”. Keeping things simple works for me, so I picked the single volume option.
Next you need to decide what type of RAID this pool should use. The UI is smart enough to alert you if you pick a RAID type that is incompatible with the remaining unallocated drives. eg. RAID 5 requires a minimum of 3 drives (and they should all be the same size), so with just two drives currently mounted that wasn’t one to pick for me.
In fact if you’re not sure which one to pick (and I wasn’t clear what the difference was between some of the options), the RAID Calculator might be worth a visit.
The DS1621xs+ supports the following RAID types:
- RAID 1
- RAID 5
- RAID 6
- RAID 10
- RAID 0
- RAID F1 (intended for SSDs)
I was surprised to learn that ‘SHR’ (Synology Hybrid RAID) is not on that list. I suspect this is related to the target market. A small/medium business should be in a position to afford enough disks to pick one of the industry standard RAID levels (5/6/10). But bear that in mind - if you were hoping for the flexibility that SHR offers then another model might be a better choice.
Next up you can assign available drives to this pool. Depending on the RAID type you’ll be prompted to assign the minimum number of drives (and can assign more if any are available).
Next you can do a drive check. If you trust your drives (eg. if you’ve used them before without issue) I guess you could skip this, but I wanted to be thorough. It does take a while though, from a few hours to possibly even overnight depending on how fast and/or large your disks are.
Once you’ve selected all those settings you get a chance to confirm you want to proceed.
The pool is now being created, with the drive verification going on in the background.
In my case it took around 2.5 hours for both 4TB drives to be verified (and pleasingly no problems were found!). Once the pool is available you can create a volume. I just created one that took up the entire pool, and chose the ‘Btrfs’ file system. Nice that the Volume Creation Wizard is one case where it does provide a link to extra info on file systems.
I had been wondering when I’d be able to make use of the SSD drives. Using SSDs for caching is a feature that isn’t found on every Synology model, but the DS1621xs+ is on the list and one of the few that comes with built-in M.2 slots (so you don’t use up a SATA drive bay).
Creating an SSD cache involves selecting the SSD drives and associating the cache with a specific volume. There’s also the SSD Cache Advisor. I suspect the advisor is more useful if you’re adding a cache to a system that’s been up and running for a while as then it will have some history to base its analysis on. In my case as my volume was newly created it just recommended a really small cache.
One thing to watch out for with caches: if you want a read-write cache, you’ll need to have at least two SSDs. That’s because they must be configured as RAID 1, 5 or 6. If you only want a read-only cache then you can get away with a single SSD. This isn’t really called out in the SSD Cache Creation Wizard but is mentioned in the SSD Cache help page.
I have to select both drives.
With two SSDs and the earlier choice of a read-write cache, RAID 1 is my only option:
A chance to confirm how much space to allocate. I chose the maximum.
I’ve included a screenshot of the SSD cache status here. (This was actually taken a few days after the initial setup - I wanted to give it a chance to see a bit of data, though as you can see my usage so far is still pretty light)
The final thing I wanted to do was go back and configure those other two drives. I decided to put these drives in a separate storage pool as I wasn’t sure how good they were. I went through the same process using the Storage Pool Creation Wizard. When I got to the “Choose drives” step and dragged the first available drive over, I got this popup.
That’s a helpful warning, and in a business setting if I saw that I’d be inclined to destroy that drive and replace it with a new one. It isn’t worth the risk. If you really wanted to keep using the drive then this is definitely the time to make sure you have ‘Perform drive check’ selected.
With the volumes in place and caching enabled things are looking good. Next up we’ll go and explore the Package Center and start adding some useful functionality.
Just look at all the packages to choose from (and this doesn’t include the 3rd party ones)!
Note that links to Amazon are sponsored links
My new laptop came with a 512GB SSD. As I mentioned previously, you can select a higher capacity when you order but I decided to pick the lowest size and then upgrade it myself with a 3rd-party SSD, specifically a Samsung 970 EVO Plus SSD NVMe M.2 2TB.
I chose the 970 EVO Plus as it was the largest capacity SSD that Samsung makes that is compatible with the Dell Latitude 7420 with a price that was still within budget. Specifically it is an M.2 2280 form factor with a PCIe Gen 3.0 x 4 NVMe 1.3 interface.
Special thanks to fellow MVP and SixPivot colleague Bill Chesnut who lent me an SSD caddy so I could use disk cloning software to clone the data from the old SSD onto the new 970.
Accessing the existing SSD in the 7420 is pretty easy. There’s a few standard Phillips-head screws to loosen to take the base off (yay, no Torx screws like my 9550 had!), then two smaller screws to remove the cover/heatsink of the existing SSD.
Pull out the old SSD and slide in the new one, screw everything back in place and power up.
Strangely, the first time I did this, Windows was a little unhappy and decided my disk had errors and was going to spend a really long time fixing them.
I stopped that after a while (as yes, it was serious when it said it might take over an hour!) and decided to try the clone again. This time just to be safe I ran a chkdsk (specifcially
chkdsk /f /scan /forceofflinefix /b c: on the original disk just to ensure everything was good before cloning. The second clone was fine and Windows booted up just like you’d expect, but instead of a 256 drive, I now had almost 2TB!
So it’s obviously bigger, but is it better?
I used HDTune to run some benchmarks to find out. All of these were run with as few other processes/services running, A/C power was connected and laptop performance configured to run at ‘Maximum performance’.
Here’s the original SSD. I must say I was a bit surprised to see the variations here - quite inconsistent.
Dropping in the Samsung SSD (with no other changes), and things are looking a bit better already.
|Transfer Rate Minimum
|Transfer Rate Maximum
|Transfer Rate Average
But can we do any better?
Samsung NVM Express Driver
Samsung make the Samsung NVM Express Driver available to install. Would that eak out a few extra MB/sec? I tried installing it but it failed with an error that it wasn’t compatible with the current hardware.
Turns out if your device is in ‘RAID’ storage mode (as mine was set to by default), then it won’t install. You need to switch to ‘AHCI’ mode first. That setting change must be made in the BIOS (UEFI) settings:
Historically making this kind of change might cause you to end up with an unbootable system, but in my case Windows was able to automatically load the ‘Standard NVM Express Controller’ and the system booted correctly. (Truth be told I did actually boot into Safe mode after making the change, but I didn’t need to do anything specific once I got there so I suspect it wasn’t really necessary).
Restarting back up and we’re now running in AHCI mode. Let’s just check if there’s any change in performance from just doing that:
|Transfer Rate Minimum
|Transfer Rate Maximum
|Transfer Rate Average
Pretty much in the same ballpark as before. But now I was able to successfully install the Samsung NVM Express Driver:
After another reboot, time for the final HDTune benchmark..
|Transfer Rate Minimum
|Transfer Rate Maximum
|Transfer Rate Average
Nothing dramatic, but a small improvement, which I’m happy to take 😀
Samsung also have their ‘Samsung Magician’ software that you can use to configure and review your Samsumg SSD drives. It also comes with a benchmark. No idea how accurate it is but here’s the results:
Interestingly it does seem like more features and information are available in Samsung Magician once the Samsung NVM Express Driver is installed.
So in the end, I think it turned out well.
This morning I was trying to create a SQL client alias. An application assumes my local SQL Server instance is named ‘SQLEXPRESS’ but I’ve installed SQL Server 2019 Developer Edition (which installs as the default instance).
I opened SQL Server 2019 Configuration Manager and created an alias.. It didn’t work. I can never remember whether to use the ‘32bit’ configuration or the other one! I tried both, still no success.
I checked Google to see I was doing it right. Yep, tried this and still no luck.
Let’s double-check how my local instance is configured..
Whoops! Aliases can only use ‘Named Pipes’ or ‘TCP/IP’ and both of those protocols are disabled.
Let’s enable TCP/IP
And enable binding to 1433 on localhost for IPv4 and IPv6. I won’t bother with the other IP addresses as I’m not planning to share this instance externally.
Restart the SQL service and we’re all good.
Oh.. and for my own reference, the ‘32bit’ client alias did the trick for SSMS, but for a .NET application I needed also add a non-32bit alias too.