Asigra BLM Archiving – Align the value of your data with the cost to protect it

Years ago, we treated all data as being equal. All data originated on one type of storage and stayed there until it was deleted. We now understand that not all data is created equal. Some types of data are more important than others, or accessed more frequently than others. Backup Lifecycle Management (BLM), defines the BLM concept where data is created on one storage system, then migrated to less expensive storage systems as it ages.

Asigra Backup Tiers

tiered data

For example:

Data that is 2 minutes old is highly valued.
Data that is 2 months old may be of interest but is not as highly valued.
Data that is 2 years old may be needed for records but it is not critical to the daily functioning of the company.
DS-System – Primary Storage-Business-Critical Operational Data

Business Critical Operation Data contains the files, databases, email systems, etc., that are needed for operations on a day-to-day basis. All data that is critical to business operations data should be stored in the DS-System Tier.

BLM Archiver – Policy based retention of older data

Large file servers or other large repositories of potentially older data can be moved to BLM, Cost savings are the primary benefit by allowing storage of older data and automatic retention policies that move aged data into the lower cost tier. BLM Archiver can also be leveraged to provide storage of past generations of data while keeping the most recent version in business critical DS-System.

Managecast will help with analyzing your data to determine a protection method to best suit your recovery requirements and budget. There are many options to protect business data by strategically identifying the value and aligning the cost to protect it.

BLM Cloud Storage – For Low-Cost, Rarely Retrieved Files

Typically for files older than 1 year, BLM Cloud Storage is a method to cost effectively protect large data sets that are still needed for reference, compliance, and infrequent restores.

Files older than a specified age can be selected to move to long-term cloud storage and are generally grouped in large chucks from 250GB on up to multiple terabytes and then copied to long-term archive on disk.

Customers can utilize Amazon S3 cloud storage or use Managecast Enterprise Cloud Storage

Veeam v 8 Certificate Error When Upgrading (Authentication failed because the remote party has closed the stream)

We were setting up Veeam cloud connect infrastructure in order to provide Veeam Cloud Backup – which many of our customers were asking us to provide. Everything was going well with the installation and we started out with a self signed certificate for testing.  We then applied a certificate from a well-known Certificate Authority and it still worked fine.  We then got a notifacation from Veeam about an available update (v8 Update 3).  It is import to be on the same version or higher as clients so we went to update right away.

Upon updating to Update 3, Clients could no longer connect, getting the following error when trying to connect:

“Error: Authentication failed because the remote party has closed the stream”

This was happening immediatly when connecting and Veeam wouldnt let editing the Cloud repository continue because it did not have a certificate.

First, we tried reapplying the certificate on the service provider side. Although it completed successfully, the clients were still getting the same error.

We then tried just creating a new Self-signed certificate and that didnt work.

We thought maybe the client had to be on the same version, So I upgraded a client to version 3 and still got the same error.

Before updating Veeam, we had taken snapshots of all the Veeam components (Backup and Replication server, Cloud Gateway, WAN Accelerator, and Repository).  We reverted to before the upgrade and the clients could connect whether it was a self-signed certificate or from the Certificate Authority.

There were previous updates available, so we tried with Update 2 as well. Same results.

At this point, we opened a support ticket with Veeam. We uploaded Logs from every component and from the client side.  After they inspected the logs they came back and had us try installing update 2b, sending them logs from before and after the upgrade.  We still had the same results!

After they inspected those logs they sent me a process to try that ultimately worked.

They had us first apply a Self-Signed Certificate on the base installation of Version 8 and then upgrade to 2b, and if the self signed certificate still worked, to then apply the one from the Certificate Authority from the pfx file.

It worked!

We still had Update 3 to get to so I took another set of snapshots and upgraded to Update 3 and everything stayed working.

Hopefully this can save you some time, As I didn’t see this error documented in Veeam’s KB articles or documentation about installing certificates.

Reducing backup cost with Asigra


Data growth continues to expand at an explosive pace. More and more files are created every day. The files often get much larger over time as well. Compliance and other government mandates also require longer retention of past data which means backup data can grow even more.

Fortunately we have a lot of options when it comes to managing the size of offsite backups. This article will help you keep your backups to a reasonable size to ensure you are protecting your valuable data in the most cost effective way possible.

This information assumes you are using Asigra Cloud Backup, but it may also benefit you if using another capacity based backup system.

Not only will fixing ineffeciencies help lower your costs for offsite backup, but can also speed up the time it takes to make backups and reduce bandwidth usage!

Identify largest backup sets

Typically the largest backup sets are the backup sets that typically need the most attention from a space consumption standpoint. It does not do much good to focus our energy on small backup sets.

Run a backup set report from DS-Client (Reports menu, Backup Sets, Print or Print Preview) to determine which backup sets are consuming the most space.

Identify and make a list of the top backup sets based on “Stored Size” Column. Sometimes it’s one large backup set that, or the top 3.  A client with 100 backup sets may have a larger number of top backup sets than a customer with 3 backup sets, but generally your top 20% of your backups will consume 80% of the total space (the good ole 80/20 rule!)

Review backup logs on the largest backup sets

So, how did the largest backup sets become the largest?  Large backup sets can sometimes grow larger than expected, especially over time, because data is being backed up that is not needed. Examples could include (but not limited to):

  • Large SQL or other database dumps
  • Reorganizing file system data that is then detected as new backup data. For example, copying data from old directories to new directories.
  • Daily backup of antivirus signature updates (and old updates never removed)
  • Users copying large amounts of data to the network (family pictures, videos, backups of PC’s)
  • Backups made onto the server be backed up – nothing like backing up backups!

A great way to detect unneeded data is to review the detailed backup logs and review the list of files being backed up. Perform the following:

  1. From the DS-Client, click Logs, Activity Logs
  2. Change the From field date to several months ago to see the past activity
  3. Change the Activity field to “Backup” to only see the backup activity
  4. Change the Node/Set to the largest backup set to see only that activity
  5. Click on Find to see all backups for that specific node
  6. Once you see the list of backups, click on the “Transferred Amount” column twice to put the largest transfers at the top of the list. The list will be listed in order of the largest backup transfers over a given period with the largest transfers listed a the top.
  7. Select the largest backup (based on transfer size) in the list – this should be listed at the very top. Take note of how much data got transferred, the length of the backup and how many files got backed up.
  8. With the largest backup selected, click on the “Detailed log” button.
  9. On the next screen, click on Find to see all of the files that got backed up in that session.
  10. Click on the Size column to put the largest files on top
  11. Review the path and filename that got backed up and verify this is data that needed to be backed up.

Notes: If there are no extra large backups over a given range of dates consider reviewing “average” backups and look for data that doesn’t need to be backed up.

Managecast is highly experienced in reviewing backups. You may always choose to engage Managecast to get assistance with these items as we perform these functions frequently and know what to look for. However, you know your data best, so your intimate knowledge of your environment can also be valuable in determining if you are backing up data efficiently.

Review all backup sets and ask yourself “Does all of this data REALLY need to be offsite?”

For instance, antivirus or other utility servers may be important, but replaceable and do not contain any unique data that needs to be protected offsite. By eliminating offsite backups of these types of backup sets you may be able to reduce the total offsite data.

Consider using “local-ONLY” backup sets for data that needs to be backed up, but not critical enough to justify the cost of off-site backups. However, this may impact your recovery time objective in a significant disaster so make sure you know the pros and cons!

Some other things to check:

  1. Are you backing up Anti-virus definitions/software? Does this data need to be offsite or can you use local-only backups?
  2. Are recycle bin and temp files/folders being backed up? They can probably be excluded.
  3. Consider excluding the “System Volume Information” folder at the root of each disk drive being backed up. This is unneeded data.
  4. Are you excluding certain file extensions such as *.tmp, *.bak, *.mp3, and other file extensions that may represent non-critical data?
  5. Check the retention rules:
    1. Do all of the backup sets have a retention policy assigned?
    2. Is a schedule set to run the retention?
    3. How are you handling removal of deleted files from the backups? Check the retention for handling of removing deleted files from backups.

Are you fully leveraging archive storage?

Asigra provides for 4 different storage tiers to allow you to align the value of your data with the cost to protect it. Archiving can dramatically lower cloud backup storage costs.

Operating system data and applications can be replaced, so by using local-only backup for this type of data you can lower your overall costs.

In addition, older, static, rarely used data can be archive to dramatically reduce costs. To learn more about archiving and the different backup storage tiers, click here.