Custom Collections

What are custom collections?

Custom collections are where you can capture any data from your monitored instances and store it in custom user tables in the repository database. DBA Dash will take care of the scheduling for your collection and provide you with a script that you can tweak to handle the data import and storage of collected data. DBA Dash can also take care of the data retention, using partition switching to efficiently remove old data. DBA Dash will also create some sample custom reports to display the collected data in the application.

Depending on what the requirements are for your collection you might be able to use the provided scripts as is or you can tweak them to your own requirements.

Do you want to collect a custom metric? Use this feature instead.

Do you want a custom check for the summary dashboard? Use this feature instead.


⚠️Think about the cost of running your custom data collection. We don’t want the monitoring to become the cause of our performance issues by running an expensive query at high volume. We should also consider the data volumes and impact on the repository database. Consider what schedule and retention is appropriate for the collection. If the data doesn’t change frequently, have it run out of hours.

  • Create a stored procedure that will return the data for you custom collection.

This could be your own script or a community script. You must be able to execute it without any parameters. If you need to supply parameter values either create a new stored procedure that executes the script with the required values or set parameter default values for the stored procedure.

ℹ️ DBA Dash will try to execute the stored procedure in whatever default database is specified on your monitored instances. Typically this will be master, but you can adjust the default database in the Source tab of the service configuration tool if required.

Why a stored procedure?

Custom collections are currently limited to running a stored procedure with no parameters. This provides a degree of extra security as it prevents someone with access to edit your config file from running any adhoc SQL. This only provides limited protection as someone with access to write to the service folder could potentially exploit that to elevate their permissions.

⚠️See the security document for information on how to protect your config file and for other security related advice.

  • Grant the account that runs the DBA Dash service EXECUTE permission on the stored procedure.


Custom collections are configured using the service configuration tool.

ALL Instances

  • Click the Custom Collections button on the Options tab to configure a custom collection for ALL monitored instances.
  • You will be prompted to connect to a SQL instance which will default to one of your existing source connections. Edit the connection details (if required) and click connect. The connection is used for the stored procedure list, the preview button and to get the schema for your collection when you click the Get Script button. Once created, the collection will apply to ALL instances.
  • Goto next steps

Specific Instance

  • Go to the Source tab.
  • Find your instance in the grid.
  • Click the View/Edit link in the Custom Collections column.
  • Goto next steps

New instance

  • On the source tab, click Custom Collections button on the Other tab before adding the instance OR
  • Copy an existing connection that already has the custom collections defined using the Copy link in the Copy Connection column.

Next steps

Custom Collection Configuration

  • Select a stored procedure from the drop down. You can click the Preview button to test.
  • The collection name is set based on the procedure name. Each collection needs a unique name - You can edit this if required. The collection name is used to determine the name of the stored procedure that will handle the data import and the names of other objects related to the collection
  • Set a cron schedule. Use the links for some examples. This either needs to be a duration in seconds or a Quartz compatible cron expression. See here
  • Set the run on service start and override the default timeout if required.
  • Click Add
  • Click the Get Script button.
  • Set a value for data retention.
    Note: If you set a value of 0, partitioning won’t be used and the previous collection will be replaced instead of maintaining multiple snapshots.
  • The script is copied to the clipboard.
  • Review the generated script, edit it as required and run it against the DBA Dash repository database.
  • Click the Update button.
  • Click Save to save the configuration changes.
  • Restart the service to apply the new configuration changes.
  • The sample reports will appear the next time you load the GUI and will be populated with data when the collection runs.

Disable a collection

If you setup a custom collection for ALL instances and want to disable it for a specific instance:

  • Open the custom collections for ALL instances. Click the Copy link for the collection you want to disable. (This just saves the procedure and collection name for us. You can enter manually instead if you prefer)
  • On the source tab, click the View/Edit button in the Custom Collections column for the instance. 💡Tip: Use the search feature to find your instance.
  • The stored procedure and collection name should be set. Click the Disable link to remove the schedule.
  • Click Add
  • Click Update

The collection defined at the instance level will override the server level options for the collection. No schedule is defined so the collection will be disabled.


sp_Blitz is a popular community health check script for your SQL instances. There is some overlap with checks provided by DBA Dash, but sp_Blitz has additional checks. sp_Blitz provides it’s information in a single report with links that provide advice on each check. DBA Dash is more suited to support your daily DBA checks and sp_Blitz is ideal to run when setting up a new server, working on a SQL instance for the first time and to run occasionally for SQL instances.

Capturing sp_Blitz for all your SQL instances is just a case of following the steps above. You can probably set the retention to 0, but the option is available to keep previous snapshots if you prefer. The default reports are OK but with a bit of work you can have a summary dashboard with drill down links like this:

sp_Blitz custom collection

For sp_Blitz you can setup initially using the steps above and replace the reports with the customized versions from here

🔒Note: sp_Blitz requires sysadmin permissions to run. See the Security section in the discussion post for a workaround to run with minimal permissions.

You can find more custom report examples here and also post your own custom collection and report examples.