🌙
 

Subscribe to the Taegis™ XDR Documentation RSS Feed at .

Learn more about RSS readers or RSS browser extensions.

On-Premises HA Data Collector

data collectors integrations on-premises secureworks


A Highly Available (HA) Collector is a collector deployed across multiple hosts. HA Collectors are deployed in active-active, multi-master configurations. Because of this multi-node configuration, it eliminates a single point of failure. This means all nodes are up and running and if a node goes down, the other nodes detect the failure and compensate.

HA Collectors consist of three nodes. At least two nodes must be active for the collector to be online.

There are two deployment models for on-premises collectors:

For DHCP collectors, DHCP reservations must be created so the IPs don’t change once they are assigned.

The same firewall rules as non-HA Collectors must be in-place for HA Collectors, as well as additional rules for inter-node communication. The following ports must be open between the collector nodes:

Because the cluster runs in a multi-master configuration, it’s important that all the nodes on the cluster have low-latency connections to one another.

Note

Third-party tools or applications cannot be installed on any XDR Collector.

Create and Download Your XDR HA Collector

You can preconfigure, create, and download an on-premises HA Collector for your environment directly in Secureworks® Taegis™ XDR from Integrations→Data Collectors.

  1. From the XDR left-hand side navigation, select Integrations→ Data Collectors. This page displays the collectors your organization has configured.

  2. Select Actions→Add Collector from the top right. The Add Collector modal displays.

Add New Collector

Add New Collector

  1. Select On-Premises as the collector type and then select Next.

  2. Complete the following fields:

  1. From the HA Cluster drop-down menu, choose Yes.

  2. For HA cluster with static IP address configuration, select Static, choose Configure, and follow the steps in the following Static section. For HA cluster with DHCP configuration, select DHCP and follow the steps in the following DHCP section.

Static

  1. For static IP configuration, you must configure three nodes; for each node, fill in the following:

    • Hostname — The hostname of the collector
    • NTP Servers — (Optional) Specify your own NTP server Domain Name or IP address using the following format: pool.ntp.org or xxx.xxx.xxx.xxx
    • IP Address — The IP Address of the node
    • Subnet Mask — The netmask of the node
    • Default Gateway — The default gateway for the node
    • Preferred and Alternate DNS Server — The DNS servers for the node
  2. When all required fields for all nodes are complete, select Create Nodes and proceed to Step 9.

Add Static HA Collector

Add Static HA Collector

Dhcp

  1. For DHCP configuration, fill in the following:

    • NTP Servers — (Optional) Specify your own NTP server Domain Name or IP address using the following format: pool.ntp.org or xxx.xxx.xxx.xxx
    • Hostname — The hostname of the collector
    • IP Range — The CIDR block that specifies the range of addresses the collector nodes are allocated from. Enter one or more CIDR blocks separated by commas. If the exact IP addresses are known, then the IP addresses can be entered instead of CIDR blocks. CIDR blocks and IP addresses can be intermixed as necessary by separating multiple entries with commas. This range is used to adjust collector settings so that node-to-node communication works. Examples:
      • If the DHCP address range is 192.168.22.0 - 192.168.22.255, then the CIDR block should be: 192.168.22.0/24.
      • If the nodes could be in either of two ranges like 192.168.22.0 - 192.168.22.255 or 192.168.15.0 - 192.168.15.255, then you can specify both CIDR blocks separated by a comma: 192.168.22.0/24,192.168.15.0/24
      • If exact IP addresses are known, then you can specify those as well: 192.168.22.15,192.168.22.16,192.168.22.17
  2. When all required fields complete, select Create Collector and proceed to Step 9.

Add DHCP HA Collector

Add DHCP HA Collector

Note

Default and custom NTP settings are only used during initial Data Collector setup. Once connectivity is established, the Data Collector synchronizes time via the XDR backend connection.

  1. The Install Collector modal displays the following files available for download:

    • One or more .iso that contain the configuration files for your collector; one .iso for DHCP configurations used for each host, and three for static configurations, one for each host.
    • An .ova that is the virtual machine that the collector runs in.
    • A .vhd, which is a .zip collection (downloaded filename is: ctpx-collector.zip) of Microsoft HyperV disk images. This is an alternative to the .ova file.

Save these files; you’ll use them during the installation process described below.

HA Collector Files for Static and DHCP Configurations

HA Collector Files for Static and DHCP Configurations

Installation

The On-Premises Data Collector is a virtual machine appliance that must be installed in your hypervisor environment in order to collect data and transmit it to the XDR Infrastructure. It can be preconfigured and downloaded in XDR from Integrations > Data Collectors and installed in a vsphere and/or hyperv environment. Once the appropriate information is provided, the collector is customized, built, and configured to DHCP or static IP addressing depending on your selection.

Once complete, an .iso cdrom image containing your client certificate/credentials and disk image in the form of .ova (vsphere) or .vhdx (hyper-v) is available for download from XDR. You are required to attach the .iso (cdrom image) to the collector VM on boot.

Once booted, the three-node appliance registers with XDR and the status of the connection displays in the XDR Console.

Note

Recommended virtual environment versions for the XDR On-Premises Data Collector are vSphere ESXi 6.7 or later or Hyper-V 8.0 in Windows Server 2016 or later.

Data Collection and Network Access Requirements

The following is a reference architecture for data collection and provides an overview of the collector’s network access requirements.

Regions

XDR Regional Configuration

Some configuration specifics of XDR depend on the region you are deployed in (US1, US2, US3, EU).

HA Data Collection Architecture

HA Data Collection Architecture

The collector acts by default as a Syslog forwarder and collects security log data. All logs that are sent to the collector are collected and transmitted using rapid batching to the XDR infrastructure. The collector listens and transmits data on the following ports, and you must allow access from inside your organization to outside on the following hostname and ports.

When configuring an HA Collector for syslog, a load balancer must be configured to provide a Virtual IP for all incoming syslog traffic. Behind the load balancer are three collector nodes. Each collector node accepts syslog traffic. The load balancer forwards the traffic from its VIP to the three downstream nodes. It accepts UDP traffic on port 514 and TCP traffic on Port 601 and forwards it to the same ports on the three collector nodes.

Note

All Syslog data is forwarded to XDR by way of a secure mTLS connection using TLS 1.3.

Taegis™ XDR Collectors forward Syslog data in batches. The frequency is optimized according to batch size and time since last forward.

Hostname and port configurations are as follows:

Outbound https on Port 443 — TCP

XDR API

NTP servers: 123/UDP Outbound

The XDR On-Premises Data Collector uses the following standard NTP servers:

Note

Windows DC servers cannot serve as NTP servers for non-domain hosts such as the On-Premises Data Collector.

Inbound—Syslog

Outbound Device APIs

Inter-Node Ports

The following ports must be open between the collector nodes:

Connectivity Requirements for Data Collectors

Regions

XDR Regional Configuration

Some configuration specifics of XDR depend on the region you are deployed in (US1, US2, US3, EU).

Any device that uses its own SSL certificate, including Cloud-based and On-Premises Data Collectors, must safelist the following destination IP addresses or domains in order to avoid conflict. If using an AWS data collector, please refer to the AWS table.

For Most Data Collectors

Source Destination Port/Protocol Notes
Data Collector IP or hostname
US1
collector.ctpx.secureworks.com
18.217.45.178/32
3.16.4.173/32
18.224.219.97/32
13.59.146.90/32
3.16.16.254/32
18.223.74.238/32
US2
collector.delta.taegis.secureworks.com
52.14.113.127/32
3.141.73.137/32
3.136.78.106/32
US3
collector.foxtrot.taegis.secureworks.com
44.229.101.49
35.166.77.47
34.214.135.78
EU
collector.echo.taegis.secureworks.com
18.158.143.139/32
35.159.14.37/32
52.59.37.234/32
TCP/443 Safelisting device access to XDR
Data Collector IP or hostname NTP severs IP/Hostnames provided during provisioning UDP/123 Safelisting device access to NTP servers

This rule is only necessary when custom NTP servers are provided during provisioning.
Data Collector IP or hostname 0.pool.ntp.org
1.pool.ntp.org
2.pool.ntp.org
3.pool.ntp.org
UDP/123 Safelisting device access to default NTP server.

This rule is only necessary when custom NTP servers are not provided during provisioning.
Data Collector IP or hostname DNS server IPs provided during provisioning UDP/53
TCP/53
Safelisting device access to DNS servers

Note

If using local NTP, the access must be safelisted both to and from the data collector on those networks.

For AWS Data Collectors

Source Destination Port/Protocol Notes
AWS Data Collector IP or hostname
US1
collector.ctpx.secureworks.com
18.217.45.178/32
3.16.4.173/32
18.224.219.97/32
13.59.146.90/32
3.16.16.254/32
18.223.74.238/32
US2
collector.delta.taegis.secureworks.com
52.14.113.127/32
3.141.73.137/32
3.136.78.106/32
US3
collector.foxtrot.taegis.secureworks.com
44.229.101.49
35.166.77.47
34.214.135.78
EU
collector.echo.taegis.secureworks.com
18.158.143.139/32
35.159.14.37/32
52.59.37.234/32
TCP/443 Safelisting device access to Taegis XDR via hostname
AWS Data Collector IP or hostname NTP severs IP/Hostnames provided during provisioning UDP/123 Safelisting device access to NTP servers

This rule is only necessary when custom NTP servers are provided during provisioning.
AWS Data Collector IP or hostname 169.254.169.123 UDP/123 Safelisting device access to default NTP server.

This rule is only necessary when custom NTP servers are not provided during provisioning.
AWS Data Collector IP or hostname DNS server IPs provided during provisioning UDP/53
TCP/53
Safelisting device access to DNS servers

Proxy Support

Cloud-based and On-Premises Data Collectors attempt to discover local proxy settings on the host if they are unable to connect directly to the internet.

Cloud-based and On-Premises Data Collectors also support a hard-coded proxy. If you need to create a data collector that contains a hard-coded proxy, please submit a support request with the following required information:

If the proxy is configured but is unavailable or not reachable, the data collector will fall back to a direct connection.

Note

Cloud-based and On-Premises Data Collectors do not support hard-coded authenticated proxies at this time. A proxy with man in the middle (MITM) capability needs to safelist the above network connections.

Spool Log Cache

A 200GB spool log holds data when the forwarding connection to XDR is slowed or temporarily unavailable.

Virtual Machine Requirements

Important

Data Collectors only support x86-64-v2 micro-architecture level. The x86-64-v2 micro-architecture defines specific features on top of the x86-64 baseline architecture that is found on most modern CPUs that have a feature level matching the 2008 Intel Nehalem architecture. These features provide compile-time optimizations for increased speed and performance. Virtual Machine configurations that do not meet this requirement will fail to boot.

Virtual Machine Requirements:

Configuration

XDR steps you through collector configuration and provides an .iso with credentials and configuration for the collector and an .ova virtual image that you then deploy in your environment.

Set Up On-Premises Data Collector with vSphere

The following steps you through installation of the XDR On-Premises Data Collector. It is assumed you have the preconfigured .iso and .ova files on hand. If you still need those, navigate to Integrations > Data Collectors in XDR or contact your XDR representative.

For HA Collectors, you need three VMs. Complete the following Steps 1-14 three times. DHCP collectors have just one .iso file that are used for each VM. Each collector with a static IP address has a different .iso file.

Install the On-Premises Data Collector on VMware vSphere

  1. Make sure you have the .iso and .ova files you need to run the collector appliance. You can preconfigure and download those in XDR from Integrations > Data Collectors.

  2. Connect to your vSphere Web Console.

  3. Navigate to Virtual Machines and select Create / Register VM.

Create or Register a VM

Create or Register a VM

  1. From Select Creation Type, select Deploy a virtual machine from an OVF or OVA file.

Select Creation Type

Select Creation Type

  1. Give your new VM a name such as taegis-xdr-collector and then select the .ova file you downloaded from XDR (see Step 1 above).

Select XDR .ova File

Select XDR .ova File

  1. Select the appropriate datastore where you want to store the VM’s disk images.

Important

Choose a datastore with at least 220GB of free space for the VM. 20GB is needed for the primary and 200GB for the secondary drive.

 

  1. Choose your preferred network and disk configurations.
  2. In the Deployment Options section, make sure to uncheck Power On Automatically. Review the configuration and then select Finish.

Steps 1-8

Steps 1-8

  1. Navigate to Storage and choose Datastore Browser.

Select Datastore Browser

Select Datastore Browser

  1. Select the datastore where you want to store the XDR On-Premises Data Collector configuration .iso (see Step 1).
  2. Click Upload and select the configuration .iso.
  3. Navigate back to Virtual Machines, and right-click the VM to bring up the context menu. Choose Edit Settings from the context menu.

Edit Settings

Edit Settings

  1. Change the CDROM device from Client Device to Datastore ISO File.

Select Datastore .iso

Select Datastore .iso

  1. The Datastore Browser opens. From there select the .iso you uploaded in Step 11 and click Save to finalize the changes.

Note

DHCP HA Collectors have only one .iso file. HA Collectors with static IP addresses have three. In the case of static collectors, you must use each .iso once so that each VM gets assigned a different .iso file. In the case of DHCP collectors, use the same .iso file on every VM.

Once you have completed Step 14, go back to Step 1 to create the next node. Once you have stood up the three nodes, you are done creating your VMs. After the three VMs have been configured, proceed to Step 15.

  1. The VM is now ready to be powered on.

Once a VM comes online, it must stay online until all VMs have joined. This is required as the cluster does not allow new VMs to join the HA cluster if any other VM is down or unreachable. In other words, bring the first VM up, leave it running, then bring the second VM up. Then, leaving both these VMs up, bring up the third VM. Don’t turn off any of the VMs until all three have joined.

Note

When deploying an onsite collector the .iso must be mounted at first boot to configure the XDR Collector. After the XDR Collector shows "READY" in XDR, the .iso can be dismounted.

Set Up On-Premises Data Collector with Hyper-V

The following steps you through installation of the XDR On-Premises Data Collector using a Hyper-V environment. It is assumed you have the preconfigured .iso and .ova files on hand. If you still need those, navigate to Integrations > Data Collectors in XDR or contact your XDR representative.

Install the On-Premises Data Collector on Hyper-V

  1. It is assumed you have the preconfigured files on hand. If you still need those, navigate to Integrations > Data Collectors in XDR or contact your XDR representative. Unzip the ctpx_collector.zip file downloaded from XDR when you created your On-Premises Data Collector.

Tip

Place ctpx_collector.vhdx and ctpx-collector-disk2.vhdx in the same folder with the .iso file(s) as it needs to be mounted to the CD drive on the Hyper-V machine.

Unzip your ctpx-collector download

Unzip your ctpx-collector.zip download

  1. Open the Hyper-V Manager and select New, then select Virtual Machine... from the context popup in the upper right pane.

Create a New Virtual Machine

Create a New Virtual Machine

  1. The New Virtual Machine Wizard displays. Select Next.

Virtual Machine Wizard

Virtual Machine Wizard

The Specify Name and Location dialog displays.

  1. Specify a name for the Virtual Machine then select Next.

Virtual Machine Name

Virtual Machine Name

The Specify Generation dialog displays.

  1. Choose Generation 1 then select Next.

Specify Generation

Specify Generation

The Assign Memory dialog displays.

  1. Assign memory (see Virtual Machine Requirements above for recommended settings) and select Next.

Assign Memory

Assign Memory

The Configure Networking dialog displays.

  1. Choose the network adapter you are going to use with the virtual machine, then select Next.

Configure Networking

Configure Networking

The Connect Virtual Hard Disk dialog displays.

  1. Select Use an existing virtual hard disk and then browse to the location you chose in Step 1. Choose the ctpx-collector.vhdx file, then select Next.

Connect Virtual Hard Disk

Connect Virtual Hard Disk

The Completing the New Virtual Machine Wizard dialog displays.

  1. Review the summary information and click Finish.

Completing the New Virtual Machine Wizard

Completing the New Virtual Machine Wizard

  1. From the Hyper-V Manager, select the virtual machine in the Virtual Machines pane (upper center), then select Settings in the lower right pane.

Hyper-V Manager

Hyper-V Manager

  1. Select Processor from the Settings menu in the upper left-hand pane (see Virtual Machine Requirements above for recommended settings). Do NOT select OK or Apply at this step.

Hyper-V Processor

Hyper-V Processor

  1. Next, open IDE Controller 0 right below Processor. The IDE Controller details display in the right-hand pane. Choose Hard Drive then select Add.

IDE Controller 0

IDE Controller 0

  1. Another Hard Drive appears under IDE Controller 0 in the left pane; it should be automatically selected. Select Browse..., navigate to the folder from Step 1, and choose ctpx-collector-disk2.vhdx. Do NOT select OK or Apply at this step.

Second Hard Drive

Second Hard Drive

  1. Click on DVD Drive under IDE Controller 1 in the upper left pane. Choose Image file:, then Browse, navigate to the folder from Step 1, and choose ctpx-collector.iso. Do NOT select OK or Apply at this step.

ctpx-collector.iso

ctpx-collector.iso

  1. Review the Processor, IDE Controller 0, and IDE Controller 1 entries on the left pane. Verify the following selections:

Processor

IDE Controller 0

IDE Controller 1:

Verify Settings

Verify Settings

Once verified, select OK. The Settings menu closes.

Note

DHCP HA Collectors have only one .iso file. HA Collectors with static IP addresses have three. In the case of static collectors, you must use each .iso once so that each VM gets assigned a different .iso file. In the case of DHCP collectors, use the same .iso file on every VM.

Once you have completed Step 15, go back to Step 1 to create the next node. Once you have stood up the three nodes, you are done creating your VMs. After the three VMs have been configured, move on to Step 16.

  1. The VMs are ready to be powered on.

Select the virtual machine on under Virtual Machines in the Hyper-V Manager, right-click or otherwise bring up the context menu, and select Start.

Once a VM comes online, it must stay online until all VMs have joined. This is required as the cluster does not allow new VMs to join the HA cluster if any other VM is down or unreachable. In other words, bring the first VM up, leave it running, then bring the second VM up. Then, leaving both these VMs up, bring up the third VM. Don’t turn off any of the VMs until all three have joined.

Start

Start

  1. The XDR On-Premises Data Collector is now running.

Maintenance or Downtime

It is a requirement that any maintenance that results in downtime be performed one node at a time. After a node has been taken down for maintenance, it must be brought back online and it must have successfully rejoined the cluster before bringing the next one down. You can execute the show cluster status command in the device console to verify it has rejoined.

IP Address Changes

Currently HA Collectors with statically configured IP addresses can't change their IP Address. If an IP address needs to be changed, a new HA Collector must be provisioned.

For HA Collectors with IP addresses allocated via DHCP reservation, only one node can change its IP address at a time. Also, all other nodes in the cluster must be up and running for the node to be added back to the cluster with its new IP address.

Access Troubleshooting Console

The Admiral console allows you to access information about a deployed XDR Collector locally. The tools provided within Admiral assist in device setup and troubleshooting of common problems such as network connectivity. For more information, see Admiral Console.

 

On this page: