Skip to main content

Frequently Asked Questions

Solutions to some common questions can be found here and in our (Knowledge Base)[]. If you can't find your answer, please reach out to and let us know.

Which cloud providers does Cado support?

Cado supports native deployments in AWS and Azure. If deployed in AWS, Cado supports AWS as well as cross-cloud acquisitions from GCP and Azure.

Does Cado support memory capture?

Yes. See Memory for more details.

Why do I see "Installing Cado Response" instead of the login page?

On first boot, the platform will take about 10 minutes to install. During that time, you will see "Installing Cado Response". After the one-time installation process completes, you will be presented with a login screen. During an upgrade, you will also see "Installing Cado Response" until the upgrade is complete.

I deployed the Cado platform. How do I log in?

Navigate to the Cado instance IP https://<Cado_IP>. You can initially login with the username admin and the password i-xxx where i-xxx is the ID of your instance. After first login, you will be asked to change your password. Note that for Azure, the password is the very long Resource ID. The Resource ID can be found within your Azure Portal by navigating to Home > Virtual Machines, clicking on the Cado VM name, then clicking Properties on the left navigation bar. Scroll down to find the Resource ID.

Can I allow or restrict IP ranges from accessing my Cado instance?

Yes, and it is strongly recommended to follow the principle of least privilege and restrict access to only those IPs which require access. In AWS, navigate to your AWS EC2 Console, click on your Cado Instance ID, navigate to the Security tab, then click on the "Security Group" link, then "Edit Inbound Rules". Here you can add and remove access rules for the Cado instance. Note that you can also restrict access to HTTPS and SSH during the deployment process via the CloudFormation Template. In Azure, navigate to your Azure Portal and find the Network Security Group where your Cado VM is deployed. From there, click "Inbound security rules" and you can Add and Delete access rules. If you have deployed with an ALB then you'll need to update the inbound rules for the ALB secuirty group intead. Please note that the ALB will only allow HTTP/HTTPS access - SSH rules will need to be configured in the EC2 security group.

How can I control the lifecycle of data stored by the Cado platform?

AWS supports this through Object Locks and Lifecycle Rules associated with your S3 bucket. Please see the AWS documentation for more details.

Azure supports this through Life Cycle Management Policies for your blob storage. Please see the Azure documentation for more details.

Does the Cado platform store sensitive customer data?

Whether sensitive customer data is stored within the Cado platform is completely dependent on what data sources and evidence is imported into the software. In all cases, data is either stored within the database on the main Cado instance, within EFS, within a customers' S3 bucket or Azure storage. Since the Cado platform is deployed within the customers' cloud environment, the customer has control over the life cycle management of the data (example: using Life Cycle management policies) as well as who has access to potential sensitive data via VPC and Subnet rules.

Can I monitor the health and proper function of the Cado platform?

Yes. Within Cado, you can see the platform health and statistics by navigating to the "Platform" tab.

This lists Free Disk Space, Available Memory, Total Memory, CPU Usage and pipeline processing details. In addition, roundtrip response latency for the API /api/v2/system/status endpoint should be less than or equal to 200ms.

Can I estimate the costs for running the Cado platform in AWS?

Yes, please see Cost Management for details.

How can I capture Logs or a Support Bundle from the Cado platform?

The following methods for getting Logs to Cado are available, in order of ease:

  1. Try sending logs to Cado automatically through the UI by navigating to the Help screen. Click on Send Logs to Cado and this will automatically package up a bundle of logs and send them directly to Cado over HTTPS
  2. If that fails, click on Download Logs and upload the zip file to the Customer Portal
  3. If 1 and 2 fail, or you cannot log into the platform then open an SSH session to the Cado Platform (you'll need the key you used during deployment) and execute the command sudo tar -cvzf /var/log/cado_logs.tar.gz /var/log then upload the zip file to the Customer Portal.

How do I contact Support?

You can contact Cado Support via email at or reach out to your account manager for more details.

What are the Cado Service Level Agreements?

The Cado Service Level Agreements (SLAs) are part of customers' end-user license agreement. Please refer to your EULA for more details.

Does the Cado platform support processing of EC2s or AMIs with EBS Encryption?

Yes. As of build, the kms:CreateGrant and kms:Decrypt permissions are enabled within the Cado role by default. If you upgraded from a build earlier than, you may need to add these permissions manually.

If vulnerabilities are discovered within your software, are customers notified?

Yes. Updates and notifications are provided to customers if vulnerablities or vulnerable packages are identified within our software.

How do I expand my available data storage within the Cado platform in AWS?

Below are the steps for expanding the Cado platform data storage volume after deployment. This follows the steps as outline by AWS' documentation: Extend a Linux file system after resizing a volume


Before continuing, create a snapshot of the CadoResponseDataVolume volume which has the device name /dev/sdh. This will ensure you have a backup of the data volume to restore, in case you encounter an issue.

  1. Extend the /dev/sdh storage volume via the AWS Console by:
    • Navigating to your Cado instance
    • Click the Storage tab
    • Click on the volume instance ID for Device Name /dev/sdh
    • Select the CadoResponseDataVolume
    • Click Actions
    • Select Modify Volume
    • Enter the new volume size
    • Click Modify
  2. SSH into the Cado instance. example: ssh -i "us-east-2.pem"
  3. Run df -hT. Note the Avail space for the /dev/nvme1n1 filesystem mounted on /home/admin/data. This should display the old volume size.
  4. Run lsblk and note the nvme1n1 volume SIZE mounted on /home/admin/data. This should display the new volume size.
  5. To extend the volume and make it available to the OS, run sudo xfs_growfs -d /home/admin/data
  6. Run df -hT again and note the Avail space for the /dev/nvme1n1 filesystem mounted on /home/admin/data This should now display the new volume size.