Skip to main content
Skip table of contents

Event Playbooks

Overview

The Event Playbook uses D3 SOAR's codeless playbook editor to allow you to create standardized workflows for ingested events from different integrations. You can create multiple event playbooks and select which playbook to run for each scheduled event ingestion. An event playbook can be triggered to run in two different stages of an event's life cycle in D3 SOAR: On Event Ingestion, and After Event Dismissal.

You can create, view, and edit all event playbooks in Configurations > Event Playbooks. Event playbooks can help SOC Managers automate standard procedures applied to ingested events and reduce time spent by an analyst executing these procedures manually.

This document will provide all the information needed to help you build and use an event playbook in D3 SOAR.

UI Elements

This section outlines all of the notable UI features of the event playbook. Each element will be explained in more detail in the following Features section.

Action Bar

Breadcrumbs

The relative path of your location within Configurations

Shared tag (Multi-tenancy only)

Represents the sharing status of this playbook from the master vSOC

Concurrent Users

All the current users viewing/editing this playbook are displayed here in their initials.

Draft/Live status

Represents the playbook’s draft/live status

Actions

The list of actions you can perform on an event playbook:

  1. View Live / View Draft

  2. Submit

  3. More Actions

    1. Clone

    2. Replace

    3. Delete

    4. Version History

    5. Command References

Task Panel

Add New Tasks

Drag and drop one of the 9 tasks types into the editor to build your playbook workflow

Quick-add Tasks

Drag and drop specific command tasks into the editor to build your playbook workflow

Main Editor

Viewing Options

  1. Overview

  2. Zoom In/Out

  3. Fit to screen

  4. Print

  5. Refresh

Local Shared Data

A data storage area where you can save reusable information for different tasks to utilize

Trigger Output Data

Stored event data that is returned after the event playbook has run.

A wildcard search through all the playbook tasks currently in the editor

Test

Test the playbook with sample event data to verify your workflow

Task Detail Panel

View the detailed configurations and inputs of each task

Features

Triggers

In an event playbook, each task chain must be connected to a trigger. Each task chain is only run when the trigger condition is satisfied. Currently, event playbooks have two triggers:

  1. On Event Ingestion – Tasks connected to this trigger run when raw event data is ingested and field mapping, artifact extraction, and system reputation check is completed

  2. After Event Dismissal – Tasks connected to this trigger run immediately after an event is dismissed

You should use these triggers to help you classify which workflows to run depending on the stage of the event life cycle an event is currently in.

Draft/Live Modes

There are two playbook modes in the event playbook:

Draft

Playbooks that are in Draft mode are editable. This mode is meant for you to test and make changes to your playbooks without impacting the playbooks you are currently using in a live environment.

Live

Playbooks in Live mode are the playbooks you can use in your live environment. These playbooks are read-only as they are being actively used.

By default, all newly created playbooks are in Draft mode. Once you are satisfied with your playbook configuration, you can switch the Playbook mode from draft to live by pressing the Submit button in the top right.

Collaboration and Concurrent Users

When collaborating with other users on a playbook, it’s essential to know if you are on the same playbook and what changes each of you is making. Seeing concurrent users is also important when working independently on a playbook so that you are aware of where other users are, and no crucial data is unknowingly overridden.

Concurrent Users at Playbook Level

Every user currently working on the same playbook will appear on the toolbar as coloured circles with their initials. When you hover over a specific user, you can see their name. The toolbar can display up to five users’ initials. However, if the total number of concurrent users exceeds five, the 5th circle will show you how many more people are here on this playbook. When you hover over that number, you will see a list of their names.

Concurrent Users at Task Level

When a user is on a specific playbook task, whether they’re viewing or editing it, you will see an outline in their colour around the task they're on. When you hover over the task, it tells you exactly who the user is.

In the case of multiple users on one task, you can hover over the task to see everyone’s name.

Concurrent Users at Field Level

To view a specific field that a user is editing, click on the task to open up the task configuration panel. The field being edited will be outlined in the editor's colour. When you hover over the field, you will see the editor's full name at the bottom right corner. However, concurrent users at field level can only be seen when the playbook is in draft mode.

In the case of multiple users on the same field, you can hover over the field to see everyone’s name. Fields are automatically saved when you click away from the input box.

Alert

Please be aware that when multiple users are editing the same field, the last saved edit will override previous saved edits.

Concurrent Editing and Saving

View Draft / View Live

As mentioned in the above section, a playbook can have two modes, a live mode, and a draft mode. When another user clicks on the Edit button while you are viewing a playbook, a draft mode is created for the other user to edit. You will remain on the live mode so that your viewing is not interrupted. If you want to edit or view the changes made by the other user, you can do so by clicking on the View Draft button. You can exit the draft mode at any time by clicking on the View Live button.

Submit Changes with Concurrent Editors

When you’ve finished editing and want to make the changes go live, the system will notify you if another user is still editing a field. You can choose to cancel your submission or submit it anyway. However, doing the latter may cause data loss or overridden changes on the fields they were editing.

In the event that other users submits their edits to go live while you are still editing a field, you will see the following notification:

You can then choose to see the newest live view or open a draft view of the most recent live view to continue your edits.

Version History

A version of the playbook is saved every time it goes live. Each version contains the DateTime the playbook went live and the user who performed this action. You can access all previous versions of this playbook by clicking on the More button located in the top right and clicking on Version History.

To restore to a previous version of a playbook

  1. Make sure this playbook has gone Live at least once

  2. Click on More button > Version History

  3. Click on Export this version for the version you want to restore to

  4. Navigate back to the playbook and make sure the playbook is in Draft mode. You can do so by clicking on the Edit button if the playbook is still in live mode.

  5. Click on More button > Replace Playbook

  6. Browse and select the playbook XML file that was exported in step 3

  7. Click Replace

Result

The playbook has been replaced with the selected version.

Local Shared Data

The local shared data allows you to store data in JSON format for tasks to reuse in a playbook. This storage area is local – data stored here is only accessible by this playbook and this playbook only. This feature may be helpful in situations where there are data (e.g. URL reputations) you want to be accessible by any tasks regardless of their playbook path.

There are four utility commands in D3 SOAR to help you manipulate data in the local shared data:

Command

Description

Input

Output

Add Root Key for Local Shared Data

Adds new JSON object data to the local shared data

JSON object data to be saved and the JSON path to define where it is saved

Stored the data to the defined JSON path in local shared data

Remove Field from Local Shared Data

Deletes specified fields from the local shared data

Removes field with specified JSON Path from Local Shared Data in playbook

A log of what data was deleted

Extract URLs from Local Shared Data

Extracts all URLs within the local shared data

None

All URLs in JSON array format in the local shared data

Get Local Shared Data with JSON Path

Retrieve JSON data from local shared data by the specified JSON path

JSON path to retrieve data from

The JSON object saved in the specified path

To add and retrieve data from local shared data

In this example, we want to save URL reputation data from a task to the local shared data and add this data to a global list in the system.

  1. Drag the command task Add Root Key for Local Shared Data and drop it right after the Check URL Reputation task

  2. In the command input area of Add Root Key for Local Shared Data:

    1. Input

      1. Select the path of the output of the Check URL Reputation task

    2. JSON Path

      1. Enter $.URLReputationData

  3. In Add to Global List command task, make sure Dynamic input is enabled and click on

  4. Select Local Shared Data from the data source dropdown

  5. Select the path $.URLReputationData

  6. Test the playbook

Result

The Add to Global List command successfully retrieves URL reputation data from the Local Shared Data.

Trigger Output Data

After running the Event Playbook, the event data structure will change as a result of data processing. The returned data is stored as the latest event playbook output data in JSON format. This is known as the trigger output data.

Reader Note

Only Event Playbook has the trigger output data feature. Data is only stored after running the test with event raw data.

Store trigger output data

You can test run the event playbook by choosing a trigger and storing the subsequent trigger output data. Below is a detailed step-by-step on how to populate the trigger output data.

  1. Navigate to Configuration

  2. Click on Event Playbook tab

  3. Open an Event Playbook

  4. Click on the Run button

  5. Build Test Data

  6. Select an Integration & Connection

Click on Next and Build

After running the playbook, data will display inside the trigger output data.

Modify and get trigger output data

You can edit the trigger output data to optimize your work in two ways

  • Utility Command

  • Passdown Task

Utility Command

There are two built-in utility commands to help you get or update trigger output data.

Command

Description

Input

Output

Get Trigger Output Data with JSON Path

Get Trigger Output Data with input JSON Path

JSON path to retrieve data from

The JSON object saved in the specified path

Update Root Key Value for Trigger Output Data with JSON Path

Adds a specified root key for the input Trigger Data

JSON object data to be saved and the JSON path to define where it is saved

Store data to the defined JSON path in trigger output data

Update Root Key Value for Trigger Output Data with JSON Path

In this example, we want to update the Data field within the trigger output data.

  1. Drag a command task Update Root Key Value for Trigger Output Data with JSON Path and drop it right after the On Event Ingestion task.

  2. In the command task, select the source type, and input the JSON path and updated value.

    1. Source Type - Select the source type to update

    2. JSON Path - Input the path you want to update the value. In this case, our path will map to the $.Data value within the trigger output data.

Reader Note

If the JSON path is the same, the value will be replaced. Otherwise the new path and value will be added to the end of trigger output data.

c. Updated Value - Input the value to update and save in the trigger output data

  1. Save the task and run the test.

Result

The data inside the trigger output data will be updated to “this is a sample.”

Passdown Task

The passdown task allows the user to pass playbook returned data or custom data and write it other destinations. In the case of trigger output data, new data can be written and updated into the existing trigger output data. This can be done by selecting the destination field as the trigger output data.

  1. Drag the Passdown task and drop it right after the On Event Ingestion task.

  2. In the command task, select the destination field, and input the destination JSON path and data.

    1. Destination Field - Select the destination fields you want to make a change to. In this case, we will update the trigger output data.

    2. Destination JSON Path - Input the specific path you want to update. We will write the new data into the $.Data path within the trigger output field.

Reader Note

If the JSON path exists inside of the trigger output data, the value will be replaced with the data field. Otherwise the new path and value will be added to the end of trigger output data.

c. Data - Input the desired data and save in the trigger output data

  1. Run the Task

Result

The “data” value inside the trigger output data will be updated to “test”.

View trigger output data

Format Builder

Select the trigger output data as a data source on the top left of the window and click on a specific path to view its details.

Playbook Test

The event playbook allows you to test your workflow while you are editing it in Draft mode. You can use this feature to test the validity of your playbook while you are building it. The playbook editor offers testing at three different levels:

  1. Playbook Level

  2. Task Level

  3. Input Level

Playbook Level

You can test your entire event playbook workflow to see if everything is working as intended by clicking on the Test ▶ button in the action bar. For an event playbook, you will have to first build a sample set of event data before testing. Since the event playbooks rely on triggers to trigger the execution, you will also need to select the trigger type while building the sample data. Once the sample data has been built, the playbook will run from the source node on the left all the way to the rightmost node.

After your first test run, you might see various icon(s) appear in each Task card. These icons represent the current status of the Playbook task, and you can click on the icon(s) to view the Playbook Task Details in the pop-up window.

Task Status

Description

Run

This status is only available when you are testing a playbook. This allows you to test individual tasks to check if they’ve been configured correctly without the data from prerequisite tasks.

Queued

This status means the task is in a queue and waiting for its turn to run.

Done

The green checkmark means the task was successfully run.

Required

Suggested

Required tasks are represented by an asterisk and suggested tasks have a light bulb. Both of these statuses are referred to as a Pending Task. Required Tasks require your manual response, whereas this action is optional for Suggested Tasks.

Running

When you see this icon, it means the Task flow is currently being executed.

Skipped

The Suggested Task has been skipped, instead of being executed.

Error

Indicates that an error has occurred while the Task was running. You can rerun these Tasks after correcting the error.

Rerun

The icon appears when the task has been rerun.

Wait For Reply

This status appears when the Task is a nested Playbook and there are additional Tasks to be completed within this Task before it is updated as Done.

You can click on the icons to view the task details and status. Within the pop-up window, the top panel

You can click on the icons to view the task details and status. Within the pop-up window, the top panel highlights the Task Type, Assignee, Due Date, Status, Run Mode and action buttons of the playbook task. If not run, the first tab within this window presents the Input tab. You can review the contextual data in the Input tab, and update the input fields before running or skipping the Playbook Action. To run the task, click on the “Run/Done” button.

When you choose to skip a task, you can skip to the next task in the flow without having to run the previous task.

To test an event playbook
  1. Click on Test

  2. Select your trigger type and click on Build Test Data

Example: On Event Ingestion

  1. Select your integration to retrieve sample event data from and select a connection

  2. Click on Next

  3. Fill in your fetch event search parameters and click on Next

  4. Review the raw event data fetched. If you are satisfied with the event data, click on Build

  5. Click on Run Test

Result

The event playbook will run the tasks connected to the On Event Ingestion trigger path.

Task Level

On the task level, you can test a single task with specified input parameters and view the resulting output. Before testing at the task level, you will have to test the entire playbook at least once first to ensure there is data to test with. This can be done by clicking on the Test button in the action bar. Testing a task can help you ensure that it is well configured before moving on to creating the next step of your workflow.

To test a task
  1. Click on the ✅ icon to check Playbook Task Details.

  1. Click on the Rerun Button.

Input Level

On the Input level, you can test your inputs to make sure it satisfies the command parameters. Before testing at the input level, you will have to test the Playbook at least once first to ensure there is data to test with. Testing an input is especially useful when you are configuring a dynamic input - testing here can ensure the correct JSON path has been picked, or the correct data transformation has been applied.

To test an Input in a Command Task
  1. Select any Command Task.

  2. Click on the Test button for the specific input you would like to test. Make sure the dynamic toggle is on, otherwise the Test button will not be visible.

To test an Input in a Data Formatter Task
  1. Select any Data Formatter Task.

  2. Click the Test icon on the right task panel.

Tasks

Playbook Tasks are the building blocks of a playbook. Tasks take in input data, perform data processing, and return output data. These tasks are connected to each other using control flow, and data flows from left to right starting from the source node. There are 9 Playbook Tasks with which you can use to build or edit a Playbook:

  1. Command Task

  2. Interaction Task

  3. Conditional Task

  4. Data Formatter Task

  5. Stage Task

  6. SLA Task

  7. REST API Task

  8. Passdown Task

  9. Merge Task

To use a playbook task, simply drag and drop the task you want to use from the playbook side panel into the Main Editor. A new node will be created and connected to the preceding node via a line.

Task Data Flow

Each Task can take in input data, perform data processing, and return output data. You can link Tasks together in the Main Editor with arrows that represent data Control Flow. The following section will explain the data flow of Tasks.

Input

A Task can take in both static and dynamic inputs.

A static input means that the input is fixed and will not change. For example, in a Fetch Event Command Task, you can set a fixed start/end date for the Events you are fetching for. However, if your workflow requires a moving time window for the Fetch Event command to fetch from, you will need to use a dynamic input.

There are 5 types of dynamic input methods for a Task. They are:

  1. Dynamic Input

  2. Data Formatter

  3. HTML Editor / Data Formatter

  4. DateTime / Data Formatter

  5. Dropdown

Dynamic Input

Dynamic Input is the most common input type in a Playbook - it allows for both static and dynamic data to be used as an input for a Task. This feature can help you transfer data from one Task to another in a Playbook and chain Tasks together.

Static Data: Enter in the values that match the data type into the text area

Dynamic Data: Enable the dynamic toggle and enter the corresponding JSON Path of the dynamic data. This JSON Path should point to a data source that is dynamically updated in the Playbook. Optionally, you can apply regular expressions on this data through the Regex Filter tab to filter the JSON data.

Data Formatter

Data Formatter input allows you to use formatted dynamic data as an input for a Task. It is similar to Dynamic Input - but in addition, the Data Formatter allows you to perform complex data transformation on the data with our library of Jinja filters. With this, you can customize and create very specific dynamic inputs for a Task. The Quick Start guide in the Data Formatter contains more information on how to use this feature.

HTML Editor / Data Formatter

As the name suggests, this input method is the combination of HTML and Data Formatter. This input type is the same as Data Formatter, with the added ability to switch between the Data Formatter editor and a HTML editor. This allows you to build customized HTML templates that contain formatted dynamic fields that update automatically.

Date / Time Input

Date / Time input allows you to select a specific Date and Time as a static input for a Task. If your workflow requires a dynamic Date/Time input, you have the option to use the Data Formatter to select and format a dynamic input as the Date/Time input for a Task.

Dropdown Input

Static Data: Dropdown input allows you to select a value from a predefined list as an input for a Task.

Dynamic Data: You can define a JSON path or formatted data that outputs a value corresponding to one of the dropdown options. The playbook will dynamically select the matching value (from the JSON path or formatted data) in the dropdown when run.

Control Flow

All playbook tasks have an order of execution that determines which playbook task runs first, how often they run, whether they loop, etc. You can set a task to run if a condition is satisfied, thus creating branching flows; tasks can also run in a loop. The order of execution of playbook tasks is called a Control Flow.

There are a few types of Control Flow used in D3: sequencing, conditional branching, looping, wait for all, wait for any, and nesting.

Control Flows are represented by a directional line within the playbook. A playbook task will be triggered to run after the directly preceding task finished running. All playbook tasks can only be triggered and run once, except for when you set tasks to run in a loop.

There are a few examples of different types of Control Flow that are often used in D3:

Control Flow Type

Example

Sequencing

In the Playbook View, Playbook Tasks are executed from left to right. Any Task that has completed is marked as Done.

Initial Analysis has a forward arrow that links to Start Analysis. From Start Analysis, two arrows branch out to 2 different Tasks.

Once the Initial Analysis has been executed, the following Tasks are triggered to run.

Wait for All

Tasks from different branches can be connected, and converge to one single Task.

When you set that converged Task as Wait For All, it will only run when all the preceding Tasks connected to it have finished running.

When the Stage Task (Summary) is set as Wait for All, all three preceding tasks will need to be completed for the Summary Stage Task to run.

Wait for Any

When you set that converged Task as Wait For Any, it can run when any one of the preceding Tasks connected to it has finished running.

You can set a Task to either option:

Wait for Any, Run Once

Wait for Any, Run Always

The “URL analysis completed” Task is set to Wait for Any. When either of its two preceding tasks is completed, the “URL analysis completed” Task will be triggered.

Note: a Task will only be triggered once. As such, any preceding Tasks that are completed will not trigger the following Playbook Task again.

Conditional Branching

This flow is created when you set up options or conditions in the Conditional Task. Satisfying each option will trigger a flow accordingly, allowing you to create multiple branches from one original Task.

The Conditional Task “Has Source IP addresses” shows a branching workflow based on a simple True or False logic.

When an option is selected, the following task(s) will be triggered. For example, if option True is satisfied, the task Init Source IP counter will be triggered.

Condition-Controlled Loops

An option that is branched out from a Conditional Task can be linked back to a preceding task, creating a loop that runs until a condition is met. A limit can be set for the number of loops that can be run in a Playbook to avoid infinite loops.

In the example below, the Delay Execution for 1 minute and Check Status Tasks will be repeated until the option in Check if Search is completed is True.

Nesting

A command that uses the Codeless Playbook Implementation method can be reused as a nested Command Task in another playbook.

By nesting commands, you can organize and divide large playbooks into smaller and more manageable commands.

Nested playbooks are differentiated from normal command tasks by the link icon that appears on the task node. Click this icon to access the underlying Playbook workflow.

The example below shows an Event Playbook using a nested Command with Playbook

Task Options

Every task has a few options for its run, to ensure the smooth configuration of the playbook. This can include: Active, Auto Run, Required, Stop on Error, and Long Run.

Active

Activate or deactivate the task

Auto Run

This Task will proceed automatically after the run of the previous task.

When Auto-run is checked, the Required, Assigned To and Due Time fields will be hidden.

Required

The task will be required to be manually executed by an assigned user or group of users.
Required Tasks will appear in the Pending Tasks tab of the Investigation Dashboard

Stop on Error

Stop the playbook run if this task returns an error.

Stop on error is a good option for debugging your playbook. If you plan to re-submit your playbook, you can replace and re-run the new playbook on existing event playbooks that have stopped.

This can be done in the Event Playbook Viewer.

Long Run

Improves the performance of parallel processes of playbook tasks. Recommended only for resource-intensive tasks.

Reader Note

The Long Run checkbox is only available to Python-implemented commands. Long Run-enabled playbook tasks are exempted from taking up quota in the resource pool – improving the performance of parallel computing in a playbook. If this setting is not enabled, the Python task will default to a runtime of 10 minutes, after which the process will be stopped. By default, there is no runtime limit if this setting is enabled. It is recommended to only enable this feature when necessary, i.e., for resource-intensive tasks.

Output

There are five types of task output. The output of each task in runtime is yielded in the formats: Output Data, Context Data, Result, Return, and Raw Data. If the playbook task encounters an error, there will also be an Error tab. After running a playbook, you can click on each task to view the task output details.

Key Fields

The key pieces of information that are extracted from the output data.

Context Data

Context data is the raw data that has been formatted into a structured JSON object format, which can then be stored in the Playbook for data sharing

Return Data

Return data shows the success/failure status of a Task for Integration Commands.

For transformation and condition Commands, the Return data is specified with corresponding output data types.

Raw Data

Raw data shows the unformatted and raw information of external events provided by third-party vendors.

Result

Result data is formatted in a reader-friendly HTML table.

Error Data

Error data of each task can be retrieved as an input of an error-handling Task.

Task Types

As mentioned above, there are 9 different types of tasks that you can use in playbooks. Each task has its own unique function and use case, which we will discuss below.

Command Task

Command task is one of the most important task types in D3 SOAR. The command task allows you to utilize D3’s library of built-in commands and 2000+ integration commands to perform tasks including data processing, data transformation, system actions and much more. It is essential to every playbook as it connects our system with third-party integrations.

Our system has four different types of commands:

  1. Third-party integration commands

  2. System Commands

  3. Utility Commands

  4. Cyber Commands

In addition to the built-in commands, you can build your own custom commands with Python or D3’s Codeless Playbook Editor.

All Command Tasks have the following basic elements:

  1. Task Info & Action Bar: Displays the Task’s name, type, and actions you can take (Save or Delete).

  2. Task Name

  3. Task Specific Panel: Displays task-specific controls. This area changes depending on what type of Task you selected. In this case, it shows the Command inputs because the Command Task is selected.

  4. Task Options: Optional selections that can be made to the command.

  5. Assigned To: Assign a User, Role or Group to execute the Task.

  6. Due Time: Define the time period for the Task to be executed. Tasks that haven’t been executed after this time period become overdue.

  7. Run Mode: Select from the dropdown list with 2 main types of Control Flows:

    1. Wait for Any, Run Always / Once

    2. Wait for All

Types of Commands

Command Type

Details

Examples

Integration Commands

Commands available from third-party providers

- Get URL Report - VirusTotal

- Get IP Risk Score - Maxmind

- Create Ticket - ServiceNow

- Block URLs - Palo Alto Firewall

- Query DataDog Events - DataDog

Basic Utility

Commands that enable simple processing on the following data types:

Text

Number

Boolean

DateTime

Text Array

Number Array

Boolean Array

DateTime Array

JSON Array

JSON Object

File

File Array

- Append Text
- Contains Number
- Contains Text
- Convert Number to Text
- Ends with
- Less or equal

Cyber Utility

Commands that perform cyber-specific data processing.

- Extract Artifacts From Excel File
- Check Email Authenticity
- Extract URLs from JSON Object

System Utility

Predefined Commands that perform actions on system objects.

- Close Incident
- Add to Global List
- Generate Summary Report
- Resume SLA
- Send Email

Nesting

A playbook functioning as a custom command with inputs and outputs which can be used in one or many other playbooks is called a nested playbook. It is very common that many playbooks share one or more the same small segments of the workflow. In order to reduce the repetitive configuration of the workflows, nested playbooks can be configured to accommodate the need for reusability of playbooks. One of these use cases is IP enrichment across multiple threat intelligence tools.

To create a nested Playbook

The following steps demonstrate the process of creating a sample nested playbook for IP enrichment:

Step 1: Create a New Utility Command

  • Navigate to Configurations > Utility Command

  • Add a new command by clicking on +

  • Fill in the following details:

    • Command Name
      ipEnrichment

    • Implementation
      Codeless Playbook

    • Command Category
      Cyber Utility

Step 2: Configure Command Setting, Inputs and Outputs

  • Ensure that Command Task under Features is selected to make it available for use as a Playbook command.

  • Navigate to Overview > Inputs tab, and add a new Input Parameter. Configure the fields as shown below:

    • Parameter Name
      Ip

    • Display Name
      IP

    • Parameter Type
      Text

    • Is Required?
      No

    • Parameter Index
      0

    • Input Type
      Text Template

    • Default Value
      (Leave empty)

    • Description
      (Leave empty)

    • Sample Data
      8.8.8.8

  • Navigate to Overview > Outputs tab, and select the Raw Data subtab

  • Change the Type to JSON Object

Step 3: Configure Playbook

  • Click on the New Task tab

  • In the Playbook editor, drag and drop 3 commands tasks connecting to first node, and name the commands:

    • IP Reports - VirusTotal

    • IP Reputation - Recorded Future

    • MISP - Search IP

  • Within each command’s input parameter, choose corresponding connection, make it “Active” and “Auto Run”, and enable “Dynamic” JSON Path to {{ $.PlaybookData.DataSource.commandInput.ip }}

  • Add a Data Formatter task to consolidate the result as shown below, and ensure the “Run Mode” is set to “Wait for All”. This will ensure that this Data Formatter task only executes once the three previous commands finish execution.

    CODE
    { "MaliciousCount_VT": {{PlaybookData | jsonpath ('$.["IP Reports - VirusTotal"].outputData.MaliciousCounts') | first}}, "RiskLevel_RF": "{{PlaybookData | jsonpath ('$.["IP Reputation - Recorded Future"].rawData.data.results[*].risklevel')}}", "ThreatLevel_MISP": "{{PlaybookData | jsonpath ('$.["MISP - Search IP"].rawData.response.Attribute[*].Event.threat_level')}}"    
    
    }

  • Add a passdown command to the end of the playbook to set the output parameter Raw Data in Step 1 with content that has been constructed in the previous step.

Step 4: Publish and use the Playbook

  • Click on Submit to commit and publish the Playbook command

  • In other Playbooks (event/incident), you can now use this new command within a command task. The link sign 🔗 on the command playbook indicates that this command is using a nested playbook.

Data Formatter Task

The Data Formatter Task lets you perform complex data transformations to playbook data. You can use D3’s built-in data filter to format and select specific playbook data and view the resulting output. The resulting reformatted output can be used as inputs for subsequent playbook tasks.

The example below demonstrates how you can add a data formatter to an event playbook to clean up URL data output from a preceding task. We will create a script in data formatter that removes duplicate URLs and ensures each URL is in a standardized format.

  1. Give the task a relevant name that indicates what it does. In this example, we will name the task “URL Data Cleansing”.

  2. Input a custom Jinja script to select and format data. You can use the format builder to help you build a script by selecting the data source and filters you would like to use. Refer to the Quick Start guide for in-depth documentation on the available filters. In the following example, the script that cleanses the raw data and parses it into a JSON array:

    CODE
    {% set urlString = PlaybookData | jsonpath('$.DataSource.incident.Artifacts[*].URLAddress[*].Url') %}
    {% set result%}
    [
    {% for urlString in PlaybookData | jsonpath('$.["Extract URLs"].contextData') %}
        {% if urlString | last | equals("/") %}
        "{{result | plus (urlString[0:-1]) }}",
        {% else %}
        "{{result | plus (urlString)}}",
        {% endif %}
    {% endfor %}
    ]
    {% endset %}
    {{result | unique}}

    You can click Test to run and view your script’s output. The Data Formatter Output pop-up displays a bug log to help you troubleshoot if necessary. For our example, the output data was:

    CODE
    [
      "http://amazon.com",
      "http://google.com",
      "http://abc.com",
      "http://xmr.pool.minergate.com"
    ]
  3. Configure the control flow for the command task. Use the tick boxes to configure if the command task is active, if it will auto run, and if it will stop when an error occurs. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the command task is active, and will always auto-run when its preceding tasks have been completed.

  4. Include a description for the data formatter task for future playbook editing reference if desired. We have not included one for this example.

  5. Click the gray check mark button to confirm and create the command task.

  6. Done!

Reader Note

  • In Jinja, data surrounded by double curly brackets {{ and }} indicate that it is dynamic data. When you use the format builder to select a data path for dynamic data, double curly brackets will surround the selected path by default. An example of input data for a Data Formatter task created with the format builder may look like this:
    {
    "Title":{{PlaybookData | jsonpath('$.["Sample Data"].returnData.Title')}},
    "Description":{{PlaybookData | jsonpath('$.["Sample Data"].returnData.Description')}}
    }
    Due to the nature of Jinja’s formatting, input data with multiple sets of double curly brackets will be processed separately. This has caused some users’ input data to be formatted incorrectly by the Data Formatter. In the screenshot below, you can see that the output data is in a text format rather than a JSON format.

    Instead, it is recommended to place just one set of double curly brackets around the entire input JSON data:

    {{
    {
    "Title":PlaybookData | jsonpath('$.["Sample Data"].returnData.Title'),
    "Description":PlaybookData | jsonpath('$.["Sample Data"].returnData.Description')
    }
    }}

    The corrected input JSON object will output a properly formatted JSON object as shown in the screenshot below.

    It is best practice to place one set of double curly brackets outside of the JSON data to ensure the input data is formatted as valid JSON data.

Success

Using a data formatter task in the event playbook editor, you have added a script that checks if a URL ends with /, and if so, removes / and checks for duplicate URLs. You have configured the data formatter task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

Stage Task

The Stage Task lets you create checkpoints into a playbook’s workflow to organize playbooks. Stage tasks split the playbooks into different stages for easier management of workflows. Analysts can also use stage tasks to assign investigators to the event.

In this example, we will create a stage task in an event playbook to signify the data enrichment stage of the playbook workflow.

  1. Give the stage task a relevant name that indicates the corresponding stage of your playbook’s workflow. In this example, we will name the stage task “Data Enrichment”.

  2. Configure the control flow for the stage task. Use the tick boxes to configure if the stage task is active and if it will autorun. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the command task is active, and will always auto-run when its preceding tasks have been completed.

  3. Include a description for the stage task for future playbook editing reference if desired. We have not included one for this example.

  4. Click the gray check mark button to confirm and create the command task.

  5. Done!

Success

Using a stage task in the event playbook editor, you have organized your playbook by indicating the data enrichment stage of the playbook workflow. You have configured the stage task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

SLA Task

The SLA Task lets you set precise service-level agreement (SLA) for all tasks in the event response lifecycle. This task facilitates the reporting and monitoring of your team’s compliance with SLA guidelines.

In this example, we use an SLA Task to measure how long it takes the playbook to complete the escalation stage. This functionality is commonly used by SOC teams to ensure events are dismissed/escalated and handled within a reasonable time frame per SLAs.

  1. Give the task a relevant name that indicates the SLA agreement. In this example, we will name the task “Event Dismissal SLA Start”.

  2. Select the SLA agreement for this task using the dropdown menu. Configured SLAs will be available to select. You can add custom SLAs in the settings module of D3 SOAR. For this example, we will select Event Dismissed in 15 Minutes. This means that the time limit to respond to an event is 15 minutes.

  3. Define the SLA tracking stage for the task. With this configuration, you can select whether you want to start, pause, resume, stop or reset the SLA tracking task. For this particular SLA task in the example, we are adding the SLA task to start tracking when SLA tracking has started. We would need to add a Stop SLA task in the playbook to define when the SLA tracking will stop.

  4. Select who should be notified when the SLA is breached using the dropdown menu. Investigation teams, groups, users, and roles will be available to select. For our example, we will notify Security of the task.

  5. Configure the control flow for the SLA task. Use the tick boxes to configure if the stage task is active and if it will autorun. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the SLA task is active, and will always auto-run when its preceding tasks have been completed.

  6. Include a description for the SLA task for future playbook editing reference if desired. A brief description of the SLA is included in the example shown.

  7. Click the gray check mark button to confirm and create the command task.

  8. Done!

Success

Using an SLA task in the event playbook editor, you have added a task that will start a timer and notify the user if the response time of an event is within the 15-minute limit per the SLA in the example. You have configured the SLA task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

Conditional Task

The Conditional Task allows you to set up different possible paths a playbook workflow may go through depending on how the output of a preceding task runs through a logical test.

In this example, the Conditional Task checks which option a user has selected in an automated email sent by the D3 SOAR system inquiring if he/she has clicked on any malicious links.

  1. Give the task a relevant name that describes the conditional task. In this example, we will name the task “Did the User click on the Link?”.

  2. Use the tick boxes to set whether the task is active and automated. Both boxes are checked in this example so that the task is active and will auto-run.

  3. Select the condition type for the task (i.e. input or command). For the input condition type, you can build a custom Python script with the built-in data formatter to define the workflow for the conditionals. You can also select the command condition type, which allows you to use D3’s library of built-in and integration commands to define the workflow for the conditionals.

  4. Configure the control flow for the conditional task. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the conditional task will always auto-run when its preceding tasks have been completed.

  5. Include a description for the conditional task for future playbook editing reference if desired. We have not included one for this example.

  6. Add conditionals for the task. Conditionals correspond to the output data of the preceding playbook task. You can also change the order of the conditional options you have added. A common set of conditions is true and false, which we have used in this example.

  7. Click the gray check mark button to confirm and create the command task.

  8. Done!

Reader Note

In this example, the command condition type was selected to use the Contains Text command to check if the input text (from the preceding task’s dynamic output data) contains the specified text. Subsequent playbook tasks can then be added to each corresponding conditional option.

Success

Using a conditional task in the event playbook editor, you have added a task that checks which option the user has selected in the automated email sent by the D3 SOAR system inquiring if he/she has clicked on any malicious links by using the Contains Text command. You have configured the conditional task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run. Subsequent tasks can then be added to each configured conditional option (i.e., true and false).

Interaction Task

The Interaction Task is typically used to add a point of manual human processing into your playbook. With the Interaction Task, you can incorporate questions, prompts or notes to guide your SOC team’s decision-making process throughout an event response. Unlike other playbook tasks, an interaction task cannot be automated, as its purpose is to facilitate manual response in event response.

In this example, we will create an interaction task to prompt the investigator to manually select from a list of potentially malicious IP addresses from the dynamic output data of a preceding playbook task to block.

  1. Give the task a relevant name that describes the interaction task. In this example, we will name the task “Which of the following IP addresses do you want to block?”.

  2. Add questions, prompts, information etc. that will guide the investigator’s manual processing of the task. You can use the JSON path picker to select output data and use the Regex filter to filter the JSON data. For this example, we added a question asking the investigator which IP address to manually block from the preceding playbook task’s output dynamic data of some potentially malicious IP.

  3. Configure whether the interaction task is active and a required task. Both tick boxes are checked in our example so that the interaction task is both active and required.

  4. Assign the group, user or role responsible for processing the interaction task. We have assigned the task to the investigator for this example.

  5. Set the time constraint for the assignee to process the task. This setting can be used to align with any SLA requirements of your organization or use case. Here, we have set the time limit to process the interaction task to one hour.

  6. Configure the control flow for the interaction task. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the interaction task will always auto-run when its preceding tasks have been completed.

  7. Include a description for the command task for future playbook editing reference if desired. We have not included one for this example.

  8. Click the gray check mark button to confirm and create the command task.

  9. Done!

Success

Using an interaction task in the event playbook editor, you have added a task that will prompt an investigator on your SOC team to manually process and block IP addresses from a list of malicious IPs from the dynamic output data of a preceding playbook task. The task configured will be active and required, and must be processed within an hour. You have configured the interaction task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

Merge Task

The Merge Task lets you take in data from all preceding connected playbook tasks and combine the data based on specified conditions. Tasks that are connected after the merge task can then reuse the combined data.

In this example, we will create a merge task to combine data after an Unwind JSON Array command that deconstructed an array field from some JSON data of the risk levels of a list of IPs. The subsequent task after the Unwind JSON Array command queried the IP addresses using an integration command task. Since there was more than one data point, multiple instances of the query command task were configured to run. The merge task will now combine the data output of the different instances for use in later playbook tasks.

  1. Give the task a relevant name. In this example, we will name the task “Merge”.

  2. Set the condition to control how the data is combined. There are four conditions you can configure to control how data is combined: by playbook instance, by task name, by task instance, or by a specified JSON path. For this example, we will combine the data by playbook instance.

  3. Use the tick boxes to configure if the task is active, if it is automated, and if it will stop on error. Active and Auto Run are selected for this example.

  4. Configure the control flow for the merge task. The dropdown menu lets you select the control flow for the task. Refer to Task Data Flow for more details. The configurations in this example ensure the interaction task will always auto-run when its preceding tasks have been completed.

  5. Include a description for the command task for future playbook editing reference if desired. We have not included one for this example.

  6. Click the gray check mark button to confirm and create the command task.

  7. Done!

Success

Using a merge task in the event playbook editor, you have added a task that will combine the dynamic data output of multiple instances of a preceding playbook task. The combined data can be used for subsequent playbook tasks. The task configured will be active and automated, and will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

REST API Task

The REST API Task lets you use REST APIs to send a request to a web service to retrieve or modify data in your playbooks. You can use the five HTTP methods for CRUD (create, retrieve, update and delete) operations.

In this example, we will create a REST API task to send a webhook event to D3 SOAR with the POST method in order to create a new event entry.

  1. Give the task a relevant name that describes the REST API task. In this example, we will name the task “Send Webhook Event”.

  2. Configure the input parameters to use the REST API. This includes the request URL, the HTTP method, the request header and the body data for the request. In this example, the following values for the input parameters were configured:

    1. URL:

      CODE
      {{PlaybookData | jsonpath('$.["Construct Webhook URL"].returnData')|trim}}

      The request URL is selected using the format builder to select the data from a previous playbook task

    2. Method:
      POST
      We will use the POST request method to perform a create operation for a new event.

    3. Header:

      CODE
      {"d3key":"d3test"}

      The key-value pair is used for the API authorization.

    4. Body:

      CODE
      {% set result %}
      {
      "eventtype":"Suspicious Email",
        "username": "user1",
        "sourceip": "192.168.87.104",
        "hostname": "D3cyber7ce",
        "email": "demo@d3soar.com",
        "domain": "www.google.com",
        "destinationip": "8.8.8.8",
        "sha256": "8b2e701e91101955c73865589a4c72999aeabc11043f712e05fdb1c17c4ab19a",
        "timestamp": "{{'today' | date('YYYY/MM/DD HH:mm:ss')}}"
      }
      
      {%endset%}{{result|trim}}

      The body contains the JSON object data we want to POST into D3 SOAR.

  3. Use the tick boxes to configure if the task is active, if it is automated, and if it will stop on error. Active and Auto Run are selected for this example.

  4. Use the tick boxes to set whether the task is active and automated. Both boxes are checked in this example so that the task is active and will auto-run.

  5. Include a description for the command task for future playbook editing reference if desired. We have not included one for this example.

  6. Click the gray check mark button to confirm and create the command task.

  7. Done!

Success

Using a REST API task in the event playbook editor, you have added a task that will use a REST API to perform a create operation of an event into D3 SOAR. You have configured the REST API task so that it will automatically run in the playbook as soon as the preceding playbook tasks have successfully run.

Passdown Task

You can configure the Passdown Task to save the results of the executed Playbook Task and pass down its values into the Command Task’s output or any designated placeholder fields. This Task is only available in Event Playbooks and Commands - Codeless Playbook Implementation Method.

Selecting an Event Playbook

After setting up your event playbooks, you can select a specific event playbook to run for each integration. You configure this in two areas: Data Ingestion and File Upload.

Data Ingestion

The following instructions teach you how to configure a data ingestion schedule and select a specific event playbook to run when a fetch command is executed.

  1. Click on the Data Ingestion tab.

  2. Hover over the desired Data Source or the + icon next to the Search bar.

  3. Click on the + icon.

  1. Select Schedule as the Intake Method.

Reader Note
The availability of intake methods depends on the integration and user configuration. Some integrations have Webhook as well.

  1. (Optional) Setup Event Field Mapping

  1. Select a Connection.

  2. Set the Interval. This determines how often the schedule executes.

  3. (Optional) Select an Event Playbook

  4. Check the box if you want to receive email notifications after a certain number of failed attempts.

  5. Fill in the inputs for the Fetch Event command in Command Details. (Command details may vary depending on the integration)

  6. Click on Save & Run

Result

The new Data Source will be displayed as a card on the left hand side and the schedule will start running.

JavaScript errors detected

Please note, these errors can depend on your browser setup.

If this problem persists, please contact our support.