Skip to content

Data Productivity Cloud new user tutorial

This tutorial is for new users who wish to get started with the Data Productivity Cloud and use Designer to create data pipelines that make your data business ready faster.

In this tutorial, you will learn how to:

  • Register as a new user to the Data Productivity Cloud.
  • Choose a data platform to load data to.
  • Navigate the Designer UI, and create an orchestration pipeline to extract and load your data into your chosen data warehouse.
  • Create a transformation pipeline to transform your data.
  • Use source control to save your work.
  • Publish your branch, and schedule your working pipeline to run every week.
  • View a clear representation of data flow in your transformation pipeline using our data lineage feature.

This example will use the Jira Query component to extract data from a Jira data source. We will then perform a basic data transformation and store the results in a Snowflake destination.

Note

  • If you wish to follow along using the Jira Query component, you'll need valid login credentials to access the Atlassian portal and the Jira application. A valid account is required to authorize the Jira Query component you will be using later in the example. For more information, read Jira Query authentication guide.
  • While this tutorial extracts data from Jira, the foundational principles still apply if you're connecting to other services such as Salesforce, Facebook, Shopify, and so on.

Jira is a performance and organizational tool, typically used in software development to coordinate work. In this case, our pipeline will be used to return all outstanding "epics" in our Jira board. An outstanding epic is one that is either in-progress or not yet started. Completed and closed epics will not be returned. As part of this process, we will transform the data by using a simple filter to exclude the data we don't want to see.

When the pipeline is built, we will publish it and set it to run via a schedule. All of our work will be maintained through Designer's built-in source control. You can also view a visualization of the data flow when the pipeline is running.


Register to the Data Productivity Cloud

Before you begin, you will need to register for a Hub account. The Hub is a single platform that offers a wide range of services for all registered customers, and this is the only way you can access the Data Productivity Cloud. For more information, read Getting started.

Hub sign up

Follow the steps in Registration. During the registration process, you are required to complete three sections:

  • Create your profile.
  • Set up your account.
  • Get started.

When you get to the section where you're setting up your account, for the purpose of this tutorial, choose the Matillion trial warehouse.

Choose data platform

There is no user setup required when you choose a Matillion trial warehouse—you will be provisioned with a Snowflake data warehouse, and this is where you will load your data, ready to transform it later.

The process for creating a trial warehouse may take a couple of minutes. While you wait, feel free to watch the video on the screen that introduces the capabilities of the Data Productivity Cloud. When your account has been created, you will be a registered Hub user, and you'll be given a free trial to use with complementary credits for your user account. For more information, read free trial.


The Designer canvas

When the setup process is completed, the Designer canvas will be displayed. A project will be waiting for you labelled My first project. For more information about the basic concepts behind what has already been set up for you—a cloud data warehouse and a project—read Designer overview and Projects, respectively.

Designer canvas

Note

To become familiar with the Designer UI, read Designer UI basics.

Version control

Before we begin, at the top of the canvas in its own tab you'll notice you are currently working on the main branch. When a project was first created for you during registration, an associated Matillion-hosted Git repository was also created, with a default branch called main.

Note

It's recommended that you create another branch other than main to develop your pipelines.

Designer uses Git for version control, making it simple to collaborate and manage data pipelines in your team. Before creating a new branch, it's important to understand how Git is incorporated in the Designer interface. For context, read Git in Designer.


Create a new branch

Let's begin by creating a new branch and making sure we Branch from main.

  1. From the Designer canvas, click My first project to return to the action menu of your project.
  2. In the Branches tab, click Add new branch.
  3. Give a unique name for your new branch.
  4. Choose main as the branch to branch from.
  5. Select an environment. Environments are for configuring your connection to your data warehouse. An environment was already set up for you during registration, but for more information, read Environments.
  6. Click Create.
  7. You will be returned to the Branches tab. Find and click on your newly created branch to return to the Designer canvas on your new branch.

Build an orchestration pipeline

After you have added a new branch, What would you like to do today? will be displayed. Click the Add Orchestration pipeline tile to add our first pipeline. Create an appropriate name for your new pipeline. You could call it "extract-jira-data", and then click Add.

Create orchestration pipeline

The Designer UI helps informs you which pipeline type you have just created by displaying a blue "O" for orchestration next to the name of your pipeline. A green "T" denotes a transformation pipeline.

You're in the Designer canvas and ready to build your first pipeline. For the purpose of this tutorial let's add the Jira Query component on to the canvas. For an introduction to using connectors, read Connectors in Designer.

Extract and load Jira data

On your Designer canvas, locate and click on the Start component and then click the black + symbol to open the Add component dialog. In the search bar, type "Jira". Click Jira Query. The component will be added to the pipeline canvas with a connecting line tying your Start component to your Jira Query component.

Click the Jira Query component and you'll see the Properties panel open on the right-hand side.

As the first step in configuring the Jira Query component, you're required to authenticate to it. Choose between User/Password and User/API Token. Read Jira Query authentication guide to learn how to obtain a Jira API Token. For more information about storing tokens and passwords in Designer, read Secret definitions.

Lets start to configure the component for our Snowflake data warehouse. You can obtain the component property information from the Jira Query documentation. For the purpose of this example, we want to extract epics information from the Jira database. For specific property information, refer to the following table:

Property Value
Basic Advanced Mode Basic
Data Source Epics
Data Selection BoardID, Done, Id, Key, Name, Summary
Target Table Create a suitable name for the table that will hold the extracted data.

Note

You can leave the remainder of the properties set to their default values.

You must validate the pipeline before you can sample the data. To do this, Click Validate in the top-right of the pipeline canvas. When the validation is successful you will be able to sample the extracted data from your Epics database. Click the Sample data tab at the bottom of the pipeline canvas, and click the Sample data button.

Sample Jira data

Note

The benefit of validating and sampling your data directly from the component is that you can test your pipeline at each stage of the process without needing to complete and run it.

We want to display this data in a transformation pipeline, which we'll create now, so we can reference its name in our orchestration pipeline, using the Run Transformation component. A transformation pipeline is where you add one or more components to transform your data.


Build a transformation pipeline

Let's begin by adding a transformation pipeline.

  1. On the pipeline canvas, click the Pipelines tab in the top-left.
  2. Click Add to open the menu.
  3. Select Transformation pipeline. The Add transformation pipeline dialog will open.
  4. Enter an appropriate name for your pipeline. For example, you could name it "transform-jira-data".
  5. Click Add.

    Create transformation pipeline

  6. Return to your orchestration pipeline and add the Run Transformation component.

  7. Use the Transformation Pipeline drop-down menu to select the name of your newly created transformation pipeline.
  8. Validate and Run the pipeline. Adding the Run Transformation component means we can run our transformation pipeline as part of our orchestration pipeline. When you run a pipeline you can see its progress in the Task History tab. For more information, read Task history tab.

    Add Run Transformation pipeline

  9. Return to your transformation pipeline. We want to transform our Epics table data after extracting and loading it in to our orchestration pipeline earlier and transform it into something useful. To do this, add the Table Input component. For specific property information, refer to the following table:

Property Value
Target Table Select the target table name you set in the Jira Query component earlier.
Column Names Select all the columns from the Epics database: BoardId, Done, Id, Key, Name, and Summary.

Note

You can leave the remainder of the properties set to their default values.

Select column names

Validate the pipeline and sample the data from the Sample data tab, just like you did earlier for your orchestration pipeline. For more information, read validate, and sample the data, respectively.

Sample data from Table Input


Filter the data

We now need to add a Filter component to our transformation pipeline to exclude data we don't want to see. In this case, we will use a filter to only show Epics that are not completed, because they are in progress or have not yet been started. To do this, add the Filter component to the pipeline canvas. You can obtain the component property information from the linked Filter documentation. For specific property information, refer to the following table:

Property Value
Filter Conditions Set the filter to only show the Done Input Column to false.

In the Filter Conditions dialog, set the filter to meet the following conditions:

  • Input Column: Done
  • Qualifier: Is
  • Comparator: Equal to
  • Value: false

The Done column indicates whether an Epic has been closed, i.e. True. An open Epic will therefore have False as the value, and as such, the filter will exclude all completed Epics, and only return rows where the status is False (incomplete).

Note

You can leave the remainder of the properties set to their default values.

Filter data

Now that our filter is configured, we need to designate a table in which to store our filtered data. To do this, add the Table Output component. For specific property information, refer to the following table:

Property Value
Target Table Select the target table name you set in the Jira Query component earlier.
Column Mapping Set the Input Column and Output Column to the following: Done, Id, Name, and Summary.

Note

You can leave the remainder of the properties set to their default values.

Validate the pipeline and Run it. For more information, read validate.

Table Output data

When you run a pipeline you can see its progress in the Task History tab. For more information, read Task history tab.

Now that our pipeline is complete, the next thing to do is to push it to the main branch so that it is saved to the Matillion-hosted Git repository, and can be published for other users to see and use.


Use Git to push local changes to the main branch

Currently, these pipelines have only been created locally in our instance of the branch we're using. We now need to commit and push our workflow changes to the remote repository. After this has been done. The pipelines can be viewed and accessed by other users in our project when they pull remote changes. To do this, click the downward facing arrow in your current Git branch at the top of the page. The Git menu will be displayed. Click Commit changes. For more information, read Git commit.

Committing local changes

You can enter an optional commit message describing the work that has been done, then click Commit.

Commit changes dialog

A message will appear saying your changes have been saved to your local branch. We now want to push our local commits to the remote branch. To do this:

  1. Click Push local changes.
  2. Click Push. Another message will appear saying your local commits have been pushed to the remote branch. For more information, read Git push local changes.

Push local commits to remote

Now, let's go in to our main branch. Our final step is to merge our changes from the remote development branch to the main branch. To do this:

  1. Click Merge from branch.
  2. In the Merge from (Remote Branch) drop-down menu, make sure you select the branch you created earlier, which has our pipeline workflow saved to it. Merge into (Local Branch) will already be set to the main branch.
  3. Click Merge. For more information, read Git merge.

Merge changes

Next, we want to publish and schedule our pipeline. Publishing is only permitted on the main branch. This is another reason why it's a requirement to push our changes to main once they're finalized.


Publish a pipeline

To publish our pipelines:

  1. Make sure you're in the main branch.
  2. Click Push local changes in the main branch tab at the top. For more information, read Git push local changes.
  3. Toggle to enable Publish branch. Publishing your pipelines will make them available for scheduling, which we will be doing next.
  4. Select the Environment to publish to. In this case, select My first project-default.
  5. Click Push & Publish.

Push and Publish pipeline

Toast notifications will appear in the top-right of the Designer UI informing you your push was successful, your local commits have been pushed to the remote branch, and the publish was successful, stating your branch has been published.


Schedule a pipeline

We now have a published pipeline workflow, and we can also run our orchestration and transformation pipelines whenever we want to.

Let's create a schedule for our orchestration pipeline to extract the new Epics data in Jira, then it'll use the Run Transformation component to filter it. To do this, from the main branch in our pipelines, click the Add Schedule button in the top-right.

Click Schedule.

You will be taken away from the pipeline canvas, and the Create a new schedule page will be displayed. Enter the following schedule properties:

Property Description
Name A unique, descriptive name for the schedule. You can call it "Schedule Epics data".
Environment Select the environment configuration that the schedule will run pipelines using. In this case, select our default Snowflake environment, My first project-default.
Published pipeline Select a pipeline to run in the schedule. In this case, select our orchestration pipeline, "extract-jira-data".
Timezone Remain in the Standard tab, and choose a Timezone for the schedule to adhere to. Choose Europe/London.
Starts on Select the date and time that the schedule will be enabled. This is not necessarily the time of the first schedule run. For example, if your start date and time is 1st August 2024 at 5:00pm, with a schedule set to run every day at 8:00am, your pipeline will first run at 8:00am on 2nd August 2024.
Repeat every The interval between schedule runs. Select 1.
Unit The units of the Repeat every interval. Options are Day (the default), Week, Hour, or Minute. Select Week.
Repeat on Select the day of the week you want your schedule to run. Select Mon.
Hour Select the hour and minute you want your schedule to begin. Select 8 and 0, respectively.

Click Create.

Schedule pipeline

Your project page will be displayed. A message will appear, saying Schedule Epics data has been created. Click on the Schedules tab to see your new schedule in the list. You can disable the schedule at any time, by clicking the Enabled toggle off.

Scheduled pipeline


Pipeline data flow

Now we have working, published, pipelines, lets use Designer's Data lineage to view a clear representation of the data flow in our transformation pipeline.

To access the Data Lineage feature, follow these steps:

  1. In the top-left of the Designer, click .
  2. Click Manage, then click Pipeline Runs.
  3. The Your activity page will be displayed. Click Lineage in the panel on the left.

We can review details about our selected Jira data source, the number of columns in the Epics table we selected earlier, column names, and column types.

Data lineage

To learn more about Data Lineage, read Data lineage.