IBM VEST Workshops
40 min
Last updated 06/05/2023

101: Accessing data

Introduction

As a data scientist, you have been assigned to analyze employee attrition at your retail electronics company. In 2018, your company, facing severe attrition, built a model to predict which employees would leave the company, allowing the human resources department to proactively work to retain them. However, the model has lost some of its predictive power thanks to changing conditions such as the COVID-19 pandemic, subsequent emphasis on remote work, and supply chain disruptions.

You will need to work with data from across the enterprise to develop potential replacement models. You will deploy and evaluate those models, and track them throughout their lifecycle, gathering metrics and metadata to help determine if and when they are ready to be deployed into production.

Prerequisites

Note: The content below can be completed entirely using free services; however, you may exceed your monthly free data allowance for Watson Studio unless you disable active runtimes as shown at the end of this lab. If possible, use a paid account instance of Watson Studio. Otherwise, you may need to wait until your data credits refresh to finish the lab.

Before you can proceed, you need to set up your environment. This involves importing a project and setting up its machine learning service. This lab has been built on Cloud Pak for Data as a Service (Cloud). For on-premises environments, platform details such as API key creation and data access via object storage will differ, but the vast majority of the lab is still applicable. It is expected that you complete this lab on the cloud.

1. Download the project file

Throughout this lab, you will use a Watson Studio project. This project has been prebuilt, and is provided for you in a zip file that is located in the Data and AI Live Demos GitHub repository. Before proceeding, please download the Trusted-AI-L3-Tech-Lab.zip project file.

2. Log in to Cloud Pak for Data

If you have not already done so, using the instructions and the screenshots below, log in to IBM Cloud Pak for Data. You will use your personal Cloud Pak for Data as a Service account in the Dallas region to do the lab.

login_screen
  1. Using the Work with resources in this IBM Cloud region: dropdown, select the Dallas region.

    • This is very important because all of the cloud services that you were instructed to provision in the Getting started instructions, should be provisioned in the Dallas region.
  2. Enter your IBM ID that you used to sign up for Cloud Pak for Data.

    • This should be your personal IBM Cloud ID. If you have logged in to Cloud Pak for Data previously and checked the Remember IBMid check box, it may already be filled in for you. If you have multiple Cloud Pak for Data accounts you can click inside the IBMid area above the Continue button and a list of available IBM IDs will be displayed for you to choose from.
  3. Click the Continue button

If your IBM ID is recognized as an IBM w3id (internal IBM w3 intranet id), you will be prompted to enter your password using the w3id on IBM Security Verify dialog:

sso_login
  1. Enter the password for your IBM w3id.
  2. Click the Sign in button.

If your IBM Id is not an IBM w3id, you will be prompted to enter your password using the standard IBM Security Verify dialog:

nonibm_login
  1. Enter the password for your IBM ID.
  2. Click the Log in button

3. Validate Your Environment

You will be brought to the Cloud Pak for Data home page. If at any point during these instructions you get a pop-up message from Cloud Pak for Data asking to guide or assist you, dismiss it and simply follow the steps in the instructions.

Make sure you are in your personal cloud account

personal _account

You can be a member of many cloud accounts. This lab is done from your personal cloud account. The account displayed should show Your cloud account ID - Your name. For example, in the screen shot above, the personal account is 1792871 - Rick Buglio. If your personal account is selected, like the screen shot above, you can proceed to the next step. If not, follow the instructions below to switch to your personal account. Note that the full menu bar including your account details and region may only appear if your browser window is maximized.

techzone_account

In the screenshot above, the user is in the Techzone Outcomes cloud account and needs to switch to their personal account. This is a common scenario because IBM sellers and business partners could still have a reservation in the Data and AI Live Demos (Techzone Outcomes) cloud account to do the Data Fabric Level 2 sales demos. When you login to Cloud Pak for Data, it remembers the last account you logged into, so it's a good practice to check what account you are in before proceeding.

  1. Using the Cloud account dropdown on the toolbar, select your personal cloud account.

Make sure you are in the Dallas region

This is very important. Based on the Getting started instructions, you should have your services provisioned in the Dallas region.

You cannot work across cloud service regions in Cloud Pak for Data as a Service. Only global services, like Cloud Object Storage, can be used across cloud service regions. Therefore, your services must all be in the same region for you to complete this lab.

london

In the screen shot above, the user is in the London region and needs to switch to the Dallas region. If your region is not set to the Dallas region perform the following steps:

  1. Using the Region dropdown on the toolbar, select the Dallas region.

4. Import the lab project

WorkWithData
  1. From the Watson Studio Cloud welcome screen, click on Work with data.
  2. Click on Create a project from a sample or file.
DropinProject
  1. In the Upload file section, drop in the Trusted-AI-L3-Tech-Lab.zip file you've just downloaded.
  2. In the Name field, give the project a name. For example: "Trusted AI L3 Tech Lab".
  3. Use the Select storage service dropdown to select your Cloud Object Storage instance.
  4. Click on Create. The project will be imported, which will take about a minute.
view_import_summary
  1. When the import has finished, click on View import summary.
successful_import
  1. Verify that all assets have been successfully imported. If they have, click Close and move to step 11. If one or more assets has failed to import, proceed to step 9.
delete_project
  1. If one or more assets failed to import, go to your list of projects and select the checkbox associated with the new project.
  2. Click the blue Delete button in the top right and follow the prompts on screen to confirm deletion. Return to the new project screen and repeat this section, starting from step 2.
  3. Once your project has been successfully imported, you may proceed with the rest of the lab.

5. Add your Watson Machine Learning (WML) service

The WML service was created earlier as described on the Getting started page.

AddWMLService
  1. In your new project, select the Manage tab.
  2. Select Services & integrations from the left menu bar.
  3. Click on the blue Associate service button on the right.
AssociateSvc
  1. Click on the checkbox associated with the Machine learning service you provisioned. Note that you may have multiple services listed here, so be sure to select one with Machine Learning in the Type column of the table.
  2. Click on the blue Associate button at the lower right of the screen.

Accessing data

Data science is nothing without data. In this section, you will learn how to make data assets available in a project and how to access the data through a notebook. Data can also be accessed through other available tools, which is covered in the No-code and low-code solutions section of this lab.

Cloud Pak for Data accesses data through two forms: files and connections. Files are not limited to the comma separated values (csv) format. For example, data can come as archives in a compressed zip or tar format.

Connections are access points to different repositories. Here is a partial list of the connections available in Cloud Pak for Data as a Service:

connections

1. Add a connection to a project

In the Getting started section of the labs, you were instructed to create a Db2 service. This is the database service used in this lab. It's a perfect example of creating and accessing a connection.

Enterprises have critical data sets spread over multiple repositories; to add more complexity, those data assets can reside on-premises or in the cloud. It is critical for data science practitioners to have easy access to this data, while keeping the data in its original location to reduce cost and complexity. Copying the data leads to confusion, where, for example, users would not know if they are using the most up-to-date data. As the number of data copies increases with the number of projects in the enterprise, it leads to confusion on which copy represents the real version of the truth. Over time, different copies of the same data leads to partial versions of the truth, which makes it nearly impossible to decipher.

As shown earlier, Cloud Pak for Data supports a wide variety of connection types. Let's see how to create a connection.

  1. Get the credentials for your Db2 service. Open the Services and software resource list from the IBM Cloud console.
  2. Click on the Db2 service.
Db2CredentialPage
  1. In the left side panel, select Service credentials.
  2. If there are no credentials, click on the blue New credential button, then click on Add.
  3. Select the twistie for the credentials.
  4. Click the copy icon to copy your credentials to your clipboard. Paste them into a text editor, so the specific elements can be referenced in the Db2 connection creation below.

2. Create a database connection

Heads up! Parts of this section will be on the quiz.

In this section, you use the credential created (or found) in the previous step to create a connection to a Db2 database.

This example uses the administrator username and password of the administrator. In a real world situation, it would be better to create an additional user and limit its privileges. This is what was done for the use of the "Data Warehouse" connection that is part of this project.

  1. From your list of projects, open the one you imported in a previous step by clicking on its name.
AssetTab
  1. Select the Assets tab.
  2. Click on the blue New asset button.
  3. Click on the Connection tile.
connection_type
  1. From the list of connection types, scroll down and select IBM Db2 (use this specific name from all the Db2-type possibilities)
  2. Click the blue Select button.
  3. Enter a name for your connection (suggested: Db2).

For steps 8-12 below, you will be copying values from the credentials you pasted into a text editor into the form. Copy the values between the quotations from the credentials text. The values you will need are database, hostname, port, username, and password. They can be found in the credentials text as follows:

Db2Json

Copy the values one at a time, and paste them into the corresponding fields:

db_fields
  1. Scroll down to the Database field and enter the database credential element (default bludb).
  2. In the Hostname or IP address field, enter the hostname credential element.
  3. In the Port field, enter the port credential element
  4. Scroll down to the Credentials section. In the Username field, enter the username credential element. The username also serves as the schema name.
  5. In the Password field, enter the password credential element.
  6. Scroll down to the Certificates section and select the checkbox for Port is SSL-enabled. While adding SSL credentials was not strictly necessary for this connection, selecting the option allows you to add a level of security by allowing the use of the connection only from trusted hosts.
  7. Click Test connection in the upper right. If the test was unsuccessful, verify that you have correctly entered the credentials.
  8. Once the connection test is successful, click the blue Create button in the bottom right.
  9. Click Create in the pop-up window to confirm.

3. Create connected data

Connected data allows the use of data coming from any data repository as if it were a file, without copying it to a new location. This provides a way for data scientists and other team members to access the exact data they want instead of having to use SQL or other languages to get to that data.

FilterResources
  1. To create and populate a table, begin by opening the IBM Cloud Resource list and clicking on the Db2 instance.

Use the filter by Name and type "Db2" to quickly get to this resource

GoToUI
  1. Select the Manage tab.
  2. Click on the Go to UI button. The IBM Db2 on Cloud dashboard opens in a new browser tab.
Db2Icons
  1. Click on the SQL icon on the left.
run_all
  1. Enter the following SQL statement:
CREATE TABLE testing ( col1 int, col2 int, col3 int );
sql
  1. Click on the blue Run all button at the top right.
  2. Using a local text editor, create and save a file named test.csv, on your computer with four rows of three columns in the following format:
col1,col2,col3 8,1,6 3,5,7 4,9,2
txt
DataIcon
  1. Click on the Data icon just below the SQL icon.
file_selection
  1. Drag and drop the file from your hard drive into the file selection box. You will see it appear under Selected file on the right when uploaded.
  2. Click the blue Next button in the bottom right.
create_table
  1. Select your schema. It is the username from the credential above and should be the only schema listed.
  2. Click the New table button in the Table cell. Name your table TESTING.
  3. Click Create to create the table.
  4. Click Next three times to skip through the remaining screens until you see the Begin Load button.
  5. Click Begin Load.
LoadCompleted

This should complete with three rows read and three rows loaded. You can close this tab. In the next step you return to your project asset tab in Cloud Pak for Data as a Service (Watson Studio).

4. Create a connected data asset pointing to the table

  1. Return to the tab with your Watson Studio project. If you have closed the tab, open your list of projects and click on the name of your project to open it.
ImportAsset
  1. Select the Assets tab.
  2. Click on Import asset.
connected_data_import
  1. Select Connected data.
  2. Click on the Db2 connection (or the name you gave to your connection).
  3. Select your Db2 username as the schema name.
  4. Select the TESTING table and click the blue Select button in the lower right.
  5. Keep the default name and click Import.

5. Accessing files and connections through a notebook

Now that you have multiple files you can use, let's see how they can be accessed in a Jupyter notebook. Access to these files can also be done using tools such as AutoAI and SPSS Modeler. This will be explored later in this lab.

For a quick tutorial on Jupyter notebooks, click here. While in the notebook, you can use the blue Play button or press Shift + Enter on your keyboard to execute the current cell. Executing cells will be marked with an asterisk in the brackets to the left of the cell. When the cell has finished executing, the asterisk will be replaced with a number. When executing cells in the notebook, it's often best to wait until a cell has finished before executing the next one, which allows you to address any errors in the code.

In the previous (optional) section, you imported a dataset that includes masked columns. As you'll see in this notebook, these columns are unmasked when using a database connection (in this case, Db2). To preserve the masking, you need to use a Watson Query (data virtualization) connection instead.

First let's start by opening the notebook we'll be working with in this step. To get to the notebook, you start from within your project under the Assets tab.

  1. Click on the three vertical dots at the end of the line for 01-Data access lab. A drop down menu appears.
  2. Click on Edit. Once the notebook is open, follow the instructions provided.
EditNotebook

Now that you're in the editor, one key activity in the Data access notebook is something called Code Snippets. To access this capability, from a notebook in edit mode, you need to open the data tab. To do so, follow these steps:

  1. The figure below has a red circle around the code icon. Once you click this icon the data tab will then open on the right.
  2. This tab will have a button titled Select data from project. Go ahead and click this button.
  3. A new menu will pop open allowing you to select which data items you would like to connect. Start by clicking Data asset
  4. Another section will now display. Now click the modeling_records_2022.csv file name.
  5. You can now click the Select button in this menu. The Code Snippets sidebar will then reappear.
  6. In the Code Snippets sidebar your selected data item and the options for how to load it ("pandas DataFrame", "Credentials") will now be visible.

You'll notice there is a button titled Insert code to cell at the bottom of the sidebar now. Don't click this yet. Follow the steps in the notebook to place your cursor in the proper cell before inserting the code snippet.

InsertToCodeSelectAsset

Now that you know how to load data into a notebook make sure you read through the rest of notebook and complete each step.

Stopping active runtimes

After you complete the edit/run session in the notebook, Watson Studio will keep the notebook runtime active until the usage timeout is reached. This is convenient if you have multiple notebooks using the same environment in a project as they all use the same runtime. However, it can cause the limited capacity unit hours (CUH) in a lite plan instance to be used up too quickly in the lab.

To find and stop runtime environments in a project when you complete a lab with a Juypter notebook (like this lab), follow these steps.

ProjectEnvironments
  1. In the Cloud Pak for Data dashboard, click on the Project name breadcrumb to return to the project overview Trusted AI L3 Tech Lab.
  2. Click on the Manage tab.
  3. Select Environments from the left. You should see 1 active runtime that was started when you opened the notebook editing session.
StopActiveRuntime
  1. Select the active runtime in the checkbox
  2. Click on the button on the right to stop the runtime instance.
  3. When prompted, confirm by clicking on the Stop button.

Complete these tasks after each lab where you use a Juypter notebook and your Watson Studio CUH usage will be optimized and saved for future use. These steps apply to all runtimes, including Spark and GPU runtime environments (these have a default timeout of 3 hours).