DP-100: Designing and Implementing a Data Science Solution on Azure

9%

Question 11

You have written a script that uses the Scikit-Learn framework to train a model.
Which framework-specific estimator should you use to run the script as an experiment?
PyTorch
Tensorflow
SKLearn




Answer is SKLearn. To run a scikit-learn training script as an experiment, use the generic Estimator estimator or a SKLearn estimator.

Question 12

You have run an experiment to train a model.
You want the model to be stored in the workspace, and available to other experiments and published services.
What should you do?
Register the model in the workspace.
Save the model as a file in a Compute Instance.
Save the experiment script as a notebook.




Answer is Register the model in the workspace. To store a model in the workspace, register it.

Question 13

You have uploaded some data files to a folder in a blob container, and registered the blob container as a datastore in your Azure Machine Learning workspace.
You want to run a script as an experiment that loads the data files and trains a model.
What should you do?
Save the experiment script in the same blob folder as the data files.
Create a data reference for the datastore location and pass it to the script as a parameter.
Create global variables for the Azure Storage account name and key in the experiment script.




Answer is Create a data reference for the datastore location and pass it to the script as a parameter.

To access a path in a datastore in an experiment script, you must create a data reference and pass it to the script as a parameter. The script can then read data from the data reference parameter just like a local file path.

Question 14

You have a CSV file containing structured data that you want to use to train a model.
You upload the file to a folder in an Azure Storage blob container, for which a datastore is defined in your workspace. Now you want to create a dataset for the data so that it can be easily used as a Pandas dataframe.
Which kind of dataset should you create?
A file dataset
A tabular dataset




Answer is A tabular dataset. Use a tabular dataset for structured data that you want to work with in a Pandas dataframe.

Question 15

You have registered a dataset in your workspace.
You want to use the dataset in an experiment script that is run using an estimator.
What should you do?
Pass the dataset as a named input to the estimator.
Create a data reference for the datastore location where the dataset data is stored, and pass it to the script as a parameter.
Use the dataset to save the data as a CSV file in the experiment script folder before running the experiment.




Answer is Pass the dataset as a named input to the estimator. To access a dataset in an experiment script, pass the dataset as a named input to the estimator.

Question 16

You need to create a compute target for training experiments that require a graphical processing unit (GPU).
You want to be able to scale the compute so that multiple nodes are started automatically as required.
Which kind of compute target should you create.
Compute Instance
Compute Cluster
Inference Cluster




Answer is Compute Cluster. Use a compute cluster to create multiple nodes of GPU-enabled VMs that are started automatically as needed.

Question 17

You are using an estimator to run an experiment, and you want to run it on a compute cluster named training-cluster-1. Which property of the estimator should you set to run the experiment on training-cluster-1?
compute_target = 'training-cluster-1'
environment_definition = 'training-cluster-1'
source_directory = 'training-cluster-1'




Answer is compute_target = 'training-cluster-1' To specify an compute target for an estimator, use the compute_target parameter.

Question 18

You are creating a pipeline that includes a step to train a model using an estimator.
Which kind of step should you define in the pipeline for this task.
PythonScriptStep
DatabricksStep
EstimatorStep




Answer is EstimatorStep. Use an EstimatorStep to run an estimator in a pipeline.

Question 19

You are creating a pipeline that includes two steps.
Step 1 preprocesses some data, and step 2 uses the preprocessed data to train a model.
What type of object should you use to pass data from step 1 to step 2 and create a dependency between these steps?
Datastore
PipelineData
Data Reference




Answer is PipelineData. To pass data between steps in a pipeline, use a PipelineData object.

Question 20

You have used the Python SDK for Azure Machine Learning to create a pipeline that trains a model.
What do you need to do so that a client application can invoke the pipeline through an HTTP REST endpoint?
Rename the pipeline to pipeline_name-production.
Publish the pipeline.
Create an inference cluster compute target.




Answer is Publish the pipeline. You must publish a pipeline to create an endpoint.

< Previous PageNext Page >

Quick access to all questions in this exam