You have an Azure subscription that contains an Azure Synapse Analytics workspace and a user named Used.
You need to ensure that User1 can review the Azure Synapse Analytics database templates from the gallery. The solution must follow the principle of least privilege.
Which role should you assign to User1?
You have an Azure Synapse Analytics dedicated SQL pod.
You need to create a pipeline that will execute a stored procedure in the dedicated SQL pool and use the returned result set as the input (or a downstream activity. The solution must minimize development effort.
Which Type of activity should you use in the pipeline?
You have an Azure subscription that contains an Azure Synapse Analytics workspace named ws1 and an Azure Cosmos D6 database account named Cosmos1 Costmos1 contains a container named container 1 and ws1 contains a serverless1 SQL pool.
you need to ensure that you can Query the data in container by using the serverless1 SQL pool.
Which three actions should you perform? Each correct answer presents part of the solution
NOTE: Each correct selection is worth one point.
You have an Azure Data Factory pipeline named pipeline1 that is invoked by a tumbling window trigger named Trigger1. Trigger1 has a recurrence of 60 minutes.
You need to ensure that pipeline1 will execute only if the previous execution completes successfully.
How should you configure the self-dependency for Trigger1?
Tumbling window self-dependency properties
In scenarios where the trigger shouldn't proceed to the next window until the preceding window is successfully completed, build a self-dependency. A self-dependency trigger that's dependent on the success of earlier runs of itself within the preceding hour will have the properties indicated in the following code.
Example code:
'name': 'DemoSelfDependency',
'properties': {
'runtimeState': 'Started',
'pipeline': {
'pipelineReference': {
'referenceName': 'Demo',
'type': 'PipelineReference'
}
},
'type': 'TumblingWindowTrigger',
'typeProperties': {
'frequency': 'Hour',
'interval': 1,
'startTime': '2018-10-04T00:00:00Z',
'delay': '00:01:00',
'maxConcurrency': 50,
'retryPolicy': {
'intervalInSeconds': 30
},
'dependsOn': [
{
'type': 'SelfDependencyTumblingWindowTriggerReference',
'size': '01:00:00',
'offset': '-01:00:00'
}
]
}
}
}
You have an Azure subscription that contains an Azure Data Factory data pipeline named Pipeline1, a Log Analytics workspace named LA1, and a storage account named account1.
You need to retain pipeline-run data for 90 days. The solution must meet the following requirements:
* The pipeline-run data must be removed automatically after 90 days.
* Ongoing costs must be minimized.
Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.
Submit Cancel
Currently there are no comments in this discussion, be the first to comment!