R Basics-26 - Coding-Iterations-2 (Hands On) - ClassiCoder
Iterations
Perform the following tasks:
. Write a function sum_whole that find the sum of the first N (passed as arguments) whole numbers. Assign the result to the variable ans and return it.
Iterations-2-R-Basic
Iterations-2-R-Basic-Solution
42 comments
UnknownJune 10, 2022 at 6:02 AM
Test cases is not passed
AnonymousJuly 11, 2022 at 3:25 AM
not run
AnonymousApril 17, 2023 at 1:13 AM
You are designing an Azure Synapse Analytics solution that will be used to analyze patient outcome data from a hospital.
Which database template should you use?
Select only one answer.
Healthcare Insurance
Healthcare Provider
Life Insurance & Annuities
Pharmaceuticals
AnonymousApril 17, 2023 at 1:14 AM
You create a Microsoft Purview account and add an Azure SQL Database data source that has data lineage scan enabled.
You assign a managed identity for the Microsoft Purview account and the db_owner role for the database.
After scanning the data source, you are unable to obtain any lineage data for the tables in the database.
You need to create lineage data for the tables.
What should you do?
Select only one answer.
Use SQL authentication.
Use a user-managed service principal.
Create a master key in the database.
Create a certificate in the database.
AnonymousApril 17, 2023 at 1:16 AM
You have a database named DB1 and a data warehouse named DW1.
You need to ensure that all changes to DB1 are stored in DW1.The solution must capture the new value and the existing value and store each value as a new record.
What should you include in the solution?
Select only one answer.
change data capture
change tracking
transactional replication
merge replication
AnonymousApril 17, 2023 at 1:17 AM
You have a database named DB1 and a data warehouse named DW1.
You need to ensure that all changes to DB1 are stored in DW1. The solution must meet the following requirements:
Identify that a row has changed, but not the final value of the row. Minimize the performance impact on the source system. What should you include in the solution?
Select only one answer.
change data capture
change tracking
snapshot replication
merge replication
AnonymousApril 17, 2023 at 1:18 AM
You plan to implement a data storage solution for a healthcare provider.
You need to ensure that the solution follows industry best practices and is designed in the minimum amount of time.
What should you use?
Select only one answer.
Azure Synapse Analytics database templates
Azure Quickstart guides
Azure Data Factory
Azure Resource Manager (ARM) templates
AnonymousApril 17, 2023 at 1:02 PM
You need to implement encryption at rest by using transparent data encryption (TDE).
You implement a master key.
What should you do next?
Select only one answer.
Create a certificate that is protected by the master key.
Create a database encryption key.
Turn on the database encryption process.
Back up the master database.
AnonymousApril 17, 2023 at 1:03 PM
You are implementing an application that queries a table named Purchase in an Azure SQL database.
The application must show data only for the currently signed-in user.
You use row-level security (RLS), implement a security policy, and implement a function that uses a filter predicate.
Users in the marketing department report that they cannot see their data.
What should you do to ensure that the marketing department users can see their data?
Select only one answer.
Grant the SELECT permission on the Purchase table to the Marketing users.
Grant the SELECT permission on the function to the Marketing users.
Implement a blocking predicate.
Rebuild the function with SCHEMABINDING=OFF.
AnonymousApril 17, 2023 at 1:05 PM
You need to grant a user access to write data to an Azure Data Lake Storage Gen2 account.
Which security technology should you use to grant the access?
Select only one answer.
RBAC
NTFS
ACL
OAuth 2.0 Bearer Tokens
AnonymousApril 17, 2023 at 1:07 PM
You have a Microsoft 365 subscription and an Azure subscription that are linked to an Azure AD tenant. The Azure subscription includes an Azure SQL database named DB1.
You need to ensure that when Azure AD users connect to DB1, they are prompted for multi-factor authentication (MFA). The solution must ensure that the users are not required to use MFA when accessing Microsoft Office 365.
What should you configure?
Select only one answer.
Conditional Access
Azure AD
Azure SQL Database
Per-User MFA
AnonymousApril 17, 2023 at 1:09 PM
You have an Azure Synapse Analytics workspace.
You need to measure the performance of SQL queries running on the dedicated SQL pool.
Which two actions achieve the goal? Each correct answer presents a complete solution
Select all answers that apply.
From the Monitor page of Azure Synapse Studio, review the SQL requests tab.
Query the sys.dm_pdw_exec_request view.
Query the sys.dm_pdw_exec_sessions view.
From the Monitor page of Azure Synapse Studio, review the Pipeline runs tab.
AnonymousApril 17, 2023 at 1:11 PM
You have a job that aggregates data over a five-second tumbling window.
You are monitoring the job and notice that the SU (Memory) % Utilization metric is more than 80 percent, and the Backlogged Input Events metric shows values greater than 0.
What should you do to resolve the performance issue?
Select only one answer.
Increase the number of the Streaming Units (SU).
Change the compatibility level.
Create a user-defined aggregate to perform the aggregation.
Change the tumbling window to a snapshot window.
AnonymousApril 17, 2023 at 1:13 PM
You have an Azure Data Factory named ADF1.
You need to review Data Factory pipeline runtimes for the last seven days. The solution must provide a graphical view of the data.
What should you use?
Select only one answer.
the Gantt view of the pipeline runs
the List view of the pipeline runs
the Dashboard view of the pipeline runs
the Overview tab of Azure Data Factory Studio
AnonymousApril 17, 2023 at 1:14 PM
You have an Azure subscription that uses Microsoft Defender for Storage.
A storage anomaly is detected in Defender for Storage.
You need to identify the account that performed the suspicious action.
Which two activity details can be used to identify the account? Each correct answer presents a complete solution.
Select all answers that apply.
application
IP address
device name
storage account
AnonymousApril 17, 2023 at 1:16 PM
You have an Apache Spark pool in Azure Synapse Analytics.
You run a notebook that creates a DataFrame containing a large amount of data.
You need to preserve the DataFrame in memory.
Which two transformations can you use? Each correct answer presents a complete solution.
Select all answers that apply.
cache()
persist()
take()
write()
AnonymousApril 17, 2023 at 1:17 PM
You monitor an Azure Data Factory pipeline that occasionally fails.
You need to implement an alert that will contain failed pipeline run metrics. The solution must minimize development effort.
Which two actions achieve the goal? Each correct answer presents a complete solution.
Select all answers that apply.
From Azure portal, create an alert and add the metrics.
From the Monitor page of Azure Data Factory Studio, create an alert.
Implement a Web activity in the pipeline.
Implement a WebHook activity in the pipeline.
AnonymousApril 17, 2023 at 1:22 PM
You have an Azure Synapse Analytics workspace that includes a table named Table1.
You are evaluating the use of a clustered columnstore index.
What is the minimum recommended number of rows for clustered columnstore indexes?
Select only one answer.
600,000
6 million
60 million
600 million
AnonymousApril 17, 2023 at 1:24 PM
You have an Azure Synapse Analytics workspace.
You need to build a materialized view.
Which two items should be included in the SELECT statement of the view? Each correct answer presents part of the solution.
Select all answers that apply.
an aggregate function
the GROUP BY clause
the HAVING clause
a subquery
the OPTION clause
AnonymousApril 17, 2023 at 1:26 PM
You have a solution that transforms data in an ELT process. The solution includes a column named fulladdress that contains address data in the following format: [street address], [city] [state] [postal code].
You need to ensure that the solution will support real-time processing and batch processing of events.
Which two services should you include in the solution? Each correct answer presents part of the solution.
Select all answers that apply.
Azure Event Hubs
Azure Stream Analytics
Azure Cosmos DB
Azure Data Factory
AnonymousApril 17, 2023 at 1:33 PM
You have a Delta Lake solution that contains a table named table1.
You need to roll back the contents of table1 to 24 hours ago.
Which command should you run?
Select only one answer.
RESTORE TABLE employee TO TIMESTAMP AS OF current_timestamp() - INTERVAL '24' HOUR;
VACUUM employee RETAIN 24;
COPY INTO employee1
ALTER TABLE employee
AnonymousApril 17, 2023 at 1:37 PM
You have an Azure Stream Analytics solution that receives data from multiple thermostats in a building.
You need to write a query that returns the average temperature per device every five minutes.
Which windowing function should you use?
Select only one answer.
TumblingWindow
HoppingWindow
SlidingWindow
SessionWindow
AnonymousApril 17, 2023 at 1:37 PM
You create an Azure Stream Analytics job. You run the job for five hours.
You review the logs and notice multiple instances of the following message.
{"message Time":"2019-02-04 17:11:52Z","error":null, "message":"First Occurred: 02/04/2019 17:11:48 | Resource Name: ASAjob | Message: Source 'ASAjob' had 24 data errors of kind 'LateInputEvent' between processing times '2019-02-04T17:10:49.7250696Z' and '2019-02-04T17:11:48.7563961Z'. Input event with application timestamp '2019-02-04T17:05:51.6050000' and arrival time '2019-02-04T17:10:44.3090000' was sent later than configured tolerance.","type":"DiagnosticMessage","correlation ID":"49efa148-4asd-4fe0-869d-a40ba4d7ef3b"}
You need to ensure that these events are not dropped.
What should you do?
Select only one answer.
Increase the tolerance for late arrivals.
Increase the tolerance for out-of-order events.
I…
AnonymousApril 17, 2023 at 1:41 PM
You have an Azure Stream Analytics job named Job1.
Job1 runs continuously and executes non-parallelized queries.
You need to minimize the impact of Azure node updates on Job1. The solution must minimize costs.
To what should you increase the Scale Units (SUs)?
Select only one answer.
2
3
6
12
AnonymousApril 17, 2023 at 1:42 PM
You are building a data pipeline solution in Azure Data Factory.
You need to ensure that the solution outputs the data in the source format and in the report format.
Which data processing technique should you use?
Select only one answer.
ETL
ELT
ELTL
AnonymousApril 17, 2023 at 1:44 PM
You are creating an Azure Data Factory pipeline.
You need to store the passwords used to connect to resources.
Where should you store the passwords?
Select only one answer.
Azure Key Vault
Data Factory
Azure SQL Database
Azure Repos
AnonymousApril 17, 2023 at 1:45 PM
You are configuring Azure Data Factory to be used in a CI/CD deployment process.
You need to minimize the administrative tasks required by using global parameters.
Which global parameters should you configure?
Select only one answer.
server names within a connection object
execution schedule
sink task name
target database version number
AnonymousApril 17, 2023 at 2:28 PM
You have an Azure Data Lake Storage account named store.dfs.core.windows.net and an Apache Spark notebook named Notebook1.
You plan to use Notebook1 to load and transform data in store.dfs.core.windows.net.
You need to configure the connection string for Notebook1.
You have an Azure subscription that contains an Azure Synapse Analytics workspace.
You use the workspace to perform ELT activities that can take up to 30 minutes to complete.
You develop an Azure function to stop the compute resources used by Azure Synapse Analytics during periods of zero activity.
You notice that it can take more than 20 minutes for the compute resources to stop.
You need to minimize the time it takes to stop the compute resources. The solution must minimize the impact on running transactions.
How should you change the function?
Select only one answer.
Add a timer to wait 20 minutes before stopping the compute resources.
Check the sys.dm_operation_status dynamic management view until no transactions are active in the database before stopping the compute resources.
Close all connections to the database before stopping the compute resources.
Set the database to READ_ONLY before stopping the compute resources.
AnonymousApril 17, 2023 at 2:35 PM
You design an Azure Data Factory data flow activity to move large amounts of data from text files to an Azure Synapse Analytics database. You add a data flow script to your data flow. The data flow in the designer has the following tasks:
DistinctRows: Aggregate data by using myCols that produce columns. source1: Import data from DelimitedText1. derivedColumn1: Create and update the C1 columns. select1: Rename derivedColumn1 as select1 with columns C1. sink1: Add a sink dataset. You need to ensure that all the rows in source1 are deduplicated.
What should you do?
Select only one answer.
Create a new aggregate task after source1 and copy the script to the aggregate task.
Change the incoming stream for DistinctRows to source1.
Change the incoming stream for derivedColumn1 to DistinctRows.
Create a new flowlet task after source1.
AnonymousApril 17, 2023 at 2:39 PM
You have 500 IoT devices and an Azure subscription.
You plan to build a data pipeline that will process real-time data from the devices.
You need to ensure that the devices can send messages to the subscription.
What should you deploy?
Select only one answer.
an Azure event hub
an Azure Stream Analytics workspace
an Azure Storage account
AnonymousApril 17, 2023 at 2:41 PM
You are building an Azure Stream Analytics pipeline.
You need to ensure that data in the pipeline is analyzed using fixed five-minute windows.
Which windowing function should you use?
Select only one answer.
TumblingWindow
HoppingWindow
SlidingWindow
SessionWindow
AnonymousApril 17, 2023 at 2:45 PM
You have an Azure Stream Analytics job named Job1. Job1 is configured to use one Streaming Unit (SU) and can be parallelized for up to three nodes.
You need to ensure there are three nodes available for the job.
What is the minimum number of SUs you should configure?
Select only one answer.
3
6
18
24
AnonymousApril 17, 2023 at 2:45 PM
You have an Azure Stream Analytics job named Job1. Job1 is configured to use one Streaming Unit (SU) and can be parallelized for up to three nodes.
You need to ensure there are three nodes available for the job.
What is the minimum number of SUs you should configure?
Select only one answer.
3
6
18
24
AnonymousApril 17, 2023 at 2:51 PM
You have an Azure Data Factory named datafactory1.
You configure datafacotry1 to use Git for source control.
You make changes to an existing pipeline.
When you try to publish the changes, you notice the following message displayed when you hover over the Publish All button.
Publish from ADF Studio is disabled to avoid overwriting automated deployments. If required you can change publish setting in Git configuration.
You need to allow publishing from the portal.
What should you do?
Select only one answer.
Change the Automated publish config setting.
Select Override live mode in the Git Configuration.
Use a git client to merge the collaboration branch into the live branch.
Use the browser to create a pull request.
AnonymousApril 17, 2023 at 2:54 PM
You plan to deploy Azure Data Factory.
You need to set the global parameters that will be used by the deployment.
Which two methods can you use? Each correct answer presents a complete solution.
Select all answers that apply.
an Azure Resource Manager (ARM) template
a PowerShell script
a Python script
Azure Data Studio
AnonymousApril 17, 2023 at 2:57 PM
You have an Azure Data Lake Storage Gen2 account.
You grant developers Read and Write permissions by using ACLs to the files in path \root\input\cleaned\.
The developers report that they cannot open the files.
How should you modify the permissions to ensure that the developers can open the files?
Select only one answer.
Grant Execute permissions to all folders.
Grant Execute permissions to the root folder only.
Add Execute permissions to the files.
Add Contributor permission to the developers.
AnonymousApril 17, 2023 at 3:03 PM
You need to store information about failed Azure Data Factory pipelines for three months.
Which three actions should you perform? Each correct answer presents part of the solution.
Select all answers that apply.
Create a storage account that has a lifecycle policy.
Create a Log Analytics workspace.
Add diagnostic settings and add Log Analytics as a target.
Add diagnostic settings and add Azure Event Hubs as a target.
From the Monitor page of Azure Synapse Studio, review the Pipeline runs tab.
AnonymousApril 17, 2023 at 3:09 PM
You have an Azure subscription that contains the following resources:
An Azure Synapse Analytics workspace named app1-syn An Azure Data Lake Storage Gen2 account named app1synstg A file system named data in app1synstg You upload a file named Data.parquet to app1synstg.
You need to query the first 100 rows of Data.parquet by using a SQL serverless pool.
Which query should you run?
Select only one answer.
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1synstg.dfs.core.windows.net/data/NYCTripSmall.parquet', FORMAT = 'PARQUET' ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1-syn.dfs.core.windows.net/data/NYCTripSmall.parquet', FORMAT = 'PARQUET' ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1synstg.dfs.core.windows.net/data/NYCTripSmall.parquet', SINGLE_CLOB ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1-syn.dfs.core.windows.net/data/NYCTripSmall.parquet', SINGLE_CLOB ) as result
AnonymousApril 17, 2023 at 3:15 PM
You have an Azure subscription that uses Microsoft Purview.
You need to identify which assets have been cataloged by Microsoft Purview.
What should you use?
Select only one answer.
the Microsoft Purview governance portal
Azure Data Studio
Azure Data Factory
the Microsoft Purview compliance portal
AnonymousMay 19, 2023 at 5:47 AM
sum_whole <- function(N){ if (N < 0) { stop("N must be a non-negative integer.") }
42 comments
Which database template should you use?
Select only one answer.
Healthcare Insurance
Healthcare Provider
Life Insurance & Annuities
Pharmaceuticals
You assign a managed identity for the Microsoft Purview account and the db_owner role for the database.
After scanning the data source, you are unable to obtain any lineage data for the tables in the database.
You need to create lineage data for the tables.
What should you do?
Select only one answer.
Use SQL authentication.
Use a user-managed service principal.
Create a master key in the database.
Create a certificate in the database.
You need to ensure that all changes to DB1 are stored in DW1.The solution must capture the new value and the existing value and store each value as a new record.
What should you include in the solution?
Select only one answer.
change data capture
change tracking
transactional replication
merge replication
You need to ensure that all changes to DB1 are stored in DW1. The solution must meet the following requirements:
Identify that a row has changed, but not the final value of the row.
Minimize the performance impact on the source system.
What should you include in the solution?
Select only one answer.
change data capture
change tracking
snapshot replication
merge replication
You need to ensure that the solution follows industry best practices and is designed in the minimum amount of time.
What should you use?
Select only one answer.
Azure Synapse Analytics database templates
Azure Quickstart guides
Azure Data Factory
Azure Resource Manager (ARM) templates
You implement a master key.
What should you do next?
Select only one answer.
Create a certificate that is protected by the master key.
Create a database encryption key.
Turn on the database encryption process.
Back up the master database.
The application must show data only for the currently signed-in user.
You use row-level security (RLS), implement a security policy, and implement a function that uses a filter predicate.
Users in the marketing department report that they cannot see their data.
What should you do to ensure that the marketing department users can see their data?
Select only one answer.
Grant the SELECT permission on the Purchase table to the Marketing users.
Grant the SELECT permission on the function to the Marketing users.
Implement a blocking predicate.
Rebuild the function with SCHEMABINDING=OFF.
Which security technology should you use to grant the access?
Select only one answer.
RBAC
NTFS
ACL
OAuth 2.0 Bearer Tokens
You need to ensure that when Azure AD users connect to DB1, they are prompted for multi-factor authentication (MFA). The solution must ensure that the users are not required to use MFA when accessing Microsoft Office 365.
What should you configure?
Select only one answer.
Conditional Access
Azure AD
Azure SQL Database
Per-User MFA
You need to measure the performance of SQL queries running on the dedicated SQL pool.
Which two actions achieve the goal? Each correct answer presents a complete solution
Select all answers that apply.
From the Monitor page of Azure Synapse Studio, review the SQL requests tab.
Query the sys.dm_pdw_exec_request view.
Query the sys.dm_pdw_exec_sessions view.
From the Monitor page of Azure Synapse Studio, review the Pipeline runs tab.
You are monitoring the job and notice that the SU (Memory) % Utilization metric is more than 80 percent, and the Backlogged Input Events metric shows values greater than 0.
What should you do to resolve the performance issue?
Select only one answer.
Increase the number of the Streaming Units (SU).
Change the compatibility level.
Create a user-defined aggregate to perform the aggregation.
Change the tumbling window to a snapshot window.
You need to review Data Factory pipeline runtimes for the last seven days. The solution must provide a graphical view of the data.
What should you use?
Select only one answer.
the Gantt view of the pipeline runs
the List view of the pipeline runs
the Dashboard view of the pipeline runs
the Overview tab of Azure Data Factory Studio
A storage anomaly is detected in Defender for Storage.
You need to identify the account that performed the suspicious action.
Which two activity details can be used to identify the account? Each correct answer presents a complete solution.
Select all answers that apply.
application
IP address
device name
storage account
You run a notebook that creates a DataFrame containing a large amount of data.
You need to preserve the DataFrame in memory.
Which two transformations can you use? Each correct answer presents a complete solution.
Select all answers that apply.
cache()
persist()
take()
write()
You need to implement an alert that will contain failed pipeline run metrics. The solution must minimize development effort.
Which two actions achieve the goal? Each correct answer presents a complete solution.
Select all answers that apply.
From Azure portal, create an alert and add the metrics.
From the Monitor page of Azure Data Factory Studio, create an alert.
Implement a Web activity in the pipeline.
Implement a WebHook activity in the pipeline.
You are evaluating the use of a clustered columnstore index.
What is the minimum recommended number of rows for clustered columnstore indexes?
Select only one answer.
600,000
6 million
60 million
600 million
You need to build a materialized view.
Which two items should be included in the SELECT statement of the view? Each correct answer presents part of the solution.
Select all answers that apply.
an aggregate function
the GROUP BY clause
the HAVING clause
a subquery
the OPTION clause
You need to retrieve the street address.
Which T-SQL statement should you use?
Select only one answer.
LEFT(fulladdress,CHARINDEX(‘,’,fulladdress))
LEFT(fulladdress,CHARINDEX(‘,’,fulladdress)-1)
SUBSTRING(fulladdress,1,16)
SUBSTRING(fulladdress,1,CHARINDEX(‘,’,fulladdress))
You need to ensure that the solution will support real-time processing and batch processing of events.
Which two services should you include in the solution? Each correct answer presents part of the solution.
Select all answers that apply.
Azure Event Hubs
Azure Stream Analytics
Azure Cosmos DB
Azure Data Factory
You need to roll back the contents of table1 to 24 hours ago.
Which command should you run?
Select only one answer.
RESTORE TABLE employee TO TIMESTAMP AS OF current_timestamp() - INTERVAL '24' HOUR;
VACUUM employee RETAIN 24;
COPY INTO employee1
ALTER TABLE employee
You need to write a query that returns the average temperature per device every five minutes.
Which windowing function should you use?
Select only one answer.
TumblingWindow
HoppingWindow
SlidingWindow
SessionWindow
You review the logs and notice multiple instances of the following message.
{"message Time":"2019-02-04 17:11:52Z","error":null, "message":"First Occurred: 02/04/2019 17:11:48 | Resource Name: ASAjob | Message: Source 'ASAjob' had 24 data errors of kind 'LateInputEvent' between processing times '2019-02-04T17:10:49.7250696Z' and '2019-02-04T17:11:48.7563961Z'. Input event with application timestamp '2019-02-04T17:05:51.6050000' and arrival time '2019-02-04T17:10:44.3090000' was sent later than configured tolerance.","type":"DiagnosticMessage","correlation ID":"49efa148-4asd-4fe0-869d-a40ba4d7ef3b"}
You need to ensure that these events are not dropped.
What should you do?
Select only one answer.
Increase the tolerance for late arrivals.
Increase the tolerance for out-of-order events.
I…
Job1 runs continuously and executes non-parallelized queries.
You need to minimize the impact of Azure node updates on Job1. The solution must minimize costs.
To what should you increase the Scale Units (SUs)?
Select only one answer.
2
3
6
12
You need to ensure that the solution outputs the data in the source format and in the report format.
Which data processing technique should you use?
Select only one answer.
ETL
ELT
ELTL
You need to store the passwords used to connect to resources.
Where should you store the passwords?
Select only one answer.
Azure Key Vault
Data Factory
Azure SQL Database
Azure Repos
You need to minimize the administrative tasks required by using global parameters.
Which global parameters should you configure?
Select only one answer.
server names within a connection object
execution schedule
sink task name
target database version number
You plan to use Notebook1 to load and transform data in store.dfs.core.windows.net.
You need to configure the connection string for Notebook1.
Which URI should you use?
Select only one answer.
abfss://container@store.dfs.core.windows.net/products.csv
https://container@store.dfs.core.windows.net/products.csv
adf://container@store.dfs.core.windows.net/products.csv
dbfs://container@store.dfs.core.windows.net/products.csv
You use the workspace to perform ELT activities that can take up to 30 minutes to complete.
You develop an Azure function to stop the compute resources used by Azure Synapse Analytics during periods of zero activity.
You notice that it can take more than 20 minutes for the compute resources to stop.
You need to minimize the time it takes to stop the compute resources. The solution must minimize the impact on running transactions.
How should you change the function?
Select only one answer.
Add a timer to wait 20 minutes before stopping the compute resources.
Check the sys.dm_operation_status dynamic management view until no transactions are active in the database before stopping the compute resources.
Close all connections to the database before stopping the compute resources.
Set the database to READ_ONLY before stopping the compute resources.
DistinctRows: Aggregate data by using myCols that produce columns.
source1: Import data from DelimitedText1.
derivedColumn1: Create and update the C1 columns.
select1: Rename derivedColumn1 as select1 with columns C1.
sink1: Add a sink dataset.
You need to ensure that all the rows in source1 are deduplicated.
What should you do?
Select only one answer.
Create a new aggregate task after source1 and copy the script to the aggregate task.
Change the incoming stream for DistinctRows to source1.
Change the incoming stream for derivedColumn1 to DistinctRows.
Create a new flowlet task after source1.
You plan to build a data pipeline that will process real-time data from the devices.
You need to ensure that the devices can send messages to the subscription.
What should you deploy?
Select only one answer.
an Azure event hub
an Azure Stream Analytics workspace
an Azure Storage account
You need to ensure that data in the pipeline is analyzed using fixed five-minute windows.
Which windowing function should you use?
Select only one answer.
TumblingWindow
HoppingWindow
SlidingWindow
SessionWindow
You need to ensure there are three nodes available for the job.
What is the minimum number of SUs you should configure?
Select only one answer.
3
6
18
24
You need to ensure there are three nodes available for the job.
What is the minimum number of SUs you should configure?
Select only one answer.
3
6
18
24
You configure datafacotry1 to use Git for source control.
You make changes to an existing pipeline.
When you try to publish the changes, you notice the following message displayed when you hover over the Publish All button.
Publish from ADF Studio is disabled to avoid overwriting automated deployments. If required you can change publish setting in Git configuration.
You need to allow publishing from the portal.
What should you do?
Select only one answer.
Change the Automated publish config setting.
Select Override live mode in the Git Configuration.
Use a git client to merge the collaboration branch into the live branch.
Use the browser to create a pull request.
You need to set the global parameters that will be used by the deployment.
Which two methods can you use? Each correct answer presents a complete solution.
Select all answers that apply.
an Azure Resource Manager (ARM) template
a PowerShell script
a Python script
Azure Data Studio
You grant developers Read and Write permissions by using ACLs to the files in path \root\input\cleaned\.
The developers report that they cannot open the files.
How should you modify the permissions to ensure that the developers can open the files?
Select only one answer.
Grant Execute permissions to all folders.
Grant Execute permissions to the root folder only.
Add Execute permissions to the files.
Add Contributor permission to the developers.
Which three actions should you perform? Each correct answer presents part of the solution.
Select all answers that apply.
Create a storage account that has a lifecycle policy.
Create a Log Analytics workspace.
Add diagnostic settings and add Log Analytics as a target.
Add diagnostic settings and add Azure Event Hubs as a target.
From the Monitor page of Azure Synapse Studio, review the Pipeline runs tab.
An Azure Synapse Analytics workspace named app1-syn
An Azure Data Lake Storage Gen2 account named app1synstg
A file system named data in app1synstg
You upload a file named Data.parquet to app1synstg.
You need to query the first 100 rows of Data.parquet by using a SQL serverless pool.
Which query should you run?
Select only one answer.
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1synstg.dfs.core.windows.net/data/NYCTripSmall.parquet', FORMAT = 'PARQUET' ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1-syn.dfs.core.windows.net/data/NYCTripSmall.parquet', FORMAT = 'PARQUET' ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1synstg.dfs.core.windows.net/data/NYCTripSmall.parquet', SINGLE_CLOB ) as result
SELECT TOP 100 * FROM OPENROWSET( BULK 'https://app1-syn.dfs.core.windows.net/data/NYCTripSmall.parquet', SINGLE_CLOB ) as result
You need to identify which assets have been cataloged by Microsoft Purview.
What should you use?
Select only one answer.
the Microsoft Purview governance portal
Azure Data Studio
Azure Data Factory
the Microsoft Purview compliance portal
if (N < 0) {
stop("N must be a non-negative integer.")
}
sum <- 0
for(i in 1:N) {
sum<-sum+i
}
and <- sum
return(and)
}