Practice Exam /DP600 – MICROSOFT Fabrics Preparation With 350 Questions
And Correct Answers With Rationales| Already Graded A+
You want to use Apache Spark to explore data interactively in Microsoft Fabric. What
should you create?
a. Spark job definition.
b. Data Factory pipeline.
c. notebook.
c
You need to use Spark to analyze data in a CSV file. What's the simplest way to accomplish
this goal?
a. Load the file into a dataframe.
b. Import the data into a table in a warehouse
c. Convert the data to Parquet format.
a
Which method is used to split the data across folders when saving a dataframe?
a. splitBy
b. distributeBy
c. partitionBy
c
For each table, the lakehouse stores a folder containing Parquet data files and a ... folder in
which transaction details are logged in JSON format.
a. _delta_Log
b. log
c. parquet
a
What is atomicity in ACID transactions?
a. transactions leave the database in a consistent state
b. transactions complete as a single unit of work
c. in-process transactions can't interfere with one another
, 2
d. when a transaction completes, the changes it made are persisted
b
What is consistency in ACID transactions?
a. transactions leave the database in a consistent state
b. transactions complete as a single unit of work
c. in-process transactions can't interfere with one another
d. when a transaction completes, the changes it made are persisted
a
What is durability in ACID transactions?
a. when a transaction completes, the changes it made are persisted
b. transactions complete as a single unit of work
c. in-process transactions can't interfere with one another
d. transactions leave the database in a consistent state
a
What is isolation in ACID transactions?
a. when a transaction completes, the changes it made are persisted
b. transactions complete as a single unit of work
c. in-process transactions can't interfere with one another
d. transactions leave the database in a consistent state
c
Dataframe that's saved as a ... table means that the table definition in the metastore and the
underlying data files are both managed by the Spark runtime for the Fabric lakehouse.
Deleting the table will also delete the underlying files from the Tables storage location for
the lakehouse.
a. managed
b. external
a
, 3
... tables: in which the relational table definition in the metastore is mapped to an
alternative file storage location.
a. managed
b. external
b
Deleting an external table from the lakehouse metastore does not delete the associated data
files.
a. TRUE
b. FALSE
a
When using a delta table as a streaming source, only ... operations can be included in the
stream. Data modifications will cause an error unless you specify the ignoreChanges or
ignoreDeletes option.
a. insert
b. select
c. append
c
The ... option is used to write a checkpoint file that tracks the state of the stream
processing. This file enables you to recover from failure at the point where stream
processing left off.
a. Location
b. checkpointLocation
c. checkpointGeo
d. checkpointGeoLocation
b
You can save the dataframe as a delta table by using the ... method
a. saveAsDelta
b. saveAsTable
c. saveAsDeltaParquet
, 4
a
Which of the following descriptions best fits Delta Lake?
a. Spark API for exporting data from a relational database into CSV files.
b. relational storage layer for Spark that supports tables based on Parquet files.
c. synchronization solution that replicates data between SQL Server and SPark.
b
You've loaded a Spark dataframe with data, that you now want to use in a delta table.
What format should you use to write the dataframe to storage?
a. CSV
b. PARQUET
c. DELTA
c
You have a managed table based on a folder that contains data files in delta format. If you
drop the table, what happens?
a. The table metadata and data files are deleted.
b. The table definition is removed from the metastore, but the data files remain intact.
c. The table definition remains in the metastore, but the data files are deleted.
a
There are two broad categories of activity in a pipeline.
a. Data transformation activities
b. Control flow activities
c. Control transformation activities
a, b
Activities that encapsulate data transfer operations, including simple Copy Data activities
that extract data from a source and load it to a destination, and more complex Data
Flow activities that encapsulate dataflows (Gen2) that apply transformations to the data as
it is transferred. This is called
a. Data transformation activities
b. Control flow activities