DP-900 Latest QUESTIONS with Correct
Verified Solutions
Four common types of non-relational databases
Key-value, document, column-family, graph
Which type of processing system is used to support live and line-of-business applications?
OLTP (online transactional processing)
Data warehouse
A relational database that is optimized for read operations
Describe the four steps of a common architecture for an OLAP system
1. Data files may be stored in a central data lake for analysis.
2. An extract, transform, and load (ETL) process copies data from files and OLTP databases into
a data warehouse that is optimized for read activity. Commonly, a data warehouse schema is
based on fact tables that contain numeric values you want to analyze (for example, sales
amounts), with related dimension tables that represent the entities by which you want to measure
them (for example, customer or product),
3. Data in the data warehouse may be aggregated and loaded into an online analytical processing
(OLAP) model, or cube. Aggregated numeric values (measures) from fact tables are calculated
for intersections of dimensions from dimension tables. For example, sales revenue might be
totaled by date, customer, and product.
4. The data in the data lake, data warehouse, and analytical model can be queried to produce
reports, visualizations, and dashboards.
,Between an OLAP and OLTP system, which of the two would be denormalized?
OLAP. Denormalization would allow for duplicate values, but would also make queries perform
faster.
An OLTP system would need to be normalized because it uses the least amount of storage and
optimizes for updates, inserts, and deletes.
Which data integration service allows you to orchestrate data flow without coding?
Data Factory
Which two services allow you to pre-process a large volume of data by using Scala?
a serverless Apache Spark pool in Azure Synapse Analytics
and Databricks
Which service allows you to perform near real-time analytics on the operational data
stored in Azure Cosmos DB?
Azure Synapse
True or False:
Batch processing is used to handle complex analysis.
True!
Which service allows you to perform on-demand analysis of large volumes of data from
text logs, websites and IoT devices by using a common querying language for all the data
sources?
, Data Explorer is used for the analysis of large amounts of text log data, websites, and IoT
devices and uses a common querying language. Data Lake Storage Gen2 is a data source, Azure
Stream Analytics is used to define streaming jobs, apply a perpetual query, and write the results
to an output. Azure Cosmos DB stores data.
What should you create in a data model to allow users to drill up and drill down in a
report?
A hierarchy enables drill up and drill down in a dimension. A dimension enables navigation, but
hierarchy is used to drill up and down in a dimension. Fact tables have values. Cubes are not
created in Microsoft PowerBI.
Which type of data structure should you use to optimize create, read, update, and delete
(CRUD) operations for data saved in a multi-column tabular format?
A relational database is the best option for CRUD operations and uses the least amount of
storage space. A key/value store is used for simple lookups based on a single key to obtain a
single value. A document database uses unstructured data such as JSON, and is optimized for
retrieval, not CRUD operations. A graph database is used to store hierarchical data, such as
organizational charts that have nodes and edges.
Which type of database should you use to store sequential data in the fastest way possible?
time series
Which two types of file store data in columnar format?
Parquet and ORC
Which two attributes are characteristics of a transactional data workload?
Verified Solutions
Four common types of non-relational databases
Key-value, document, column-family, graph
Which type of processing system is used to support live and line-of-business applications?
OLTP (online transactional processing)
Data warehouse
A relational database that is optimized for read operations
Describe the four steps of a common architecture for an OLAP system
1. Data files may be stored in a central data lake for analysis.
2. An extract, transform, and load (ETL) process copies data from files and OLTP databases into
a data warehouse that is optimized for read activity. Commonly, a data warehouse schema is
based on fact tables that contain numeric values you want to analyze (for example, sales
amounts), with related dimension tables that represent the entities by which you want to measure
them (for example, customer or product),
3. Data in the data warehouse may be aggregated and loaded into an online analytical processing
(OLAP) model, or cube. Aggregated numeric values (measures) from fact tables are calculated
for intersections of dimensions from dimension tables. For example, sales revenue might be
totaled by date, customer, and product.
4. The data in the data lake, data warehouse, and analytical model can be queried to produce
reports, visualizations, and dashboards.
,Between an OLAP and OLTP system, which of the two would be denormalized?
OLAP. Denormalization would allow for duplicate values, but would also make queries perform
faster.
An OLTP system would need to be normalized because it uses the least amount of storage and
optimizes for updates, inserts, and deletes.
Which data integration service allows you to orchestrate data flow without coding?
Data Factory
Which two services allow you to pre-process a large volume of data by using Scala?
a serverless Apache Spark pool in Azure Synapse Analytics
and Databricks
Which service allows you to perform near real-time analytics on the operational data
stored in Azure Cosmos DB?
Azure Synapse
True or False:
Batch processing is used to handle complex analysis.
True!
Which service allows you to perform on-demand analysis of large volumes of data from
text logs, websites and IoT devices by using a common querying language for all the data
sources?
, Data Explorer is used for the analysis of large amounts of text log data, websites, and IoT
devices and uses a common querying language. Data Lake Storage Gen2 is a data source, Azure
Stream Analytics is used to define streaming jobs, apply a perpetual query, and write the results
to an output. Azure Cosmos DB stores data.
What should you create in a data model to allow users to drill up and drill down in a
report?
A hierarchy enables drill up and drill down in a dimension. A dimension enables navigation, but
hierarchy is used to drill up and down in a dimension. Fact tables have values. Cubes are not
created in Microsoft PowerBI.
Which type of data structure should you use to optimize create, read, update, and delete
(CRUD) operations for data saved in a multi-column tabular format?
A relational database is the best option for CRUD operations and uses the least amount of
storage space. A key/value store is used for simple lookups based on a single key to obtain a
single value. A document database uses unstructured data such as JSON, and is optimized for
retrieval, not CRUD operations. A graph database is used to store hierarchical data, such as
organizational charts that have nodes and edges.
Which type of database should you use to store sequential data in the fastest way possible?
time series
Which two types of file store data in columnar format?
Parquet and ORC
Which two attributes are characteristics of a transactional data workload?