Free Microsoft DP-600 Exam Braindumps (page: 6)

You have a Fabric tenant that contains JSON files in OneLake. The files have one billion items. You plan to perform time series analysis of the items.

You need to transform the data, visualize the data to find insights, perform anomaly detection, and share the insights with other business users. The solution must meet the following requirements:
Use parallel processing. Minimize the duplication of data.

Minimize how long it takes to load the data.

What should you use to transform and visualize the data?

  1. the PySpark library in a Fabric notebook
  2. the pandas library in a Fabric notebook
  3. a Microsoft Power BI report that uses core visuals

Answer(s): A

Explanation:

PySpark vs Pandas Performance
Pyspark has been created to help us work with big data on distributed systems. On the other hand, the pandas module is used to manipulate and analyze datasets up to a few GigaBytes (Less than 10 GB to be specific).

So, PySpark, when used with a distributed computing system, gives better performance than pandas. Pyspark also uses resilient distributed datasets (RDDs) to work parallel on the data. Hence, it performs better than pandas.

NOTE: PySpark is a Python library that provides an interface for Apache Spark. Spark is an open-source framework for big data processing. Spark is built to process large amounts of data quickly by distributing computing tasks across a cluster of machines.

PySpark allows us to use Apache Spark and its ecosystem of libraries, such as Spark SQL for working with structured data.

We can also use Spark MLlib for machine learning and GraphX for graph processing using Pyspark in Python.

PySpark supports many data sources, including Hadoop Distributed File System (HDFS), Apache Cassandra, and Amazon S3.

Along with the data processing capabilities, we can also use pyspark with popular Python libraries such as NumPy and Pandas.


Reference:

https://www.codeconquest.com/blog/pyspark-vs-pandas-performance-memory-consumption-and-use-cases



You have a Fabric tenant that contains two workspaces named Workspace1 and Workspace2 and a user named User1.

You need to ensure that User1 can perform the following tasks: Create a new domain.

Create two subdomains named subdomain1 and subdomain2. Assign Workspace1 to subdomain1.

Assign Workspace2 to subdomain2.

The solution must follow the principle of least privilege. Which role should you assign to User1?

  1. domain admin
  2. domain contributor
  3. Fabric admin
  4. workspace Admin

Answer(s): A

Explanation:

To achieve the tasks described, User1 needs permissions to manage domains and assign workspaces to subdomains. Here’s a breakdown of the required tasks and the permissions needed:
Create a new domain:
This requires the ability to manage domains, which is specifically granted by the domain admin role.

Create two subdomains (subdomain1 and subdomain2):
Subdomains fall under domain management, so the domain admin role is required.

Assign workspaces (Workspace1 and Workspace2) to subdomains:
Assigning workspaces to subdomains is also a domain management task, which the domain admin role permits.



HOTSPOT (Drag and Drop is not supported)
You have a Fabric tenant that contains three users named User1, User2, and User3. The tenant contains a security group named Group1. User1 and User3 are members of Group1.

The tenant contains the workspaces shown in the following table.



The tenant contains the domains shown in the following table.



User1 creates a new workspace named Workspace3. You assign Domain1 as the default domain of Group1.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Hot Area:

  1. See Explanation section for answer.

Answer(s): A

Explanation:




User2 is assigned the Contributor role for Workspace3 - No
User2 is not a member of Group1, and Workspace3 is created by User1. Since Workspace3 is assigned to Domain1 (default domain of Group1), only members of Group1 will have permissions based on their role in the domain. User2 is not part of Group1, so they have no role in Workspace3.

User3 is assigned the Viewer role for Workspace3 - No
User3 is a member of Group1, and the default domain (Domain1) is assigned to Group1. However, there is no indication that User3 has been explicitly granted the Viewer role in Workspace3. If permissions were inherited, User3 would have the default role for Domain1, but the problem does not specify this explicitly, so
we assume no Viewer role is assigned.

User3 is assigned the Contributor role for Workspace1 - No
Workspace1 is explicitly assigned to User1 as the admin. There is no indication that User3 has any permissions for Workspace1. Being a member of Group1 does not grant automatic Contributor access to a workspace unless explicitly configured.



You have a Fabric warehouse named Warehouse1 that contains a table named Table1. Table1 contains customer data.

You need to implement row-level security (RLS) for Table1. The solution must ensure that users can see only their respective data.

Which two objects should you create? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

  1. DATABASE ROLE
  2. STORED PROCEDURE
  3. CONSTRAINT
  4. FUNCTION
  5. SECURITY POLICY

Answer(s): A,E

Explanation:

A database role is used to assign permissions to users or groups. In the context of RLS, you create roles that map to specific user groups or individuals, determining which rows they can access.

A security policy is used to enforce row-level security. This is done by creating a filter predicate that limits the rows returned based on a condition, such as the user's identity or a specific column value.



You are the administrator of a Fabric workspace that contains a lakehouse named Lakehouse1. Lakehouse1 contains the following tables:
Table1: A Delta table created by using a shortcut Table2: An external table created by using Spark Table3: A managed table
You plan to connect to Lakehouse1 by using its SQL endpoint. What will you be able to do after connecting to Lakehouse1?

  1. Read Table3.
  2. Update the data Table3.
  3. Read Table2.
  4. Update the data in Table1.

Answer(s): A



Viewing page 6 of 41
Viewing questions 21 - 24 out of 198 questions



Post your Comments and Discuss Microsoft DP-600 exam prep with other Community members:

DP-600 Exam Discussions & Posts