Weekend Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70percent

Snowflake DSA-C02 SnowPro Advanced: Data Scientist Certification Exam Exam Practice Test

Demo: 17 questions
Total 65 questions

SnowPro Advanced: Data Scientist Certification Exam Questions and Answers

Question 1

Which ones are the correct rules while using a data science model created via External function in Snowflake?

Options:

A.

External functions return a value. The returned value can be a compound value, such as a VARIANT that contains JSON.

B.

External functions can be overloaded.

C.

An external function can appear in any clause of a SQL statement in which other types of UDF can appear.

D.

External functions can accept Model parameters.

Question 2

Which type of Python UDFs let you define Python functions that receive batches of input rows as Pandas DataFrames and return batches of results as Pandas arrays or Series?

Options:

A.

MPP Python UDFs

B.

Scaler Python UDFs

C.

Vectorized Python UDFs

D.

Hybrid Python UDFs

Question 3

What is the formula for measuring skewness in a dataset?

Options:

A.

MEAN - MEDIAN

B.

MODE - MEDIAN

C.

(3(MEAN - MEDIAN))/ STANDARD DEVIATION

D.

(MEAN - MODE)/ STANDARD DEVIATION

Question 4

Which one is the incorrect option to share data in Snowflake?

Options:

A.

a Listing, in which you offer a share and additional metadata as a data product to one or more accounts.

B.

a Direct Marketplace, in which you directly share specific database objects (a share) to another account in your region using Snowflake Marketplace.

C.

a Direct Share, in which you directly share specific database objects (a share) to anoth-er account in your region.

D.

a Data Exchange, in which you set up and manage a group of accounts and offer a share to that group.

Question 5

The most widely used metrics and tools to assess a classification model are:

Options:

A.

Confusion matrix

B.

Cost-sensitive accuracy

C.

Area under the ROC curve

D.

All of the above

Question 6

Which of the learning methodology applies conditional probability of all the variables with respec-tive the dependent variable?

Options:

A.

Reinforcement learning

B.

Unsupervised learning

C.

Artificial learning

D.

Supervised learning

Question 7

Which of the following is a useful tool for gaining insights into the relationship between features and predictions?

Options:

A.

numpy plots

B.

sklearn plots

C.

Partial dependence plots(PDP)

D.

FULL dependence plots (FDP)

Question 8

Which one of the following is not the key component while designing External functions within Snowflake?

Options:

A.

Remote Service

B.

API Integration

C.

UDF Service

D.

Proxy Service

Question 9

Which of the following Snowflake parameter can be used to Automatically Suspend Tasks which are running Data science pipelines after specified Failed Runs?

Options:

A.

SUSPEND_TASK

B.

SUSPEND_TASK_AUTO_NUM_FAILURES

C.

SUSPEND_TASK_AFTER_NUM_FAILURES

D.

There is none as such available.

Question 10

Which of the following cross validation versions may not be suitable for very large datasets with hundreds of thousands of samples?

Options:

A.

k-fold cross-validation

B.

Leave-one-out cross-validation

C.

Holdout method

D.

All of the above

Question 11

What is the risk with tuning hyper-parameters using a test dataset?

Options:

A.

Model will overfit the test set

B.

Model will underfit the test set

C.

Model will overfit the training set

D.

Model will perform balanced

Question 12

Select the Correct Statements regarding Normalization?

Options:

A.

Normalization technique uses minimum and max values for scaling of model.

B.

Normalization technique uses mean and standard deviation for scaling of model.

C.

Scikit-Learn provides a transformer RecommendedScaler for Normalization.

D.

Normalization got affected by outliers.

Question 13

Mark the correct steps for saving the contents of a DataFrame to aSnowflake table as part of Moving Data from Spark to Snowflake?

Options:

A.

Step 1.Use the PUT() method of the DataFrame to construct a DataFrameWriter.

Step 2.Specify SNOWFLAKE_SOURCE_NAME using the NAME() method.

Step 3.Use the dbtable option to specify the table to which data is written.

Step 4.Specify the connector options using either the option() or options() method.

Step 5.Use the save() method to specify the save mode for the content.

B.

Step 1.Use the PUT() method of the DataFrame to construct a DataFrameWriter.

Step 2.Specify SNOWFLAKE_SOURCE_NAME using the format() method.

Step 3.Specify the connector options using either the option() or options() method.

Step 4.Use the dbtable option to specify the table to which data is written.

Step 5.Use the save() method to specify the save mode for the content.

C.

Step 1.Use the write() method of the DataFrame to construct a DataFrameWriter.

Step 2.Specify SNOWFLAKE_SOURCE_NAME using the format() method.

Step 3.Specify the connector options using either the option() or options() method.

Step 4.Use the dbtable option to specify the table to which data is written.

Step 5.Use the mode() method to specify the save mode for the content.

(Correct)

D.

Step 1.Use the writer() method of the DataFrame to construct a DataFrameWriter.

Step 2.Specify SNOWFLAKE_SOURCE_NAME using the format() method.

Step 3.Use the dbtable option to specify the table to which data is written.

Step 4.Specify the connector options using either the option() or options() method.

Step 5.Use the save() method to specify the save mode for the content.

Question 14

Which command is used to install Jupyter Notebook?

Options:

A.

pip install jupyter

B.

pip install notebook

C.

pip install jupyter-notebook

D.

pip install nbconvert

Question 15

Mark the incorrect statement regarding usage of Snowflake Stream & Tasks?

Options:

A.

Snowflake automatically resizes and scales the compute resources for serverless tasks.

B.

Snowflake ensures only one instance of a task with a schedule (i.e. a standalone task or the root task in a DAG) is executed at a given time. If a task is still running when the next scheduled execution time occurs, then that scheduled time is skipped.

C.

Streams support repeatable read isolation.

D.

An standard-only stream tracks row inserts only.

Question 16

Which ones are the type of visualization used for Data exploration in Data Science?

Options:

A.

Heat Maps

B.

Newton AI

C.

Feature Distribution by Class

D.

2D-Density Plots

E.

Sand Visualization

Question 17

Data Scientist used streams in ELT (extract, load, transform) processes where new data inserted in-to a staging table is tracked by a stream. A set of SQL statements transform and insert the stream contents into a set of production tables. Raw data is coming in the JSON format, but for analysis he needs to transform it into relational columns in the production tables. which of the following Data transformation SQL function he can used to achieve the same?

Options:

A.

He could not apply Transformation on Stream table data.

B.

lateral flatten()

C.

METADATA$ACTION ()

D.

Transpose()

Demo: 17 questions
Total 65 questions