Winter Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

ARA-C01 Exam Dumps - SnowPro Advanced: Architect Certification Exam

Go to page:
Question # 9

An Architect needs to design a data unloading strategy for Snowflake, that will be used with the COPY INTO command.

Which configuration is valid?

A.

Location of files: Snowflake internal location

. File formats: CSV, XML

. File encoding: UTF-8

. Encryption: 128-bit

B.

Location of files: Amazon S3

. File formats: CSV, JSON

. File encoding: Latin-1 (ISO-8859)

. Encryption: 128-bit

C.

Location of files: Google Cloud Storage

. File formats: Parquet

. File encoding: UTF-8

· Compression: gzip

D.

Location of files: Azure ADLS

. File formats: JSON, XML, Avro, Parquet, ORC

. Compression: bzip2

. Encryption: User-supplied key

Full Access
Question # 10

Based on the architecture in the image, how can the data from DB1 be copied into TBL2? (Select TWO).

A)

B)

C)

D)

E)

A.

Option A

B.

Option B

C.

Option C

D.

Option D

E.

Option E

Full Access
Question # 11

A company’s daily Snowflake workload consists of a huge number of concurrent queries triggered between 9pm and 11pm. At the individual level, these queries are smaller statements that get completed within a short time period.

What configuration can the company’s Architect implement to enhance the performance of this workload? (Choose two.)

A.

Enable a multi-clustered virtual warehouse in maximized mode during the workload duration.

B.

Set the MAX_CONCURRENCY_LEVEL to a higher value than its default value of 8 at the virtual warehouse level.

C.

Increase the size of the virtual warehouse to size X-Large.

D.

Reduce the amount of data that is being processed through this workload.

E.

Set the connection timeout to a higher value than its default.

Full Access
Question # 12

Which of the following ingestion methods can be used to load near real-time data by using the messaging services provided by a cloud provider?

A.

Snowflake Connector for Kafka

B.

Snowflake streams

C.

Snowpipe

D.

Spark

Full Access
Question # 13

Which Snowflake architecture recommendation needs multiple Snowflake accounts for implementation?

A.

Enable a disaster recovery strategy across multiple cloud providers.

B.

Create external stages pointing to cloud providers and regions other than the region hosting the Snowflake account.

C.

Enable zero-copy cloning among the development, test, and production environments.

D.

Enable separation of the development, test, and production environments.

Full Access
Question # 14

A Snowflake Architect is designing an application and tenancy strategy for an organization where strong legal isolation rules as well as multi-tenancy are requirements.

Which approach will meet these requirements if Role-Based Access Policies (RBAC) is a viable option for isolating tenants?

A.

Create accounts for each tenant in the Snowflake organization.

B.

Create an object for each tenant strategy if row level security is viable for isolating tenants.

C.

Create an object for each tenant strategy if row level security is not viable for isolating tenants.

D.

Create a multi-tenant table strategy if row level security is not viable for isolating tenants.

Full Access
Question # 15

What are characteristics of the use of transactions in Snowflake? (Select TWO).

A.

Explicit transactions can contain DDL, DML, and query statements.

B.

The autocommit setting can be changed inside a stored procedure.

C.

A transaction can be started explicitly by executing a begin work statement and end explicitly by executing a commit work statement.

D.

A transaction can be started explicitly by executing a begin transaction statement and end explicitly by executing an end transaction statement.

E.

Explicit transactions should contain only DML statements and query statements. All DDL statements implicitly commit active transactions.

Full Access
Question # 16

A company is using Snowflake in Azure in the Netherlands. The company analyst team also has data in JSON format that is stored in an Amazon S3 bucket in the AWS Singapore region that the team wants to analyze.

The Architect has been given the following requirements:

1. Provide access to frequently changing data

2. Keep egress costs to a minimum

3. Maintain low latency

How can these requirements be met with the LEAST amount of operational overhead?

A.

Use a materialized view on top of an external table against the S3 bucket in AWS Singapore.

B.

Use an external table against the S3 bucket in AWS Singapore and copy the data into transient tables.

C.

Copy the data between providers from S3 to Azure Blob storage to collocate, then use Snowpipe for data ingestion.

D.

Use AWS Transfer Family to replicate data between the S3 bucket in AWS Singapore and an Azure Netherlands Blob storage, then use an external table against the Blob storage.

Full Access
Go to page: