Winter Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

H13-723_V2.0 Exam Dumps - HCIP-Big Data Developer V2.0

Question # 4

FusionInsight HDin, useStreamingofLinuxWhen submitting a topology in command line mode, you need to

first use a possessedStreamingUser with submit permissionkinitway authentication

A.

True

B.

False

Full Access
Question # 5

forHBase rowkeyThe design principles described below are correct? (multiple choice)

A.

Attribute values with high port access weights are placed inRowKeyFront

B.

Attribute values with high port access weights are placed inRowkeysecond half

C.

The attribute value with good dispersion is placed inRowkeyfirst half

D.

The attribute value with good dispersion is placed inRowkeysecond half

Full Access
Question # 6

LoaderIf the job execution fails, the data imported during the running of this job will not be deleted automatically.

must be deleted manually.

A.

True

B.

False

Full Access
Question # 7

FusionInsightHDin, aboutHivepartition (partition) function, which is wrong as described below?

A.

Partition fields are defined when the table is created

B.

There can only be one partition field, and multi-level partitions cannot be created

C.

Using partitions can reduce the data scan range of some queries, thereby improving query efficiency

D.

Partition fields can be used aswherecondition of the sentence

Full Access
Question # 8

aboutStreamingthe topology (Topology), which of the following descriptions is wrong?

A.

OneTopologyis made up of a group ofSpoutcomponents andBotcomponents passStream GroupingsA connected directed acyclic graph (DAG)

B.

Topologywill run until it is explicitlykill

C.

Business logic is encapsulated inTopologymiddle

D.

OneTopologyOnly one can be specified to startWorkerprocess

Full Access
Question # 9

TowardsHBaseincrease in the clusterRegionServerhost, the original cluster must be stopped first, becauseHBase

Dynamic expansion is not supported.

A.

True

B.

False

Full Access
Question # 10

aboutHiveandHadoopThe relationship of other components, the following description is wrong?

A.

Hivefinally store the data inHDFSmiddle

B.

HQLable to passMapReduceperform tasks

C.

HiveYesHadoopThe platform's data warehouse tools

D.

HiverightHBasehave strong dependencies

Full Access
Question # 11

aboutKafkaThe characteristics of the following description are correct? (multiple choice)

A.

KafkaIt is a high-throughput, distributed, publish-subscribe-based messaging system

B.

KafkaPersistence of messages

C.

KafkaApplicable to offline and online message consumption scenarios

D.

Kafkaguarantee eachPartitionmessages in order

Full Access
Question # 12

FusionInsightHD which components are provided externallySQLor classSQLability?

(multiple choice)

A.

HBase

B.

Hive

C.

Spark

D.

Redis

Full Access
Question # 13

existMapReduceDuring application development,setMapOutputCompressorClassWhat is the role of classes?

A.

set upmapinput compression method

B.

set upmapHow the output is delivered

C.

set upmapoutput compression method

D.

set upreduceoutput compression method

Full Access
Question # 14

Below aboutZooKeeperThe description of the feature is wrong ()?

A.

Updates sent by the client are applied in the order in which they were sent.

B.

A message to be read by more than halfserverreceive, it will be able to successfully write to disk.

C.

Message updates can only succeed or fail, with no intermediate states.

D.

ZooKeeperThe number of nodes must be an odd number.

Full Access
Question # 15

due toSparkis a memory-based computing engine, therefore, aSparkThe amount of data the app can handle Can't give more than thisSparkThe total memory of the application.

A.

True

B.

False

Full Access
Question # 16

existSparkIn application development, which of the following codes can correctly count words?

A.

val counts=textFile.map(line=>line.split(" ')).map(-rd=>(word,1)).reduceByKey(_+_)

B.

val counts=textFile.flatMap(line=>line.split(" ")).map(word=>(word,1)).reduceByKey(_+_)

C.

val counts=textFile.map(line=>line.split(" ")).map(word=>(word,1)).groupByKey()

D.

val counts=textFile.flatmap(line=>line.split(" ")).map(-rd=>(word,1)).reduceByKey()

Full Access
Question # 17
A.

JDB, Cinterface

B.

ODB, Cinterface

C.

Pythoninterface

D.

Rubyinterface

Full Access
Question # 18

FusionInsight HDin, belonging toStreamingWhat are the roles of the service? (multiple choice)

A.

Nimbus

B.

Supervisor

C.

Broker

D.

quorumpeer

Full Access
Question # 19

When a task is submitted toYARNIn the cluster, which of the following options will generate for this taskTindivual JOPO

A.

ResourceManager

B.

container

C.

AppMaster

D.

NodeManager

Full Access
Question # 20

Suppose there is an application that needs to be accessed frequentlyOracleThe user table in the database, in order to improve performance, introduceRedisto cache

account information.

For this scene,RedisWhich of the following is the best data structure choice for ?

A.

list

B.

set

C.

hash

D.

sorted set

Full Access
Question # 21

HDFSIt adopts a "write once, read many" file access model. So it is recommended that a file be created, written and

After closing, do not modify it again.

A.

True

B.

False

Full Access
Question # 22

FusionInsight HDWhat are the supported ways to submitOozieOperation? (multiple choice)

A.

ShellOrder

B.

Java API

C.

Hue UI

D.

Rest API

Full Access
Question # 23

There are the following business scenarios: User online log files have been stored inHDFSabove, the log file content format the formula is: each online record has three fields, namely name, gender, and online time, and the fields are separated by ",";

It is required to print out all female netizens who spend more than two hours online. Which of the following code snippets can achieve

The above business scenario? (multiple choice)

A.

sc.textFile("/data/file/path”).map(.split(",")).map(p=>FemaleInfo(p

(O),p(1),p(2).trim.toInt)).toDF.registerTemp Table("FemaleinfoTable”)sqlContext.sql

("select name,sum(stay Time)as stay Time from FemaleInfoTable where gender=

female'group by name”).filter("stay Time >=120”).collect).foreach(printin)

B.

sc.textFile("/data/file/path").map(.split(",")).map(p=>FemaleInfo

(p(O),p(1),p(2).trim.toInt)).toDF.registerTemp Table("Female infoTable")

sqlContext.sq(l "select name, sum(stay Time)as stay Time from FemaleInfoTable where

gender=female).filter(“stay Time >=120”).collect).foreach(printin)

C.

val text=sc.textFile("/data/fle/path")

val data=text.filter(_.contains("female"))

Val fermaleData:RDD(String,Int) =data.map{line=>val t=line.split(',')

(t(0),t(2).toInt)

}.reduceByKey(_+_)

Val result=femaleData.filter(line=>line.2>120)

Rusult.collect.map(x=>x._1 + ',' + x._2).foreach(println)

D.

val text=sc.textFile("/data/fle/path”)

val data=text.filter(_.contains("female”))

Val fermaleData:RDD(String,Int) =data.map(line=>val t=line.split(',')

(t(0),t(2).toInt)

}

Val result=femaleData.filter(line=>line.2>120)

Rusult.collect.map(x=>x._1 +','+ x._2).foreach(println)

Full Access
Question # 24

FusionInsightHDin the system,HiveWhat are the supported interface types? (multiple choice)

A.

mouthJDB, Cinterface

B.

mouthODB, Cinterface

C.

mouthPythoninterface

D.

mouthRubyinterface

Full Access
Question # 25

RedisIn cluster mode, useJAVAWhich of the following is a best practice for application development?

A.

useJedisobject connectionRedis Server,transferJedisofset.qetWaiting for business operation interface to process data operate.

B.

useJedisPoolobject createdJedisObject pool, each business operation starts fromJedisPoolget one Jedisobject, then call,Jedisofset.getWait for the business operation interface to perform data operations.

C.

useredisClusterobject connectionRedis Server,transferJedisClusterofset.getother business operations interface for data manipulation; in multithreaded applications, multiple threads share oneJedisClusterobject.

D.

useJedisClusterobject connectionRedis Server,transferJedisClusterofset,getother business operations interfacet.

In multithreaded applications, each thread creates aJedisClusterobject.

Full Access
Question # 26

TowardsHBaseincrease in the clusterRegionServerhost, the original cluster must be stopped first, becauseHBaseDynamic expansion is not supported.

A.

True

B.

False

Full Access
Question # 27

Fusionmnsight HDproductSoDuring application development, you canSolr Admin UIxrightCollection

One-shot verification is performed. Draw aboutSolr AdminUIIs the statement correct? (multiple choice)

A.

Click Tree under Cloud to view metadata such as Collections, configuration sets, and ive nodes in Solr loud information

B.

Click Cloud to view the Replica distribution and status of each Shard under each Collection

C.

Enter the Core Oveview interface where Calectiopeyshard Replicaf is installed, and you can view the Replca index Real Qunwen You Department relies on storing size and location information

D.

Users of the Sor user group, Solr admin role, and Super group can access the Solr Admin UI

Full Access
Question # 28

Which of the following components must depend onZooKeeperto run()?

(multiple choice)

A.

HDFS

B.

HBase

C.

Spark

D.

YARN(streaming)

Full Access
Question # 29

aboutFusionInsight HDplatformHiveservice, itsWebHCatDevelopment interface, the following description does not

the correct one is?

A.

Support based onRESTquery request

B.

WebHCatThe return data format isXML (Json)

C.

WebHCatbased onHTTPandHTTPS

D.

Agreement to provide services to the outside world able to passWebHCatCreate tables, query, etc.

Full Access
Question # 30

FusionInsight HDin, aboutLoaderIs the job description correct?

A.

LoaderSubmit the job toARNAfter execution, if at this timeLoaderIf the service is abnormal, this job executes line failed.

B.

LoaderSubmit the job toYARAfter a small execution, if a certainMapperTask execution fails, can be automatically Try again.

C.

LoaderIf the job fails to execute, garbage data will be generated, which needs to be cleared manually by the user.

D.

LoderSubmit an assignment toYARAfter the job is executed, no other jobs can be submitted until the job is executed Industry.

Full Access
Question # 31

Solris a high-performance, basedLucenefull-text search service.SolrrightLuceneexpanded,

provides a ratioLuceneA richer query language and a powerful full-text search function are implemented, with a high degree of reliability.

Extensibility. At the same time fromSolr 4.0Version starts, supportsSolrCloudmodel.

A.

True

B.

False

Full Access
Question # 32

Oozieofworkflow.xmlfile is the workflow configuration file

A.

True

B.

False

Full Access
Question # 33

coordinator.xmlis the configuration file responsible for scheduling the workflow

A.

True

B.

False

Full Access
Question # 34

The user wants to develop an application to accessHBaseservice, first apply for a machine accounttester1, and give it sets accessHBaseservice permissions; then download the account'skeytabfile, and use that account to run the application

use. Subsequent applications of this application will increase business requirements and require access toHiveservice, then givingtester1User additionally granted access

askHiveAfter the component permissions, you need to re-download the user'skeytabdocument.

A.

True

B.

False

Full Access
Question # 35

FusionInsight HD a completeStreaming CQLThe application contains at least which of the following part? (multiple choice)

A.

Create an input stream

B.

Create output stream

C.

Business action statements (such asSELECT,INSERT,JIONWait)

D.

Apply commit statement

Full Access
Question # 36

FlinkusecheckpointThe mechanism guarantees fault tolerance in the operation of the application.

A.

True

B.

False

Full Access
Question # 37

FusionInsight HD assuming a topology that setsspoutConcurrency is3,bolt1Concurrency

for2,bok2Well degree is3.workerThe number is2,Sobolt1ofexecutorexistworkerhow to divide Cloth?

A.

twoworkerone of theworkerdistribute two

B.

twoworkerone for each

C.

no fixed rules

D.

twoworkerBoth are distributed

Full Access
Question # 38

HDFSRuntime,(NameNodeAt startup,)NameNodewill file system metadata from

The disk is all loaded into memory, so the total number of files that the file system can store is limited byNameNodeof memory

capacity.

A.

True

B.

False

Full Access
Question # 39

existFusionInsight HDThe key to the product38-by5Taste+Taste4Tian QiyaSWhen the armor is developed, you can pass the

passSolr Admin UIxrightCollectionDo some verification. Below (multiple choice)

A.

Click Tree under Cloud to view cafrCollections, configuration sets and ive nodes in Solr Cloud and other metadata information

B.

Click cloud to view the distribution and status of the replicas of each shard under the backup callion

C.

Enter the Collectionyshard Replicats that you and Confidence Check 1dKepca's (ore-Oevew Open, you can view the actual number of documents submitted for the Relica index

D.

The solr user group, solr, admin, and .Rrg can be used by the sr public and the .Rrg

Full Access
Question # 40

TowardsHBaseincrease in the clusterRegionServerhost, the original cluster must be stopped first, becauseHBase

Dynamic expansion is not supported.

A.

True

B.

False

Full Access
Question # 41

FSpark Streamingapplication, in aJVM, there can only be one at a time

StreamingContextin active state

A.

True

B.

False

Full Access
Question # 42

FusionInsigt HD which of the following components can be used for data collection? (multiple choice)

A.

mouthFlume

B.

mouthLoader

C.

mouthHBase

D.

mouthMapReduce

Full Access
Question # 43

FusionInsight HDsystem, aboutHiveofJDB, CInterface type, which of the following descriptions is correct?

A.

In a network isolation environment,ZooKeeperofIPCan be configured as a businessIP, can also be configured to manageIP.

B.

HiveUnlike traditional databases, soJDB, CThe established connection does not need to be closed.

C.

In a secure cluster environment, establish andHiveBefore connecting, you need to use unified authentication to log in.

D.

useJDB, CsubmitSQL, you canSQLA semicolon ";" is added after the statement.

Full Access
Question # 44

existFlumemiddle,sourceWhat is the main function of the function module?

A.

Get data and convert raw data into data objects that you process yourself

B.

Cache data and save data in memory or files according to different reliability policies

C.

Output data to the destination, support multiple output protocols

D.

Split the data and send the data to different destinations according to the characteristics of the data

Full Access
Question # 45

Spark streamingAs a micro-batch stream processing engine, it converts the data of each time slice

for oneRDDa partition in the

A.

True

B.

False

Full Access
Question # 46

existHBaseIn table design, it is common to multiple businessesrowkey, what are the possible countermeasures for situations that lead to data access contradictions?

slightly? (multiple choice)

A.

compromise method (rowkeyDesign takes into accountAandBbusiness needs)

B.

Redundancy method (use space for time factory efficiency)

C.

Create secondary index

D.

pre-scoreregion

Full Access
Question # 47

Which of the following is notMapReducespecialty?

A.

easy to program

B.

good scalability

C.

real-time computing

D.

High fault tolerance

Full Access
Question # 48

when aMapReduceWhen the application is executed, which of the following actions ismapoccurred before the stage?

A.

split

B.

combine

C.

partition

D.

sort

Full Access
Question # 49

aboutKafkaThe characteristics of the following description are correct? (multiple choice)

A.

KafkaIt is a high-throughput, distributed, publish-subscribe-based messaging system

B.

KafkaPersistence of messages

C.

KafkaApplicable to offline and online message consumption scenarios

D.

Kafkaguarantee eachPartitionmessages in order

Full Access
Question # 50

Below aboutZooKeeperThe statement is wrong ().

A.

If ZooKeeperIf there is an interruption in the process of synchronizing the message, after the failure is recovered, the transmission status before the failure can be adjusted according to the the state continues to synchronize, that is, support for resuming transmission from a breakpoint.

B.

ZooKeeperUse a custom atomic message protocol to ensure the consistency of node data in the entire system.

C.

ZooKeeperThe cluster is elected at startupLeaderRole.

D.

LeaderAfter a node receives a data change request, it first writes to disk and then writes to memory.

Full Access
Question # 51

HDFSIt adopts a "write once, read many" file access model. Therefore, it is recommended that a document

After creating and writing the profit, do not modify it again.

A.

True

B.

False

Full Access
Question # 52

For multiple business sharingrowkey, What are the countermeasures for situations that lead to conflicting data access?

(multiple choice)

A.

Mouth compromise (rowkeyDesign takes into accountAandBbusiness needs)

B.

Port redundancy method (exchange space for time/effectiveness)

C.

Create a secondary index

D.

mouth pre-scoreregion

Full Access
Question # 53

Does a project require Internet access to a certain area? ? Save and search the full text of these Internet records? ? information, with

to prevent crime in the region.

In this scenario, which of the following options is the best?

A.

existSolrCreate an index and save the data, and return all the data during full-text search.

B.

existHBasestore data on theHBaseThe filtering characteristics of , satisfy fuzzy matching query.

C.

When storing data, inSolrindex on theHBaseStore complete data? ? when, throughSolrFull-text search to obtain and record

record key information, through the key information inHBasefor the full record.

Full Access
Question # 54

existFusionInsight HDin, useSparkSQL, which of the following methods (or tools) can be used to performSQLstatement?

(multiple choice)

A.

JDB, C

B.

spark-beeline

C.

spark-sql

Full Access
Question # 55

existFusionInsight HDmiddle,SparkBy defaultYARNAs a cluster resource management system.

A.

True

B.

False

Full Access
Question # 56

existSpark, which of the following statements about broadcast variables is correct?

A.

broadcast variable in everyTasksave a copy

B.

broadcast variable in eachExecutorsave a copy

C.

Broadcast variables are read-only

D.

broadcast variable viaSparkContext.broadcast() method to generate

Full Access
Question # 57

there are the following scenarios: new data is generated by the online system every day500G, it is necessary to analyze these data by day, week,

Monthly and other dimensions for statistical summary. ask if it is suitable for useHiveWhat kind of table to handle?

A.

Bucket table, bucket according to day

B.

Partition table, partitioned by day

C.

skew table, skew field select date, skew value random

D.

partition+Bucket table, partitioned by time, each partition2560buckets

Full Access
Question # 58

aboutStreamingthe topology (Topology), which of the following descriptions is wrong?

A.

OneTopologyis made up of a group ofSpoutcomponents andBoltcomponents passStream GroupingsA connected directed acyclic graph

(DAG)

B.

Topologywill run until it is explicitlykill

C.

Business logic is encapsulatedTopologymiddle

D.

OneTopologyOnly one can be specified to startWorkerprocess

Full Access
Question # 59

existFusionInsight HDcluster, aboutkinitOperation command, which of the following statements is wrong? (multiple choice)

A.

Only the HMI account can be used.

B.

Only the machine account can be used.

C.

A client does not support the simultaneous use of multiple accounts.

D.

The ticket obtained by executing this command is intwenty fourIt will time out when it is small and needs to be executed againkinitcommand to log in again.

Full Access
Question # 60

FusionInsight HDofHive, user-definedUDFcan andHiveBuilt-inUDFduplicate name, in this case,

will use user-definedUDF.

A.

True

B.

False

Full Access
Question # 61

existSpark, assuminglinesIs anDStreamobject,filterStatements can be filtered out80%data for the following two

The correct statement is:

X: lines.filter(…).groupByKey(…)

Y: lines.groupByKey(…).filter(…)

A.

XCompareYhigher performance

B.

XCompareYlower performance

C.

XandYsame as performance

D.

UncertaintyXandYdifference in performance

Full Access
Question # 62

HDFSRuntime,NameNodewill load all the metadata of the file system from disk into memory, so the file system can

The total number of files stored is limited byNameNodememory capacity.

A.

True

B.

False

Full Access
Question # 63

FusionInsight HD a completeStreaming CQLWhich of the following parts does the application contain at least? (multiple choice)

A.

Create an input stream

B.

Create output stream

C.

Business action statements (such asSELECT,INSERT,JIONWait)

D.

Apply commit statement

Full Access
Question # 64

existFusionInsight HDin the cluster,FlumeWhich service does not support writing collected data to the cluster?

A.

HDFS

B.

HBase

C.

Kafka

D.

Redis

Full Access
Question # 65

aboutStreamingdisaster recovery capability, which of the following statements is correct? (multiple choice)

A.

SupervisorAfter the process exits, it can beNimbusAutomatically identify and pull up without affecting running business.

B.

WorkerCan be used after abnormal exitSupervisorPulls up automatically without manual intervention.

C.

When a node fails, the tasks on the node will be reassigned to other normal nodes without manual intervention.

D.

NimbusAfter the failure, the backupNimbusIt will take over automatically without affecting the running business.

Full Access
Question # 66

In useSolrWhen performing a full-text search, you canwtThe parameter specifies the response format of the query result.

aboutSolrThe response format of the query result, which of the following statements is wrong?

A.

supportCSVandJSON

B.

supportCSV,JSONandHTML

C.

supportCSV,JSONandXML

Full Access
Question # 67

HDFSIn application development of , which of the following areHDFSInterfaces supported by the service? (multiple choice)

A.

BufferedOutputStream.write

B.

BufferedOutputStream.flush

C.

FileSystem.create

D.

FileSystem.append

Full Access
Question # 68

Which of the following measures can improveHBasequery performance? (multiple choice)

A.

pre-scoreregion,MakeregionEvenly distributed, improve concurrency

B.

Query frequent attributes are placed inrowkeyThe front part

C.

Use secondary indexes to adapt to more query scenarios

D.

Data read at the same time are stored in the samecellmiddle

Full Access
Question # 69

FusionInsight HD in real-time processing scenarios, what computing frameworks are available? (multiple choice)

A.

Spark Streaming

B.

Streaming

C.

MapReduce

D.

HDFS

Full Access
Question # 70

when aMapReduceWhen the application is executed, which of the following actions ismapoccurred before the stage?

A.

split

B.

combine

C.

partition

D.

sort

Full Access
Question # 71

YarnWhich role is to manage individual node resources (CPU/Memory)of?

A.

NodeManager

B.

Resource Manager

C.

Data Node

D.

NameNode

Full Access
Question # 72

forSpark Streamingapplication, in aJVM, there can only be one at a timeStreamingContextactive condition.

A.

True

B.

False

Full Access
Question # 73

RedisofLISTData structure, suitable for which of the following scenarios? (multiple choice)

A.

Build queuing systems, such as message queues

B.

uniqOperations, such as getting the ranking value of all data in a certain period of time

C.

get latestNdata operations: for example, for a certain microblog, get the latest10comments

D.

Simulate stack operations

Full Access
Question # 74

TowardsHBaseincrease in the clusterRegionServerhost, the original cluster must be stopped first, becauseHBaseDynamic expansion is not supported.

A.

True

B.

False

Full Access
Question # 75

FusionInsight HDin, useStreamingofLinuxWhen submitting a topology in command line mode, you need to first use a StreamingUser with submit permissionkinitway authentication

A.

True

B.

False

Full Access
Question # 76

FusionInsight HDmiddle,StreamingWhich of the following scenarios is applicable? (multiple choice)

A.

Streaming data monitoring

B.

Real-time visit statistics of the website

C.

Offline log analysis

D.

Traffic flow analysis

Full Access
Question # 77

existFusionInsight HDin, aboutHivepartition (partition) function, which is wrong as described below?

A.

Partition fields are defined when the table is created

B.

There can only be one partition field, and multi-level partitions cannot be created

C.

Using partitions can reduce the data scan range of some queries, thereby improving query efficiency

D.

Partition fields can be used aswherecondition of the sentence

Full Access
Question # 78

existFusionInsight HD where can I viewMapReduceThe result of running the application?

A.

FusionInsight Managerofwebinterface

B.

Loader Webinterface

C.

Solr webinterface

D.

Yarn Webinterface

Full Access