In the operation process of F1ink, the role responsible for applying for resources is
In the Fusioninsight Manager interface, which of the following options is not included in the operation of the loader?
In an MRS cluster, which of the following components does Spark mainly interact with?
As shown in the figure, the following description of the message read by the Kafka message consumer Consumeri is wrong?
In the task scheduling process of YARN, which of the following is the task that ApplicationMastert is responsible for?
When creating a Loaderf job, in which of the following steps can the filter type be set?
FusionlnsightHD uses the HBase client to write 10 pieces of data in batches. A Regionserver node contains 2 Regions of the table, A and B, respectively. Two of the 10 pieces of data belong to A. 4 belong to B. Clearly write How many RPC requests do I need to send to the Regionserver to enter these 10 pieces of data?
Which of the following functions can the Kafka Cluster Mirroring tool achieve?
In Hadoop, if yarn.scheduler.capacity.root.QueueA. minimum-user-limit-percenti is set to 50, which of the following statements is wrong?
In many small file scenarios, Spark will start many tasks. When there is a Shuffle operation in the SQL logic, the number of hash buckets will be greatly increased, which will seriously affect the performance. In Fusioninsight, scenarios for small files usually use the( )Operator to merge partitioni generated by small files in Tabler, reduce the number of partitions, avoid generating too many hash buckets during shuffle, and improve performance?
When the Loader of Fusioninsight HD creates a job, what is the function of the connector?
Which of the following programming languages is Spark implemented in?
AC
B.C++
C. JAVA
D. Scala
When the loader in Fusioninsight HD imports files from the SFTP server, which of the following file types does not require encoding conversion and data conversion and is the fastest?
In the Fusioninsight HD system, which component does the flume data flow not need to pass through in the node?
Channel supports transactions and provides weaker order guarantees. Any number of Sources and Sinks can be connected.
Data in FlumeCompression characteristics are mainlyYesFor which of the following purposes?
After the Loadert job in Fusioninsight HD is submitted to YARN, the job cannot be manually stopped.
When a Spark application is running, if a certain task fails to run, the entire app fails to run.
Streaming of Fusioninsight HD is weakly dependent on Zookeeper, even if Zookeeperd fails, Streaming can provide services normally.
The MOB data in HBaser is directly stored in the HFile format on HDFS. Then, the address information and size M information of this file are stored as value in the store that manages HBase. These files are centrally managed through tools. This can greatly reduce the frequency of compation and split of HBase and improve performance.
Kafka is a distributed message publishing and subscription system. It only forwards messages and does not save messages.
Spark on YARN-clienti is suitable for production environment because it can see the output of APP faster.
Fl? ?in transmissiondata process. Data can be simply filtered.hostwantby setting( )at Source? ?between (writeCa?Before)ordon't caredata is filtered.
In the Fusionlnisght HD system, the number of Partitions and the number of replicas must be set when creating the Topich of Kafkal. Setting up multiple replicas can enhance the disaster tolerance of the Kafka service.
In order to consider performance optimization. It is recommended to deploy LdapServer and KrbServer on the same node in all clusters
Loaderf only supports data import and export between relational databases and HBase.
In the MRS service, the unavailability of the Zookeeper service will result in the unavailability of the kafka service.
HDFS supports large file storage, and supports multiple users' write operations to the same file, as well as modification at any position of the file.
The combine of MapReduce in the Map phase is a pre-grouping process and is optional.
Streaming:Event listening is mainly implemented through which of the following services provided by Zookeeper?
When planning and deploying a Fusionlnsight cluster, it is recommended that the management node be best deployed( ), the control node needs to be deployed at least( )Piece,Data nodes need to be deployed at least( )Piece.
Which of the following descriptions about HBase. Secondary Index is correct
The scheme authentication method of single-choice Zookeeper does not include which of the following?( )
What can the unified certification management system of the new mainstream merchants consist of?
HDFS of HadoopE is a distributed file system. Which of the following application scenarios is suitable for data storage and management?
About Spark SQL&Hive difference and connection, which of the following statements is correct?
In the Fusioninsight HD system, it fails to view the topology or submit the topology using the Shelle command of the Streaming client. Which of the following positioning methods are correct?
In a Fusioninsight HD V100R002C60 cluster, which of the following components need to be partitioned for metadata?
In the process of using Flume to transmit data, in order to prevent data loss due to the restart of the Flume process. Which of the following Channel types can be used
If HDFS is deployed with a single name node, what limitations may exist? (multiple choice)
To enable the log aggregation function of the Yam component in the Hadoop platform, which parameter needs to be configured?
There are 230 nodes in the FusionlnsightHD cluster of a university. Which of the following processes should be deployed on the data nodes during cluster planning?
ADataNode
B. NameNode
C. NodeManager
D. RegionServer
E.DBServer
F.SolrServer
SELECT aa.salarybB. address FROM employee aa JoiN SELECT adress FROM employee info where provine='zhejiang') What types of operations does bb ONaa.nanme=bB. name contain?
Which components in FusioninsightHD can Zookeeper provide distributed management support for?
Which of the following components does the platform architecture of Huawei's big data solution include?
The following figure shows the label storage strategy of HDFS. Observe the figure below, which data nodes will HBasel data be stored on
In the HDFS federated environment, which of the following contents are included in the NameSpace
Are the following descriptions correct about the HBase file storage module (HBase FileStream, HFS for short)?
Which of the descriptions of the Loader job in FusionlnsightHD is correct?
What information does a Key Value format in the HBase data file HFiler contain?
Which of the following belongs to DDL (Data Definition Language) in Hive SQL?
Which of the following statements about data management and architecture faced by traditional enterprises is correct?
Which statement is correct about worker (worker process), Executor (thread) and task (task)?
Which of the following statements about atomicity in Zooeeper's key features is correct?
Fusionlnsight HD cluster upgrade, which of the following descriptions are correct?
Fl?Me is akind of distributed. highreliable and highavailableofClothesservice. useto be effectivecompare?,polymerization?Move a lot of logs? ?.
In Huawei Fusioninsight, HBase's table design tool, connection pool management and enhanced $DK can simplify the business development of complex data tables.
In Streaming, exactly one message reliability level is achieved through the ACK mechanism.
Under the HDFS federation mechanism, metadata between NameNodes is not shared.
The following figure shows the storage locations of files A, B, C, and D. A and B are related, and their storage locations conform to the Colocation strategy.
When creating a Loader job, in which of the following steps can the number of Maps be set?
The secondary index provides HBase with the ability to index according to the value of some columns. The secondary index first searches the index table, and then locates the position in the data table, without full table scan
When using Loaderi for data import and export, data processing must go through the Reducel stage
According to the way of data acquisition in Flue, Source can be divided into driven Sourcet and( )Source.