- The Tensor Boost Engine (TBE) is a processor enablement layer built by Huawei for deep neural networks and Ascend AI processor operators library and a highly automated operators development kits.
True
False
- In Tensorflow 2.0 tf.data contains method to write data in TFRecord Format .
True
False
- Which of the following statement are through about hidden layers
- The more hidden layer the stronger the identification capability
- Excessive hidden layer may cause overfitting
- The more hidden layer the better
- Insufficient hidden layer may cause overfitting
Ans: AB
- which of the following are common kernel functions in support of vector machine (SVM)
A.Gausian kernel function
B.Sigmoid kernel function
C.Relu kernel function
D.Polynomial kernel function
Ans : ABD
5 which of the following development modes are supported by the ModelArts training platform
- In- Cloud development (codelab)
- In- Cloud development (Notebook +SDK)
- On- premises development (PyCharm +PyCharm toolkit)
- On – premises development (IDE+SDK)
Ans: ABCD
- Backed by HUAWEI CLOUD ' accumulated knowledge and expertise in AI Big data and other cutting edge technologies, GeoGenius offers a one stop AI development cloud platform for remote sensing data processing mining and management. Which of the following are the main functions of GeoGenus?
- Wheather forecasting
- Afforestation
- Emergency response and disaster prevention
- Smart transport infrastructure
Ans: AC
- Sensors and IOT technologies support for the () of the four element of AI data;Data;DATA
- The average value of tensors generated using tf.random.normal ([4], 2,1, tf.float32) is (). (Fill the blank with a number with a number). 2
- In mindspore if you want to define a function for collecting information after each step is complete, you need to inherit the () class. Callback; callback
- HUAWEI Cloud () helps enterprises build invoice management systems and self service reimburstment systems. ( Fill in the blank with the abbreviation of the service name). OCR; ocr
- In an ADAGRAD optimiser, the initial value of the cummulative square gradient is () (Fill in the blank with a number). 0
- MindSpore.ops.Gradoperation is the first order derivative method in Mindspore to compute gradients. It contains multiple parameters. When get_all is set to (), the derivatives of all parameters are computed. True
- The spartial relationship of image is that the local pixels are closely related and the pixels that are far away from each other are weakly related.
True
false
- The tree model used at the bottom layer of the random forest and GBDT algorithm is only the classification and regression Tree (CART).
True
false.
- Typically, a larger K value in the K nearest neighbour (KNN) algorithm reduces the impact of noise on classification but make the boundary between classes less obvious.
True
false.
- if a 32×32 image (without padding) is input and 5 x 5 kernel is used for convolution with a step of 2, the size of the output image is 13 x 13.
True
false
- The model Arts inference platform implements fast and efficient model inference and image recognition.
True
False
- Grid search randomly selects hyper parameter combinations therefore, it has high efficient and accuracy.
True
False
- The softplus function has a continous derivative and defines a smooth curved surface.
True
False
- A decision selects a label from features of the training data to work as the node splitting standard. Different label selection standards generate different decision tree algorithm.
True
False
- Assuming the number of hyperparameter is the same, stochastic descent (SGD) combined with manual adjustment achieves a better effect than aqdaptive learning rates.
True
False
22.Filter methods for feature selection are restricted they do not tend to select redundant variables because they consider the relationship between features.
True
False
- A model constructed using machine learning algorithms can be a functions that conforms to actual data distribution.
True
False
- Creative work, such as writing, will not be easily replicated by AI
True
False
- In the deep neural network, the back propagation algorithm uses the L&3×27; rule to perform layer by layer derivation and find the optimal parameters
True
False
- Device, edge cloud can respectively refers to smartphone . Huawei intelligent edge devices and Huawei CLOUD
True
False
- The XOR problem can be solved by using the deep neural network.
True
False
- Mindspore accelerated model convergence through automatic parallelism and second order optimization
True
False
- The PIL module in Tensorflow 2.0 is used to perform operation on images
True
False
- HUAWEI 3.0 supports the development of P50 smartphones
True
False
- Powered by Ascend AI processors, the Huawei Atlas AI computing platforms delivers product that come in various forms, such as module, card, edge statioons, secure and clusters, to build AI solutions for all scenarios across the device edge and processor
True
False
- Which of the followings statement is true about the low latency design of CPU
- The low clock frequency reduces the latency
- Complex logic control units increase the latency
C . A CPU is equipped with few ALU’s and caches. These cache merge access request to the same data and then access the DRAM reducing the latency
D.The powerful ALU can complete the computation in a short clock cycle.
Ans: D
- Which of the following statement is false about the ReLU function.
- The ReLU function is not differentiable at X=0 and a derivative is forcibly defined at this point.
- The ReLU function effectively alleviates the vanishing gradient problem
- The surface defined at the zero point of the ReLU function is not smooth enough in some regression problems
- Compared with sigmoid and tahn, the convergence of the ReLU function is slow
Ans: D
- Which of the following statement is false about gradient descent algorithm.
- The global gradient descent is more stable than the stochastic gradient descent (SGD)
- Whwn there are too many samples and GPU are not used for parallel computing , the convergence process of the global descent is time consuming.
- When GPU’S are not used for parallel computing, the Mini- Batch Gradient Decsent (MGBD take less time than SGD to complete an epoch
- Each time the global descent updates its weight, all training samples ne to be calculated.
Ans: C
- " knowledge representation is the unique method of representing knowlegde using a set of symbols in a structure that can be understood by compute squot; which of the following is true about this statement.
- This statement is false: The method of representing knowledge is not unique
- This statement is correct: The coding method of knowledge representation is unique
- This statement is correct: The knowledge representation can support expert system
- This statement is false: Knowledge representation cannot be used for expert rules or fuzzy inference
Ans: A
- Which of the followings is the correct shape of sensor ([[0,1) (2,3]] (4,5] (6, 77)]]?
A.(3, 3, 2)
- (3,2,4)
- (2,3,4)
- (2,2,2)
Ans: D
- Which of the following statements is true about variance & bias
- A model with low bias but high variance has high robustness
- A model with high bias but low variance has high precision
- A model with low bias and variance has low precision
- A model with high bias and variance performs poorly and will not be used.
Ans: D
- Which of the following statement is true about the ReLU function.
- The ReLU function is differentiable at X = 0 and the derivatives is 1
- Neuron nerosis does not
- There is no upper bound, so training is relatively easy to diverge
- The gradient is always 1 and the vanishing gradient problem can be perfectly solved
Ans: C
- Which of the following is false about the softmax function?
- The softmax function turns a vector of K real number into another vector of K real number
- The value of each element in the vector obtained from the softmax function ranges from 0 to 1
- The elements of the vector obtained from the softmax function do not add up to 1
- The softmax function is often used as the output layer of a multiclass classification task
Ans: C
- A generative adversarial network is like a game system. The generator produces fake samples, and the discriminator tries to distinguish red data from the data created by the generator. What is the ideal result.
- The generator is trained to fool the discrimitor
- The discriminator cannot distinguish real data from the data created by the generator
- The discriminator distinguishes real data from the data created by the generator
- Develop a high precision discriminator and a generator that cannot fool the discriminator
Ans: B
- Which of the following is the main function of HUAWEI CLOUD GeoGenius?
- Ecological detectin
- Heat demand forecast
- Real- time traffic detection
- Traffic Incident detection
- Which of the followings is the reasoning method of the production systems that draws conclusion through a rule library.
- Forward
- Backward
- Bidirectional
- Random
Ans: A
- Dirty data refers to data with quality problems. Which of the following statements is false about the data quality
- Noise: Data contains incorrect record or exceptions.
- Incompleteness: Data lacks attributes or contains some missing values
- Unprocessd: Data for which feature engineering has not been performed
- Inconsisitency: Data contains conflicting records
Ans: C
- Which of the followings is not a main function of Huawei CLOUD GeoGenius?
- Natural resource survey
- Ecological environment detection
- Weather forecast
- Real-Time traffic detections
Ans: D
- Which of the following statement is true about the loss functions typically used in deep learning
- Quadratic cost functions are usually for classification while cross entropy cost functions are used for clustering.
- The quadratic cost function reflects the gap between the target output and the actual outout
- The quadratic cost function reflects the gap between two probability distributions
- The purpose of training is to minimize the loss function.
Ans: B
- Which of the following is not a network operator in Mindspore
- Conv2D
- SGD
- Control Depend
- Softmax
Ans: C
- Which of the following product can be used by a company that wants to provide intelligent customer service?
- RDB
- Conversational BOT
- OBS
- GES
Ans: B
- Which of the following statement is true about the chip enablement layer in the Acsend AI software?
- The Ascend AI software stack encapsulates the framework involking and offline model generation capabilities .
- The Ascend AI software stack is the hardware compute basis of Ascend AI processors
- The Ascend AI software stack is nearest to the underlying compute resources and output operator layer task and the hardware.
- The Ascend AI software stack can invoke on appropriate deep learning framework based on the factors of related algorithms.
Ans: C
- Which of the following statements is true about rule based approaches?
- Rules are obtained through machine learning
- Rules based approaches use explicit programimg to solve problems
- Rule based approaches use sample for training
- Decision making rules are complex or difficult to desribe
Ans: B
- Which of the following statements are true about ensemble learning?
- Ensemble learning uses the decision tree as the bottom layer algorithm
- The voting policy of ensemble learning can be the averaging method
- A batch of features are randomly selected for subtree training in a random forest
- Xgboost supports parallel tarining like the random forest.
Ans: BC
- Which of the following statements are false about the Gated Recurrent Unit (GRU)
- Unlike long short term memory, GRU merges the cell state and hidden state.
- GRU is a variant of convolutionary Neural network (CNN)
- GRU combines the forget gate and update gates into a single input gate.
- GRU assembles the output of the previoud time step with the input of the current time step.
Ans: BC
- Which of the following use the Da Vinci architecture?
- Ascend 310
- RTX 3080
- Kunpeng 920
- Ascend 910
Ans: AD
- Which of the following statement are false about the universal engine of the Ascend 310 software stack.
- Provides general capabilities needed for neural network training.
- Provides compute resources and executes specific computing tasks
- Provides compute resources for Ascend AI
- Outputs tasks at the operation layer for hardware
Ans: ABCD
- Which of the following are part o minspore
- Mind Expression API layer
- Mind compile’s real-time graph – level compilation subsystem .
- CUDA and third party chips
- MindData
Ans: ABD
- Which of the followings statement are false about common optimizers
- One of the advantage of Adagrad optimizers is that the parameters update operation does not end too early
- When the momentum optimizer is used, parameter are updated by using the same learning rate, but momentum coefficient keep changing with each iteration
- Adam sets different learning rates for different parameter
- The RMSPROP Optimizer introduces an attenuation coefficient to ensure a certain proportion of gradient attenuation in each round.
Ans: AB
- Which of the following are steps of the BPTT algorithm
- Calculating the output of each neuron (forward)
- Calculating the error of each neuron (backward)
- Calculating the gradient of each weight
- Calculating the layer of neuron
Ans: ABCD
- Which of the following statement are true about the mindspore componets
- Communication: Processes data flow communication between the CPU and memory
- Dataset: Dataset processing module that loads and preprocesses data
C: NN: Neural Network cells in Mindspore that define loss functions and optimizes
D: Train: relates to training model and model quantization module
Ans: BC
- In Tensor flow2.0 which of the following cannot be used to create all zero tensor?
- tf.zeros ()
- tf.fill ( (1,3,0)
- tf.zeros –frame()
- tf.zeros tuple()
Ans: CD
- The accuracy of a trained model is high on a training data set but low on a test dataset. In this case which of the followings operations can be perfomed
- Use the sigmoid activation function
- Perform data asymetric
- Add regular terms
Ans: BC
- Which of the followings are major advantage of kera
- Easy to use
- Low level coding style
- Modular and combination
- Easy to
Ans: ACD
- Which of the following statements are true about common activation functions in deep learning.
- The tahn function is symetric with respect to the origin and the mean of its ouput is closer to 0
- During training of a deep neural network, the sigmoid, tanh and softsign functions cannot prevent vanishing gradient
Ans: AB
Thank You Mr. Debo for this