What are the mainstream deep learning open source tools? (Multiple choice)
A.
TensorFlow
B.
Caffe
C.
Torch
D.
Theano
E.
Scikit..Jeam
TensorFlow
Caffe
Torch
Theano
After the data has completed the feature engineering operation, in the process of constructing the model, which of the following options is not a step in the decision tree construction process?
A.
Pruning
B.
Feature selection
C.
Data cleaning
D.
Decision tree generation
Data cleaning
Image label service returned tag There can be more than one.
A.
TRUE
B.
FALSE
TRUE
Which of the following options is not the Tensorflow build process?
A.
Building a calculation chart
B.
Input tensor
C.
Generate a session
D.
Update weights
Update weights
Atlas 800 AI There are multiple models of servers, of which the one based on the Kunpeng processor platform is?
A.
Atlas 800 model: 9000
B.
Atlas 800 model: 3000
C.
Atlas 800 model: 3010
Atlas 800 model: 3000
Voice recognition refers to the recognition of audio data as text data.
A.
TRUE
B.
FALSE
TRUE
enter 32*32 Image with size 5*5 The step size of the convolution kernel is 1 Convolution calculation, output image Size is:
A.
28*23
B.
28*28
C.
29*29
D.
23*23
28*28
Faced with the challenge of achieving efficient distributed training for ultra-large scale models, MindSpore is handled as?
A.
Automatic parallel
B.
Serial
C.
Manual parallel
Automatic parallel
On-Device Execution, that is, the entire image is offloaded and executed, and the computing power of the Yiteng chip can be fully utilized, which can greatly reduce the interaction overhead, thereby increasing the accelerator occupancy rate. On Device The following description is wrong?
A.
MindSpore Realize decentralized autonomy through adaptive graph optimization driven by gradient data A11 Reduce, Gradient aggregation is in step, and calculation and communication are fully streamlined.
B.
Challenges of model execution under super chip computing power: Memory wall
problems, high interaction overhead, and difficulty in data supply. Partly in Host Executed, partly in Device Execution, interaction overhead is even much greater than execution overhead, resulting in low accelerator occupancy.
C.
MindSpore Through the chip-oriented depth map optimization technology, the
synchronization wait is less, and the "data computing communication" is maximized. The parallelism of “trust”, compared with training performance Host Side view scheduling method is flat.
D.
The challenge of distributed gradient aggregation under super chip computing power:ReslNet50 Single iteration 20ms Time will be generated The synchronization overhead of heart control and the communication overhead of frequent synchronization. Traditional methods require 3 Synchronization completed A11 Reduce, Data-driven method autonomy A11 Reduce, No control overhead.
MindSpore Through the chip-oriented depth map optimization technology, the
synchronization wait is less, and the "data computing communication" is maximized. The parallelism of “trust”, compared with training performance Host Side view scheduling method is flat.
PyTorch Which of the following functions does not have?
A.
Inline keras
B.
Support dynamic graph
C.
Automatic derivative
D.
GPU accelerate
Inline keras
Page 5 out of 37 Pages |
Previous |