C++ API Reference for Intel® Data Analytics Acceleration Library 2018 Update 2
This is the complete list of members for DistributedInput< step1Local >, including all inherited members.
add(InputCollectionId id, size_t key, const data_management::TensorPtr &value) | Input | |
Argument() | Argument | inline |
Argument(const size_t n) | Argument | |
Argument(const Argument &other) | Argument | protected |
check(const daal::algorithms::Parameter *par, int method) const DAAL_C11_OVERRIDE | DistributedInput< step1Local > | virtual |
get(Step1LocalInputId id) const | DistributedInput< step1Local > | |
daal::algorithms::neural_networks::training::interface1::Input::get(InputId id) const | Input | |
daal::algorithms::neural_networks::training::interface1::Input::get(InputCollectionId id) const | Input | |
daal::algorithms::neural_networks::training::interface1::Input::get(InputCollectionId id, size_t key) const | Input | |
daal::algorithms::interface1::Input::get(size_t index) const | Argument | protected |
getStorage(Argument &a) | Argument | protectedstatic |
getStorage(const Argument &a) | Argument | protectedstatic |
daal::algorithms::interface1::Input::Input() | Input | inline |
daal::algorithms::interface1::Input::Input(const Input &other) | Input | inlineprotected |
operator<<(const data_management::SerializationIfacePtr &val) | Argument | inline |
set(Step1LocalInputId id, const ModelPtr &value) | DistributedInput< step1Local > | |
daal::algorithms::neural_networks::training::interface1::Input::set(InputId id, const data_management::TensorPtr &value) | Input | |
daal::algorithms::neural_networks::training::interface1::Input::set(InputCollectionId id, const data_management::KeyValueDataCollectionPtr &value) | Input | |
daal::algorithms::interface1::Input::set(size_t index, const data_management::SerializationIfacePtr &value) | Argument | protected |
setStorage(const data_management::DataCollectionPtr &storage) | Argument | protected |
size() const | Argument | inline |
For more complete information about compiler optimizations, see our Optimization Notice.