diff options
author | luz.paz <luzpaz@users.noreply.github.com> | 2018-02-12 07:07:39 -0500 |
---|---|---|
committer | luz.paz <luzpaz@users.noreply.github.com> | 2018-02-12 07:09:43 -0500 |
commit | 5718d09e392a79881ba723376ea9af211a3e7b3f (patch) | |
tree | 917d88a502affb4b35297e2e9c732be18c484896 /modules/dnn | |
parent | b67523550f7c130332e5b5e538397897b195cb98 (diff) | |
download | opencv-5718d09e392a79881ba723376ea9af211a3e7b3f.tar.gz opencv-5718d09e392a79881ba723376ea9af211a3e7b3f.tar.bz2 opencv-5718d09e392a79881ba723376ea9af211a3e7b3f.zip |
Misc. modules/ typos
Found via `codespell`
Diffstat (limited to 'modules/dnn')
-rw-r--r-- | modules/dnn/CMakeLists.txt | 2 | ||||
-rw-r--r-- | modules/dnn/include/opencv2/dnn.hpp | 2 | ||||
-rw-r--r-- | modules/dnn/include/opencv2/dnn/all_layers.hpp | 18 | ||||
-rw-r--r-- | modules/dnn/include/opencv2/dnn/dnn.hpp | 4 | ||||
-rw-r--r-- | modules/dnn/src/caffe/opencv-caffe.proto | 2 | ||||
-rw-r--r-- | modules/dnn/src/dnn.cpp | 6 | ||||
-rw-r--r-- | modules/dnn/src/layers/concat_layer.cpp | 2 | ||||
-rw-r--r-- | modules/dnn/src/ocl4dnn/src/math_functions.cpp | 2 | ||||
-rw-r--r-- | modules/dnn/src/opencl/conv_layer_spatial.cl | 4 | ||||
-rw-r--r-- | modules/dnn/src/tensorflow/tf_importer.cpp | 4 | ||||
-rw-r--r-- | modules/dnn/test/cityscapes_semsegm_test_enet.py | 4 | ||||
-rw-r--r-- | modules/dnn/test/imagenet_cls_test_alexnet.py | 4 | ||||
-rw-r--r-- | modules/dnn/test/imagenet_cls_test_googlenet.py | 4 | ||||
-rw-r--r-- | modules/dnn/test/imagenet_cls_test_inception.py | 4 | ||||
-rw-r--r-- | modules/dnn/test/pascal_semsegm_test_fcn.py | 2 |
15 files changed, 32 insertions, 32 deletions
diff --git a/modules/dnn/CMakeLists.txt b/modules/dnn/CMakeLists.txt index 27717e1f5c..7971046851 100644 --- a/modules/dnn/CMakeLists.txt +++ b/modules/dnn/CMakeLists.txt @@ -41,7 +41,7 @@ endif() add_definitions(-DHAVE_PROTOBUF=1) -#supress warnings in autogenerated caffe.pb.* files +#suppress warnings in autogenerated caffe.pb.* files ocv_warnings_disable(CMAKE_CXX_FLAGS -Wunused-parameter -Wundef -Wignored-qualifiers -Wno-enum-compare -Wdeprecated-declarations diff --git a/modules/dnn/include/opencv2/dnn.hpp b/modules/dnn/include/opencv2/dnn.hpp index 690a82ab84..57a564bf11 100644 --- a/modules/dnn/include/opencv2/dnn.hpp +++ b/modules/dnn/include/opencv2/dnn.hpp @@ -53,7 +53,7 @@ - API for new layers creation, layers are building bricks of neural networks; - set of built-in most-useful Layers; - API to constuct and modify comprehensive neural networks from layers; - - functionality for loading serialized networks models from differnet frameworks. + - functionality for loading serialized networks models from different frameworks. Functionality of this module is designed only for forward pass computations (i. e. network testing). A network training is in principle not supported. diff --git a/modules/dnn/include/opencv2/dnn/all_layers.hpp b/modules/dnn/include/opencv2/dnn/all_layers.hpp index 6741efaac1..061d184db6 100644 --- a/modules/dnn/include/opencv2/dnn/all_layers.hpp +++ b/modules/dnn/include/opencv2/dnn/all_layers.hpp @@ -51,13 +51,13 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN /** @defgroup dnnLayerList Partial List of Implemented Layers @{ - This subsection of dnn module contains information about bult-in layers and their descriptions. + This subsection of dnn module contains information about built-in layers and their descriptions. - Classes listed here, in fact, provides C++ API for creating intances of bult-in layers. + Classes listed here, in fact, provides C++ API for creating instances of built-in layers. In addition to this way of layers instantiation, there is a more common factory API (see @ref dnnLayerFactory), it allows to create layers dynamically (by name) and register new ones. - You can use both API, but factory API is less convinient for native C++ programming and basically designed for use inside importers (see @ref readNetFromCaffe(), @ref readNetFromTorch(), @ref readNetFromTensorflow()). + You can use both API, but factory API is less convenient for native C++ programming and basically designed for use inside importers (see @ref readNetFromCaffe(), @ref readNetFromTorch(), @ref readNetFromTensorflow()). - Bult-in layers partially reproduce functionality of corresponding Caffe and Torch7 layers. + Built-in layers partially reproduce functionality of corresponding Caffe and Torch7 layers. In partuclar, the following layers and Caffe importer were tested to reproduce <a href="http://caffe.berkeleyvision.org/tutorial/layers.html">Caffe</a> functionality: - Convolution - Deconvolution @@ -125,12 +125,12 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN virtual void setOutShape(const MatShape &outTailShape = MatShape()) = 0; /** @deprecated Use flag `produce_cell_output` in LayerParams. - * @brief Specifies either interpet first dimension of input blob as timestamp dimenion either as sample. + * @brief Specifies either interpret first dimension of input blob as timestamp dimenion either as sample. * - * If flag is set to true then shape of input blob will be interpeted as [`T`, `N`, `[data dims]`] where `T` specifies number of timpestamps, `N` is number of independent streams. + * If flag is set to true then shape of input blob will be interpreted as [`T`, `N`, `[data dims]`] where `T` specifies number of timestamps, `N` is number of independent streams. * In this case each forward() call will iterate through `T` timestamps and update layer's state `T` times. * - * If flag is set to false then shape of input blob will be interpeted as [`N`, `[data dims]`]. + * If flag is set to false then shape of input blob will be interpreted as [`N`, `[data dims]`]. * In this case each forward() call will make one iteration and produce one timestamp with shape [`N`, `[out dims]`]. */ CV_DEPRECATED virtual void setUseTimstampsDim(bool use = true) = 0; @@ -146,7 +146,7 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN * @param output contains computed outputs: @f$h_t@f$ (and @f$c_t@f$ if setProduceCellOutput() flag was set to true). * * If setUseTimstampsDim() is set to true then @p input[0] should has at least two dimensions with the following shape: [`T`, `N`, `[data dims]`], - * where `T` specifies number of timpestamps, `N` is number of independent streams (i.e. @f$ x_{t_0 + t}^{stream} @f$ is stored inside @p input[0][t, stream, ...]). + * where `T` specifies number of timestamps, `N` is number of independent streams (i.e. @f$ x_{t_0 + t}^{stream} @f$ is stored inside @p input[0][t, stream, ...]). * * If setUseTimstampsDim() is set to fase then @p input[0] should contain single timestamp, its shape should has form [`N`, `[data dims]`] with at least one dimension. * (i.e. @f$ x_{t}^{stream} @f$ is stored inside @p input[0][stream, ...]). @@ -328,7 +328,7 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN * @param begin Vector of start indices * @param size Vector of sizes * - * More convinient numpy-like slice. One and only output blob + * More convenient numpy-like slice. One and only output blob * is a slice `input[begin[0]:begin[0]+size[0], begin[1]:begin[1]+size[1], ...]` * * 3. Torch mode diff --git a/modules/dnn/include/opencv2/dnn/dnn.hpp b/modules/dnn/include/opencv2/dnn/dnn.hpp index 15c41b3079..4ad303594e 100644 --- a/modules/dnn/include/opencv2/dnn/dnn.hpp +++ b/modules/dnn/include/opencv2/dnn/dnn.hpp @@ -691,7 +691,7 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN * @param swapRB flag which indicates that swap first and last channels * in 3-channel image is necessary. * @param crop flag which indicates whether image will be cropped after resize or not - * @details if @p crop is true, input image is resized so one side after resize is equal to corresponing + * @details if @p crop is true, input image is resized so one side after resize is equal to corresponding * dimension in @p size and another one is equal or larger. Then, crop from the center is performed. * If @p crop is false, direct resize without cropping and preserving aspect ratio is performed. * @returns 4-dimansional Mat with NCHW dimensions order. @@ -719,7 +719,7 @@ CV__DNN_EXPERIMENTAL_NS_BEGIN * @param swapRB flag which indicates that swap first and last channels * in 3-channel image is necessary. * @param crop flag which indicates whether image will be cropped after resize or not - * @details if @p crop is true, input image is resized so one side after resize is equal to corresponing + * @details if @p crop is true, input image is resized so one side after resize is equal to corresponding * dimension in @p size and another one is equal or larger. Then, crop from the center is performed. * If @p crop is false, direct resize without cropping and preserving aspect ratio is performed. * @returns 4-dimansional Mat with NCHW dimensions order. diff --git a/modules/dnn/src/caffe/opencv-caffe.proto b/modules/dnn/src/caffe/opencv-caffe.proto index 88aaa86c22..41cd46bb1c 100644 --- a/modules/dnn/src/caffe/opencv-caffe.proto +++ b/modules/dnn/src/caffe/opencv-caffe.proto @@ -131,7 +131,7 @@ message PriorBoxParameter { // Variance for adjusting the prior bboxes. repeated float variance = 6; // By default, we calculate img_height, img_width, step_x, step_y based on - // bottom[0] (feat) and bottom[1] (img). Unless these values are explicitely + // bottom[0] (feat) and bottom[1] (img). Unless these values are explicitly // provided. // Explicitly provide the img_size. optional uint32 img_size = 7; diff --git a/modules/dnn/src/dnn.cpp b/modules/dnn/src/dnn.cpp index be4767b28a..b66fb4236d 100644 --- a/modules/dnn/src/dnn.cpp +++ b/modules/dnn/src/dnn.cpp @@ -58,7 +58,7 @@ namespace cv { namespace dnn { CV__DNN_EXPERIMENTAL_NS_BEGIN -// this option is usefull to run valgrind memory errors detection +// this option is useful to run valgrind memory errors detection static bool DNN_DISABLE_MEMORY_OPTIMIZATIONS = utils::getConfigurationParameterBool("OPENCV_DNN_DISABLE_MEMORY_OPTIMIZATIONS", false); using std::vector; @@ -911,7 +911,7 @@ struct Net::Impl int id = getLayerId(layerName); if (id < 0) - CV_Error(Error::StsError, "Requsted layer \"" + layerName + "\" not found"); + CV_Error(Error::StsError, "Requested layer \"" + layerName + "\" not found"); return getLayerData(id); } @@ -1897,7 +1897,7 @@ struct Net::Impl if ((size_t)pin.oid >= ld.outputBlobs.size()) { CV_Error(Error::StsOutOfRange, format("Layer \"%s\" produce only %d outputs, " - "the #%d was requsted", ld.name.c_str(), + "the #%d was requested", ld.name.c_str(), ld.outputBlobs.size(), pin.oid)); } if (preferableTarget != DNN_TARGET_CPU) diff --git a/modules/dnn/src/layers/concat_layer.cpp b/modules/dnn/src/layers/concat_layer.cpp index f8c0a27dbe..63b722ee9a 100644 --- a/modules/dnn/src/layers/concat_layer.cpp +++ b/modules/dnn/src/layers/concat_layer.cpp @@ -88,7 +88,7 @@ public: for (int curAxis = 0; curAxis < outputs[0].size(); curAxis++) { if (curAxis != cAxis && outputs[0][curAxis] != curShape[curAxis]) - CV_Error(Error::StsBadSize, "Inconsitent shape for ConcatLayer"); + CV_Error(Error::StsBadSize, "Inconsistent shape for ConcatLayer"); } } diff --git a/modules/dnn/src/ocl4dnn/src/math_functions.cpp b/modules/dnn/src/ocl4dnn/src/math_functions.cpp index c52a8a93c9..05cfd509b9 100644 --- a/modules/dnn/src/ocl4dnn/src/math_functions.cpp +++ b/modules/dnn/src/ocl4dnn/src/math_functions.cpp @@ -185,7 +185,7 @@ static bool ocl4dnnFastImageGEMM(const CBLAS_TRANSPOSE TransA, int blockC_height = blocksize; int use_buffer_indicator = 8; - // To fix the edge problem casued by the sub group block read. + // To fix the edge problem caused by the sub group block read. // we have to pad the image if it's not multiple of tile. // just padding one line is enough as the sub group block read // will clamp to edge according to the spec. diff --git a/modules/dnn/src/opencl/conv_layer_spatial.cl b/modules/dnn/src/opencl/conv_layer_spatial.cl index 3369c6c971..e31d173d75 100644 --- a/modules/dnn/src/opencl/conv_layer_spatial.cl +++ b/modules/dnn/src/opencl/conv_layer_spatial.cl @@ -188,7 +188,7 @@ __kernel void ConvolveBasic( #define VLOAD4(_v, _p) do { _v = vload4(0, _p); } while(0) // Each work-item computes a OUT_BLOCK_WIDTH * OUT_BLOCK_HEIGHT region of one output map. -// Each work-group (which will be mapped to 1 SIMD16/SIMD8 EU thread) will compute 16/8 different feature maps, but each feature map is for the same region of the imput image. +// Each work-group (which will be mapped to 1 SIMD16/SIMD8 EU thread) will compute 16/8 different feature maps, but each feature map is for the same region of the input image. // NDRange: (output_width+pad)/ OUT_BLOCK_WIDTH, (output_height+pad)/OUT_BLOCK_HEIGHT, NUM_FILTERS/OUT_BLOCK_DEPTH // NOTE: for beignet this reqd_work_group_size does not guarantee that SIMD16 mode will be used, the compiler could choose to use two SIMD8 threads, and if that happens the code will break. @@ -220,7 +220,7 @@ convolve_simd( int in_addr; - // find weights adress of given neuron (lid is index) + // find weights address of given neuron (lid is index) unsigned int weight_addr = (fmg % (ALIGNED_NUM_FILTERS/SIMD_SIZE)) * INPUT_DEPTH * KERNEL_WIDTH * KERNEL_HEIGHT * SIMD_SIZE + lid; for(int i=0;i<OUT_BLOCK_SIZE;i++) { diff --git a/modules/dnn/src/tensorflow/tf_importer.cpp b/modules/dnn/src/tensorflow/tf_importer.cpp index ccb028ba1c..bc112d3560 100644 --- a/modules/dnn/src/tensorflow/tf_importer.cpp +++ b/modules/dnn/src/tensorflow/tf_importer.cpp @@ -1096,9 +1096,9 @@ void TFImporter::populateNet(Net dstNet) dstNet.setInputsNames(netInputs); } else if (type == "Split") { - // TODO: determing axis index remapping by input dimensions order of input blob + // TODO: determining axis index remapping by input dimensions order of input blob // TODO: slicing input may be Const op - // TODO: slicing kernels for convolutions - in current implenmentation it is impossible + // TODO: slicing kernels for convolutions - in current implementation it is impossible // TODO: add parsing num of slices parameter CV_Assert(layer.input_size() == 2); // num_split diff --git a/modules/dnn/test/cityscapes_semsegm_test_enet.py b/modules/dnn/test/cityscapes_semsegm_test_enet.py index 5f0b4ba625..27070d3360 100644 --- a/modules/dnn/test/cityscapes_semsegm_test_enet.py +++ b/modules/dnn/test/cityscapes_semsegm_test_enet.py @@ -8,11 +8,11 @@ try: import cv2 as cv except ImportError: raise ImportError('Can\'t find OpenCV Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') + 'configure environment variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') try: import torch except ImportError: - raise ImportError('Can\'t find pytorch. Please intall it by following instructions on the official site') + raise ImportError('Can\'t find pytorch. Please install it by following instructions on the official site') from torch.utils.serialization import load_lua from pascal_semsegm_test_fcn import eval_segm_result, get_conf_mat, get_metrics, DatasetImageFetch, SemSegmEvaluation diff --git a/modules/dnn/test/imagenet_cls_test_alexnet.py b/modules/dnn/test/imagenet_cls_test_alexnet.py index 46623b8ddf..49597d9540 100644 --- a/modules/dnn/test/imagenet_cls_test_alexnet.py +++ b/modules/dnn/test/imagenet_cls_test_alexnet.py @@ -9,12 +9,12 @@ try: import caffe except ImportError: raise ImportError('Can\'t find Caffe Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "git/caffe/python" directory') + 'configure environment variable PYTHONPATH to "git/caffe/python" directory') try: import cv2 as cv except ImportError: raise ImportError('Can\'t find OpenCV Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') + 'configure environment variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') class DataFetch(object): diff --git a/modules/dnn/test/imagenet_cls_test_googlenet.py b/modules/dnn/test/imagenet_cls_test_googlenet.py index 6c7305b022..28f1abc2ff 100644 --- a/modules/dnn/test/imagenet_cls_test_googlenet.py +++ b/modules/dnn/test/imagenet_cls_test_googlenet.py @@ -7,12 +7,12 @@ try: import caffe except ImportError: raise ImportError('Can\'t find Caffe Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "git/caffe/python" directory') + 'configure environment variable PYTHONPATH to "git/caffe/python" directory') try: import cv2 as cv except ImportError: raise ImportError('Can\'t find OpenCV Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') + 'configure environment variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') if __name__ == "__main__": parser = argparse.ArgumentParser() diff --git a/modules/dnn/test/imagenet_cls_test_inception.py b/modules/dnn/test/imagenet_cls_test_inception.py index d6f0c55209..70017195be 100644 --- a/modules/dnn/test/imagenet_cls_test_inception.py +++ b/modules/dnn/test/imagenet_cls_test_inception.py @@ -9,10 +9,10 @@ try: import cv2 as cv except ImportError: raise ImportError('Can\'t find OpenCV Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') + 'configure environment variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') # If you've got an exception "Cannot load libmkl_avx.so or libmkl_def.so" or similar, try to export next variable -# before runnigng the script: +# before running the script: # LD_PRELOAD=/opt/intel/mkl/lib/intel64/libmkl_core.so:/opt/intel/mkl/lib/intel64/libmkl_sequential.so diff --git a/modules/dnn/test/pascal_semsegm_test_fcn.py b/modules/dnn/test/pascal_semsegm_test_fcn.py index d855786a36..90fb050e0a 100644 --- a/modules/dnn/test/pascal_semsegm_test_fcn.py +++ b/modules/dnn/test/pascal_semsegm_test_fcn.py @@ -9,7 +9,7 @@ try: import cv2 as cv except ImportError: raise ImportError('Can\'t find OpenCV Python module. If you\'ve built it from sources without installation, ' - 'configure environemnt variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') + 'configure environment variable PYTHONPATH to "opencv_build_dir/lib" directory (with "python3" subdirectory if required)') def get_metrics(conf_mat): |