mirror of
https://github.com/opencv/opencv_contrib.git
synced 2025-10-19 11:21:39 +08:00
remove some non-ascii symbols
This commit is contained in:
@@ -1,10 +1,10 @@
|
|||||||
Improved Background-Foreground Segmentation Methods
|
Improved Background-Foreground Segmentation Methods
|
||||||
===================================================
|
===================================================
|
||||||
|
|
||||||
This algorithm combines statistical background image estimation and per-pixel Bayesian segmentation. It[1] was introduced by Andrew B. Godbehere, Akihiro Matsukawa, Ken Goldberg in 2012. As per the paper, the system ran a successful interactive audio art installation called “Are We There Yet?” from March 31 - July 31 2011 at the Contemporary Jewish Museum in San Francisco, California.
|
This algorithm combines statistical background image estimation and per-pixel Bayesian segmentation. It[1] was introduced by Andrew B. Godbehere, Akihiro Matsukawa, Ken Goldberg in 2012. As per the paper, the system ran a successful interactive audio art installation called "Are We There Yet?" from March 31 - July 31 2011 at the Contemporary Jewish Museum in San Francisco, California.
|
||||||
|
|
||||||
It uses first few (120 by default) frames for background modelling. It employs probabilistic foreground segmentation algorithm that identifies possible foreground objects using Bayesian inference. The estimates are adaptive; newer observations are more heavily weighted than old observations to accommodate variable illumination. Several morphological filtering operations like closing and opening are done to remove unwanted noise. You will get a black window during first few frames.
|
It uses first few (120 by default) frames for background modelling. It employs probabilistic foreground segmentation algorithm that identifies possible foreground objects using Bayesian inference. The estimates are adaptive; newer observations are more heavily weighted than old observations to accommodate variable illumination. Several morphological filtering operations like closing and opening are done to remove unwanted noise. You will get a black window during first few frames.
|
||||||
|
|
||||||
References
|
References
|
||||||
----------
|
----------
|
||||||
[1]: A.B. Godbehere, A. Matsukawa, K. Goldberg. Visual tracking of human visitors under variable-lighting conditions for a responsive audio art installation. American Control Conference. (2012), pp. 4305–4312
|
[1]: A.B. Godbehere, A. Matsukawa, K. Goldberg. Visual tracking of human visitors under variable-lighting conditions for a responsive audio art installation. American Control Conference. (2012), pp. 4305–4312
|
||||||
|
@@ -485,7 +485,7 @@ Implements loading dataset:
|
|||||||
|
|
||||||
"VOT 2015 dataset comprises 60 short sequences showing various objects in challenging backgrounds.
|
"VOT 2015 dataset comprises 60 short sequences showing various objects in challenging backgrounds.
|
||||||
The sequences were chosen from a large pool of sequences including the ALOV dataset, OTB2 dataset,
|
The sequences were chosen from a large pool of sequences including the ALOV dataset, OTB2 dataset,
|
||||||
non-tracking datasets, Computer Vision Online, Professor Bob Fisher’s Image Database, Videezy,
|
non-tracking datasets, Computer Vision Online, Professor Bob Fisher's Image Database, Videezy,
|
||||||
Center for Research in Computer Vision, University of Central Florida, USA, NYU Center for Genomics
|
Center for Research in Computer Vision, University of Central Florida, USA, NYU Center for Genomics
|
||||||
and Systems Biology, Data Wrangling, Open Access Directory and Learning and Recognition in Vision
|
and Systems Biology, Data Wrangling, Open Access Directory and Learning and Recognition in Vision
|
||||||
Group, INRIA, France. The VOT sequence selection protocol was applied to obtain a representative
|
Group, INRIA, France. The VOT sequence selection protocol was applied to obtain a representative
|
||||||
|
@@ -70,7 +70,7 @@ which is available since the 2.4 release. I suggest you take a look at its descr
|
|||||||
|
|
||||||
Algorithm provides the following features for all derived classes:
|
Algorithm provides the following features for all derived classes:
|
||||||
|
|
||||||
- So called “virtual constructor”. That is, each Algorithm derivative is registered at program
|
- So called "virtual constructor". That is, each Algorithm derivative is registered at program
|
||||||
start and you can get the list of registered algorithms and create instance of a particular
|
start and you can get the list of registered algorithms and create instance of a particular
|
||||||
algorithm by its name (see Algorithm::create). If you plan to add your own algorithms, it is
|
algorithm by its name (see Algorithm::create). If you plan to add your own algorithms, it is
|
||||||
good practice to add a unique prefix to your algorithms to distinguish them from other
|
good practice to add a unique prefix to your algorithms to distinguish them from other
|
||||||
|
@@ -52,7 +52,7 @@
|
|||||||
}
|
}
|
||||||
|
|
||||||
@incollection{IPMU2012,
|
@incollection{IPMU2012,
|
||||||
title={$F^1$-transform edge detector inspired by canny’s algorithm},
|
title={$F^1$-transform edge detector inspired by canny's algorithm},
|
||||||
author={Perfilieva, Irina and Hod'{\'a}kov{\'a}, Petra and Hurtík, Petr},
|
author={Perfilieva, Irina and Hod'{\'a}kov{\'a}, Petra and Hurtík, Petr},
|
||||||
booktitle={Advances on Computational Intelligence},
|
booktitle={Advances on Computational Intelligence},
|
||||||
pages={230--239},
|
pages={230--239},
|
||||||
@@ -75,4 +75,4 @@
|
|||||||
pages={235--240},
|
pages={235--240},
|
||||||
year={2015},
|
year={2015},
|
||||||
organization={IEEE}
|
organization={IEEE}
|
||||||
}
|
}
|
||||||
|
@@ -93,7 +93,7 @@ class CV_EXPORTS_W StaticSaliency : public virtual Saliency
|
|||||||
targets, a segmentation by clustering is performed, using *K-means algorithm*. Then, to gain a
|
targets, a segmentation by clustering is performed, using *K-means algorithm*. Then, to gain a
|
||||||
binary representation of clustered saliency map, since values of the map can vary according to
|
binary representation of clustered saliency map, since values of the map can vary according to
|
||||||
the characteristics of frame under analysis, it is not convenient to use a fixed threshold. So,
|
the characteristics of frame under analysis, it is not convenient to use a fixed threshold. So,
|
||||||
*Otsu’s algorithm* is used, which assumes that the image to be thresholded contains two classes
|
*Otsu's algorithm* is used, which assumes that the image to be thresholded contains two classes
|
||||||
of pixels or bi-modal histograms (e.g. foreground and back-ground pixels); later on, the
|
of pixels or bi-modal histograms (e.g. foreground and back-ground pixels); later on, the
|
||||||
algorithm calculates the optimal threshold separating those two classes, so that their
|
algorithm calculates the optimal threshold separating those two classes, so that their
|
||||||
intra-class variance is minimal.
|
intra-class variance is minimal.
|
||||||
|
@@ -77,7 +77,7 @@ void FindCandidateMatches(const FeatureSet &left,
|
|||||||
// method.
|
// method.
|
||||||
// I.E: A match is considered as strong if the following test is true :
|
// I.E: A match is considered as strong if the following test is true :
|
||||||
// I.E distance[0] < fRatio * distances[1].
|
// I.E distance[0] < fRatio * distances[1].
|
||||||
// From David Lowe “Distinctive Image Features from Scale-Invariant Keypoints”.
|
// From David Lowe "Distinctive Image Features from Scale-Invariant Keypoints".
|
||||||
// You can use David Lowe's magic ratio (0.6 or 0.8).
|
// You can use David Lowe's magic ratio (0.6 or 0.8).
|
||||||
// 0.8 allow to remove 90% of the false matches while discarding less than 5%
|
// 0.8 allow to remove 90% of the false matches while discarding less than 5%
|
||||||
// of the correct matches.
|
// of the correct matches.
|
||||||
|
@@ -137,7 +137,7 @@ class CV_EXPORTS_W GrayCodePattern : public StructuredLightPattern
|
|||||||
* @param patternImages The pattern images acquired by the camera, stored in a grayscale vector < Mat >.
|
* @param patternImages The pattern images acquired by the camera, stored in a grayscale vector < Mat >.
|
||||||
* @param x x coordinate of the image pixel.
|
* @param x x coordinate of the image pixel.
|
||||||
* @param y y coordinate of the image pixel.
|
* @param y y coordinate of the image pixel.
|
||||||
* @param projPix Projector's pixel corresponding to the camera's pixel: projPix.x and projPix.y are the image coordinates of the projector’s pixel corresponding to the pixel being decoded in a camera.
|
* @param projPix Projector's pixel corresponding to the camera's pixel: projPix.x and projPix.y are the image coordinates of the projector's pixel corresponding to the pixel being decoded in a camera.
|
||||||
*/
|
*/
|
||||||
CV_WRAP
|
CV_WRAP
|
||||||
virtual bool getProjPixel( InputArrayOfArrays patternImages, int x, int y, Point &projPix ) const = 0;
|
virtual bool getProjPixel( InputArrayOfArrays patternImages, int x, int y, Point &projPix ) const = 0;
|
||||||
@@ -146,4 +146,4 @@ class CV_EXPORTS_W GrayCodePattern : public StructuredLightPattern
|
|||||||
//! @}
|
//! @}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
@@ -53,7 +53,7 @@ namespace structured_light {
|
|||||||
// other algorithms can be implemented
|
// other algorithms can be implemented
|
||||||
enum
|
enum
|
||||||
{
|
{
|
||||||
DECODE_3D_UNDERWORLD = 0 //!< Kyriakos Herakleous, Charalambos Poullis. “3DUNDERWORLD-SLS: An Open-Source Structured-Light Scanning System for Rapid Geometry Acquisition”, arXiv preprint arXiv:1406.6595 (2014).
|
DECODE_3D_UNDERWORLD = 0 //!< Kyriakos Herakleous, Charalambos Poullis. "3DUNDERWORLD-SLS: An Open-Source Structured-Light Scanning System for Rapid Geometry Acquisition", arXiv preprint arXiv:1406.6595 (2014).
|
||||||
};
|
};
|
||||||
|
|
||||||
/** @brief Abstract base class for generating and decoding structured light patterns.
|
/** @brief Abstract base class for generating and decoding structured light patterns.
|
||||||
@@ -88,4 +88,4 @@ class CV_EXPORTS_W StructuredLightPattern : public virtual Algorithm
|
|||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
@@ -5,7 +5,7 @@
|
|||||||
#ifndef __OPENCV_TEXT_TEXTDETECTOR_HPP__
|
#ifndef __OPENCV_TEXT_TEXTDETECTOR_HPP__
|
||||||
#define __OPENCV_TEXT_TEXTDETECTOR_HPP__
|
#define __OPENCV_TEXT_TEXTDETECTOR_HPP__
|
||||||
|
|
||||||
#include"ocr.hpp"
|
#include "ocr.hpp"
|
||||||
|
|
||||||
namespace cv
|
namespace cv
|
||||||
{
|
{
|
||||||
|
@@ -113,4 +113,4 @@ CMAKE_OPTIONS='-DBUILD_PERF_TESTS:BOOL=OFF -DBUILD_TESTS:BOOL=OFF -DBUILD_DOCS:B
|
|||||||
@endcode
|
@endcode
|
||||||
-# now we need the language files from tesseract. either clone https://github.com/tesseract-ocr/tessdata, or copy only those language files you need to a folder (example c:\\lib\\install\\tesseract\\tessdata). If you don't want to add a new folder you must copy language file in same folder than your executable
|
-# now we need the language files from tesseract. either clone https://github.com/tesseract-ocr/tessdata, or copy only those language files you need to a folder (example c:\\lib\\install\\tesseract\\tessdata). If you don't want to add a new folder you must copy language file in same folder than your executable
|
||||||
-# if you created a new folder, then you must add a new variable, TESSDATA_PREFIX with the value c:\\lib\\install\\tessdata to your system's environment
|
-# if you created a new folder, then you must add a new variable, TESSDATA_PREFIX with the value c:\\lib\\install\\tessdata to your system's environment
|
||||||
-# add c:\\Lib\\install\\leptonica\\bin and c:\\Lib\\install\\tesseract\\bin to your PATH environment. If you don't want to modify the PATH then copy tesseract400.dll and leptonica-1.74.4.dll to the same folder than your exe file.
|
-# add c:\\Lib\\install\\leptonica\\bin and c:\\Lib\\install\\tesseract\\bin to your PATH environment. If you don't want to modify the PATH then copy tesseract400.dll and leptonica-1.74.4.dll to the same folder than your exe file.
|
||||||
|
@@ -1171,7 +1171,7 @@ class CV_EXPORTS_W TrackerMedianFlow : public Tracker
|
|||||||
tracking, learning and detection.
|
tracking, learning and detection.
|
||||||
|
|
||||||
The tracker follows the object from frame to frame. The detector localizes all appearances that
|
The tracker follows the object from frame to frame. The detector localizes all appearances that
|
||||||
have been observed so far and corrects the tracker if necessary. The learning estimates detector’s
|
have been observed so far and corrects the tracker if necessary. The learning estimates detector's
|
||||||
errors and updates it to avoid these errors in the future. The implementation is based on @cite TLD .
|
errors and updates it to avoid these errors in the future. The implementation is based on @cite TLD .
|
||||||
|
|
||||||
The Median Flow algorithm (see cv::TrackerMedianFlow) was chosen as a tracking component in this
|
The Median Flow algorithm (see cv::TrackerMedianFlow) was chosen as a tracking component in this
|
||||||
@@ -1435,7 +1435,7 @@ public:
|
|||||||
the long-term tracking task into tracking, learning and detection.
|
the long-term tracking task into tracking, learning and detection.
|
||||||
|
|
||||||
The tracker follows the object from frame to frame. The detector localizes all appearances that
|
The tracker follows the object from frame to frame. The detector localizes all appearances that
|
||||||
have been observed so far and corrects the tracker if necessary. The learning estimates detector’s
|
have been observed so far and corrects the tracker if necessary. The learning estimates detector's
|
||||||
errors and updates it to avoid these errors in the future. The implementation is based on @cite TLD .
|
errors and updates it to avoid these errors in the future. The implementation is based on @cite TLD .
|
||||||
|
|
||||||
The Median Flow algorithm (see cv::TrackerMedianFlow) was chosen as a tracking component in this
|
The Median Flow algorithm (see cv::TrackerMedianFlow) was chosen as a tracking component in this
|
||||||
|
Reference in New Issue
Block a user