|
|
- Grant Wood
- 8 years ago
- Views:
Transcription
1 Fast and Robust Moving Object Segmentation Technique for MPEG-4 Object-based Coding and Functionality Ju Guo, Jongwon Kim and C.-C. Jay Kuo Integrated Media Systems Center and Department of Electrical Engineering-Systems University of Southern California, Los Angeles, CA ABSTRACT Video object segmentation is an important component for object-based video coding schemes such as MPEG-4. A fast and robust video segmentation technique, which aims at ecient foreground and background separation via eective combination of motion and color information, is proposed in this work. First, a non-parametric gradientbased iterative color clustering algorithm, called the mean shift algorithm, is employed to provide robust dominant color regions according to color similarity. With the dominant color information from previous frames as the initial guess for the next frame, the amount of computational time can be reduce to 50%. Next, moving regions are identied by a motion detection method, which is developed based on the frame intensity dierence to circumvent the motion estimation complexity for the whole frame. Only moving regions are further merged or split according to a region-based ane motion model. Furthermore, sizes, colors, and motion information of homogeneous regions are tracked to increase temporal and spatial consistency of extracted objects. The proposed system is evaluated for several typical MPEG-4 test sequences. It provides very consistent and accurate object boundaries throughout the entire test sequences. Keywords: video segmentation, color segmentation, mean shift algorithm, ane motion, spatial segmentation, motion detection. 1 INTRODUCTION An object- or content-based coding scheme plays a signicant role in the success of the second generation video coding [1]. The object-based coding has the potential to provide a more accurate video representation at the very low bit rate video. It also allows content based functionalities such as object manipulation. In the recent development of the MPEG-4 standard [2], video coding is handled by the object unit, namely, the video object plane (VOP). VOP represents one snap shot of an object in video. For each VOP, the motion, texture, and shape information is coded in separate bit streams. This allows separate modication and manipulation of each VOP and supports the content based functionality. Thus, video segmentation, which aims at the exact separation of moving objects from the background, becomes the foundation of content based video coding, among many other interesting applications. Even though the image and video segmentation problem has been studied for more than thirty years, it is still considered one of the most challenging image processing tasks, and demands creative solutions for major breakthrough. Most existing video segmentation algorithms attempt to exploit the temporal and spatial coherence information inherent in the image sequence to achieve foreground/background separation. Temporal segmentation can identify moving objects since most moving objects have distinct motion patterns from the background. Spatial segmentation can determine object boundaries accurately if underlying objects have a dierent visual appearance (such as the color or the gray level intensity) from the background. An ecient combination of spatial-temporal segmentation
2 modules can lead to a more promising solution to the segmentation problem. It is desirable to develop an automatic segmentation algorithm that requires no user assistance and interaction. In addition, the availability of a fast implementation is also one basic requirement, which is especially needed for real time applications. A fast and robust video segmentation technique is proposed in this work. It can be roughly described below. First, a non-parametric gradient-based iterative color clustering algorithm, called the mean shift algorithm, is employed to provide robust dominant color regions according to color similarity. With the dominant color information from previous frames as the initial guess for the next frame, the amount of computational time can be reduce to 50%. Next, moving regions are identied by a motion detection method, which is developed based on the frame intensity dierence to circumvent the motion estimation complexity for the whole frame. Only moving regions are further merged or split according to a region-based ane motion model. Furthermore, sizes, colors, and motion information of homogeneous regions are tracked to increase temporal and spatial consistency of extracted objects. The paper is organized as follows. We rst review previous segmentation work in Section 2. A general description of the proposed segmentation algorithm is given in Section 3. Video segmentation results for MPEG-4 test video are presented in Section 4. They are compared with results of the three algorithms recommended in MPEG-4. Concluding remarks are given in Section Three MPEG-4 Algorithms 2 Review of Previous Work Up to now, three algorithms for automatic video segmentation have been proposed in MPEG-4 visual standard [3]. They are temporal segmentation from Fondazione Ugo Bordoni (FUB), temporal segmentation from University of Hannover (UH), and spatial-temporal segmentation from Electronics and Telecommunications Research Institute (ETRI). All of these algorithms classify pixels in an image sequence into two classes, i.e. moving objects (foreground) and background. In the algorithm proposed by ETRI, images are rst simplied by morphological lters. These lters remove regions that are smaller than a given size but preserve contours of remaining objects. The morphological gradient operator is used to estimate the luminance gradient. The region boundary is obtained by the watershed algorithm, where the similarity measure is obtained from the combination of the luminance gradient and the motion eld. Finally, similar regions are merged based on graph theory. The algorithm of UH uses two successive frame dierences to obtain a change detection mask. The uncovered background is removed by a hierarchical block matcher. The region boundaries of the change detection mask are adapted to luminance edges to improve the segmentation accuracy with respect to object boundaries. For the algorithm from FUB, a group of frames is rst selected, and dierences of each frame in the group with respect to the rst frame are evaluated. A robust fourth-order statistic test of frame dierences is performed to detect the change areas. Motion is estimated to remove the uncovered background area. Morphological open and close operators are used to rene region boundaries. In the automatic video segmentation framework, statistic change detection and motion estimation are used in the temporal domain, while the luminance based morphological operation and the watershed algorithm can be used to segment objects within an image in the space domain [3], [4]. Due to the complexity of segmentation modules, these algorithms are not suitable for the real-time implementation. Also, the segmentation result is still unsatisfactory for several typical videos, where the shape of the object can not be precisely dened. 2.2 Integrated Spatial-Temporal Approach The integration of temporal- and spatial-segmentation results can improve the performance with an increased complexity. Bouthemy and Francois [5] proposed a technique to simultaneously estimate the spatial-temporal segmentation and motion by adopting a Markov Random Field (MRF) model and Bayesian estimation. Since the MRF model is constructed in terms of local constraints on the luminance intensity and motion, the spatial information and motion can be taken into account simultaneously, and motion estimation and segmentation can be optimized jointly. However, the disadvantage of this technique is its high complexity. The spatial and temporal segmentation steps can be iteratively performed to reduce the complexity [6]. Next, since the human visual system (HVS) is very sensitive totheedgeandcontour information, exact extraction of object boundaries is crucial for
3 visual quality of segmented results. More visual information should be used to make spatial segmentation robust and consistent. 2.3 Motivation and Summary of the Proposed Approach Among many visual cues, the color information has not yet been fully exploited in video segmentation, since it is often perceived that human eyes are not too sensitive to the chrominance components, e.g. the UV data in the YUV-format video and the contribution from the color information is treated as the second order eect. Furthermore, additional computational complexities are required for color processing. We believe that the color information does play an important role in object identication and recognition in the human visual system (HVS), and it is worthwhile to include the information in the computation. Zhong and Chang [7] applied color segmentation to separate images into homogeneous regions, and tracked them along time for content-based video query. A simple uniform quantization in the L? u? v? color space was used in their work. Kanai used the uniform quantization in the HSV color space for the image segmentation [8]. The uniform color quantization was adopted in both work to reduce the complexity of segmentation. In this work, we focus on automatic video segmentation by proposing a fast and adaptive algorithm with a reduced complexity in both spatial and temporal domains. A fast yet robust adaptive color segmentation based on the mean shift color clustering algorithm is applied in the spatial domain. The mean shift algorithm has been generalized by Cheng[9] for clustering data, and used by Comaniciu and Meer for color segmentation [10]. For the k-means clustering method, it is dicult to choose the initial number of classes. By using the mean shift algorithm, the number of dominant colors can be determined automatically. Here, we develop a non-parametric gradient-based algorithm that provides a simple iterative method to determine the local density maximum. The number of color classes in the current frame can be used as the initial guess of color classes for the next frame. This helps in reducing the computational complexity of color segmentation. For the temporal domain, a noise robust higher order statistic motion detection algorithm and a color region based ane motion model is employed. After separating an image frame into homogeneous spatial regions, we determine whether each region belongs to the background or the foreground by motion detection. Only moving regions are further merged or split using the region based ane motion model. The six parameters of the ane motion model are estimated for each region. Regions with similar motion parameters are merged. Regions that do not t to the ane motion model well are split. The size, color, and motion information of each region is tracked to increase the consistency of extracted objects. The system is applied to segment several MPEG-4 test video clips. We have observed accurate object boundaries and the temporal and spatial consistency from experimental results. 3 PROPOSED VIDEO SEGMENTATION ALGORITHM The block diagram of the proposed automatic video segmentation algorithm is given in Fig. 1. It consists of 4stages. At the rst stage, the global motion compensation (GMC) procedure is performed. That is, the global motion of image sequences is estimated using the six-parameter ane model. With this information, images can be aligned accordingly. At the second stage, the mean shift color segmentation algorithm is used to partition an image into homogeneous regions. This is basically a spatial domain segmentation approach. At the third stage, we attempt to use temporal information for segmentation. A statistical motion detection approach is used to determine whether each homogeneous region is moving or not. Only for moving regions, we apply the ane motion model for motion parameter estimation. Regions can be merged and split according to the consistency of their motion parameters. The size, color and motion data of nal segmented regions are tracked. At the last stage, the morphological open and close lters are used to smooth object boundaries and eliminate small regions. These building blocks of the proposed algorithm are detailed below. 3.1 Global Motion Compensation A six-parameter ane motion model is used to estimate the global motion, which is due to camera movement such as panning, zooming, and rotation. Motion vectors are estimated based on block matching of pixel block. Motion vectors with a large matching residual are rejected. Six parameters of the ane motion model are
4 Figure 1: The block diagram of the proposed automatic video segmentation algorithm. estimated using the macroblock based motion vectors. Once the global motion is detected and estimated, multiple frames are aligned accordingly to reduce the eect of the camera movement. 3.2 Color Segmentation The intensity distribution of each color component can be viewed as a probability density function. The mean shift vector is the dierence between the mean of the probability function onalocalarea and the center of this region. In terms of mathematics, the mean shift vector associated with a region S ~x centered on ~x can be written as: ~V (~x) = R ~y2s p(~y)(~y ; ~x)d~y R ~x ~y2s p(~y)d~y ~x where p() is the probability density function. The mean shift algorithm says that the mean shift vector is proportional to the gradient of the probability density rp(~x), and reciprocal to the probability density p(~x), i.e. ~V (~x) =c rp(~x) p(~x) where c is a constant. Since the mean shift vector is along the direction of the probability density maximum, we can exploit this property to nd the actual location of the density maximum. In implementing the mean shift algorithm, the size of the search window can be made adaptive to an image by setting the radius proportional to
5 the trace of the global covariance matrix of the given image. By moving search windows in the color space using the mean shift vector iteratively, one dominant color can be located. After removing all colors inside the converged search window, one can repeat the mean shift algorithm again to locate the second dominant color. This process can be repeated several times to identify a few major dominant colors. The uniform color space L? u? v? was used by Comaniciu et al. [10] for color segmentation due to its perceptual homogeneity. To reduce the computational complexity, we use the YUV space for color segmentation since original video data are stored in the YUV format. The obtained results are comparable with those based in the L? u? v? space. Dominant colors of the current frame are used as the initial guess of dominant colors in the next frame. Due to the similarity of adjacent frames, the mean shift algorithm often converges in one or two iterations, thus reducing the computational time signicantly. Color segmentation also uses the spatial relation of pixels as a constraint [10] as described below. For each frame, dominant colors are rst generated by the mean shift algorithm. Then, all pixels are classied according to their distance to dominant colors. A relative small distance is used as a threshold to determine which classes the pixel is belong to in the beginning. Afterwards, the threshold is doubled. Only the pixel that has a smaller distance to the dominant color and has one of its neighboring pixels assigned to the same class can be classied to this class. Finally, unassigned pixels are classied to its nearest neighboring region. 3.3 Motion Detection and Estimation A robust motion detection method based on the frame dierence calculation is used to determine whether homogeneous regions are moving or not [3]. Since the statistical behavior of inter-frame dierences produced by object movement strongly deviates from the Gaussian model, a fourth-order statistic adaptive detection of the non-gaussian signal is performed. For each pixel at (x y), its fourth order moments ^m d (x y) isevaluated as ^m d (x y) = 1 9 X (s t)2w(x y) (d(s t) ; ^(x y)) 4 where d(x y) is the inter-frame dierence, W (x y) is 3 3 window centered at (x y) and ^(x y) is the sample mean of d(x y) inside window W (x y), i.e. ^(x y) = 1 9 X d(s t): (s t)2w(x y) Each pixel at (x y) is determined to be associated with the still background or the change region according to its fourth moment ^m d (x y). The change regions obtained from the higher statistic estimation include the uncovered background. The block matching algorithm is applied to the fourth order moment maps of frame dierences in order to remove the uncovered background. Pixels that have null displacements are reassigned to the background. For each homogeneous region, if 85% of pixels are identied as moving pixels, the region is identiedasmoving. Only for moving regions, the motion vector eld is estimated by using hierarchical block matching methods inside the regions. Each motion vector is labeled with a condent factor based on the matching mean square error. The ane motion model is described as: (x y) (x y) = a1 + a2x + a3y a4 + a5x + a6y where (x y) and(x y) are the motion vectors at horizontal and vertical directions, a1 a6 are constant parameters. The 6 parameters (a1 a6) of the ane model are estimated by using motion vectors with high condent factors. The obtained parameters are tested for the whole region by calculating the mean square error E dened as E R = 1 N X (x y)2r [(x y) ; ^(x y)] 2 +[(x y) ; ^(x y)] 2
6 where N is the number of pixels inside region R, ^(x y) and^(x y) are the estimated motion vectors by the ane model. When the mean square error E R for the region R is above a certain threshold, it implies that the ane motion model cannot describe the motion of that region well so that the region should be split according to the motion information. Thus, the mean square error E will be reduced. Regions with similar motion and color are merged together. The motion similarity measure is determined on 6 parameter space, which is 6X S(R1 R2) = (a R1 i i=1 ; a R2 i ) 2 where a R1 i and a R2 i are the ane model parameters for region R1 and R2, respectively. If S(R1 R2) is small, the region R1 and R2 are merged together. Moving objects are projected to the next frame according to their ane motion models. The projected region boundaries are aligned to the current region boundaries by region matching. For unmatched regions, the change of detection is used to nd moving regions. For each new moving region, we repeat the process of motion estimation, region splitting and merging. This process allows the detection of newly appeared objects in the scene. 3.4 Postprocessing The object masks obtained from the spatial and temporal segmentation sometimes have irregularity in the boundaries, such small gulfs, or isthmi, due to temporal and spatial signal uctuations. This will give visually annoying appearance and also increase the shape coding cost. We use the morphological open and close operators to remove the gulfs and isthmi, and to smooth the object boundaries to increase the shape coding eciency. A circular structuring element with 2 pixel radius is used in the morphological open and close operation. 4 EXPERIMENTAL RESULT Two MPEG-4 QCIF sequences, i.e. \Akiyo" and \Mother and daughter", are used to test the proposed algorithm. For the \Akiyo" sequence, there is only a small motion activity in the head and shoulder regions. The original 10th and 20th image frames are shown in Fig. 2(a). The results of color segmentation are given in Fig. 2(b). We can clearly see that each image is segmented into a few regions. For example, Akiyo is segmented into the hair region, the facial region, and the shoulder region. Each region has a well-aligned boundary corresponding to the real object. The motion detection algorithm identies the moving region, which is given in Fig. 2(c). The boundary is not well detected as compared with the real object boundary by using the motion information only. By incorporating the spatial color segmentation result, the nal segmentation result is much improved as shown in Fig. 2(d). For the \Mother and daughter" sequence, there are more head and hand motion activities than \Akiyo". The results of color segmentation is shown in Fig. 3(b), for two dierent frames (i.e. the 20th and 250th frames). More regions are obtained from color segmentation. All these regions are identied as belonging to either the background or the foreground. Regions, such as mother's head and shoulder, daughter's hair, shoulder and face, have contours which correspond to real objects. These objects, identied by motion detection and dened by color regions, were accurately segmented from the background as given in Fig. 3(d). Although many segmentation algorithms have been proposed, it is still a very dicult problem to evaluate the quality of the generated video objects. In MPEG-4, only subjective evaluation by tape viewing was adopted to decide the quality of segmentation results. It is desirable to use to an objective measure by comparing the segmented object with the reference object. Two criteria, i.e. spatial accuracy and temporal coherency of the video object, are important measure for the quality of a certain algorithm. Recently, Wollborn and Mech [11] proposed a simple pixel-based quality measure. The spatial distortion of an estimated binary video object mask at frame t is dened as d(a est t A ref t )= P(x y) Aest t (x y) A ref t (x y) P(x y) Aref t (x y)
7 (a) (b) (c) (d) Figure 2: The segmentation results of the \Akiyo" QCIF sequence with respect to the 10th and the 20th frames: (a) the original images, (b) the color segmentation results, (c) the motion detection results and (d) the nal results.
8 (a) (b) (c) (d) Figure 3: The segmentation results of the \Mother and daughter" QCIF sequence with respect to the 20th and the 250th frames: (a) the original images, (b) the color segmentation results, (c) the motion detection results and (d) the nal results.
9 where A ref t and A est t are the reference binary object mask and the estimated one at frame t, respectively, and is the binary \xor" operation. Temporal coherency is measured by (t) =d(a t A t;1) where A t and A t;1 are the binary mask at frame t and t ; 1, respectively. Temporal coherency est (t) for the estimated binary mask A est should be compared to temporal coherency ref (t) of the reference mask. Any signicant deviation from the reference indicates a bad temporal coherency. The segmentation results of this paper are evaluated using both criteria. The results of "Akiyo" QCIF sequence are shown in Fig. 4(a) and Fig 4(b). For the reference mask, the hand segmented mask from the MPEG-4 test material distribution is utilized. In Fig. 4(a), the dot line is obtained by using higher statistic motion detection only and the solid line is the proposed scheme. We can see that the spatial accuracy is much improved by using the color segmentation algorithm. The error is less than 2% in most frames. In Fig. 4(b), the solid line denotes the reference mask, the dot line the proposed scheme, and the dash line the motion detection using the high order statistic method only. The temporal coherency curve also closely follows the one in the reference mask. Since the reference object mask for the "Mother and Daughter" QCIF sequence is not available, only the temporal coherency is evaluated and plotted in g. 5. With the proposed color segmentation, the curve of temporal coherency is much smoother than the one with only motion detection, which demonstrates a better performance Spatial Accuracy Temporal Coherency Frame No. (a) Frame No. (b) Figure 4: The objective evaluation of the "Akiyo" QCIF sequence object mask for (a) spatial accuracy and (b) temporal coherency. We can see from these results that temporal segmentation can identify moving regions while spatial segmentation provides the important information of object boundaries. Our algorithm exploits spatial information of color similarity and obtain the accurate region boundary automatically. Since the human visual system is very sensitive to edge information, our segmentation results provide better visual quality than those in MPEG-4 due to the more accurate boundary information (see Fig. 6 and Fig. 7). 5 CONCLUSION A new video segmentation algorithm for the MPEG-4 object based coding was proposed in this work. The proposed segmentation scheme can lead to fast object segmentation. Also, the color segmentation combined with the region-based motion detection gives a very accurate video segmentation result. The performance of the proposed segmentation scheme was demonstrated via several experimental results. 6 REFERENCES [1] L. Torres and M. Kunt, Video Coding (The Second Generation Approach), Kluwer Academic Publishers, 1996.
10 Temporal Coherency Frame No. Figure 5: The objective evaluation of the "Mother and Daughter" QCIF sequence object mask by temporal coherency, where the solid line denotes the result of the proposed method while the dot line uses only motion detection with higher order statistics. (a) (b) (c) (d) Figure 6: Performance comparison of MPEG-4 automatic segmentation methods and the proposed method for the \Akiyo" QCIF sequence: (a) MPEG-4 FUB algorithm [3], (b) MPEG-4 ETRI algorithm [3], (c) MPEG-4 UH algorithm [3] and (d) the proposed algorithm.
11 (a) (b) (c) (d) Figure 7: Performance comparison of MPEG-4 automatic segmentation methods and the proposed method for the \Mother and daughter" QCIF sequence: (a) MPEG-4 FUB algorithm [3], (b) MPEG-4 ETRI algorithm [3], (c) MPEG-4 UH algorithm [3] and (d) the proposed algorithm.
12 [2] "Information Technology - Coding of Audio-Visual Objects:Visual", Doc.ISO/IEC Final Committee Draft, May [3] J. Ohm, Ed.,"Core experiments on multifunctional and advanced layered coding aspects of MPEG-4 video", Doc. ISO/IEC JTC1/SC29/WG11 N2176, May [4] C. Gu and M.G. Lee, \Semantic video object segmentation and tracking using mathematical morphology and perspective motion model," IEEE International Conference on Image Processing, Santa Barbara, CA, Oct [5] P. Bouthemy and E. Francois, \Motion segmentation and qualitative dynamic scene analysis from an image sequence", Int. Journal of Computer Vision, Vol. 10, pp , [6] F. Dufaux and F. Moscheni, \Spatio-temporal segmentation based on motion and static segmentation", IEEE International Conference on Image Processing, Washington, Oct [7] D. Zhong and S.F. Chang, \Video object model and segmentation for content-based video indexing," in IEEE International Symposium on Circuits and Systems, Hong Kong, June [8] Y. Kanai, \Image segmentation using intensity and color information," in Visual Communications and Image Processing, Jan [9] Y. Cheng, \Mean shift, mode seeking, and clustering," in IEEE Trans. Pattern Anal. Machine Intell., Vol.17, pp , [10] D. Comaniciu and P. Meer, \Robust analysis of feature space: color image segmentation,\ in Computer Vision and Pattern Recognition, San Juan, Puerto Rico, June, [11] M. Wollborn and R. Mech, \Rened procedure for objective evaluation of video object generation algorithms," Doc. ISO/IEC JTC1/SC29/WG11 M3448, March, 1998.
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
More informationEdge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
More informationTemplate-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
More informationBandwidth Adaptation for MPEG-4 Video Streaming over the Internet
DICTA2002: Digital Image Computing Techniques and Applications, 21--22 January 2002, Melbourne, Australia Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet K. Ramkishor James. P. Mammen
More informationColour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
More informationVision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu gramos@clemson.edu dshu@clemson.edu Abstract A vehicle tracking and grouping algorithm is presented in this work
More informationA Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 emre.ugur@ceng.metu.edu.tr Abstract The main objective of this project is the study of a learning based method
More informationPeter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany
RATE-DISTORTION-EFFICIENT VIDEO COMPRESSION USING A 3-D HEAD MODEL Peter Eisert, Thomas Wiegand and Bernd Girod Telecommunications Laboratory University of Erlangen-Nuremberg Cauerstrasse 7, 91058 Erlangen,
More informationDIGITAL video is an integral part of many newly emerging
782 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 14, NO. 6, JUNE 2004 Video Object Segmentation Using Bayes-Based Temporal Tracking and Trajectory-Based Region Merging Vasileios
More informationSegmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
More informationReal Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
More informationSpeed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA sungchun@usc.edu, nevatia@usc.edu Abstract. A speed
More informationFace Locating and Tracking for Human{Computer Interaction. Carnegie Mellon University. Pittsburgh, PA 15213
Face Locating and Tracking for Human{Computer Interaction Martin Hunke Alex Waibel School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Abstract Eective Human-to-Human communication
More informationTracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
More informationDetermining optimal window size for texture feature extraction methods
IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec
More informationA Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
More informationCOMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
More informationFeature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
More informationLABELING IN VIDEO DATABASES. School of Electrical and Computer Engineering. 1285 Electrical Engineering Building. in the video database.
FACE DETECTION FOR PSEUDO-SEMANTIC LABELING IN VIDEO DATABASES Alberto Albiol Departamento de Comunicaciones Universidad Politçecnica de Valencia Valencia, Spain email: alalbiol@dcom.upv.es Charles A.
More informationPerformance Analysis and Comparison of JM 15.1 and Intel IPP H.264 Encoder and Decoder
Performance Analysis and Comparison of 15.1 and H.264 Encoder and Decoder K.V.Suchethan Swaroop and K.R.Rao, IEEE Fellow Department of Electrical Engineering, University of Texas at Arlington Arlington,
More information1 Example of Time Series Analysis by SSA 1
1 Example of Time Series Analysis by SSA 1 Let us illustrate the 'Caterpillar'-SSA technique [1] by the example of time series analysis. Consider the time series FORT (monthly volumes of fortied wine sales
More informationCloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool
Cloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool Joseph C. Tsai Foundation of Computer Science Lab. The University of Aizu Fukushima-ken, Japan jctsai@u-aizu.ac.jp Abstract Video
More informationThe Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
More informationAnalecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
More informationTracking People in 2 1 2 D. Ismail Haritaoglu, David Harwood and Larry S. Davis. University of Maryland. College Park, MD 20742, USA
W 4 S: A Real-Time System for Detecting and Tracking People in 2 1 2 D Ismail Haritaoglu, David Harwood and Larry S. Davis Computer Vision Laboratory University of Maryland College Park, MD 20742, USA
More informationParametric Comparison of H.264 with Existing Video Standards
Parametric Comparison of H.264 with Existing Video Standards Sumit Bhardwaj Department of Electronics and Communication Engineering Amity School of Engineering, Noida, Uttar Pradesh,INDIA Jyoti Bhardwaj
More informationWATERMARKING FOR IMAGE AUTHENTICATION
WATERMARKING FOR IMAGE AUTHENTICATION Min Wu Bede Liu Department of Electrical Engineering Princeton University, Princeton, NJ 08544, USA Fax: +1-609-258-3745 {minwu, liu}@ee.princeton.edu ABSTRACT A data
More informationA Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
More informationA System for Capturing High Resolution Images
A System for Capturing High Resolution Images G.Voyatzis, G.Angelopoulos, A.Bors and I.Pitas Department of Informatics University of Thessaloniki BOX 451, 54006 Thessaloniki GREECE e-mail: pitas@zeus.csd.auth.gr
More informationHANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT
International Journal of Scientific and Research Publications, Volume 2, Issue 4, April 2012 1 HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT Akhil Gupta, Akash Rathi, Dr. Y. Radhika
More informationSignature Segmentation from Machine Printed Documents using Conditional Random Field
2011 International Conference on Document Analysis and Recognition Signature Segmentation from Machine Printed Documents using Conditional Random Field Ranju Mandal Computer Vision and Pattern Recognition
More informationMean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
More informationMedical Image Segmentation of PACS System Image Post-processing *
Medical Image Segmentation of PACS System Image Post-processing * Lv Jie, Xiong Chun-rong, and Xie Miao Department of Professional Technical Institute, Yulin Normal University, Yulin Guangxi 537000, China
More informationAutomatic parameter regulation for a tracking system with an auto-critical function
Automatic parameter regulation for a tracking system with an auto-critical function Daniela Hall INRIA Rhône-Alpes, St. Ismier, France Email: Daniela.Hall@inrialpes.fr Abstract In this article we propose
More informationIntroduction to Logistic Regression
OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction
More informationTracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey mustafa.teke@gmail.com b Department of Computer
More informationANIMATION a system for animation scene and contents creation, retrieval and display
ANIMATION a system for animation scene and contents creation, retrieval and display Peter L. Stanchev Kettering University ABSTRACT There is an increasing interest in the computer animation. The most of
More informationA Method of Caption Detection in News Video
3rd International Conference on Multimedia Technology(ICMT 3) A Method of Caption Detection in News Video He HUANG, Ping SHI Abstract. News video is one of the most important media for people to get information.
More informationAutomatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 jkiske@stanford.edu 1. Introduction As autonomous vehicles become more popular,
More informationA Dynamic Approach to Extract Texts and Captions from Videos
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
More informationGraphic Design. Background: The part of an artwork that appears to be farthest from the viewer, or in the distance of the scene.
Graphic Design Active Layer- When you create multi layers for your images the active layer, or the only one that will be affected by your actions, is the one with a blue background in your layers palette.
More informationAssessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
More informationSegmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
More informationDetection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
More informationAutomatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Kenji Yamashiro, Daisuke Deguchi, Tomokazu Takahashi,2, Ichiro Ide, Hiroshi Murase, Kazunori Higuchi 3,
More informationCircle detection and tracking speed-up based on change-driven image processing
Circle detection and tracking speed-up based on change-driven image processing Fernando Pardo, Jose A. Boluda, Julio C. Sosa Departamento de Informática, Universidad de Valencia Avda. Vicente Andres Estelles
More informationRobert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
More informationVideo compression: Performance of available codec software
Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes
More informationIntra-Prediction Mode Decision for H.264 in Two Steps Song-Hak Ri, Joern Ostermann
Intra-Prediction Mode Decision for H.264 in Two Steps Song-Hak Ri, Joern Ostermann Institut für Informationsverarbeitung, University of Hannover Appelstr 9a, D-30167 Hannover, Germany Abstract. Two fast
More informationEnhanced LIC Pencil Filter
Enhanced LIC Pencil Filter Shigefumi Yamamoto, Xiaoyang Mao, Kenji Tanii, Atsumi Imamiya University of Yamanashi {daisy@media.yamanashi.ac.jp, mao@media.yamanashi.ac.jp, imamiya@media.yamanashi.ac.jp}
More informationModule 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur
Module 8 VIDEO CODING STANDARDS Version ECE IIT, Kharagpur Lesson H. andh.3 Standards Version ECE IIT, Kharagpur Lesson Objectives At the end of this lesson the students should be able to :. State the
More informationVideo Affective Content Recognition Based on Genetic Algorithm Combined HMM
Video Affective Content Recognition Based on Genetic Algorithm Combined HMM Kai Sun and Junqing Yu Computer College of Science & Technology, Huazhong University of Science & Technology, Wuhan 430074, China
More informationText Localization & Segmentation in Images, Web Pages and Videos Media Mining I
Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I Multimedia Computing, Universität Augsburg Rainer.Lienhart@informatik.uni-augsburg.de www.multimedia-computing.{de,org} PSNR_Y
More informationA ROBUST BACKGROUND REMOVAL ALGORTIHMS
A ROBUST BACKGROUND REMOVAL ALGORTIHMS USING FUZZY C-MEANS CLUSTERING ABSTRACT S.Lakshmi 1 and Dr.V.Sankaranarayanan 2 1 Jeppiaar Engineering College, Chennai lakshmi1503@gmail.com 2 Director, Crescent
More informationHow To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
More informationClassification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
More informationEXTRACTION OF UNCONSTRAINED CAPTION TEXT FROM GENERAL-PURPOSE VIDEO
The Pennsylvania State University The Graduate School Department of Computer Science and Engineering EXTRACTION OF UNCONSTRAINED CAPTION TEXT FROM GENERAL-PURPOSE VIDEO A Thesis in Computer Science and
More informationUNIVERSITY OF CENTRAL FLORIDA AT TRECVID 2003. Yun Zhai, Zeeshan Rasheed, Mubarak Shah
UNIVERSITY OF CENTRAL FLORIDA AT TRECVID 2003 Yun Zhai, Zeeshan Rasheed, Mubarak Shah Computer Vision Laboratory School of Computer Science University of Central Florida, Orlando, Florida ABSTRACT In this
More informationA Memory Reduction Method in Pricing American Options Raymond H. Chan Yong Chen y K. M. Yeung z Abstract This paper concerns with the pricing of American options by simulation methods. In the traditional
More informationWhat is Visualization? Information Visualization An Overview. Information Visualization. Definitions
What is Visualization? Information Visualization An Overview Jonathan I. Maletic, Ph.D. Computer Science Kent State University Visualize/Visualization: To form a mental image or vision of [some
More informationH 261. Video Compression 1: H 261 Multimedia Systems (Module 4 Lesson 2) H 261 Coding Basics. Sources: Summary:
Video Compression : 6 Multimedia Systems (Module Lesson ) Summary: 6 Coding Compress color motion video into a low-rate bit stream at following resolutions: QCIF (76 x ) CIF ( x 88) Inter and Intra Frame
More informationA Real Time Hand Tracking System for Interactive Applications
A Real Time Hand Tracking System for Interactive Applications Siddharth Swarup Rautaray Indian Institute of Information Technology Allahabad ABSTRACT In vision based hand tracking systems color plays an
More informationEnvironmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
More informationPractical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
More informationNorbert Schuff Professor of Radiology VA Medical Center and UCSF Norbert.schuff@ucsf.edu
Norbert Schuff Professor of Radiology Medical Center and UCSF Norbert.schuff@ucsf.edu Medical Imaging Informatics 2012, N.Schuff Course # 170.03 Slide 1/67 Overview Definitions Role of Segmentation Segmentation
More informationBig Data: Image & Video Analytics
Big Data: Image & Video Analytics How it could support Archiving & Indexing & Searching Dieter Haas, IBM Deutschland GmbH The Big Data Wave 60% of internet traffic is multimedia content (images and videos)
More informationSimultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong a28wong@uwaterloo.ca William Bishop wdbishop@uwaterloo.ca Department of Electrical and Computer Engineering University
More informationCloud tracking with optical flow for short-term solar forecasting
Cloud tracking with optical flow for short-term solar forecasting Philip Wood-Bradley, José Zapata, John Pye Solar Thermal Group, Australian National University, Canberra, Australia Corresponding author:
More informationVery Low Frame-Rate Video Streaming For Face-to-Face Teleconference
Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Jue Wang, Michael F. Cohen Department of Electrical Engineering, University of Washington Microsoft Research Abstract Providing the best
More informationA HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER. Figure 1. Basic structure of an encoder.
A HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER Manoj Kumar 1 Mohammad Zubair 1 1 IBM T.J. Watson Research Center, Yorktown Hgts, NY, USA ABSTRACT The MPEG/Audio is a standard for both
More informationM3039 MPEG 97/ January 1998
INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND ASSOCIATED AUDIO INFORMATION ISO/IEC JTC1/SC29/WG11 M3039
More informationREAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
More informationLOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE. indhubatchvsa@gmail.com
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE 1 S.Manikandan, 2 S.Abirami, 2 R.Indumathi, 2 R.Nandhini, 2 T.Nanthini 1 Assistant Professor, VSA group of institution, Salem. 2 BE(ECE), VSA
More informationBehavior Analysis in Crowded Environments. XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011
Behavior Analysis in Crowded Environments XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011 Behavior Analysis in Sparse Scenes Zelnik-Manor & Irani CVPR
More informationBinary Partition Tree as an Efficient Representation for Image Processing, Segmentation, and Information Retrieval
IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 9, NO. 4, APRIL 2000 561 Binary Partition Tree as an Efficient Representation for Image Processing, Segmentation, and Information Retrieval Philippe Salembier,
More informationHow To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
More informationEfficient Background Subtraction and Shadow Removal Technique for Multiple Human object Tracking
ISSN: 2321-7782 (Online) Volume 1, Issue 7, December 2013 International Journal of Advance Research in Computer Science and Management Studies Research Paper Available online at: www.ijarcsms.com Efficient
More informationHuman behavior analysis from videos using optical flow
L a b o r a t o i r e I n f o r m a t i q u e F o n d a m e n t a l e d e L i l l e Human behavior analysis from videos using optical flow Yassine Benabbas Directeur de thèse : Chabane Djeraba Multitel
More informationSubspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China zhaoming1999@zju.edu.cn Stan Z. Li Microsoft
More informationMethodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: macario.cordel@dlsu.edu.ph
More informationAutomatic Restoration Algorithms for 35mm film
P. Schallauer, A. Pinz, W. Haas. Automatic Restoration Algorithms for 35mm film. To be published in Videre, Journal of Computer Vision Research, web: http://mitpress.mit.edu/videre.html, 1999. Automatic
More informationMPEG-4 Natural Video Coding - An overview
MPEG-4 Natural Video Coding - An overview Touradj Ebrahimi* and Caspar Horne** *Signal Processing Laboratory **Mediamatics Inc. Swiss Federal Institute of Technology EPFL 48430 Lakeview Blvd 1015 Lausanne,
More informationImage Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner (tanner@vision.ee.ethz.ch) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
More informationFigure 1: Relation between codec, data containers and compression algorithms.
Video Compression Djordje Mitrovic University of Edinburgh This document deals with the issues of video compression. The algorithm, which is used by the MPEG standards, will be elucidated upon in order
More informationAN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
More informationHow To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
More informationFace Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
More informationInteractive Video Segmentation based on Quasi-Flat Zones
Interactive Video Segmentation based on Quasi-Flat Zones Jonathan Weber University of Strasbourg LSIIT Pôle API - Bd Sébastien Brant 67412 Illkirch CEDEX FRANCE Email: j.weber@unistra.fr Sébastien Lefèvre
More informationA Study of Automatic License Plate Recognition Algorithms and Techniques
A Study of Automatic License Plate Recognition Algorithms and Techniques Nima Asadi Intelligent Embedded Systems Mälardalen University Västerås, Sweden nai10001@student.mdh.se ABSTRACT One of the most
More informationMultimedia asset management for post-production
Multimedia asset management for post-production Ciaran Wills Department of Computing Science University of Glasgow, Glasgow G12 8QQ, UK cjw@dcs.gla.ac.uk Abstract This paper describes the ongoing development
More informationOptimal matching problem in detection and recognition performance evaluation
Pattern Recognition 35 (2002) 2125 2139 www.elsevier.com/locate/patcog Optimal matching problem in detection and recognition performance evaluation Gang Liu a;, Robert M. Haralick b a Cognex Corporation,
More informationAn Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
More informationVisualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
More informationAccurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
More informationVideo Authentication for H.264/AVC using Digital Signature Standard and Secure Hash Algorithm
Video Authentication for H.264/AVC using Digital Signature Standard and Secure Hash Algorithm Nandakishore Ramaswamy Qualcomm Inc 5775 Morehouse Dr, Sam Diego, CA 92122. USA nandakishore@qualcomm.com K.
More informationHow To Improve Performance Of The H264 Video Codec On A Video Card With A Motion Estimation Algorithm
Implementation of H.264 Video Codec for Block Matching Algorithms Vivek Sinha 1, Dr. K. S. Geetha 2 1 Student of Master of Technology, Communication Systems, Department of ECE, R.V. College of Engineering,
More informationAn Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2
An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2 Assistant Professor, Dept of ECE, Bethlahem Institute of Engineering, Karungal, Tamilnadu,
More informationCalculation of Minimum Distances. Minimum Distance to Means. Σi i = 1
Minimum Distance to Means Similar to Parallelepiped classifier, but instead of bounding areas, the user supplies spectral class means in n-dimensional space and the algorithm calculates the distance between
More informationSuper-resolution method based on edge feature for high resolution imaging
Science Journal of Circuits, Systems and Signal Processing 2014; 3(6-1): 24-29 Published online December 26, 2014 (http://www.sciencepublishinggroup.com/j/cssp) doi: 10.11648/j.cssp.s.2014030601.14 ISSN:
More informationTo determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt)
Polytechnic University, Dept. Electrical and Computer Engineering EL6123 --- Video Processing, S12 (Prof. Yao Wang) Solution to Midterm Exam Closed Book, 1 sheet of notes (double sided) allowed 1. (5 pt)
More information