Person Reidentification by Probabilistic Relative Distance Comparison


 Beverley Boyd
 1 years ago
 Views:
Transcription
1 Person Redentfcaton by Probablstc Relatve Dstance Comparson WeSh Zheng 1,2, Shaogang Gong 2, and Tao Xang 2 1 School of Informaton Scence and Technology, Sun Yatsen Unversty, Chna 2 School of Electronc Engneerng and Computer Scence, Queen Mary Unversty of London, UK Abstract Matchng people across nonoverlappng camera vews, known as person redentfcaton, s challengng due to the lack of spatal and temporal constrants and large vsual appearance changes caused by varatons n vew angle, lghtng, background clutter and occluson. To address these challenges, most prevous approaches am to extract vsual features that are both dstnctve and stable under appearance changes. However, most vsual features and ther combnatons under realstc condtons are nether stable nor dstnctve thus should not be used ndscrmnately. In ths paper, we propose to formulate person redentfcaton as a dstance learnng problem, whch ams to learn the optmal dstance that can maxmses matchng accuracy regardless the choce of representaton. To that end, we ntroduce a novel Probablstc Relatve Dstance Comparson (PRDC) model, whch dffers from most exstng dstance learnng methods n that, rather than mnmsng ntraclass varaton whlst maxmsng ntraclass varaton, t ams to maxmse the probablty of a par of true match havng a smaller dstance than that of a wrong match par. Ths makes our model more tolerant to appearance changes and less susceptble to model overfttng. Extensve experments are carred out to demonstrate that 1) by formulatng the person redentfcaton problem as a dstance learnng problem, notable mprovement on matchng accuracy can be obtaned aganst conventonal person redentfcaton technques, whch s partcularly sgnfcant when the tranng sample sze s small; and 2) our PRDC outperforms not only exstng dstance learnng methods but also alternatve learnng methods based on boostng and learnng to rank. 1. Introducton There has been an ncreasng nterest n matchng people across dsjont camera vews n a multcamera system, known as the person redentfcaton problem [10, 7, 14, 8, Most of ths work was done when the frst author was at QMUL. Fgure 1. Typcal examples of appearance changes caused by crossvew varatons n vew angle, lghtng, background clutter and occluson. Each column shows two mages of the same person from two dfferent camera vews. 3]. For understandng behavour of people n a large area of publc space covered by multple nooverlappng cameras, t s crtcal that when a target dsappears from one vew, he/she can be dentfed n another vew among a crowd of people. Despte the best efforts from computer vson researchers n the past 5 years, the person redentfcaton problem remans largely unsolved. Specfcally, n a busy uncontrolled envronment montored by cameras from a dstance, person verfcaton relyng upon bometrcs such as face and gat s nfeasble or unrelable. Wthout accurate temporal and spatal constrants gven the typcally large gaps between camera vews, vsual appearance features alone, extracted manly from clothng, are ntrnscally weak for matchng people (e.g. most people n wnter wear dark clothes). In addton, a person s appearance often undergoes large varatons across dfferent camera vews due to sgnfcant changes n vew angle, lghtng, background clutter and occluson (see Fg. 1), resultng n dfferent people appearng more alke than that of the same person across dfferent camera vews (see Fgs. 4 and 5). Most exstng studes have tred to address the above problems by seekng a more dstnctve and stable feature representaton of people s appearance, rangng wdely from color hstogram [10, 7], graph model [4], spatal cooccurrence representaton model [14], prncpal axs hstogram [8], rectangle regon hstogram [2], to combnatons of multple features [7, 3]. After feature extracton, exstng methods smply choose a standard dstance mea 649
2 sure such as l 1 norm [14], l 2 norm based dstance [8], or Bahattacharyya dstance [7]. However under severe vewng condton changes that can cause sgnfcant ntraobject appearance varaton (e.g. vew angle, lghtng, occluson), computng a set of features that are both dstnctve and stable under all condton changes s extremely hard f not mpossble under realstc condtons. Moreover, gven that certan features could be more relable than others under a certan condton, applyng a standard dstance measure s undesrable as t essentally treats all features equally wthout dscardng bad features selectvely n each ndvdual matchng crcumstance. In ths paper, we propose to formulate person redentfcaton as a dstance learnng problem whch ams to learn the optmal dstance metrc that can maxmse matchng accuracy regardless the choce of representaton. To that end, a novel Probablstc Relatve Dstance Comparson (PRDC) model s proposed. The objectve functon used by PRDC ams to maxmse the probablty of a par of true match (.e. two true mages of person A) havng a smaller dstance than that of a par of related wrong match (.e. two mages of person A and B respectvely). Ths s n contrast wth that of a conventonal dstance learnng approach, whch ams to mnmse ntraclass varaton n an absolute sense (.e. makng all mages of person A more smlar) whlst maxmsng nterclass varaton (.e. makng all mages of person A and B more dssmlar). Our approach s motvated by the nature of our problem. Specfcally, the person redentfcaton problem has three characterstcs: 1) the ntraclass varaton can be large and mportantly can be sgnfcantly vared for dfferent classes as t s caused by dfferent condton changes (see Fg. 1); 2) the nterclass varaton also vares drastcally across dfferent pars of classes; and 3) annotatng matched people across camera vews s tedous and typcally only lmted number of classes (people) are avalable for tranng wth each class contanng only a handful of mages of a person from dfferent camera vews (.e. undersamplng for buldng a representatve class dstrbuton). By explorng a relatve dstance comparson model probablstcally, our model s more tolerant to the large ntra/nterclass varaton and severe overlappng of dfferent classes n a multdmensonal feature space. Furthermore, due to the thrd characterstcs of undersamplng, a model could be easly overftted f t s learned by mnmsng ntraclass dstance and maxmsng nterclass dstance smultaneously by brutal force. In contrast, our approach s able to learn a dstance wth much reduced complexty thus allevatng the overfttng problem, as valdated by our extensve experments. Related work Although t has not been exploted for person redentfcaton, dstance learnng s a wellstuded problem wth a large number of methods reported n the lterature [16, 5, 17, 7, 17, 12, 15, 9, 1]. However, most of them suffer from the overfttng problem as explaned above. Recently, a few approaches attempt to allevate the problem by ncorporatng the dea of relatve dstance comparson as our PRDC model [12, 15, 9]. However, n these works, the relatve dstance comparson s not quantfed probablstcally, and mportantly s used as an optmsaton constrant rather than objectve functon. Therefore these approaches, ether mplctly [12, 9] or explctly [15] stll am to learn a dstance by whch each class becomes more compact whlst beng more separable from each other n an absolute sense. We demonstrate through experments that they reman susceptble to overfttng for person redentfcaton. There have been a couple of feature selecton based methods proposed specfcally for person redentfcaton [7, 11]. Gray et al. [7] proposed to use boostng to select a subset of optmal features for matchng people. However, n a boostng framework, good features are only selected sequentally and ndependently n the orgnal feature space where dfferent classes can be heavly overlapped. Such selecton may not be globally optmal. Rather than selectng features ndvdually and ndependently (local selecton), we am to learn an optmal dstance measure for all features jontly va dstance learnng (global selecton). An alternatve global selecton approach was developed based on RankSVM [11]. By formulatng the person redentfcaton as a rankng problem, the RankSVM approach shares the sprt of relatve comparson n our model. Nevertheless, our approach s more prncpled and tractable than the RankSVM n that 1) PRDC s a secondorder feature selecton approach whereas RankSVM s a frstorder one whch s not able to explot correlatons of dfferent features; 2) although RankSVM allevates the overfttng problem by fusng a rankng error functon wth a large margn functon n ts objectve functon, the probablstc formulaton of our objectve functon makes PRDC more tolerant to large ntra and nterclass varatons and data sparsty; 3) tunng the crtcal free parameter of RankSVM that determnes the weght between the margn functon and the rankng error functon s computatonally costly and can be suboptmal gven lmted data. In contrast, our PRDC model does not such a problem. We demonstrate the advantage of our approach over both the Boostng [7] and RankSVM [11] based methods through experments. The man contrbutons of ths work are twofold. 1) We formulate the person redentfcaton problem as a dstance learnng problem, whch leads to noteworthy mprovement on redentfcaton accuracy. To the best of our knowledge, t has not been nvestgated before. 2) We propose a probablstc relatve dstance comparson based method that overcomes the lmtatons of exstng dstance learnng methods when appled to person redentfcaton. 650
3 2. Probablstc Relatve Dstance Comparson for Person Redentfcaton Let us formally cast the person redentfcaton problem nto the followng dstance learnng problem. For an mage z of person A, we wsh to learn a redentfcaton model to successfully dentfy another mage z of the same person captured elsewhere n space and tme. Ths s acheved by learnng a dstance functon f(, ) so that f(z, z ) < f(z, z ), where z s an mage of any other person except A. To that end, gven a tranng set { (z, y ) }, where z Z s a multdmensonal feature vector representng the appearance of a person n one vew and y s ts class label (person ID), we defne a parwse set O = {O = (x p, xn )}, where each element of a parwse data O tself s computed usng a par of sample feature vectors. More specfcally, x p s a dfference vector computed between a par of relevant samples (of the same class/person) and x n s a dfference vector from a par of related rrelevant samples,.e. only one sample for computng x n s one of the two relevant samples for computng x p and the other s a msmatch from another class. The dfference vector x between any two samples z and z s computed by x = d(z, z ), z, z Z (1) where d s an entrywse dfference functon that outputs a dfference vector between z and z. The specfc form of functon d wll be descrbed n Sec Gven the parwse set O, a dstance functon f can be learned based on relatve dstance comparson so that a dstance between a relevant sample par (f(x p )) s smaller than that between a related rrelevant par (f(x n )). That s f(x p ) < f(xn ) for each parwse data O. To ths end, we measure the probablty of the dstance between a relevant par beng smaller than that of a related rrelevant par as: P (f(x p ) < f(xn )) = ( 1+exp { f(x p ) f(xn ) }) 1. (2) We assume the events of dstance comparson between a relevant par and an rrelevant par,.e. f(x p ) < f(xn ), are ndependent 1. Then, based on the maxmum lkelhood prncple, the optmal functon f can be learned as follows: f = arg mn r(f, O), f r(f, O) = log( P (f(x p O ) < (3) f(xn ))). The dstance functon f s parametersed as a Mahanalobs (quadratc) based dstance functon: f(x) = x T Mx, M 0 (4) where M s a semdefnte matrx. The dstance learnng problem thus becomes learnng M usng Eqn. (3). Drectly learnng M usng semdefnte program technques s computatonally expensve for hgh dmensonal data [15]. In partcular, we found out n our experments that gven a d 1 Note that we do not assume the data are ndependent. mensonalty of thousands, typcal for vsual object representaton, a dstance learnng method based on learnng M becomes ntractable. To overcome ths problem, we perform egenvalue decomposton on M: M = AΛA T = WW T, W = AΛ 1 2, (5) where the columns of A are orthonormal egenvectors of M and the dagonals of Λ are the correspondng egenvalues. Note that W s orthogonal. Therefore, learnng a functon f s equvalent to learnng an orthogonal matrx W = (w 1,, w l,, w L ) such that W = arg mn r(w, O), s.t. W wt w j = 0, j r(w, O) = O log(1 + exp { W T x p 2 W T x n 2} ) An Iteratve Optmsaton Algorthm It s mportant to pont out that our optmsaton crteron (6) may not be a convex optmsaton problem aganst the orthogonal constrant due to the relatve comparson modellng. It means that dervng an global soluton by drectly optmsng W s not straghtforward. In ths work we formulate an teratve optmsaton algorthm to learn an optmal W, whch also ams to seek a low rank (nontrval) soluton automatcally. Ths s crtcal for reducng the model complexty thus overcomng the overfttng problem gven sparse data. Startng from an empty matrx, after teraton l, a new estmated column w l s added to W. The algorthm termnates after L teratons when a stoppng crteron s met. Each teraton conssts of two steps as follows: Step 1. Assume that after l teratons, a total of l orthogonal vectors w 1,, w l have been learned. To learn the next orthogonal vector w l+1, let a l+1 l = exp{ j=0 wt j x p,j (6) 2 w T j x n,j 2 }, (7) are the dffer where we defne w 0 = 0, and x p,l and x n,l ence vectors at the lth teraton defned as follows: x s,l = x s,l 1 w l 1 w l 1x T s,l 1, s {p, n}, = 1,, O, l 1, where w l 1 = w l 1 / w l 1. Note that we defne x s,0 = x s, s {p, n}, and w 0 = 0. Step 2. Obtan x p,l+1, x n,l+1 by Eqn. (8). Let O l+1 = {O l+1 =(x p,l+1, x n,l+1 )}. Then, learn a new optmal projecton w l+1 on O l+1 as follows: r l+1 (w, O l+1 ) (8) w l+1 = arg mn w r l+1(w, O l+1 ), where (9) = log(1 + a l+1 O l+1 exp { w T x p,l+1 2 w T x n,l+1 2} ). We seek an optmal soluton by a gradent descent method: w l+1 w l+1 λ r l+1 w l+1, λ 0, (10) 651
4 r l+1 = w l+1 O l+1 2 a l a l+1 exp { wl+1 T xp,l+1 2 wl+1 T xn,l+1 2} exp { wl+1 T xp,l+1 2 wl+1 T xn,l+1 2} ( x p,l+1 x p,l+1 T x n,l+1 x n,l+1 T ) wl+1. where λ s a step length automatcally determned at each gradent update step. Accordng to the descent drecton n Eqn. (10) the ntal value of w l+1 for the gradent descent method s set to w l+1 = O l+1 1 (x n,l+1 O l+1 x p,l+1 ). (11) Note that the update n Eqn. (8) deducts nformaton from each sample x s,l 1 affected by w l 1 as wl 1 T xs,l = 0, so that the next learned vector w l wll only quantfy the part of the data left from the last step,.e. x s,l. In addton, ndcates the trends n the change of dstance measures a l+1 for x p and xn over prevous teratons and serve as a pror weght for learnng w l. The teraton of the algorthm (for l > 1) s termnated when the followng crteron s met: r l (w l, O l ) r l+1 (w l+1, O l+1 ) < ε. (12) where ε s a small tolerance value set to 10 6 n ths work. The algorthm s summarsed n Algorthm 1. Algorthm 1: Learnng the PRDC model Data: O = {O = (x p, xn )}, ε > 0 begn w 0 0, w 0 0; x s,0 x s, s {p, n}, O 0 O; l 0; whle 1 do Compute a l+1 by Eqn. (7); Compute x s,l+1, s {p, n} by Eqn. (8); O l+1 {O l+1 = (x p,l+1, x n,l+1 )}; Estmate w l+1 usng Eqn. (9); w l+1 = w l+1 ; w l+1 f (l > 1)&(r l (w l, O l ) r l+1 (w l+1, O l+1 ) < ε) then break; end l l + 1; end end Output: W = [ ] w 1,, w l 2.2. Theoretcal Valdaton The followng two theorems valdate that the proposed teratve optmsaton algorthm learns a set of orthogonal projectons {w l } that teratvely decrease the objectve functon n Crteron (6). Theorem 1. The learned vectors w l, l = 1,, L, are orthogonal to each other. Proof. Assume that l 1 orthogonal vectors {w j } l 1 j=1 have been learned. Let w l be the optmal soluton of Crteron (9) at the l teraton. Frst, we know that w l s n the range space 2 of {x p,l } {x n,l } accordng to Eqns. (10) and (11),.e. w l span{x s,l, = 1,, O, s {p, n}}. Second, accordng to Eqn. (8), we have w T j x s,j+1 = 0, s {p, n}, j = 1,, l 1 span{x s,l, = 1,, O, s {p, n}} span{x s,l 1, = 1,, O, s {p, n}} span{x s,0, = 1,, O, s {p, n}} Hence, w l s orthogonal to w j, j = 1,, l 1. (13) Theorem 2. r(w l+1, O l+1 ) r(w l, O l ), where W l = (w 1,, w l ), l 1. That s, the algorthm teratvely decreases the objectve functon value. Proof. Let w l+1 be the optmal soluton of Eqn. (9). By Theorem 1, t s easy to prove that for any j 1, wj T x s,j = wj T x s,0 = wj T x s, s {p, n}. Hence we have r l+1 (w l+1, O l+1 ) = O l+1 log(1 + a l+1 = r(w l+1, O) exp { wl+1 T xp,l+1 2 wl+1 T xn,l+1 2} ) Also r l+1 (0, O l+1 ) = r(w l, O). Snce w l+1 s the mnmal soluton, we have r l+1 (w l+1, O l+1 ) r l+1 (0, O l+1 ), and therefore r(w l+1, O) r(w l, O). Snce Crteron (9) may not be convex, a local optmum could be obtaned n each teraton of our algorthm. However, even f the computaton was trapped n a local mnmum of Eqn. (9) at the l + 1 teraton, Theorem 2 s stll vald f r l+1 (w l+1, O l+1 ) r l (w l, O l ), otherwse the algorthm wll be termnated by the stoppng crteron (12). To allevate the local optmum problem at each teraton, multple ntalsatons could also be deployed n practce Learnng n an Absolute Data Dfference Space To compute the data dfference vector x defned n Eqn. (1), most exstng dstance learnng methods use the followng entrywse dfference functon x = d(z, z ) = z z (14) to learn M = WW T n the normal data dfference space denoted by DZ = { x j = z z j z, z j Z }. The learned dstance functon s thus wrtten as: f(x j ) = (z z j ) T M(z z j ) = W T x j 2. (15) In ths work, we compute the dfference vector by the followng entrywse absolute dfference functon: x = d(z, z ) = z z, x(k) = z(k) z (k). (16) 2 It can be explored by Lagrangan equaton for Eqn. (9) for a nonzero wl. 652
5 where z(k) s the kth element of the sample feature vector. M s thus learned n an absolute data dfference space, denoted by { DZ = xj = z z j z, z j Z }, and our dstance functon becomes: f( x j ) = z z j T M z z j = W T x j 2. (17) We now explan why learnng n an absolute data dfference space s more sutable to our relatve comparson model. Frst, we note that: z (k) z j (k) (z (k) z j (k) (18) (z (k) z j (k)) (z (k) z j (k)), hence we have x j x j. x j x j, where. s an entrywse. As x j, x j 0, we thus can prove xj x j xj x j. (19) Ths suggests that the varaton of x j gven the same sample space Z s always less than that of x j. Specfcally, f z, z j, z j are from the same class, the ntraclass varaton s smaller n DZ than n DZ. On the other hand, f z j and z j belong to a dfferent class as z, the varaton of nterclass dfferences s also more compact n the absolute data dfference space. Snce the varatons of both relevant and rrelevant sample dfferences x p and x n are smaller, the learned dstance functon usng Eqn. (6) would yeld more consstent dstance comparson results therefore benefttng our PRDC model. Specally, for the same semdefnte matrx M, the Cauchy nequalty suggests upper( W T ( x j x j ) ) upper( W T (x j x j ) ), where upper( ) s the upper bound operaton. Ths ndcates that n the latent subspace nduced by W, the maxmum varaton of x j T M x j s lower than that of x T j Mx j. We show notable beneft of learnng PRDC n an absolute data dfference space n our experments Feature Representaton Our PRDC model can be appled regardless of the choce of appearance feature representaton of people. However, n order to beneft from dfferent and complementary nformaton captured by dfferent features, we start wth a mxture of colour and texture hstogram features smlar to those used n [7] and let our model automatcally dscover an optmal feature dstance. Specfcally, we dvded a person mage nto sx horzontal strpes. For each strpe, the RGB, YCbCr, HSV color features and two types of texture features extracted by Schmd and Gabor flters were computed and represented as hstograms. In total 29 feature channels were constructed for each strpe and each feature channel was represented by a 16 dmensonal hstogram vector. Each person mage was thus represented by a feature vector n a 2784 dmensonal feature space Z. Snce the features computed for ths representaton nclude lowlevel features wdely used by exstng person redentfcaton technques, ths representaton s consdered as generc and representatve. 3. Experments Datasets and settngs. Two publcally avalable person redentfcaton datasets, LIDS MultpleCamera Trackng Scenaro (MCTS) [18, 13] and VIPeR [6], were used for evaluaton. In the LIDS MCTS dataset, whch was captured ndoor at a busy arport arrval hall, there are 119 people wth a total 476 person mages captured by multple nonoverlappng cameras wth an average of 4 mages for each person. Many of these mages undergo large llumnaton change and are subject to occlusons (see Fg. 4). The VIPeR dataset s the largest person redentfcaton dataset avalable consstng of 632 people captured outdoor wth two mages for each person. Vewpont change s the most sgnfcant cause of appearance change wth most of the matched mage pars contanng one front/back vew and one sdevew (see Fg. 5). In our experments, for each dataset, we randomly selected all mages of p people (classes) to set up the test set, and the rest were used for tranng. Each test set was composed of a gallery set and a probe set. The gallery set conssted of one mage for each person, and the remanng mages were used as the probe set. Ths procedure was repeated 10 tmes. Durng tranng, a par of mages of each person formed a relevant par, and one mage of hm/her and one of another person n the tranng set formed a related rrelevant par, and together they form the parwse set O defned n Sec. 2. For evaluaton, we use the average cumulatve match characterstc (CMC) curves [6] over 10 trals to show the ranked matchng rates. A rank r matchng rate ndcates the percentage of the probe mages wth correct matches found n the top r ranks aganst the p gallery mages. Rank 1 matchng rate s thus the correct matchng/recognton rate. Note that n practce, although a hgh rank 1 matchng rate s crtcal, the top r ranked matchng rate wth a small r value s also mportant because the top matched mages wll normally be verfed by a human operator [6]. PRDC vs. NonLearnng based Dstances. We frst compared our PRDC wth nonlearnng based l 1 norm dstance and Bhattacharyya dstance whch were used by most exstng person redentfcaton work. Our results (Fgs. 2 and 3, Tables 1 and 2) show clearly that wth the proposed PRDC, the matchng performance for both datasets s mproved notably, more so when the number of people n the test pool ncreases (.e. tranng set sze decreases). The mprovement s partcularly dramatc on the VIPeR dataset. In partcular, Table 2 shows that a 4fold ncrease n correct matchng rate (r = 1) s obtaned aganst both l 1 norm and Bhattacharyya dstances when p = 316. The results valdate the mportance of performng dstance learnng. Examples of matchng people usng PRDC for both datasets are shown n Fgs. 4 and 5 respectvely. 653
6 Matchng Rate (%) LIDS 75 PRDC 65 Adaboost 55 ITM LMNN 45 MCC 35 Xng s L1 Norm Rank Score (a) p = 50 Matchng Rate (%) LIDS Rank Score (b) p = 80 Fgure 2. Performance comparson usng CMC curves on LIDS MCTS dataset. Methods p = 30 p = 50 p = 80 r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 PRDC Adaboost LMNN ITM MCC Xng s L1norm Bhat Table 1. Top ranked matchng rate (%) on LIDS MCTS. p s sze of the gallery set (larger p means smaller tranng set) and r s the rank. PRDC vs. Alternatve Learnng Methods. We also compared PRDC wth 5 alternatve dscrmnant learnng based approaches. These nclude 4 popular dstance learnng methods, namely Xng s method [16], LMNN [15], ITM [1] and MCC [5], and a method specfcally desgned for person redentfcaton based on Adaboost [7]. Among the 4 dstance learnng methods, only LMNN explots relatve dstance comparson. But as mentoned n Sec. 1, t s used as an optmsaton constrant rather than the man objectve functon whch s also not formulated probablstcally. MCC s smlar to PRDC n that a probablstc model s used but t s not a relatve dstance comparson based method. Note that snce MCC needs to select the best dmenson for matchng, we performed crossvaldaton by selectng ts value n {[1 : 1 : 10], d}, where d s the maxmum rank MCC can learn. Among the 5, the only method that learns n an absolute data dfferent space s Adaboost. Our results (Fgs. 2 and 3, Tables 1 and 2) show clearly that our model yelds the best rank 1 matchng rate and overall much superor performance compared to the compared models. The advantage of PRDC s partcularly apparent when a tranng set s small (learnng becomes more dffcult) and a test set s large ndcated by the value of p (matchng becomes harder). Table 2 shows that on VIPeR when 100 people are used for learnng and 532 people for testng (p = 532), the correct matchng rate for PRDC (and MCC) s almost more than doubled aganst any alternatve dstance learnng methods. Partcularly, beneftng from beng a probablstc model, MCC gves the most comparable results to PRDC when the tranng set s large. However, ts performance degrades dramatcally when the sze of tranng data decreases (see columns under p = 80 n Table 1 and p = 532 n Table 2). Ths suggests that overfttng to lmted tranng data s the man reason for the nferor performance of the compared alternatve learnng approaches. PRDC vs. RankSVM. Dfferent from PRDC, RankSVM has a free parameter whch determnes the relatve weghts between the margn functon and the rankng error functon [11]. In our experment, we crossvaldated the parameter n {0.0001, 0.005, 0.001, 0.05, 0.1, 0.5, 1, 10, 100, 1000}. As shown n Tables 3 and 4, the two methods all perform very well aganst other compared algorthms and our PRDC yelds overall better performance especally at lower rank matchng rate and gven less tranng data. The better performance of PRDC s due to the probablstc modellng and a secondorder rather than frstorder feature selecton. It s also noted that tunng the free parameter for RankSVM s not a trval task and the performance can be senstve to the tunng especally gven sparse data, whle PRDC does not have ths problem. In addton RankSVM s computatonally more expensve (see detals later). Effect of learnng n an Absolute Data Dfference Space. We have shown n Sec. 2.3 that n theory our relatve dstance comparson learnng method can beneft from learnng n an absolute data dfference space. To valdate ths expermentally, we compare PRDC wth PRDC raw whch learns n the normal data dfference space DZ (see Sec. 2.3). The result n Table 5 ndcates that learnng n an absolute data dfference space does mprove the matchng performance. Note that most exstng dstance learnng models are based on learnng n the normal data dfference space DZ. It s possble to reformulate some of them n 654
7 Matchng Rate (%) VIPeR PRDC Adaboost 50 ITM 40 LMNN 30 MCC 20 Xng s 10 L1 Norm Rank Score (a) p = 316 Matchng Rate (%) VIPeR Rank Score (b) p = 532 Fgure 3. Performance comparson usng CMC curves on VIPeR dataset. Methods p = 316 p = 432 p = 532 r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 PRDC Adaboost LMNN ITM MCC Xng s L1norm Bhat Table 2. Top ranked matchng rate (%) on VIPeR. p s the number of classes n the testng set; r s the rank. Rank PRDC RankSVM r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 p = p = p = Table 3. PRDC vs. RankSVM (%) on LIDS. Rank PRDC RankSVM r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 p = p = p = Table 4. PRDC vs. RankSVM (%) on VIPeR. Methods LIDS, (p = 50) VIPeR (p = 316) r = 1 r = 5 r = 10 r = 20 r = 1 r = 5 r = 10 r = 20 PRDC PRDC raw ITM abs MCC abs Table 5. Effect of learnng n an absolute data dfference space. Methods LIDS MCTS VIPeR p = 30 p = 50 p = 80 p = 316 p = 432 p = 532 rank(w) Table 6. Average Rank of W Learned by PRDC. order to learn n an absolute data dfference space. In Table 5 we show that when ITM and MCC are learned n the absolute data dfference space DZ, termed as ITM abs and MCC abs respectvely, ther performances become worse as compared to ther results n Tables 1 and 2. Ths ndcates that the absolute dfferent space s more sutable for our relatve comparson dstance learnng. Computatonal cost. Though PRDC s teratve, t has relatvely low cost n practce. In our experments, for VIPeR wth p = 316, t took around 15 mnutes for an Intel dualcore 2.93GHz CPU and 48GB RAM to learn PRDC for each tral. We observed that the low cost of PRDC s partally due to ts ablty to seek a sutable low rank of W (.e. converge wthn very few teratons) as shown n Table 6. For comparson, among the other compared methods, Adaboost s the most costly and took over 7 hours for each tral. For the 4 compared dstance learnng methods, PCA dmensonalty reducton must be performed otherwse they becomes ntractable gven the hgh dmensonal feature space. For the RankSVM method, each tral took around 2.5 hours due to parameter tunng. 4. Concluson We have proposed a new approach for person redentfcaton based on probablstc relatve dstance comparson whch ams to learn an sutable optmal dstance measure gven large ntra and nterclass appearance varatons and sparse data. Our experments demonstrate that 1) by formulatng person redentfcaton as a dstance learnng problem, clear mprovement n matchng performance can be obtaned and the mprovement s more sgnfcant when tranng sample sze s small, and (2) our PRDC outperforms not only exstng dstance learnng methods but also alternatve learnng methods based on boostng and learnng to rank. Acknowledgements Ths research was partally funded by the EU FP7 project SAMURAI wth grant no Dr. WeSh Zheng was also addtonally supported by the 985 project n Sun Yatsen Unversty wth grant no References [1] J. Davs, B. Kuls, P. Jan, S. Sra, and I. Dhllon. Informatontheoretc metrc learnng. In ICML,
8 Fgure 4. Examples of Person Redentfcaton on LIDS MCTS usng PRDC. In each row, the leftmost mage s the probe, mages n the mddle are the top 20 matched gallery mages wth a hghlghted red box for the correctly matched, and the rghtmost shows a true match Fgure 5. Examples of Person Redentfcaton on VIPeR usng PRDC [2] P. Dollar, Z. Tu, H. Tao, and S. Belonge. Feature mnng for mage classfcaton. In CVPR, [3] M. Farenzena, L. Bazzan, A. Perna, M. Crstan, and V. Murno. Person redentfcaton by symmetrydrven accumulaton of local features. In CVPR, [4] N. Ghessar, T. Sebastan, and R. Hartley. Person redentfcaton usng spatotemporal appearance. In CVPR, [5] A. Globerson and S. Rowes. Metrc learnng by collapsng classes. In NIPS, [6] D. Gray, S. Brennan, and H. Tao. Evaluatng appearance models for recognton, reacquston, and trackng. In IEEE Internatonal workshop on performance evaluaton of trackng and survellance, [7] D. Gray and H. Tao. Vewpont nvarant pedestran recognton wth an ensemble of localzed features. In ECCV, [8] W. Hu, M. Hu, X. Zhou, J. Lou, T. Tan, and S. Maybank. Prncpal axsbased correspondence between multple cameras for people trackng. PAMI, 28(4): , [9] J. Lee, R. Jn, and A. Jan. Rankbased dstance metrc learnng: An applcaton to mage retreval. In CVPR, [10] U. Park, A. Jan, I. Ktahara, K. Kogure, and N. Hagta. Vse: Vsual search engne usng multple networked cameras. In ICPR, [11] B. Prosser, W.S. Zheng, S. Gong, and T. Xang. Person redentfcaton by support vector rankng. In BMVC, [12] M. Schultz and T. Joachms. Learnng a dstance metrc from relatve comparsons. In NIPS, [13] UK. Home Offce LIDS multple camera trackng scenaro defnton [14] X. Wang, G. Doretto, T. Sebastan, J. Rttscher, and P. Tu. Shape and appearance context modelng. In ICCV, [15] K. Wenberger, J. Bltzer, and L. Saul. Dstance metrc learnng for large margn nearest neghbor classfcaton. In NIPS, [16] E. Xng, A. Ng, M. Jordan, and S. Russell. Dstance metrc learnng, wth applcaton to clusterng wth sdenformaton. In NIPS, [17] L. Yang, R. Jn, R. Sukthankar, and Y. Lu. An effcent algorthm for local dstance metrc learnng. In AAAI, [18] W.S. Zheng, S. Gong, and T. Xang. Assocatng groups of people. In BMVC,
Dropout: A Simple Way to Prevent Neural Networks from Overfitting
Journal of Machne Learnng Research 15 (2014) 19291958 Submtted 11/13; Publshed 6/14 Dropout: A Smple Way to Prevent Neural Networks from Overfttng Ntsh Srvastava Geoffrey Hnton Alex Krzhevsky Ilya Sutskever
More informationFace Alignment through Subspace Constrained MeanShifts
Face Algnment through Subspace Constraned MeanShfts Jason M. Saragh, Smon Lucey, Jeffrey F. Cohn The Robotcs Insttute, Carnege Mellon Unversty Pttsburgh, PA 15213, USA {jsaragh,slucey,jeffcohn}@cs.cmu.edu
More informationWho are you with and Where are you going?
Who are you wth and Where are you gong? Kota Yamaguch Alexander C. Berg Lus E. Ortz Tamara L. Berg Stony Brook Unversty Stony Brook Unversty, NY 11794, USA {kyamagu, aberg, leortz, tlberg}@cs.stonybrook.edu
More informationBoosting as a Regularized Path to a Maximum Margin Classifier
Journal of Machne Learnng Research 5 (2004) 941 973 Submtted 5/03; Revsed 10/03; Publshed 8/04 Boostng as a Regularzed Path to a Maxmum Margn Classfer Saharon Rosset Data Analytcs Research Group IBM T.J.
More informationAlgebraic Point Set Surfaces
Algebrac Pont Set Surfaces Gae l Guennebaud Markus Gross ETH Zurch Fgure : Illustraton of the central features of our algebrac MLS framework From left to rght: effcent handlng of very complex pont sets,
More informationSequential DOE via dynamic programming
IIE Transactons (00) 34, 1087 1100 Sequental DOE va dynamc programmng IRAD BENGAL 1 and MICHAEL CARAMANIS 1 Department of Industral Engneerng, Tel Avv Unversty, Ramat Avv, Tel Avv 69978, Israel Emal:
More informationEnsembling Neural Networks: Many Could Be Better Than All
Artfcal Intellgence, 22, vol.37, no.2, pp.239263. @Elsever Ensemblng eural etworks: Many Could Be Better Than All ZhHua Zhou*, Janxn Wu, We Tang atonal Laboratory for ovel Software Technology, anng
More informationSVO: Fast SemiDirect Monocular Visual Odometry
SVO: Fast SemDrect Monocular Vsual Odometry Chrstan Forster, Mata Pzzol, Davde Scaramuzza Abstract We propose a semdrect monocular vsual odometry algorthm that s precse, robust, and faster than current
More informationSupport vector domain description
Pattern Recognton Letters 20 (1999) 1191±1199 www.elsever.nl/locate/patrec Support vector doman descrpton Davd M.J. Tax *,1, Robert P.W. Dun Pattern Recognton Group, Faculty of Appled Scence, Delft Unversty
More information(Almost) No Label No Cry
(Almost) No Label No Cry Gorgo Patrn,, Rchard Nock,, Paul Rvera,, Tbero Caetano,3,4 Australan Natonal Unversty, NICTA, Unversty of New South Wales 3, Ambata 4 Sydney, NSW, Australa {namesurname}@anueduau
More informationDocumentation for the TIMES Model PART I
Energy Technology Systems Analyss Programme http://www.etsap.org/tools.htm Documentaton for the TIMES Model PART I Aprl 2005 Authors: Rchard Loulou Uwe Remne Amt Kanuda Antt Lehtla Gary Goldsten 1 General
More informationMANY of the problems that arise in early vision can be
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 26, NO. 2, FEBRUARY 2004 147 What Energy Functons Can Be Mnmzed va Graph Cuts? Vladmr Kolmogorov, Member, IEEE, and Ramn Zabh, Member,
More informationBRNO UNIVERSITY OF TECHNOLOGY
BRNO UNIVERSITY OF TECHNOLOGY FACULTY OF INFORMATION TECHNOLOGY DEPARTMENT OF INTELLIGENT SYSTEMS ALGORITHMIC AND MATHEMATICAL PRINCIPLES OF AUTOMATIC NUMBER PLATE RECOGNITION SYSTEMS B.SC. THESIS AUTHOR
More informationComplete Fairness in Secure TwoParty Computation
Complete Farness n Secure TwoParty Computaton S. Dov Gordon Carmt Hazay Jonathan Katz Yehuda Lndell Abstract In the settng of secure twoparty computaton, two mutually dstrustng partes wsh to compute
More informationAsRigidAsPossible Image Registration for Handdrawn Cartoon Animations
AsRgdAsPossble Image Regstraton for Handdrawn Cartoon Anmatons Danel Sýkora Trnty College Dubln John Dnglana Trnty College Dubln Steven Collns Trnty College Dubln source target our approach [Papenberg
More informationDo Firms Maximize? Evidence from Professional Football
Do Frms Maxmze? Evdence from Professonal Football Davd Romer Unversty of Calforna, Berkeley and Natonal Bureau of Economc Research Ths paper examnes a sngle, narrow decson the choce on fourth down n the
More informationAsRigidAsPossible Shape Manipulation
AsRgdAsPossble Shape Manpulaton akeo Igarash 1, 3 omer Moscovch John F. Hughes 1 he Unversty of okyo Brown Unversty 3 PRESO, JS Abstract We present an nteractve system that lets a user move and deform
More informationDP5: A Private Presence Service
DP5: A Prvate Presence Servce Nkta Borsov Unversty of Illnos at UrbanaChampagn, Unted States nkta@llnos.edu George Danezs Unversty College London, Unted Kngdom g.danezs@ucl.ac.uk Ian Goldberg Unversty
More informationcan basic entrepreneurship transform the economic lives of the poor?
can basc entrepreneurshp transform the economc lves of the poor? Orana Bandera, Robn Burgess, Narayan Das, Selm Gulesc, Imran Rasul, Munsh Sulaman Aprl 2013 Abstract The world s poorest people lack captal
More informationAssessing health efficiency across countries with a twostep and bootstrap analysis *
Assessng health effcency across countres wth a twostep and bootstrap analyss * Antóno Afonso # $ and Mguel St. Aubyn # February 2007 Abstract We estmate a semparametrc model of health producton process
More informationTrueSkill Through Time: Revisiting the History of Chess
TrueSkll Through Tme: Revstng the Hstory of Chess Perre Dangauther INRIA Rhone Alpes Grenoble, France perre.dangauther@mag.fr Ralf Herbrch Mcrosoft Research Ltd. Cambrdge, UK rherb@mcrosoft.com Tom Mnka
More informationStable Distributions, Pseudorandom Generators, Embeddings, and Data Stream Computation
Stable Dstrbutons, Pseudorandom Generators, Embeddngs, and Data Stream Computaton PIOTR INDYK MIT, Cambrdge, Massachusetts Abstract. In ths artcle, we show several results obtaned by combnng the use of
More informationEVERY GOOD REGULATOR OF A SYSTEM MUST BE A MODEL OF THAT SYSTEM 1
Int. J. Systems Sc., 1970, vol. 1, No. 2, 8997 EVERY GOOD REGULATOR OF A SYSTEM MUST BE A MODEL OF THAT SYSTEM 1 Roger C. Conant Department of Informaton Engneerng, Unversty of Illnos, Box 4348, Chcago,
More informationThe Developing World Is Poorer Than We Thought, But No Less Successful in the Fight against Poverty
Publc Dsclosure Authorzed Pol c y Re s e a rc h Wo r k n g Pa p e r 4703 WPS4703 Publc Dsclosure Authorzed Publc Dsclosure Authorzed The Developng World Is Poorer Than We Thought, But No Less Successful
More informationEffect of a spectrum of relaxation times on the capillary thinning of a filament of elastic liquid
J. NonNewtonan Flud Mech., 72 (1997) 31 53 Effect of a spectrum of relaxaton tmes on the capllary thnnng of a flament of elastc lqud V.M. Entov a, E.J. Hnch b, * a Laboratory of Appled Contnuum Mechancs,
More informationAlpha if Deleted and Loss in Criterion Validity 1. Appeared in British Journal of Mathematical and Statistical Psychology, 2008, 61, 275285
Alpha f Deleted and Loss n Crteron Valdty Appeared n Brtsh Journal of Mathematcal and Statstcal Psychology, 2008, 6, 275285 Alpha f Item Deleted: A Note on Crteron Valdty Loss n Scale Revson f Maxmsng
More information4.3.3 Some Studies in Machine Learning Using the Game of Checkers
4.3.3 Some Studes n Machne Learnng Usng the Game of Checkers 535 Some Studes n Machne Learnng Usng the Game of Checkers Arthur L. Samuel Abstract: Two machnelearnng procedures have been nvestgated n some
More informationTurbulence Models and Their Application to Complex Flows R. H. Nichols University of Alabama at Birmingham
Turbulence Models and Ther Applcaton to Complex Flows R. H. Nchols Unversty of Alabama at Brmngham Revson 4.01 CONTENTS Page 1.0 Introducton 1.1 An Introducton to Turbulent Flow 11 1. Transton to Turbulent
More informationUniform topologies on types
Theoretcal Economcs 5 (00), 445 478 555756/000445 Unform topologes on types YChun Chen Department of Economcs, Natonal Unversty of Sngapore Alfredo D Tllo IGIER and Department of Economcs, Unverstà Lug
More informationDISCUSSION PAPER. Is There a Rationale for OutputBased Rebating of Environmental Levies? Alain L. Bernard, Carolyn Fischer, and Alan Fox
DISCUSSION PAPER October 00; revsed October 006 RFF DP 03 REV Is There a Ratonale for OutputBased Rebatng of Envronmental Leves? Alan L. Bernard, Carolyn Fscher, and Alan Fox 66 P St. NW Washngton, DC
More information