All Downloads are FREE. Search and download functionalities are using the official Maven repository.

.terrierteam.jtreceval.0.0.2.source-code.trec_eval-linux-amd64 Maven / Gradle / Ivy

ELF>`
@@?{@8@%"@@@@@??@@@@?8?8 @@b@b?%?( (@(@b(@b??@@  P?tdXXBXB||Q?td/lib64/ld-linux-x86-64.so.2GNU	""#$)?9?2?????7?????lp?  ??L?`2?{??$??	??7??:??????!???%X??chu????v(.?E???eb??ebQ?eb??eblibm.so.6__gmon_start___Jv_RegisterClassesloglog2explibc.so.6fflushexitoptindstrncpyputcharreallocabortmmapreadstdoutfputslseekmalloc__ctype_b_locoptargstderrmunmapgetopt_longindexfwriteatofatolcloseopenfprintfqsortstrcmp__libc_start_mainsnprintffreeGLIBC_2.3GLIBC_2.2.5;0ii
 ui	*ui	*?Ab?eb$?eb%?eb#?eb"?Ab?Ab?AbBbBbBbBb	 Bb
(Bb0Bb8Bb
@BbHBbPBbXBb`BbhBbpBbxBb?Bb?Bb?Bb?Bb?Bb?Bb?Bb?Bb?Bb?Bb?Bb ?Bb!H???C????DH????5z6"?%|6"@?%z6"h??????%r6"h??????%j6"h??????%b6"h?????%Z6"h?????%R6"h?????%J6"h?????%B6"h?p????%:6"h?`????%26"h	?P????%*6"h
?@????%"6"h?0????%6"h? ????%6"h
?????%
6"h?????%6"h??????%?5"h??????%?5"h??????%?5"h??????%?5"h?????%?5"h?????%?5"h?????%?5"h?????%?5"h?p????%?5"h?`????%?5"h?P????%?5"h?@????%?5"h?0????%?5"h? ????%?5"h?????%?5"h????1?I??^H??H???PTI???OAH???OAH??8@???????H??H?14"H??t??H??Ð????????????UH??SH???=X"uD?@bH-@bH??H?X?H??W"H9?vH??H??W"??@bH??W"H9?w???W"H??[??UH?=2"H??t?H??t? @bI???A???Ð?UH??H????????H??????H?E?H?E?H?E?HDž???HDž(???HDž0???HDž???HDž ???HDž???HDž???HDž8???HDž@???HDžH??????HDžP????gAHDžX????gAHDž`???????????H?H?????????H??h???H??h???H??u
?????H??h???H??Dž????H??????H????????????I??? Cb??gA?????E?}?????E??D??x?????x???2????x???H???iA??H?E??H?R3"H?=?U"??gA??D?????????H?????H??H??????U???H?5?U"H?????????u'H??U"H?=?U"??gA???????????H?E?????H??(???H??H??(????????H?=\U"????H??@????????HDž????????H?54U"H??????
????H??0???H??H??0???????H?=U"????H??8???????H?=?T"????H??H????n???H??T"H??P????[???H??T"H??X????H???HDž ????8???H??`???H??H??`???H??T"H?E?????H?5?T"H?=?1"?1????????H?}?t*H??????
???uDž|????????:??V????(T"??;?????tH?5&T"H?=1"???????(?????S"Hc?H??H?????H?H?E??B??S"??S"Hc?H??H?????H?H?E??B??S"H?E??mH?E?H??H?? EbH??P?????????uIH?E?H??H??0EbH?U?H?u?H??????Ѓ??u6H?=sS"H?U??hA???????n???H?E?H?:3"H9E?|?H?-3"H9E?|'H??P???H?=)S"? hA??r?????(???H?E??mH?E?H??H???EbH??X??????????uIH?E?H??H???EbH?U?H?u?H??????Ѓ??u6H?=?R"H?U??hA??????????H?E?H??2"H9E?|?H??2"H9E?|'H??X???H?=qR"?HhA???????p???H??`???H??t+H??p???H?u?H?????詪???uDž|????????H?}?u;H??????zhA?_???u%H?
?Q"?&???hA?!?????????HDž?????hAHDž????HDž????HDž????HDž????H?E??{H?E?H?ŠFbH?@HH???u`H?E?H?ŠFbH?HH?E?H?4ŠFbH??????H??????у??u/H?E?H?ŠFbH?H?=AQ"??hA???????@???H?E?H??3"H9E??t???H??????H?????X???H??????H??????H??u
??????H??????H??H??H??????H??????H??H????H??????H??????HDž????H?E??H?????H??t'H?U?H?E?H??H?H?0H??????e???????H?E??=H?M?H?U?H??H?H?H??H?H?0H?U?H?E?H??H?H?8??????tH?E?H?E?H;E??H?E?H;E??uH?E?? H??????H?E?H??H¸H?BH?E?H??????H;E??H?U?H?E?H??H?H?H??????H?E??H?E?H?ŠFbH?@HH????H?E?H?ŠFbL?HH?E?L?ŠFbH?U?H?E?H??L?H?M?H?U?H??H?H?H??H?4H??????H?????I??L??L??A?у??u/H?E?H?ŠFbH?H?=O"??hA??^????????H?E?H??1"H9E??>???H??`???H??t'H??????H??p???諯???uDž|????????H?E???H?E?H?ŠFbH?@HH????H?E?H?ŠFbL?@ H?E?H?4ŠFbH??????H??????H?????A?Ѓ??u/H?E?H?ŠFbH?H?=>N"?iA???????=???H?????H??t`H?E?H?ŠFbH?H0H?E?H?4ŠFbH??????H??????у??u/H?E?H?ŠFbH?H?=?M"?HiA???????????H?E?H?=0"H9E??????H??????H??H??????H?E?H?E?H;E??????H??????H??u%H?
kM"?;??xiA??????e???H?E??H?E?H?ŠFbH?@HH????H?E?H?ŠFbL?@(H?E?H?4ŠFbH??????H?U?H?????A?Ѓ??t1H?E?H?ŠFbH?H8H?E?H?4ŠFbH??????H??????у??u/H?E?H?ŠFbH?H?=?L"??iA???????????H?E?H?/"H9E??;???H?????????u%H?
cL"????iA?????
?]???H??????H??????H??????H??????H??h???H????????&?????|?????UH??H??0H?}?H?u?H?U?H?E??H?E?H?E?H?PhH?E?H??H?H?H??u?H?E?H?PhH?E?H??H?H?E?H?H?E?H?PhH?E?H??H?H?E?H?BH?E????u6H?E?H?PhH?E?H??H?H?H?=iK"??kA??????E??????&H?E?H?@hH?PH?E?H??H?H??E??E???UH??H?? H?}?H?E??7H?E?H?ŠFbH?0H?}????????uH?E?H?ŠFbH?@H?????H?E?H?b-"H9E?|?H?U-"H9E?|	?E???????E??E???UH??H??@H?}?H?u?H?E?H?E??H?E?H?E????tH?E??<.u?H?E??<.u.H?E??H?E?H?U?H?u?H?}??G??????u?E??????H?E??jH?E?H??H???KbH?}???????uIH?E?H??H???KbH?E??H?E?H?8????????u	?E??????8H?E?H?E?H?H??u??E??H?E?H?0"H9E?|?H?}??????E̋E???UH??H??0H?}?H?E?H?5?&"??kA??H???H?E??%H?E?H?pXH?E?H??H???Eb?R?????tH?E?H?z)"H9E?|?H?m)"H9E?}H?E?H??H???Eb??kA??????H?E??%H?E?H?pPH?E?H??H?? Eb???????tH?E?H??("H9E?|?H??("H9E?}H?E?H??H??(Eb??kA??z?????kA????H?E??QH?E?H?ŠFbH?@HH???u6H?E?H?E?H?ŠFbH?PH?E?H?ŠFbH?0?lA?????H?E?H??*"H9E?|?H?}?u
? lA?7??????UH??H?? H?}?H?u?H?E?H?E??H?E?H?E????tH?E??<.u?H?E????tH?E??H?E?H?U?H?E?H?BH?}??o???H??H?E?H?P??UH??H?? H?}?H?E??LH?E?H??H?? EbH?E?H?xP??????u'H?E?H??H??8Eb??҃??u?E???????H?E?H?A'"H9E?|?H?E??IH?E?H??H???EbH?E?H?xX??????u$H?E?H??H???Eb??҃??u?E??????~H?E?H?!'"H9E?|?H?E??)H?E?H??H??Fb??҃??u	?E??????>H?E?H??'"H9E?|?H?E?H?@`H??t???????u	?E???????E??E??Ð?UH??H?}?H?u?H?U?H?E?H?@H???UH??H?? H?}?H?u?H?U?H?E?H?@H?PH?u?H?? H?E?H?x??̝H??H?E?H?PH?E?H?@H??u	?E??????UH?E?H?PH?E?H?PHH?E?H?PH?E?H?@HH??H?H?E?H?H??H?BH?E?H?@H?PH?E?H?P?E??E???UH??H?? H?}?H?u?H?U?H?E?H?@H?PH?u?H?? H?E?H?x???H??H?E?H?PH?E?H?@H??u	?E??????UH?E?H?PH?E?H?PHH?E?H?PH?E?H?@HH??H?H?E?H?H??H?BH?E?H?@H?PH?E?H?P?E??E???UH??SH??HH?}?H?u?H?U?H?E?H?@hH??t\H?E?H?@hH?E??BH?E?H?0H?E?H?8??????u&H?E?H?pH?E?H?x@?W???u?E??????AH?E?H?E?H?H??u?H?E?H?@@H?@H?E?H?E?H?PH?E?H?@@H?@H?H?u?H?? H?E?H?x??ݛH??H?E?H?PH?E?H?@H??u?E???????H?E??uH?E?H?PH?E?H?@HE?H??H?H?E?H??HE?H?0H?E?H?8?H??H?CH?E?H?PH?E?H?@HE?H??H?H?H??u	?E??????QH?E?H?E?H?@@H?@H;E??u???H?E?H?PH?E?H?PHH?E?H?PH?E?H?@@H?@H?H?E?H?P?E??E?H??H[??UH??SH??HH?}?H?u?H?U?H?E?H?@hH??t\H?E?H?@hH?E??BH?E?H?0H?E?H?8???????u&H?E?H?pH?E?H?x@?????u?E??????JH?E?H?E?H?H??u?H?E?H?@@H?@H?E?H?E?H?PH?E?H?@@H?@H?H?u?H?? H?E?H?x??%?H??H?E?H?PH?E?H?@H??u?E???????H?E??~H?E?H?PH?E?H?@HE?H??H?H?E?H??HE?H?H?E?H?8H?U??E??L
H??H?CH?E?H?PH?E?H?@HE?H??H?H?H??u	?E??????QH?E?H?E?H?@@H?@H;E??l???H?E?H?PH?E?H?PHH?E?H?PH?E?H?@@H?@H?H?E?H?P?E??E?H??H[??UH??SH??8H?}?H?u?H?U?H?E?H?@hH??t\H?E?H?@hH?E??BH?E?H?0H?E?H?8?'?????u&H?E?H?pH?E?H?x@?[???u?E??????&H?E?H?E?H?H??u?H?E?H?@H?PH?u?H?? H?E?H?x???H??H?E?H?PH?E?H?@H??u?E???????H?E?H?PH?E?H?PHH?E?H?@@H?H??t=H?E?H?PH?E?H?@HH??H?H?E?H?@@H?0H?E?H?8?5H??H?C?*H?E?H?PH?E?H?@HH??H?H?E?H?H??H?BH?E?H?PH?E?H?@H??H?H?H??u	?E??????H?E?H?@H?PH?E?H?P?E??E?H??8[??UH??SH??8H?}?H?u?H?U?H?E?H?@hH??t\H?E?H?@hH?E??BH?E?H?0H?E?H?8??????u&H?E?H?pH?E?H?x@?h???u?E??????&H?E?H?E?H?H??u?H?E?H?@H?PH?u?H?? H?E?H?x???H??H?E?H?PH?E?H?@H??u?E???????H?E?H?PH?E?H?PHH?E?H?@@H?H??t=H?E?H?PH?E?H?@HH??H?H?E?H?@@H?0H?E?H?8?
H??H?C?*H?E?H?PH?E?H?@HH??H?H?E?H?H??H?BH?E?H?PH?E?H?@H??H?H?H??u	?E??????H?E?H?@H?PH?E?H?P?E??E?H??8[??UH??SH??XH?}?H?u?H?E?H?E?H?E??H?E??<,uH?E?H?E?H?E????u?H?E?H??????H?E???H?}??H??H??H??????????H??H?E?H?H?E?H?H??tH?E?H?????????H?E?H?}?u?E??????H?E?H??????H?E???H?}??H??H??H??H?PH?E?H?8H?u??y???H?U?H?E?H?BH?U?H?E?H?BH?E?H?E?H?E?H?E?H?E??CH?E??<,u3H?E??H?E?H??H??H]?H?}?????H?H?E?H?E?H??H?E?H?E?H?E????u?H?E?H??H??H]?H?}??a???H?H?E?H?E?Hc?H?}??/@?????H?E?H?H??$H?
?<"?$??H?A?????E??????oH?E??TH?U?H??H?E?H??H?H?H?E?H??HE?H?H9?u$H?
+<"?&??p?A?O????E??????H?E?H?E?H;E?|??E??E?H??X[??UH??SH??hH?}?H?u?H?E?H?E?H?E??H?E??<,uH?E?H?E?H?E????u?H?E?H??????H?E???H?}??H??H??H???????????H??H?E?H?H?E?H?H??tH?E?H????????H?E?H?}?u?E??????{H?E?H??????H?E???H?}??H??H??H??H?PH?E?H?8H?u??/???H?U?H?E?H?BH?U?H?E?H?BH?E?H?E?H?E?H?E?H?E??LH?E??<,uH?E??E??????M??Y??E??X??E?H?E?H?PH?E?H?@HH??H??H*M??E??^??????E?H?E?H?C?H??H[?Ð?UH??H?}?H?u?H?U???UH??H?? H?}?H?u?H?U?H?E?H?@`H??t[H?E?H?PH?E?H?@HH??H?H?HH?E?H?8H?E?H?PH?E?H?@HH??H?H?0H?M??E?H???ܔA???????YH?E?H?PH?E?H?@HH??H?H?HH?E?H?8H?E?H?PH?E?H?@HH??H?H?0H?M??E?H????A????????UH??H?? H?}?H?u?H?U?H?E?H?@`H??t[H?E?H?PH?E?H?@HH??H?H?HH?E?H?8H?E?H?PH?E?H?@HH??H?H?0H?M??E?H???ܔA??????VH?E?H?PH?E?H?@HH??H??@?H,?H?E?H?8H?E?H?PH?E?H?@HH??H?H?0H?????A????????UH??H??0H?}?H?u?H?U?H?E???H?E?H?@`H??tcH?E?H?PH?E?H?@HHE?H??H?H?HH?E?H?8H?E?H?PH?E?H?@HHE?H??H?H?0H?M??E?H???ܔA??)????aH?E?H?PH?E?H?@HHE?H??H?H?HH?E?H?8H?E?H?PH?E?H?@HHE?H??H?H?0H?M??E?H????A??????H?E?H?E?H?@@H?@H;E????????UH??H?}?H?u?H?U???UH??H??@H?}?H?u?H?U?H?E?H?@H????H?E?H?PH?E?H?@HH??H??@?]?H?E?H?H?E?H?E?H?PH?E?H?@HH??H?H?H?E?H?E?H?@`H??t
H?E??A?H?E?
?A?E?H?M?H?U?H?u??A?????????UH??H?? H?}?H?u?H?U?H?E?H?@H????H?E?H?@`H??t[H?E?H?PH?E?H?@HH??H?H?HH?E?H?8H?E?H?PH?E?H?@HH??H?H?0H?M??E?H??? ?A??O????VH?E?H?PH?E?H?@HH??H??@?H,?H?E?H?8H?E?H?PH?E?H?@HH??H?H?0H???1?A?????????UH??H??PH?}?H?u?H?U?H?E???H?E?H?@H????H?E?H?PH?E?H?@HHE?H??H??@?]?H?E?H?H?E?H?E?H?PH?E?H?@HHE?H??H?H?H?E?H?E?H?@`H??t
H?E??A?H?E?
?A?E?H?M?H?U?H?uп?A??/???H?E?H?PH?E?H?@HHE?H??H?H?8?????H?E?H?E?H?@@H?@H;E??!???H?E?H?@@H?H??t!H?E?H?@@H?x????H?E?H?@@H?8???????UH??H??@H?}?H?u?H?U?H?E?H?@H????H?E?H?PH?E?H?@HH??H??@?]?H?E?H?H?E?H?E?H?PH?E?H?@HH??H?H?H?E?H?E?H?@`H??t
H?E??A?H?E?
?A?E?H?M?H?U?H?u??A?????H?E?H?@@H?H??tAH?E?H?PH?E?H?@HH??H?H?8????H?E?H?@@H?8????H?E?H?@@H?x?{??????UH??SH??H?}?H?u?H?U??E?H?}????"????E??}?????}????c????E??}?~e?E??????????H??("H??("H??tE?}????(???H???t-?E?Hc؋E?Hc?H?5g("?}?????H9?u
?}????????u)H?=&("H?U??@?A??k???Dž|????????H?("H?P??E?H?H??<
t?E?H?H??H??'"H??
?E??E?H?H??H??'"H??H?E?H??'"H?E??H?E?H?}??
????H??H?E?H?E????u?H?U?H??H?H?H????????H?E?H?}?uDž|????????WH?E?H?E?H?]'"H?E??vH?M?H??H?U?H??H?u?H?}?????uOH?U?H?E?H??H)?H??H??H??H?????????H??H?PH?=?&"?x?A??-???Dž|?????????H?E?H?E?????{???H?U?H?E?H??H)?H??H??H??H?????????H??H?E?H?E?Hc?H?}???A@??V???H?E?H?E??KH?U?H??H?H?H??HE?H?0H?M?H??H?U?H??H?H?H??H?H?8??????tH?E?H?E?H?E?H;E?|?H?U?H??H?H?H?????G???H? &"H?&"H??tMH?U?H??H?H?H????????H??%"H??%"H??t"H?E?H?????????H??%"H??%"H??uDž|????????H??%"H?E?H??%"H?E?H??%"H?E?H?E???AH?E???H?U?H??H?H?H??HE?H?0H?}??!?????tH?}?t-H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E?H?U?H??H?H?H??HE?H?H?E?H?U?H?E?H?BH?U?H?E?H?H?E?H?@??AH?U?H?E?H?BH?U?H??H?H?H??HE?H?PH?E?H?H?U?H??H?H?H??HE?H?x????H??H?E?H?PH?E?H?E?H?E?H;E??????H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?U?H?E?H?H?H$"H?E?H?PH?}?????Dž|?????|???H?Ĉ[??UH??H??0H?}?H?u?H?E?H?0H?E?H?8?????E??}?t?E??E??H?E?H?pH?E?H?x?????E܋E???UH??H??@H?}?H?u?H?U?H?M?H?E?H?H?E??H?E?H?E??<
t)?????H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E?????H?H?E??H??H?H????% ??t?H?E??<
u?E??????=H?E??H?E??H?E?H?E??<
t0?Y???H?H?E??H??H?H????% ??u??H?E??)???H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E?H?E??<
t)?????H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E?????H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?9???H?H?E??H??H?H????% ??u?H?E??<
u?E??????H?U?H?E?H??H?E??????H?H?E??H??H?H????% ??t?H?E??<
t[H?E??H?E??H?E?H?E??<
t)????H?H?E??H??H?H????% ??u?H?E??<
t	?E??????H?E??H?E?H?U?H?E?H??E??E???UH??H?? "H??tH?=? "????H?? "H?? "H??tH?? "H???????H?} "H?~ "H??tH?r "H???????H?_ "H?` "H??tH?T "H??????H?A "??Ð?UH??H??H?}?H?u?H??x???H?E?H?E?H?}????+????E??}??tL?}????p???H?E?H?}?t/?E?H?u?A?A??????????H?E?H?}??u)H?=g"H?U????A?????Džt????????,H?}?H??????H?q"H?j"H??u)H?="H?U????A??c???Džt?????????H?5"H?U?H?M?H??H????H?u?H?}?????????t
?}????????u)H?=?"H?U???A?????Džt????????H??"H?P?H?E?H??<
tH??"H?U?H??
H?E?H??"H?E?H??H?E?H??"H?E??H?E?H?}??
?:???H??H?E?H?E????u?H?U?H??H?H?H?????_???H?E?H?}?uDžt?????????H?E?H?E?H?"H?E???H?E?H??H?E?H?E????t4H?E??<
t)?F???H?H?E??H??H?H????% ??u?H?E??<
uH?E?H??H?E??}H?M?H??H?U?H??H?u?H?E?H?}?I??????uOH?U?H?E?H??H)?H??H??H??H?????????H??H?PH?="?P?A??`???Džt?????????H?E?H?E????????H?U?H?E?H??H)?H??H??H??H?????????H??H?E?H?E?Hc?H?}ȹ?K@?????H?E?H?E??KH?U?H??H?H?H??HE?H?0H?M?H??H?U?H??H?H?H??H?H?8?J?????tH?E?H?E?H?E?H;E?r?H?E?H????????H?|"H?u"H??tMH?U?H??H?H?H?????X???H?A"H?:"H??t"H?E?H?????6???H?'"H? "H??uDžt????????H?
"H?E?H??"H?E?H??"H?E?H?E?{?AH?E??	H?U?H??H?H?H??HE?H?0H?}??]???????H?}?t-H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E? H?U?H??H?H?H??HE?H?H?E?H?U?H?E?H?BH?M?H?U?H?E?H?H?E?H?HH?E?H?@|?AH?U?H?E?H?BH?U?H??H?H?H??HE?H?PH?E?H?H?U?H??H?H?H??HE?H?x?????Z?H?E??@H?E?H?E?H?E?H;E??????H?U?H?E?H?@H??H)?H??H??H??H?E?H?H??x???H?E?H?H??"H??x???H?PH?}?????Džt?????t?????UH??H??0H?}?H?u?H?E?H?0H?E?H?8??????E??}?t?E??E??H?E?H?pH?E?H?x??????E܋E???UH??H??@H?}?H?u?H?U?H?M?L?E?H?E?H?H?E?H?U?H?E?H??H?E?????H?H?E??H??H?H????% ??t?H?E??<
u?E??????&H?E??H?E??H?E?H?E??<
t0輿??H?H?E??H??H?H????% ??u??H?E?茿??H?H?E??H??H?H????% ??t?H?E??<
??H?E???t?E??????H?E?H?E??<
t)?2???H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??????H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t0虾??H?H?E??H??H?H????% ??u??H?E??i???H?H?E??H??H?H????% ??t?H?E??<
??H?E???t?E??????kH?E?H?E??<
t)????H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??Խ??H?H?E??H??H?H????% ??t?H?E??<
u?E???????H?E??H?E??H?E?H?E??<
t)?v???H?H?E??H??H?H????% ??u?H?E??<
u?E??????H?U?H?E?H??H?E??$???H?H?E??H??H?H????% ??t?H?E??<
tH?E??H?E??H?E?H?E??<
u?H?E??H?E?H?U?H?E?H??E??E???UH??H?Q"H??tH?=E"舼??H?5"H?6"H??tH?*"H???b???H?"H?"H??tH?"H???"H?E?H?E??AH?E??EH?U?H??H??H?H??HE?H?0H?}?苸??????H?}?t-H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E?H?U?H??H??H?H??HE?H?H?E?H?U?H?E?H?BH?U?H?E?H?H?E?H?@?AH?U?H?E?H?BH?U?H??H??H?H??HE?H?PH?E?H?H?U?H??H??H?H??HE?H?PH?E?H?PH?U?H??H??H?H??HE?H?x ?!????Z?H?E??@H?U?H??H??H?H??HE?H?PH?E?H?PH?E? H?E?H?E?H;E??????H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?U?H?E?H?H??"H?E?H?PH?}?褶??Dž|?????|???H?Ĉ[??UH??H??0H?}?H?u?H?E?H?0H?E?H?8?׶???E??}?t?E??E??H?E?H?pH?E?H?x豶???E܋E???UH??H??PH?}?H?u?H?U?H?M?L?E?L?M?H?E?H?H?E??H?E?H?E??<
t)????H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??̵??H?H?E??H??H?H????% ??t?H?E??<
u?E???????H?E??H?E??H?E?H?E??<
t)?n???H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??3???H?H?E??H??H?H????% ??t?H?E??<
u?E??????OH?E??H?E??H?E?H?E??<
t)?մ??H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E?蚴??H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?H?U?H?E?H??H)?H??H??H?PH?=s	"?0?A?輰??Dž|????????H?E? H?E????u?H?U?H?E?H??H)?H??H??H?E?H?E?Hc?H?}??_@? ?????H?E?H?E??9H?E?H??HE?H?0H?U?H?? H?E?H??H?H?8?ͯ????tH?E?H?E?H?E?H;E?|?H?U?H??H?H?H?????????H?6	"H?/	"H??tMH?U?H??H?H?H?????Ү??H??"H??"H??t"H?E?H????谮??H??"H??"H??uDž|????????H??"H?E?H??"H?E?H??"H?E?H?E?_?AH?E???H?E?H??HE?H?0H?}???????tvH?}?t-H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E?H?E?H??HE?H?H?E?H?U?H?E?H?BH?U?H?E?H?H?E?H?@`?AH?U?H?E?H?BH?E?H??HE?H?PH?E?H?H?E?H?@f?AH?E?H??HE?H?x训???Z?H?E??@H?E?H??HE?H?PH?E?H?PH?E? H?E?H?E?H;E??????H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?U?H?E?H?H?\"H?E?H?PH?}??;???Dž|?????|???H?Ĉ[??UH??H??0H?}?H?u?H?E?H?0H?E?H?8?n????E??}?t?E??E??H?E?H?pH?E?H?x?H????E܋E???UH??H??@H?}?H?u?H?U?H?M?L?E?H?E?H?H?E??H?E?H?E??<
t)袬??H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??g???H?H?E??H??H?H????% ??t?H?E??<
u?E??????OH?E??H?E??H?E?H?E??<
t)?	???H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??Ϋ??H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?p???H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??5???H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?ת??H?H?E??H??H?H????% ??u?H?E??<
u?E??????H?U?H?E?H??H?E?腪??H?H?E??H??H?H????% ??t?H?E??<
t[H?E??H?E??H?E?H?E??<
t)?3???H?H?E??H??H?H????% ??u?H?E??<
t	?E??????H?E??H?E?H?U?H?E?H??E??E???UH??H??"H??tH?=?"謩??H??"H??"H??tH??"H??膩??H?{"H?|"H??tH?p"H???`???H?]"H?^"H??tH?R"H???:???H??"???UH??SH??H?}?H??x???H??p????E?H??x??????ͩ???E??}?????}????????E??}?~e?E?????舨??H??"H??"H??tE?}????Ө??H???t-?E?Hc؋E?Hc?H?5?"?}??2???H9?u
?}??ŧ?????u,H?=?"H??x????h?A?????Džl????????}H?E"H?P??E?H?H??<
t?E?H?H??H?""H??
?E??E?H?H??H?"H??H?E?H??"H?E??H?E?H?}??
?A???H??H?E?H?E????u?H?E?H?????o???H?E?H?}?uDžl?????????H?E?H?E?H??"H?E??pH?E?H??H?M?H??H?U?H??H?u?H?}?I??????u>H?U?H?E?H??H)?H??H??H?PH?=?"???A?????Džl????????NH?E? H?E????u?H?U?H?E?H??H)?H??H??H?E?H?E?Hc?H?}???i@? ?"???H?E?H?E?H?E???H?E?H??HE?H?0H?U?H?? H?E?H??H?H?8??????tH?E?H?E??H?E?H??HE?H?pH?U?H?? H?E?H??H?H?x誦????tH?E??fH?E?H??HE?H?pH?U?H?? H?E?H??H?H?x?r?????u5H?E?H??HE?H?PH?=_?!?ЗA?訦??Džl????????H?E?H?E?H;E?????H?U?H??H?H?H?????n???H???!H???!H??tfH?E?H?????L???H???!H???!H??tDH?E?H?????*???H?{?!H?t?!H??t"H?E?H????????H?a?!H?Z?!H??uDžl????????_H?G?!H?E?H?$?!H?E?H?!?!H?E?H??!H?E?H?E???AH?E???AH?E??H?E?H??HE?H?0H?}??%???????H?}?tVH?U?H?E?H?@H??H)?H??H??H?PH?E?H?H?E?H?E?H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E?H??HE?H?H?E?H?U?H?E?H?BH?U?H?E?H?H?E?H?@??AH?U?H?E?H?BH?E?H??HE?H?@H?E?H?U?H?E?H?B?lH?E?H??HE?H?pH?}??C?????tOH?}?t(H?U?H?E?H?@H??H)?H??H??H??H?E?H?H?E?H?E?H??HE?H?@H?E?H?U?H?E?H?BH?E?H??HE?H?PH?E?H?H?E?H??HE?H?x?t???H??H?E?H?PH?E?H?E?H?E?H;E??d???H?U?H?E?H?@H??H)?H??H??H?PH?E?H?H?U?H?E?H?@H??H)?H??H??H??H?E?H?H??p???H?E?H?H?
?!H??p???H?PH?}??????Džl?????l???H?Ę[??UH??H??0H?}?H?u?H?E?H?0H?E?H?8??????E??}?t?E??E??>H?E?H?pH?E?H?x?΢???E??}?t?E??E??H?E?H?pH?E?H?x訢???E܋E???UH??H??@H?}?H?u?H?U?H?M?L?E?H?E?H?H?E??H?E?H?E??<
t)????H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??ǡ??H?H?E??H??H?H????% ??t?H?E??<
u?E??????OH?E??H?E??H?E?H?E??<
t)?i???H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E??.???H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?Р??H?H?E??H??H?H????% ??u?H?U?H?E?H??H?E?蕠??H?H?E??H??H?H????% ??t?H?E??<
u?E??????H?E??H?E??H?E?H?E??<
t)?7???H?H?E??H??H?H????% ??u?H?E??<
u?E??????H?U?H?E?H??H?E?????H?H?E??H??H?H????% ??t?H?E??<
t[H?E??H?E??H?E?H?E??<
t)蓟??H?H?E??H??H?H????% ??u?H?E??<
t	?E??????H?E??H?E?H?U?H?E?H??E??E???UH??H?5?!H??tH?=)?!????H??!H??!H??tH??!H??????H???!H???!H??tH???!H???????H???!H???!H??tH???!H??蚞??H???!H???!H??tH???!H???t???H???!??Ð?UH??H??pH?}?H?u?H?U?H?M?H?E?H?0H?=??!襞????ugH?U?H???!H?H???!H?BH???!H?BH???!H?BH???!H?B H???!H?B(H???!H?B0H???!H?B8?E???H?E?H?p??A?(?????uH?E?H?p??A??????t'H?
?!?W??(?A?3????E??????H?E?H?H??????H?E???H?}??H??H??H??H??H?E?H?=??!H?U????fb??MH?v?!H?o?!H??u?E??????=H?U?H?E?H?0H?=L?!藝??H?E?H?@H?E?H?E?H?@H?E?H?E?H?H?E?H?=??!H?Uȹ??fb?VMH?{?!H?t?!H??t-H?=p?!H?Uȹ ??fb?)MH?V?!H?O?!H??u?E??????H?E??eH?E?H??H??H?"?!H?H?E?H?PH?E?H??H?H?H?H?E?H??H??H???!H?H?E?H?PH?E?H??H??@?AH?E?H?E?H;E?|?H?E?Hc?H???!H?ǹx@? ?5???H?E?H?@HH;E?}H?E?H?@HH?E?H?E??&H?E?H??H??H?o?!H?H?E?H??H?BH?E?H?E?H;E?|?H?E?Hc?H?C?!H?ǹ~x@? ?????H?E??|H? ?!H?P?H?E?H??H?H?0H?E?H??H??H???!H?H?8菛????u;H?E?H??H??H???!H?H?H?=s?!???A?輛???E??????H?E?H?E?H;E??v???H?E?H?@H?E?H?E?H?PH?E?H?H??H?H?E?H?E?H?@H?E?H?E??cH?E?H?@H;E?~H?E?H?@H?E?H?E?H?0H?E?H??H?8?Қ????u)H?E?H?H?=??!???A??????E??????dH?E?H?E?H;E?r?H?U?H??H?=?!???fb?JH?i?!H?b?!H??u?E??????H?E?H??H??H??H?;?!H???H?Ѹ?H?E?H?@H?E?H?E?H?PH?E?H?H??H?H?E?H?E??FH?E?H?@H??x#H?E?H?@H??H??H???!H?H?H??H?H?E?H?E?H;E?s)H?E?H??H??H??!H?H?0H?E?H?8蝙????x?H?E?H;E?s)H?E?H??H??H???!H?H?0H?E?H?8?j?????~#H?E?H??H??H???!H?H?@?????H?E?H?@H??y H?E?H??H??H???!H?H?@?????!H?E?H??H??H?d?!H?H?E?H?@H?BH?E?H?@H??x#H?E?H?@H??H??H???!H?H?H??H?H?E?H?E?H?E?H;E???????5H?E?H?@H??x#H?E?H?@H??H??H???!H?H?H??H?H?E?H?E?H;E?r?H?y?!H?~?!H?{?!H???!H???!H?.?!H?w?!H?E?H?@0H????H?E?Hc?H?p?!H?ǹlw@? ????H?E?H?E??sH?E?H??H??H?:?!H?H?PH?E?H?@@H9?|H???!H??H???!H???!H?E?H??H?H?E?H??H??H???!H?H?@H?H?E?H?E?H?E?H;E?}#H?E?H??H??H???!H?H?@H???`???H?E?H?a?!?H?E???H?x?!H?H?H?E?H??H??H?r?!H?H?@H??H?H?E?H??H??H?Q?!H?H?@H?H?E?H??H??H?4?!H?H?@H???uH???!H??H???!?mH?E?H??H??H??!H?H?@H???uH???!H??H???!?9H?E?H??H??H???!H?H?PH?E?H?@@H9?|H?h?!H??H?]?!H?E?H?E?H;E?????H?E?H?G?!H?T?!H?E??eH?S?!H?E?H??H?H?H??tEH?E?H??H?(?!H?E?H?@@H;E?(H?
?!H??!H?E?H??H?H?H?H???!H?E?H?E?H;E?~?H?U?H???!H?H???!H?BH???!H?BH???!H?BH???!H?B H???!H?B(H???!H?B0H???!H?B8?E??E???UH??H?}?H?u?H?E?H?@H??xRH?E?H?@H??xEH?E?H?PH?E?H?@H9?}	?E??????ZH?E?H?PH?E?H?@H9?~	?E???!H?H?@H???u#H?E?H??H??H??!H?H?BH??H?B?FH?E?H??H??H???!H?H?PH?E?H?@@H9?|H?E?H??H??H???!H?H?H??H?H?E?H?E?H;E??????H?E?H??H??H???!H?H?E?H?BH?E?H??H??H?m?!H?H?@ H?E???H?E?H??H??H?B?!H?H?P0H?E?H??H?H?H????H?E?H??H??H??!H?H?E?H??H?B(H?E?H?@@H;E?aH?E?H??H??H???!H?H?E?H??H??H???!H?H?p H?E?H??H??H???!H?H?P0H?E?H??H?H?H?H?A H?E?H?E?H;E??+???H?E?H?E?H?H;E??W???H?E?H?H?E?H?H?\?!H?E?H?PH?E?!H?E?H?P?E??E???UH??H?}?H?u?H?E?H?@H??xRH?E?H?@H??xEH?E?H?PH?E?H?@H9?}	?E??????ZH?E?H?PH?E?H?@H9?~	?E??T??H?=?!?U??H??([??UH??SH??8H?}?H?E?H?@H??~
H?E??A?H?E??AH?uп??A??S??H?E?H?p0??A??uS??H?E?H?p8?8?A??^S??H?E?H?p@?X?A??GS??H?E?H?pH???A??0S??H?E?H?pP???A??S??H?E?H?pX?ŞA??S??H?E?H?p`?ٞA???R??H?E?H?ph??A???R??H?E?H?pp??A??R??H?E?H?@H????H?E?H?@(H????H?}?H???B???H?E?H?p ??A??wR??H?E??H?M??E?gfff?E?ffffH?E?H??H??H??H???H??H)?H?]?H?E?H??HE?H?H??H)?H?U?H?}?u?:?A??
R??H?E?H?P(H?E?H??H???Z??7?A???Q??H?E?H?E?H?U?H?R H9??^????
?R???nH?E?H?@H??u?@?A??Q???UH?E?H?p?x?A??Q??H?E??&H?E?H?H?U?H??H?H?H??H?<?!???H?E?H?E?H?@H;E??H?=??!?S??H??8[??UH??H?? H?}?H?E?H?0???A??Q??H?E?H?PH?E?H?p?ȟA???P??H?E??,H?E?H?HH?E?H??H??H??H)?H?<?????H?E?H?E?H?H;E??H?}?H?? ??????UH??H???!H??~"H?=??!?LQ??H?i?!H???!?AH?_?!H??~H?ì!H???Q??H?@?!H?A?!H??~H???!H????P??H?"?!H?#?!H??~H???!H????P??H??!H??!H??~H?i?!H???P??H??!H??!H??~H?K?!H???P??H?ȫ!H?ɫ!H??~H?-?!H???]P??H???!H???!H??~H??!H???7P??H???!H???!H??~H??!H???P??H?n?!H?o?!H??~H?ӫ!H????O??H?P?!H?Q?!H??~H???!H????O??H?2?!H???!H??~H???!H???O??H???!H???!H??~H???!H???yO??H?v?!??Ð??UH??H??0H?}?H?u?H?U?H?M?H?E?H?H??y
H?E??YH?E?H?H;E?|
H?E?H?E??BH?E?H?H??~	H?}??
O??H?E?H?H??HU?H?E?H?H?E?H?H??H?}??N??H?E?H?E???UH??H??0H?}?H?u?H?U?M?H?E?H?H??y
H?E??H?E?H?H;E?|
H?E?H?E??tH?E?H?H??u3H?E?H?H??HU?H?E?H?H?E?H??E?H?H??H???'N??H?E??5H?E?H?H??HU?H?E?H?H?E?H??E?H?H??H??H?}???N??H?E?H?E??Ð?UH??SH??xH?}?H?u?H?U??E?H?}?????N???E??}?????}????N???E??}?~e?E??????M??H???!H???!H??tE?}?????M??H???t-?E?Hc؋E?Hc?H?5??!?}??*M??H9?u
?}??L?????u&H?=ɦ!H?U????A??N???E??????H?;?!H?P??E?H?H??<
t?E?H?H??H??!H??
?E??E?H?H??H???!H??H?E?H??!H?E??H?E?H?}??
??M??H??H?E?H?E????u?H?E?H?????mL??H?E?H?}?u?E??????H?E?H?E?H???!H?E??mH?E?H??H?M?H??H?U?H??H?u?H?}?I???!???u;H?U?H?E?H??H)?H??H??H?PH?=??!?0?A???L???E??????H?E? H?E????u?H?U?H?E?H??H)?H??H??H?E?H?E?Hc?H?}ȹ??@? ?&L??H?E?H?E??9H?E?H??HE?H?0H?U?H?? H?E?H??H?H?8??K????tH?E?H?E?H?E?H;E?|?H?U?H??H?H?H?????)K??H?r?!H?k?!H??t+H?U?H??H?H?H??????J??H???!H?8?!H??u?E??????H?(?!H?E?H??!H?E?H?E?[?AH?E???H?E?H??HE?H?0H?}???M????X??\E?????f(??E?f(??^?f(??E??X??E?H??H???H?@H????H?E?H?U?H?M?H?u?H?}?L?E?L?M?L?U?H????????????H????????????H????????????H????????????H????????????L????????????L??L?ֿp?A????H?E?H??x???H;E??7fW??E?f.???????H?U?H?E?H??H?H?8H??P???????????H??????H?E??E??XE??E??E??XE??E?fW?f.E?uz?>?M?????X??\E????f(??E?f(??^?f(??E??X??E?H??H???H?@H??~mH?E?H?U?H?M?H?u?H?}?L?E?H????????????H????????????fW?H????????????H????????????H??L?ƿ??A??w??H?E?H??x???H;E???????H?E??LH?E?H?m?H?}?x'H??P???H?U?H??H?H?H??H??@ݝ ??????ݝ ???݅ ????]?H?}?x&H??P???H?U?H??H?H?H??H?H?@H;E?|?fW??E?f.?w??E??XE??E?H??H???H?@H??~ZH?E?H?U?H?M?H?u?H?}?H????????????H????????????H????????????fW?H??H?????A??R???H?E?fW??E?f.??????fW??E?f.?w?-H??(???H?PH??0???H?@HH??H??E??^E??@H??P???H??????Dž??????????UH??SH??XH?}?H?u?H?U?H?E?H?E?H?E?H?@@H??t H?E?H?@@H?@H?E?H?E?H?@@H?@H?E?H?E?H?@(HE?H??H??H?H?H?????????H??H?E?H?H?E?H?H??u?E??????lH?E?H?E??H?E?H?H?U?H??H?H?H??H?H?E?H??HE?H?8?????H?H?E?H?H?U?H??H?H?H??H?H?E?H??HE?H?@H?BH?E?H?H?U?H??H?H?H??H?H?@H?E?H?E?H?E?H;E??a???H?E??
H?E??H?E?H?E?H;E?}%H?E?H?H?U?H??H?H?H??H?H?H;E?u?H?E?H;E?};H?E?H?H?U?H??H?H?H??H?H?E?H?P0H?E?H??H?H?H?A?H?E?H?H?U?H??H?H?H??H?H?E?H?H?E?H?H?U?H??H?H?H??H??H*E??@H?E?H?H?U?H??H?H?H??H?H?E?H?P0H?E?H??H?H?H?AH?E?H?E?H?E?H?@(H;E??????H?E?Hc?H?E?H?H?ǹ?A??d???H?E?H?@H?E??9H?E?H?pH?E?H?H?U?H??H?H?H??H?H?@H?H?E?H?PH?E?H?E?H;E?|?H?U?H?E?H?B?E??E?H??X[??UH??H?}?H?u?H?E??HH?E??@f(??\?f(??,???UH??H?}?H?u?H?E??NH?E?H?H?U?H??H?H?H??H?H?H;E?u$H?E?H?H?U?H??H?H?H??H??@?]??H?E?H?E?H?@H;E?????]?H?E?H?E??E???UH??H??H?}?H?u?H?U?H?M?L??x???H?E?H?@@H?@H?E?H?E?H?E?H?M?H?U?H?u?H?}??!]?????uDžt????????WH?E?H??uDžt?????H?E??H?E?H??HE?H?H;E?ueH??x???H?PH?E?H?@HHE?H??H??H*M?H?U?H?E?H9?HN??H*?f(??^?f(??AH?E?H?E?H?@@H?@H9E???H?U?H?E?H??H?H?H?E?H?@@H9?|H?E?H?E?H?E?H;E??N????bH?E?H??HE?H?H?E?H??x???H?PH?E?H?@HHE?H??H??H*M?H?U?H?E?H9?HN??H*?f(??^?f(??AH?E?H?E?H?@@H?@H;E??Džt?????t????Ð?UH??H??H?}?H?u?H?U?H?M?L??x???H?E?H?@@H?@H?E?H?E?H?E?H?M?H?U?H?u?H?}??U[?????uDžT????????4H?E??H?E?H??HE?H?H;E?uiH??x???H?PH?E?H?@HHE?H??H?H??X???H?}?t
??ݝ`??????ݝ`???݅`???H??X????XH?E?H?E?H?@@H?@H9E???H?U?H?E?H??H?H?H?E?H?@@H9?|H?E?H?E?H?E?H;E??J????SH??x???H?PH?E?H?@HHE?H??H?H??h???H?}?t
??ݝp??????ݝp???݅p???H??h????XH?E?H?E?H?@@H?@H;E??DžT?????T????ÐUH??H??H?}?H?u?H??x???H??p???L??h????H?E?H?M?H??x???H?u?H?}??Y?????uDžd????????H?E?H?E?H?E?H?E??~H?U?H?E?H??H?H?H????\H?U?H?E?H??H?H?H???u
H?E??9H?U?H?E?H??H?H?H??x*H?U?H?E?H??H?H?H?E?H?@@H9?}
H?E???H?E?H?}?u?E??
???X??E????H*E??E??M??i??X??]?f(??^??M??H??X??E?f(??^?f(?H?E?H??HE?HE??H*??^E?f(??Y?H?E?H???H*???f(??X?H?E?H??HE??H*?????X?f(??^?f(??Y?f(??X??E??X??E?H?E?H?E?H;E??t???H?E?H??tH?E??H*??E??^??E?H??h???H?PH??p???H?@HH??H?H?E?H?BDžd?????d????Ð??UH??H??H?}?H?u?H??x???H??p???L??h???H??p???H?@@H?@H?E?H?E?H?E??H?E?H?M?H??x???H?u?H?}??0W?????uDžd????????KH?E?H??uDžd????3H?E??H?E?H??HE?H?H;E?uVH??h???H?PH??p???H?@HHE?H??H?H?E??H*??E??^??BH?E?H??p???H?@@H?@H9E???H?U?H?E?H??H?H?H?E?H?@@H9?|*H?E??H*M?H?E?H???H*??^??E??X??E?H?E?H?E?H;E??8????=H??h???H?PH??p???H?@HHE?H??H?H?E??H*??E??^??BH?E?H??p???H?@@H?@H;E??Džd?????d????ÐUH??H???H?}?H?u?H??x???H??p???L??h???H?E??H?E?H?M?H??x???H?u?H?}??wU?????uDžT????????[H?E??H?U?H?E?H??H?H?HE?H?E?H?E?H?@@H;E??H?E?H?E?H?E?H?E???H?U?H?E?H??H?H?H?????H?U?H?E?H??H?H?H???u
H?E??H?U?H?E?H??H?H?H??x'H?U?H?E?H??H?H?H?E?H?@@H9?}H?E??rH?E?H?}?~PH?U?H?E?H9?HN??H*?H?U?H?E?H9?HN??H*??^????f(??\?f(??E??X??E???E??
???X??E?H?E?H?E?H;E??????H?E?H??tH?E??H*??E??^??E?H??h???H?PH??p???H?@HH??H?H??X????E?f.U?w??E?ݝ`?????@?ݝ`?????`??????????H???H??H???H??X???H?BH?E?H?@H??~SH??h???H?PH??p???H?@HH??H?H?@H?U?H??H?????H???H??H?????H????H?A?????DžT?????T?????UH??H?}?H?u?H?U?H?M?L?E?H?E?H?@H??M!???UH??H?? H?}?H?u?H?U?H?E?H?@H??t"H?
?M!H?E?H????A???A??/??????UH??H??H?}?H?u?H??x???H??p???L??h???H??p???H?@@H?@H?E?H?M?H??x???H?u?H?}??5R?????uDžd????????	H?E???H,?H?E?H9?HN?H?E?H?}?yH?E?H?U?H??H?=?L!???hb?Ԡ??H??L!H??L!H??uDžd????????H?E??nH?U?H?E?H??H?H?H?E?H?}?	~?E?>?1H?}?xH?E???0?E??H?}??u?E?-?H?}??u?E?.??E?]
   [-D debug_level] [-N ] [-M ] [-R rel_format] [-T results_format]
   rel_info_file  results_file 
 
Calculate and print various evaluation measures, evaluating the results  
in results_file against the relevance info in rel_info_file. 
 
There are a fair number of options, of which only the lower case options are 
normally ever used.   
 --help:
 -h: Print full help message and exit. Full help message will include
     descriptions for any measures designated by a '-m' parameter, and
     input file format descriptions for any rel_info_format given by '-R'
     and any top results_format given by '-T.'
     Thus to see all info about preference measures use
          trec_eval -h -m all_prefs -R prefs -T trec_results 
 --version:
 -v: Print version of trec_eval and exit.
 --query_eval_wanted:
 -q: In addition to summary evaluation, give evaluation for each query/topic
 --measure measure_name[.measure_params]:
 -m measure: Add 'measure' to the lists of measures to calculate and print.
    If 'measure' contains a '.', then the name of the measure is everything
    preceeding the period, and everything to the right of the period is
    assumed to be a list of parameters for the measure, separated by ','. 
    There can be multiple occurrences of the -m flag.
    'measure' can also be a nickname for a set of measures. Current 
    nicknames include 
       'official': the main measures often used by TREC
       'all_trec': all measures calculated with the standard TREC
                   results and rel_info format files.
       'set': subset of all_trec that calculates unranked values.
       'prefs': Measures not in all_trec that calculate preference measures.
 --complete_rel_info_wanted:
 -c: Average over the complete set of queries in the relevance judgements  
     instead of the queries in the intersection of relevance judgements 
     and results.  Missing queries will contribute a value of 0 to all 
     evaluation measures (which may or may not be reasonable for a  
     particular evaluation measure, but is reasonable for standard TREC 
     measures.) Default is off.
 --level_for_rel num:
 -l: Num indicates the minimum relevance judgement value needed for 
      a document to be called relevant. Used if rel_info_file contains 
      relevance judged on a multi-relevance scale.  Default is 1. 
 --nosummary:
 -n: No summary evaluation will be printed
 --Debug_level num:
 -D : Debug level.  1 and 2 used for measures, 3 and 4 for merging
     rel_info and results, 5 and 6 for input.  Currently, num can be of the
     form . and only qid will be evaluated with debug info printed.
     Default is 0.
 --Number_docs_in_coll num:
 -N : Number of docs in collection Default is MAX_LONG 
 -Max_retrieved_per_topic num:
 -M : Max number of docs per topic to use in evaluation (discard rest). 
      Default is MAX_LONG.
 --Judged_docs_only:
 -J: Calculate all values only over the judged (either relevant or  
     nonrelevant) documents.  All unjudged documents are removed from the 
     retrieved set before any calculations (possibly leaving an empty set). 
     DO NOT USE, unless you really know what you're doing - very easy to get 
     reasonable looking numbers in a file that you will later forget were 
     calculated  with the -J flag.  
 --Rel_info_format format:
 -R format: The rel_info file is assumed to be in format 'format'.  Current
    values for 'format' include 'qrels', 'prefs', 'qrels_prefs'.  Note not
    all measures can be calculated with all formats.
 --Results_format format:
 -T format: the top results_file is assumed to be in format 'format'. Current
    values for 'format' include 'trec_results'. Note not all measures can be
    calculated with all formats.
 --Zscore Zmean_file:
 -Z Zmean_file: Instead of printing the raw score for each measure, print
    a Z score instead. The score printed will be the deviation from the mean
    of the raw score, expressed in standard deviations, where the mean and
    standard deviation for each measure and query are found in Zmean_file.
    If mean is not in Zmeanfile for a measure and query, -1000000 is printed.
    Zmean_file format is ascii lines of form 
       qid  measure_name  mean  std_dev
 
 
Standard evaluation procedure:
For each of the standard TREC measures requested, a ranked list of
of relevance judgements is created corresponding to each ranked retrieved doc,
A rel judgement is set to -1 if the document was not in the pool (not in 
rel_info_file) or -2 if the document was in the pool but unjudged (some 
measures (infAP) allow the pool to be sampled instead of judged fully).  
Otherwise it is set to the value in rel_info_file. 
Most measures, but not all, will treat -1 or -2 the same as 0, 
namely nonrelevant.  Note that relevance_level is used to 
determine if the document is relevant during score calculations. 
Queries for which there is no relevance information are ignored. 
Warning: queries for which there are relevant docs but no retrieved docs 
are also ignored by default.  This allows systems to evaluate over subsets  
of the relevant docs, but means if a system improperly retrieves no docs,  
it will not be detected.  Use the -c flag to avoid this behavior. 
Usage: trec_eval [-h] [-q] {-m measure}* trec_rel_file trec_top_file
   -h: Give full help information, including other options
   -q: In addition to summary evaluation, give evaluation for each query
   -m: calculate and print measures indicated by 'measure'
       ('-m all_qrels' prints all qrels measures, '-m official' is default)
helpversionquery_eval_wantedmeasurecomplete_rel_info_wantedlevel_for_relnosummaryDebug_levelJudged_docs_onlyNumber_docs_in_collMax_retrieved_per_topicRel_info_formatResults_formatOutput_old_results_formatZscoreqrelstrec_resultshvqm:cl:nD:JN:M:R:T:oZ:trec_eval version %s
trec_eval: illegal measure '%s'
trec_eval: Quit in file '%s'
trec_eval: Illegal rel_format '%s'
trec_eval: Illegal retrieval results format '%s'
officialtrec_eval: illegal measure 'official'
alltrec_eval: Cannot initialize measure '%s'
trec_eval: Can't calculate measure '%s'
trec_eval: Can't accumulate measure '%s'
trec_eval: Can't print query measure '%s'
trec_eval: No queries with both results and relevance info
trec_eval: Can't print measure '%s'
trec_eval: cleanup failed
m@$@$@$@$@$@?@$@$@?@?@$@$@$@?@$@?@$@$@$@$@$@@$@$@$@$@$@$@$@$@.@$@$@$@$@?@$@$@$@E@?@]@?@$@?@$@$@$@$@?@trec_eval: improper measure in parameter '%s'
%s
-----------------------
Individual measure documentation for requested measures%s
%s-- No measures indicated.
   Request measure documentation using <-m measure> on command lineqrelsRel_info_file format: Standard 'qrels'
Relevance for each docno to qid is determined from rel_info_file, which 
consists of text tuples of the form 
   qid  iter  docno  rel 
giving TREC document numbers (docno, a string) and their relevance (rel,  
a non-negative integer less than 128, or -1 (unjudged)) 
to query qid (a string).  iter string field is ignored.   
Fields are separated by whitespace, string fields can contain no whitespace. 
File may contain no NULL characters. 
qrels_jgRel_info_file format: Standard 'qrels'
Relevance for each docno to qid is determined from rel_info_file, which 
consists of text tuples of the form 
   qid  ujg  docno  rel 
giving TREC document numbers (docno, a string) and their relevance (rel,  
a non-negative integer less than 128, or -1 (unjudged)) 
to query qid (a string) for a particular user judgment group. 
This allows averaging (or other operations) of appropriate evaluation measures
across multiple users, whoc may differ in their judgments. 
Fields are separated by whitespace, string fields can contain no whitespace. 
File may contain no NULL characters. 
prefsRel_info_file format: Non-standard 'prefs'
Preferences of user(s) for docs for a given qid is determined from
text_prefs_file, which consists of text tuples of the form
   qid  ujg  ujsubg  docno  rel_level
giving TREC document numbers (docno, a string) and their relevance
level (rel_level,a non-negative float) to query qid (a string) for a 
user judgment sub-group (ujsubg, a string) within a user judgment
group (ujg, a string).
Fields are separated by whitespace, string fields can contain no whitespace.
File may contain no NULL characters.

Preferences are indicated indirectly by comparing rel_level of
different docnos within the same user judgment sub group(JSG).  A
judgment sub group establishes preferences between all docnos with
non-tied rel_levels within the group. Except possibly for 0.0, the
actual values of rel_level are ignored by default; they only serve to
establish a ranking within the JSG.

If a user only expresses a preference between two docs, then that user JSG
will have 2 lines in text_prefs_file:
      qid1  ujg1  sub1 docno1  3.0
      qid1  ujg1  sub1 docno2  2.0

If a user completely ranks some small number N (5-10) of docs, then N lines 
are used.
For example:
      qid1  ujg1  sub1  docno1  3.0
      qid1  ujg1  sub1  docno2  2.0
      qid1  ujg1  sub1  docno3  0.0
      qid1  ujg1  sub1  docno4  6.0
      qid1  ujg1  sub1  docno5  0.0
      qid1  ujg1  sub1  docno6  2.0
establishes a total of 13 preferences (5 with docno4 preferred, 4 with docno1 
preferred, 2 each with docno2 and docno6 preferred).

If a given user has multiple preferences that aren't complete, the preferences
are expressed in multiple JSGs within a single JG.
For example:
      qid1  ujg1  sub1  docno1  3.0
      qid1  ujg1  sub1  docno2  2.0
      qid1  ujg1  sub1  docno3  1.0
      qid1  ujg1  sub2  docno1  2.0
      qid1  ujg1  sub2  docno2  1.0
      qid1  ujg1  sub2  docno4  3.0
expressses 5 preferences (1>2, 1>3, 2 > 3, 4>1, 4>2).  Note the duplicate
1 > 2 is not counted as a separate preference

Multiple users are indicated by different JGs.
For example:
      qid1  ujg1  sub1  docno1  3.0
      qid1  ujg1  sub1  docno2  2.0
      qid1  ujg2  sub1  docno1  0.0
      qid1  ujg2  sub1  docno3  6.0
      qid1  ujg2  sub1  docno4  2.0
      qid1  ujg2  sub2  docno1  0.0
      qid1  ujg2  sub2  docno2  8.0
expressses 5 preferences (1>2, 3>1, 4>1, 3>4, 2>1).

A Judgment Group (JG) conceptually represents preferences for a single
information need of a user at a single time.  Within a single JG, it
is an error if there are inconsistencies (doc A > doc B in one JSG,
but B > A or B == A in another).  The different JSGs within a JG are
just a mechanism tha allows expressing partial ordering within a JG.
Within a single JG, preferences are transistive:
      qid1  ujg1  sub1  docno1  3.0
      qid1  ujg1  sub1  docno2  2.0
      qid1  ujg1  sub1  docno3  1.0
      qid1  ujg1  sub2  docno2  5.0
      qid1  ujg1  sub2  docno4  4.0
expresses 5 preferences (1>2, 1>3, 2>3, 2>4, 1>4).  There is no
preference expressed between 3 and 4.

Different JGs may contain contradictory preferences, as in an earlier
example.  These disagreements are realistic and desirable: users (or
even the same user at different times) often do not agree with each
other's preferences.  Individual preference evaluation measures will
handle these contradictions (or confirmations) in different ways.

A rel_level of 0.0 by convention means that doc is non-relevant to the
topic (in that user's opinion).  it is an inconsistency (and an error)
if a doc is assigned a rel_level of 0.0 in one JSG, but a different
rel_level value in another JSG of the same JG.  Some preference
evaluation measures may handle 0.0 differently.  Thus when converting
a preference file in some other format into text_prefs format, do not
assign a rel_level of 0.0 to a docno unless it is known that docno was
considered nonrelevant.

Handling of rel_level 0.0 separately addresses the general problem
that the number of nonrelevant docs judged for a topic can be critical
to fair evaluation - adding a couple of hundred preferences involving
nonrelevant docs (out of the possibly millions or billions in a
collection) can both change the importance of the topic when averaging
and even change whether system A scores better than system B on a
topic (even given identical retrieval on the added nonrel docs).  How
to handle this correctly for preference evaluation will be an
important future research problem.
qrels_prefsRel_info_file format: Non-standard 'qrels_prefs'
The file format is exactly the same as rel_info_file format 'qrels',
however it is interpreted as a restricted 'prefs' rel_info_file.
It cannot represent some user preferences (in particular, if a single user
prefers Doc A to Doc B, and A to C, but does not express a preference
between A and C) , but it allows the standard TREC qrels file to serve as 
input for preference evaluation measures.

Read all relevance preference information from text_qrels_prefs_file.
Preferences of user(s) for docs for a given qid is determined from
text_prefs_file, which consists of text tuples of the form
   qid  ujg   docno  rel_level
giving TREC document numbers (docno, a string) and their relevance
level (rel_level,a non-negative float) to query qid (a string) for a 
 user judgment group (ujg, a string).
Fields are separated by whitespace, string fields can contain no whitespace.
File may contain no NULL characters.

Preferences are indicated indirectly by comparing rel_level of
different docnos within the same user judgment group(JG).  A
judgment group establishes preferences between all docnos with
non-tied rel_levels within the group. Except possibly for 0.0, the
actual values of rel_level are ignored by default; they only serve to
establish a ranking within the JSG.

If a user only expresses a preference between two docs, then that user JSG
will have 2 lines in text_prefs_file:
      qid1  ujg1   docno1  3.0
      qid1  ujg1   docno2  2.0

If a user completely ranks some small number N (5-10) of docs, then N lines 
are used.
For example:
      qid1  ujg1    docno1  3.0
      qid1  ujg1    docno2  2.0
      qid1  ujg1    docno3  0.0
      qid1  ujg1    docno4  6.0
      qid1  ujg1    docno5  0.0
      qid1  ujg1    docno6  2.0
establishes a total of 13 preferences (5 with docno4 preferred, 4 with docno1 
preferred, 2 each with docno2 and docno6 preferred).

A Judgment Group (JG) conceptually represents preferences for a single
information need of a user at a single time.  Within a single JG, it
is an error if there are inconsistencies (doc A > doc B in one JSG,
but B > A or B == A in another).

Different JGs may contain contradictory preferences, These
disagreements are realistic and desirable: users (or even the same
user at different times) often do not agree with each other's
preferences.  Individual preference evaluation measures will handle
these contradictions (or confirmations) in different ways.

A rel_level of 0.0 by convention means that doc is non-relevant to the
topic (in that user's opinion).  Some preference evaluation measures
may handle 0.0 differently.  Thus when converting a preference file in
some other format into text_prefs format, do not assign a rel_level of
0.0 to a docno unless it is known that docno was considered
nonrelevant.

Handling of rel_level 0.0 separately addresses the general problem
that the number of nonrelevant docs judged for a topic can be critical
to fair evaluation - adding a couple of hundred preferences involving
nonrelevant docs (out of the possibly millions or billions in a
collection) can both change the importance of the topic when averaging
and even change whether system A scores better than system B on a
topic (even given identical retrieval on the added nonrel docs).  How
to handle this correctly for preference evaluation will be an
important future research problem.
trec_resultsResults_file format: Standard 'trec_results'
Lines of results_file are of the form 
     030  Q0  ZF08-175-870  0   4238   prise1 
     qid iter   docno      rank  sim   run_id 
giving TREC document numbers (a string) retrieved by query qid  
(a string) with similarity sim (a float).  The other fields are ignored, 
with the exception that the run_id field of the last line is kept and 
output.  In particular, note that the rank field is ignored here; 
internally ranks are assigned by sorting by the sim field with ties  
broken deterministicly (using docno). 
Sim is assumed to be higher for the docs to be retrieved first. 
File may contain no NULL characters. 
Lines may contain fields after the run_id; they are ignored. 
Process for evaluating qrels and trec_resultsProcess for evaluating qrels_jg and trec_resultsProcess for evaluating prefs and trec_results   Copyright (c) 2008 - Chris Buckley. 

   Permission is granted for use and modification of this file for
   research, non-commercial purposes. 

   Process for evaluating qrels_prefs and trec_resultstrec_eval: Negative cutoff detected
trec_eval: duplicate cutoffs detected
trec_eval: malformed pair parameters in '%s'
%s_%ld%s_%3.2f%s_%sZ%-22s	%s	%6.4f
%-22s	%s	%6.4f
%-22s	%s	%ld
Z%s%-22s	%s	%6.4f
Z%-22s	%s	%6.4f
%-22s	%s	%ld
trec_eval.get_qrels: Cannot read qrels file '%s'
trec_eval.get_qrels: Malformed line %ld
qrelstrec_eval.get_results: Cannot read results file '%s'
trec_eval.get_results: Cannot copy results file '%s'
trec_eval.get_results: Cannot close results file '%s'
trec_eval.get_results: Malformed line %ld
trec_resultstrec_eval.get_prefs: Cannot read prefs file '%s'
trec_eval.get_prefs: Malformed line %ld
prefstrec_eval.get_prefs: Cannot read prefs file '%s'
trec_eval.get_qrels_prefs: Malformed line %ld
prefs0trec_eval.get_qrels: Cannot read qrels file '%s'
trec_eval.get_qrels_jg: Malformed line %ld
trec_eval.get_qrels: duplicate docs %s
qrels_jgno queryqrelstrec_resultstrec_eval.form_res_qrels: rel_info format not qrels or results format not trec_results
trec_eval.form_res_qrels: duplicate docs %strec_eval.form_res_rels: duplicate docs %s
no_queryno queryqrels_jgtrec_resultstrec_eval: rel_info format not qrels_jg or results format not trec_results
trec_eval.form_res_qrels: duplicate docs %sno_queryno queryDebug: Form_prefs starting query '%s'
Returned Cached Form_prefsprefsqrels_prefstrec_resultstrec_eval.form_prefs_info: prefs_info format not (prefs or qrels_prefs) or results format not trec_results
trec_eval.form_prefs_counts: Internal docid %ld occurs with different rel_level in same jsg
trec_eval.form_prefs_counts: doc '%s' has both 0 and non-0 rel_level assigned
trec_eval.form_prefs_counts: Pref inconsistency found
      internal rank %ld and internal rank %ld are conflicted
After input, before rankstrec_eval.form_prefs_counts: duplicate docs %sAfter -M, ranksInput, before ranksAfter marking not judgedAfter assigning docid_ranksForm_prefs: num_judged %ld, num_judged_ret %ld
Final prefsPrefs_and_ranks Dump.  num_pref_lines %ld,  %s
  %s	%s	%4.2f	%s	%3ld
Docno_results Dump.  num_results %ld, %s
  %s	%4.2f	%3ld
    EC Dump. Rel_level %4.2f. Num_docid_ranks %ld
      %3ld     Prefs_Array Dump. Num_judged %ld
      Row %3ld
          (%ld)
     %2hhd  Counts_Array Dump. Num_judged %ld
    Row %3ld
        (%ld)
   %2hd ECPrefs_array  JG Dump.  Type %s
    num_prefs_fulfilled_ret %ld
    num_prefs_possible_ret %ld
    num_prefs_fulfilled_imp %ld
    num_prefs_possible_imp %ld
    num_prefs_possible_notoccur %ld
    num_nonrel %ld
    num_nonrel_ret %ld
    num_rel %ld
    num_rel_ret %ld
    Rel_array Dump. %ld values%4.2f     JG is not initialized (0 ECs and no rel_array    Dump of %ld ECs within JG
Results_prefs Dump.  %ld Judgment Groups
  num_judged_ret %ld,  num_judged %ld
no_querytrec_eval.get_zscores: Cannot read zscores file '%s'
trec_eval.get_zscores: Malformed line %ld
runidnum_qnum_retnum_relnum_rel_retmapgm_mapRprecbprefrecip_rankiprec_at_recallPrelstringrecallinfAPgm_bprefutility11pt_avgndcgrelative_PRprec_multsuccessmap_cutndcg_cutndcg_relRndcgbinGGset_Pset_recallset_relative_Pset_mapset_Fnum_nonrel_judged_retprefs_num_prefs_possprefs_num_prefs_fulprefs_num_prefs_ful_retprefs_simpprefs_pairprefs_avgjgprefs_avgjg_Rnonrelprefs_simp_retprefs_pair_retprefs_avgjg_retprefs_avgjg_Rnonrel_retprefs_simp_impprefs_pair_impprefs_avgjg_impmap_avgjgP_avgjgRprec_mult_avgjgofficialsetall_trecall_prefsprefsqrels_jgmap    Mean Average Precision
    Precision measured after each relevant doc is retrieved, then averaged
    for the topic, and then averaged over topics (if more than one).
    This is the main single-valued number used to compare the entire rankings
    of two or more retrieval methods.  It has proven in practice to be useful
    and robust.
    The name of the measure is unfortunately inaccurate since it is 
    calculated for a single topic (and thus don't want both 'mean' and
    'average') but was dictated by common usage and the need to distiguish
    map from Precision averaged over topics (I had to give up my attempts to
    call it something else!)
    History: Developed by Chris Buckley after TREC 1.
    Cite: 'Retrieval System Evaluation', Chris Buckley and Ellen Voorhees.
    Chapter 3 in TREC: Experiment and Evaluation in Information Retrieval
    edited by Ellen Voorhees and Donna Harman.  MIT Press 2005
P    Precision at cutoffs
    Precision measured at various doc level cutoffs in the ranking.
    If the cutoff is larger than the number of docs retrieved, then
    it is assumed nonrelevant docs fill in the rest.  Eg, if a method
    retrieves 15 docs of which 4 are relevant, then P20 is 0.2 (4/20).
    Precision is a very nice user oriented measure, and a good comparison
    number for a single topic, but it does not average well. For example,
    P20 has very different expected characteristics if there 300
    total relevant docs for a topic as opposed to 10.
    Note:   trec_eval -m P.50 ...
    is different from 
            trec_eval -M 50 -m set_P ...
    in that the latter will not fill in with nonrel docs if less than 50
    docs retrieved
    Cutoffs must be positive without duplicates
    Default param: -m P.5,10,15,20,30,100,200,500,1000
num_q    Number of topics results averaged over.  May be different from
    number of topics in the results file if -c was used on the command line 
    in which case number of topics in the rel_info file is used.
num_ret    Number of documents retrieved for topic. 
    May be affected by Judged_docs_only and Max_retrieved_per_topic command
    line parameters (as are most measures).
    Summary figure is sum of individual topics, not average.
num_rel    Number of relevant documents for topic. 
    May be affected by Judged_docs_only and Max_retrieved_per_topic command
    line parameters (as are most measures).
    Summary figure is sum of individual topics, not average.
qrelsqrels_jgtrec_eval: m_num_rel: rel_info format not qrels or qrels_jg
num_rel_ret    Number of relevant documents retrieved for topic. 
    May be affected by Judged_docs_only and Max_retrieved_per_topic command
    line parameters (as are most measures).
    Summary figure is sum of individual topics, not average.
gm_map    Geometric Mean Average Precision
    This is the same measure as 'map' (see description of 'map') on an
    individual topic, but the geometric mean is calculated when averaging
    over topics.  This rewards methods that are more consistent over topics
    as opposed to methods which do very well for some topics but very poorly
    for others.
    gm_ap is reported only in the summary over all topics, not for individual
    topics.
?h㈵??>Rprec    Precision after R documents have been retrieved.
    R is the total number of relevant docs for the topic.  
    This is a good single point measure for an entire retrieval
    ranking that averages well since each topic is being averaged
    at an equivalent point in its result ranking.
    Note that this is the point that Precision = Recall.
    History: Originally developed for IR rankings by Chris Buckley
    after TREC 1, but analogs were used in other disciplines previously.
    (the point where P = R is an important one!)
    Cite: 'Retrieval System Evaluation', Chris Buckley and Ellen Voorhees.
    Chapter 3 in TREC: Experiment and Evaluation in Information Retrieval
    edited by Ellen Voorhees and Donna Harman.  MIT Press 2005
recip_rank    Reciprocal Rank of the first relevant retrieved doc.
    Measure is most useful for tasks in which there is only one relevant
    doc, or the user only wants one relevant doc.
??bpref    Main binary preference measure.
    Fraction of the top R nonrelevant docs that are retrieved after each
    relevant doc. Put another way: when looking at the R relevant docs, and
    the top R nonrelevant docs, if all relevant docs are to be preferred to
    nonrelevant docs, bpref is the fraction of the preferences that the
    ranking preserves.
    Cite: 'Retrieval Evaluation with Incomplete Information', Chris Buckley
    and Ellen Voorhees. In Proceedings of 27th SIGIR, 2004.
??iprec_at_recall    Interpolated Precision at recall cutoffs.
    This is the data shown in the standard Recall-Precision graph.
    The standard cutoffs and interpolation are needed to average data over
    multiple topics; otherwise, how is a topic with 5 relevant docs averaged
    with a topic with 3 relevant docs for graphing purposes?  The Precision 
    interpolation used here is
      Int_Prec (rankX) == MAX (Prec (rankY)) for all Y >= X.
    Default usage: -m iprec_at_recall.0,.1,.2,.3,.4,.5,.6,.7,.8,.9,1 ...
????????recall    Recall at cutoffs
    Recall (relevant retrieved / relevant) measured at various doc level
    cutoffs in the ranking. If the cutoff is larger than the number of docs
    retrieved, then it is assumed nonrelevant docs fill in the rest.
    REcall is a fine single topic measure, but does not average well.
    Cutoffs must be positive without duplicates
    Default param: -m recall.5,10,15,20,30,100,200,500,1000
Rprec_mult    Precision measured at multiples of R (num_rel).
    This is an attempt to measure topics at the same multiple milestones
    in a retrieval (see explanation of R-prec), in order to determine
    whether methods are precision oriented or recall oriented.  If method A
    dominates method B at the low multiples but performs less well at the
    high multiples then it is precision oriented (compared to B).
    Default param: -m Rprec_mult.0.2,0.4,0.6,0.8,1.0,1.2,1.4,1.6,1.8,2.0 ...
????????utility    Set utility measure
    Set evaluation based on contingency table:
                        relevant  nonrelevant
       retrieved            a          b
       nonretrieved         c          d
    where  utility = p1 * a + p2 * b + p3 * c + p4 * d
    and p1-4 are parameters (given on command line in that order).
    Conceptually, each retrieved relevant doc is worth something positive to
    a user, each retrieved nonrelevant doc has a negative worth, each 
    relevant doc not retrieved may have a negative worth, and each
    nonrelevant doc not retrieved may have a (small) positive worth.
    The overall measure is simply a weighted sum of these values.
    If p4 is non-zero, then '-N num_docs_in_coll' may also be needed - the
    standard results and rel_info files do not contain that information.
    Default usage: -m utility.1.0,-1.0,0.0,0.0 ...
    Warning: Current version summary evaluation averages over all topics;
    it could be argued that simply summing is more useful (but not backward
    compatible)
trec_eval.calc_utility: improper number of coefficients
11pt_avg    Interpolated Precision averaged over 11 recall points
    Obsolete, only use for comparisons of old runs; should use map instead.
    Average interpolated at the given recall points - default is the
    11 points being reported for ircl_prn.
    Both map and 11-pt_avg (and even R-prec) can be regarded as estimates of
    the area under the standard ircl_prn curve.
    Warning: name assumes user does not change default parameter values:
    measure name is independent of parameter values and number of parameters.
    Will actually average over all parameter values given.
    To get 3-pt_avg as in trec_eval version 8 and earlier, use
      trec_eval -m 11-pt_avg.0.2,0.5,0.8 ...
    Default usage: -m 11-pt_avg.0.0,.1,.2,.3,.4,.5,.6,.7,.8..9,1.0
trec_eval.calc_m_11ptavg: No cutoff values
????????ndcg    Normalized Discounted Cumulative Gain
    Compute a traditional nDCG measure according to Jarvelin and
    Kekalainen (ACM ToIS v. 20, pp. 422-446, 2002)
    Gain values are set to the appropriate relevance level by default.  
    The default gain can be overridden on the command line by having 
    comma separated parameters 'rel_level=gain'.
    Eg, 'trec_eval -m ndcg.1=3.5,2=9.0,4=7.0 ...'
    will give gains 3.5, 9.0, 3.0, 7.0 for relevance levels 1,2,3,4
    respectively (level 3 remains at the default).
    Gains are allowed to be 0 or negative, and relevance level 0
    can be given a gain.
    Based on an implementation by Ian Soboroff
ndcg: %ld %ld %3.1f %6.4f %3.1f %6.4f
ndcg_cut    Normalized Discounted Cumulative Gain at cutoffs.
    Compute a traditional nDCG measure according to Jarvelin and
    Kekalainen (ACM ToIS v. 20, pp. 422-446, 2002) at cutoffs.
    See comments for ndcg.
    Gain values are the relevance values in the qrels file.  For now, if you
    want different gains, change the qrels file appropriately.
    Cutoffs must be positive without duplicates
    Default params: -m ndcg_cut.5,10,15,20,30,100,200,500,1000
    Based on an implementation by Ian Soboroff
ndcg_cut: cutoff %ld dcg %6.4f
ndcg_cut:%ld %3.1f %6.4f
ndcg_cut: cutoff %ld idcg %6.4f
ndcg_cut:%ld %ld %3.1f %6.4f
Rndcg    Normalized Discounted Cumulative Gain at R levels
    Experimental measure
    Compute a traditional nDCG measure according to Jarvelin and
    Kekalainen (ACM ToIS v. 20, pp. 422-446, 2002), averaged at the various
    R level points. The R levels are the number of docs at each non-negative
    gain level in the judgments, with the gain levels sorted in decreasing
    order. Thus if there are 5 docs with gain_level 3, 3 with gain 2, 10
    with gain 1, and 50 with gain 0, then 
    Rndcg = 1/4 (ndcg_at_5 + ndcg_at_8 + ndcg_at_18 + ndcg_at_68).
    In this formulation, all unjudged docs have gain 0.0, and thus there is
    a final implied R-level change at num_retrieved.
    Idea behind Rndcg, is that the expected value of ndcg is a smoothly
    decreasing function, with discontinuities upward at each transistion
    between positive gain levels in the ideal ndcg.  Once the gain level 
    becomes 0, the expected value of ndcg then increases until all docs are
    retrieved. Thus averaging ndcg is problematic, because these transistions
    occur at different points for each topic.  Since it is not unusual for
    ndcg to start off near 1.0, decrease to 0.25, and then increase to 0.75
    at various cutoffs, the points at which ndcg is measured are important.
    
    Gain values are set to the appropriate relevance level by default.  
    The default gain can be overridden on the command line by having 
    comma separated parameters 'rel_level=gain'.
    Eg, 'trec_eval -m Rndcg.1=3.5,2=9.0,4=7.0 ...'
    will give gains 3.5, 9.0, 3.0, 7.0 for relevance levels 1,2,3,4
    respectively (level 3 remains at the default).
    Gains are allowed to be 0 or negative, and relevance level 0
    can be given a gain.
Rndcg: %ld %ld %3.1f %6.4f %3.1f %6.4f %6.4f
Rndcg: %ld %ld %3.1f %6.4f %3.1f %6.4f
ndcg_rel    Normalized Discounted Cumulative Gain averaged over rel docs
    Experimental measure
    Compute a traditional nDCG measure according to Jarvelin and
    Kekalainen (ACM ToIS v. 20, pp. 422-446, 2002), averaged at rel docs.
    Idea behind ndcg_rel, is that the expected value of ndcg is a smoothly
    decreasing function, with discontinuities upward at each transistion
    between positive gain levels in the ideal ndcg.  Once the gain level 
    becomes 0, the expected value of ndcg then increases until all rel docs are
    retrieved. Thus averaging ndcg is problematic, because these transistions
    occur at different points for each topic.  Since it is not unusual for
    ndcg to start off near 1.0, decrease to 0.25, and then increase to 0.75
    at various cutoffs, the points at which ndcg is measured are important.
    This version averages ndcg over each relevant doc, where relevant is
    defined as expected gain > 0.  If a rel doc is not retrieved, then
    ndcg for the doc is the dcg at the end of the retrieval / ideal dcg. 
    
    Gain values are set to the appropriate relevance level by default.  
    The default gain can be overridden on the command line by having 
    comma separated parameters 'rel_level=gain'.
    Eg, 'trec_eval -m ndcg_rel.1=3.5,2=9.0,4=7.0 ...'
    will give gains 3.5, 9.0, 3.0, 7.0 for relevance levels 1,2,3,4
    respectively (level 3 remains at the default).
    Gains are allowed to be 0 or negative, and relevance level 0
    can be given a gain.
ndcg_rel: %ld %ld %3.1f %6.4f %3.1f %6.4f %6.4f
ndcg_rel: %ld %ld %3.1f %6.4f %3.1f %6.4f
ndcg_rel: %ld %ld %6.4f %6.4f %6.4f
binG    Binary G
    Experimental measure. (4/10/2008)
    G is a gain related measure that combines qualities of MAP and NDCG.
    G(doc) == rel_level_gain (doc) / log2 (2+num_nonrel retrieved before doc)
    G is the average of G(doc) over all docs, normalized by
    sum (rel_level_gain).
    BinG restricts the gain to either 0 or 1 (nonrel or rel), and thus is the
    average over all rel docs of (1 / log2 (2+num_nonrel before doc))
binG: %ld %ld %6.4f
??G    Normalized Gain
    Experimental measure 4/10/2008
    G is a gain related measure that combines qualities of MAP and NDCG.
    Contribution of doc doc retrieved at rank i is 
    G(doc) == gain (doc) / log2 (2+ideal_gain(i)-results_gain(i))
    where results_gain(i) is sum gain(doc) for all docs before i
    and ideal_gain is the maximum possible results_gain(i)
    G is the sum of G(doc) over all docs, normalized by max ideal_gain.
    Gain values are set to the appropriate relevance level by default.  
    The default gain can be overridden on the command line by having 
    comma separated parameters 'rel_level=gain'.
    Eg, 'trec_eval -m G.1=3.5,2=9.0,4=7.0 ...'
    will give gains 3.5, 9.0, 3.0, 7.0 for relevance levels 1,2,3,4
    respectively (level 3 remains at the default).
    Gains are allowed to be 0 or negative, and relevance level 0
    can be given a gain.
    The idea behind G is that the contribution of a doc retrieved at i
    should not be independent of the docs before. If most docs before have
    higher gain, then the retrieval of this doc at i is nearly as good as 
    possible, and should be rewarded appropriately
G: %ld %ld %3.1f %6.4f %3.1f %6.4f %6.4f %6.4f
G: %ld %ld %3.1f %6.4f %3.1f %6.4f %6.4f
G: %ld %ld %3.1f %6.4f %3.1f %6.4f
@relative_P    Relative Precision at cutoffs
    Precision at cutoff relative to the maximum possible precision at that
    cutoff.  Equivalent to Precision up until R, and then recall after R
    Cutoffs must be positive without duplicates
    Default params: -m relative_P.5,10,15,20,30,100,200,500,1000
success    Success at cutoffs
    Success (a relevant doc has been retrieved) measured at various doc level
    cutoffs in the ranking.
    If the cutoff is larger than the number of docs retrieved, then
    it is assumed nonrelevant docs fill in the rest.
    Cutoffs must be positive without duplicates
    Default param: trec_eval -m success.1,5,10
    History: Developed by Stephen Tomlinson.
infAP    Inferred AP
    A measure that allows sampling of judgement pool: Qrels/results divided
    into unpooled, pooled_but_unjudged, pooled_judged_rel,pooled_judged_nonrel.
    My intuition of infAP:
    Assume a judgment pool with a random subset that has been judged.
    Calculate P at rel doc using only the judged higher retrieved docs,
    then average in 0's from higher docs that were not in the judgment pool.
    (Those in the pool but not judged are ignored, since they are assumed
    to be relevant in the same proportion as those judged.)
    Cite:    'Estimating Average Precision with Incomplete and Imperfect
    Judgments', Emine Yilmaz and Javed A. Aslam. CIKM 
???h㈵??>?h㈵??>map_cut    Mean Average Precision at cutoffs
    Map measured at various doc level cutoffs in the ranking.
    If the cutoff is larger than the number of docs retrieved, then
    it is assumed nonrelevant docs fill in the rest.
    Map itself is precision measured after each relevant doc is retrieved,
    averaged over all relevant docs for the topic.
    Cutoffs must be positive without duplicates
    Default param: -m map_cut.5,10,15,20,30,100,200,500,1000
gm_bpref   Binary preference (bpref), but using goemetric mean over topics
    See the explanation for 'bpref' for the base measure for a single topic.
    Gm_bpref uses the geometric mean to combine the single topic scores.
    This rewards methods that are more consistent across topics as opposed to
    high scores for some topics and low scores for others.
    Gm_bpref is printed only as a summary measure across topics, not for the
    individual topics.
gm_bpref: bpref %6.4f, gm_bpref %6.4f???h㈵??>runid    Runid given by results input file.
%-22s	%s	%s
relstring    The relevance values for the first N (default 10) retrieved docs
    are printed as a string, one character per relevance value for a doc.
    If the relevance value is between 0 and 9, it is printed.
    If the value is > 9,  '>' is printed.
    If the document was not in the pool to be judged, '-' is printed.
    if the document was in the pool, but unjudged (eg, infAP),  '.' is printed
    if the document has some other relevance value, '<' is printed.
    Measure is only printed for individual queries.
    Default usage:  -m relstring.10 
%-22s	%s	'%s'
set_P    Set Precision: num_relevant_retrieved / num_retrieved 
    Precision over all docs retrieved for a topic.
    Was known as exact_prec in earlier versions of trec_eval
    Note:   trec_eval -m P.50 ...
    is different from 
            trec_eval -M 50 -m set_P ...
    in that the latter will not fill in with nonrel docs if less than 
    50 docs retrieved
set_recall    Set Recall: num_relevant_retrieved / num_relevant 
    Recall over all docs retrieved for a topic.
    Was known as exact_recall in earlier versions of trec_evalset_relative_P    Relative Set Precision:  P / (Max possible P for this size set) 
    Relative precision over all docs retrieved for a topic.
    Was known as exact_relative_prec in earlier versions of trec_eval
    Note:   trec_eval -m relative_P.50 ...
    is different from 
            trec_eval -M 50 -m set_relative_P ...
set_map    Set map: num_relevant_retrieved**2 / (num_retrieved*num_rel)
    Unranked set map, where the precision due to all relevant retrieved docs
    is the set precision, and the precision due to all relevant not-retrieved
    docs is set to 0.
    Was known as exact_unranked_avg_prec in earlier versions of trec_eval.
    Another way of loooking at this is  Recall * Precision on the set of
    docs retrieved for a topic.
set_F      Set F measure: weighted harmonic mean of recall and precision
    set_Fx = (x+1) * P * R / (R + x*P)
    where x is the relative importance of R to P (default 1.0).
    Default usage: trec_eval -m set_F.1.0 ...
    Cite: Variant of van Rijsbergen's E measure ('Information Retrieval',
    Butterworths, 1979).
??num_nonrel_judged_ret    Number of non-relevant judged documents retrieved for topic. 
    Not an evaluation number per se, but gives details of retrieval results.
    Summary figure is sum of individual topics, not average.
prefs_num_prefs_poss    Number of possible prefs independent of whether documents retrieved
    Summary figure is sum of individual topics, not average.
prefs_num_prefs_ful    Number of prefs fulfilled
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), and counts as failure if neither A nor B retrieved.
    Summary figure is sum of individual topics, not average.
prefs_num_prefs_ful_ret    Number of prefs fulfilled among retrieved docs
    For doc pref A>B, both A nd B must be retrieved to be counted.
    Summary figure is sum of individual topics, not average.
prefs_simp    Simple ratio of preferences fulfilled to preferences possible.
    If a doc pair satisfies two preferences, both are counted.
    If preferences are conflicted for a doc pair, all are counted
    (and thus max possible score may be less than 1.0 for topic).
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), and counts as failure if neither A nor B retrieved.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_pair   Average over doc pairs of preference ratio for that pair.
    If a doc pair satisfies 3 preferences but fails 2 preferences (preferences
    from 5 different users),  then the score for doc pair is 3/5.
    Same as prefs_simp if there are no doc_pairs in multiple judgment groups.
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), and counts as failure if neither A nor B retrieved.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_avgjg    Simple ratio of preferences fulfilled to preferences possible
    within a judgment group, averaged over jgs.  I.e., rather than considering
    all preferences equal (prefs_simp), consider all judgment groups equal.
    prefs_avgjg = AVERAGE_OVER_JG (fulfilled_jg / possible_jg);
    May be useful in applications where user satisfaction is represented
    by a jg per user, and it is not desirable for many preferences expressed
    by user1 to swamp a few preferences by user2.
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), and counts as failure if neither A nor B retrieved.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_avgjg_Rnonrel    Ratio of preferences fulfilled to preferences possible within a
    judgment group, averaged over jgs, except that the number of
    nonrelevant retrieved docs (rel_level == 0.0) in each jg is set to
    R, the number of relevant retrieved docs (rel_level > 0.0) in that jg.
    
    This addresses the general problem that the number of
    nonrelevant docs judged for a topic can be critical to fair
    evaluation - adding a couple of hundred preferences involving
    nonrelevant docs (out of the possibly millions in a collection) can
    both change the importance of the topic when averaging and even
    change whether system A scores better than system B (even given
    identical retrieval on the added nonrel docs).
    
    This measure conceptually sets the number of nonrelevant retrieved
    docs of a jg to R. If the actual number, N, is less than R, then R
    * (R-N) fulfilled preferences are added.  If N is greater than R,
    then only the first R (rank order) docs in the single ec with
    rel_level = 0.0 are used and the number of preferences are
    recalculated.  
    If there is a single jg with two equivalence classes (one of them 0.0), 
    then prefs_avgjg_Rnonrel is akin to the ranked measure bpref.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_simp_ret    Simple ratio of preferences fulfilled to preferences possible among.
    the retrieved docs. 
    If a doc pair satisfies two preferences, both are counted.
    If preferences are conflicted for a doc pair, all are counted
    (and thus max possible score may be less than 1.0 for topic).
    For doc pref A>B, A and B must both be retrieved to be counted as either
    fulfilled or possible.
    pref_*_ret measures should be used for dynamic collections but are
    inferior in most other applications.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_pair_ret   Average over doc pairs of preference ratio for that pair.
    If a doc pair satisfies 3 preferences but fails 2 preferences (preferences
    from 5 different users),  then the score for doc pair is 3/5.
    Same as prefs_simp if there are no doc_pairs in multiple judgment groups.
    For doc pref A>B, A and B must both be retrieved to be counted as either
    fulfilled or possible.
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), and counts as failure if neither A nor B retrieved.
    pref_*_ret measures should be used for dynamic collections but are
    inferior in most other applications.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_avgjg_ret    Simple ratio of preferences fulfilled to preferences possible
    within a judgment group, averaged over jgs.  I.e., rather than considering
    all preferences equal (prefs_simp), consider all judgment groups equal.
    prefs_avgjg = AVERAGE_OVER_JG (fulfilled_jg / possible_jg);
    May be useful in applications where user satisfaction is represented
    by a jg per user, and it is not desirable for many preferences expressed
    by user1 to swamp a few preferences by user2.
    For doc pref A>B, A and B must both be retrieved to be counted as either
    fulfilled or possible.
    pref_*_ret measures should be used for dynamic collections but are
    inferior in most other applications.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_avgjg_Rnonrel_ret    Ratio of preferences fulfilled to preferences possible within a
    judgment group, averaged over jgs, except that the number of
    nonrelevant retrieved docs (rel_level == 0.0) in each jg is set to
    R, the number of relevant retrieved docs (rel_level > 0.0) in that jg.
    
    This addresses the general problem that the number of
    nonrelevant docs judged for a topic can be critical to fair
    evaluation - adding a couple of hundred preferences involving
    nonrelevant docs (out of the possibly millions in a collection) can
    both change the importance of the topic when averaging and even
    change whether system A scores better than system B (even given
    identical retrieval on the added nonrel docs).
    
    This measure conceptually sets the number of nonrelevant retrieved
    docs of a jg to R. If the actual number, N, is less than R, then R
    * (R-N) fulfilled preferences are added.  If N is greater than R,
    then only the first R (rank order) docs in the single ec with
    rel_level = 0.0 are used and the number of preferences are
    recalculated.  
    If there is a single jg with two equivalence classes (one of them 0.0), 
    then prefs_avgjg_Rnonrel is akin to the ranked measure bpref.
    For doc pref A>B, A and B must both be retrieved to be counted as either
    fulfilled or possible.
    pref_*_ret measures should be used for dynamic collections but are
    inferior in most other applications.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_simp_imp    Simple ratio of preferences fulfilled to preferences possible.
    If a doc pair satisfies two preferences, both are counted.
    If preferences are conflicted for a doc pair, all are counted
    (and thus max possible score may be less than 1.0 for topic).
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), but ignores pair if neither A nor B retrieved.
    pref_*_imp measures don't have any preferred applications that I know of,
    but some people like them.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_pair_imp   Average over doc pairs of preference ratio for that pair.
    If a doc pair satisfies 3 preferences but fails 2 preferences (preferences
    from 5 different users),  then the score for doc pair is 3/5.
    Same as prefs_simp if there are no doc_pairs in multiple judgment groups.
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), but ignores pair if neither A nor B retrieved.
    pref_*_imp measures don't have any preferred applications that I know of,
    but some people like them.
    Assumes '-R prefs' or '-R qrels_prefs'
prefs_avgjg_imp    Simple ratio of preferences fulfilled to preferences possible
    within a judgment group, averaged over jgs.  I.e., rather than considering
    all preferences equal (prefs_simp), consider all judgment groups equal.
    prefs_avgjg = AVERAGE_OVER_JG (fulfilled_jg / possible_jg);
    May be useful in applications where user satisfaction is represented
    by a jg per user, and it is not desirable for many preferences expressed
    by user1 to swamp a few preferences by user2.
    For doc pref A>B, this includes implied preferences (only one of A or B
    retrieved), but ignores pair if neither A nor B retrieved.
    pref_*_imp measures don't have any preferred applications that I know of,
    but some people like them.
    Assumes '-R prefs' or '-R qrels_prefs'
map_avgjg    Mean Average Precision over judgment groups 
    Precision measured after each relevant doc is retrieved, then averaged
    for the topic, and then averaged over judgement group (user) and then 
    averaged over topics (if more than one).
    Same as the workhorse measure 'map' except if there is more than one
    set of relevance judgments for this query (each set indicated by a
    different judgment group), the score will be averaged over the judgment
    groups.
Rprec_mult_avgjg    Precision measured at multiples of R(num_rel) averged over users.
    This is an attempt to measure topics at the same multiple milestones
    in a retrieval (see explanation of R-prec), in order to determine
    whether methods are precision oriented or recall oriented.  If method A
    dominates method B at the low multiples but performs less well at the
    high multiples then it is precision oriented (compared to B).
    If there is more than one judgment group (set of evalutation judgments
    of a user), then the measure is averaged over those jgs.
    Default param: 
    trec_eval -m Rprec_mult_avgjg.0.2,0.4,0.6,0.8,1.0,1.2,1.4,1.6,1.8,2.0  ...
????????P_avgjg    Precision at cutoffs, averaged over judgment groups (users)
    Precision measured at various doc level cutoffs in the ranking.
    If the cutoff is larger than the number of docs retrieved, then
    it is assumed nonrelevant docs fill in the rest.  Eg, if a method
    retrieves 15 docs of which 4 are relevant, then P20 is 0.2 (4/20).
    If there are multiple relevance judgment sets for this query, Precision
    is averaged over the judgment groups.
    Cutoffs must be positive without duplicates
    Default param: trec_eval -m P.5,10,15,20,30,100,200,500,1000
yaap    Yet Another Average Precision
    Adaptation of MAP proposed by Stephen Robertson to get a value
    that is more globally averagable than MAP.  Should be monotonic with
    MAP on a single topic, but handles extreme values better.
    log ((1 + sum_probrel)  /  (1 + num_rel - sum_probrel))
    where sum_probrel = sum over all rels of (numrel_before_it / current rank)
    Cite: 'On Smoothing Average Precision', Stephen Robertson.
    ECIR 2012, LNCS 7224, pp.158-169.  2012.
    Edited by R.Baeza-Yates et al. Springer-Verlag Berlin
??;|??????R????2?????????????0??8???X???x???????m???%?????????8 
??Xj??x????G????????????????82??X???x???????0???K??????	???8	???X	???x	????	h???	|???	????	F??
0??8
h??X
l??x
s$???
?$???
?'???
X(???
G.???.??8<2??X?2??xV8????8???L$@?A?C
E???%@?A?C
E??x'@JA?C
E???)@3A?C
E??+@?A?C
E?<?-@LA?C
E?\?/@$A?C
|0@ZA?C
?i0@?A?C
??0@?A?C
??1@?A?C
?T2@A?C
o2@vA?C
<?2@?A?C
\?3@A?C
|?3@?A?C
?54@?A?C
??4@A?C
E???5@A?C
?6@?A?C
?6@?A?C
<?7@A?C
\?8@A?C
|?8@?A?C
??9@?A?C
??:@8A?C
??;@A?C
??<@A?C
H??A@QA?C
<B@?A?C
\E@?A?C
|?E@?A?C
??K@QA?C
??K@?A?C
??O@?A?C
?4P@zA?C
H??U@QA?C
<?U@?A?C
\?Y@?A?C
|DZ@?A?C
H??_@QA?C
?h_@A?C
?pb@?A?C
?c@?A?C
H??i@wA?C
<j@A?C
\m@?A?C
|?m@?	A?C
?lw@?A?C
?x@oA?C
?~x@%A?C
??x@?A?C
Py@?A?C
<?@?A?C
\??@oA?C
|!?@%A?C
?F?@?A?C
?$?@?A?C
???@rA?C
?l?@?A?C
	b?@SA?C
<	??@?A?C
\	??@|A?C
|	?@oA?C
?	?@^A?C
?	ݯ@?A?C
?	a?@?A?C
?	?@'A?C

?@?A?C
<
Ա@oA?C
\
C?@%A?C
|
h?@&A?C
?
??@?A?C
?
??@?A?C
?
8?@?A?C
E??
,?@?A?C
E?k?@BA?C
E?<??@?A?C
E?\=?@?A?C
|к@A?C
?Լ@?A?C
?a?@?A?C
?$?@?A?C
E????@QA?C
?@A?C
<?@zA?C
\??@?A?C
|v?@?A?C
?L?@?A?C
?$?@	A?C
?0?@?A?C
???@DA?C

?@bA?C
<
h?@pA?C
\
??@pA?C
|
H?@sA?C
?
??@pA?C
?
,?@_A?C
?
??@A?C
?
??@?A?C
X?@A?C
<t?@?A?C
\?@?A?C
|??@?A?C
?h?@{A?C
H????@?A?C
???@A?C
?W?@
A?C
E?d?@0A?C
<??@?A?C
\?@A?C
|8?@?A?C
???@
A?C
E????@0A?C
??@?A?C
???@?A?C
pA
A?C
E?<}A0A?C
\?A?A?C
|4A^A?C
??A	A?C
??A
A?C
E???A0A?C
??A?A?C
`A?A?C
<,A?A?C
\?AeA?C
|<A?A?C
?A?A?C
??A.A?C
?AJA?C
?PAxA?C
?AVA?C
<A?A?C
\?A?A?C
|@A?A?C
??A?A?C
?tA?A?C
?, A%A?C
?T!A?A?C
?!AA?C
<?"A?A?C
\?#A?A?C
|?$AxA?C
?&AkA?C
?x)A?A?C
?$+A?A?C
?!.A?A?C
?4A?A?C
<?5A?A?C
\?7A2A?C
|?8AsA?C
?G;A?A?C
?@ARA?C
?hAA?A?C
?LDA?A?C
?EA?A?C
<?GA
A?C
\?KA?A?C
|0NATA?C
H?zRx???$4????J??Q??Q@??????????????????;@@
hPA???o@@@x@
6?Ab?X@?@x	???o?@???o???o>@(@bn@~@?@?@?@?@?@?@?@?@@@.@>@N@^@n@~@?@?@?@?@?@?@?@?@
@
@.
@>
@N
@?PA?PAheA?fAh?fAv?fAq?fAm?fAc?fAlgAngADgAJ.gANBgAMZgARjgATygAo?gAZ?lA?lA?<@E@knAxnAc@m@?pA?pA4P@?Y@??A??ADZ@pb@??A?A?E@?O@?lA??A??A?x@knA??A?AF?@?pA??AH?Aк@??A??Ax?Aк@?A?A??A`Zb`Mb?Mb Nb?Nb Lb?Nb@ObPb?Ob?PbMb[b?Qb?XbZb?Rb@Sb Tb?Vb?Vb?Tb Vb?Ub`Ub?Yb?Wb`Xb`[b \b?[b?\b ]b?]b?]b@^b?^b_b`_b?_b `b?`b?`b@ab?abbb`bb?bb cb?dbdb@eb4\?Ab?Ah?Ap?Ax?A??A??A??A??A??A??A??A\?Ab?Ah?Ap?Ax?A??A??A??A??A??A??A??A??A AɠAϠAؠA?A?A?A??A?A?A?A?A&?A,?A1?A3?A9?AD?AS?A[?Aa?A\?Ab?Ah?Ap?Ax?AؠA3?A9?AD?AS?A[?A\?Ab?Aw?A??A??A??AáAΡAڡA?A??A?A?A4?AC?AR?A\?Ab?Aw?A??A??A??AáAΡA\?Ab?Ab?Al?At?A??A`Hb??AJb??A?Hb??A`Jb??AKb??A`Kb??A??A[@$?@o2@?3@6@?8@????????
d???	?Lbh?Ap?A? @0?@?2@54@?7@?:@?Lb????????ЩAةA @??@o2@?@?5@?9@??????????A??A @h?@o2@?3@?6@?9@??????????A??A @??@o2@H?@?6@?9@?????????A?A @??@o2@?3@?6@?9@?????????A?A[@,?@o2@?4@?5@?8@??????????A??A[@??@o2@?3@6@?8@??????????A??A[@??@o2@?3@6@?8@????????x?A??A[@X?@o2@?3@6@?8@????????????????????????333333????????????333333??ffffff????????????????????`Pbx?A??A}"@t?@?2@54@?7@?:@?Pb????????
d???	@Qb??A??A? @?@?2@54@?7@?:@?Qb????????????????????????333333????????????333333??ffffff??????????????????@
 Rb@?AP?A}"@??@?2@54@?7@?:@pRb???????????SbH?AP?A>$@h?@o2@?3@6@?;@ Sb????????????????????????333333????????????333333??ffffff?????????????????????Sb??A??A>$@??@o2@?3@6@?;@Tb??????????A??A?%@??@o2@?3@6@?;@`hb????????
d???	?Tb??A??A? @?@?2@54@?7@?:@0Ub????????8?A@?A?%@8?@o2@?3@6@?;@?hb????????h?Ax?A?%@??@o2@?3@6@?;@?hb??????????A?A[@4Ao2@?3@6@?8@??????????A??A?%@?Ao2@?3@6@?;@?hb????????
d???	@Wb?A?A? @`A?2@54@?7@?:@?Wb????????
 Xb8?A@?A? @,A?2@54@?7@?:@@Xb??????????A??A[@?Ao2@?3@6@?8@????????
d???	 Yb??A??A? @<A?2@54@?7@?:@pYb????????p?A??A[@Ao2@?4@?5@?8@??????????A??A8@?AT2@?3@?5@A????????$@?Zb??A??A>$@PAT2@?3@?AA?Zb?????????A?A[@?Ao2@?3@6@?8@??????????A??A[@@Ao2@?3@6@?8@????????@?AP?A[@?Ao2@?3@6@?8@??????????A??A[@tAo2@?3@6@?8@???????????\b@?AH?A>$@, Ao2@?3@6@?;@?\b??????????A??A @T!Ao2@?3@?6@?9@????????x?A??A @?!Ao2@?3@?6@?9@?????????A0?A @?"Ao2@?3@?6@?9@???????? ?A8?A @?#Ao2@?3@?6@?9@??????????A?A[@?$Ao2@?3@6@?8@??????????A??A[@&Ao2@?3@6@?8@??????????A??A[@x)Ao2@?3@6@?8@????????h?A??A[@$+Ao2@?3@6@?8@??????????A??A[@?4Ao2@?3@6@?8@?????????B?B[@?5Ao2@?3@6@?8@?????????B?B[@?7Ao2@?3@6@?8@?????????B?B[@?8Ao2@?3@6@?8@?????????B?B[@@Ao2@?3@6@?8@?????????B?B[@hAAo2@?3@6@?8@????????B B[@LDAo2@?3@6@?8@????????0B@B[@?EAo2@?3@6@?8@????????????????????????333333????????????333333??ffffff??????????????????@
?cb B8B}"@?GA?2@54@?7@?:@?cb????????
d???	`db?B?B? @?KA?2@54@?7@?:@?db????????(B0B[@0NAo2@?3@6@?8@????????GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55)GCC: (GNU) 4.1.2 20080704 (Red Hat 4.1.2-55),8@?,?8@,?-T2@4,w6?3@j,???5@?,I?8@?,?R?<@?,?]?E@?
,Ei4P@
,?tDZ@?,??4A?,???5A?,??7A2,???8A=,??@AR,?hAA?,y?LDA?,???EA?,???GA
,???KA?,?0NAT??
main??\te_rel_info_formatrte_num_rel_info_format?te_results_format?te_num_results_format?te_form_inter_procs?te_num_form_inter_procs????
te_init_meas_empty=te_init_meas_s_float?te_init_meas_s_long?te_init_meas_a_float_cut_long?te_init_meas_a_float_cut_floatS
te_init_meas_s_float_p_float?
te_init_meas_s_float_p_pairL?-?qte_acc_meas_empty?te_acc_meas_sNte_acc_meas_a_cutuw6r	ite_calc_avg_meas_empty?te_calc_avg_meas_s_te_calc_avg_meas_a_cut?te_calc_avg_meas_s_gm???-	zte_print_single_meas_empty?te_print_single_meas_s_floatUte_print_single_meas_s_long?te_print_single_meas_a_cut?I?	yte_print_final_meas_empty?te_print_final_meas_s_floatMte_print_final_meas_s_long?te_print_final_meas_a_cut&	te_print_final_meas_s_float_p8?R?te_get_qrelsH
te_get_qrels_cleanupF?]?te_get_trec_results?
te_get_trec_results_cleanup8Ei?5te_get_prefs?
te_get_prefs_cleanupD?ts-te_get_qrels_prefs?
te_get_qrels_prefs_cleanup>?g=te_meas_prefs_simp_ret)??mCte_meas_prefs_pair_ret*?xMte_meas_prefs_avgjg_ret2??
?te_meas_prefs_avgjg_Rnonrel_ret)??g=te_meas_prefs_simp_imp)?mCte_meas_prefs_pair_imp*y?xMte_meas_prefs_avgjg_imp$???	v	te_meas_map_avgjg+??s
G
te_meas_Rprec_mult_avgjg"??
?	te_meas_P_avgjg?5		te_meas_yaap?6@8@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)trec_eval.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֙4unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t??45char??._flags
?#_IO_read_ptr/#_IO_read_end/#_IO_read_base/#_IO_write_base/# _IO_write_ptr/#(_IO_write_end/#0_IO_buf_base/#8_IO_buf_end/#@_IO_save_base/#H_IO_backup_base/#P_IO_save_end/#X_markers #`_chain!&#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/,#?_lock3<#?_offset<#?__pad1E-#?__pad2F-#?__pad3G-#?__pad4H-#?__pad5I?#?_modeK?#?_unused2MB#??	5
_IO_lock_t? _IO_marker?_next? #_sbuf?&#_pos??#?=
<5&?
R5&long long unsigned intlong long intdouble??option k?l?#has_argo?#flagp?#valq?#measure_name/#parameters/#MEAS_ARG??pquery_flag?#summary_flag?#debug_level?#debug_query/#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,/#Presults_format-/#Xzscore_flag.?#`meas_arg3?#hEPI4?8?9/#value<}#TREC_EVAL_VALUE=?E(@qidA/#num_queriesB?#valuesCE#num_valuesD?#max_num_valuesE?# ?TREC_EVALG??Lprintable_paramsM/#num_paramsQ?#param_valuesR-#PARAMSS\ `qida/#run_idb/#ret_formatc/#q_resultse-#RESULTSf?ihnum_q_resultsi?#max_num_q_resultsj?#resultski#ALL_RESULTSl?nqido/#rel_formatp/#q_rel_infoq-#REL_INFOr?	tnum_q_relsu?#max_num_q_relsv?#rel_infow	#?ALL_REL_INFOx?'
trec_measP|?~/#k?/#init_meas?S
#calc_meas??
#acc_meas??
# calc_avg_meas?	#(print_single_meas?)#0print_final_and_cleanup_meas?I#8meas_params?O#@eval_index??#HA
?A
G
M
?7	K'
}
?}
?
?
?
M
?
	??
	??
	?
	7	Y
?
?}
?
?
M
?
	K?
?
?}
?
?
M
	#	?
)?}
?
?
I?}
G
M
/?TREC_MEAS?7	????/#name_list??#/TREC_MEASURE_NICKNAMES?f? ???/#k?/#??#??'#?A
/#	?'?REL_INFO_FILE_FORMAT??? ???/#k?/#???#??'#??A
/?o?RESULTS_FILE_FORMAT?I
?meas?/#mean?}#stddev?}#ZSCORE_QID??T
?qid?/#num_zscores??#zscores?T
#
ZSCORES?
?
?num_q_zscores??#q_zscores֞
#Z
ALL_ZSCORES?i
}main??8@?@argc????|argv????|trec_results_file?/??all_results?o??trec_rel_info_file?/??all_rel_info?#	??~zscores_file?/??all_zscores??
??~epi????}accum_eval?K??}q_eval?K??}i???@j???Hm???Pc???\help_wanted???`measure_marked_flag???h^:@A@option_index????}long_options??	 Cb?add_meas_arg_info???@?@Lepi?A
?Xmeas?/?Pparam?/?Hi???h@mark_single_measure???@@???/?Xi???h?mark_measure??@@?epi?A
?H??/?@i???Xptr?/?`?@?@?	??h8trec_eval_help?@?@0epiA
?Hm??Xf??`m_marked??h?get_debug_level_queryC?@?@|epiBA
?X?B/?PptrD/?h ??Q??@6@?epiPA
?XiR??h
??&VersionID/	Cbhelp_message	/	Cb!stderr?&"?;/!optindI?usagew/	Cb"?}?
??#U"w~u"??
??#"??"??
?-#"X??"???
??#"F??"???"	??!stderr?&"?;/!optindI?"?}?"w~u"??"??"??"X??"???"F??"???"	???n8@8@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)formats.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval4unsigned charshort unsigned intunsigned intsigned charshort intintlong int4charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#H	?,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?H `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf??hnum_q_resultsi?#max_num_q_resultsj?#resultsk?#HALL_RESULTSlWnqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?Ztnum_q_relsu?#max_num_q_relsv?#rel_infowZ#ALL_REL_INFOx?? ?name??#	k??#	??#	w??#
??t??`???
?REL_INFO_FILE_FORMAT?zQ ?name??#	k??#	?q#	w??#
k?t?k?QRESULTS_FILE_FORMAT?? ?	???#results_info_format??#	k??#	w??#FORM_INTER_PROCSĒfloat??X1w?F?!N?? >X	 Ebte_num_rel_info_format??	?EbF?!	?Ebte_num_results_format??	?Eb >	Fbte_num_form_inter_procs?	?Fb???T2@8@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)meas_init.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֙?unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t???5char???._flags
?#_IO_read_ptr/#_IO_read_end/#_IO_read_base/#_IO_write_base/# _IO_write_ptr/#(_IO_write_end/#0_IO_buf_base/#8_IO_buf_end/#@_IO_save_base/#H_IO_backup_base/#P_IO_save_end/#X_markers#`_chain!#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/!#?_lock31#?_offset<#?__pad1E-#?__pad2F-#?__pad3G-#?__pad4H-#?__pad5I?#?_modeK?#?_unused2M7#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?#_pos??#?=15
&?G5
&long long unsigned intlong long intdouble?measure_name/#parameters/#MEAS_ARG|,pquery_flag?#summary_flag?#debug_level?#debug_query/#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,/#Presults_format-/#Xzscore_flag.?#`meas_arg3,#h?EPI4?b89/#?
?	trec_measP|~/#explanation?/#init_meas??	#calc_meas?
#acc_meas?L
# calc_avg_meas?|
#(print_single_meas??
#0print_final_and_cleanup_meas??
#8meas_params??
#@eval_index??#H?	??	?	?	2???	?	??	?	

?	?	2
.
?
??	A
??	
A
?	G
?"
q
??	
q
?	w
?R
?
??	
A
?
?
??	?	?	?
BTREC_MEAS??7te_init_meas_empty*?8@[@epi)?	?htm)7?`?)?	?X?
?te_init_meas_s_float2?[@ @`epi1?	?htm17?`?1?	?X?te_init_meas_s_longE? @? @?epiD?	?htmD7?`?D?	?X?te_init_meas_a_float_cut_longX?? @}"@?epiW?	?@tmW7???W?	???Y??PiZ??X? @C!@!^,?`?M
te_init_meas_a_float_cut_float??}"@>$@Depi??	?@tm?7?????	????M
?Pi???X?"@?"@!?,?`r?
te_init_meas_s_float_p_float??>$@?%@?epi??	?Ptm?7?H???	?@`$@?$@!?,?`bte_init_meas_s_float_p_pair???%@x'@?epi۴	?Ptm?7?H???	?@?%@Y&@!?,?`?get_long_cutoffs?x'@?)@(??
??/?????@ptr/?H?/?P???Xi??`?get_float_cutoffs<??)@?+@t?;?
??;/???=??@ptr>/?H?>/?P??M
?Xi@??`)get_float_paramsq??+@?-@??p?
?@p/???r??Hptrs/?P?s/?X?tM
?`?get_param_pairs???-@?/@???
???/??????@last_seen?5?Optr?/?P??/?X????`ucomp_long???/@0@Xptr1???hptr2???`kcomp_float??0@i0@?ptr1?M
?hptr2?M
?`?append_long?/i0@?0@??/?X????P????`??/?h@append_float?/?0@?1@<?/?X??r?P????`??/?h?append_string?/?1@T2@??/?X??/?P????`??/?hstderr?stderr??s??3@T2@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)meas_acc.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval.unsigned charshort unsigned intunsigned intsigned charshort intintlong int.charlong long unsigned intlong long intdoublexmeasure_name#parameters#MEAS_ARGC?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hxEPI4?,8name9#value<9#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# ,TREC_EVALGCLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?u `qida#run_idb#ret_formatc#q_resultse?#RESULTSf$?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#rel_infow#?ALL_REL_INFOx?	2trec_measP|name~#explanation?#init_meas?^#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?4#0print_final_and_cleanup_meas?T#8meas_params?Z#@eval_index??#H
L?LRX?9?2
??????X?????u?9d
?????X???
	???	X%?
4????
T??RX:TREC_MEAS?9
?te_acc_meas_empty?T2@o2@?epi??htm??`K??X@X?P?`
Nte_acc_meas_s'?o2@?2@ epi%??htm%??`K&??X@&X?Pte_acc_meas_a_cut1??2@?3@lepi/??Xtm/??PK0??H@0X?@i2??hn	[]?5@?3@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)meas_avg.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalRunsigned charshort unsigned intunsigned intsigned charshort intintlong intRcharlong long unsigned intlong long intdoublexmeasure_name#parameters#MEAS_ARGC?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hxEPI4?,8name9#value<9#TREC_EVAL_VALUE=?(@qidA#	qB?#valuesC?#num_valuesD?#max_num_valuesE?# ,TREC_EVALGCLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?m `qida#run_idb#ret_formatc#q_resultse?#RESULTSf?nqido#rel_formatp#q_rel_infoq?#REL_INFOr|tnum_q_relsu?#max_num_q_relsv?#rel_infow#?ALL_REL_INFOx?
*trec_measP|name~#explanation?#init_meas?V#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?,#0print_final_and_cleanup_meas?L#8meas_params?R#@eval_index??#HD?DJP?1?*??????P?
??
??
m?
1\?????P?
?????P
?,????L??JP2TREC_MEAS?1?te_calc_avg_meas_empty??3@?3@?epi??htm??`d?XevalP?P?
X_te_calc_avg_meas_s$??3@54@epi"??Xtm"??Pd#?H}#P?@q%??h?te_calc_avg_meas_a_cut2?54@?4@Pepi0??Xtm0??Pd1?H}1P?@i3??`q4??hte_calc_avg_meas_s_gmF??4@?5@?epiD??PtmD??HdE?@}EP??sumG9?XqH??`)	a??8@?5@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)meas_print_single.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdouble?measure_name	#parameters	#MEAS_ARGL?pquery_flag?#summary_flag?#debug_level?#debug_query	#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,	#Presults_format-	#Xzscore_flag.?#`meas_arg3?#h?EPI4?58name9	#value#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?"#_pos??#?D8<
-?N<
-long long unsigned intlong long intdouble?measure_name6#parameters6#MEAS_ARG?3pquery_flag#summary_flag#debug_level#debug_query6#relation_flag# average_complete_flag #(judged_docs_only_flag$#0num_docs_in_coll'#8relevance_level(#@max_num_docs_per_topic+#Hrel_info_format,6#Presults_format-6#Xzscore_flag.#`meas_arg33#h?EPI4?? `qida6#run_idb6#ret_formatc6#q_resultse4#RESULTSfD?hnum_q_resultsi#max_num_q_resultsj#resultsk?#?ALL_RESULTSl?96?E
docno6#simE#floatTEXT_RESULTS?num_text_results#max_num_text_results#text_results?#NTEXT_RESULTS_INFOb!qid"6#docno#6#sim$6#LINES%??	te_get_trec_results8??E@?K@?epi6??~text_results_file66??~all_results7??~fd9???orig_buf:6??size;???ptr<6??current_qid=6??i>??lines??	??line_ptr@?	?@num_linesA??Hnum_qidB?P?C6??q_results_ptrE??Xtext_info_ptrF?	?`text_results_ptrG??h?
comp_lines_qid_docno???K@?K@
ptr1??	?Xptr2??	?Pcmp???l?
parse_results_line???K@?O@\
start_ptr??Xqid_ptr??Pdocno_ptr??Hsim_ptr??@????ptr?6?hte_get_trec_results_cleanup???O@4P@?
stderr?"trec_results_buf06	?ebtext_info_pool1?		?ebtext_results_pool2?	fbq_results_pool3?	fbstderr?"?s

DZ@4P@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)get_prefs.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֙?unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t???5char???._flags
?#_IO_read_ptr/#_IO_read_end/#_IO_read_base/#_IO_write_base/# _IO_write_ptr/#(_IO_write_end/#0_IO_buf_base/#8_IO_buf_end/#@_IO_save_base/#H_IO_backup_base/#P_IO_save_end/#X_markers#`_chain!#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/!#?_lock31#?_offset<#?__pad1E-#?__pad2F-#?__pad3G-#?__pad4H-#?__pad5I?#?_modeK?#?_unused2M7#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?#_pos??#?=15
&?G5
&long long unsigned intlong long intdouble?measure_name/#parameters/#MEAS_ARG|,pquery_flag?#summary_flag?#debug_level?#debug_query/#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,/#Presults_format-/#Xzscore_flag.?#`meas_arg3,#h?EPI4?~nqido/#rel_formatp/#q_rel_infoq-#REL_INFOr=?tnum_q_relsu?#max_num_q_relsv?#rel_infow?#~ALL_REL_INFOx?2/?floatU 2jg3/#jsg4/#rel_level5#docno6/#TEXT_PREFS7
?9num_text_prefs:?#max_num_text_prefs;?#text_prefsqid?2#jg@2#docnoA2#relB2#LINESC??	te_get_qrels_jgU?c@?i@TepiT???~text_qrels_fileT2??~all_rel_infoT??~fdV???sizeW???ptrX2??~current_qidY2??current_jgY2??iZ???lines[?	??line_ptr\?	??num_lines]???num_qid^???num_jg^??@rel_info_ptr`??Htext_jg_info_ptra?	?Ptext_jg_ptrb??Xtext_qrels_ptrcI?`:?R
comp_lines_qid_jg_docno???i@j@?ptr1??	?Xptr2??	?Pcmp???l?
parse_qrels_line??j@m@?start_ptr???Xqid_ptr???Pjg_ptr???Hdocno_ptr???@rel_ptr????ptr?2?hte_get_qrels_jg_cleanup?m@?m@8stderr?trec_qrels_bufM2	Pfbtext_jg_info_poolN?		Xfbtext_jg_poolO?	`fbtext_qrels_poolPI	hfbrel_info_poolQ?	pfbstderr?>
?Ny@?m@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)form_res_rels.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֝Runsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t??R9char?>?._flags
?#_IO_read_ptr3#_IO_read_end3#_IO_read_base3#_IO_write_base3# _IO_write_ptr3#(_IO_write_end3#0_IO_buf_base3#8_IO_buf_end3#@_IO_save_base3#H_IO_backup_base3#P_IO_save_end3#X_markers#`_chain!#h_fileno#?#p_flags2'?#t_old_offset)
#x_cur_column-?#?_vtable_offset.?#?_shortbuf/%#?_lock35#?_offset<#?__pad1E1#?__pad2F1#?__pad3G1#?__pad4H1#?__pad5I?#?_modeK?#?_unused2M;#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?#_pos??#?A59
*?K9
*long long unsigned intlong long intdouble?measure_name3#parameters3#MEAS_ARG?0pquery_flag?#summary_flag?#debug_level?#debug_query3#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,3#Presults_format-3#Xzscore_flag.?#`meas_arg30#h?EPI4?? `qida3#run_idb3#ret_formatc3#q_resultse1#RESULTSfA?nqido3#rel_formatp3#q_rel_infoq1#REL_INFOr??6??:
docno3#sim:#floatTEXT_RESULTS?num_text_results?#max_num_text_results?#text_results?#CTEXT_RESULTS_INFOW?docno3#rel?#TEXT_QRELS?]num_text_qrels?#max_num_text_qrels ?#text_qrels"]#?TEXT_QRELS_INFO$	4	@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(GU4	#0results_rel_listY4	#8?RES_RELS]z:	?	 !docno"3#sim#:#rank$?#rel%?#DOCNO_INFO&P	?
te_form_res_rels;??m@lw@?epi9???rel_info9???results:??res_rels:J	??i??@text_results_info@?
?Htrec_qrelsA?
?Pqrels_ptrC]?Xend_qrelsC]?`?t@lu@rrl???h?c0comp_rank_judged?lw@x@?ptr10?hptr20?`?	?comp_sim_docno?x@~x@ptr10?hptr20?`?comp_docno#?~x@?x@hptr1!0?hptr2"0?`te_form_res_rels_cleanup*??x@Ny@?stderr?current_query)3	?Fbmax_current_query*?	?fbG-4		?fbmax_rel_levels.?	?fbsaved_res_rels/:		?fbranked_rel_list04		gbmax_ranked_rel_list1?	?fbdocno_info40	gbmax_docno_info5?	?fbstderr?'g?"?@Py@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)form_res_rels_jg.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֠xunsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t?__off64_t?x<char?d?._flags
?#_IO_read_ptr6#_IO_read_end6#_IO_read_base6#_IO_write_base6# _IO_write_ptr6#(_IO_write_end6#0_IO_buf_base6#8_IO_buf_end6#@_IO_save_base6#H_IO_backup_base6#P_IO_save_end6#X_markers#`_chain!"#h_fileno#?#p_flags2'?#t_old_offset)
#x_cur_column-?#?_vtable_offset.?#?_shortbuf/(#?_lock38#?_offset<#?__pad1E4#?__pad2F4#?__pad3G4#?__pad4H4#?__pad5I?#?_modeK?#?_unused2M>#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?"#_pos??#?D8<
-?N<
-long long unsigned intlong long intdouble?measure_name6#parameters6#MEAS_ARG?3pquery_flag#summary_flag#debug_level#debug_query6#relation_flag# average_complete_flag #(judged_docs_only_flag$#0num_docs_in_coll'#8relevance_level(#@max_num_docs_per_topic+#Hrel_info_format,6#Presults_format-6#Xzscore_flag.#`meas_arg33#h?EPI4?? `qida6#run_idb6#ret_formatc6#q_resultse4#RESULTSfD?nqido6#rel_formatp6#q_rel_infoq4#REL_INFOr??9??=
docno6#sim=#floatTEXT_RESULTS?num_text_results#max_num_text_results#text_results?#FTEXT_RESULTS_INFOZ?docno6#rel#TEXT_QRELS??I&num_text_qrels'#text_qrels(#TEXT_QRELS_JG*?,num_text_qrels_jg-#text_qrels_jg.?#ITEXT_QRELS_JG_INFO/^u	@Enum_rel_retG#num_retJ#num_nonpoolL#num_unjudged_in_poolN#num_relS# num_rel_levelsT#(mUu	#0results_rel_listYu	#8RES_RELS]??	`qida6#num_jgsb#jgsc?	#{	RES_RELS_JGd?	
 "docno#6#sim$=#rank%#rel&#DOCNO_INFO'?	?te_form_res_rels_jg;?Py@?@epi9???rel_info9??results:??res_rels:???i<??num_results=??jg>??text_results_info@???trec_qrelsA??@qrels_ptrC?Hend_qrelsC?Pmax_relD?Xrel_level_ptrFu	?`
?@Y?@rrl??h?	???comp_rank_judged??@??@Lptr1??hptr2??`
Acomp_sim_docno(???@!?@?ptr1&??hptr2'??`?comp_docno4?!?@F?@?ptr12??hptr23??`te_form_res_rels_jg_cleanup:?F?@"?@0stderr?"current_query*6	?Fbmax_current_query+	gbm.u		8gbmax_rel_levels/	gbjgs0?		@gbnum_jgs1	Hgbmax_num_jgs2	 gbranked_rel_list3u		Pgbmax_ranked_rel_list4	(gbdocno_info5?	Xgbmax_docno_info6	0gbstderr?"E??Ѽ@$?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)form_prefs_counts.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֡?unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t?__off64_t??=char???._flags
?#_IO_read_ptr7#_IO_read_end7#_IO_read_base7#_IO_write_base7# _IO_write_ptr7#(_IO_write_end7#0_IO_buf_base7#8_IO_buf_end7#@_IO_save_base7#H_IO_backup_base7#P_IO_save_end7#X_markers)#`_chain!/#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/5#?_lock3E#?_offset<#?__pad1E5#?__pad2F5#?__pad3G5#?__pad4H5#?__pad5I?#?_modeK?#?_unused2MK#???	_IO_lock_t?
)_IO_marker?_next?)#_sbuf?/#_pos??#?EE=
.?[=
.long long unsigned intlong long intdouble?measure_name7#parameters7#MEAS_ARG?@pquery_flag#summary_flag#debug_level#debug_query7#relation_flag# average_complete_flag #(judged_docs_only_flag$#0num_docs_in_coll'#8relevance_level(#@max_num_docs_per_topic+#Hrel_info_format,7#Presults_format-7#Xzscore_flag.#`meas_arg3@#h?EPI4?? `qida7#run_idb7#ret_formatc7#q_resultse5#RESULTSfQ?nqido7#rel_formatp7#q_rel_infoq5#REL_INFOr?F??H
7#simH#floatTEXT_RESULTS#?num_text_results#max_num_text_results#text_results?#QTEXT_RESULTS_INFOe 2jg37#jsg47#B5H#67#TEXT_PREFS7??9num_text_prefs:#max_num_text_prefs;#text_prefs?Xj>?`?debug_print_counts_arrayMk?@??@ caM?PiN?XjN?`debug_print_jg^??@=?@ljg^U?Pi_?`bdebug_print_results_prefs?=?@к@?rp??
?Xi??hte_form_pref_counts_cleanup??к@Ѽ@stdout?/stderr?/current_query?7	?Fbmax_current_query?	`gb3?	?gb??	?gbnum_jgs?	?gbjgs?U	?gbmax_num_jgs?	hgbrank_pool??	?gbmax_rank_pool?	pgbec_pool??
	?gbmax_ec_pool?	xgbca_pool?o		?gbmax_ca_pool?	?gbca_ptr_pool?u		?gbmax_ca_ptr_pool?	?gbpa_pool??	hbmax_pa_pool?	?gbpa_ptr_pool??	hbmax_pa_ptr_pool?	?gbrel_pool??
	hbmax_rel_pool?	?gb??Q	hbmax_prefs_and_ranks?	?gbdocno_results?6	 hbmax_docno_results?	?gbtemp_pa_pool??	(hbmax_temp_pa_pool?	0hbtemp_pa_ptr_pool??	8hbmax_temp_pa_ptr_pool?	@hbsaved_num_judged?	?gbstdout?/stderr?/K~?"?@Լ@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)utility_pool.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֜Zunsigned charshort unsigned intunsigned intsigned charshort intintlong intZcharlong long unsigned intlong long intdoublefloat??te_chk_and_mallocԼ@a?@Pptr?h	LX?`needed??Xsize??P
?
?Ite_chk_and_realloc,a?@"?@?ptr*?h	L*X?`needed+??Xsize+I?T
??
B??@$?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)get_zscores.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֛uunsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t??u7char?l?._flags
?#_IO_read_ptr1#_IO_read_end1#_IO_read_base1#_IO_write_base1# _IO_write_ptr1#(_IO_write_end1#0_IO_buf_base1#8_IO_buf_end1#@_IO_save_base1#H_IO_backup_base1#P_IO_save_end1#X_markers"#`_chain!(#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/.#?_lock3>#?_offset<#?__pad1E/#?__pad2F/#?__pad3G/#?__pad4H/#?__pad5I?#?_modeK?#?_unused2MD#??	7
_IO_lock_t?"_IO_marker?_next?"#_sbuf?(#_pos??#??
>7(?
T7(long long unsigned intlong long intdouble?measure_name1#parameters1#MEAS_ARG?9pquery_flag?#summary_flag?#debug_level?#debug_query1#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,1#Presults_format-1#Xzscore_flag.?#`meas_arg39#h?EPI4?P	?1??meas?1#mean?#stddev?#ZSCORE_QID?[??qid?1#num_zscores??#zscores??#?ZSCORESӥ.?num_q_zscores??#q_zscores?.#?ALL_ZSCORES??float? 0qid11#meas21#mean31#stddev41#LINES5P?te_get_zscoresE?$?@??@?epiCJ??zscores_fileC???all_zscoresD???fdF???sizeG???ptrH1??current_qidI1??iJ???linesK???line_ptrL??@num_linesM??Hnum_qidN??Ptext_zscores_ptrP??Xzscores_ptrQ.?`4?K	comp_lines_qid_meas????@?@4ptr1???Xptr2???Pcmp???l?	parse_zscore_line???@?@?start_ptr?U?Xqid_ptr?U?Pmeas_ptr?U?Hmean_ptr?U?@stddev_ptr?U??ptr?1?hte_get_zscores_cleanup???@??@?stderr?(trec_zscores_buf>1	Hhbtext_zscores_pool??	Phbzscores_pool@.	Xhbstderr?(?+?#?@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)convert_zscores.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdouble{8name9#value`
??`
?@`
-A`
AB`
?C`
WD`
?E`
?F`
GG`
=H`
!I`
nJ`
?K`
sL`
?M`
MN`
?O`
?P`
3Q`
?R`
S`
?T`
?U`
?V`
?W`
?X`
RY`
?Z`
r[`
\`
D]`
?^`
?_`
-``
ea`
_b`
`c`

d`
(e`
f`
?g`
?h`
?i`
j`
?k`
?l`
?m`
?n`
?o`
}p`u
u
?3`
]re
?
?off_names??
	`Hb?
?"trec_names??
	?Hb?
?set_names??
	Jb?prefs_names?
	`JbG?prefs_off_name?7	Kbw?qrels_jg_names?g	`Kb???
?̗
=`
I>`
??`
?@`
-A`
AB`
?C`
WD`
?E`
?F`
GG`
=H`
!I`
nJ`
?K`
sL`
?M`
MN`
?O`
?P`
3Q`
?R`
S`
?T`
?U`
?V`
?W`
?X`
RY`
?Z`
r[`
\`
D]`
?^`
?_`
-``
ea`
_b`
`c`

d`
(e`
f`
?g`
?h`
?i`
j`
?k`
?l`
?m`
?n`
?o`
}p`]re
	?Fbte_num_trec_measures??	@Hb?̗	?Kbte_num_trec_measure_nicknames??	Lb.	??-?@$?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_map.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoubleumeasure_name?#parameters?#MEAS_ARG@?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#huEPI4?)8name9?#value<6#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# )TREC_EVALG@Lprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?r `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf!?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
*trec_measP|name~?#explanation??#init_meas?V#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?,#0print_final_and_cleanup_meas?L#8meas_params?R#@eval_index??#HD?DJP?1?*??????P?
??
??
r?
1\?????P?
?????P
?,????L??JP2TREC_MEAS?1float3@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU3#0results_rel_listY3#8?RES_RELS]r	te_calc_map-?$?@-?@?epi+????+???results+???~tm,	??~eval,P??~res_rels.9??sum/6?Xrel_so_far0??`i1??h	
Xte_meas_mapX	 Lb?	???@0?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_P.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublesmeasure_name?#parameters?#MEAS_ARG>?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hsEPI4?'8name9?#value<4#TREC_EVAL_VALUE=??(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 'TREC_EVALG>Lprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?p `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOrtnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
(trec_measP|name~?#explanation??#init_meas?T#calc_meas??#acc_meas??# calc_avg_meas?
#(print_single_meas?*#0print_final_and_cleanup_meas?J#8meas_params?P#@eval_index??#HB?BHN?/?(~?~???N?
??
??
p?
/Z??~??N?
????~??N
?*?~??J?~HN0TREC_MEAS?/float1@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU1#0results_rel_listY1#8?RES_RELS]p	te_calc_P3?0?@??@Hepi1~???1???results1???~tm2	??~eval2N??~cutoffs41?Pcutoff_index5??Xi6??`res_rels77??rel_so_far8??h%	
V:	??long_cutoff_array*		?Lbdefault_P_cutoffs	?Lbte_meas_PV	Mb???f?@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_num_q.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#			B?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGB
Lprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?l `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr{tnum_q_relsu?#max_num_q_relsv?#		w#?ALL_REL_INFOx?
$trec_measP|name~?#explanation??#init_meas?P#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?&#0print_final_and_cleanup_meas?F#8meas_params?L#@eval_index??#H>?>DJ?+?$z?z???J?
??
??
l?
+V??z??J?
????z??J
?&?z??F?zDJ,
TREC_MEAS?+float?te_calc_num_q&???@?@?epi$z?h	$??`results%??Xtm%??Peval%J?H?
Rxte_calc_avg_num_q0??@f?@?epi.z?Xtm.??Pall_rel_info/??Haccum_eval/J?@		1??hte_meas_num_qR	`Mb	????@h?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_num_ret.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval	unsigned charshort unsigned intunsigned intsigned charshort intintlong int	charlong long unsigned intlong long intdoubleymeasure_name#parameters#MEAS_ARGD?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hyEPI4?-8name9#value<:#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# -TREC_EVALGDLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?v `qida#run_idb#ret_formatc#q_resultse?#RESULTSf%?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	'	w#?ALL_REL_INFOx?
.trec_measP|name~#explanation?#init_meas?Z#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?0#0print_final_and_cleanup_meas?P#8meas_params?V#@eval_index??#HH?HNT?5?.??????T?
??
??
v?
5`?????T?
?????T
!?0????P??NT6TREC_MEAS?5float7@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU7#0results_rel_listY7#8?RES_RELS]v?te_calc_num_ret$?h?@??@,epi"???'	"???results#???tm#???eval#T??res_rels%=???
\te_meas_num_ret\	?Mb??E??@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_num_rel.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֙^	unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t??^	5char?0	?._flags
?#_IO_read_ptr/#_IO_read_end/#_IO_read_base/#_IO_write_base/# _IO_write_ptr/#(_IO_write_end/#0_IO_buf_base/#8_IO_buf_end/#@_IO_save_base/#H_IO_backup_base/#P_IO_save_end/#X_markers#`_chain!#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/!#?_lock31#?_offset<#?__pad1E-#?__pad2F-#?__pad3G-#?__pad4H-#?__pad5I?#?_modeK?#?_unused2M7#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?#_pos??#?=15
&?G5
&long long unsigned intlong long intdouble?measure_name/#parameters/#MEAS_ARG|,pquery_flag?#summary_flag?#debug_level?#debug_query/#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,/#Presults_format-/#Xzscore_flag.?#`meas_arg3,#h?EPI4?e8name9/#value#TREC_EVAL_VALUE=	?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 1TREC_EVALGHLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?z `qida#run_idb#ret_formatc#q_resultse#RESULTSf)?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?	w#?ALL_REL_INFOx?
2trec_measP|name~#explanation?#init_meas?^#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?4#0print_final_and_cleanup_meas?T#8meas_params?Z#@eval_index??#HL?LRX?9?2??????X?
??
??
z?
9d?????X?
??	???	X
%?4????T??RX:TREC_MEAS?9float;@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU;#0results_rel_listY;#8?RES_RELS]z?te_calc_num_rel_ret%???@,?@ epi#????	#???results$???tm$???eval$X??res_rels&A???
`te_meas_num_rel_ret`	?Nb7	???@,?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_gm_map.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?	unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	charlong long unsigned intlong long intdoublexmeasure_name#parameters#MEAS_ARGC?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hxEPI4?,8name9#value<9#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# ,TREC_EVALGCLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?u `qida#run_idb#ret_formatc#q_resultse?#RESULTSf$?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?	w#?ALL_REL_INFOx?
-trec_measP|name~#explanation?#init_meas?Y#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?/#0print_final_and_cleanup_meas?O#8meas_params?U#@eval_index??#HG?GMS?4?-??????S?
??
??
u?
4_?????S?
?????S

 ?/????O??MS5TREC_MEAS?4float6@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU6#0results_rel_listY6#8?RES_RELS]u	te_calc_gm_map(?,?@??@\ epi&????	&???results'???~tm'	??~eval'S??~res_rels)<??sum*9?Xrel_so_far+??`i,??h	
[te_meas_gm_map[	?Nb?	?9??@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_Rprec.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?	unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?	w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]t	te_calc_Rprec,???@??@? epi*????	*???results*???~tm+	??~eval+R??~res_rels-;??num_to_look_at.??Xrel_so_far/??`i0??h	
Zte_meas_RprecZ	@Ob 	??W?@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_recip_rank.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?	unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	
charlong long unsigned intlong long intdouble|measure_name#parameters#MEAS_ARGG?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#h|EPI4?08name9#value<=#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 0TREC_EVALGGLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?y `qida#run_idb#ret_formatc#q_resultse#RESULTSf(?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?	w#?ALL_REL_INFOx?
1trec_measP|name~#explanation?#init_meas?]#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?3#0print_final_and_cleanup_meas?S#8meas_params?Y#@eval_index??#HK?KQW?8?1??????W?
??
??
y?
8c?????W?
?????W
$?3????S??QW9TREC_MEAS?8float:@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU:#0results_rel_listY:#8?RES_RELS]y?te_calc_recip_rank%???@W?@? epi"????	"???results#???tm#???eval$W??~res_rels&@??i'??h?
_te_meas_recip_rank_	?Ob?	?q?@X?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_bpref.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?	unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?	w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]tZ	te_calc_bpref(?X?@q?@@!epi&???~?	&???~results&???~tm'Z	??~eval'R??~res_rels);??j*??@nonrel_so_far+??Hrel_so_far+??Ppool_unjudged_so_far+??Xnum_nonrel,??`bpref-8?h`	
Zte_meas_bprefZ	Pb
???@t?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_iprec_at_recall.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval
unsigned charshort unsigned intunsigned intsigned charshort intintlong int
charlong long unsigned intlong long intdouble?measure_name	#parameters	#MEAS_ARGL?pquery_flag?#summary_flag?#debug_level?#debug_query	#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,	#Presults_format-	#Xzscore_flag.?#`meas_arg3?#h?EPI4?58name9	#valueTREC_MEAS?=float?@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU?#0results_rel_listY?#8?RES_RELS]~u	te_calc_iprec_at_recall.?t?@?@?!epi+???~
+???~results,???~tm,u	??~eval-\??~cutoff_percents/?	??cutoffs0??@current_cut1??Hrr2E??~rel_so_far3??Pi4??Xprecis5B?`int_precis5B?h{	
dB?	B
float_cutoff_array?		`Pbdefault_iprec_at_recall_cutoffs	?Pbte_meas_iprec_at_recalld	?Pb?	? K ??@?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_recall.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval
unsigned charshort unsigned intunsigned intsigned charshort intintlong int
charlong long unsigned intlong long intdoublexmeasure_name#parameters#MEAS_ARGC?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hxEPI4?,8name9#value<9#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# ,TREC_EVALGCLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?u `qida#run_idb#ret_formatc#q_resultse?#RESULTSf$?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	-
w#?ALL_REL_INFOx?
-trec_measP|name~#explanation?#init_meas?Y#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?/#0print_final_and_cleanup_meas?O#8meas_params?U#@eval_index??#HG?GMS?4?-??????S?
??
??
u?
4_?????S?
?????S

 ?/????O??MS5TREC_MEAS?4float6@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU6#0results_rel_listY6#8?RES_RELS]u)	te_calc_recall+??@??@?!epi)???-
)???results*???~tm*)	??~eval*S??~cutoffs,6?Pcutoff_index-??Xi.??`res_rels/<??rel_so_far0??h/	
[D	??long_cutoff_array4		@Qbdefault_recall_cutoffs	?Qbte_meas_recall[	?Qb?	?!? h?@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_Rprec_mult.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval6
unsigned charshort unsigned intunsigned intsigned charshort intintlong int6

charlong long unsigned intlong long intdouble|measure_name#parameters#MEAS_ARGG?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#h|EPI4?08name9#value<=#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 0TREC_EVALGGLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?y `qida#run_idb#ret_formatc#q_resultse#RESULTSf(?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	H
w#?ALL_REL_INFOx?
1trec_measP|name~#explanation?#init_meas?]#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?3#0print_final_and_cleanup_meas?S#8meas_params?Y#@eval_index??#HK?KQW?8?1??????W?
??
??
y?
8c?????W?
?????W
$?3????S??QW9TREC_MEAS?8float:@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU:#0results_rel_listY:#8?RES_RELS]yk	te_calc_Rprec_mult.???@h?@$"epi+???~H
+???~results,???~tm,k	??~eval-W??~cutoff_percents/v	??cutoffs0:?@current_cut1??Hrr2@??~rel_so_far3??Pi4??Xprecis5=?`int_precis5=?hq	
_=?	=?	Rprec_cutoff_array|		 Rbdefault_Rprec_cutoffs	pRbte_meas_Rprec_mult_	?Rb??"i!??@h?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_utility.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalsize_t֙Z
unsigned charshort unsigned intunsigned intsigned charshort intintlong int__off_t??__off64_t??Z
5char?Q
?._flags
?#_IO_read_ptr/#_IO_read_end/#_IO_read_base/#_IO_write_base/# _IO_write_ptr/#(_IO_write_end/#0_IO_buf_base/#8_IO_buf_end/#@_IO_save_base/#H_IO_backup_base/#P_IO_save_end/#X_markers#`_chain!#h_fileno#?#p_flags2'?#t_old_offset)#x_cur_column-?#?_vtable_offset.?#?_shortbuf/!#?_lock31#?_offset<#?__pad1E-#?__pad2F-#?__pad3G-#?__pad4H-#?__pad5I?#?_modeK?#?_unused2M7#?	_IO_lock_t?
_IO_marker?_next?#_sbuf?#_pos??#?=15
&?G5
&long long unsigned intlong long intdouble?measure_name/#parameters/#MEAS_ARG|,pquery_flag?#summary_flag?#debug_level?#debug_query/#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,/#Presults_format-/#Xzscore_flag.?#`meas_arg3,#h?EPI4?e8name9/#value26
'?H6
'long long unsigned intlong long intdouble?measure_name0#parameters0#MEAS_ARG}-pquery_flag?#summary_flag?#debug_level?#debug_query0#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,0#Presults_format-0#Xzscore_flag.?#`meas_arg3-#h?EPI4?f8name90#value?(@qidA0#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# fTREC_EVALG}PLprintable_paramsM0#num_paramsQ?#param_valuesR.#PARAMSS? `qida0#run_idb0#ret_formatc0#q_resultse.#RESULTSf^?nqido0#rel_formatp0#q_rel_infoq.#REL_INFOr?Ttnum_q_relsu?#max_num_q_relsv?#?
wT#?ALL_REL_INFOx
g	trec_measP|name~0#explanation?0#init_meas??	#calc_meas??	#acc_meas?
# calc_avg_meas?I
#(print_single_meas?i
#0print_final_and_cleanup_meas??
#8meas_params??
#@eval_index??#H?	??	?	?	3n?g	?	??	?	?	?	?	?	3?	??	??	n?	
??	?	
?	
??	>
??	?	>
?	D
Z
i
??	?	
O
?
??	?	?	o
PTREC_MEAS?nfloatp@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUp#0results_rel_listYp#8?RES_RELS]?
?te_calc_11ptavg1???@??@?"epi/?	??~?
/?	??~results0?	??~tm0???~eval0?	??~cutoff_percents2???cutoffs3p??current_cut4??@rr5v??~rel_so_far6??Hi7??Pprecis8s?Xint_precis8s?`sum9s?h??
sstderr??s
'
float_cutoff_array?	?Sbdefault_11ptavg_cutoffsP	Tbstderr?te_meas_11pt_avg?
	 Tb8%q#?@??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_ndcg.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?
unsigned charshort unsigned intunsigned intsigned charshort intintlong int?
charlong long unsigned intlong long intdoublevmeasure_name?#parameters?#MEAS_ARGA?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hvEPI4?*8name9?#value<7#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# *TREC_EVALGALprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?JTnameU?#valueV7#FLOAT_PARAM_PAIRW"? `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSfbnqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?Xtnum_q_relsu?#max_num_q_relsv?#	?
wX#ALL_REL_INFOx
ktrec_measP|name~?#explanation??#init_meas??#calc_meas??#acc_meas?# calc_avg_meas?M#(print_single_meas?m#0print_final_and_cleanup_meas??#8meas_params??#@eval_index??#H??????r?k????????
??
?
??
r?????
??B???B?H
^#m???S?????sTREC_MEAS?rfloatt@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUt#0results_rel_listYt#8?RES_RELS]??-	?
.?#	?
/?#gain07#REL_GAIN1?	3rel_gains4	#	?
5?#total_num_at_levels6?#?GAINS7?L
te_calc_ndcgA???@W?@#epi????~?
????~results@???~tm@L
??~eval@???~?
Bz??~results_gainC7??results_dcgC7?@ideal_gainD7?Hideal_dcgD7?Pcur_levelE??X?
E??`iF??hgainsG	??~R

??
setup_gains??W?@d?@T#tm?L
???
??
??gains???pairs?
?@num_pairs???Hi???Pj???X?
???`
z	J`comp_rel_gain??d?@??@?#ptr1?	?hptr2?	?`?get_gain?7??@?@?#?
ʴ?Xgainsʹ?Pi???h
??
	default_ndcg_gains	`hbte_meas_ndcg?	?Tb?	6&?$5?@?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_ndcg_cut.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?
unsigned charshort unsigned intunsigned intsigned charshort intintlong int?
charlong long unsigned intlong long intdoublezmeasure_name#parameters#MEAS_ARGE?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hzEPI4?.8name9#value<;#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# .TREC_EVALGELprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?w `qida#run_idb#ret_formatc#q_resultse#RESULTSf&?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?
w#?ALL_REL_INFOx?
/trec_measP|name~#explanation?#init_meas?[#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?1#0print_final_and_cleanup_meas?Q#8meas_params?W#@eval_index??#HI?IOU?6?/??????U?
??
??
w?
6a?????U?
?????U
"?1????Q??OU7TREC_MEAS?6float8@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU8#0results_rel_listY8#8?RES_RELS]wq	te_calc_ndcg_cut-??@5?@8$epi+???~?
+???~results,???~tm,q	??~eval,U??~cutoffs.8??cutoff_index/???res_rels0>??~gain1;?@sum1;?Hideal_dcg2;?Pcur_lvl3??Xlvl_count3??`i4??hw	
]?	??long_cutoff_array|		?Tbdefault_ndcg_cutoffs	0Ubte_meas_ndcg_cut]	`UbR7'P%??@8?@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_Rndcg.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?
unsigned charshort unsigned intunsigned intsigned charshort intintlong int?
charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?KTnameU?#valueV8#FLOAT_PARAM_PAIRW#? `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSfcnqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?Ytnum_q_relsu?#max_num_q_relsv?#	+wY#ALL_REL_INFOx
ltrec_measP|name~?#explanation??#init_meas??#calc_meas??#acc_meas?# calc_avg_meas?N#(print_single_meas?n#0print_final_and_cleanup_meas??#8meas_params??#@eval_index??#H??????s?l????????
??
?
??
s?????
??C???C?I
_$n???T?????tTREC_MEAS?sfloatu@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUu#0results_rel_listYu#8?RES_RELS]??=	4>?#	??#gain@8#REL_GAINA?	Crel_gainsD	#	E?#total_num_at_levelsF?#?GAINSG??
te_calc_RndcgQ?8?@??@?$epiO???~+O???~resultsP???~tmP?
??~evalP???~"R{??~results_gainS8??results_dcgS8??old_ideal_gainT8??ideal_gainT8??ideal_dcgT8?@sumU8?Hnum_changed_ideal_gainV??Pcur_levelW??XW??`iX??hgainsY	??~?

?Isetup_gains????@??@?$tm??
??"?I??gains?T??pairs?Z?@num_pairs???Hi???Pj???X???`O
{	K?comp_rel_gain????@?@%ptr1?	?hptr2?	?`get_gain?8?@??@h%4??Xgains??Pi???h
?
	default_ndcg_gains	?hbte_meas_Rndcg?	?Ub95({&3A??@GNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_ndcg_rel.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval>unsigned charshort unsigned intunsigned intsigned charshort intintlong int>charlong long unsigned intlong long intdoublezmeasure_name#parameters#MEAS_ARGE?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hzEPI4?.8name9#value<;#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# .TREC_EVALGELprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?NTnameU#valueV;#FLOAT_PARAM_PAIRW&? `qida#run_idb#ret_formatc#q_resultse#RESULTSffnqido#rel_formatp#q_rel_infoq#REL_INFOr?\tnum_q_relsu?#max_num_q_relsv?#	|w\#ALL_REL_INFOx
otrec_measP|name~#explanation?#init_meas??#calc_meas??#acc_meas?!# calc_avg_meas?Q#(print_single_meas?q#0print_final_and_cleanup_meas??#8meas_params??#@eval_index??#H??????v?o????????
??
?
??
v?????
??F???F?L
b'q???W?????wTREC_MEAS?vfloatp@E	ZG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUp#0results_rel_listYp#8?RES_RELS]??9	?:?#	f;?#gain<;#REL_GAIN=?	?rel_gains@	#	PA?#total_num_at_levelsB?#?GAINSC?}
te_calc_ndcg_relM???@pA?%epiK???~|K???~resultsL???~tmL}
??~evalL???~sNv??~results_gainO;??results_dcgO;??ideal_gainP;??ideal_dcgP;??sumQ;?@ZR??Hnum_relS??Pcur_levelT??XfT??`iU??hgainsV	??~?

?-setup_gains??pA}A&tm?}
??s?-??gains?8??pairs?>?@num_pairs???Hi???Pj???XP???`3
v	N?comp_rel_gain??}A?AL&ptr1?	?hptr2?	?`?get_gain?;?A3A?&????Xgains???Pi???h
??
	default_ndcg_gains	?hbte_meas_ndcg_rel?	 Vb1	3)?'?A4AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_binG.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublevmeasure_name?#parameters?#MEAS_ARGA?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hvEPI4?*8name9?#value<7#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# *TREC_EVALGALprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?s `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf"?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
+trec_measP|name~?#explanation??#init_meas?W#calc_meas??#acc_meas??# calc_avg_meas?
#(print_single_meas?-#0print_final_and_cleanup_meas?M#8meas_params?S#@eval_index??#HE?EKQ?2?+??????Q?
??
??
s?
2]?????Q?
?????Q
?-????M??KQ3TREC_MEAS?2float4@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU4#0results_rel_listY4#8?RES_RELS]s
	te_calc_binG'?4A?A?&epi%????%???results%???~tm&
	??~eval&Q??~res_rels(:??sum)7?Xrel_so_far*??`i+??h	
Yte_meas_binGY	?Vb5"*!(`A?AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_G.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublesmeasure_name?#parameters?#MEAS_ARG>?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hsEPI4?'8name9?#value<4#TREC_EVAL_VALUE=??(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 'TREC_EVALG>Lprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?GTnameU?#valueV4#FLOAT_PARAM_PAIRW? `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf_nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?Utnum_q_relsu?#max_num_q_relsv?#	?wU#ALL_REL_INFOx
htrec_measP|name~?#explanation??#init_meas??#calc_meas??#acc_meas?# calc_avg_meas?J#(print_single_meas?j#0print_final_and_cleanup_meas??#8meas_params??#@eval_index??#H??????o?h????????
??
?
??
o?????
????????E
[ j???P?????pTREC_MEAS?ofloatq@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUq#0results_rel_listYq#8?RES_RELS]??5	?6?#	?7?#gain84#REL_GAIN9?	;rel_gains<	#	?=?#total_num_at_levels>?#?GAINS???
te_calc_GI??A?A0'epiG???~?G???~resultsH???~tmH?
??~evalH???~?Jw??~results_gainK4??sum_resultsK4??ideal_gainL4??sum_idealL4??sum_costM4?@min_costM4?Hresults_gN4?Pcur_levelO??X?O??`iP??hgainsQ	??~?

?3setup_gains???A?A|'tm??
????3??gains?>??pairs?D?@num_pairs???Hi???Pj???X????`9
w	G?comp_rel_gain???A?A?'ptr1?	?hptr2?	?`?get_gain?4?A`A(????Xgains???Pi???h
??
	default_G_gains	?hbte_meas_G?	?Vb?	 +:)*A`AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_rel_P.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]t!	te_calc_rel_P)?`A*A`(epi'???'???results'???~tm(!	??~eval(R??~cutoffs*5?Pcutoff_index+??Xi,??`rr-;??rel_so_far.??h'	
Z<	??long_cutoff_array,		@Wbdefault_relative_P_cutoffs	?Wbte_meas_relative_PZ	?Wb?	!,?)?A,AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_success.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval
unsigned charshort unsigned intunsigned intsigned charshort intintlong int
charlong long unsigned intlong long intdoubleymeasure_name#parameters#MEAS_ARGD?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hyEPI4?-8name9#value<:#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# -TREC_EVALGDLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?v `qida#run_idb#ret_formatc#q_resultse?#RESULTSf%?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	w#?ALL_REL_INFOx?
.trec_measP|name~#explanation?#init_meas?Z#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?0#0print_final_and_cleanup_meas?P#8meas_params?V#@eval_index??#HH?HNT?5?.??????T?
??
??
v?
5`?????T?
?????T
!?0????P??NT6TREC_MEAS?5float7@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU7#0results_rel_listY7#8?RES_RELS]v+	te_calc_success+?,A?A?(epi)???)???results*???~tm*+	??~eval*T??~cutoffs,7?Pcutoff_index-??Xi.??`res_rels/=??rel_so_far0??h1	
\F	??success_cutoff_array6		 Xbdefault_success_cutoffs	@Xbte_meas_success\	`Xb?	"->*9A?AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_infap.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval%unsigned charshort unsigned intunsigned intsigned charshort intintlong int%charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	7w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]te	te_calc_infap+??A9A?(epi)???~7)???~results)???~tm*e	??~eval*R??~res_rels,;??j-??@nonrel_so_far.??Hrel_so_far.??Ppool_unjudged_so_far.??Xinf_ap/8?`?A?AfjM8?hk	
Zte_meas_infAPZ	?Xb?	.?*A<AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_map_cut.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval@unsigned charshort unsigned intunsigned intsigned charshort intintlong int@charlong long unsigned intlong long intdoubleymeasure_name#parameters#MEAS_ARGD?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hyEPI4?-8name9#value<:#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# -TREC_EVALGDLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?v `qida#run_idb#ret_formatc#q_resultse?#RESULTSf%?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	Rw#?ALL_REL_INFOx?
.trec_measP|name~#explanation?#init_meas?Z#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?0#0print_final_and_cleanup_meas?P#8meas_params?V#@eval_index??#HH?HNT?5?.??????T?
??
??
v?
5`?????T?
?????T
!?0????P??NT6TREC_MEAS?5float7@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU7#0results_rel_listY7#8?RES_RELS]v9	te_calc_map_cut+?<AAD)epi)???~R)???~results*???~tm*9	??~eval*T??~cutoffs,7?Hcutoff_index-??Pi.??Xres_rels/=??rel_so_far0??`sum1:?h?	
\T	??long_cutoff_arrayD		 Ybdefault_map_cutoffs	pYbte_meas_map_cut\	?Yb?	/O+?AAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_gm_bpref.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval[unsigned charshort unsigned intunsigned intsigned charshort intintlong int[charlong long unsigned intlong long intdoublezmeasure_name#parameters#MEAS_ARGE?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hzEPI4?.8name9#value<;#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# .TREC_EVALGELprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?w `qida#run_idb#ret_formatc#q_resultse#RESULTSf&?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	mw#?ALL_REL_INFOx?
/trec_measP|name~#explanation?#init_meas?[#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?1#0print_final_and_cleanup_meas?Q#8meas_params?W#@eval_index??#HI?IOU?6?/??????U?
??
??
w?
6a?????U?
?????U
"?1????Q??OU7TREC_MEAS?6float8@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU8#0results_rel_listY8#8?RES_RELS]w`	te_calc_gm_bpref'?A?A?)epi%???~m%???~results&???~tm&`	??~eval&U??~res_rels(>??j)??@nonrel_so_far*??Hrel_so_far*??Ppool_unjudged_so_far*??Xnum_nonrel+??`bpref,;?hf	
]te_meas_gm_bpref]	Zb?
0?+PA?AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_runid.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalvunsigned charshort unsigned intunsigned intsigned charshort intintlong intvcharlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float?te_calc_runid$??AA?)epi"??h?"??`results#??Xtm#??Peval#R?H?
ZRte_print_runid+?APA(*epi*??htm*R?`eval*R?XZrunid ?	?hbte_meas_runidZ	`Zb?
?0;,?APAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_relstring.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	charlong long unsigned intlong long intdouble{measure_name#parameters#MEAS_ARGF?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#h{EPI4?/8name9#value<<#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# /TREC_EVALGFLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?x `qida#run_idb#ret_formatc#q_resultse#RESULTSf'?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
0trec_measP|name~#explanation?#init_meas?\#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?2#0print_final_and_cleanup_meas?R#8meas_params?X#@eval_index??#HJ?JPV?7?0??????V?
??
??
x?
7b?????V?
?????V

#?2????R??PV8TREC_MEAS?7float9@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU9#0results_rel_listY9#8?RES_RELS]x<	te_calc_relstring3?PA?At*epi1???~?1???~results2???~tm2<	??~eval2V??~params4G	?Hres_rels5???i6??Plen6??X,A?AcC	?grelD??hB	
^<?	te_print_q_relstringS??AA?*epiR??htmR<	?`evalR??X
te_print_relstring^?A?A+epi]??htm]
?`eval]V?X^!
<?relstring_len
	?Zbdefault_relstring_params	?Zbcurrent_string	?hbstring_len?	?hbte_meas_relstring^	[b	2?,>A?AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_set_P.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]t?te_calc_set_P(??A>AX+epi&????&???results&???tm'???eval'R??res_rels);???
Zte_meas_set_PZ	`[b	?29-?A@AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_set_recall.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?
charlong long unsigned intlong long intdouble|measure_name#parameters#MEAS_ARGG?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#h|EPI4?08name9#value<=#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# 0TREC_EVALGGLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?y `qida#run_idb#ret_formatc#q_resultse#RESULTSf(?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
1trec_measP|name~#explanation?#init_meas?]#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?3#0print_final_and_cleanup_meas?S#8meas_params?Y#@eval_index??#HK?KQW?8?1??????W?
??
??
y?
8c?????W?
?????W
$?3????S??QW9TREC_MEAS?8float:@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU:#0results_rel_listY:#8?RES_RELS]y?te_calc_set_recall%?@A?A?+epi"????"???results#???tm#???eval$W??res_rels&@???
_te_meas_set_recall_	?[b	?3?-sA?AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_set_rel_P.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?	charlong long unsigned intlong long intdouble{measure_name#parameters#MEAS_ARGF?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#h{EPI4?/8name9#value<<#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# /TREC_EVALGFLprintable_paramsM#num_paramsQ?#param_valuesR#PARAMSS?x `qida#run_idb#ret_formatc#q_resultse#RESULTSf'?nqido#rel_formatp#q_rel_infoq#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
0trec_measP|name~#explanation?#init_meas?\#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?2#0print_final_and_cleanup_meas?R#8meas_params?X#@eval_index??#HJ?JPV?7?0??????V?
??
??
x?
7b?????V?
?????V

#?2????R??PV8TREC_MEAS?7float9@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU9#0results_rel_listY9#8?RES_RELS]x?te_calc_set_relative_P(??AsA?+epi%????%???results&???tm&???eval'V??res_rels)????
^te_meas_set_relative_P^	 \b	?4.) AtAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_set_map.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?charlong long unsigned intlong long intdoubleymeasure_name#parameters#MEAS_ARGD?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hyEPI4?-8name9#value<:#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# -TREC_EVALGDLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?v `qida#run_idb#ret_formatc#q_resultse?#RESULTSf%?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	
w#?ALL_REL_INFOx?
.trec_measP|name~#explanation?#init_meas?Z#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?0#0print_final_and_cleanup_meas?P#8meas_params?V#@eval_index??#HH?HNT?5?.??????T?
??
??
v?
5`?????T?
?????T
!?0????P??NT6TREC_MEAS?5float7@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU7#0results_rel_listY7#8?RES_RELS]v?te_calc_set_map'?tA) A<,epi%???
%???results&???tm&???eval&T??res_rels(=???
\te_meas_set_map\	?\b?	?5{.Q!A, AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_set_F.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval
unsigned charshort unsigned intunsigned intsigned charshort intintlong int
charlong long unsigned intlong long intdoublewmeasure_name?#parameters?#MEAS_ARGB?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hwEPI4?+8name9?#value<8#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# +TREC_EVALGBLprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?t `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf#?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	*
w#?ALL_REL_INFOx?
,trec_measP|name~?#explanation??#init_meas?X#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?.#0print_final_and_cleanup_meas?N#8meas_params?T#@eval_index??#HF?FLR?3?,??????R?
??
??
t?
3^?????R?
?????R	
?.????N??LR4TREC_MEAS?3float5@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU5#0results_rel_listY5#8?RES_RELS]t	te_calc_set_F*?, AQ!A?,epi(???*
(???results(???~tm)	??~eval)R??~params+	?Pres_rels,;??x-8?XP-8?`R-8?h	
Z83	8?set_F_param_array#		?\bdefault_set_F_params	?\bte_meas_set_FZ	 ]b5	?6?.?!AT!AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_num_nonrel_judged_ret.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval3
unsigned charshort unsigned intunsigned intsigned charshort intintlong int3
charlong long unsigned intlong long intdouble?measure_name#parameters#MEAS_ARGRpquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3#h?EPI4?;8name9#value#0print_final_and_cleanup_meas?^#8meas_params?d#@eval_index??#HV?V\bC?<??????b?
?
??
??
Cn?????b?
?????b
/?>????$^??\bD%TREC_MEAS?CfloatE@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsUE#0results_rel_listYE#8?RES_RELS]??te_calc_num_nonrel_judged_ret%?T!A?!A?,epi"???E
"???results#???tm#???eval$b??res_rels&K??	
jte_meas_num_nonrel_judged_retj	?]bg?7a/?"A?!AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_num_prefs_poss.c/users/tr.craigm/src/trec_eval9_from_github/trec_evald
unsigned charshort unsigned intunsigned intsigned charshort intintlong intd
char"?long long unsigned intlong long intdouble?measure_name#parameters#MEAS_ARG]
pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3
#h?EPI4?F8name9#value?nqido#rel_formatp#q_rel_infoq#REL_INFOr?4tnum_q_relsu?#max_num_q_relsv?#	v
w4#?ALL_REL_INFOx?
Gtrec_measP|name~#explanation?#init_meas?s#calc_meas??#acc_meas??# calc_avg_meas?)#(print_single_meas?I#0print_final_and_cleanup_meas?i#8meas_params?o#@eval_index??#Ha?agmN?G??????m?
?
??
??
Ny?????m?
?????m$
:?I????/i??gmO0TREC_MEAS?Nfloat??orel_levelp?#num_in_ecq?#docid_rankss?#ECv??	N
?"#array?#	Y
??#PREFS_ARRAY??c?	N
?c#array?i#	Y
??#?cCOUNTS_ARRAY?.?	x?ecs??	#num_ecs??#prefs_array?#rel_array??	#(num_prefs_fulfilled_ret??#0num_prefs_possible_ret??#8num_prefs_fulfilled_imp??#@num_prefs_possible_imp??#Hnum_prefs_possible_notoccur??#Pnum_nonrel??#Xnum_nonrel_ret??#`num_rel??#hnum_rel_ret??#p??JG҃T
8?num_jgs??#jgs?T
#	Y
??#num_judged_ret??#pref_counts?o# ?	RESULTS_PREFS??	0te_calc_prefs_num_prefs_poss$??!A?"A -epi!???v
!???results"???tm"0??eval#m??~results_prefs%Z
??i&??`poss'??h6
ute_meas_prefs_num_prefs_possu	?]bc?8?/?#A?"AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_num_prefs_ful.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?
unsigned charshort unsigned intunsigned intsigned charshort intintlong int?
char!?long long unsigned intlong long intdouble?measure_name
#parameters
#MEAS_ARG\pquery_flag?#summary_flag?#debug_level?#debug_query
#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,
#Presults_format-
#Xzscore_flag.?#`meas_arg3#h?EPI4?E8name9
#valuetrec_measP|name~#explanation?#init_meas?j#calc_meas??#acc_meas??# calc_avg_meas? #(print_single_meas?@#0print_final_and_cleanup_meas?`#8meas_params?f#@eval_index??#HX?X^d
E?>??????d?
?
??
??
Ep?????d?
?????d
1?@????&`??^dF'TREC_MEAS?Efloat??orel_levelp}#num_in_ecq?#docid_rankss?#ECv??	C?#array?#	N??#PREFS_ARRAY??Z?	C?Z#array?`#	N??#?ZCOUNTS_ARRAY?%?	x?ecs??	#num_ecs??#prefs_array?#rel_array??	#(num_prefs_fulfilled_ret??#0num_prefs_possible_ret??#8num_prefs_fulfilled_imp??#@num_prefs_possible_imp??#Hnum_prefs_possible_notoccur??#Pnum_nonrel??#Xnum_nonrel_ret??#`num_rel??#hnum_rel_ret??#p?}JG?zK
8?num_jgs??#jgs?K
#	N??#num_judged_ret??#pref_counts?f# ?	RESULTS_PREFS??	:te_calc_prefs_avgjg,?x)A#+A?.epi)???k)???results*???~tm*:??~eval+d??~results_prefs-Q
??i.??Pful/??Xposs/??`sum0J?h@
lte_meas_prefs_avgjgl	?_b
I=?2?4A$+AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_avgjg_Rnonrel.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?char!?long long unsigned intlong long intdouble?measure_name
#parameters
#MEAS_ARG\pquery_flag?#summary_flag?#debug_level?#debug_query
#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,
#Presults_format-
#Xzscore_flag.?#`meas_arg3#h?EPI4?E8name9
#valueR?XR???`N???hnum_ful@???@???W
t?recalculatel!.A?4A4/jgj???~?j???~ret_num_fulj???~ret_num_possk???~num_fulm???n???_i.AW1Aec1s???ec2s???ptr1t???ptr2t???new_nonrel_ecv???~W1A?4Ai???@j???Hfirst_discarded_nonrel???Pa??Xnum_nonrel_seen???`????h?
?	
?te_meas_prefs_avgjg_Rnonrelt	 `bcp>?3?5A?4AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_simp_ret.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?char?long long unsigned intlong long intdouble?measure_name#parameters#MEAS_ARGWpquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3#h?EPI4?@8name9#value
ote_meas_prefs_pair_reto	?`btN@?4?8A?7AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_avgjg_ret.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval5unsigned charshort unsigned intunsigned intsigned charshort intintlong int5char?long long unsigned intlong long intdouble?measure_name	#parameters	#MEAS_ARGXpquery_flag?#summary_flag?#debug_level?#debug_query	#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,	#Presults_format-	#Xzscore_flag.?#`meas_arg3#h?EPI4?A8name9	#valuep??~rp@R
??iA??PsumBV?XRC??`NC??hnum_fulD???[D???_
x?recalculatemG;A@A?0jgk???~?k???~ret_num_fulk???~ret_num_possl???~num_fuln???[o???f?;AW>Aec1t???ec2t???ptr1u???ptr2u??@new_nonrel_ecw???W>A??Ai???Hj???Pfirst_discarded_nonrel???Xa??`num_nonrel_seen???h?
?	
?te_meas_prefs_avgjg_Rnonrel_retx	?abcdB@6fAA@AGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_simp_imp.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval?unsigned charshort unsigned intunsigned intsigned charshort intintlong int?char?long long unsigned intlong long intdouble?measure_name#parameters#MEAS_ARGWpquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3#h?EPI4?@8name9#value
ote_meas_prefs_pair_impo	`bbtBDT7?EALDAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_prefs_avgjg_imp.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalunsigned charshort unsigned intunsigned intsigned charshort intintlong intchar?long long unsigned intlong long intdouble?measure_name	#parameters	#MEAS_ARGXpquery_flag?#summary_flag?#debug_level?#debug_query	#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,	#Presults_format-	#Xzscore_flag.?#`meas_arg3#h?EPI4?A8name9	#valuew#?ALL_REL_INFOx?
0trec_measP|name~#explanation?#init_meas?\#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?2#0print_final_and_cleanup_meas?R#8meas_params?X#@eval_index??#HJ?JPV?7?0??????V?
??
??
x?
7b?????V?
?????V

#?2????R??PV8TREC_MEAS?7float9@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU9#0results_rel_listY9#8?RES_RELS]x?`qida#num_jgsb?#jgsc?#?RES_RELS_JGdOk	te_calc_map_avgjg(??EA?GA?1epi&???>&???results'???tm'k	??eval'V??rr)???sum*<?Prel_so_far+??Xi,??`jg,??hq	
^te_meas_map_avgjg^	 cbo
 F`8?KA?GAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_Rprec_mult_avgjg.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalGunsigned charshort unsigned intunsigned intsigned charshort intintlong intGcharlong long unsigned intlong long intdouble?measure_name
#parameters
#MEAS_ARGM?pquery_flag?#summary_flag?#debug_level?#debug_query
#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,
#Presults_format-
#Xzscore_flag.?#`meas_arg3?#h?EPI4?68name9
#value?7??????]?
?
??
?
>i?????]?
?????]
*?9????Y??W]? TREC_MEAS?>float@@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU@#0results_rel_listY@#8?RES_RELS]?`qida
#num_jgsb?#jgsc?#FRES_RELS_JGdV?	te_calc_Rprec_mult_avgjg1??GA?KA,2epi.???Y.???results/???~tm/?	??~eval0]??~cutoff_percents2?	??cutoffs3@??current_cut4??@rr5???rel_so_far6??Hi7??Pjg7??Xprecis8C?`int_precis8C?h?	
eC?	C	Rprec_cutoff_array?		?cbdefault_Rprec_avgjg_cutoffs 	?cbte_meas_Rprec_mult_avgjge	db
!G90NA?KAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_P_avgjg.c/users/tr.craigm/src/trec_eval9_from_github/trec_evalbunsigned charshort unsigned intunsigned intsigned charshort intintlong intbcharlong long unsigned intlong long intdoubleymeasure_name#parameters#MEAS_ARGD?pquery_flag?#summary_flag?#debug_level?#debug_query#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,#Presults_format-#Xzscore_flag.?#`meas_arg3?#hyEPI4?-8name9#value<:#TREC_EVAL_VALUE=?(@qidA#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# -TREC_EVALGDLprintable_paramsM#num_paramsQ?#param_valuesR?#PARAMSS?v `qida#run_idb#ret_formatc#q_resultse?#RESULTSf%?nqido#rel_formatp#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	tw#?ALL_REL_INFOx?
.trec_measP|name~#explanation?#init_meas?Z#calc_meas??#acc_meas??# calc_avg_meas?#(print_single_meas?0#0print_final_and_cleanup_meas?P#8meas_params?V#@eval_index??#HH?HNT?5?.??????T?
??
??
v?
5`?????T?
?????T
!?0????P??NT6TREC_MEAS?5float7@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU7#0results_rel_listY7#8?RES_RELS]v?`qida#num_jgsb?#jgsc?#=RES_RELS_JGdM?	te_calc_P_avgjg-??KA0NAx2epi+???t+???results,???tm,?	??eval,T??cutoffs.7?Hcutoff_index/??Pi0??Xjg0??`rr1???rel_so_far2??h?	
\?	??long_cutoff_array?		`dbdefault_P_avgjg_cutoffs	?dbte_meas_P_avgjg\	?db1	"H?9?OA0NAGNU C 4.1.2 20080704 (Red Hat 4.1.2-55)m_yaap.c/users/tr.craigm/src/trec_eval9_from_github/trec_eval}unsigned charshort unsigned intunsigned intsigned charshort intintlong int}charlong long unsigned intlong long intdoublevmeasure_name?#parameters?#MEAS_ARGA?pquery_flag?#summary_flag?#debug_level?#debug_query?#relation_flag?# average_complete_flag ?#(judged_docs_only_flag$?#0num_docs_in_coll'?#8relevance_level(?#@max_num_docs_per_topic+?#Hrel_info_format,?#Presults_format-?#Xzscore_flag.?#`meas_arg3?#hvEPI4?*8name9?#value<7#TREC_EVAL_VALUE=?(@qidA?#num_queriesB?#valuesC?#num_valuesD?#max_num_valuesE?# *TREC_EVALGALprintable_paramsM?#num_paramsQ?#param_valuesR?#PARAMSS?s `qida?#run_idb?#ret_formatc?#q_resultse?#RESULTSf"?nqido?#rel_formatp?#q_rel_infoq?#REL_INFOr?tnum_q_relsu?#max_num_q_relsv?#	?w#?ALL_REL_INFOx?
+trec_measP|name~?#explanation??#init_meas?W#calc_meas??#acc_meas??# calc_avg_meas?
#(print_single_meas?-#0print_final_and_cleanup_meas?M#8meas_params?S#@eval_index??#HE?EKQ?2?+??????Q?
??
??
s?
2]?????Q?
?????Q
?-????M??KQ3TREC_MEAS?2float4@Enum_rel_retG?#num_retJ?#num_nonpoolL?#num_unjudged_in_poolN?#num_relS?# num_rel_levelsT?#(rel_levelsU4#0results_rel_listY4#8?RES_RELS]s
	te_calc_yaap'?0NA?OA?2epi%???~?%???~results%???~tm&
	??~eval&Q??~res_rels(:??sum)7?Hrel_so_far*??Pi+??X	
Yte_meas_yaapY	@eb%:;I$>$>I:;
:;I8
	&I
:;:;
:;I8

I!I/
:;I8
:;'III.?:;I@:;I
4:;I
.:;'I@:;I
4:;I
:;I
4:;I
.:;'@ .:;'I@!4:;I?<"4:;I?<#!%$>$>I:;
:;I8
:;I	
:;I8

'III
I!I/4:;I?<4:;I?<4:;I?
4:;I?
4:;I?
4:;I?
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;
:;I8
'II&I.?:;'I@:;I
:;I
4:;I
4:;I
.:;'I@:;I
4:;I
4:;I
:;I
4:;I?<%$>$>I:;
:;I8
:;I	:;
'II&I
.?:;'I@:;I
:;I
.?:;'I@4:;I
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.?:;'I@:;I
:;I
4:;I
4:;I
.?:;'I@%$>$>I:;
:;I8
:;I	:;
'II&I
.?:;'I@:;I
.?:;'I@4:;I
%$>$>I:;
:;I8
:;I	:;
'II&I
.?:;'I@:;I
4:;I
.?:;'I@%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;.?:;'I@:;I
4:;I
.:;'I@.?:;I@4:;I?<%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;.?:;'I@:;I
4:;I
4:;I
.:;'I@:;I
.?:;I@4:;I?<%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;.?:;'I@:;I
4:;I
.:;'I@.:;'I@:;I
4:;I
.?:;I@4:;I?<%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;.?:;'I@:;I
4:;I
.:;'I@.?:;I@4:;I?<%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;.?:;'I@:;I
4:;I
.:;'I@.?:;I@4:;I?<%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;&I
:;I8
.?:;'I@:;I
4:;I
.:;I@:;I
.?:;I@4:;I?<4:;I
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;&I
:;I8
.?:;'I@:;I
4:;I
.:;I@:;I
.?:;I@4:;I?<4:;I
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;&I
:;I8
.?:;'I@:;I
:;I
4:;I
4:;I
.:;'I@:;I
:;I
4:;I
4:;I
.:;'@.:;I@.?:;I@4:;I?<%:;I$>$>I.?:;'I@:;I
	:;I

&I%:;I$>$>I:;
:;I8
	&I
:;:;
:;I8

I!I/:;.?:;'I@:;I
4:;I
.:;'I@.?:;I@4:;I?<%$>$>I&I:;
:;I8
:;I	
:;I8

.?:;'I@:;I
:;I

4:;I
4:;I
.:;'I@%$>$>I:;
:;I8
:;I	:;
'II&I
4:;I?<I!I/4:;I
4:;I?
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;
:;I8
'II&I.:;'I@:;I
:;I
4:;I
4:;I
4:;I?<4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;
:;I8
'II&I.:;'I@:;I
:;I
4:;I
4:;I?<4:;I?
%:;I$>$>I:;
:;I8
	:;
:;
:;I8
I
!I/:;
:;I8
'II&I.:;'I@:;I
:;I
4:;I
4:;I?<4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
.:;'@4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I
.:;'@4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
I!I/4:;I?
%$>$>I:;
:;I8
:;I	
:;I8

:;'II
&I.:;'I@:;I
:;I
4:;I
4:;I?
j??
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includetrec_eval.ctrec_eval.hgetopt.hstdio.hlibio.htypes.hstddef.h	8@?y????????????????)???)?)Y0??![u??YZ!Z/Z?Z/Z!Z/Z/Z?Z?[?Z!?[/?u?/?$$??%?y? ?????%?y? ??????vY??7?Y1%y?
X%?%ױ?3?8?ڟ?"???d%w?X?u???9%?=%o?XB ?.?????g%w?X??????2C?#???-?ן2???=+&ɑ?2?????Y???//?U?w? ?2??Z??$?̟?$?ˠ?Y?.	 u?Y2?%???=2?????'????'???$?咟5/?
formats.ctrec_eval.hQ??
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includemeas_init.ctrec_eval.hstdio.hlibio.htypes.hstddef.h	8@)?Y3>>??*=?3>>??*=?3Oɻ/g??yX
??I??C$???ןyOɻ/g??yX
??I??L$???ן
tMɻ/g??yX?>???=*?=?yKɻ/g??yX?>???=*?=?z????_?9??????u????????+???x????_?9??????u)??)??2???x????_?9????u)??)???y?K??Y?h??K0tJ?g??_?9??ɭuYحu*w?<*???x?Y1?u?u??2?-Y?"?2-Y?(?1?u?VY?"?]0?
meas_acc.ctrec_eval.h	T2@=Y4=[Y4??gjYv0?
meas_avg.ctrec_eval.h	?3@=Y4=?ɮuLY4>?ɮu?TkY5??ɮ???H9Y|9?
meas_print_single.ctrec_eval.h	?5@ ?Y5=?[YY4=?[VY4>??cya?Y?8?
meas_print_final.ctrec_eval.h	?8@?Y5=?Y4=?[VY5???y$	???Y4=????Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeget_qrels.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	?<@?vt????YL[??W?#僭/"@?yX	?*???AU?x歭???%u#YZ??#?)pX.#????Y??3u?;?0g?;0gZ;?0g?;g?0??;>???3K????宻宻?Y???
/usr/lib/gcc/x86_64-redhat-linux/4.1.2/include/usr/include/usr/include/bitsget_trec_results.ctrec_eval.hstddef.htrec_format.hstdio.hlibio.htypes.h	?E@7???h?????u????Z#??W?#僭[K??0)@?rX?*???AU?o歭???)u#YZ??/?+oX.#?"??1?Y??3???0g?;0?9?0g?;0?9?0g?;g?0??"???2K????宻宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeget_prefs.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	4P@?vt????YL[??W?$僭YA@?xX
?*???CU?x歭???*u#YZ??#?","nX.#????Y??3?;?0g?;?0g?;?0g?;?0g?;g?0??;>???4K????宻宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeget_qrels_prefs.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	DZ@?vt????YL[??W??僭/-/?yX	?????/U?x歭????u#YZ/?#g?"unX.#????Y??3??;?0g?;?0g?;?0g?;g?0??;>???2K????宻宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeget_qrels_jg.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	c@??t????YL[??W??僭/-/?yX	?????/Y?1v1&s?.?歭?????u$YY#Z/?#=??u#Z=?g?aX".$#?!???Y?u??3??;?0g?;?0g?;?0g?;g?0??;>???2K????宻宻宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeform_res_rels.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	?m@:
th[?,??0-?h???Z??00??׽????</?.???Y/׻???xX
?1?$????#T33?_????#jX./?#U????????/'!:-??I?=?>'w .??????(?[?4??K?K??ɑɑ?4??????4?/3K?????宻宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeform_res_rels_jg.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	Py@:
th???,??0-?h???????00??׽????</?
.???;/׻U?
X<?/?#?;???T33?5?????jX./??U????59??Y'?M-?K%?\?&?#'w?.???2??a?.?X???4??K?K??ɑɑ?4??????4?/2K?????宻孮?宻?Y???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeform_prefs_counts.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	$?@?'??!!!9??K??9-????
??,??/?0'(>1(wXX	???筭???!!!9??????׃?4?,AY?+?m?*K+j?0'Y?????1Y(/YQ&3X?,AY?+?m?*K+j???3??????%??Y7xD
.!??3󻾻?????????$0?WN*E6E4$Z$[(/??0???Z[?Y?Zfh??3?3>h?a?1?#?N?&k??w?g?/?V/4????w?g?/?V/7????x?2ggggggggg"?h?N?+?w?.ɼg??1?w?Y؟'"?2K?????宻宻宻宻宻宻宻宻宻宻宻?Y?`?
/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeutility_pool.cstddef.h	Լ@u??ɟ??K?3g??ɟ?K?K$o??
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includeget_zscores.ctrec_eval.hstdio.hlibio.htypes.hstddef.h	$?@?L
t????YL[??W?????/-/?yX	?????/U?V??????u5Z/??g&&qX.5?摟w?Y??3??;?0g?;?0g?;?0g?;g?0??;>???3K????宻?Y?7?
convert_zscores.ctrec_eval.h	??@?z???0??+?gV,??ptf?2????Y?$g?g?x?
ֻ2????Y?$g?g?x?
ֻ60?
measures.ctrec_eval.hl>?
m_map.ctrec_eval.htrec_format.h	$?@,"?惑??Y%?u1?w<?
m_P.ctrec_eval.htrec_format.h	0?@2"????M<??vX
./IU%?X/?
m_num_q.ctrec_eval.h	??@%u%Y5=?ɮ#YdB?
m_num_ret.ctrec_eval.htrec_format.h	h?@#???%???
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includem_num_rel.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	??@%???%?6?ɼ??W%??Vka%??-V?D?k?X#?hF?
m_num_rel_ret.ctrec_eval.htrec_format.h	??@$???%?qA?
m_gm_map.ctrec_eval.htrec_format.h	,?@'"?惑??Y%?uhp?m@?
m_Rprec.ctrec_eval.htrec_format.h	??@+"??/u??V?:?lE?
m_recip_rank.ctrec_eval.htrec_format.h	??@$??????5??@?
m_bpref.ctrec_eval.htrec_format.h	X?@'(????e6??????Y[8V3[uPhX.?h%??J?
m_iprec_at_recall.ctrec_eval.htrec_format.h	t?@-.6#?%?@?g/W?#????!?'/)V?wX
f/)Vy??~A?
m_recall.ctrec_eval.htrec_format.h	?@*"?????M>i?vX
./>U%??E?
m_Rprec_mult.ctrec_eval.htrec_format.h	??@-.6#?%?@?#?g/OU݃??!?/)V?'wX
f?????
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includem_utility.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	h?@4#???!????
??
/usr/include/usr/include/bits/usr/lib/gcc/x86_64-redhat-linux/4.1.2/includem_11pt_avg.ctrec_eval.htrec_format.hstdio.hlibio.htypes.hstddef.h	??@0.5?=??#?%?@?g/W?#????!?'/?V?wX
f/?Vy???
??
m_ndcg.ctrec_eval.htrec_format.h	??@?.#??摑????2?5Y0?Y?-5?oi&Z2?5?ZxX
?ZY0?Y?-5?ZqX.-??2K?????9???30$T5?>?;#'6uXf黟4统w?"2??%$1??C?
m_ndcg_cut.ctrec_eval.htrec_format.h	?@,./?#??M)??&?5q9./)?1S	??????Y/Y???S?&˭5?j@X\S?1zX	??'@?
m_Rndcg.ctrec_eval.htrec_format.h	8?@?.?&???摑??????2Y0?Y?-??Z??55?f|%Z2?5?ZxX
=?[ZY0?Y?-??Z?5?ZhX.2??2K?????9???30$T5?>?;#'6uXf黟4统w?"2??%$1?*C?
m_ndcg_rel.ctrec_eval.htrec_format.h	??@?.???&??摑????2?5Y0?Y?-Y5?Z?h|%Z2?5?Z?mrXZY0?Y?-Y5?ZoX.,?V2??2K?????9???30$T5?>?;#'6uXf黟4统w?"2??%$1?t??
m_binG.ctrec_eval.htrec_format.h	4A&"?惑??Y<?*

.u1?<?
m_G.ctrec_eval.htrec_format.h	?A?
.&??摑???????2?Y0?Y?-????>=g?)Z2???>?mtXZY0?Y?-??ZqX.-??2K?????9???30$T5?>?;#'6uXf黟4统w?"2??%$1?@?
m_rel_P.ctrec_eval.htrec_format.h	`A("?????MJ??vX
.//JT&?}B?
m_success.ctrec_eval.htrec_format.h	,A*"????MN??vX
./NU%??@?
m_infap.ctrec_eval.htrec_format.h	?A*(??惃????Y[8V^[u??i?.?h%??B?
m_map_cut.ctrec_eval.htrec_format.h	<A*(/???????M8??Yu%./8UO??C?
m_gm_bpref.ctrec_eval.htrec_format.h	A&(????e6??????Y[8V3[uPhX.?ks?S?T/?
m_runid.ctrec_eval.h	?A#u?Y2=?"Y?D?
m_relstring.ctrec_eval.htrec_format.h	PA2(2????1?g?/??Kx<
?"?2=;Y3=??????Yc@?
m_set_P.ctrec_eval.htrec_format.h	?A'????:?hE?
m_set_recall.ctrec_eval.htrec_format.h	@A$????:?hD?
m_set_rel_P.ctrec_eval.htrec_format.h	?A'???!F?fB?
m_set_map.ctrec_eval.htrec_format.h	tA&???!X?l@?
m_set_F.ctrec_eval.htrec_format.h	, A)"????##?Z?rP?
m_num_nonrel_judged_ret.ctrec_eval.htrec_format.h	T!A$???T?~O?
m_prefs_num_prefs_poss.ctrec_eval.htrec_format.h	?!A#?????&&&3#?yN?
m_prefs_num_prefs_ful.ctrec_eval.htrec_format.h	?"A%?????&&?#?zR?
m_prefs_num_prefs_ful_ret.ctrec_eval.htrec_format.h	?#A$?????&?#?{E?
m_prefs_simp.ctrec_eval.htrec_format.h	?$A(?????&&&&&
6u5??E?
m_prefs_pair.ctrec_eval.htrec_format.h	&A("?????PY?-.??PY?-.??LU?	?u/??F?
m_prefs_avgjg.ctrec_eval.htrec_format.h	x)A+"???&&&&&ux?.1?4N?
m_prefs_avgjg_Rnonrel.ctrec_eval.htrec_format.h	$+A:(???&&?&.&..@@&/m$.4?
.$??AL?'Z'0#V2SWXWX???'0?0#V2SWVWX????'YG??3?/?S??/?S(?/?Sm?X?3?/?S(?/?Su?.???uI?
m_prefs_simp_ret.ctrec_eval.htrec_format.h	?4A+????&&?u5??I?
m_prefs_pair_ret.ctrec_eval.htrec_format.h	?5A,"?????PY?-.u/?}J?
m_prefs_avgjg_ret.ctrec_eval.htrec_format.h	?7A-"?撟&&u?	?1?R?
m_prefs_avgjg_Rnonrel_ret.ctrec_eval.htrec_format.h	?8A>(???&&?&&@@&/p$.4?
.Ƀ?C?$[$10vSV[V_?f?$1?10vSV#V_fȻ??$Y??0?,?S??,?Su?.???|I?
m_prefs_simp_imp.ctrec_eval.htrec_format.h	@A*?????&&&&5u5??I?
m_prefs_pair_imp.ctrec_eval.htrec_format.h	hAA*"?????PY?-.??PY?-.u/??J?
m_prefs_avgjg_imp.ctrec_eval.htrec_format.h	LDA,"???&&&&uy?.1?~D?
m_map_avgjg.ctrec_eval.htrec_format.h	?EA'???Ƀ??0Y%?uu`.?O??K?
m_Rprec_mult_avgjg.ctrec_eval.htrec_format.h	?GA0"???"?ɟPiB?=Y?U3g??!?/MV?4wXpf.??0]W%???B?
m_P_avgjg.ctrec_eval.htrec_format.h	?KA,????Ƀ??M^?0vX
#/lUpX.??WW%?l??
m_yaap.ctrec_eval.htrec_format.h	0NA&)?惑??Y%?q?
????x?8@rA?C
?@?A?C
?@|A?C
@A?C
@?A?C
?@nA?C
?@@A?C
????x??8@#A?C
?[@?A?C
? @?A?C
$?? @?A?C
E?$?}"@?A?C
E?$?>$@?A?C
E?$??%@?A?C
E?$?x'@JA?C
E?$??)@3A?C
E?$??+@?A?C
E?$??-@LA?C
E???/@$A?C
?0@ZA?C
?i0@?A?C
??0@?A?C
??1@?A?C
????x?PT2@A?C
Po2@vA?C
P?2@?A?C
????x???3@A?C
??3@?A?C
?54@?A?C
$??4@A?C
E?????x?h?5@A?C
h6@?A?C
h?6@?A?C
h?7@A?C
????x??8@A?C
?8@?A?C
?9@?A?C
?:@8A?C
?;@A?C
????x?$??<@A?C
H???A@QA?C
?B@?A?C
?E@?A?C
????x?X?E@?A?C
X?K@QA?C
X?K@?A?C
X?O@?A?C
????x?$?4P@zA?C
H???U@QA?C
??U@?A?C
??Y@?A?C
????x?$?DZ@?A?C
H??_@QA?C
?h_@A?C
?pb@?A?C
????x?$0c@?A?C
H?0?i@wA?C
0j@A?C
0m@?A?C
????x???m@?	A?C
?lw@?A?C
?x@oA?C
?~x@%A?C
??x@?A?C
????x??	Py@?A?C
?	?@?A?C
?	??@oA?C
?	!?@%A?C
?	F?@?A?C
????x?@
$?@?A?C
@
??@rA?C
@
l?@?A?C
@
b?@SA?C
@
??@?A?C
@
??@|A?C
@
?@oA?C
@
?@^A?C
@
ݯ@?A?C
@
a?@?A?C
@
?@'A?C
@
?@?A?C
@
Ա@oA?C
@
C?@%A?C
@
h?@&A?C
@
??@?A?C
@
??@?A?C
$@
8?@?A?C
E?$@
,?@?A?C
E?$@
k?@BA?C
E?$@
??@?A?C
E?@
=?@?A?C
@
к@A?C
????x?X
Լ@?A?C
X
a?@?A?C
????x?$?
$?@?A?C
E??
??@QA?C
?
?@A?C
?
?@zA?C
????x?P??@?A?C
Pv?@?A?C
PL?@?A?C
????x??$?@	A?C
????x?0?@?A?C
????x?8??@DA?C
8?@bA?C
????x??h?@pA?C
????x????@pA?C
?H?@sA?C
????x? ??@pA?C
????x?X,?@_A?C
????x????@A?C
????x????@?A?C
????x?X?@A?C
????x?8t?@?A?C
????x?p?@?A?C
????x????@?A?C
????x?$?h?@{A?C
H?????x? ??@?A?C
????x?X??@A?C
$XW?@
A?C
E?Xd?@0A?C
X??@?A?C
????x???@A?C
????x?08?@?A?C
$0??@
A?C
E?0??@0A?C
0?@?A?C
????x????@?A?C
$?pA
A?C
E??}A0A?C
??A?A?C
????x?p4A^A?C
????x???A	A?C
$??A
A?C
E???A0A?C
??A?A?C
????x?H`A?A?C
????x??,A?A?C
????x???AeA?C
????x??<A?A?C
????x?(A?A?C
????x?`?A.A?C
`AJA?C
????x??PAxA?C
??AVA?C
?A?A?C
????x?0?A?A?C
????x?h@A?A?C
????x???A?A?C
????x??tA?A?C
????x?, A%A?C
????x?HT!A?A?C
????x???!AA?C
????x???"A?A?C
????x???#A?A?C
????x?(?$AxA?C
????x?`&AkA?C
????x??x)A?A?C
????x??$+A?A?C
?!.A?A?C
????x?(?4A?A?C
????x?`?5A?A?C
????x???7A2A?C
????x???8AsA?C
?G;A?A?C
????x?(@ARA?C
????x?`hAA?A?C
????x??LDA?A?C
????x???EA?A?C
????x??GA
A?C
????x?@?KA?A?C
????x?$x0NATA?C
H?_IO_FILEte_form_inter_procste_num_rel_info_formatlong unsigned intte_results_formatte_rel_info_formatexplanationte_trec_measureste_num_trec_measure_nicknamescleanupte_num_trec_measuresget_filete_num_results_formatte_num_form_inter_procsnameoptargte_trec_measure_nicknameste_form_inter_procslong unsigned intte_results_formatte_rel_info_formatexplanationcleanupget_filerel_info_formatvaluesstart_ptrcutoffs_IO_FILEvaluefull_namenum_paramslong unsigned intnum_cutoffsevalparamslength_requirednameparam_stringmeas_arg_ptrlong unsigned intaccum_evalq_evallong unsigned intall_rel_infonum_queriesaccum_evallong unsigned intlong unsigned int_IO_FILElong unsigned intrun_id_ptr_IO_FILElong unsigned int_IO_FILElong unsigned int_IO_FILElong unsigned int_IO_FILElong unsigned int_IO_FILErel_levelslong unsigned int_IO_FILErel_levelslong unsigned intfull_arraynum_judgedresults_prefs_IO_FILEtrec_prefslocationlong unsigned intprefs_and_rankstext_results_infonum_prefsptr1rank_pool_ptrdocnoptr2num_resultsnum_judged_retrel_levelcurrent_boundlong unsigned int_IO_FILElong unsigned intall_zscoreslong unsigned intzscoreste_meas_infAPte_meas_num_nonrel_judged_rette_meas_prefs_avgjg_impte_meas_prefs_pair_impte_meas_runidte_meas_prefs_simp_impte_meas_prefs_num_prefs_fulte_meas_num_qte_meas_Rprecte_meas_prefs_num_prefs_ful_rette_meas_ndcgte_meas_gm_mapte_meas_Rndcgte_meas_set_relative_Pte_meas_P_avgjgte_meas_recip_rankte_meas_ndcg_cutlong unsigned intte_meas_ndcg_relte_meas_set_mapte_meas_num_rel_rette_meas_mapte_meas_utilityte_trec_measureste_meas_recallte_meas_yaapte_meas_successte_meas_relative_Pte_meas_prefs_avgjg_rette_meas_num_rette_meas_prefs_pair_rette_meas_map_avgjgte_meas_Rprec_multte_meas_prefs_simp_rette_meas_prefs_avgjg_Rnonrelte_meas_set_Fte_meas_set_Pte_meas_prefs_pairte_meas_gm_bprefte_meas_map_cutte_meas_binGte_meas_prefs_num_prefs_posste_meas_bprefte_meas_num_relte_meas_prefs_avgjg_Rnonrel_rette_meas_11pt_avgte_meas_prefs_avgjgte_meas_relstringte_meas_Gte_meas_Pte_meas_iprec_at_recallte_meas_prefs_simpte_meas_set_recallte_meas_Rprec_mult_avgjgte_trec_measure_nicknameslong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infonum_querieslong unsigned intrel_info_IO_FILEtext_qrelstrec_qrelsnum_text_qrelslong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_info_IO_FILElong unsigned intrel_info_IO_FILElong unsigned intrel_infolong unsigned intnum_gainsnum_at_levelres_relsrel_inforel_levellong unsigned intrel_infolong unsigned intnum_gainsnum_at_levelres_relsrel_inforel_levellong unsigned intnum_gainsnum_rel_retnum_at_levelres_relsrel_inforel_levellong unsigned intrel_infolong unsigned intnum_gainsnum_at_levelres_relsrel_inforel_levellong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_possnum_judgedlong unsigned intrel_infonum_judged_retfull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_possnum_judgedlong unsigned intrel_infonum_judged_retfull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infofull_arraynum_judgedlong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infolong unsigned intrel_infowwrvrswsvwvRvRSwSVwV?v??w??w??
v?
?
w?
?
w?
PvPQwQTwT?v??w??w??vww#v#$w$'w'?v??w??w??v??w??w?EvEFwFIwIvw
w
?v??w??w?@v@AwADwD?
v?
?
w?
?
w?
?v??w??w?gvghwhkwk?v??w??w??v??w??w?1v12w25w5?v??w??w?RvRSwSVwVvwwvww?v??w??w?4vwwvww?v??w??w?hvhiwilwljvwwvww?v??w??w??v??w??w??vwwvww?v??w??w??v??w??w??v??w??w??vwwvwwXvXYwY\w\JvJKwKNwN?vww?v??w??w?@v@AwADwD?	v?	?	w?	?	w?	?
vwwzvz{w{~w~?v??w??w?p	vp	q	wq	t	wt	
vww?v??w??w?$v$%w%(w(,v,-w-0w0?vww?v??w??w??v??w??w?
v

w

w
?
vww?	v?	?	w?	?	w?	7
v7
8
w8
;
w;
?
v?
?
w?
?
w?
?
v?
?
w?
?
w?
vvww?v??w??w?bvbcwcfwf?v??w??w??v??w??w??
vww?v??w??w?HvHIwILwL>v>?w?BwB?v??w??w?pvpqwqtwt?v??w??w?[ v[ \ w\ _ w_ ?(v?(?(w?(?(w?(=)v=)>)w>)A)wA)?)v?)?)w?)?)w?)?)v?)?)w?)?)w?)?*v?*?*w?*?*w?*+v+ +w +#+w#+D+vD+E+wE+H+wH+j+vj+k+wk+n+wn+a,va,b,wb,e,we,-v--w--w-.v.	.w	..w.G/vG/H/wH/K/wK/?0v?0?0w?0?0w?03v33w33w3?3v?3?3w?3?3w?3?5vww?v??w??w?Nvww?v??w??w??v??w??w??v??w??w?nvww?v??w??w??v??w??w??vww	vww?vwwDvDEwEHwH?vwwpvwwpvpqwqtwt?vwwpvww_vwwvww?vwwvww?vww?vww?vww{vww?vwwv?w??w??v??w??w??v??w??w?B	vwwvww?v??w??w??	v?	?	w?	?	w?	?	v?	?	w?	?	w?	\
vww?v??w??w??	v?	?	w?	?	w?	
v

w

w
?
vww^vww	v	
w

w
	v		w		w	F	vF	G	wG	J	wJ	?	vww?vww?vwwevww?vww?vww.v./w/2w2xvwwxvxywy|w|?v??w??w?]vww?vww?vww?vww?vww%vww?vwwvww?vww?vwwxvwwkvww?vww?v??w?w?	vww?vww?vww2vwwsvstwtwww=vwwRvww?vww?vww?vww
vww?vwwTv.symtab.strtab.shstrtab.interp.note.ABI-tag.gnu.hash.dynsym.dynstr.gnu.version.gnu.version_r.rela.dyn.rela.plt.init.text.fini.rodata.eh_frame_hdr.eh_frame.ctors.dtors.jcr.dynamic.got.got.plt.data.bss.comment.debug_aranges.debug_pubnames.debug_info.debug_abbrev.debug_line.debug_frame.debug_str.debug_loc@#@ 1???o@@@4;x@x?C@6K???o>@>LX???o?@?Pg?@?xqX@X?{@@@vX@X?`
@`
C?hPAhP?xPAxP???XBX|??"B?"??@b@?@b@? @b @?(@b(@???Ab?A??Ab?A??Bb?B?" ??eb?eP ??e?
?8s??(?>
f?@??I$??:0??=?6?H G30zSȄ?1$u	x?Q@@@@x@@>@?@?@	X@
@@X@`
@
hPAxPAXB?"B@b@b @b(@b?Ab?Ab?Bb?eb !?
@??@b*@b8 @bE?ebS?ebb?
@x@???@b??8B? @b?0PA????Cb?Cb?Cb? Cb?@	?@n@?.?@@6?@?H?@|\??f??rx'@J?i0@???)@3??0@???+@???1@???-@L??/@$?0@Z??????????/??;?ebJ?ebY?ebi?ebwB@???A@Q?????ebJ?eb?fb?fb??K@???K@Q???fbJfb fbi(fb ?U@???U@Q1??0fbJ8fb@fbiHfbCh_@?_@QZ??;PfbiXfb{`fbYhfbipfbwj@??i@w?????Fb??fb??fb??fb??fb?fb@gb!gb,x@o;~x@%F?fbQlw@?b????Fb?gb?gbu gb?(gb?0gb?Hgb?@gbPgb!Xgb,??@o;!?@%F8gbQ?@??????Fb?`gbuhgb?pgb?xgb??gb??gb??gb??gb??gb?gb?gb&?gb??gb??gb7?gbB?gbQ?gbY?gbehbu?gb}?gb??@^?hb?hb?hb?a?@??b?@S???@r?=?@??l?@?
ݯ@???@?.??@|@(hbM0hb^8hbo@hb??@o? hb,Ա@o???@?;C?@%??@'???@??h?@&?@?18?@?@,?@?Xk?@Bq??@????????Hhb?Phb?Xhb??@???@Q???v?@?L?@???#`Hbh-?Hb8Jb`B`Jb?NKbH]`Kb0l??t$?@	?????LbH??Lb?0?@???????@D??@b????h?@p???	??@p	H?@s*	??:	??@pN	??Y	,?@_h	??r	??@?	???	??@??	???	X?@?	???	`PbX?	?Pb
t?@?
???@QbH$
?Qb;
?@?J
??Y
 RbPl
pRb?
??@??
???
Sb ?
 Sb?
h?@{?
???	?SbX?
Tb??@???`hb-??@:W?@
F??@?Od?@0]????TbHj0Ub?@????hb?8?@?:??@
F?@?O??@0????hb???@?:pA
F?A?O}A0????4A^?????hb??A	:?A
F?A?O?A0????@WbH?Wb!`A?/??; XbP@Xbh,A?x????Ae???? YbH?pYb?<A?????A??????A.?AJ
?hb
??
?Zb'
?Zb@
?hbK
PAx]
?AVr
A??
?hb?
???
?A??
???
@A??
???
?A??
???
tA????\b+?\b@, A%N??hT!A??????!A?????"A?????#A?,??;?$AxN??]&Akp???x)A?????$+A??!.A??????4A?????5A?(??<?7A2T??p?8As?G;A?????@AR????hAA?????LDA????EA?0??Y
?cbPE?cba?GA
z???`dbH??db??KA?????0NAT?@b?@b??Ab@b@b(@b9@bL(@bU?Fb?f ?Bbq???bbP?o2@v??_bP??? Eb???OA??	`
@`XbP l3`[bPA ]bPO?b}"@??? @??Py@?? ? ????NbP???:@8(
hPA.C_bPV`bbPmZbP~????$??NbP????Eb? @? LbP?\bP*?E@?>?PbPV?8@?r4P@z?7@??^bP?	??VbP??2@?? \bPxPAT2@#?5?XbPC?TbPP?ebdbbP{?Bb??UbP??`bP??Eb?Lb?7??VbP`_bP??@?-?dbP=?Kb`W?x@?p?abP???ObP?PbP?@SbP??3@??8@?PA?<@:1?DFb?X@be?OA?u@abP??eb? TbP?8@#??Fb?F?@???[bPpb@?'>$@?D?9@?_m@?w?3@???Eb ?`MbP?$?@???RbP??Y@???[bP?%@?(?m@?	9E@?N?a???ebm6@?? `bP?!??`bP?a?@????@ebPc@?MbP@^bP:$?@?L[@?a%u VbP????hb??O@??к@????4@?Լ@?DZ@?54@?* NbP:?ebN?5@i cbP{?@z?c?`ZbP?dbP?u??MbP??]bP@ObP`UbP/???eb6?K@Hb`?YbPp?;@????QbP?(????eb?8@r?
@@??]bP?"?6@?>?WbPcall_gmon_startcrtstuff.c__CTOR_LIST____DTOR_LIST____JCR_LIST__dtor_idx.6147completed.6145__do_global_dtors_auxframe_dummy__CTOR_END____FRAME_END____JCR_END____do_global_ctors_auxtrec_eval.cVersionIDhelp_messageusagelong_options.5036mark_measureget_debug_level_querytrec_eval_helpcleanupadd_meas_arg_infomark_single_measureformats.cmeas_init.cget_long_cutoffsappend_longget_float_cutoffsappend_floatget_float_paramsappend_stringget_param_pairscomp_longcomp_floatmeas_acc.cmeas_avg.cmeas_print_single.cmeas_print_final.cget_qrels.ctrec_qrels_buftext_info_pooltext_qrels_poolrel_info_poolparse_qrels_linecomp_lines_qid_docnoget_trec_results.ctrec_results_buftext_results_poolq_results_poolparse_results_lineget_prefs.ctrec_prefs_buftext_prefs_poolparse_prefs_lineget_qrels_prefs.cparse_qrels_prefs_lineget_qrels_jg.ctext_jg_info_pooltext_jg_poolcomp_lines_qid_jg_docnoform_res_rels.ccurrent_querymax_current_querymax_rel_levelsmax_ranked_rel_listmax_docno_infosaved_res_relsranked_rel_listdocno_infocomp_sim_docnocomp_docnorel_levelscomp_rank_judgedform_res_rels_jg.cmax_num_jgsnum_jgsjgsform_prefs_counts.cmax_rank_poolmax_ec_poolmax_ca_poolmax_ca_ptr_poolmax_pa_poolmax_pa_ptr_poolmax_rel_poolmax_prefs_and_ranksmax_docno_resultssaved_num_judgednum_judgednum_judged_retca_poolca_ptr_poolprefs_and_ranksec_poolrank_poolform_prefs_and_ranksrel_poolpa_poolpa_ptr_poolinit_counts_arrayform_jg_paform_jg_ecdebug_print_results_prefsadd_ec_pref_to_jginit_prefs_arrayadd_transitivesadd_pa_pref_to_jgtemp_pa_poolmax_temp_pa_pooltemp_pa_ptr_poolmax_temp_pa_ptr_poolmult_and_check_changedocno_resultsdebug_print_docno_resultscomp_prefs_and_ranks_docnodebug_print_prefs_and_rankscomp_results_inc_rankcomp_prefs_and_ranks_jg_rel_leveldebug_print_ecdebug_print_prefs_arraydebug_print_counts_arraydebug_print_jgutility_pool.cget_zscores.ctrec_zscores_buftext_zscores_poolzscores_poolparse_zscore_linecomp_lines_qid_measconvert_zscores.cfind_qidfind_measmeasures.coff_namestrec_namesset_namesprefs_namesprefs_off_nameqrels_jg_namesm_map.cte_calc_mapm_P.clong_cutoff_arraydefault_P_cutoffste_calc_Pm_num_q.cte_calc_num_qte_calc_avg_num_qm_num_ret.cte_calc_num_retm_num_rel.cte_calc_num_relte_calc_avg_num_relm_num_rel_ret.cte_calc_num_rel_retm_gm_map.cte_calc_gm_mapm_Rprec.cte_calc_Rprecm_recip_rank.cte_calc_recip_rankm_bpref.cte_calc_bprefm_iprec_at_recall.cfloat_cutoff_arraydefault_iprec_at_recall_cutoffste_calc_iprec_at_recallm_recall.cdefault_recall_cutoffste_calc_recallm_Rprec_mult.cRprec_cutoff_arraydefault_Rprec_cutoffste_calc_Rprec_multm_utility.cutility_param_arraydefault_utility_paramste_calc_utilitym_11pt_avg.cdefault_11ptavg_cutoffste_calc_11ptavgm_ndcg.cdefault_ndcg_gainste_calc_ndcgsetup_gainsget_gaincomp_rel_gainm_ndcg_cut.cdefault_ndcg_cutoffste_calc_ndcg_cutm_Rndcg.cte_calc_Rndcgm_ndcg_rel.cte_calc_ndcg_relm_binG.cte_calc_binGm_G.cdefault_G_gainste_calc_Gm_rel_P.cdefault_relative_P_cutoffste_calc_rel_Pm_success.csuccess_cutoff_arraydefault_success_cutoffste_calc_successm_infap.cte_calc_infapm_map_cut.cdefault_map_cutoffste_calc_map_cutm_gm_bpref.cte_calc_gm_bprefm_runid.cte_calc_runidte_print_runidrunidm_relstring.crelstring_lendefault_relstring_paramsstring_lente_calc_relstringte_print_q_relstringte_print_relstringcurrent_stringm_set_P.cte_calc_set_Pm_set_recall.cte_calc_set_recallm_set_rel_P.cte_calc_set_relative_Pm_set_map.cte_calc_set_mapm_set_F.cset_F_param_arraydefault_set_F_paramste_calc_set_Fm_num_nonrel_judged_ret.cte_calc_num_nonrel_judged_retm_prefs_num_prefs_poss.cte_calc_prefs_num_prefs_possm_prefs_num_prefs_ful.cte_calc_prefs_num_prefs_fulm_prefs_num_prefs_ful_ret.cte_calc_prefs_num_prefs_ful_retm_prefs_simp.cte_calc_prefs_simpm_prefs_pair.cte_calc_prefs_pairm_prefs_avgjg.cte_calc_prefs_avgjgm_prefs_avgjg_Rnonrel.cte_calc_prefs_avgjg_Rnonrelrecalculatem_prefs_simp_ret.cte_calc_prefs_simp_retm_prefs_pair_ret.cte_calc_prefs_pair_retm_prefs_avgjg_ret.cte_calc_prefs_avgjg_retm_prefs_avgjg_Rnonrel_ret.cte_calc_prefs_avgjg_Rnonrel_retm_prefs_simp_imp.cte_calc_prefs_simp_impm_prefs_pair_imp.cte_calc_prefs_pair_impm_prefs_avgjg_imp.cte_calc_prefs_avgjg_impm_map_avgjg.cte_calc_map_avgjgm_Rprec_mult_avgjg.cdefault_Rprec_avgjg_cutoffste_calc_Rprec_mult_avgjgm_P_avgjg.cdefault_P_avgjg_cutoffste_calc_P_avgjgm_yaap.cte_calc_yaap__preinit_array_start__fini_array_end_GLOBAL_OFFSET_TABLE___preinit_array_end__fini_array_start__init_array_end__init_array_start_DYNAMICte_trec_measuresdata_startexp@@GLIBC_2.2.5te_meas_prefs_avgjg_impte_acc_meas_ste_meas_prefs_avgjgprintf@@GLIBC_2.2.5te_rel_info_format__libc_csu_finisnprintf@@GLIBC_2.2.5_startte_meas_successclose@@GLIBC_2.2.5te_meas_set_Pte_meas_set_Fabort@@GLIBC_2.2.5te_init_meas_a_float_cut_floatte_init_meas_a_float_cut_longte_form_res_rels_jg__gmon_start___Jv_RegisterClassesputs@@GLIBC_2.2.5te_meas_num_rel_retexit@@GLIBC_2.2.5te_print_final_meas_a_cut_finiputchar@@GLIBC_2.2.5te_meas_prefs_simpte_meas_prefs_pair_impte_meas_gm_bpreflog2@@GLIBC_2.2.5read@@GLIBC_2.2.5malloc@@GLIBC_2.2.5te_meas_gm_map__libc_start_main@@GLIBC_2.2.5te_num_results_formatte_init_meas_s_longte_meas_mapte_meas_set_mapte_get_trec_resultste_meas_iprec_at_recallte_print_final_meas_s_floatte_get_prefste_print_single_meas_a_cutte_meas_prefs_num_prefs_ful_retatof@@GLIBC_2.2.5te_meas_binGte_acc_meas_a_cutte_meas_set_relative_P_IO_stdin_usedte_acc_meas_emptyfree@@GLIBC_2.2.5te_meas_infAPte_meas_ndcgoptind@@GLIBC_2.2.5te_meas_prefs_simp_imp__data_startte_meas_Rndcgte_meas_prefs_simp_rette_num_rel_info_formatte_num_trec_measure_nicknames__ctype_b_loc@@GLIBC_2.3te_meas_Gte_meas_prefs_pairte_convert_to_zscorete_meas_P_avgjgte_trec_measure_nicknameste_form_res_rels_cleanupte_meas_prefs_avgjg_Rnonrel_retatol@@GLIBC_2.2.5te_meas_recip_rankte_meas_bprefte_meas_utilityte_calc_avg_meas_emptyte_print_final_meas_empty__dso_handlete_get_qrelsfputs@@GLIBC_2.2.5lseek@@GLIBC_2.2.5te_form_inter_procs__DTOR_END____libc_csu_initte_meas_prefs_avgjg_retoptarg@@GLIBC_2.2.5te_meas_11pt_avgte_init_meas_emptyte_num_form_inter_procste_form_res_rels_jg_cleanupte_meas_set_recallte_get_qrels_prefs_cleanupte_init_meas_s_float_p_floatte_print_final_meas_s_longte_get_qrels_jg_cleanupte_calc_avg_meas_ste_results_formatte_meas_num_qte_get_zscoreste_meas_Rprec_multte_get_prefs_cleanupgetopt_long@@GLIBC_2.2.5te_meas_relstringte_init_meas_s_float_p_pairte_form_res_relste_get_qrels_cleanupqsort@@GLIBC_2.2.5__bss_startte_print_single_meas_s_floatte_meas_prefs_avgjg_Rnonrelstrcmp@@GLIBC_2.2.5te_meas_prefs_pair_rette_chk_and_reallocindex@@GLIBC_2.2.5te_meas_yaapte_get_qrels_jgte_meas_Pte_meas_prefs_num_prefs_fulform_prefs_countste_init_meas_s_floatmunmap@@GLIBC_2.2.5te_meas_ndcg_rel_endte_get_trec_results_cleanupte_form_pref_counts_cleanupstrncpy@@GLIBC_2.2.5te_calc_avg_meas_s_gmte_chk_and_mallocte_get_qrels_prefste_calc_avg_meas_a_cutte_meas_num_relstderr@@GLIBC_2.2.5te_print_single_meas_emptyte_meas_map_avgjgte_get_zscores_cleanupfwrite@@GLIBC_2.2.5te_meas_runidte_meas_Rprec_mult_avgjgrealloc@@GLIBC_2.2.5te_meas_num_rette_meas_num_nonrel_judged_rette_meas_Rprecte_meas_ndcg_cut_edatafprintf@@GLIBC_2.2.5te_num_trec_measureste_meas_map_cutte_print_final_meas_s_float_popen@@GLIBC_2.2.5te_meas_recallmmap@@GLIBC_2.2.5log@@GLIBC_2.2.5stdout@@GLIBC_2.2.5main_initte_meas_prefs_num_prefs_possfflush@@GLIBC_2.2.5te_print_single_meas_s_longte_meas_relative_P




© 2015 - 2025 Weber Informatics LLC | Privacy Policy