Categories and Subject Descriptors J.4 [Computer Applications] SOCIAL AND BEHAVIORAL SCIENC
Mesh 医学主题词表

Mesh 医学主题词表The following is a table of the categories and subcategories of the tree structure chart:A。
AnatomyA1.Body ns___A3.Digestive SystemA4.Respiratory SystemA5.Urogenital SystemA6.Endocrine SystemA7.vascular SystemA8.Nervous SystemA9.Sense OrgansA10.TissuesA11.CellsA12.Fluids and nsA13.Animal Structures___A15.___A16.Embryonic StructuresB。
OrganismsB1.InvertebratesB2.___B3.BacteriaB4.VirusesB5.Algae and FungiB6.PlantsB7.ArchaeaC。
Diseases___C2.Virus DiseasesC3.Parasitic DiseasesC4.NeoplasmsC5.Musculoskeletal Diseases C6.Digestive System Diseases ___C8.Respiratory Tract Diseases___C10.Nervous System DiseasesC11.Eye Diseases______ of the tree structure chart。
The categories are Anatomy。
Organisms。
and Diseases。
with ___ each。
The Anatomy category includes Body ns。
Musculoskeletal System。
Digestive System。
Respiratory System。
Association_for_Computing_Machinery__ACM__SIG_Proceedings

A Sample ACM SIG Proceedings Paper in LaTeX[Extended Abstract]Author One∗Institute OneAddress One author.one@Author TwoInstitute T woAddress T wo author.two@Author ThreeInstitute ThreeAddress Three author.three@Author FourInstitute FourAddress Four author.four@ABSTRACTAbstract text.Abstract text.Abstract text.Abstract text. Abstract text.Abstract text.Abstract text. Categories and Subject DescriptorsH.4[Information Systems Applications]:Miscellaneous;D.2.8[Software Engineering]:Metrics—complexity mea-sures,performance measuresGeneral TermsTheoryKeywordsACM proceedings,L A T E X,text tagging1.INTRODUCTIONSample text.Sample text.Sample text.Sample text.Sam-ple text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text. Sample text.Sample text.Sample text.Sample text.Sam-ple text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text. Sample text.Sample text.Sample text.Sample text.Sam-ple text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text.Sample text. Sample text.Sample text.Sample text.Citation of Einstein paper[1].2.RESULTS AND DISCUSSIONSLorem ipsum dolor sit amet,consectetur adipiscing elit,sed do eiusmod tempor incididunt ut labore et dolore magna ali-qua.Ut enim ad minim veniam,quis nostrud exercitation ∗Now on postdoctoral fellow at ABC University ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.Excepteur sint occae-cat cupidatat non proident,sunt in culpa qui officia deserunt mollit anim id est laborum.Lorem ipsum dolor sit amet,consectetur adipiscing elit,sed do eiusmod tempor incididunt ut labore et dolore magna ali-qua.Ut enim ad minim veniam,quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.Excepteur sint occae-cat cupidatat non proident,sunt in culpa qui officia deserunt mollit anim id est laborum.Lorem ipsum dolor sit amet,consectetur adipiscing elit,sed do eiusmod tempor incididunt ut labore et dolore magna ali-qua.Ut enim ad minim veniam,quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.Excepteur sint occae-cat cupidatat non proident,sunt in culpa qui officia deserunt mollit anim id est laborum.Lorem ipsum dolor sit amet,consectetur adipiscing elit,sed do eiusmod tempor incididunt ut labore et dolore magna ali-qua.Ut enim ad minim veniam,quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur.Excepteur sint occae-cat cupidatat non proident,sunt in culpa qui officia deserunt mollit anim id est laborum.3.REFERENCES[1]A.Einstein.Zur Elektrodynamik bewegter K¨o rper.(German)[On the electrodynamics of moving bodies].Annalen der Physik,322(10):891–921,1905.。
General Terms

The University of Amsterdam at WebCLEF2006Krisztian Balog Maarten de RijkeISLA,University of Amsterdamkbalog,mdr@science.uva.nlAbstractOur aim for our participation in WebCLEF2006was to investigate the robustness ofinformation retrieval techniques to crosslingual retrieval,such as compact documentrepresentations,and query reformulation techniques.Our focus was on the mixedmonolingual task.Apart from the proper preprocessing and transformation of variousencodings,we did not apply any language-specific techniques.Instead,the targetdomain metafield was used in some of our runs.A standard combSUM combinationusing Min-Max normalization was used to combine runs,based on a separate contentand title indexes of documents.We found that the combination is effective only forthe human generated topics.Query reformulation techniques can be used to improveretrieval performance,as witnessed by our best scoring configuration,however thesetechniques are not yet beneficial to all different kinds of topics.Categories and Subject DescriptorsH.3[Information Storage and Retrieval]:H.3.1Content Analysis and Indexing;H.3.3Infor-mation Search and Retrieval;H.3.4Systems and Software;H.3.7Digital Libraries;H.2.3[Database Managment]:Languages—Query LanguagesGeneral TermsMeasurement,Performance,ExperimentationKeywordsWeb retrieval,Multilingual retrieval1IntroductionThe world wide web is a natural setting for cross-lingual information retrieval,since the web content is essentially multilingual.On the other hand,web data is much noisier than traditional collections, eg.newswire or newspaper data,which originated from a single source.Also,the linguistic variety in the collection makes it harder to apply language-dependent processing methods,eg.stemming algorithms.Moreover,the size of the web only allows for methods that scale well.We investigate a range of approaches to crosslingual web retrieval using the test suite of the CLEF2006WebCLEF track,featuring a stream of known-item topics in various languages.The topics are a mixture of human generated(manually)and automatically generated topics.Our focus is on the mixed monolingual task.Our aim for our participation in WebCLEF2006was to investigate the robustness of information retrieval techniques,such as compact document repre-sentations(titles or incoming anchor-texts),and query reformulation techniques.The remainder of the paper is organized as follows.In Section2we describe our retrieval system as well as the specific approaches we applied.In Section3we describe the runs that we submitted,while the results of those runs are detailed in Section4.We conclude in Section5.2System DescriptionOur retrieval system is based on the Lucene engine[4].For our ranking,we used the default similarity measure of Lucene,i.e.,for a collection D, document d and query q containing terms t i:sim(q,d)=t∈q tf t,q·idf tnorm q·tf t,d·idf tnorm d·coord q,d·weight t,wheretf t,X=freq(t,X),idf t=1+log|D| freq(t,D),norm d= |d|,coord q,d=|q∩d||q|,andnorm q=t∈qtf t,q·idf t2.We did not apply any stemming nor did we use a stopword list.We applied case-folding and normalized marked characters to their unmarked counterparts,i.e.,mapping´a to a,¨o to o,æto ae,ˆıto i,etc.The only language specific processing we did was a transformation of the multiple Russian and Greek encodings into an ASCII transliteration.We extracted the full text from the documents,together with the title and anchorfields,and created three separate indexes:•Content:Index of the full document text.•Title:Index of all<title>fields.•Anchors:Index of all incoming anchor-texts.We performed three base runs using the separate indexes.We evaluated the base runs using the WebCLEF2005topics,and decided to use only the content and title indexes.2.1Run combinationsWe experimented with the combination of content and title runs,using standard combination methods as introduced by Fox and Shaw[1]:combMAX(take the maximal similarity score of the individual runs);combMIN(take the minimal similarity score of the individual runs);combSUM(take the sum of the similarity scores of the individual runs);combANZ(take the sum of the similarity scores of the individual runs,and divide by the number of non-zero entries);combMNZ(take the sum of the similarity scores of the individual runs,and multiply by the number of non-zero entries); and combMED(take the median similarity score of the individual runs).Fox and Shaw[1]found combSUM to be the best performing combination method.Kamps and de Rijke[2]conducted extensive experiments with the Fox and Shaw combination rules for nine european languages,and demonstrated that combination can lead into significant improvements.Moreover,they proved that the effectiveness of combining retrieval strategies differs between En-glish and other European languages.In Kamps and de Rijke[2],combSUM emerges as the best combination rule,confirming Fox and Shaw’sfindings.Similarity score distributions may differ radically across runs.We apply the combination methods to normalized similarity scores.That is,we follow Lee[3]and normalize retrieval scores into a[0,1]range,using the minimal and maximal similarity scores(min-max normalization):s =s−minmax−min,(1)where s denotes the original retrieval score,and min(max)is the minimal(maximal)score over all topics in the run.For the WebCLEF2005topics the best performance was achieved using the combSUM combi-nation rule,which is in conjunction with thefindings in[1,2],therefore we used that method for our WebCLEF2006submissions.2.2Query reformulationIn addition to our run combination experiments,we conducted experiments to measure the effect of phrase and query operations.We tested query-rewrite heuristics using phrases and word n-grams.Phrases In this straightforward mechanism we simply add the topic terms as a phrase to the topic.For example,for the topic WC0601907,with title“diana memorial fountain”,the query becomes:diana memorial fountain“diana memorial fountain”.Our intuition is that rewarding documents that contain the whole topic as a phrase,not only the individual terms,would be beneficial to retrieval performance.N-grams In our approach every word n-gram from the query is added to the query as a phrase with weight n.This means that longer phrases get bigger ing the previous topic as an example,the query becomes:diana memorial fountain“diana memorial”2“memorial fountain”2“diana memorial fountain”3,where the number in the upper index denotes the weight attached to the phrase(the weight of the individual terms is1).3RunsWe submittedfive runs to the mixed monolingual task:Baseline Base run using the content only index.Comb Combination of the content and title runs,using the CombSUM method.CombMeta The Comb run,using the target domain metafield.We restrict our search to the corresponding domain.CombPhrase The CombMeta run,using the Phrase query reformulation technique. CombNboost The CombMeta run,using the N-grams query reformulation technique.4ResultsTable1lists our results in terms of mean reciprocal rank.Runs are listed along the left-hand side,while the labels indicate either all topics(all)or various subsets:automatically generated (auto)—further subdivided into automatically generated using the unigram generator(auto-u) and automatically generated using the bigram generator(auto-b)—and manual(manual),which is further subdivided into new manual topics and old manual topics.Significance testing was done using the two-tailed Wilcoxon Matched-pair Signed-Ranks Test, where we look for improvements at a significance level of0.05(1),0.001(2),and0.0001(3). Signficant differences noted in Table1are with respect to the Baseline run.Table1:Submission results(Mean Reciprocal Rank)runID all auto auto-u auto-b manual man-n man-oBaseline0.16940.12530.13970.11100.39340.47870.3391Comb0.168510.120830.139430.10210.41120.49520.3578CombMeta0.194730.150530.167030.134130.418830.510810.36031 CombPhrase0.200130.157030.163930.150030.41900.51380.3587CombNboost0.195430.158630.159530.157630.38260.48910.3148Combination of the content and title runs(Comb)increased performance only for the manual topics.The use of the title tag does not help when the topics are automatically generated. Instead of employing a language detection method,we simply used the target domain metafield. The CombMeta run improved the retrieval performance significantly for all subsets of topics.Our query reformulation techniques show mixed,but promising results.The best overall score was achieved when the topic,as a phrase,was added to the query(CombPhrase).The comparison of CombPhrase vs CombMeta reveals that they achieved similar scores for all subsets of topics,except for the automatic topics using the bigram generator,where the query reformulation technique was clearly beneficial.The n-gram query reformulation technique(CombNboost)further improved the results of the auto-b topics,but hurt accuracy on all other subsets,especially on the manual topics. The CombPhrase run demonstrates that even a very simple query reformulation technique can be used to improve retrieval scores.However,we need to further investigate how to automatically detect whether it is beneficial to use such techniques(and if yes,which technique to apply)for a given a topic.Comparing the various subsets of topics,we see that the automatic topics have proven to be more difficult than the manual ones.Also,the new manual topics seem to be more appropriate for known-item search than the old manual topics.There is a clear ranking among the various subsets of topics,and this ranking is independent from the applied methods:man−n man−o auto−u auto−b.5ConclusionsOur aim for our participation in WebCLEF2006was to investigate the robustness of information retrieval techniques to crosslingual web retrieval.The only language-specific processing we applied was the transformation of various encodings into an ASCII transliteration.We did not apply any stemming nor did we use a stopword list.We indexed the collection by extracting the full text and the titlefields from the documents.A standard combSUM combination using Min-Max normalization was used to combine the runs based on the content and title indexes.We found that the combination is effective only for the human generated topics,using the titlefield did not improve performance when the topics are automatically generated.Significant improvements (+15%MRR)were achieved by using the target domain metafield.We also investigated the effect of query reformulation techniques.We found,that even very simple methods can improve retrieval performance,however these techniques are not yet beneficial to retrieval for all subsets of topics.Although it may be too early to talk about a solved problem,effective and robust web retrieval techniques seem to carry over to the mixed monolingual setting.6AcknowledgmentsKrisztian Balog was supported by the Netherlands Organisation for Scientific Research(NWO)un-der project numbers220-80-001,600.-065.-120and612.000.106.Maarten de Rijke was supported by NWO under project numbers017.001.190,220-80-001,264-70-050,354-20-005,600.-065.-120, 612-13-001,612.000.106,612.066.302,612.069.006,640.001.501,and640.002.501. References[1]E.Fox and bination of multiple searches.In The Second Text REtrieval Con-ference(TREC-2),pages243–252.National Institute for Standards and Technology.NIST Special Publication500-215,1994.[2]J.Kamps and M.de Rijke.The effectiveness of combining information retrieval strategiesfor European languages.In Proceedings of the2004ACM Symposium on Applied Computing, pages1073–1077,2004.[3]bining multiple evidence from different properties of weighting schemes.InSIGIR’95:Proceedings of the18th annual international ACM SIGIR conference on Research and development in information retrieval,pages180–188,New York,NY,USA,1995.ACM Press.ISBN0-89791-714-6.doi:/10.1145/215206.215358.[4]Lucene.The Lucene search engine,2005./.。
科技外文文献原文

AMBULANT:A Fast,Multi-Platform Open Source SML Player Dick C.A. Bulterman, Jack Jansen, Kleanthis Kleanthous, Kees Blom and Daniel Benden CWI: Centrum voor Wiskunde en InformaticaKruislaan 4131098 SJ Amsterdam, The Netherlands +31 20 592 43 00 Dick.Bulterman@cwi.nl ABSTRACTThis paper provides an overview of the Ambulant Open SMIL player. Unlike other SMIL implementations, the Ambulant Player is a reconfigureable SMIL engine that can be customized for use as an experimental media player core.The Ambulant Player is a reference SMIL engine that can be integrated in a wide variety of media player projects. This paper starts with an overview of our motivations for creating a new SMIL engine then discusses the architecture of the Ambulant Core (including the scalability and custom integration features of the player).We close with a discussion of our implementation experiences with Ambulant instances for Windows,Mac and Linux versions for desktop and PDA devices.Categories and Subject Descriptors H.5.1 Multimedia Information Systems [Evaluation]H.5.4 Hypertext/Hypermedia [Navigation]. General TermsExperimentation, Performance, V erification KeywordsSMIL, Player, Open-Source, Demos1.MOTIV ATIONThe Ambulant Open SMIL Player is an open-source, full featured SMIL 2.0 player. It is intended to be used within the researcher community (in and outside our institute) in projects that need source code access to a production-quality SMIL player environment. It may also be used as a stand-alone SMIL player for applications that do not need proprietary mediaformats.The player supports a range of SMIL 2.0 profiles ( including desktop and mobile configurations) and is available in distributions for Linux, Macintosh, and Windows systems ranging from desktop devices to PDA and handheld computers. While several SMIL player implementationsexist,including the RealPlayer [4], InternetExplorer [5], PocketSMIL [7],GRiNS [6],X-SMILES [8] and various proprietary implementations for mobile devices, we developed Ambulant for three reasons:Permission to make digital or hard copiesof all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish,to post on servers or to redistribute tolists,requires prior specific permissionand/or a fee.'MM' 04, October 10-16, 2004, New Y ork, New Y ork, USA.Copyright 2004 ACM 1-58113-893-8/04/0010...$5.00.•N one of the existi ng SMIL players provides a complete and correct SMIL 2.0 implementation. The Ambulant player implements all of SMIL, based on the SMIL 2.0 Language profile plus extensions to support advanced animation and the needs of the mobile variant used by the 3GPP/PSS-6 SMIL specification [9]. •A ll commercial SMIL players are geared to the presentation of proprietary media. The Ambulant player uses open-source media codecs and open-source network transfer protocols, so that the player can be easily customized foruse in a wide range of researchprojects.• Our goal is to build a platform that will encourage the development of comparable multimedia research output.By providing what we expect will be a standard baseline player, other researchers and developmentorganizations can concentrate on integratingextensions to the basic player (either in terms of new media codecs or new network control algorithms). These extensions can then be shared by others.In contrast to the Helix client architecture [10], which also moved to a GPL core in mid-2004, the Ambulant player supports a wider range of SMIL target application architectures,it provides a more complete and correct implementation of the SMIL language,it provides much better performance on low-resource devices and it provides a more extensible media player architecture. It also provides an implementation that includes all of the media codecs as part of the open client infrastructure.The Ambulant target community is not viewers of media content, but developers of multimedia infrastructures, protocols and networks. Our goal has been to augument the existing partial SMIL implementations produced by many groups with a complete implementation that supports even the exotic features of the SMIL language. The following sections provide an introduction to the architecture of the player and describe the state of the various Ambulant implementations. We then discuss how the Ambulant Core can be re-purposed in other projects. We start with a discussion of Ambulant 's functional support for SMIL.2.FUNCTIONAL SUPPORT FOR SMIL 2.0The SMIL 2.0 recommendation [1] defines 10 functional groups that are used to structure the standard '5s0+ modules. These modules define the approximately 30 XML elements and 150 attributes that make up the SMIL 2.0 language. In addition to defining modules, the SMIL 2.0 specification also defines a number of SMIL profiles: collection of elements, attributes and attribute values that are targeted to meet the needs of a particular implementation community. Common profiles include the full SMIL 2.0 Language, SMIL Basic, 3GPP SMIL,XHTML+SMIL and SMIL 1.0 profiles.A review of these profiles is beyond the scope of this paper(see [2]), but a key concern of Ambulant ' sdevelopment has been to provide a player core that can be used to support a wide range of SMIL target profiles with custom player components.This has resulted in an architecture that allows nearly all aspects of the player to be plug-replaceable via open interfaces. In this way, tailored layout, scheduling, media processing and interaction modules can be configured to meet the needs of individual profile requirements. The Ambulant player is the only player that supports this architecture.The Ambulant player provides a direct implementation of the SMIL 2.0 Language profile, plus extensions that provide enhanced support for animation and timing control. Compared with other commercial and non-commercial players, the Ambulant player implements not only a core scheduling engine, it also provides complete support for SMIL layout,interaction, content control and networking facilities.Ambulant provides the most complete implementation of the SMIL language available to date.3.AMBULANT ARCHITECTUREThis section provides an overview of the architecture of the Ambulant core. While this discussion is high-level, it will provide sufficient detail to demonstrate the applicability of Ambulant to a wide range of projects. The sections below consider thehigh-level interface structure, the common services layer and the player com mon core architecture.3.1The High-Level Interface StructureFigure 1 shows the highest level player abstract ion. The player core support top-level con trol exter nal entry points (in clud ing play/stop/pause) and in turn man ages a collection of external factories that provide in terfaces to data sources (both for sta ndard and pseudo-media), GUI and window system interfaces and in terfaces to ren derers. Unlike other players that treat SMIL as a datatype [4],[10], the Ambula nt en gi ne has acen tral role in in teractio n with the input/output/scree n/devices in terfaces.This architecture allows the types of entry points (and the moment of evaluation) to be customized and separated from the various data-sources and renderers. This is important forintegration with environments that may use non-SMIL layout or special device in terface process ing.Figuit 1 k Ambulaittliigk-ljtwLstruchm.3.2The Common Services LayerFigure 2 shows a set of com mon services that are supplied for the player to operate. These in clude operati ng systems in terfaces, draw ing systems in terfaces and support for baseli ne XML fun ctio ns.All of these services are provided by Ambulant; they may also be integrated into other player-related projects or they may be replaced by new service components that are optimized for particular devices or algorithms. Hsurt 2. Amldant Common [Services Liwr/3.3The Player Common CoreFigure 3 shows a slightly abstracted view ofthe Ambula nt com mon core architecture. The view is essentially that of a single instanceof the Ambula nt player. Although only oneclass object is shown for eachservice,multiple interchangeable implementations have been developed for all objects (except the DOM tree) during theplayer 'development. As an example,multiple schedulers have bee n developed to match the fun cti onalcapabilities of various SMIL profiles.Arrows in the figure denote that one abstract class depends on the services offered by the other abstract class. Stacked boxes denote that a si ngle in sta nce of the player will con tain in sta nces of multiple con crete classes impleme nting that abstract class: one for audio, one for images, etc. All of the stacked-box abstract classes come with a factory function to create the in sta nces of the required con crete class.The bulk of the player implementation is architected to be platform in depe ndent. As we will discuss, this platform in depe ndent component has already been reused for five separate player impleme ntati ons. The platform dependent portions of the player include support for actual ren deri ng, UI in teract ion and datasource processing and control. When the player is active, there is asingle instanee of the scheduler and layout manager, both of which depend on the DOM tree object. Multiple instances of data source and playable objects are created. These in teract with multiple abstract rendering surfaces. The playable abstract class is the scheduler in terface (play, stop) for a media no de, while the renderer abstract class is the drawing in terface (redraw). Note that not all playables are ren derers (audio, SMIL ani mati on). The architecture has bee n desig ned to have all comp onents be replaceable, both in terms of an alter native impleme ntati on of a give n set of functionality and in terms of a complete re-purposing of the player components. In this way, the Ambulant core can be migrated to being a special purpose SMIL engine or a non-SMIL engine (such as support for MPEG-4 or other sta ndards).The abstract in terfaces provided by the player do not require a “ SMIL on Top” model of docume nt process ing. The abstract in terface can be used with other high-level control 4.1 Implementation PlatformsSMIL profiles have been defined for a widerange of platforms and devices, ranging fromdesktop implementations to mobile devices. Inorder to support our research on distributedmodels (such as in an XHTML+SMIL implementation), or to control non-SMILlower-level rendering (such as timed text).Note that in order to improve readability of theillustrati on, all auxiliary classes (threadi ng, geometry and color han dli ng, etc.) and several classes that were not important for general un dersta nding (player driver engine, transitions, etc.) have been left out of the diagram.4. IMPLEMENTATION EXPERIENCESThis sectio nwill briefly review ourimpleme ntatio n experie nces with theAmbula nt player. We discuss the implementation platforms used during SMIL ' s development and describe a set of test documents that were created to test the fun cti on ality of the Ambula nt player core. We con clude with a discussi on on the performa nee of the Ambula nt player.SMIL document extensions and to provide a player that was useful for other research efforts, we decided to provide a wide range of SMIL implementations for the Ambulant project. The Ambulant core is available as a single C++ source distribution that provides support for the following platforms:•Linux: our source distributi on in elude makefiles that are used with the RH-8 distribution of Linux. We provide support for media using the FF-MPEG suite [11]. The player interface is built using the Qt toolkit [12]. •Macintosh:Ambulant supports Mac OS X 10.3. Media rendering support is available via the internal Quicktime API and via FF-MPEG . The player user interface uses standard Mac conventions and support (Coca). •Windows: Ambulant provides conventional Win32 support for current generation Windows platforms. It has been most extensivelytested with XP (Home,Professional and TabletPC) and Windows-2000. Media rendering include third-party and local support for imaging and continuous media. Networking and user interface support are provided using platform-embeddedlibraries.•PocketPC: Ambulant supports PocketPC-2000,PocketPC-2002andWindows Mobile 2003 systems. The PocketPC implementations provide support for basic imaging, audio and text facilities.•Linux PDA support:Ambulant provides support for the Zaurus Linux-PDA. Media support is provided via the FF-MPEG library and UI support is provide via Qt. Media support includes audio, images and simple text.In each of these implementations, our initial focus has been on providing support for SMIL scheduling and control functions.We have not optimized media renderer support in the Ambulant 1.0 releases, but expect to provide enhanced support in future versions. 4.2 Demos and Test SuitesIn order to validate the Ambulant player implementation beyond that available with the standard SMIL test suite [3], several demo and test documents have been distributed with the player core. The principal demos include: •Welcome: A short presentation that exercises basic timing,media rendering, transformations and animation.•NYC: a short slideshow in desktop and mobile configurations that exercises scheduling, transformation and media rendering.•News: a complex interactive news document that tests linking, event-based activation, advanced layout, timing and media integration. Like NYC, this demo support differentiated mobile and desktop configurations.•Links: a suite of linking and interaction test cases.•Flashlight: an interactive user'sguide that tests presentation customization using custom test attributes and linking/interaction support. These and other demos are distributed as part of the Ambulant player web site [13].4.3Performance EvaluationThe goal of the Ambulant implementation was to provide a complete and fast SMIL player. We used a C++ implementation core instead of Java or Python because our experience had shownthat on small devices (which we feel hold significant interest for future research), the efficiency of the implementation still plays a dominant role. Our goal was to be able to read, parse, model and schedule a 300-node news presentation in less than two seconds on desktop and mobile platforms. This goal was achieved for all of the target platforms used in the player project. By comparison, the same presentation on the Oratrix GRiNS PocketPC player took 28 seconds to read, parse and schedule. (The Real PocketPC SMIL player and the PocketSMIL players were not able to parseand schedule the document at all because of their limited SMIL language support.)In terms of SMIL language performance, our goal was to provide a complete implementation of the SMIL 2.0 Language profile[14]. Where other players have implemented subsets of this profile,Ambulant has managed to implement the entire SMIL 2.0 feature set with two exceptions: first, we currently do not support the prefetch elements of the content control modules; second, we provide only single top-level window support in the platform-dependent player interfaces. Prefetch was not supported because of the close association of an implementation with a given streaming architecture. The use of multiple top-level windows, while supported in our other SMIL implementation, was not included in version 1.0 of Ambulant because of pending working on multi-screen mobile devices. Both of these feature are expected to be supported in the next release of Ambulant.5.CURRENT STATUS AND AVAILABILITYT his paper describes version 1.0 of the Ambulant player, which was released on July 12, 2004. (This version is also known as the Ambulant/O release of the player.) Feature releases and platform tuning are expected to occur in the summer of 2004. The current release of Ambulant is always available via our SourceForge links [13], along with pointers to the most recent demonstrators and test suites.The W3C started its SMIL 2.1 standardization in May, 2004.At the same time, the W3C' s timed text working group is completing itsfirst public working draft. We will support both of these activities in upcoming Ambulant releases.6.CONCLUSIONSWhile SMIL support is becoming ubiquitous (in no small part due to its acceptance within the mobile community), the availability of open-source SMIL players has been limited. This has meant that any group wishing to investigate multimedia extensions or high-/low-level user or rendering support has had to make a considerable investment in developing a core SMIL engine.We expect that by providing a high-performance, high-quality and complete SMIL implementation in an open environment, both our own research and the research agendas of others can be served. By providing a flexible player framework, extensions from new user interfaces to new rendering engines or content control infrastructures can be easily supported.7.ACKNOWLEDGEMENTS This work was supported by the Stichting NLnet in Amsterdam.8.REFERENCES[1]W3C,SMIL Specification,/AudioVideo.[2]Bulterman,D.C.A and Rutledge, L.,SMIL 2.0:Interactive Multimedia for Weband Mobile Devices, Springer, 2004.[3]W3C,SMIL2.0 Standard Testsuite,/2001/SMIL20/testsuite/[4]RealNetworks,The RealPlayer 10,/[5]Microsoft,HTML+Time in InternetExplorer 6,/workshop/author/behaviors/time.asp[6]Oratrix, The GRiNS 2.0 SMIL Player./[7]INRIA,The PocketSMIL 2.0 Player,wam.inrialpes.fr/software/pocketsmil/. [8],X-SMILES: An Open XML-Browser for ExoticApplications./[9]3GPP Consortium,The Third-GenerationPartnership Project(3GPP)SMIL PSS-6Profile./ftp/Specs/archive/26_series/26.246/ 26246-003.zip[10]Helix Community,The Helix Player./.[11]FFMPEG ,FF-MPEG:A Complete Solution forRecording,Converting and Streaming Audioand Video./[12]Trolltech,Qtopia:The QT Palmtop/[13]Ambulant Project,The Ambulant 1.0 Open Source SMIL 2.0Player, /.[14]Bulterman,D.C.A.,A Linking andInteraction Evaluation Test Set for SMIL,Proc. ACM Hypertext 2004, SantaCruz,August, 2004.。
mindlessormindfultechnology无意识或注意技术

Below is given annual work summary, do not need friends can download after editor deleted Welcome to visit againXXXX annual work summaryDear every leader, colleagues:Look back end of XXXX, XXXX years of work, have the joy of success in your work, have a collaboration with colleagues, working hard, also have disappointed when encountered difficulties and setbacks. Imperceptible in tense and orderly to be over a year, a year, under the loving care and guidance of the leadership of the company, under the support and help of colleagues, through their own efforts, various aspects have made certain progress, better to complete the job. For better work, sum up experience and lessons, will now work a brief summary.To continuously strengthen learning, improve their comprehensive quality. With good comprehensive quality is the precondition of completes the labor of duty and conditions. A year always put learning in the important position, trying to improve their comprehensive quality. Continuous learning professional skills, learn from surrounding colleagues with rich work experience, equip themselves with knowledge, the expanded aspect of knowledge, efforts to improve their comprehensive quality.The second Do best, strictly perform their responsibilities. Set up the company, to maximize the customer to the satisfaction of the company's products, do a good job in technical services and product promotion to the company. And collected on the properties of the products of the company, in order to make improvement in time, make the products better meet the using demand of the scene.Three to learn to be good at communication, coordinating assistance. On‐site technical service personnel should not only have strong professional technology, should also have good communication ability, a lot of a product due to improper operation to appear problem, but often not customers reflect the quality of no, so this time we need to find out the crux, and customer communication, standardized operation, to avoid customer's mistrust of the products and even the damage of the company's image. Some experiences in the past work, mentality is very important in the work, work to have passion, keep the smile of sunshine, can close the distance between people, easy to communicate with the customer. Do better in the daily work to communicate with customers and achieve customer satisfaction, excellent technical service every time, on behalf of the customer on our products much a understanding and trust.Fourth, we need to continue to learn professional knowledge, do practical grasp skilled operation. Over the past year, through continuous learning and fumble, studied the gas generation, collection and methods, gradually familiar with and master the company introduced the working principle, operation method of gas machine. With the help of the department leaders and colleagues, familiar with and master the launch of the division principle, debugging method of the control system, and to wuhan Chen Guchong garbage power plant of gas machine control system transformation, learn to debug, accumulated some experience. All in all, over the past year, did some work, have also made some achievements, but the results can only represent the past, there are some problems to work, can't meet the higher requirements. In the future work, I must develop the oneself advantage, lack of correct, foster strengths and circumvent weaknesses, for greater achievements. Looking forward to XXXX years of work, I'll be more efforts, constant progress in their jobs, make greater achievements. Every year I have progress, the growth of believe will get greater returns, I will my biggest contribution to the development of the company, believe in yourself do better next year!I wish you all work study progress in the year to come.Mindless or Mindful Technology?Yvonne RogersUCLIC, University College London, London, UKAbstractWe are increasingly living in our digital bubbles. Even when physically together – as families and friends in our living rooms, outdoors and public places – we have our eyes glued to our own phones, tablets and laptops. The new generation of ‘all about me’ health and fitness gadgets, that is becoming more mainstream, is making it worse. Do we really need smart shoes that tell us when we are being lazy and glasses that tell us what we can and cannot eat? Is this what we want from technology – ever more forms of digital narcissism, virtual nagging and data addiction? In contrast, I argue for a radical rethink of our relationship with future digital technologies. One that inspires us, through shared devices, tools and data, to be more creative, playful and thoughtful of each other and our surrounding environments.Categories and Subject DescriptorsH.5.2 User InterfacesKeywordsMindful technology; vision; creative technology, HCIShort BioYvonne Rogers is a Professorof Interaction Design, thedirector of UCLIC and adeputy head of the ComputerScience department at UCL.Her research interests are in theareas of ubiquitous computing,interaction design and human-computer interaction. A centraltheme is how to designinteractive technologies thatcan enhance life by augmenting and extending everyday, learning and work activities. This involves informing, building and evaluating novel user experiences through creating and assembling a diversity of pervasive technologies. Yvonne is also the PI at UCL for the Intel Collaborative Research Institute on Sustainable Connected Cities (ICRI Cities) which was launched in October 2012 as a joint collaboration with Imperial College.She was awarded a prestigious EPSRC dream fellowship rethinking the relationship between ageing, computing and creativity. She is a visiting professor at the Open University and Indiana University. She has spent sabbaticals at Stanford, Apple, Queensland University, University of Cape Town, University of Melbourne, QUT and UC San Diego. Central to her work is a critical stance towards how visions, theories and frameworks shape the fields of HCI, cognitive science and Ubicomp. She has been instrumental in promulgating new theories (e.g., external cognition), alternative methodologies (e.g., in the wild studies) and far-reaching research agendas (e.g., “Being Human: HCI in 2020” manifesto). She has also published a monograph (2012) called “HCI Theory: Classical, Modern and Contemporary.”From 2006-2011, Yvonne was professor of HCI in the Computing Department at the OU, where she set up the Pervasive Interaction Lab. From 2003-2006, she was a professor in Informatics at Indiana University. Prior to this, she spent 11 years at the former School of Cognitive and Computing Sciences at Sussex University.Yvonne was one of the principal investigators on the UK Equator Project (2000-2007) where she pioneered ubiquitous learning. She has published widely, beginning with her PhD work on graphical interfaces to her recent work on public visualisations and behavioural change. She is one of the authors of the definitive textbook on Interaction Design and HCI now in its 3rd edition that has sold over 150,000 copies worldwide and has been translated into 6 languages. She is a Fellow of the British Computer Society and the ACM's CHI Academy: “an honorary group of individuals who have made substantial contributions to the field of human-computer interaction. These are the principal leaders of the field, whose efforts have shaped the disciplines and/or industry, and led the research and/or innovation in human-computer interaction.”Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s). Copyright is held by the author/owner(s).EICS’14,June 17–20, 2014, Rome, Italy.ACM 978-1-4503-2725-1/14/06./10.1145/2607023.2611428。
Information Management

DIRECT-INFOA Media Monitoring System for Sponsorship TrackingG. Kienast, H. Rehatschek,A. HortiJOANNEUM RESEARCH Institute of Information Systems & Information Management Steyrergasse 17, A-8010 Graz+43(316)876-1182 gert.kienast@joanneum.atS. Busemann, T. DeclerckDFKI GmbHLanguage Technology LabStuhlsatzenhausweg 3D-66123 Saarbrücken+49(681)302-5282stephan.busemann@dfki.deV. Hahn, R. CavetFraunhofer Institut für GraphischeDatenverarbeitungFraunhoferstraße 5D-64283 Darmstadt+49(6151)155-612volker.hahn@igd.fhg.deABSTRACTDIRECT-INFO is a system for media monitoring applied to the field of sponsorship tracking. Significant parts of TV streams and electronic press feeds are automatically selected and subsequently monitored to find appearances of the name or logo of a sponsoring company in connection with the sponsored party. For this purpose basic features are fully automatically extracted from TV and press and thereafter fused to semantically meaningful reports to support executive decision makers. Extracted features include logos, positive & negative mentions of a brand or product, multimodal video segmentation, speech-to-text transcripts, detected topics and genre classification. In this paper we first describe the technical workflow and architecture of the DIRECT-INFO system and then present the main innovations in four key areas.Categories and Subject DescriptorsJ.0 [General Computer Applications]General TermsAlgorithms, Economics, Management, Measurement KeywordsMedia monitoring, sponsorship tracking, logo recognition, object detection, natural language processing, multimodal video segmentation, information fusion.1.INTRODUCTIONSport sponsors are spending significant amounts of money to be publicly associated with a team or an athlete. Knowledge about how often a sponsor is mentioned in connection with the sponsored party is a direct indicator for executive managers to estimate whether to continue sponsorship or not. The sponsored party can use this information in order to further motivate the sponsor to invest. Such information is gathered through global advertisement expenditure measurement, which is performed by media monitoring companies.This type of business intelligence is a very complex task, which is currently performed manually in the sector of sponsorship tracking, therefore it is very expensive. These companies employ a huge staff to continuously monitor a set of media channels ranging from TV and radio to press and internet. Technology support on this work is currently mainly on the data management side after the information from the media monitored have been extracted by humans.The focus of the DIRECT-INFO project is to create an application for semi-automatic sponsorship tracking in the area of media monitoring which shall speed up the manual process. This means a sponsor wants to know how often his brand is mentioned in connection with the sponsored company. The simple detection of a brand in one modality (e.g. video) is not sufficient to meet the requirements of this business case. Multimodal analysis and fusion – as implemented within DIRECT-INFO – is needed in order to fulfill these requirements.1.1Benefits for service providers & end users Reducing labor-intensive monitoring work of TV & newspapers by providing a semi-automatic system can greatly reduce costs for media information companies and their customers. Human intervention is mainly limited to verification and validation of results. Both end users and service providers benefit from having the desired information available more quickly.Information extracted from heterogeneous media sources (TV, newspaper, internet…) can be accessed through a common interface. Examples for possible end user requests are:•The overall time a company logo was visible in a given broadcast•Whether a company (their representatives, sponsors or products) appeared in positive or negative contexts in agiven period in the media.In addition to sponsorship tracking also other application scenarios are possible, e.g. the monitoring of political election campaigns. Questions arising may include: How much coverage did a party or a politician get in the news or talk shows? Was s/hementioned in a positive or negative way? What about competitive candidates?2.RELATED WORKCurrently there exist only semi-automatic commercial solutions and ongoing research projects for dedicated business cases in the media monitoring domain. According to our knowledge the only multimodal analysis system which targets sponsorship tracking has been developed by the R&D project DIRECT-INFO. For the very specific (unimodal) task of brand detection the R&D project DETECT and commercial products from SPORTSi, Spikenet Technology, and Omniperception can be identified. Furthermore it has to be mentioned that there are many (manual) service providers in the market, however, very few technological solution providers.DETECT is a completed EC project which aimed to automatically extract features from digital video streams. In particular logo appearances were detected and based on these appearances statistics for the sponsoring party were generated [5].SPORTSi™ is a commercial software marketed by tnssport [6]. It tracks television exposure and value of brands within sport events. The software can measure exposure from advertising boards. The SPORTSi™ system is being developed with BrandTraX Inc., based on software developed by Bell Labs, the R&D department of Lucent Technologies. tnssport offers media monitoring as a service for the business cases: coverage analysis, audience data interpretation and quantity/quality of sponsor exposure. Recently Margaux Matrix [9] began marketing a system developed by Spikenet Technology [10]. The initial target of this application is to track and to generate statistics for logos within Formula 1 race broadcasts. An initial evaluation of the demo system seems to indicate that the method is accurate, but extremely sensitive to a variety of threshold values. They claim a detection accuracy of 97.3% with a false positive rate of just 0.1. Careful parameter adjusting needs to be carried out for each logo class and broadcast to get these optimal results.Magellan™ [11] Automatic Brand Analysis and Logo Detection software provides the tools necessary to automate the chronometry process for Sports programs. Magellan™ enables significant improvements in manual workflow throughput. The Automatic Brand Analysis and Logo Detection are based upon purpose-built proprietary image recognition algorithms. Magellan™ searches through frames of video to automatically detect the exposure of a brand image or a logo. The Magellan system by OmniPerception claims to be a robust solution to logo tracking, which offers real-time multiple brand/logo detection in video streams. It handles occlusion, overlap, rotation, and can localize the logo within the frame. It is currently not possible to tell any details of how the algorithm works or how accurate it truly is.3.THE DIRECT-INFO SYSTEMThis section gives a short introduction on the main innovation of DIRECT-INFO (achieved both on the level of basic analysis as well as in fusion of heterogeneous low-level information to high-level semantic results) and then describes the components and their roles in the application. 3.1Main Innovations3.1.1Logo DetectionAn algorithm for logo recognition based on receptive fields [6], [7] has been developed by the DETECT project. This approach for logo recognition has been shown to be invariant against affine transformations, perspective transformations, occlusions and motion blur. Matching is performed by histogram intersection, which compares the histogram of computed receptive fields in a detected region of interest within a frame versus the histograms loaded from the model database.Although the receptive fields approach is very fast the number of false positive detections is rather high depending on the logo. Furthermore the receptive fields approach is unable to recognize multiple adjacent logo occurrences as such. Adjacent occurrences are merged to one object recognition.To overcome these drawbacks the “Scale Invariant Feature Transform” (SIFT) [1], [3] is now used and has been enhanced for logo recognition in DIRECT-INFO. SIFT is able to recognize adjacent logos as independent occurrences and application of this algorithm results in a higher precision rate. Additionally the SIFT algorithm has been extended to improve object recognition rates (see section 4 for details). The first improvement handles recognition of an object if just one interest point has matched. In this case the surrounding of the candidate object location is investigated in detail. The second improvement applies the same strategy in the temporal domain: If an object is recognized with less confidence in the next frame at the same location, the SIFT implementation examines the candidate region in more detail. 3.1.2Multimodal SegmentationThe aim of the multimodal segmentation is the unsupervised extraction of meaningful semantic information from broadcasted video taking advantage of the media's multimodality. Since semantic is not independent of context, the goal is to detect and extract logical entities (scenes) from the video stream on top of results coming from the classification of basic event sequences.In contrast to most solutions for video analysis, which are still focusing on one modality, the multi modal scene classification approach is based on the analysis of different kinds of information channels:•Visual modality including artificial (graphics) and natural content (video)•Audio modality, which includes environmental sounds as well as music, jingles etc.•Text modality: text overlays, which provide semantic information.The usage of multimodal analysis in video raises the question about what should be analyzed in the video stream and how could it be done. Regarding human beings the process of perception is a pre-conscious level of cognition (“signal level”); it organizes the incoming sensoric signals into information instances such as objects and events. This perceptual organization is then taken over by higher cognitive levels in order to be enriched by knowledge, so that we can become aware of what is present in the world around us. Because object recognition is still a hard task, event detection and modeling is the more promising way towardsautomatic semantic annotation and description of multimodal broadcasts [14].3.1.3Text AnalysisIn order to detect positive or negative mentions of brand names in the news, DIRECT-INFO applies linguistics-based text analysis tools. Going that far beyond keyword matching is new and highly promising, given a solid coverage of the languages in question (English and Italian). Text is available from different media: •Speech input transformed to corresponding written text•Newspaper transformed from PDF to plain text•Text from on-line tickers reporting on a sports eventAll sources are unrestricted and error-prone. Both these facts call for fallback strategies to more shallow analysis techniques. Clearly the performance depends on the correct recognition of the key brand names. The alternative use of statistical approaches is hardly possible, as annotated data supporting the classification process are not available and expensive to provide. In chapter 5 we concentrate on the linguistic analysis of newspaper text.3.1.4Information FusionAfter basic analysis is finished all results are stored in a common MPEG-7 document, which can be accessed by all components through the MPEG-7 document server.Heterogeneous multi-media information in MPEG-7 does usually not directly answer customer needs (section 1.1). However, automatically relating pieces of information in appropriate ways can indeed answer such needs. A novel type of human-machine interface has been designed and implemented to support the media analyst in interpreting fused Appearances of company information and making this information available to the end user. Starting from basic MPEG-7 data, rule-based fusion operators can be used, modified and parameterized by the media analyst in order to create the complex appearances the end user is interested in. Using a comfortable use case management system combined with extensive search and retrieval functionality, the media analyst can view, interpret and publish the fused appearances for presentation to the end user. The system operates semi-automatically – the media analyst finally decides which appearances are relevant and will be published for a particular end user.The MPEG-7 parser, the database, the fusion component and the facts management interface to the media analyst, which we also call “set-up application”, are implemented in Python using the open source Zope/Plone framework[14], [22].3.2Pilot Use CaseFor the use case targeted in the evaluation phase of the project we are interested in the Juventus soccer team and its sponsors Nike, Tamoil and Sky Sports. Appearances and mentions of the sponsors in connection with Juventus will be monitored.The material to be monitored includes TV recordings from the Italian TV channels Sky Sports and Italia 1of December 17-19, 2004 and Italian daily newspapers Gazzetta dello Sport and Corriere della Sera of December 18-20, 2004 covering a soccer match of Juventus against Inter Milan including pre- and post-game coverage. 3.3WorkflowIn order to meet the requirements of sponsorship tracking the following workflow has been identified. This workflow can be easily configured by the user and adapted to other usage scenarios.1.The Acquisition Component records video chunks ofconstant length (a few minutes) & EPG information and notifies the central content analysis controller (CAC) on their availability. Based on EPG information the Content Essence Repository (CER) and the Content Analysis Controller (CAC) prepare “Semantic Blocks”, each represented by an MPEG-2 essence file and an MPEG-7 document for metadata. A semantic block is a closed TV broadcast, e.g. a soccer game or a talk show. A semantic block may be assembled from one or more recorded video chunks.2.For each semantic block the CAC starts an analysissubsystem that performs automatic genre classification on this semantic block in order to get another indicator – next to the EPG information – if the current semantic block is to be considered “relevant” for the given use case. Based on a condensed result of the genre classification and the EPG information the CAC decides whether the semantic block is analyzed or not, e.g. a soccer game is relevant for sponsorship tracking while a soap opera is not.3.If the semantic block is relevant for analysis, the CAC passesit to the further analysis subsystems according to the user defined workflow. For sponsorship tracking these subsystems perform logo detection, video segmentation, text analysis and topic detection based on a speech-to-text transcript. All analysis results are stored in the MPEG-7 document of the semantic block.4.After analysis the user performs a “quality check” via theMPEG-7 Result Editor. The user may directly modify and/or approve the results or restart analysis with changed parameters or start and end time of semantic blocks.5.After the quality check results are passed to the FusionComponent. It automatically reduces the complexity of the various low-level results by creating rule-based “Appearances” of a sponsor, team or brand. Based on user interaction appearances may be classified and edited manually. The fused results are stored in a local database. 6.When a specific customer request comes in the user queriesthe database and prepares the data relevant for the specific customer.7.The customer accesses the data via a web-based interfaceproviding access to the relevant appearances of his brand, as well as to statistics.In parallel there is a simplified workflow for processing PDF editions of newspapers. Via a (web-) subscription interface PDF files are downloaded automatically once per day. Semantic analysis is restricted to text analysis and logo detection in images.3.4System ArchitectureThe system has been designed in a modular way to be easily reconfigurable. Between the individual components web service interfaces have been defined and implemented. Figure 1 shows theindividual components as well as the data- and control-flowbetween them.Figure 1. DIRECT-INFO system architecture.The Acquisition & Content Essence Repository (CER) captures TV streams, Electronic Programming Guide (EPG) information and downloads digital newspaper subscriptions from the internet. The analysis subsystems access essence data via a shared network folder.The Content Analysis Controller (CAC) is the central control logic in the DIRECT-INFO system which manages the analysis workflow. It is notified by the CER about newly available semantic blocks and distributes the analysis jobs according to a pre-defined workflow. After analysis the results can be verified via the integrated MPEG-7 Result Editor.A set of Analysis Subsystems is responsible for the extraction of low- and mid-level features from the content. The subsystems expose a web service interface towards the CAC to accept control commands as to start or stop a job. One job is always performed on one entire semantic block. Status (progress, success, error) are reported back to the CAC, while the results are stored in the MPEG-7 document server. The analysis subsystems include Genre Classification , Logo & Object Detection , Speech-to-Text Engine including a Topic Detection , Text Analysis , Multimodal Segmentation and Video Summary Generation .The MPEG-7 Document Server is the central repository for all metadata generated within the DIRECT-INFO system. It handles concurrent access to the MPEG-7 documents by the subsystems. The CAC creates an MPEG-7 document per semantic block which is then filled up with the results provided by the analysis subsystems.The Fusion Component analyzes the MPEG-7 document to identify the appearances of brand names and computes complex appearances involving results of multiple subsystems. The fusion rules can be parameterized. The results are verified and published for end user delivery. These activities are carried out by the media analyst using an interface we call the Set-up Application . The set-up application also includes the management of user data and user requests.The Delivery System is the interface for final users including an easy access to all information that is relevant for them at different levels of detail. It presents the analysis results as bulletin showing graphical illustrations of appearances including graphs, trend charts, etc. for different time periods. Furthermore it gives users the ability to browse single appearances providing related information about time and volume of appearance, positive/negative classification, keywords describing the content of the video and videos representing the appearance itself. Additionally, the delivery system provides an alert system to inform customers immediately via SMS, MMS or email in case that an important event has been detected.The analysis workflow can easily be customized to fit other use cases by adding or replacing the analysis subsystems. The analysis workflow (which subsystems run in which order) and the analysis parameters (e.g. logos to search for) are configured via the CAC.4. OBJECT AND LOGO DETECTIONThe task of logo detection is closely related to detecting known planar objects in still and moving images, with some special requirements. Since logos vary in size , can be rotated and are subject to different lightning conditions the algorithm shall be invariant against all these factors. Logos may be partly occluded or on non-rigid surfaces (a player’s shirt) so a logo shall still be detected, even if only parts of it are visible/matched. Logos shall be found in still images, nevertheless additional information gained from an image sequence may be used to improve its results. Some logos (e.g. the Nike “checkmark”) are only defined by their shape and may appear in any color. Hence the algorithm must not rely only on color features. Ideally the algorithm should be configurable to be able to trade off quality vs. speed.Based on performing a (paper-based) evaluation, using the results from [4] and some practical tests the SIFT algorithm (Scale Invariant Feature Transform) by David Lowe [1], [2], [3] waschosen to be used as it meets these requirements best.Figure 2. Logo template (top), matched key points (white lines)and detected logo appearances (white boxes).During the DIRECT-INFO project the SIFT algorithm has been enhanced. SIFT key points are tracked to circumvent variations of logo recognition hits between frames. If an object has been recognized with high confidence in one frame, the surrounding of the estimated object position in the following frame is primarily investigated. If the object is recognized with less confidence (fewer hits) in the next frame, it can be assumed to be the same object occurrence. A two-pass approach has been introduced by splitting the object recognition in a “rough” and “fine” step. If only one single descriptor hit has been detected in the rough step, the fine step looks with more sensitive parameters in the surrounding of the hit in order to detect more matching descriptors supporting or rejecting the thesis of an object appearance.Figure 2 shows a typical result of the logo detection performed on one video frame. As not all appearances of logos are found in every single frame, intraframe information is used to bridge those gaps to be able to track logo appearance over their entire visibility time. Problems arise due to interlaced TV content, motion blurring and the small size of logos. The size of the logo appearances in Figure 2 are on the lower limit of what currently can be detected.However, the SIFT algorithm is not invariant against strong perspective distortions of logos. However, this can be overcome by learning synthetically distorted logo models, which are created automatically during the learning step.Several performance improvements have been made allowing processing of five frames per second using half PAL resolution (360x288 pixels) on a standard (single CPU) PC. If necessary real-time can be achieved by only processing every n th frame (depending on quality requirements) and/or using multi-processor hardware. In our application real-time is not a necessity, since the DIRECT-INFO system performs a pre-filtering step based on the relevancy of a semantic block, reducing the daily feed of 24 hours of material per channel to about six hours to be fed into analysis.5.LINGUISTIC TEXT ANALYSISIn this section we describe a set of subsequent processing steps from shallow to deeper linguistic levels, as well as tools to support this analysis. Thanks to pre-existing large-scale linguistic resources this new approach to opinion extraction yields promising results on free newspaper text.As the newspaper articles are available in PDF format, some additional processes are needed to access the textual source. A tool is currently being implemented that reconstructs the logical units of the PDF document by heuristically interpreting fonts, positions and keywords. In the current system, we just extract the text from the HTML document, accepting any errors this may involve.The textual elements are then morpho-syntactically processed, so that they can be associated with their part of speech (POS) and morphological properties (like number, gender). This is achieved by combining a huge lexical resource taken from the multi-lingual indexing software IDX with rule-based POS disambiguation tools. We decided to use IDX – a further development of the system described in [13] – instead of a classical morphological analyzer since IDX offers the possibility to encode additional information with the tokens, such as translations in various languages and synonyms. For instance, the synonym set for the name of the soccer team “Juventus Turin” also includes “Juve” and “bianconeri”, which allows us to relate these surface forms to the said team. In an application that detects opinions about an entity, this feature is for sure crucial, since it allows covering all known synonyms of the official naming of the entity.The information delivered by IDX is refined by a POS disambigu-ation tool. A word like “comando” may be either a noun or a verb. Which one is correct in a given context can be told quite reliably by investigating the POS of the left and of the right neighbor.All morpho-syntactic information is then passed to the syntactic SCHUG analyzer (Shallow and CHunk-based Unification Grammar, [12]) that provides for a dependency analysis of the textual input up to the detection of grammatical functions (e.g. subject, direct object). The following SCHUG sample analysis shows the annotation provided for the sentential level, not going into the details of the linguistic analysis of the individual noun phrases and prepositional phrases (NP, PP).<LING_INFO BOS="0" EOS="17" STRING="Per ora comanda la Juventus di Fabio Capello, che affronta con quattro punti di vantaggio sul Milan la volata per il titolo inverno."><CLAUSE id="1" BOC="0" EOC="2" MARKER="C"POLARITY="positive"><PP_ADJUNCT FRAG="0">Per ora</PP_ADJUNCT> <PRED">comanda</PRED><SUBJ FRAG="2">la Juventus di FabioCapello</SUBJ></CLAUSE>...The sentence analyzed (under STRING) has been segmented into “clauses”, the first one being “Per ora comanda la Juventus di Fabio Capello”. A clause corresponds more or less to a semantic unit of the whole sentence, where we expect one verbal component to be present. In this first clause, SCHUG has identified three main components: the predicate (PRED, “comanda”), the subject (SUBJ, “la Juventus di Fabio Capello”) and a modifying prepositional phrase (PP_ADJUNCT).The POLARITY tag included in the annotation above is purely syntactic information: it tells us if a negative particle word (“no”, “none”, “never”, etc.) has been used or not. In our example this is not the case. This information supports the calculus on positive/negative mention described below.From manual annotation work on newspaper articles on soccer a list of expressions having typically a positive or negative connotation has emerged. These expressions are listed in a special lexicon, modifiable by the media analyst, which is consulted by SCHUG while processing the document:comando => {POS => Noun, INT => "positive"} dominare => {POS => Verb, INT => "positive"} stanco => {POS => Adj, INT => "negative"} For instance, all forms of the noun stem “comando” have a positive connotation (in the soccer domain).In our example we find the verb “comandare”, which is encoded in our lexicon, just like the noun “comando”. We can now calculate the assessment expressed in the sentential clause on the Juventus team: it is positive, since the POLARITY tag is positivefor the whole clause and the subject of the clause has no negative expression in it.The heuristics we use for detecting the positive/negative mentions are encoded as a two-level rule system based on the results of the different stages of linguistic analysis:1. Calculate positive/negative at the level of linguisticfragments, e.g. within noun phrases (NPs), on the base of the dependency structure. For example:If mention(MOD) = positive & mention(Head_Noun) = negative => mention(NP) = negative 2. Then calculate positive/negative at the sentential levels,based on the mention properties of the grammatical relations. For example:If mention(SUBJ) = positive & mention(PRED) = negative & no_other_arguments_present=> mention(Sentence) = negative The second heuristic would apply to a sentence like “The team of Fabio Capello lost the game in the last five minutes”.6. MULTIMODAL SEGMENTATIONThe multimodal analysis approach aims to achieve a segmentation of the broadcasted video stream into logical units, which are annotated with semantic information derived by the classification of the visual and audio context (scene modeling) and with theinformation derived through an OCR engine.Figure 3. Workflow of the multimodal segmentation process. For the description and identification of scenes, the detection of several basic events belonging to the different modalities is necessary:•Transition edits in the visual modality as hard cuts or gradual transitions, which results in the segmentation of the video stream in coherent shots.• Overlayed text events represented by uninterrupted textual expressions.•Cuts in the auditory layout, representing changes in the sound signal. This includes transitions from silence to music, speaker changes, classification of audio segments into speech, music and noise.Starting from basic events in video corresponding to the pure perceptual level as shots, noise, music, text-overlays, etc., the multimodal scene classification approach aims on theidentification of logical units of coherent content. The segmented units are annotated using predefined semantic attributes that are derived automatically from the underlying event model describing the context of the extracted scene (e.g. attribute "named face" through combination of a detected face and a textual overlay within the same region). For news broadcasts the system intends to differentiate e.g. complete news stories, anchorman, interviews, trailers and advertisements as for sport broadcasts especially football it is envisaged to differentiate trailers, background reporting, interviews, highlights and the game itself. The multimodal event analysis additionally is accompanied with the results of the OCR engine that recognizes the results of the basic text overlay detection and passes it via the MPEG-7 document server to the text analysis module.In the following sections the different entities of the multimodal analysis approach will be described.6.1 Basic Event DetectionFor multimodal analysis the main goal of the Shot Detection is the identification of visual transitions in a continuous video stream. The transitions form the syntactic structure of the visual modality separating the entire stream into single shots. Shots are generally considered as the elementary units constituting a video from an author’s or director’s point of view. Detecting shot boundaries thus means to recover those elementary video units, which in turn provide the basis, together with all other basic event detectors, for the scene and story segmentation. The shot detection approach used for the DIRECT-INFO system does not rely only on the automatic detection of hard cuts in the visual modality. It is furthermore capable to analyze and identify all kinds of visual transitions including cuts, fades, dissolves and wipes. It therefore consists of two independent parallel loops analyzing simultaneously hard cuts and special effects and consolidating the results of both for the shot segmentation.The Face Detection continuously provides the number of appearing frontal faces and their position in each frame within a color video stream. This information is very valuable for annotating videos with semantic information as the genre of the video (like newscast or commercial), which can in many cases be directly determined if the positions of the appearing faces are known. For example, a video containing a single, almost unmoving face could hint for a newscast with the detected face belonging to the anchorperson. The detection is achieved by combination of a fast initial detection module (skin color filtering) with a verification approach based on SVM classification. The output of the face detection is provided to the scene modeling module for further processing and classification issues.The Text Detection process consists of three different parts: The first part concentrates on the detection of edges in the video image representing the boundaries of overlaid characters. As with the edge based feature, only high contrast characters could be found. The text detection uses a second independent approach that is based on the segmentation of images into coherent regions. In a third part the results coming from both modules are combined for the final determination of text regions.After the detection and extraction of each individual text region, the resulting sub images are preprocessed in order to create a black on white image representation for further processing through an OCR engine. The outputs of the text detection module。
Categories and Subject Descriptors
Searching the Web Using Composed Pages Ramakrishna Varadarajan Vagelis Hristidis Tao LiFlorida International University{ramakrishna,vagelis,taoli}@Categories and Subject Descriptors:H.3.3 [Information Search and Retrieval]General Terms: Algorithms, Performance Keywords: Composed pages1.INTRODUCTIONGiven a user keyword query, current Web search engines return a list of pages ranked by their “goodness” with respect to the query. However, this technique misses results whose contents are distributed across multiple physical pages and are connected via hyperlinks and frames [3]. That is, it is often the case that no single page contains all query keywords. Li et al. [3] make a first step towards this problem by returning a tree of hyperlinked pages that collectively contain all query keywords. The limitation of this approach is that it operates at the page-level granularity, which ignores the specific context where the keywords are found within the pages. More importantly, it is cumbersome for the user to locate the most desirable tree of pages due to the amount of data in each page tree and a large number of page trees.We propose a technique called composed pages that given a keyword query, generates new pages containing all query keywords on-the-fly. We view a web page as a set of interconnected text fragments. The composed pages are generated by stitching together appropriate fragments from hyperlinked Web pages, and retain links to the original Web pages. To rank the composed pages we consider both the hyperlink structure of the original pages, as well as the associations between the fragments within each page. In addition, we propose heuristic algorithms to efficiently generate the top composed pages. Experiments are conducted to empirically evaluate the effectiveness of the proposed algorithms. In summary, our contributions are listed as follows: (i) we introduce composed pages to improve the quality of search; composed pages are designed in a way that they can be viewed as a regular page but also describe the structure of the original pages and have links back to them, (ii) we rank the composed pages based on both the hyperlink structure of the original pages, and the associations between the text fragments within each page, and (iii) we propose efficient heuristic algorithms to compute top composed pages using the uniformity factor. The effectiveness of these algorithms is shown and evaluated experimentally. 2.FRAMEWORKLet D={d1,d2,,…,d n} be a set of web pages d1,d2,,…,d n. Alsolet size(d i)be the length of d i in number of words. Termfrequency tf(d,w) of term (word) w in a web page d is thenumber of occurrences of w in d. Inverse documentfrequency idf(w)is the inverse of the number of web pagescontaining term w in them. The web graph G W(V W,E W) of aset of web pages d1,d2,,…,d n is defined as follows: A node v i∈V W, is created for each web page d i in D. An edge e(v i,v j)∈E W is added between nodes v i,v j∈V W if there is a hyperlink between v i and v j. Figure 1 shows a web graph. Thehyperlinks between pages are depicted in the web graph asedges. The nodes in the graph represent the web pages andinside those nodes, the text fragments, into which that webpage has been split up using html tag parsing, are displayed(see [5]).In contrast to previous works on web search [3,4], wego beyond the page granularity. To do so, we view each pageas a set of text fragments connected through semanticassociations. The page graph G D(V D,E D) of a web page d isdefined as follows: (a) d is split to a set of non-overlapping text fragments t(v), each corresponding to a node v∈V D.(b) An edge e(u,v)∈E D is added between nodes u,v∈V D if there is an association between t(u) and t(v) in d. Figure 2 shows the page graph for Page 1 of Figure 1. As denoted in Figure 1, page 1 is split into 7 text fragments and each one is represented by a node. An edge between two nodes denotes semantic associations. Higher weights denote greater association. In this work nodes and edges of the page graph are assigned weights using both query-dependent and independent factors (see [5]). The semantic association between the nodes is used to compute the edge weights (query-independent) while the relevance of a node to the query is used to define the node weight (query-dependent).A keyword query Q is a set of keywords Q={w1,…,w m}.A search result of a keyword query is a subtree of the webgraph, consisting of pages d1,…,d l, where a subtree s i of thepage graph G Di of d i is associated with each d i. A result is total−all query keywords are contained in the text fragments−and minimal−by removing any text fragment a query keyword is missed. For example, Table 1 shows the Top-3 search results for the query “Graduate Research Scholarships” on the Web graph of Figure 1.3.RANKING SEARCH RESULTSProblem 1 (Find Top-k Search Results).Given a webgraph G W, the page graphs G D for all pages in G W, and akeyword query Q, find the k search results R with maximumScore(R).The computation of Score(R) is based on the followingprinciples. First, search results R involving fewer pages areranked higher [3]. Second, the scores of the subtrees ofCopyright is held by the author/owner(s).SIGIR’06, August 6–11, 2006, Seattle, Washington, USA. ACM 1-59593-369-7/06/0008.Figure 2: A page graph of Page 1 of Figure 1.Table 1: Top-3 search results for query “Graduate ResearchScholarships”Search Resultsthe page graphs of the constituting pages of are combined usinga monotonic aggregate function to compute the score of the searchresult. A modification of the expanding search algorithm of [1] isused where a heuristic value combining the Information Retrieval(IR) score, the PageRank score [4], and the inverse of theuniformity factor (uf)of a page is used to determine the nextexpansion page. The uf is high for pages that focus on a single orfew topics and low for pages with many topics. The uf iscomputed using the edge weights of the page graph of a page(high average edge weights imply high uf). The intuition behindexpanding according tothe inverse uf is that among pages withsimilar IR scores, pages with low uf are more likely to contain ashort focused text fragment relevant to the query keywords.Figure 3 shows the quality of the results of our heuristic search vs.the quality of the results of the non-heuristic expanding search [1](a random page is chosen for expansion since hyperlinks are un-weighted) compared to the optimal exhaustive search. Themodified Spearman’s rho metric [2] is used to compare two Top-kFigure 3: Quality Experiments using Spearman’s rho.4.REFERENCES[1]G. Bhalotia, C. Nakhe, A. Hulgeri, S. Chakrabarti and S,Sudarshan:Keyword Searching and Browsing in Databases using BANKS.ICDE, 2002[2]Ronald Fagin, Ravi Kumar, and D. Sivakumar: Comparing top-klists. SODA, 2003[3]W.S. Li, K. S. Candan, Q. Vu and D. Agrawal: Retrieving andOrganizing Web Pages by "Information Unit". WWW, 2001[4]L. Page, S. Brin, R. Motwani, and T. Winograd: The pagerankcitation ranking: Bringing order to the web. Technical report,Stanford University, 1998[5]R. Varadarajan, V Hristidis: Structure-Based Query-SpecificDocument Summarization. CIKM, 2005。
“On the Internet, Nobody Knows You’re a Dog”
and Design
The JikesXen Java Server PlatformGeorgios GousiosAthens University of Economics and Businessgousiosg@aueb.grAbstractThe purpose of the JVM is to abstract the Java language from the hardware and software platforms it runs on.For this rea-son,the JVM uses services offered by the host operating sys-tem in order to implement identical services for the Java lan-guage.The obvious duplication of effort in service provision and resource management between the JVM and the operat-ing system has a measurable cost on the performance of Java programs.In my PhD research,I try tofind ways of min-imizing the cost of sharing resources between the OS and the JVM,by identifying and removing unnecessary software layers.Categories and Subject Descriptors C.0[General]:Hard-ware/software interfaces; C.4[Performance of Systems]: Performance Attributes; D.4.7[Operating Systems]:Orga-nization and DesignGeneral Terms Performance,LanguagesKeywords JVM,Performance,Operating System,Virtual Machine1.Background and MotivationThe raison d’ˆe tre of contemporary JVM s is to virtualize the Java language execution environment to allow Java pro-grams to run unmodified on various software and hardware platforms.For this reason,the JVM uses services provided by the host OS and makes them available to the executing program through the Java core libraries.Table1provides an overview of those services.The JVM is a software representation of a hardware ar-chitecture that can execute a specific format of input pro-grams.Being such,it offers virtual hardware devices such as a stack-based processor and access to temporary storage (memory)through bytecode instructions.The JVM is not a general purpose machine,though:its machine code includesCopyright is held by the author/owner(s).OOPSLA’07,October21–25,2007,Montr´e al,Qu´e bec,Canada.ACM978-1-59593-786-5/07/0010.support for high level constructs such as threads and classes,while memory is managed automatically.On the other hand, I/O is handled by the OS and access to I/O services is pro-vided to Java programs through library functions.The ob-servation that the JVM is both a provider and a consumer ofservices leads to the question of whether the JVM can assumethe role of the resource manager.Generic architectures sacrifice absolute speed in favor ofversatility,expandability and modularity.A question thatemerges is whether the services provided by the OS arestrictly necessary for the JVM to execute Java programsand,if not,what will be the increase in the performanceof the JVM if it is modified so as to manage the computingresources it provides to programs internally.Part of my workis to assess the necessity of certain services offered by the OS to the operation of the JVM and to measure their effect on the JVM performance,to get an estimate of the possible speedup that could be expected if the JVM replaced in the OS in the role of the resource provider/broker,in the context of purpose-specific systems.My initialfindings show that Java pays a large price to the OS for services it does not necessarily require.All standard OS s are designed for a specific purpose:to protect programs from accessing each other’s memory and to protect I/O re-sources from being accessed concurrently.As others have shown[2],if the execution environment is based on a type-safe language and therefore protects itself against invalid memory access,then memory protection,as it is currently implemented by OS s,is mostly unnecessary.My currently unpublishedfindings show that the mechanisms employed by JVM s to overcome the restrictions placed to them by the OS API s can deteriorate their performance,mainly when ex-ecuting memory and I/O intensive applications.2.Our ResearchThe specific question my research tries to answer is the fol-lowing:Can the performance of Java server platforms be improved by allowing them to manage directly the comput-ing resources they require?For the purposes of my research,I make the following assumptions—based on worked published by other re-searchers:Resources JVM System Library OS Kernel JikesXenCPU Java to Native Thread Map-ping Native to Kernel ThreadMappingThread Resource Alloca-tion,Thread SchedulingMultiplexing Java threadsto CPU s,Thread initializa-tionMemory Object Allocation andGarbage Collection Memory Allocation andDeallocation from theprocess address spaceMemory protection,PageTable Manipulation,Mem-ory AllocationObject Allocation andGarbage CollectionI/O Java to System Library I/OMapping,Protect Java frommisbehaving I/O Provide I/O abstractions System call handling,Uni-fied access mechanisms toclasses of devicesDriver infrastructure,I/OmultiplexingTable1.Resource management tasks in various levels of the Java execution stack and in JikesXen JVM•A JVM is able to manage efficiently computational re-sources by scheduling Java threads to physical processors internally.Multiprocessing can be implemented within the JVM very efficiently[1].•Virtual memory is deteriorating the performance of garbage collection[4]and thus it should be omitted in favor of a single,non-segmented address space.•A specially modified JVM can be run directly on hard-ware with minimal supporting native code[3],and there-fore the unsafe part of the system can be minimized down to a very thin hardware abstraction layer.To investigate my research question,I am in the pro-cess of designing and building the JikesXen virtual machine, which is a port of JikesRVM to the Xen VMM platform. JikesXen is based on a very thin layer of native code,the nanokernel,whose main purpose is to receive VMM inter-rupts and forward them to the Java space and also to initial-ize the VMM on boot.I use parts of the Xen-provided Mini OS demonstration operating system as the JikesXen nanokernel. The system itself is essentially a blob that lumps together the JikesRVM core image and stripped down version of the classpath which are already pre-compiled to native code dur-ing the JikesRVM build phase.JikesXen does not require a device driver infrastructure;instead,the Java core libraries will use the devices exported by the VMM directly,through adapter classes for each device.The adapter classes are also responsible to prevent concurrent access to shared resources, such as I/O-ports using Java based mutual exclusion prim-itives(e.g.synchronized blocks).It also implements other functionality relevant to the device class such as caching, network protocols,file systems and character streams.Ta-ble1presents an overview of the functionality of JikesXen as a resource manager.My work falls into the bare metal JVM research stream. But how does it differ from already existing approaches? The distinctive characteristic of my system is that it does not run directly on hardware.The use of the Xen VMM for taking care of the hardware intricacies,allows me to focus on more important issues than communicating with the hardware,such as providing a copy-less data path from the hardware to the application.Additionally,since all access to hardware is performed through the classpath,I plan to use the JVM locking mechanisms to serialize access to it. This will render the requirement for a resource manager, or a kernel,obsolete.Finally,my system’s focus being a single multithreaded application,I do not consider a full-fledged process isolation mechanism.Instead,we base our object sharing models on proven lightweight object sharing mechanisms(isolates[1]).3.ConclusionsMy work is in its early stages of development.I am currently constructing a build system that will compile the classpath and JikesRVM in a single binary image.I have already suc-ceeded in loading the JikesRVM boot image in the Xen hy-pervisor,but at the moment the system does nothing useful as important subsystems remain to be implemented. AcknowledgmentsThis work is partially funded by the Greek Secretariat of Re-search and Technology thought,the Operational Programme “COMPETITIVENES”,measure8.3.1(PENED),and is co-financed by the European Social Funds(75%)and by na-tional sources(25%)and partially by the European Com-munity’s Sixth Framework Programme under the contract IST-2005-033331“Software Quality Observatory for Open Source Software(SQO-OSS)”.References[1]Grzegorz Czajkowski,Laurent Dayn`e s,and Ben Titzer.Amulti-user virtual machine.In Proceedings of the General Track:2003USENIX Annual Technical Conference,pages 85–98,San Antonio,Texas,USA,ENIX. [2]Galen Hunt et al.An overview of the Singularity project.Microsoft Research Technical Report MSR-TR-2005-135MSR-TR-2005-135,Microsoft Research,2005.[3]Georgios Gousios.Jikesnode:A Java operating system.Master’s thesis,University of Manchester,September2004. [4]Matthew Hertz,Yi Feng,and Emery D.Berger.Garbagecollection without paging.In PLDI’05:Proceedings of the 2005ACM SIGPLAN conference on Programming language design and implementation,pages143–153,New York,NY, USA,2005.ACM Press.。
广域存储的一致性和稳定性COPS理论
Princeton University, † Intel Labs, ‡ Carnegie Mellon University
ABSTRACT
To appear in Proceedings of the 23rd ACM Symposium on Operating Systems Principles (SOSP’11)
Don’t Settle for Eventual:
Scalable Causal Consistency for Wide-Area Storage with COPS1.INTRODUCTION
Categories and Subject Descriptors
C.2.4 [Computer Systems Organization]: Distributed Systems
General Terms
Design, Experimentation, Performance
Distributed data stores are a fundamental building block of modern Internet services. Ideally, these data stores would be strongly consistent, always available for reads and writes, and able to continue operating during network partitions. The CAP Theorem, unfortunately, proves it impossible to create a system that achieves all three [13, 23]. Instead, modern web services have chosen overwhelmingly to embrace availability and partition tolerance at the cost of strong consistency [16, 20, 30]. This is perhaps not surprising, given that this choice also enables these systems to provide low latency for client operations and high scalability. Further, many of the earlier high-scale Internet services, typically focusing on web search, saw little reason for stronger consistency, although this position is changing with the rise of interactive services such as social networking applications [46]. We refer to systems with these four properties—Availability, low Latency, Partition-tolerance, and high Scalability—as ALPS systems. Given that ALPS systems must sacrifice strong consistency (i.e., linearizability), we seek the strongest consistency model that is achievable under these constraints. Stronger consistency is desirable because it makes systems easier for a programmer to reason about. In this paper, we consider causal consistency with convergent conflict handling, which we refer to as causal+ consistency. Many previous systems believed to implement the weaker causal consistency [10, 41] actually implement the more useful causal+ consistency, though none do so in a scalable manner. The causal component of causal+ consistency ensures that the data store respects the causal dependencies between operations [31]. Consider a scenario where a user uploads a picture to a web site, the picture is saved, and then a reference to it is added to that user’s album. The reference “depends on” the picture being saved. Under causal+ consistency, these dependencies are always satisfied. Programmers never have to deal with the situation where they can get the reference to the picture but not the picture itself, unlike in systems with weaker guarantees, such as eventual consistency. The convergent conflict handling component of causal+ consistency ensures that replicas never permanently diverge and that conflicting updates to the same key are dealt with identically at all sites. When combined with causal consistency, this property ensures that clients see only progressively newer versions of keys. In comparison, eventually consistent systems may expose versions out of order. By combining causal consistency and convergent conflict handling, causal+ consistency ensures clients see a causally-correct, conflict-free, and always-progressing data store. Our COPS system (Clusters of Order-Preserving Servers) provides causal+ consistency and is designed to support complex online applications that are hosted from a small number of large-scale datacenters, each of which is composed of front-end servers (clients of
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Experimental Study of Market Reputation MechanismsKay-Yut Chen Hewlett-Packard Labs Kay-Yut.Chen@Tad HoggHewlett-Packard LabsTad.Hogg@Nathan WoznyCalifornia Institute of TechnologyABSTRACTWe experimentally compare low-information, high-information and self-reporting reputation mechanisms. The results indicate players strategically reacted to the reputation mechanisms, with higher information mechanisms increasing market efficiency. Categories and Subject DescriptorsJ.4 [Computer Applications]: SOCIAL AND BEHAVIORAL SCIENCESGeneral TermsEconomics, Experimentation, Human FactorsKeywordsexperimental economics, reputation, markets1. Reputation and MarketsReputations can help ensure promised actions are taken without the expense of external enforcement or third party monitoring [6]. The Internet and subsequent development of e-commerce allow an increasing number of small players to engage in buying and selling. However, this environment increases the importance of establishing trust in a market where everyone can choose to be anonymous. One approach is eBay’s feedback mechanism in which participants rate the performance of the other party in their transactions.Establishing trust is possible through repeated interactions, e.g., the tit-for-tat strategy in the iterated Prisoner’s Dilemma [1]. The effects of various reputation mechanisms have been experimentally studied in simplified contexts. One approach [5] uses the “trust game” among pairs of players, where one player can choose to send money to a second, this amount is then substantially increased and the second player can choose to share some of that gain with the first player. By removing many of the complexities involved in market transactions, this game can study the effect of different policies of past behaviors revealed to participants. For questions involving reputation's effect on market efficiency, more complex experimental scenarios are needed. One example is a fixed-price market in which sellers have the option of not fulfilling their contracts [2]. Players are placed in pairs by the experimenter, and randomly assigned roles of buyer or seller. The buyer chooses whether to send the purchase price to the seller and, if so, the seller then decides whether to deliver the purchase or just keep the buyer's money. Revealing the seller's history of fulfillment then provides reputation information for the buyer. A similar one-sided market, but allowing variable prices, arises in an experimental study of a principal/agent market [4]. In this case the agents choose their level of service after principals have committed to their contracts. In this experiment, the agents face a market choice, maximizing the difference between their fee and expense on providing their service, but the payoff for the principal is probabilistic rather than directly determined by the choice made by the agent.In the e-commerce context, although a given person may engage in many transactions, only a few may be with a particular person. In such cases, people face the question of whether a proposed transaction is likely to proceed as agreed. Moreover, unlike the Prisoner’s Dilemma, people can choose not to do business with those deemed untrustworthy, or offer different terms based on the perceived level of trust. Furthermore, individuals usually cannot spread risk among many transactions, unlike large companies. This may lead to risk-averse people avoiding transactions that could benefit both parties, and thus result in lower market efficiency. Thus an important question is to what extent reputation mechanisms can aid this process. Analysis of eBay-like markets suggests that while feedback has desirable features in theory [3], such a market may not be efficient, fair and stable in practice.To help address these issues, we designed experiments to provide a broader set of endogenous choices for the players than in the prior experimental work described above. First, the players can explicitly decide who they wish to do business with rather than being paired with a single other player by the experimenter. Thus we can examine whether people choose to use reputation information to ostracize those with low reputations or give them poor prices based on their higher perceived risk. Second, both buyers and sellers make fulfillment choices and so reputations are relevant to both sides of the market. In the context of self-reported information, this allows players the opportunity to choose to misreport their experience as possible punishment for a poor report on their own reputation. More generally, it allows for the formation of clusters of mutually high-reputation trading arrangements. Third, our experiments include a full market so prices and trading volumes are determined endogenously, providing a broader view of the macroeconomic consequences of different information policies than is possible in more restricted scenarios. On the other hand, the additional complexity of a full market and players selecting who to do business with makes for a more complicated theoretical analysis.We choose to study reputation in the laboratory due to the uncontrollable nature of real world marketplaces. The key issues in designing reputation mechanisms are the nature of information to collect, its reliability and how it is disseminated to the participants. This paper reports on an experimental comparison of mechanisms revealing differing amounts of information on the transaction history of the players. We found that subjects responded to amount of information strategically. In addition, we observed the endogenously generated consequences for market price and volume.COPYRIGHT IS HELD BY THE AUTHOR/OWNER(S). EC’04, MAY 17–20, 2004, NEW YORK, NEW YORK, USA. ACM 1-58113-711-0/04/0005.2. Experimental DesignLaboratory economic experiments involve small groups interacting for only a few hours. Thus experiments with reputation mechanisms require a simple underlying market so participants can quickly build up enough transaction history to distinguish behaviors. Furthermore, the value of a high reputation decreases toward the end of an experiment, leading experienced subjects to not fulfill contracts. We thus need enough time to see at least some behavior relatively unaffected by this end-of-game effect.The main component of the experiment is an exchange economyof a single homogenous good. Each unit of good a buyer purchased can be redeemed for a pre-determined amount of money. Each unit of good a seller sold costs a pre-determined amount of money. In each period, buyers and sellers receive tables listing their redemption values and costs, respectively. The aggregate supply and demand is kept constant across periods, and this is told to participants before each experiment. However, redemption values and costs were assigned to random individualsin each period.We chose a discrete double auction as the market institution. Each period consists of a fixed number of rounds. Buyers and sellers took turns making offers and accepting offers made by others. Players were permitted to have only one offer at a time, although they may offer to buy or sell multiple units. In addition, we allowed the subjects to filter their offer, i.e., allowing only some agents to accept their offer. When an offer was accepted, it became a contract – an agreement for the seller to send the goods and for the buyer to send payment. The contracts were not binding. After the last round of exchanging offers, players were given a list of contracts they had accepted. They then decided whether or not to fulfill each contract. We included noise, as a fixed probability that either payments or goods are lost in transit. The information policy determines the past transaction behavior made available to participants. We used three information policies, announced at the start of each experiment:1. Low information: Players knew only their own transactions.2. High information: Players were told about all transactions that took place between any buyer and any seller.3. Self-reported ratings: After contract fulfillment, players could give trading partners a positive or negative rating. A value-weighted score of the feedback is then made public.Prior to the experiment, participants completed a short training program, with self-check quizzes, via the web (/econexperiment/MarketInfo/instructions.htm.)3. ResultsWe conducted 8 experimental sessions. The first had 8 subjects. The rest had at least 12 subjects. In all experiments, market prices converged reasonably well to equilibrium within 3 periods. While rapid convergence is expected for the market [7], this may not have been the case with the our addition of non-binding contracts.As expected, all experiments exhibit strong end-game effects. Subjects were told when the game would end two periods ahead of time. Fulfillment decreased sharply around 4 periods before the end of an experiment.We expect people are more likely to fulfill contracts when more information about their actions is available to people deciding whether to transact future business with them. This matches our observation of lower fulfillment with the low information policy, where subjects can safely reason that only people with whom they had unfulfilled contracts know of their “bad” record. Thus, not fulfilling a contract is less costly to future business than the case with full information. This is also strong evidence that people react to strategic implications of how information is used by other people. The figure shows an example of our experimental results.Fulfillment rates in the self-reporting treatment are similar to the high information treatment. This is consistent with observations on Internet auction sites such as eBay where the amount of negative feedback is only about 1%. This is not direct evidence that cheating is rare since feedback does not have to be 100% accurate. Similarly, reports in our experiment were not always accurate. However, they were, on average, a good representation of how likely a subject is going to fulfill contracts.Our experimental design allowed us to also observe endogenous macroeconomic effects: when fulfillment decreased, the contracted volume also increased but prices remained nearly constant. Market efficiencies responded to the level of fulfillment in a highly nonlinear fashion.4. References[1] R. Axelrod and W. Hampton, The Evolution of Cooperation, Science211:1390-1396 (1981).[2] G. Bolton et al., How Effective are Online Reputation Mechanisms?, Max Planck Institute paper 25-2002 (2002)[3] C. Dellarocas, Analyzing the Economic Efficiency of eBay-like Online Reputation Mechanism, in Proc of 3rd ACM Conf on E-Commerce, Oct 2001.[4] D. DeJong, R. Forsythe and R. Lundholm, Ripoffs, Lemons and Reputation Formation in Agency Relationships: A Laboratory Market Study, Journal of Finance40:809-823 (1985).[5] C. Kesser, Experimental Games for the Design of Reputation Management Systems, IBM Systems J. 42:498 (2003).[6] D. Klein, Reputation: Studies in the Voluntary Elicitation of Good Conduct, Univ. of Michigan Press 1997.[7] V. Smith, Bargaining and Market Behavior: Essays in Experimental Economics, Cambridge Univ. Press, 2000.。