Show simple item record

dc.contributor.advisorGrauman, Kristen Lorraine, 1979-en
dc.creatorVijayanarasimhan, Sudheendraen
dc.date.accessioned2011-06-02T14:33:43Zen
dc.date.accessioned2011-06-02T14:34:23Zen
dc.date.available2011-06-02T14:33:43Zen
dc.date.available2011-06-02T14:34:23Zen
dc.date.issued2011-05en
dc.date.submittedMay 2011en
dc.identifier.urihttp://hdl.handle.net/2152/ETD-UT-2011-05-3014en
dc.descriptiontexten
dc.description.abstractVisual recognition research develops algorithms and representations to autonomously recognize visual entities such as objects, actions, and attributes. The traditional protocol involves manually collecting training image examples, annotating them in specific ways, and then learning models to explain the annotated examples. However, this is a rather limited way to transfer human knowledge to visual recognition systems, particularly considering the immense number of visual concepts that are to be learned. I propose new forms of active learning that facilitate large-scale transfer of human knowledge to visual recognition systems in a cost-effective way. The approach is cost-effective in the sense that the division of labor between the machine learner and the human annotators respects any cues regarding which annotations would be easy (or hard) for either party to provide. The approach is large-scale in that it can deal with a large number of annotation types, multiple human annotators, and huge pools of unlabeled data. In particular, I consider three important aspects of the problem: (1) cost-sensitive multi-level active learning, where the expected informativeness of any candidate image annotation is weighed against the predicted cost of obtaining it in order to choose the best annotation at every iteration. (2) budgeted batch active learning, a novel active learning setting that perfectly suits automatic learning from crowd-sourcing services where there are multiple annotators and each annotation task may vary in difficulty. (3) sub-linear time active learning, where one needs to retrieve those points that are most informative to a classifier in time that is sub-linear in the number of unlabeled examples, i.e., without having to exhaustively scan the entire collection. Using the proposed solutions for each aspect, I then demonstrate a complete end-to-end active learning system for scalable, autonomous, online learning of object detectors. The approach provides state-of-the-art recognition and detection results, while using minimal total manual effort. Overall, my work enables recognition systems that continuously improve their knowledge of the world by learning to ask the right questions of human supervisors.en
dc.format.mimetypeapplication/pdfen
dc.language.isoengen
dc.subjectArtificial intelligenceen
dc.subjectActive learningen
dc.subjectObject recognitionen
dc.subjectObject detectionen
dc.subjectCost-sensitive learningen
dc.subjectMulti-level learningen
dc.subjectBudgeted learningen
dc.subjectLarge-scale active learningen
dc.subjectLive learningen
dc.subjectMachine learningen
dc.subjectVisual recognition systemen
dc.titleActive visual category learningen
dc.date.updated2011-06-02T14:34:23Zen
dc.contributor.committeeMemberDhillon, Inderjit S.en
dc.contributor.committeeMemberAggarwal, J K.en
dc.contributor.committeeMemberMooney, Raymond J.en
dc.contributor.committeeMemberTorralba, Antonioen
dc.description.departmentComputer Sciencesen
dc.type.genrethesisen
thesis.degree.departmentComputer Sciencesen
thesis.degree.disciplineComputer Scienceen
thesis.degree.grantorUniversity of Texas at Austinen
thesis.degree.levelDoctoralen
thesis.degree.nameDoctor of Philosophyen


Files in this item

Icon

This item appears in the following Collection(s)

Show simple item record