4
VizWiz Dataset Browser: A Tool for Visualizing Machine Learning Datasets Nilavra Bhattacharya * , Danna Gurari University of Texas at Austin * [email protected] Figure 1: Screenshot of our visualization tool showing an example of a visual-question, alongside different types of annota- tions that can be searched for. This example was obtained by searching for the word “milk” within the question text. Abstract We present a visualization tool to exhaustively search and browse through a set of large-scale machine learn- ing datasets. Built on the top of the VizWiz dataset, our dataset browser tool has the potential to support and en- able a variety of qualitative and quantitative research, and open new directions for visualizing and researching with multimodal information. The tool is publicly available at https://vizwiz.org/browse. 1. Introduction A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal- ysis [7]. Researchers often want to immerse themselves in the data. For datasets containing thousands of images and annotations, there is no straightforward way to do this. Vi- sualization efforts involve writing very specific programs or scripts to generate plots, such as bar charts containing counts of different categories, or sunburst diagrams show- ing relative proportions of different annotations. How- 1 arXiv:1912.09336v1 [cs.LG] 19 Dec 2019

arXiv:1912.09336v1 [cs.LG] 19 Dec 2019 · A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: arXiv:1912.09336v1 [cs.LG] 19 Dec 2019 · A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers

VizWiz Dataset Browser: A Tool for Visualizing Machine Learning Datasets

Nilavra Bhattacharya∗, Danna GurariUniversity of Texas at Austin

[email protected]

Figure 1: Screenshot of our visualization tool showing an example of a visual-question, alongside different types of annota-tions that can be searched for. This example was obtained by searching for the word “milk” within the question text.

Abstract

We present a visualization tool to exhaustively searchand browse through a set of large-scale machine learn-ing datasets. Built on the top of the VizWiz dataset, ourdataset browser tool has the potential to support and en-able a variety of qualitative and quantitative research, andopen new directions for visualizing and researching withmultimodal information. The tool is publicly available athttps://vizwiz.org/browse.

1. Introduction

A major challenge of working with large-scale machinelearning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers often want to immerse themselves inthe data. For datasets containing thousands of images andannotations, there is no straightforward way to do this. Vi-sualization efforts involve writing very specific programsor scripts to generate plots, such as bar charts containingcounts of different categories, or sunburst diagrams show-ing relative proportions of different annotations. How-

1

arX

iv:1

912.

0933

6v1

[cs

.LG

] 1

9 D

ec 2

019

Page 2: arXiv:1912.09336v1 [cs.LG] 19 Dec 2019 · A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers

ever, these visualization attempts can produce aggregatedresults only, thereby hiding interesting examples. Even fordata cleaning and quality control purposes, manually goingthrough each individual image and its annotation is tedious,and is prone to human error.

To overcome these challenges, we developed the VizWizdataset browser. The VizWiz dataset originates from peoplewho are blind, who used mobile phones to snap photos andrecord questions about them (e.g., “what type of beverageis in this bottle?” or “has the milk expired?”), and containsimages paired with questions about the image [5]. Subse-quent research has generated a variety of annotations on topthe VizWiz dataset. These include: ten crowdsourced an-swers to each visual question [6], reasons explaining whythe ten answers can differ, if they do [4]; captions for de-scribing the images to users with visual impairments; themultitude of skills needed by an AI system to automaticallyanswer the visual question; quality issues present in the im-ages (since they were captured by users who could not seethe photo they were capturing), and whether text is presentin the image. As more and more annotations were beingcollected, we felt the need to view all these different kindsof rich data in a single platform, in order to get a holisticview of the information contained within these datasets.

2. Design and Implementation

The VizWiz Dataset Browser is a single-page web-application built on the Linux - Apache - MariaDB - PHP(LAMP) stack. It supports searching for textual annota-tions, and filtering for categorical annotations. The mainpurpose of the tool is to view images, and search for thoseimages using the ‘meta-data’ provided by the annotations.To scale effortlessly with an increasing variety of annota-tions, we decided to keep the search functionalities on theleft side of the screen, in its own independently scrollablesection. By not opting for a horizontal layout of the searchand filter options, we can display more dynamic informa-tion above the fold. Similar design choices are employedby popular eCommerce websites which display numerousfilters on their search-results page [1, 2, 3].

2.1. Visualization Section

Figure 1 shows a screenshot of the main information vi-sualization area. The image and the textual annotations:(a) question, (b) ten answers, and (c) five captions, are dis-played in their natural form, while the categorical anno-tations: (d) answer-difference reasons, (e) skills, and (f)quality issues, are displayed as one-dimensional heatmaps,based on how many crowdworkers (out of 5) selected a cat-egorical label.

2.2. Summary of Results

The top portion of the visualization section shows a sum-mary of the search results. This includes the number of totalimages found for the current search and/or filter query, andthe range of images shown on the current page. To supportminimal page loading times, we decided to show a max-imum of 50 images per page. Users can choose to viewthe thumbnails of all the images displayed on the currentpage (as shown in Figure 2) by clicking on ‘Expand Sum-mary of Images’. Clicking on a thumbnail image withinthe ‘Summary of Images’ section will take the user to thedetails-section of the image.

Figure 2: The summary section shows an overview of thedifferent images returned for the search or filter query.Clicking a thumbnail image lets the user view the detailsof the image, as in Figure 1. This example was obtained bysearching for the word “glass” in the question.

2.3. Searching for Images by Textual Annotations

Text searching capabilities are present for searching forwords and phrases within the visual question, the ten an-swers, and the five crowdsourced captions. Full-text search-ing is powered by MariaDB relational database1. Addition-ally, users can search for an image using its specific file-name. These search capabilities are shown in Figure 3.

2.4. Filtering Images by Categorical Annotations

The visualization tool can be used to filter images basedon the different types of categorical annotations available:

1https://mariadb.com/kb/en/library/full-text-index-overview

2

Page 3: arXiv:1912.09336v1 [cs.LG] 19 Dec 2019 · A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers

Figure 3: Different ways to search for images using textualannotations. Users can search for words and phrases withinthe question, the ten answers, and the five captions.

Figure 4: Filtering for images using categorical annotations.The screenshot shows the labels for the answer-differencedataset [4].

(a) answer-difference reasons, (b) skills, and (c) quality is-sues. This functionality proves to be useful when we wantto explore relationships between the different datasets. Forexample, selecting DFF (Difficult Question) as an answer-difference reason, and ROT (image needs to be rotated)as an image-quality issue, we can view the specific caseswhere the visual questions are difficult to answer becausethe images need to be rotated. The filtering capabilities forthe answer-difference reasons are shown in Figure 4.

2.5. Ordering of Search Results

Figure 5: Various options for ordering the search results.

The search results can be ordered (sorted) using the op-tions shown in Figure 5. When searching for textual an-notations (words or phrases in the question, answers, orcaptions), the result are sorted in decreasing order of thenumber of matched words in the annotation. ‘Diversityof answers’ orders the results based on how different theten answers are, using the Shannon Entropy of the ten an-swers. For categorical annotations (answer-difference rea-sons, skills, quality issues, text-presence), the results areranked based on how many crowdworkers (out of five) an-notated the images using the chosen categorical labels.

2.6. Toggling Display of Annotations

Figure 6: Options to hide or show different datasets.

Viewing all the different annotations at once can be over-whelming. Often, the user may want to selectively viewcertain annotations (e.g., for taking screenshots). For thispurpose, the ‘View’ section, as shown in Figure 6, can beused to hide or show the different datasets as desired.

3

Page 4: arXiv:1912.09336v1 [cs.LG] 19 Dec 2019 · A major challenge of working with large-scale machine learning datasets is the difficulty of exploratory data anal-ysis [7]. Researchers

3. ConclusionIn summary, the VizWiz Dataset Browser can prove to

be a useful tool to search, filter, and visualize multiple largedatasets. It is already being used to aid a variety of ongoingresearch efforts in the domains of computer vision, acces-sibility, and human-computer interaction. We are hopefulthat future researchers who choose to work with the VizWizdataset will find the tool useful for answering interestingresearch questions.

AcknowledgementsWe thank the crowdworkers for providing the annotations.We thank Kenneth R. Fleischmann, Meredith Morris, EdCutrell, and Abigale Stangl for their valuable feedbackabout this tool and paper. This work is supported in partby funding from the National Science Foundation (IIS-1755593) and Microsoft.

References[1] Amazon.com: Online Shopping. https://www.amazon.

com. [Online; accessed 11-Dec-2019]. 2[2] eBay Inc. https://www.ebay.com. [Online; accessed

11-Dec-2019]. 2[3] Walmart Inc. https://www.walmart.com. [Online; ac-

cessed 11-Dec-2019]. 2[4] N. Bhattacharya, Q. Li, and D. Gurari. Why does a visual

question have different answers? In Proceedings of the IEEEInternational Conference on Computer Vision, pages 4271–4280, 2019. 2, 3

[5] J. P. Bigham, C. Jayant, H. Ji, G. Little, A. Miller, R. C. Miller,R. Miller, A. Tatarowicz, B. White, and S. White. VizWiz:Nearly real-time answers to visual questions. In Proceedingsof the 23nd Annual ACM Symposium on User Interface Soft-ware and Technology, pages 333–342. ACM, 2010. 2

[6] D. Gurari, Q. Li, A. J. Stangl, A. Guo, C. Lin, K. Grauman,J. Luo, and J. P. Bigham. VizWiz Grand Challenge: Answer-ing Visual Questions from Blind People. In Proceedings ofthe IEEE Conference on Computer Vision and Pattern Recog-nition, pages 3608–3617, 2018. 2

[7] J. W. Tukey. Exploratory Data Analysis. Addison-WesleyPublishing Company, 1977. 1

4