- Image zooming. Users wanted to zoom into the details of an image, especially when they couldn't get close enough to an object physically.
- Map of the museum/exhibit. Especially to locate objects that are on the device, or to tap on parts of the floor and get the objects that are located there.
- Artifact search. Typing in part of the name/description of an artifact, and getting a list of relevant artifacts. This was especially desired when in front of an artifact that didn't have an object code (or when they didn't detect the object code label).
- At the beginning of the session, a tutorial/guide on how to use the application/what to use it for.
- List of digital "nuggets". A number of users wanted a full list (on the device) of all things digital that they could look at for the exhibition on the device, whether it be audio, video, images, extra information, stories, etc.
- A way of narrowing down lists of artifacts. For instance, if user knows that the artifact they're looking for is a painting, some way of narrowing the list down to paintings.
- More narrative behind an artifact. Instead of more technical information.
- Something more visual, less textual. Some users noted being anxious to see something more visual/exciting, and less textual.
- Note that says whether you're allowed to touch/interact with specific artifacts. Some artifacts/interactives are meant to be touched, many others are not.
- Removal of one-media, two-tap redundancy. All/most artifacts only had one piece of media--it would've been better to play it after one tap instead of requiring two taps (one to expand the panel, the other to actually play).
- Location-awarenessaware browsing. P: "When you walk around, maybe it can show on the screen what the possibilities are, an image of what you can see from your position, instead of thinking, 'Where's the object code? Is there an object code?', and instead of saying, 'Well I saw number 8 and 10, but where were the others?'"