Visual searching with the Web Perception Toolkit

Wouldn’t it be great if users could search your site using their camera? Imagine
this. Your site is Razor McShaveyface. Your customers tell you they have trouble
finding the right cartridges for their razor when they reorder. They don’t know
the right …

Wouldn’t it be great if users could search your site using their camera? Imagine
this. Your site is Razor McShaveyface. Your customers tell you they have trouble
finding the right cartridges for their razor when they reorder. They don’t know
the right keywords for your product search. And let’s be honest, they probably
never will.

What if they never need to? What if they could point their phone’s camera at the
UPC code on package and your site could present them with the right cartridge
and a big red "reorder" button?

Think of other ways you can use a camera on a site. Imagine a site that
supports in-store price checking. Imagine getting information about a museum
exhibit or historical marker. Imagine identifying real-world landmarks in games
like geocaching or scavenger hunts.

The Web Perception Toolkit
makes these camera-based scenarios possible. In some cases you can even create
an experience without writing code.

How does it work?

The open-source Web Perception Toolkit helps you add visual search to your
website. It passes a device camera stream through a set of detectors that map
real-world objects, here called "targets", to content on your site. This mapping is defined
using Structured Data (JSON-LD) on your site. With this data, you can present the right information in a customizable
UI.

I’ll show you enough of this to give you a taste of how it works. For a complete
explanation, check out the Getting
Started
guide, the toolkit
reference
, the I/O Sandbox demo and the sample demos.

Structured data

The toolkit can’t find just any target in the camera’s view. You must provide it
with linked JSON data for the targets you want it to recognize. This data also
contains information about those targets that will be shown to the user.

The data is all you need to create a user experience like the one in the image
below. If you do nothing else, the Web Perception Toolkit can identify targets,
then show and hide cards based on the information provided in the data. Try this
for yourself using our artifact-map
demo
.

The default interface is available by using just the linked data.
The default interface.

Add data to your site with a JSON linked data file, included using a <script>
tag and the "application/ld+json" MIME type.

<script type="application/ld+json" src="//path/to/your/sitemap.jsonld">

The file itself looks something like this:

[
{
"@context": "https://schema.googleapis.com/",
"@type": "ARArtifact",
"arTarget": {
"@type": "Barcode",
"text": "012345678912"
},
"arContent": {
"@type": "WebPage",
"url": "http://localhost:8080/demo/artifact-map/products/product1.html",
"name": "Product 1",
"description": "This is a product with a barcode",
"image": "http://localhost:8080/demo/artifact-map/products/product1.png"
}
}
]

The user experience

What if you want more than the default user experience? The toolkit gives you
livecycle events, Card and Button objects for crafting the user experience
around those events, and an easy way to style the cards. I’m going to show a
little of this with code based losely on our Getting
Started
guide.

The most important lifecycle event is PerceivedResults, which is fired every
time a target is found. A target can be a real-world object or a marker such as
a bar code or QR code.

The process for responding to this event is the same as for any other event with
an exception already alluded to. If you don’t implement the event, a user
interface is automatically created using structured data. To override this
behavior start your event handler by callingevent.preventDefault().

const container = document.querySelector('.container');
async function onPerceivedResults(event) {
// preventDefault() to stop default result Card from showing.
event.preventDefault();
// Process the event.
}
window.addEventListener(PerceptionToolkit.Events.PerceivedResults, onPerceivedResults);

Let’s look at the event more closely. The event itself contains arrays of
markers and targets that it has both found and lost. When targets are found in
the world, the even fires and passes found objects in event.found. Similarly,
when targets pass from the camera view the event fires again, passing lost
objects in event.lost. This helps account for hand and marker movements:
cameras not held steadily enough, dropped markers, that kind of thing.

async function onPerceivedResults(event) {
// preventDefault() to stop default result Card from showing
event.preventDefault();
if (container.childNodes.length > 0) { return; }
const { found, lost } = event.detail;
// Deal with lost and found objects.
}

Next, you show an appropriate card based on what the toolkit found.

async function onPerceivedResults(event) {
event.preventDefault();
if (container.childNodes.length > 0) { return; }
const { found, lost } = event.detail;
if (found.length === 0 && lost.length === 0) {
// Object not found.
// Show a card with an offer to show the catalog.
} else if (found.length > 0) {
// Object found.
// Show a card with a reorder button.
}
}

Adding cards and buttons is simply a matter of instantiating them and appending
them to a parent object. For example:

const { Card } = PerceptionToolkit.Elements;
const card = new Card();
card.src = 'Your message here.'
container.appendChild(card)'

Finally, here’s what the whole thing looks like. Notice the conveniences I’ve
added to the user experience. Whether the marker is found or not, I provide
one-click access to what I think is most useful in the circumstances.

async function onPerceivedResults(event) {
// preventDefault() to stop default result Card from showing
event.preventDefault();
if (container.childNodes.length > 0) { return; }
const { found, lost } = event.detail;
const { ActionButton, Card } = PerceptionToolkit.Elements;
if (found.length === 0 && lost.length === 0) {
//Make a view catalog button.
const button = new ActionButton();
button.label = 'View catalog';
button.addEventListener('click', () => {
card.close();
//Run code to launch a catalog.
});
//Make a card for the button.
const card = new Card();
card.src = 'We wish we could help, but that\'s not our razor. Would you like to see our catalog?';
card.appendChild(button);
//Tell the toolkit it does not keep the card around
// if it finds something it recognizes.
card.dataset.notRecognized = true;
container.appendChild(card);
} else if (found.length > 0) {
//Make a reorder button.
const button = new ActionButton();
button.label = 'Reorder';
botton.addEventListener('click', () => {
card.close();
//Run code to reorder.
})
const card = new Card();
card.src = found[0].content;
card.appendChild(button);
container.appendChild(card);
}
}

Formatting cards

The Web Perception Toolkit provides built-in formatting for cards and buttons
with the default stylesheet. But you can easily add your own. The provided
Card and ActionButton objects contain style properties (among many others)
that let you put your organizational stamp on the look and feel. To include the
default stylesheet, add a <link> element to your page.

<link rel="stylesheet" href="//path/to/toolkit/styles/perception-toolkit.css">

Conclusion

As I said at the top, this is not an exhaustive look at the Web Perception
Toolkit
. Hopefully it gives
you a sense of how easy it is to add visual searching to a website. Learn more
with the Getting Started guide
and the sample
demos
.
Dig in to the toolkit
documentation
to learn what it
can do.


Print Share Comment Cite Upload Translate
APA
Joe Medley | Sciencx (2024-03-28T21:09:11+00:00) » Visual searching with the Web Perception Toolkit. Retrieved from https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/.
MLA
" » Visual searching with the Web Perception Toolkit." Joe Medley | Sciencx - Tuesday May 7, 2019, https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/
HARVARD
Joe Medley | Sciencx Tuesday May 7, 2019 » Visual searching with the Web Perception Toolkit., viewed 2024-03-28T21:09:11+00:00,<https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/>
VANCOUVER
Joe Medley | Sciencx - » Visual searching with the Web Perception Toolkit. [Internet]. [Accessed 2024-03-28T21:09:11+00:00]. Available from: https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/
CHICAGO
" » Visual searching with the Web Perception Toolkit." Joe Medley | Sciencx - Accessed 2024-03-28T21:09:11+00:00. https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/
IEEE
" » Visual searching with the Web Perception Toolkit." Joe Medley | Sciencx [Online]. Available: https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/. [Accessed: 2024-03-28T21:09:11+00:00]
rf:citation
» Visual searching with the Web Perception Toolkit | Joe Medley | Sciencx | https://www.scien.cx/2019/05/07/visual-searching-with-the-web-perception-toolkit/ | 2024-03-28T21:09:11+00:00
https://github.com/addpipe/simple-recorderjs-demo