Want to explore "what if"?

What we learnt from analysing Pinterest data with machine learning

Written by RiaRole: Our Former Digital Content Specialist

Heads up: Our Ideas Factory has been refreshed, levelled up, and grown-up into Alphero Intelligence. Some of our old posts are pretty cool tho'. Check this one out.

Mobile phone showing a grid of photos. Photo by charlesdeluvio on Unsplash
  • What data can we extract from a Pinterest board that can be used to filter search results, or inform new product ideas?
  • Image information: Who gives you more? Pinterest or Google?
  • We’re going to need a load more training material to teach machines about Scandi design.

Pinterest self-identifies as a visual discovery engine. It has long established itself as the go-to platform for inspiration boards — whether you’re painting a portrait, designing a costume, or decorating your home. There are over 200 billion pins saved on Pinterest; two-thirds of which are related to brands and products. It’s a haven for product discovery.

So, naturally, we wanted to see what data we could extract from a Pinterest board to inform new product ideas.

From inspiration boards to real-world purchases

Could we train a machine learning model on a particular Pinterest board, to find our inspiration items at a hardware or furniture store?

pinterest hardware store
Turning Courtyard Garden inspiration into real products

What information can we get straight from Pinterest?

There’s a limited amount of data that the Pinterest API will return.

Code snippet from Pinterest API
Not much info!

As you can see, most of the interesting contextual data is in the notes. If we analyse the notes of each Pin pinned to a Pinterest Board, surely we can get some interesting stuff back, right?

Contextual data from notes in Pinterest API
Turning this...
Snapshot of Pinterest dataset
... into this!

This gives us way more insight into some of the keywords commonly associated with the Board and its Pins, but the information that is really relevant is buried into general classification which is complex to sift through.

What can we discover about an image with Google Cloud Vision?

Compared to the surface-level data generously provided to us by Pinterest themselves, it would be fair to hypothesise that we could derive more detailed insights using Google Cloud’s pre-trained machine learning models.

Showing what information a user can get from a Pinterest Pin
Taking a Pin, and getting a whole bunch of information from it. Source Pinterest

We had way more success extracting data from individual pins than from entire Pinterest boards. At a Pin-level, it was much easier to extract a colour palette and objects from a photo using computer vision. But these labels and object classifications are still way too broad to accurately use for online shopping purposes. Chair? Plant? Table? Well, yeah…

What else could we do with Pinterest?

Pinterest has its own tool for finding visually similar results, but it also has its mishaps.

Pinterest visually showing similar results for a search
Close enough — Let’s just hang door knobs from the kitchen ceiling!
User search on Pinterest for kitchen lights
We can do better. Source Pinterest

If we have a detailed enough image, consisting of a number of objects, we can use the Pinterest API to extract more granular data.

But as for taking those labels and chucking them into online stores as search queries?

search queries on Esty vs Amazon
Source Etsy and Amazon

It doesn’t go well.

Machines can recognise things, but just can’t categorise them accurately with the limited amount of training we were able to do based on Pinterest. Sometimes the style is more relevant than the literal object — for example, Scandinavian design. We can return more meaningful results via computer vision than we can with frequency analysis, but it needs refinement. Iterating parts of images over and over might also yield better results. We’re going to need a load more training material to teach machines about Scandi.

Regardless, even if we can get perfect information from an image, it doesn’t matter if our catalogue doesn’t return results based on that information. Search performance is important and proper tagging within a catalogue will dictate the meaningfulness of results. That’s our takeaway.

Written by RiaRole: Our Former Digital Content Specialist