Get fresh updates from Hortonworks by email

Once a month, receive latest insights, trends, analytics information and knowledge of Big Data.

cta

Get Started

cloud

Ready to Get Started?

Download sandbox

How can we help you?

closeClose button
August 05, 2016
prev slideNext slide

Demo #1 Play-by-Play: Data Hacks & Demos @ #HS16SJ

Match image to an identifier, correlate with data and initiate personalized, real time electronic convo with customer in store

During the 1st demo of the Data Hacks & Demos session, at Hadoop Summit San Jose, Jeremy Dyer modelled the scenario of a customer walking into a store, where a retailer can find out who they are, what kind of shopper they are, what they are interested in, so they can engage and create the most appealing offer.

So what did Jeremy tell the audience?

Physical stores are losing a lot of perishable data, but it doesn’t have to be that way

Most digital businesses can capture your clickstream – what you are surfing for on the web, who you are, what are your interests. But there is an entire physical world of interactions that cannot be tracked, or necessarily correlated with your “digital identity”. In fact, for physical retailers, it can be very difficult to capture and interpret any of this information. People are interacting with products, with their phone, with each other. But why is this difficult? How does a brick and mortar store associate you with a digital identity before you checkout? How can a physical store know if they are stocking the right products that appeal to their customer base? Are potential customers price shopping while in the store?

And being able to figure this out isn’t a “science fiction” type moment from Star Trek or The Jetsons. All the technology needed to do this exists today – WiFi, iBeacons, closed circuit security,  cell phones, and Apache NiFi!  Together, all this enables a real-time capability to collect data, make it more valuable and analyze it.

During the three days of Hadoop Summit, the audience participated in a live demo that had a stationary Thunderbolt display capturing images of Hadoop Summit attendees, along with a mobile cell phone and backpack roaming the venue capturing attendees presence. The images were matched to a bar code on the bottom of their badge, and matched to a snapshot of their face, and then sent into the cloud.

This is the essence of an edge system, capturing raw data from the point of origin with Apache NiFi , very focused on sensing information. Then this data is flowed into another system  – a more back office system using Apache NiFi and more, where you have the most resources to both acquire and combine information. Armed with all this information you can decide if you want to interact with the customer in the store  – perhaps send a coupon or send some kind of audio feedback.

How did edge processing of images work?

To demonstrate how it all worked, Jeremy uploaded an image of himself and ran it through the edge processing system – Apache NiFi. Apache NiFii takes the data and sends it to a back-end server which is listening for data. When it receives the image file, it run a detection algorithm tied to the bar code system, and looks up the first name, last name, then it combines it with information it has correlated to the bar code and the electronic voice speaks  “Je-re-mee from Hor-ton-works, how was your tra-vel from At-lan-ta?”

The  electronic voice speaks  “Je-re-mee from Hor-ton-works, how was your tra-vel from At-lan-ta?”

That was the 1st demo of Data Hacks & Demos at Hadoop Summit San Jose. The 2nd demo about Apache NiFi and Spark for facial recognition is up next in this blog series. In the meantime, to get started with building something like this yourself, check out these links:

Demo 1 Greet Customer Apache NiFi Hack Demos Hortonworks HS16SJ

Tags:

Leave a Reply

Your email address will not be published. Required fields are marked *