Facebook’s Artificial Intelligence Has The Ability to Search Photos by Content

The term artificial intelligence was coined 60 year ago. But now its starting to deliver. Lumos’s computer vision platform was initially used to improve the experience for visually impaired members of the Facebook community. Lumos is now powering image content search for all users. What does this means to you? You can now search for images on Facebook with key words that describe the contents of a photo, rather than being limited by tags and captions.

How does this work? It starts with the huge task of computational training. For the object recognition used in Facebook’s image search, the artificial intelligence (AI) system started with a small set of 130,000 public photos shared on Facebook. Using the annotated photos the system could learn which pixel patterns correspond to particular subjects. It then went on to use the tens of millions of photos on Facebook. So what this means is that the caption-reading technology trained a deep neural network though public photos shared on Facebook. The model essentially matches search descriptors to features pulled from photos with some degree of probability. You can now search for photos based on Facebook AI’s assessment of their content, not just based on how humans happened to describe the photos with text when they posted them.

How could this be used? Say you were searching on a dress you really liked in a video. Using the search it could be related back to something on Marketplace or even connect you directly with an ad-partner to improve customer experiences while keeping revenue growth afloat. So it seems it can help both customers, customer experience and companies selling things as well as ad partners.

What else is new? Facebook released the text-to-speech tool last April for visually impaired users so they could use the tools to understand the contents of photos. Then, the system could tell you that a photo involved a stage and lights, but it wasn’t very good at relating actions to objects. But now the Facebook team has improved that painstakingly labeling 130,000 photos pulled from the platform. Facebook trained a computer vision model to identify 12 actions happening in the photos. So for instance, instead of just hearing it was “a stage,” the blind person would hear “people playing instruments” or “people dancing on a stage” or “people walking” or “people riding horses.” This provides contextually relevancy that was before not possible.

You could imagine one day being able to upload a photo of your morning bagel and this technology could identify the nutritional value of that bagel because we were able to detect, segment, and identify what was in the picture.

So it seems the race is on for services not just for image recognition, but speech recognition, machine-driven translation, natural language understanding, and more. What’s your favorite AI vendor?

@Drnatalie, VP, Program Executive, Salesforce ITC

Share

Has The Age of George Jetson IoT Time Come? Alex Was the Star of CES

Alexa Voice Service (AVS) is the software that allows owners to control compatible devices with their voice. From the various  reports it was estimated there were 700–1,100 Alexa-controllable products at CES. And the Amazon / Alexa logo was everywhere at CES.

Is the Age of George Jetson here? In a smart home, everything from the the HVAC to the TV to window shades can be controlled. However it’s not easy to really have a whole house of Artificial Intelligence (AI) controlled devices. Why? Many of the IoT-enabled devices don’t talk to other devices if they are made by different manufacturers. Opps! The IoT world awaits THE killer app, like Apple Homekit or Google Home. We are still waiting for them to provide all encompassing, unified smart “home.”

The Amazon Echo is a hands-free speaker controlled with your voice. It connects to the Alexa Voice Service to provide information, news, play music, report on sports scores, deliver weather reports… The uses for AVS and Alexa are limited only by your imagination.

When something is connected to Alexa, the device instantly becomes pseudo-interoperable. Interoperable technology is not an evolutionarily stable strategy for most IoT manufacturers. Interoperability is the ability of different information technology systems and software applications to communicate, exchange data, and use the information that has been exchanged to do something.

What CES showed us is that voice control seems to be the unifying app for IoT. And Alexa is the biggest name in voice control. Smart devices are generally controlled with apps. If there is an app to control the smart device, the app allows AVS to directly control the smart device. So you could say, “Alexa, tell Crestron I’d like to turn the lights on in the bedroom” (for your Crestron) or “Alexa, I would like to turn the heat on the downstairs thermostat to 70 degrees” (for your Iris Smart Home System). It’s easy to see the value of voice control in so many ordinary situations. What’s interesting about AVS is that even though Crestron and Iris have nothing to do with one another, you can control them both with your voice.

Alexa has finely tuned automatic speech recognition (ASR) and natural language understanding (NLU) engines that recognize and respond to voice requests instantly. Alexa is always getting smarter with new capabilities and services through machine learning, regular API updates, feature launches, and custom skills from the Alexa Skills Kit (ASK.) The AVS API  is a programming language agnostic service that makes it easy to integrate Alexa into your devices, services, and applications. And it’s free.

And you can create meaningful user experiences for an endless variety of use cases with Alexa Voice Service (AVS); Amazon’s intelligent voice recognition and natural language understanding service. AVS includes a full range of features, including smart home control, streaming music content, news, timers… and can be added to any connected device that has a microphone and speaker.

But while Alexa has a head start, Google Home, an Echo competitor, is very likely to quickly catch up. Google Home though, works with a completely different set of protocols and has different “awake” words. These are command words that make it pay attention and carry out the request. It seems that we may need to learn to speak to different systems in different ways – perhaps we’ll need lessons in Alexa speak and Google speak as well as and Siri and Cortana speak!

So is the Age of George Jetson here yet? Sort of. What will be interesting is to see if there is a start-up that will pull all of this together so that us regular humans don’t need to become AI experts to connect and use the technology.

Dr. Natalie Petouhoff, VP and Principal Analyst, Constellation Research

Covering customer-facing applications

 

Share

Capgemini Collaborates with Celaton on Artificial Intelligence in the Cloud

What’s the Partnership Between Capgemini And Celaton Mean to Your Company? Capgemini, consulting, technology and outsourcing services, has announced a new global collaboration with Celaton, a specialist Artificial Intelligence (AI) company, to license and use its inSTREAM, cognitive learning technology. The 3 year contract, signed between Capgemini and Celaton, will extend Capgemini’s already strong automation capabilities, help to drive further efficiencies and add Artificial Intelligence to Capgemini’s Business Services solution portfolio.

What Does Celaton’s inSTREAM Software Do? It streamlines the handling of unstructured unpredictable (and structured) content such as correspondence, claims, complaints and invoices that organizations receive by email, social media, fax and paper. This minimizes the need for human intervention and ensures that only accurate, relevant and structured data enters business systems. Unique to inSTREAM is its ability to learn through the natural consequence of processing information and collaborating with people. Capgemini’s extensive knowledge and experience in business process services will also enable Celaton to accelerate and improve inSTREAM’s capabilities.

What Will The Partnership Provide For Clients? The cooperation will enable Capgemini to increase efficiency, shorten turnaround times and enhance quality in areas where incoming documents and queries need to be processed, improving overall customer satisfaction. At a time when more and more customers expect the use of AI and modern automation tools, the alliance will help Capgemini’s Business Services advance their market leading use of automation and AI for its core business. Earlier this year, Capgemini introduced an Autonomic Platform-as-a-Service (PaaS) offering founded on best of breed technologies to deliver intelligent automation solutions on-demand for enterprises. The Autonomic PaaS aims to improve the predictability of organizations’ operations across their infrastructure, applications and business processes. The Celaton agreement is a further commitment from Capgemini to develop advanced client solutions using intelligent automation, cognitive and AI technologies.

Is This Offered in a SaaS or Cloud Mode? The addition of Celaton inSTREAM expands Capgemini’s Business Services’ extensive Software-as-a-Service (SaaS) portfolio with an artificial intelligence-based processing solution for incoming unstructured content –which is driven by its global automation Centers of Excellence. It is an important element in ensuring the delivery of maximum value to its customers.

Notes From The Executives: Lee Beardmore, VP and Capgemini’s Business Services Chief Technology Officer said, “There is significant industry debate on how cognitive computing and artificial intelligence will impact the BPO market. We are taking our delivery from debate to global implementation and are proud to partner with Celaton as a leading vendor in the business process AI space. Building on the introduction of Capgemini’s Autonomic Platform-as-a-Service, Celaton’s technology extends the penetration of cognitive computing into our delivery of business process services.”

Andrew Anderson, CEO of Celaton said, “I am delighted that Celaton and Capgemini have committed to this global partnership. The transformational impact of AI has been proven with many organizations and yet this emerging technology is often greeted with scepticism. Capgemini’s global reach and credibility will have an impact on the perception and adoption of AI and I’m very excited that Capgemini’s customers will soon be able to realize its significant benefits.”

My POV: AI is very important to the emerging capabilities of company’s to add cognitive computing into the delivery of business processes of discerning unstructured content. And with social and digital content abounding, there is no storage of unstructured content. And there is unlimited potential in the value of this unstructured content if it can be harnessed. This duo will give brands that opportunity.

@DrNatalie Petouhoff, VP and Principal Analyst, Constellation Research

Covering Customer-facing Applications That Drive Better Business Results
Share