Slide 1

Slide 1 text

not smart, just human! Manu Rink Technical Evangelist @ A Microsoft Cognitive Services boot camp Source

Slide 2

Slide 2 text

Disclaimer Basic understanding of web based services and using them. Yes, we will see a bit of source code! ~LVL300 It’s not impolite to leave now – no kitten memes to miss here!

Slide 3

Slide 3 text

a bit of theory or “just every talk needs a bit of boring dust in front of the cool stuff”

Slide 4

Slide 4 text

Cognition is "the mental action or process of acquiring knowledge and understanding through thought, experience, and the senses." It encompasses processes such as knowledge, attention, memory and working memory, judgment and evaluation, reasoning and "computation", problem solving and decision making, comprehension and production of language. Source: https://en.wikipedia.org/wiki/Cognition

Slide 5

Slide 5 text

Let the API do the work for you… ... what do you want to do?

Slide 6

Slide 6 text

the Api or Use Cognitive Services to help humans face their challenges … with machines!

Slide 7

Slide 7 text

Let the API do the work for you… ... what do you want to do?

Slide 8

Slide 8 text

Computer Vision API Analyse Describe Thumbnail OCR [

Slide 9

Slide 9 text

Computer Vision API Documentation: https://www.microsoft.com/cognitive-services/en-us/computer-vision-api/documentation API Reference https://dev.projectoxford.ai/docs/services/56f91f2d778daf23d8ec6739/operations/56f91f2e778daf14a499e1fa Analyse an image https://api.projectoxford.ai/vision/v1.0/analyze[?visualFeatures][&details] URL parameters Visual features Categories, Tags, Description, Faces, ImageType, Color, Adult Details Currently just “Celebrities” is supported Header Content-Type application/json, application/octet-stream, multipart/form-data Ocp-Apim-Subscription-Key Get your key from “My account” at https://www.microsoft.com/cognitive-services/. You might have to create an account first.

Slide 10

Slide 10 text

Computer Vision API JSON result for analyzed image • Details “Celebrities” • Visual features “Categories, Tags, Description, Adult”

Slide 11

Slide 11 text

Computer Vision API 1.0 Describe an image https://api.projectoxford.ai/vision/v1.0/describe[?maxCandidates] URL parameters - maxCandidates Header - see “Analyse” Get Thumbnail https://api.projectoxford.ai/vision/v1.0/generateThumbnail[?width][&height][&smartCropping] URL parameters – width, height, smartCropping Header - see “Analyse” OCR https://api.projectoxford.ai/vision/v1.0/ocr[?language][&detectOrientation ] URL parameters – language, detectOrientation Header - see “Analyse”

Slide 12

Slide 12 text

Computer Vision API On the left: JSON result for OCR On the right: JSON result for Describe

Slide 13

Slide 13 text

Y A X B Demo Computer Vision API

Slide 14

Slide 14 text

Recognition in images Recognition in videos Recognition with rectangles [ Emotion API

Slide 15

Slide 15 text

emotion API beta Documentation: https://www.microsoft.com/cognitive-services/en-us/emotion-api/documentation API Reference https://dev.projectoxford.ai/docs/services/56f91f2d778daf23d8ec6739/operations/56f91f2e778daf14a499e1fa Emotion Recognition https://api.projectoxford.ai/emotion/v1.0/recognize Header Content-Type application/json, application/octet-stream, multipart/form-data Ocp-Apim-Subscription-Key Get your key from “My account” at https://www.microsoft.com/cognitive-services/. You might have to create an account first.

Slide 16

Slide 16 text

emotion API beta JSON result for Emotion Recognition of an image. For every detected face the API returns • the face rectangle • the list of emotions with scores

Slide 17

Slide 17 text

Emotion API beta Emotion Recognition in videos https://api.projectoxford.ai/emotion/v1.0/recognizeinvideo[?outputStyle] URL parameters – outputStyle [aggregate, perFrame] Header - see “Recognition” Result on 202 - video operation status/result as URL Emotion Recognition with Face Rectangles https://api.projectoxford.ai/emotion/v1.0/recognize?faceRectangles={faceRectangles} URL parameters – faceRectangles (left, top, width, height) Header - see “Recognition” Recognition in Video Operation Result https://api.projectoxford.ai/emotion/v1.0/operations/{oid}] URL parameters – oid (URL from Emotion Recognition in videos) Header - see “Recognition” Result: Status of recognition operation. On SUCCEEDED -> JSON can be retrieved from processingResult field. https://www.microsoft.com/cognitive-services/en-us/emotion-api/documentation/howtocallemotionforvideo

Slide 18

Slide 18 text

Y A X B Demo Emotion API

Slide 19

Slide 19 text

Face API Detect Verify Identify Find Similar [

Slide 20

Slide 20 text

Face API 1.0 Documentation: https://www.microsoft.com/cognitive-services/en-us/face-api/documentation/overview API Reference https://dev.projectoxford.ai/docs/services/563879b61984550e40cbbe8d/operations/563879b61984550f30395236 Detect Faces https://api.projectoxford.ai/face/v1.0/detect[?returnFaceId][&returnFaceLandmarks][&returnFaceAttributes] URL parameters returnFaceId faceId needed if face should later be attached to a person returnFaceLandmarks get position of e.g. eyes, pupils, nose, eyebrows,… returnFaceAttributes get attributes “age, gender, smile, facialHair, headPose, glasses” for a face Header Content-Type application/json, application/octet-stream, multipart/form-data Ocp-Apim-Subscription-Key Get your key from “My account” at https://www.microsoft.com/cognitive-services/. You might have to create an account first.

Slide 21

Slide 21 text

Face API 1.0 JSON result for Face Detection of an image. For every detected face the API returns • the faceId • the list of faceLandmarks • requested attributes of the face

Slide 22

Slide 22 text

Face API 1.0 Find Similar Faces https://api.projectoxford.ai/face/v1.0/findsimilars URL parameters – faceId, faceListId, faceIds, maxNumOfCandidatesReturned, mode [matchPerson, matchFace] Header - see “Detect” Verify a face https://api.projectoxford.ai/face/v1.0/verify Request Body Face2Face Verification: faceId1, faceId2 Face2Person Verification: faceId, personGroupId, personId Header - see “Detect” Identify a face https://api.projectoxford.ai/face/v1.0/identify Request Body – faceIds, personGroupId, maxNumOfCandidatesReturned, confidenceThreshold Header - see “Detect”

Slide 23

Slide 23 text

Face API 1.0 APIs for creating, populating, training and maintaining persons, person groups and face lists

Slide 24

Slide 24 text

Face API 1.0 So..., how exactly does the verification and identification of faces to persons work? [*13] [*14]

Slide 25

Slide 25 text

Y A X B Demo Face API

Slide 26

Slide 26 text

Putting itall together or Go build your own product!

Slide 27

Slide 27 text

See Hear Read Understand Speech People Intention Topics Video Indexer

Slide 28

Slide 28 text

No content

Slide 29

Slide 29 text

Full Feature API Support

Slide 30

Slide 30 text

Y A X B Demo Video Indexer

Slide 31

Slide 31 text

nearly the end or ...just one more thing!

Slide 32

Slide 32 text

Let’s get started make our tech more human! [*vid1] Disney Research created a telepresence robot which feels human in his interactions. Link to the paper: https://s3-us-west- 1.amazonaws.com/disneyresearch/wp- content/uploads/20160503162533/A-Hybrid- Hydrostatic-Transmission-and-Human-Safe- Haptic-Telepresence-Robot-Paper.pdf

Slide 33

Slide 33 text

How to get STARTed?!

Slide 34

Slide 34 text

if (! self.haveAzureAccount) { openWebpage(https://azure.microsoft.com ) } else { let pass = Manu.instance.grabCredit(“AzurePass”) self.azureAccount.redeem(pass) } openWebpage(https://azure.microsoft.com/en- us/services/cognitive-services) self.doEpicShit()

Slide 35

Slide 35 text

Merci :) Manu Rink Technical Evangelist [email protected] @codeprincess says