Learn more >
DeployableObject Detection in Images
Get this modelTry the API Try the web app Try in a Node-RED flow Try in CodePen
By IBM Developer Staff | Published December 12, 2018
Artificial intelligenceDeep learningVisual recognitionImage Feature ExtractionObject Detection in Images
This model detects humans and their poses in a given image. The model first detects the humans in the input image and then identifies the body parts, including nose, neck, eyes, shoulders, elbows, wrists, hips, knees, and ankles. Next, each pair of associated body parts is connected by a pose line. The pose lines are assembled into full body poses for each of the humans detected in the image. The model is based on the TF implementation of OpenPose model.
This model can be deployed using the following mechanisms:
docker run -it -p 5000:5000 codait/max-human-pose-estimator
kubectl apply -f https://raw.githubusercontent.com/IBM/MAX-Human-Pose-Estimator/master/max-human-pose-estimator.yaml
You can test or use this model
Once deployed, you can test the model from the command line. For example if running locally:
curl -F "image=@samples/p3.jpg" -XPOST http://localhost:5000/model/predict
You should see a JSON response like that below:
The information returned from the model can be used to construct and visualize pose lines for the humans detected in the image, such as shown in the example below. For more details see the GitHub README.
Complete the node-red-contrib-model-asset-exchange module setup instructions and import the human-pose-estimator getting started flow.
Learn how to send an image to the model and how to render the results in CodePen.
Get an overview of computer vision with deep learning and learn how it can help your applications recognize what an…
Artificial intelligenceDeep learning+
This learning path gives you an understanding and working knowledge of Watson Visual Recognition. It explains the basics of Visual…
This article walks you through the basics of the Watson Visual Recognition service, such as how to get credentials and…
Back to top