Joel Pitt: Building & Calling Face Services Through SingularityNETmachine-learning facial-recognition blockchain artificial-intelligence algorithms
Our team has been seeding SingularityNET with valuable face-related services, enabling both existing and emerging solutions.
SummaryThis post reviews our initial collection of human face-related services that are being launched on SingularityNET.The first half covers the capabilities of our services. The second half covers the interactions with the SingularityNET smart-contracts via our command line tool. This allows you to create and fund jobs with which to make calls to these example services.The code is in the face-services repository.
Seeding Our Network With Flexible Face-Related Services
Why are a bunch of human-face related services useful and interesting for SingularityNET?
The specific services that are being implemented are described below.Note: The images used are test images in the repository, with attribution in the README.
Face-Related Services Being Developed on SingularityNET
1. Face Detection
Given an RGB image, face detection returns bounding boxes where faces are found. This sounds like a simple problem, but there have been many approaches to solving this, each with their own trade-offs.
For a long time the Haar-cascade (available in OpenCV) was popular, but dlib also implements a HOG (histogram of gradients) detector, and more recently an convolutional neural network (CNN) detector. The latter is much more robust to non-frontal face orientation than previous methods, but requires more computation.
2. Face landmarks
A bounding box is a great start to working with faces — but there’s a lot of variety in face structure and where the parts of a face end up inside said bounding box. This could be due to rotation, different face structure, and the expression on the person’s face. That’s where landmarks come in!
Given an RGB image, and the bounding box for one or more faces, the goal is to find the pixel location of various key points or “landmarks” on the face, e.g. the corners of the eyes, the tip of the nose, etc. Shown in the image is a 68-point model, but dlib also has a 5-point model which is sufficient for some applications.
3. Face alignment
One thing you can do with landmarks is align the face to a canonical orientation. A bounding box doesn’t let you do this by itself because its edges are always axis aligned. Why is alignment useful? Things like face recognition work a lot better if face images are aligned, other techniques like 3d reconstruction can also benefit from this as it removes one source of variability and lets a model focus on what it’s good at — instead of also trying to learn how to rescale and orientate the face.
So, the summary is: Given an RGB image, and a bounding box where there is a face, find the landmarks and then use these to rotate and scale the face to a canonical size and orientation. The difference from the original image in this example isn’t substantial. But if we take this following image of a model posing (another image in the test set of creative commons images), it’s more clear that the rotation is corrected:
4. Face Recognition
So how can we tell apart the two people in our example image? Face recognition deals with the challenge of identity: Given an RGB image, and a bounding box where there is a face, we return a vector of 128 floats describing the identity of the face.
Note this doesn’t tell you who the person is, it just differentiates this face from other faces by mapping them to different locations in a 128 dimensional space. Faces belonging to different people should be further away in this 128-d space than two photos of the same person.Note— While facial recognition is powerful, democratizing AI through our decentralized network helps ensure machine learning won’t support covert surveillance.
SingularityNET Services as Wrappers
I want to be clear that this isn’t original research by SingularityNET (though we have plenty of that, like trying to learn invariants and unsupervised language learning). Instead, this is an example of taking existing algorithms and making them available on SingularityNET. This helps us build a rich ecosystem that covers both existing and emerging machine learning models and solutions.
Calling each of these algorithms has been made almost as easy as a function call thanks to the marvelous dlib library by Davis King, and the hard work of the many contributors to OpenCV. Rather than explain how to do these calls yourself from python, here are some guides that others have already written:
- Face detection with OpenCV and deep learning.
- Real-time facial landmark detection with OpenCV, Python, and dlib
- DLib also comes with great short and easy to follow python examples
The thing we’re interested in here is how to expose these calls on a network of AI services. Currently the alpha only supports JSON-RPC for direct communication with a service (there will be other options in future, and you can easily work around this if it’s too limiting — a topic for future post). Our task is to think about each function call and wrap it as a JSON-RPC method. Here is the find_face method from the face detection service:
The actual face detection and algorithm selection is wrapped in face_detect(...), so this is mostly about marshaling data.
About the most complicated thing here, assuming you’ve seen python async keyword, is the base64 decoding of the image.
If async and await are new to you, there are a few guides that are helpful (e.g. https://snarky.ca/how-the-heck-does-async-await-work-in-python-3-5/ ). In short, when you call an async function, it isn’t actually executed and instead you are returned a “coroutine”. It only gets executed when you pass it to an event loop, which is done in the below code (here jsonrpc_handler is the coroutine):
Now, the code here isn’t as simple as an example could be, because I also added support for a grpc interface and wanted to ensure the two servers played well together. If it’s unclear then a minimal but complete example is available in the alpha-service. This doesn’t require executing the event loop explicitly, because it only has a json-rpc interface and as such it can just rely on the aiohttp and jsonrpcserver libraries to manage the event loop for us.
Once you have you request and response cycle wrapped in a json-rpc method, you can get on to interacting with the blockchain…
We’ve made some recent improvements to our Alpha, and soon, we’ll be releasing a detailed tutorial on how to start using it for free. Stay tuned!
Our team believes that developer ergonomics are crucial to making SingularityNET successful, so while we currently need to call a CLI, and there is currently no schema defining the JSON-RPC endpoints, it’s a first step.
In the future, we’ll release improvements that allows developers to publish a specification for a service’s API, and an SDK that removes the need for the CLI to make service calls. The end goal is that it should be as easy as making a function call!
Be sure to join our Community Forum, which allows you to chat directly with our AI team, as well as developers and researchers from around the world.
Building & Calling Face Services Through SingularityNET was originally published in SingularityNET on Medium, where people are continuing the conversation by highlighting and responding to this story.
Original article was created by: Joel Pitt at blog.singularitynet.ioDisclaimer: This article should not be taken as, and is not intended to provide, investment advice. Please conduct your own thorough research before investing in any cryptocurrency or ICO.
Interested in Cryptocurrencies and ICO's?
Follow our telegram channel for daily cryptomarket reports!Join @cointrends
|1 week ago||blockchain|
|1 week ago||blockchain|
|2 weeks ago||machine-learning|
|2 weeks ago||machine-learning|
|2 weeks ago||artificial-intelligence|
|2 weeks ago||artificial-intelligence|
|1 month ago||algorithms|
|2 months ago||algorithms|
|10 months ago||facial-recognition|
Stay on top of Altcoins and ICO trends.
Subscribe to our free Weekly Cryptomarket report
Delivered once a week, strongly to your inbox.Subscribe to our mailing list
October 04, 2018
Liquid Democracy for Distributed AI SystemsCreating Reputation Consensus and Liquid Democracy based on an open ledger protocol, for distributed AI systems.IntroductionHistory has made it abundantly clear that an effective and reliable level of governance is a necessary element for any...From: Aigents with Anton Kolonin
October 02, 2018
If simulating the brain is proving tricky, why don’t we try decoding it?A Faith Misplaced“There’s a good reason the first flying machines weren’t mechanical bats: people tried that, and they were terrible.” — Dan RobitzskiIn the current AI Spring, many people and corporations...From: Ahmad Matyana
September 28, 2018
In the third part of this series, we focus on how to use Aigents Graph Analysis to visualize and better understand Blockchain and Social Networks.In part one of our four-part series, we started with an introductory overview of different types of graphs, referring to examples rendering these...From: Aigents with Anton Kolonin
September 25, 2018
We’re joining forces with UNESCO to teach children about technologies that will shape the coming decades.Roughly four decades ago, the first steps to create the “network of networks” were a reality — leading us to the World Wide Web of today. About a decade ago, the first iPhone was...From: Ibby Benali
September 20, 2018
The New KingmakersIn this first part of a three-part series, we discuss the strategic monopolization of AI talent by the giants of the tech industry and how an increase in speed, scale and ambition is resulting in an AI talent war never seen before.Et Tu, Github?Developers are the builders of...From: Arif Khan
September 17, 2018
In the second of a series of interviews with our SingularityNET team members, we speak to Tim Richmond from the marketing team and find out about his love for music, what it means for him to be part of SingularityNET’s community first philosophy and his journey with AI.Hi Tim! Tell us a bit...From: Akram Yacob