Apple Now Has A Neural Network API

Unlock Apple's machine learning with a few lines of code

Apple's WWDC conference comes with a slew of new announcements and APIs.

WWDC 2016

Apple's WWDC conference comes with a slew of new announcements and APIs.Apple

Neural networks are all the rage recently—they're a way to get software to make decisions based on an overwhelming amount of data and the fundamental component of deep learning, which allows computers to recognize photos, speech, and text with unparalleled accuracy.

Apple, which has traditionally kept its artificial intelligence research under wraps, but now is allowing developers to build neural networks by calling on the company's simple API.

Those developing with Apple's neural networks, called Basic Neural Network Subroutines, won't be able to train on their own data. Instead Apple has pretrained them for certain tasks, and from the documentation the API seems very focused towards image recognition.

"For example, the initial input might be an image and the inference might be that it’s an image of a dinosaur," documentation says, in the only concrete use case given.

The API will run on macOS, iOS, tvOS, and watchOS, and is optimized for each device's CPU.

Follow all of our WWDC 2016 coverage here.