Machine learning framework for both deep learning and traditional algorithms
NeoML is an end-to-end machine learning framework that allows you to build, train, and deploy ML models. This framework is used by ABBYY engineers for computer vision and natural language processing tasks, including image preprocessing, classification, document layout analysis, OCR, and data extraction from structured and unstructured documents.
Key features:
The full С++ library version has been tested on the platforms:
Target OS | Compiler | Architecture |
---|---|---|
Windows 7+ (CPU and GPU) | MSVC 2019+ | x86, x86_64 |
Ubuntu 14+ (CPU) | gcc 5.4+ | x86_64 |
MacOS 10.11+ (CPU) | Apple clang 12+ | arm64, x86_64 |
iOS 11+ (CPU, GPU) | Apple clang 12+ | arm64-v8a, x86_64 |
Android 5.0+ (CPU), Android 7.0+ (GPU) | clang 7+ | armeabi-v7a, arm64-v8a, x86, x86_64 |
The inference Java and Objective-C library versions have been tested on the platforms:
Target OS | Compiler | Architecture |
---|---|---|
iOS 11+ (CPU, GPU) | Apple clang 12+ | arm64-v8a, x86_64 |
Android 5.0+ (CPU), Android 7.0+ (GPU) | clang 7+ | armeabi-v7a, arm64-v8a, x86, x86_64 |
The library is built with CMake (recommended versions 3.18 and later).
For best CPU performance on Windows, Linux and macOS we use Intel MKL.
When processing on a GPU, you can optionally use CUDA (version 11.2 upd.1) on Windows or Linux and Vulkan (version 1.1.130 and later) on Windows, Linux or Android.
We also use Google Test for testing and Google Protocol Buffers for working with ONNX model format.
We use very convinous generator of JIT code xbyak for speeding up some convolutions on x86_64 processors.
See here for instructions on building the C++ library version for different platforms.
See here for instructions on building the Java and Objective-C versions that would only run the trained neural networks.
Several tutorials with sample code will help you start working with the library:
The library was developed with these principles in mind:
The user interface is completely separated from the low-level calculations implemented by a math engine.
The only thing you have to do is to specify at the start the type of the math engine that will be used for calculations. You can also choose to select the math engine automatically, based on the device configuration detected.
The rest of your machine learning code will be the same regardless of the math engine you choose.
Each network works with one math engine instance, and all its layers should have been created with the same math engine. If you have chosen a GPU math engine, it will perform all calculations. This means you may not choose to use a CPU for “light” calculations like adding vectors and a GPU for “heavy” calculations like multiplying matrices. We have introduced this restriction to avoid unnecessary synchronizations and data exchange between devices.
The math engine interface is thread-safe; the same instance may be used in different networks and different threads.
Note that this may entail some synchronization overhead.
However, the neural network implementation is not thread-safe; the network may run only in one thread.
NeoML library also works with the models created by other frameworks, as long as they support the ONNX format. See the description of import API. However, you cannot export a NeoML-trained model into ONNX format.
The library uses its own binary format (implemented by CArchive
, CArchiveFile
) to save and load the trained models.
Processing on GPU often helps significantly improve performance of mathematical operations. The NeoML library uses GPU both for training and running the models. This is an optional setting and depends on the hardware and software capabilities of your system.
To work on GPU, the library requires:
The NeoML library originates in ABBYY internal infrastructure. For various reasons ABBYY uses a cross-platform framework called FineObj. Because of this, the open library version uses some of this framework primitives. See the common classes description.
NeoML contains two C++ libraries:
The library provides C++ objects that implement various high-level algorithms. It consists of several parts:
The math engine used for calculations is a separate module that implements the low-level mathematical functions used in the algorithms library. The user can also call these functions but usually never needs to.
This module has different implementations for different platforms. In particular, there is an implementation that uses a GPU for calculations.
The math engine is also a set of C++ interfaces described here.
See the extensive documentation of the Python module on readthedocs.io.
To work with the inference version of the library in Java and Kotlin we provide a Java interface.
To work with the inference version of the library in Swift and Objective-C we provide an Objective-C interface.
Copyright © 2016-2020 ABBYY Production LLC. Licensed under the Apache License, Version 2.0. See the license file.