Mobile AI Compute Engine (or MACE for short) is a deep learning inference framework optimized for
mobile heterogeneous computing platforms. The design focuses on the following
targets:
Performance
The runtime is highly optimized with NEON, OpenCL and Hexagon, and
Winograd algorithm is introduced to
speed up the convolution operations. Besides the fast inference speed, the
initialization part is also intensively optimized to be faster.
Power consumption
Chip dependent power options like big.LITTLE scheduling, Adreno GPU hints are
included as advanced APIs.
Responsiveness
UI responsiveness guarantee is sometimes obligatory when running a model.
Mechanism like automatically breaking OpenCL kernel into small units is
introduced to allow better preemption for the UI rendering task.
Memory usage and library footprint
Graph level memory allocation optimization and buffer reuse are supported.
The core library tries to keep minium external dependencies to keep the
library footprint small.
Model protection
Model protection is the highest priority feature from the beginning of
the design. Various techniques are introduced like coverting models to C++
code and literal obfuscations.
Platform coverage
A good coverage of recent Qualcomm, MediaTek, Pinecone and other ARM based
chips. CPU runtime is also compatible with most POSIX systems and
archetectures with limited performance.
MACE Model Zoo contains
several common neural networks and models which will be built daily against a list of mobile
phones. The benchmark result can be found in the CI result page.
Any kind of contributions are welcome. For bug reports, feature requests,
please just open an issue without any hesitance. For code contributions, it's
strongly suggested to open an issue for discussion first. For more details,
please refer to the contribution guide.
MACE depends on several open source projects located in
third_party directory. Particularly, we learned a lot from
the following projects during the development: