The Transform Technology Summits start October 13th with Low-Code/No Code: Enabling Enterprise Agility. Register now!
Xnor.ai today launched AI2Go, a platform that lets developers and manufacturers make prebuilt AI models optimized for on-device artificial intelligence. AI2Go is designed for state-of-the-art edge computing in devices like cameras, drones, and sensors.
The platform comes with hundreds of models made especially for smart home, security, auto, entertainment, and surveillance devices. The service was built to alleviate challenges that can arise when attempting to make AI for edge use cases — like latency, power consumption, or a limited amount of available memory.
Models can be made with a few clicks and lines of code and constraint settings can be tuned to manage things like memory usage. Models are also customized for various use cases and infused with an inference engine.
“With version zero, people can specify these constraints and get a model and download it. All of those models are already pretrained; [people] just need to grab it and use it,” Xnor CEO Ali Farhadi told VentureBeat in a phone interview. “Version 1 will enable functionalities to let people bring their own training data for custom models, and with the second version developers will be able to bring in already trained models and optimize them for the edge.”
Embedded AI has become an increasingly popular way to deploy intelligence without cloud or internet connection and to ensure user privacy. Smaller models can also allow developers and manufacturers to consider lower cost or commodity hardware for their devices.
Earlier this year, Xnor demonstrated that it can create a computer vision model small enough to fit on an FPGA chip powered by a single solar cell.
Xnor will continue to offer enterprise services for manufacturers and customers, and AI2Go models will come with free evaluation license agreements.
A number of hardware and software solutions for edge computing have been introduced in recent months, such as Nvidia’s Jetson Nano — its lowest-cost Jetson edge AI chip to date — and Googler’s TensorFlow Lite 1.0 for embedded devices, both of which arrived in March. Qualcomm introduced its Cloud AI 100 chip for edge inference in April.
VentureBeatVentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative technology and transact. Our site delivers essential information on data technologies and strategies to guide you as you lead your organizations. We invite you to become a member of our community, to access:
- up-to-date information on the subjects of interest to you
- our newsletters
- gated thought-leader content and discounted access to our prized events, such as Transform 2021: Learn More
- networking features, and more