Friday, November 24, 2017
Search
  
Submit your own News for
inclusion in our Site.
Click here...
Breaking News
Samsung Foundry in Advanced Discussions With New Customers for 7nm Chips
Tesla Finished Installing the World's Largest Mega-battery in Australia Within 100 Days
Apple Applies for Patent on Foldable Display
HP Patches Code execution Bug in Enterprise Printers
YouTube Takes More Steps to Tackle Down Videos Inappropriate for Minors
Broadcom Considering Increasing Qualcomm Bid
Facebook Will Notify you if you Have Followed Russian Propaganda
Russia Threatens to Block Ads on Google
Active Discussions
Which of these DVD media are the best, most durable?
How to back up a PS2 DL game
Copy a protected DVD?
roxio issues with xp pro
Help make DVDInfoPro better with dvdinfomantis!!!
menu making
Optiarc AD-7260S review
cdrw trouble
 Home > News > General Computing > AMD, In...
Last 7 Days News : SU MO TU WE TH FR SA All News

Tuesday, October 10, 2017
AMD, Intel, ARM, IBM and Others Support the Open Neural Network Exchange Format for AI


AMD, ARM, Huawei, IBM and Intel have announced their support for the Open Neural Network Exchange (ONNX) format, which was co-developed by Microsoft and Facebook in order to reduce friction for developing and deploying AI.

Introduced last month, the Open Neural Network Exchange (ONNX) format is a standard for representing deep learning models that enables models to be transferred between frameworks (PyTorch, Caffe2, and Cognitive Toolkit). ONNX is the first step toward an open ecosystem where AI developers can easily move between tools and choose the combination that is best for them.

Standardization is good for both the compute industry and for developers because it enables a level of interoperability between various products and frameworks, while streamlining the path from development to production.

By joining the project, Intel plans to further expand the choices developers have on top of frameworks powered by the Intel Nervana Graph library and deployment through the company's Deep Learning Deployment Toolkit.

Intel plans to enable users to convert ONNX models to and from Intel Nervana Graph models, giving users an even broader selection of choice in their deep learning toolkits.

Arm is already engaged to accelerate Caffe2 for its Arm Cortex-A CPUs as well as for Arm Mali GPU-based devices which currently use the Facebook application.



Previous
Next
QLC NAND Flash to Succeed TLC NAND Next Year        All News        Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging
Apple Enters into Deal Content With Spielberg, NBCUniversal: report     General Computing News      Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging

Get RSS feed Easy Print E-Mail this Message

Related News
Intel Advances Artificial Intelligence With Nervana Neural Network Processor
Intel's New Loihi Self-Learning Chip Promises to Accelerate Artificial Intelligence
Facebook Opens AI Research in Montreal
IFA: Huawei Says it Has an Edge Over Apple With New Kirin 970 Mobile AI Chipset
Microsoft Unveils Project Brainwave for Real-time AI
Intel Introduces Movidius Neural Compute Stick
IBM Revamps Unit Around Watson Artificial Intelligence, Microsoft Creates New AI Lab
Microsoft Used AI to Help Crack Down on Tech Support Scams
Intel Invests in Three AI Companies
Microsoft AI Masters Pac-Man
LG Establishes New Robotics and AI research Divisions
Google to Retire the AlphaGo AI Program

Most Popular News
 
Home | News | All News | Reviews | Articles | Guides | Download | Expert Area | Forum | Site Info
Site best viewed at 1024x768+ - CDRINFO.COM 1998-2017 - All rights reserved -
Privacy policy - Contact Us .