Wednesday, July 18, 2018
Search
  
Submit your own News for
inclusion in our Site.
Click here...
Breaking News
IBM Backs Stronghold Digital Stable Coin
Samsung Adds Deezer's HiFi Lossless Audio to Select Soundbar Models
Amazon.com Faced Problems During Prime Day
Pioneer to Release New Blu-ray Disc Burners For Long-term Archiving
Walmart to Use Microsoft's Cloud Tech
Micron and Intel to End Their 3D XPoint Joint Development Partnership
Samsung Announces First 8Gb LPDDR5 DRAM for 5G and AI-powered Mobile Applications
Roku Debuts Roku TV Wireless Speakers
Active Discussions
Which of these DVD media are the best, most durable?
How to back up a PS2 DL game
Copy a protected DVD?
roxio issues with xp pro
Help make DVDInfoPro better with dvdinfomantis!!!
menu making
Optiarc AD-7260S review
cdrw trouble
 Home > News > General Computing > AMD, In...
Last 7 Days News : SU MO TU WE TH FR SA All News

Tuesday, October 10, 2017
AMD, Intel, ARM, IBM and Others Support the Open Neural Network Exchange Format for AI


AMD, ARM, Huawei, IBM and Intel have announced their support for the Open Neural Network Exchange (ONNX) format, which was co-developed by Microsoft and Facebook in order to reduce friction for developing and deploying AI.

Introduced last month, the Open Neural Network Exchange (ONNX) format is a standard for representing deep learning models that enables models to be transferred between frameworks (PyTorch, Caffe2, and Cognitive Toolkit). ONNX is the first step toward an open ecosystem where AI developers can easily move between tools and choose the combination that is best for them.

Standardization is good for both the compute industry and for developers because it enables a level of interoperability between various products and frameworks, while streamlining the path from development to production.

By joining the project, Intel plans to further expand the choices developers have on top of frameworks powered by the Intel Nervana Graph library and deployment through the company's Deep Learning Deployment Toolkit.

Intel plans to enable users to convert ONNX models to and from Intel Nervana Graph models, giving users an even broader selection of choice in their deep learning toolkits.

Arm is already engaged to accelerate Caffe2 for its Arm Cortex-A CPUs as well as for Arm Mali GPU-based devices which currently use the Facebook application.



Previous
Next
QLC NAND Flash to Succeed TLC NAND Next Year        All News        Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging
Apple Enters into Deal Content With Spielberg, NBCUniversal: report     General Computing News      Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging

Get RSS feed Easy Print E-Mail this Message

Related News
Microsoft Calls for Public Regulation of AI Face Recognition Software
Facebook Launches AR Ads
Samsung Wins at Two Top Global AI Machine Reading Comprehension Challenges
Baidu Unveils High-Performance Kunlun AI Chip, AI Partnerships With Intel
AI Algorithms Crushed Human Players in Dota 2 Video Game
Intel to Bring Silicon-Based Security to AI and Blockchain Workloads
Nvidia Uses AI to Produce High-quality, 240fps Slow-motion Video From 30fps Source
IBM's AI Machine Learns to Debate Humans
Google Uses Deep Learning to Predict When a Patient Will Die
Intel to Showcase AI and HPC Demos at ISC
Deep Mind's Neural Scene Rendering System Predicts 3D Surroundings Using Its Own Sensors
Samsung Launches Fund to Invest in AI Startups

Most Popular News
 
Home | News | All News | Reviews | Articles | Guides | Download | Expert Area | Forum | Site Info
Site best viewed at 1024x768+ - CDRINFO.COM 1998-2018 - All rights reserved -
Privacy policy - Contact Us .