Cookie Consent by Free Privacy Policy Generator ๐Ÿ“Œ TensorFlow 2 MLPerf submissions demonstrate best-in-class performance on Google Cloud

๐Ÿ  Team IT Security News

TSecurity.de ist eine Online-Plattform, die sich auf die Bereitstellung von Informationen,alle 15 Minuten neuste Nachrichten, Bildungsressourcen und Dienstleistungen rund um das Thema IT-Sicherheit spezialisiert hat.
Ob es sich um aktuelle Nachrichten, Fachartikel, Blogbeitrรคge, Webinare, Tutorials, oder Tipps & Tricks handelt, TSecurity.de bietet seinen Nutzern einen umfassenden รœberblick รผber die wichtigsten Aspekte der IT-Sicherheit in einer sich stรคndig verรคndernden digitalen Welt.

16.12.2023 - TIP: Wer den Cookie Consent Banner akzeptiert, kann z.B. von Englisch nach Deutsch รผbersetzen, erst Englisch auswรคhlen dann wieder Deutsch!

Google Android Playstore Download Button fรผr Team IT Security



๐Ÿ“š TensorFlow 2 MLPerf submissions demonstrate best-in-class performance on Google Cloud


๐Ÿ’ก Newskategorie: AI Videos
๐Ÿ”— Quelle: blog.tensorflow.org

Posted by Pankaj Kanwar, Peter Brandt, and Zongwei Zhou from the TensorFlow Team

MLPerf, the industry standard for measuring machine learning performance, has released the latest benchmark results from the MLPerf Training v0.7 round. Weโ€™re happy to share that Googleโ€™s submissions demonstrate leading top-line performance (fastest time to reach target quality), with the ability to scale up to 4,000+ accelerators and the flexibility of the TensorFlow 2 developer experience on Google Cloud.

In this blog post, weโ€™ll explore the TensorFlow 2 MLPerf submissions, which showcase how enterprises can run valuable workloads that MLPerf represents on cutting-edge ML accelerators in Google Cloud, including widely deployed generations of GPUs and Cloud TPUs. Our accompanying blog post highlights our record-setting large-scale training results.

TensorFlow 2: designed for performance and usability

At the TensorFlow Developer Summit earlier this year, we highlighted that TensorFlow 2 would emphasize usability and real-world performance. When competing to win benchmarks, engineers have often relied on low-level API calls and hardware-specific code that may not be practical in everyday enterprise settings. With TensorFlow 2, we aim to provide high performance out of the box with more straightforward code, avoiding the significant issues that low-level optimizations can cause with respect to code reusability, code health, and engineering productivity.
Time to converge (in minutes) using Google Cloud VMs with 8 NVIDIA V100 GPUs from Googleโ€™s MLPerf Training v0.7 Closed submission in the โ€œAvailableโ€ category.
TensorFlowโ€™s Keras APIs (see this collection of guides) offer usability and portability across a wide array of hardware architectures. For example, model developers can use the Keras mixed precision API and Distribution Strategy API to enable the same codebase to run on multiple hardware platforms with minimal friction. Googleโ€™s MLPerf submissions in the Available-in-Cloud category were implemented using these APIs. These submissions demonstrate that near-identical TensorFlow code written using high level Keras APIs can deliver high performance across the two leading widely-available ML accelerator platforms in the industry: NVIDIAโ€™s V100 GPUs and Googleโ€™s Cloud TPU v3 Pods.

Note: All results shown in the charts are retrieved from www.mlperf.org on July 29, 2020. MLPerf name and logo are trademarks. See www.mlperf.org for more information. Results shown: 0.7-1 and 0.7-2.
Time to convergence (in minutes) using Google Cloud TPU v3 Pod slices containing 16 TPU chips from Googleโ€™s MLPerf Training v0.7 Closed submission in the โ€œAvailableโ€ category.

Looking under the hood: performance enhancements with XLA

Googleโ€™s submissions on GPUs and on Cloud TPU Pods leverage the XLA compiler to optimize TensorFlow performance. XLA is a core part of the TPU compiler stack, and it can optionally be enabled for GPU. XLA is a graph-based just-in-time compiler that performs a variety of different types of whole-program optimizations, including extensive fusion of ML operations.

Operator fusion reduces the memory capacity and bandwidth requirements for ML models. Furthermore, fusion reduces the launch overhead of operations, particularly on GPUs. Overall, XLA optimizations are general, portable, interoperate well with cuDNN and cuBLAS libraries, and can often provide a compelling alternative to writing low-level kernels by hand.

Googleโ€™s TensorFlow 2 submissions in the Available-in-Cloud category use the @tf.function API introduced in TensorFlow 2.0. The @tf.function API offers a simple way to enable XLA selectively, providing fine-grained control over exactly which functions will be compiled.

The performance improvements delivered by XLA are impressive: on a Google Cloud VM with 8 Volta V100 GPUs attached (each with 16 GB of GPU memory), XLA boosts BERT training throughput from 23.1 sequences per second to 168 sequences per second, a ~7x improvement. XLA also increases the runnable batch size per GPU by 5X. Reduced memory usage by XLA also enables advanced training techniques such as gradient accumulation.
Impact of enabling XLA (in minutes) on the BERT model using 8 V100 GPUs on Google Cloud as demonstrated by Googleโ€™s MLPerf Training 0.7 Closed submission compared to unverified MLPerf results on the same system with optimization(s) disabled.

State-of-the-art accelerators on Google Cloud

Google Cloud is the only public-cloud platform that provides access to both state-of-the-art GPUs and Cloud TPUs, which allows AI researchers and data scientists the freedom to choose the right hardware for every task.

Cutting-edge models such as BERT, which are extensively used within Google and industry-wide for a variety of natural language processing tasks, can now be trained on Google Cloud leveraging the same infrastructure that is used for training internal workloads within Google. Using Google Cloud, you can train BERT for 3 million sequences on a Cloud TPU v3 Pod slice with 16 TPU chips in under an hour at a total cost of under $32.

Conclusion

Google's MLPerf 0.7 Training submissions showcase the performance, usability, and portability of TensorFlow 2 across state-of-the-art ML accelerator hardware. Get started today with the usability and power of TensorFlow 2 on Google Cloud GPUs, Google Cloud TPUs, and TensorFlow Enterprise with Google Cloud Deep Learning VMs.

Acknowledgements

The MLPerf submission on GPUs is the result of a close collaboration with NVIDIA. Weโ€™d like to thank all engineers at NVIDIA who helped us with this submission.
...



๐Ÿ“Œ AI industryโ€™s performance benchmark, MLPerf, for the first time also measures the energy that machine learning consumes


๐Ÿ“ˆ 36.37 Punkte

๐Ÿ“Œ Train your TensorFlow model on Google Cloud using TensorFlow Cloud


๐Ÿ“ˆ 31.7 Punkte

๐Ÿ“Œ Google, Nvidia split top marks in MLPerf AI training benchmark


๐Ÿ“ˆ 31.05 Punkte

๐Ÿ“Œ Train TensorFlow models at cloud scale with TensorFlow Cloud | Demo


๐Ÿ“ˆ 29.7 Punkte

๐Ÿ“Œ Graphcore & MLPerf: AI-Prozessor schlรคgt Nvidia bei Preis und Leistung


๐Ÿ“ˆ 29.05 Punkte

๐Ÿ“Œ Neural Magic's sparsity, Nvidia's Hopper, and Alibaba's network among firsts in latest MLPerf AI benchmarks


๐Ÿ“ˆ 29.05 Punkte

๐Ÿ“Œ MLPerf Benchmark fรผr AI: Intel ist ein Gewinner, aber Nvidia gewinnt noch viel mehr


๐Ÿ“ˆ 29.05 Punkte

๐Ÿ“Œ Intel und Nvidia in MLPerf: In AMDs Abwesenheit debรผtiert H200 und Gaudi2 stichelt


๐Ÿ“ˆ 29.05 Punkte

๐Ÿ“Œ TensorFlow Enterprise: Productionizing TensorFlow with Google Cloud (TF Dev Summit '20)


๐Ÿ“ˆ 27.81 Punkte

๐Ÿ“Œ Und Microsoft so: Cloud, Cloud, Cloud, Cloud, Cloud, Cloud, Cloud


๐Ÿ“ˆ 27.22 Punkte

๐Ÿ“Œ 25 Best CDN Providers 2019 (sorted by best ent, best small biz, best budget and best free CDNs)


๐Ÿ“ˆ 23.5 Punkte

๐Ÿ“Œ CapOneMe โ€“ a vulnerable cloud environment to demonstrate the Capital One breach


๐Ÿ“ˆ 23.44 Punkte

๐Ÿ“Œ The Linux Foundation develops certification to enable cloud native pros demonstrate competence


๐Ÿ“ˆ 23.44 Punkte

๐Ÿ“Œ Optimizing Cloud Performance: An In-Depth Guide to Cloud Performance Testing and its Benefits


๐Ÿ“ˆ 22.41 Punkte

๐Ÿ“Œ Submissions now open: Indie games programs to help developers grow with Google Play


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Submissions now open: Indie games programs to help developers grow with Google Play


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Google To Stop Supporting Public URL Submissions To Its Search Index


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Google Asks Three Outside Antivirus Firms To Start Scanning Submissions To Android's Play Store


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Grow your game on Google Play with the Indie Games Accelerator - submissions now open


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Grow your game on Google Play with the Indie Games Accelerator - submissions now open


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Grow your game on Google Play with the Indie Games Accelerator - submissions now open


๐Ÿ“ˆ 22.2 Punkte

๐Ÿ“Œ Deep Learning: Tensorflow Lite wird noch kleiner als Tensorflow Mobile


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ TensorFlow 101 (Really Awesome Intro Into TensorFlow)


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Evaluating TensorFlow models with TensorFlow Model Analysis


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15195: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15210: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15204: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15201: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15205: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15198: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15203: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Medium CVE-2020-15207: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ TensorFlow Roadshow Videos available in Spanish [Videos de la gira global de TensorFlow disponibles en espaรฑol]


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ TensorFlow operation fusion in the TensorFlow Lite converter


๐Ÿ“ˆ 21.92 Punkte

๐Ÿ“Œ Low CVE-2020-15197: Tensorflow Tensorflow


๐Ÿ“ˆ 21.92 Punkte











matomo