Speed is all you need: On-device acceleration of large diffusion models via GPU-aware optimizations

Favorite Posted by Juhyun Lee and Raman Sarokin, Software Engineers, Core Systems & Experiences The proliferation of large diffusion models for image generation has led to a significant increase in model size and inference workloads. On-device ML inference in mobile environments requires meticulous performance optimization and consideration of trade-offs due

Read More
Shared by Google AI Technology June 15, 2023

Now is the time to define Open Source AI

Favorite You’ve seen so many announcements about new releases of AI large language models, foundational models, tooling, services all claiming to be “open” or “Open Source”. But none of them are shared with the same conditions. Look carefully and you’ll find that each adds some restriction somewhere. That’s far from

Read More
Shared by voicesofopensource June 14, 2023

8 ways Google Lens can help make your life easier

Favorite Lens makes it easy to search what you see and explore the world around you — including the new ability to search for skin conditions. View Original Source (blog.google/technology/ai/) Here.

Reconstructing indoor spaces with NeRF

Favorite Marcos Seefelder, Software Engineer, and Daniel Duckworth, Research Software Engineer, Google Research When choosing a venue, we often find ourselves with questions like the following: Does this restaurant have the right vibe for a date? Is there good outdoor seating? Are there enough screens to watch the game? While

Read More
Shared by Google AI Technology June 14, 2023

6 Gmail AI features to help save you time

Favorite These AI-powered Gmail features can make your email experience even faster, easier and more organized. View Original Source (blog.google/technology/ai/) Here.

AWS Inferentia2 builds on AWS Inferentia1 by delivering 4x higher throughput and 10x lower latency

Favorite The size of the machine learning (ML) models––large language models (LLMs) and foundation models (FMs)––is growing fast year-over-year, and these models need faster and more powerful accelerators, especially for generative AI. AWS Inferentia2 was designed from the ground up to deliver higher performance while lowering the cost of LLMs

Read More
Shared by AWS Machine Learning June 14, 2023