Saltar e ir al contenido
  • En la empresa
  • Soluciones Para Sectores Concretos
  • Valor Para Clientes Y Partners
  • Producto Y Tecnología
  • Liderazgo de Opinión
Languages
  • Deutsch
  • 한국어
  • Español
  • Français
  • English
  • Italiano
  • 日本語
  • Português
  • Deutsch
  • 한국어
  • Español
  • Français
  • English
  • Italiano
  • 日本語
  • Português
  • En la empresa
  • Soluciones Para Sectores Concretos
  • Valor Para Clientes Y Partners
  • Producto Y Tecnología
  • Liderazgo de Opinión
  • Deutsch
  • 한국어
  • Español
  • Français
  • English
  • Italiano
  • 日本語
  • Português
  • Visión General
    • Por qué Snowflake
    • Historias de clientes
    • Partners
    • Servicios
  • Visión General
    • Data Cloud
    • Descripción general de la plataforma
    • Snowflake Marketplace
    • Desarrollado por Snowflake
    • Demostración en directo
  • Workloads
    • Collaboration
    • Data Science & ML
    • Cybersecurity
    • Applications
    • Data Warehouse
    • Data Lake
    • Data Engineering
    • Unistore
  • Precios
    • Opciones de precios
  • Por sector
    • Publicidad, medios de comunicación y entretenimiento
    • Servicios financieros
    • Sanidad y ciencias de la vida
    • Fabricación
    • Sector Público
    • Retail / CPG
    • Tecnología
  • Conozca
    • Biblioteca de recursos
    • Documentación
    • Demostración en directo
    • Formación
  • Conectar
    • Blog
    • Comunidad
    • Eventos
    • Webinars
    • Podcast
  • Visión General
    • Sobre Snowflake
    • Los inversores
    • Liderazgo y dirección
    • Carreras
Autor
Arun Agarwal
Arun Agarwal
Chase Ginther
Chase Ginther
Harshal Pimpalkhute
Harshal Pimpalkhute
Share
Subscribe
Jul 23, 2024

Meta’s Llama 3.1 405B Now Available for Enterprise App Development in Snowflake Cortex AI

  • Producto y tecnología
    • IA y ML:
Meta’s Llama 3.1 405B Now Available for Enterprise App Development in Snowflake Cortex AI

Today, Snowflake is excited to announce that the Llama 3.1 collection of multilingual large language models (LLMs) are now available in Snowflake Cortex AI, providing enterprises with secure, serverless access to Meta’s most advanced open source model. Snowflake offers the largest context window of any vendor, at 128k, for the Llama 3.1 collection of models. 

Using Cortex AI, Snowflake customers can securely use their data with Meta’s newest open source models for AI app development. Llama 3.1 405B, along with 8B and 70B models, are now available for serverless inference, with fine-tuning for the entire Llama 3.1 collection coming soon. Additionally, the Llama 3 8B and 70B models will continue to be available for both serverless inference and fine-tuning (public preview). 

Built with top-quality trust and safety features for reliable, secure use, the newest Llama 3.1 collection of multilingual LLMs makes cutting-edge technology easily accessible, ready to serve a vibrant community of generative AI users with broad platform support.

Llama 3.1 405B at a Glance

Llama is an accessible, open LLM designed for developers, researchers and businesses to build, experiment and responsibly scale their generative AI ideas. Llama 3.1 405B is the first open source model that performs on par with leading proprietary AI models in general knowledge, steerability, math, tool use and multilingual translation, among other capabilities. 

Based on its enhanced abilities, Llama 3.1 405B should be evaluated for use cases such as:

  • Long-document processing: With the larger context window of 128k tokens, customers can summarize, analyze and run other natural language processing tasks without having to break up or chunk long documents. 
  • Advanced multilingual apps: Models are optimized for dialogue across 10 languages, providing a robust solution for apps that use several different languages, including Spanish, Portuguese, Italian and German.
  • Synthetic data generation: Customers can generate synthetic data and apply it in post-training to improve smaller Llama models.
  • Distillation recipe: Customers can use distillation to create small, efficient models from large ones like Llama 3.1 405B. Access to scripts and notebooks, along with detailed methods and examples, allows for self-distillation. These smaller models offer comparable performance at lower costs and reduced latency, ideal for resource-constrained environments.

The new collection of Llama 3.1 models is truly community-focused, with a broad reach across the ecosystem. Llama models have been downloaded hundreds of millions of times and are supported by thousands of community projects. From cloud providers to startups, the world is building with Llama, solving complex enterprise use cases globally.

Minimizing risk with robust safety features

As Snowflake continues to provide AI tools to its enterprise customers, we have learned that making generative AI easy to use is pointless if it isn’t safe. As generative AI applications move into production and user bases expand, the risk of harmful interactions increases. Implementing robust safety features without compromising scalability or increasing costs is essential. We are excited to share that the Llama 3.1 model collection is dedicated to trust and safety, releasing models and tools that promote community collaboration and standardize safety practices in generative AI.

“Harnessing generative AI depends on safety and trust, and the Snowflake platform provides us with the necessary assurances to innovate and leverage industry-leading large language models at scale,” said Ryan Klapper, an AI leader at Hakkoda. “Meta’s Llama models within Snowflake Cortex AI are a powerful combination that will unlock even more opportunities for us to service internal RAG-based applications. Through these applications, stakeholders can more seamlessly interact with internal knowledge centers, ensuring that they have access to accurate and relevant information whenever needed.”

For more details on model benchmarks and other use cases, visit the Llama 3.1 405B announcement from Meta. 

Snowflake open sources its optimizations for fine-tuning and long context inference

To deliver efficient and cost-effective inference and fine-tuning of 100+ billion-parameter models, the Snowflake AI Research team has been working on a stack for massive inference and fine-tuning used in Cortex AI. Today, with the launch of Llama 3.1 405B from Meta, this effort has become more relevant than ever. 

Meta’s OSS release of the Llama 3.1 collection inspired our Snowflake AI Research team to open source more of our innovations. To help the AI community leverage and build on top of this powerful but massive model, we are open sourcing the Massive LLM Inference System Optimization and the Massive LLM Fine-Tuning System Optimization stacks.

You can learn more about these in Snowflake’s engineering blog on Inference System Optimization and Fine-tuning System Optimization.

Build using Llama 3.1 405B: With Cortex AI you can run inference on your data in tables using familiar SQL syntax. Upload a sample table (e.g., customer survey results) with the relevant columns, such as Customer ID, Customer Name and Customer Feedback. You can use the following SQL function to create a summary column of the Customer Feedback.

Availability: The Llama 3.1 405B model is available in AWS us-west-2, and it will be accessible to more customers in the coming weeks. You can access the availability matrix for future updates. For more information on Cortex AI, including available models and usage instructions, please refer to Snowflake’s documentation center. 

We are excited to see how you use Llama 3.1 405B to build AI applications for your specific use cases. Visit the Snowflake Cortex AI homepage to learn more and try it for yourself with our quickstart guide. 

Share

Related Content

  • Producto y tecnología
    • IA y ML:
May 10, 2024

Snowflake Arctic, el mejor LLM para IA empresarial: inteligente de manera eficiente y verdaderamente abierto

Utilizar large language models (LLM) para crear inteligencia empresarial de la máxima calidad siempre ha consumido muchos recursos y ha resultado tan prohibitivo que, a menudo, cuesta entre decenas y…

Explore
Leer más
  • Producto y tecnología
    • IA y ML:
May 16, 2024

Snowflake Arctic Cookbook Series: Instruction-Tuning Arctic 

On April 24, we released Snowflake Arctic with a key goal in mind: to be…

Find Out How
Leer más

Get Snowflake’s Guide
To Gen AI and LLMs

Download Here

Snowflake Inc.
  • PLATAFORMA
    • Data Cloud Platform
    • Arquitectura
    • Precios
    • Data Marketplace
  • SOLUCIONES
    • Snowflake para sanidad y ciencias de la salud
    • Snowflake para servicios financieros
    • Snowflake para Marketing Analytics
    • Snowflake para el comercio minorista
    • Snowflake para el sector educativo
  • Recursos
    • Biblioteca de recursos
    • Webinars
    • Documentación
    • Comunidad
    • Asuntos legales
  • Explorar
    • Noticias
    • Blog
    • Tendencias
  • Acerca de
    • Acerca de Snowflake
    • Liderazgo y dirección
    • Partners
    • Empleo
    • Contacto

Sign up for Snowflake Communications

Thanks for signing up!

  • Privacy Notice
  • Site Terms
  • Cookie Settings

© 2024 Snowflake Inc. All Rights Reserved