AI อะไรเนี่ย

Tools

Accelerate LLM Fine-Tuning with Unstructured Data using SageMaker and S3

Accelerate LLM Fine-Tuning with Unstructured Data using SageMaker and S3

What's New: Seamless LLM Fine-Tuning with SageMaker and S3

AWS has unveiled a powerful integration between Amazon SageMaker Unified Studio and Amazon S3 general purpose buckets, designed to revolutionize how teams handle unstructured data for machine learning and data analytics. This update makes it significantly easier to leverage the vast amounts of unstructured data stored in S3, directly within your SageMaker workflows. It's a game-changer for data scientists and ML engineers looking to streamline their data pipelines and accelerate model development.

This integration simplifies what was once a complex process, enabling a more fluid and efficient journey from raw data to a production-ready model. By bringing your S3 data directly into SageMaker Unified Studio, you can spend less time on data wrangling and more time on actual model innovation.

Dive Deep: The Llama 3.2 VQA Example

To demonstrate the power of this new integration, AWS showcased a practical use case: fine-tuning the Llama 3.2 11B Vision Instruct model for visual question answering (VQA). This involves providing the model with an image and a question, then receiving an accurate answer, like identifying a transaction date from a receipt. The base Llama 3.2 11B Vision Instruct model, readily available via Amazon SageMaker JumpStart, already achieves an impressive Average Normalized Levenshtein Similarity (ANLS) score of 85.3% on the DocVQA dataset right out of the box.

To further boost performance for specific VQA tasks, fine-tuning was performed using the DocVQA dataset from Hugging Face, which boasts 39,500 rows of rich training data. During this process, three distinct fine-tuned model versions were created, utilizing varying dataset sizes: 1,000, 5,000, and 10,000 images. Amazon SageMaker's fully managed serverless MLflow was used throughout to meticulously track experimentation and evaluate the accuracy improvements of each version, ensuring transparent and reproducible results. For more details on this integration and demonstration, check out the Accelerating LLM Fine-Tuning with SageMaker Unified Studio and S3 blog post.

Streamlined Workflow and Key Requirements

One of the most compelling aspects of this solution is how Amazon SageMaker Unified Studio orchestrates the entire end-to-end process. From initial data ingestion and preprocessing to model development and metric evaluation, everything is managed within a unified environment. This holistic approach significantly reduces operational overhead and enhances collaboration among teams working on ML projects.

While the workflow is highly streamlined, it's important to note a key requirement for training jobs: they necessitate p4de.24xlarge compute instances. This often requires submitting a service quota increase request to AWS to ensure you have the necessary resources for your fine-tuning endeavors. This specific compute power is crucial for efficiently handling the large language models and extensive datasets involved in advanced fine-tuning. Learn more about the setup and workflow by visiting the AWS Machine Learning Blog.

Get Started Today

The new integration between Amazon SageMaker Unified Studio and S3 general purpose buckets marks a significant leap forward for anyone working with LLMs and unstructured data. By providing a streamlined, end-to-end platform for fine-tuning models like Llama 3.2, AWS empowers developers and data scientists to achieve higher accuracy and accelerate their AI initiatives. This is an essential tool for organizations looking to gain deeper insights from their vast data repositories.

Read more: Accelerating LLM Fine-Tuning with SageMaker Unified Studio and S3 and discover how to accelerate your LLM fine-tuning projects today.