How to use KerasHub with Hugging Face

Estimated read time 2 min read

Post Content

​ Did you know you can run a PyTorch checkpoint on a JAX backend? In this video, Yufeng shows you how KerasHub allows you to mix and match model architectures with pre-trained weights from Hugging Face Hub seamlessly.

As the AI landscape expands, developers often struggle with framework lock-in. You might like a specific model architecture but find the perfect fine-tuned weights were trained using a different framework.

Join us as we break down the difference between “Model Architecture” and “Model Weights” and demonstrate the power of KerasHub. We will walk through three live coding examples using popular Large Language Models (LLMs) to show how you can leverage the massive Hugging Face ecosystem while utilizing your preferred backend (JAX, TensorFlow, or PyTorch).

Topics we’ll cover:
– The difference between “architecture” (code) and “weights” (checkpoints)
– How KerasHub bridges the gap between frameworks
– Loading a cybersecurity-focused Mistral model on JAX
– Running a fine-tuned Llama 3.1 checkpoint
– Deploying a multilingual Gemma translator

Resources:
Get started with KerasHub →https://goo.gle/3Yz15mh
Read the Blog Pos → https://goo.gle/3LHQZws
Check out the Code (Colab) → https://goo.gle/4sF3ucI
Keras Documentation →https://goo.gle/4qknlwh

Chapters:
0:00 Introduction: Mixing Architectures and Weights
0:50 Concept: Model Architecture vs. Model Weights
2:00 What is KerasHub?
2:27 Using Hugging Face Hub Checkpoints
3:12 Coding Demo: Setting the Backend
3:24 Demo: Running Mistral on JAX
4:03 Demo: Running Llama 3.1
4:28 Demo: Running Gemma for Translation
4:46 Summary & Conclusion

Subscribe to Google for Developers → https://goo.gle/developers

Speaker: Yufeng Guo,
Products Mentioned: Keras, Gem   Read More Google for Developers 

You May Also Like

More From Author