Share and discover more about AI with social posts from the community.huggingface/OpenAi
I've been testing a new open-source AI image generation model, FLUX.1 [dev], today. I'm quite impressed! This could mark a goodbye to Midjourney and Stable Diffusion. The images were done locally on an RTX 3090, for around 40 seconds per image.
Thank you Robin&Black Forest Labs!
Databricks ❤️ Hugging Face: up to 40% faster training and tuning of Large Language Models
Generative AI has been taking the world by storm. As the data and AI company, we have been on this journey with the release of the open source large language model Dolly, as well as the internally crowdsourced dataset licensed for research and commercial use that we used to fine-tune it, the databricks-dolly-15k. Both the model and dataset are available on Hugging Face. We’ve learned a lot throughout this process, and today we’re excited to announce our first of many official commits to the Hugging Face codebase that allows users to easily create a Hugging Face Dataset from an Apache Spark dataframe.
https://huggingface.co/blog/databricks-case-study Databricks ❤️ Hugging Face: up to 40% faster training and tuning of Large Language Models
Rocket Money x Hugging Face: Scaling Volatile ML Models in Production
"We discovered that they were not just service providers, but partners who were invested in our goals and outcomes” - Nicolas Kuzak, Senior ML Engineer at Rocket Money.
Scaling and Maintaining ML Models in Production Without an MLOps Team
We created Rocket Money (a personal finance app formerly known as Truebill) to help users improve their financial wellbeing. Users link their bank accounts to the app which then classifies and categorizes their transactions, identifying recurring patterns to provide a consolidated, comprehensive view of their personal financial life. A critical stage of transaction processing is detecting known merchants and services, some of which Rocket Money can cancel and negotiate the cost of for members. This detection starts with the transformation of short, often truncated and cryptically formatted transaction strings into classes we can use to enrich our product experience.
https://huggingface.co/blog/rocketmoney-case-study Rocket Money x Hugging Face: Scaling Volatile ML Models in Production​
Fetch Consolidates AI Tools and Saves 30% Development Time with Hugging Face on AWS
xecutive Summary
Fetch, a consumer rewards company, developed about 15 different AI tools to help it receive, route, read, process, analyze, and store receipts uploaded by users. The company has more than 18 million active monthly users for its shopping rewards app. Fetch wanted to rebuild its AI-powered platform and, using Amazon Web Services (AWS) and with the support of AWS Partner Hugging Face, moved from using third-party applications to developing its own tools to gain better insights about customers. Consumers scan receipts —or forward electronic receipts— to receive rewards points for their purchases. Businesses can offer special rewards to users, such as extra points for purchasing a particular product. The company can now process more than 11 million receipts per day faster and gets better data.
https://huggingface.co/blog/fetch-eap-case-study Fetch Consolidates AI Tools and Saves 30% Development Time with Hugging Face on AWS
Ryght’s Journey to Empower Healthcare and Life Sciences with Expert Support from Hugging Face
Who is Ryght?
Ryght is building an enterprise-grade generative AI platform tailored for the healthcare and life sciences sectors. Today is their official launch of Ryght Preview, now publicly available for all.

Life science companies are amassing a wealth of data from diverse sources (lab data, EMR, genomics, claims, pharmacy, clinical, etc.), but analysis of that data is archaic, requiring large teams for everything from simple queries to developing useful ML models. There is huge demand for actionable knowledge to drive drug development, clinical trials, and commercial activity, and the rise of precision medicine is only accelerating this demand.

Ryght Laptophttps://huggingface.co/blog/ryght-case-study
Going multimodal: How Prezi is leveraging the Hub and the Expert Support Program to accelerate their ML roadmap
Everybody knows that a great visual is worth a thousand words. The team at Prezi, a visual communications software company, is putting this insight into practice with their Prezi presentations that combine images and text in highly dynamic presentations.
Prezi has joined the Hugging Face Expert Support Program to fully leverage modern machine learning's potential. Over the past months, Hugging Face has supported Prezi in integrating smaller, more efficient open-source models into their ML workflows. This cooperation started at a perfect time, as multimodal models are becoming increasingly capable.https://huggingface.co/blog/prezi-case-study Going multimodal: How Prezi is leveraging the Hub and the Expert Support Program to accelerate their ML roadmap
XLSCOUT Unveils ParaEmbed 2.0: a Powerful Embedding Model Tailored for Patents and IP with Expert Support from Hugging Face
XLSCOUT, a Toronto-based leader in the use of AI in intellectual property (IP), has developed a powerful proprietary embedding model called ParaEmbed 2.0 stemming from an ambitious collaboration with Hugging Face’s Expert Support Program. The collaboration focuses on applying state-of-the-art AI technologies and open-source models to enhance the understanding and analysis of complex patent documents including patent-specific terminology, context, and relationships. This allows XLSCOUT’s products to offer the best performance for drafting patent applications, patent invalidation searches, and ensuring ideas are novel compared to previously available patents and literature.https://huggingface.co/blog/xlscout-case-study XLSCOUT Unveils ParaEmbed 2.0: a Powerful Embedding Model Tailored for Patents and IP with Expert Support from Hugging Face
A SmolLM - blazingly fast and remarkably powerful
TL;DR
This blog post introduces SmolLM, a family of state-of-the-art small models with 135M, 360M, and 1.7B parameters, trained on a new high-quality dataset. It covers data curation, model evaluation, and usage.

Introduction
There is increasing interest in small language models that can operate on local devices. This trend involves techniques such as distillation or quantization to compress large models, as well as training small models from scratch on large datasets. These approaches enable novel applications while dramatically reducing inference costs and improving user privacy.

Microsoft's Phi series, Alibaba's Qwen2 (less than 2B), and Meta's MobileLLM demonstrate that small models can achieve impressive results when designed and trained thoughtfully. However, most of the details about the data curation and training of these models are not publicly available.https://huggingface.co/blog/smollm SmolLM - blazingly fast and remarkably powerful
Powerful ASR + diarization + speculative decoding with Hugging Face
Whisper is one of the best open source speech recognition models and definitely the one most widely used. Hugging Face Inference Endpoints make it very easy to deploy any Whisper model out of the box. However, if you’d like to introduce additional features, like a diarization pipeline to identify speakers, or assisted generation for speculative decoding, things get trickier. The reason is that you need to combine Whisper with additional models, while still exposing a single API endpoint.https://huggingface.co/blog/asr-diarization Powerful ASR + diarization + speculative decoding with Hugging Face Inference Endpoints
BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Model card for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone).

BLIP.gif
Pull figure from BLIP official repo
About-Open source Claude Artifacts – built with Llama 3.1 405B
Llama Coder
An open source Claude Artifacts – generate small apps with one prompt. Powered by Llama 3 405B & Together.ai.

Tech stack
Llama 3.1 405B from Meta for the LLM
Together AI for LLM inference
Sandpack for the code sandbox
Next.js app router with Tailwind
Helicone for observability
Plausible for website analyticshttps://github.com/nutlope/llamacoder
App is running again after a quick restart.

sourceoftruthdata/sot_autotrain_dreambooth_v1.1
This is the new SOTA image generation model that is super popular this week after auraflow! 😮
🌟Save traffic:
1. flux-schnell is a 4-step model, which can be used commercially. I think it can run with 16G video memory (welcome to correct); but is the quality improvement that big (roughly equivalent to SD3 medium?)
2. flux-dev is a 20-step model, which can be used by individuals, with super good quality, no need to write tags, can understand natural language, and the quality is close to MJ (about 90%?); FP16 model requires 24G video memory; FP8 model theoretically requires 12G video memory, but I haven't tested the actual usage and quality, welcome everyone's feedback.
3. flux-pro is not open source, only API calls.

black-forest-labs – Replicate

🧐black-forest-labs provides multiple advanced image generation models, including top prompt following, visual quality, image detail and output diversity models.

➡️Release link: https://replicate.com/black-forest-labs
➡️comfyui: https://comfyanonymous.github.io/ComfyUI_examples/flux/

Highlights
🌟 flux-pro: Cutting-edge image generation model with excellent prompt following and image quality.
⚡️ flux-schnell: The fastest image generation model designed for local development and personal use, with 74.6K runs.
🤖 flux-dev: A 12 billion parameter modified flow transformer for generating images from text descriptions, with 23.5K runs.
XLabs-AI/flux-controlnet-canny · Hugging Face
This repository provides a checkpoint with trained ControlNet Canny model for FLUX.1-dev model by Black Forest Labs

Training details
XLabs AI team is happy to publish fune-tuning Flux scripts, including:

LoRA 🔥
ControlNet 🔥
See our github for train script and train configs.

Training dataset
https://huggingface.co/XLabs-AI/flux-controlnet-canny?continueFlag=581a750ca897b60a8587e69e05615925 XLabs-AI/flux-controlnet-canny · Hugging Face
Flux-based controlnet model has been released!

XLabs-AI/flux-controlnet-canny · Hugging Face
🧐 This is a ControlNet Canny training checkpoint for the FLUX.1-dev model, supporting image generation tasks.
➡️Link: Web link
Key points
📦 This model is based on Black Forest Labs' FLUX.1-dev, and fine-tuning scripts for LoRA and ControlNet are provided.
📝 The training dataset contains images and corresponding JSON files with text prompts.
🖥 Inference usage examples include command line operations to support image generation based on prompts.
🔒 The model is distributed using the FLUX.1 [dev] non-commercial license.