Share and discover more about AI with social posts from the community.huggingface/OpenAi
Segment Anything 2 Demo-meta
SAM 2 from Meta FAIR is the first unified model for real-time, promptable object segmentation in images & videos. Using the model in our web-based demo you can segment, track and apply effects to objects in video in just a few clicks.
https://sam2.metademolab.com/
SAM 2 from Meta FAIR is the first unified model for real-time, promptable object segmentation in images & videos. Using the model in our web-based demo you can segment, track and apply effects to objects in video in just a few clicks.
https://sam2.metademolab.com/
Download the new lightweight Gemma 2 (2B) model weights from
@huggingface
. Explore the #Gemmaverse with new tools like ShieldGemma and Gemma Scope for safer and more interpretable AI → https://huggingface.co/google/gemma-2-2b?linkId=10571204
@huggingface
. Explore the #Gemmaverse with new tools like ShieldGemma and Gemma Scope for safer and more interpretable AI → https://huggingface.co/google/gemma-2-2b?linkId=10571204
Really cool to see that SF3D is trending on Huggingface. They created an amazing system for setting up the demos super easily and even extending Gradio was fairly straightforward - I’ve done a relightable viewer for it.
https://huggingface.co/spaces/stabilityai/stable-fast-3d and viewer https://pypi.org/project/gradio-litmodel3d/
https://huggingface.co/spaces/stabilityai/stable-fast-3d and viewer https://pypi.org/project/gradio-litmodel3d/
Introducing Idefics 3 8B Llama 3, Apache 2.0 licensed VLM with enhanced Document QA capabilities! ⚡️
> Vision backbone: SigLip, Text backbone: Llama 3.1 8B
> Text + Image input w/ text output
> 8.5B parameter model
> Supports up to 10K context
> Apache 2.0 licensed
> DocVQA
link:https://huggingface.co/HuggingFaceM4/Idefics3-8B-Llama3
> Vision backbone: SigLip, Text backbone: Llama 3.1 8B
> Text + Image input w/ text output
> 8.5B parameter model
> Supports up to 10K context
> Apache 2.0 licensed
> DocVQA
link:https://huggingface.co/HuggingFaceM4/Idefics3-8B-Llama3
New multimodal release: Idefics3!
Adding vision to Llama 3.1 8b 👀
Strong improvement over April's Idefics2: +14 points on DocVQA, +6 points on MathVista 🧠
Interleave up to 60 images with text! 🤯
Comparable performance to the unreleased Llama 3.1 8B multimodal 🦾
8B-parameters: runs natively in one A100 🤏
Open license: Apache 2.0 🤗
Transparent training data: Ethically sourced datasets, built for the community 🥳Use it today with our branch of transformers: https://huggingface.co/HuggingFaceM4/Idefics3-8B-Llama3
and our open weights:
Adding vision to Llama 3.1 8b 👀
Strong improvement over April's Idefics2: +14 points on DocVQA, +6 points on MathVista 🧠
Interleave up to 60 images with text! 🤯
Comparable performance to the unreleased Llama 3.1 8B multimodal 🦾
8B-parameters: runs natively in one A100 🤏
Open license: Apache 2.0 🤗
Transparent training data: Ethically sourced datasets, built for the community 🥳Use it today with our branch of transformers: https://huggingface.co/HuggingFaceM4/Idefics3-8B-Llama3
and our open weights:
Haiyan Zhang:Fortifying Teams with AI and Optimized Workflows
Last week, I had an opportunity to speak at SIGGRAPH, one of the computer graphics industry’s premier events that focuses on research, education, and skill development. I spoke with Munkhtsetseg Nandigjav, Associate Dean School of Animation & Motion at Savannah College of Art and Design, about my role as the General Manager for Gaming AI at Microsoft Gaming, our Responsible AI framework, and the ways that leaders can support their teams with AI to adapt to an ever-changing industry landscape.
Before I dive into the specifics of AI for Gaming and how I believe it can help change the industry we love for the better, I want to share a bit about my own background and why this matters so much to me.https://developer.microsoft.com/en-us/games/articles/2024/08/fortifying-teams-with-ai-and-optimized-workflows/
Last week, I had an opportunity to speak at SIGGRAPH, one of the computer graphics industry’s premier events that focuses on research, education, and skill development. I spoke with Munkhtsetseg Nandigjav, Associate Dean School of Animation & Motion at Savannah College of Art and Design, about my role as the General Manager for Gaming AI at Microsoft Gaming, our Responsible AI framework, and the ways that leaders can support their teams with AI to adapt to an ever-changing industry landscape.
Before I dive into the specifics of AI for Gaming and how I believe it can help change the industry we love for the better, I want to share a bit about my own background and why this matters so much to me.https://developer.microsoft.com/en-us/games/articles/2024/08/fortifying-teams-with-ai-and-optimized-workflows/
New strategies in fight against AI deepfakes form google
Google (NASDAQ: GOOGL) has introduced new policy updates to intensify its fight against artificial intelligence (AI)- generated content portraying individuals in explicit contexts without their permission.
In a statement, the tech giant disclosed that it will demote results of explicit deepfakes in Google Search to protect victims from bad actors amid a spike in offensive incidents. Google says the latest tools against deepfakes are an improvement on its existing policies with the most drastic change being the ease of filing complaints.
While victims have always enjoyed the right to request takedowns of non-consensual fake content from Google Search, the latest improvements allow for easy reporting of offensive websites. Google’s statement disclosed that the company will remove duplicates of the derogatory content on the web, building on its experiments with other illicit content.
“These efforts are designed to give people added peace of mind, especially if they’re concerned about similar content about them popping up in the future,” read the statement.
The second weapon in Google’s arsenal against deepfakes is an improvement in the Search ranking system. Google believes its decision to build systems to rank quality information at the top of Search may be “the best protection against harmful content.”
Going forward, the search giant unveiled plans to push AI-generated NSFW (not safe for work) content lower on its rankings to stifle its distribution. For searches involving specific names, Google says it will promote high-quality, non-explicit content to drown out the exposure to AI-generated deepfakes.
There are plans to outrightly demote websites that have a slew of reports against them for AI deepfakes, smothering their circulation and distribution from the source.
The combination of the features is poised to reduce incidents by up to 70%, but the company notes that the fight is far from finished. For now, Google continues to grapple with deepfakes that are consensual from those made without the approval of an individual, as search engines are unable to make the distinction.
“These changes are major updates to our protections on Search, but there’s more work to do to address this issue, and we’ll keep developing new solutions to help people affected by this content,” said Google.https://coingeek.com/google-unveils-new-strategies-in-fight-against-ai-deepfakes/
Google (NASDAQ: GOOGL) has introduced new policy updates to intensify its fight against artificial intelligence (AI)- generated content portraying individuals in explicit contexts without their permission.
In a statement, the tech giant disclosed that it will demote results of explicit deepfakes in Google Search to protect victims from bad actors amid a spike in offensive incidents. Google says the latest tools against deepfakes are an improvement on its existing policies with the most drastic change being the ease of filing complaints.
While victims have always enjoyed the right to request takedowns of non-consensual fake content from Google Search, the latest improvements allow for easy reporting of offensive websites. Google’s statement disclosed that the company will remove duplicates of the derogatory content on the web, building on its experiments with other illicit content.
“These efforts are designed to give people added peace of mind, especially if they’re concerned about similar content about them popping up in the future,” read the statement.
The second weapon in Google’s arsenal against deepfakes is an improvement in the Search ranking system. Google believes its decision to build systems to rank quality information at the top of Search may be “the best protection against harmful content.”
Going forward, the search giant unveiled plans to push AI-generated NSFW (not safe for work) content lower on its rankings to stifle its distribution. For searches involving specific names, Google says it will promote high-quality, non-explicit content to drown out the exposure to AI-generated deepfakes.
There are plans to outrightly demote websites that have a slew of reports against them for AI deepfakes, smothering their circulation and distribution from the source.
The combination of the features is poised to reduce incidents by up to 70%, but the company notes that the fight is far from finished. For now, Google continues to grapple with deepfakes that are consensual from those made without the approval of an individual, as search engines are unable to make the distinction.
“These changes are major updates to our protections on Search, but there’s more work to do to address this issue, and we’ll keep developing new solutions to help people affected by this content,” said Google.https://coingeek.com/google-unveils-new-strategies-in-fight-against-ai-deepfakes/
SwarmUI startup and creation speed
Maybe because it is based on ComfyUI, but SwarmUI startup and creation speed is quite fast
I can't simply compare it with the A1111 version of SD-webui, but for now, if you want to run SDXL, this is much more comfortable
Maybe because it is based on ComfyUI, but SwarmUI startup and creation speed is quite fast
I can't simply compare it with the A1111 version of SD-webui, but for now, if you want to run SDXL, this is much more comfortable
FLUX.1 [schnell] is a 12 billion parameter rectifier transformer
that generates images from text descriptions.
It has many features and uses, but also some limitations and prohibited uses.
Highlights
Powerful generation capabilities: can generate images from text descriptions, output quality is cutting-edge, prompt following is competitive, and generates high-quality images in 1 to 4 steps.
Training method: trained by latent adversarial diffusion distillation.
License and use: licensed under apache-2.0 for personal, scientific and commercial purposes, with reference implementation, sample code and API endpoints.
Usage: can be used through specific github repositories, ComfyUI, Diffusers, etc., and corresponding code examples are given.
Limitations: cannot provide factual information, may amplify social biases, may not match prompt generation output, prompt following is affected by prompt style.
Prohibited use: It cannot be used to violate laws and regulations, harm minors, generate false and harmful information, disseminate personally identifiable information, harass or bully others, create illegal content, make fully automated decisions that affect personal legal rights, or create false information on a large scale.
https://huggingface.co/black-forest-labs/FLUX.1-schnell
that generates images from text descriptions.
It has many features and uses, but also some limitations and prohibited uses.
Highlights
Powerful generation capabilities: can generate images from text descriptions, output quality is cutting-edge, prompt following is competitive, and generates high-quality images in 1 to 4 steps.
Training method: trained by latent adversarial diffusion distillation.
License and use: licensed under apache-2.0 for personal, scientific and commercial purposes, with reference implementation, sample code and API endpoints.
Usage: can be used through specific github repositories, ComfyUI, Diffusers, etc., and corresponding code examples are given.
Limitations: cannot provide factual information, may amplify social biases, may not match prompt generation output, prompt following is affected by prompt style.
Prohibited use: It cannot be used to violate laws and regulations, harm minors, generate false and harmful information, disseminate personally identifiable information, harass or bully others, create illegal content, make fully automated decisions that affect personal legal rights, or create false information on a large scale.
https://huggingface.co/black-forest-labs/FLUX.1-schnell
FLUX.1 [dev] is a 12 billion parameter rectifier transformer that can generate images from text descriptions.
Highlights
Powerful generation capabilities: It can generate images from text descriptions, and the output quality is cutting-edge, second only to the FLUX.1 [pro] model.
Excellent performance: It has competitive prompt following capabilities, matching the performance of closed-source alternatives.
Efficient training: Training is improved through guided distillation.
Open weight usage: Open weights to promote new scientific research and empower artists to develop innovative workflows.
Multiple usage paths: It provides reference implementations, sample code, can be obtained from multiple sources through API, can also be used for local reasoning in Comfy UI, and can be used with diffusers library.
Restrictions on use: It cannot provide factual information, may amplify social bias, may not match prompts, and is greatly affected by prompt style.
Prohibited uses: It clarifies a series of prohibited uses that violate laws and regulations, harm others, etc.
Followed license: Follows the FLUX.1 [dev] non-commercial license.
https://huggingface.co/black-forest-labs/FLUX.1-dev
Highlights
Powerful generation capabilities: It can generate images from text descriptions, and the output quality is cutting-edge, second only to the FLUX.1 [pro] model.
Excellent performance: It has competitive prompt following capabilities, matching the performance of closed-source alternatives.
Efficient training: Training is improved through guided distillation.
Open weight usage: Open weights to promote new scientific research and empower artists to develop innovative workflows.
Multiple usage paths: It provides reference implementations, sample code, can be obtained from multiple sources through API, can also be used for local reasoning in Comfy UI, and can be used with diffusers library.
Restrictions on use: It cannot provide factual information, may amplify social bias, may not match prompts, and is greatly affected by prompt style.
Prohibited uses: It clarifies a series of prohibited uses that violate laws and regulations, harm others, etc.
Followed license: Follows the FLUX.1 [dev] non-commercial license.
https://huggingface.co/black-forest-labs/FLUX.1-dev
this is amazing because Flux Schnell is, well, super-schnell (aka fast) and amazing at prompt following, while the IPA SDXL step gives it all the texture and style you would ever need
you can use this in the Glif Chrome Extension (pick the Flux Guided Style Transfer preset) or on glif:
https://chromewebstore.google.com/detail/glif-remix-the-web-with-a/abfbooehhdjcgmbmcpkcebcmpfnlingo
or here directly: https://glif.app/@fab1an/glifs/clzjsqemt00006mh13re8uu3b
you can use this in the Glif Chrome Extension (pick the Flux Guided Style Transfer preset) or on glif:
https://chromewebstore.google.com/detail/glif-remix-the-web-with-a/abfbooehhdjcgmbmcpkcebcmpfnlingo
or here directly: https://glif.app/@fab1an/glifs/clzjsqemt00006mh13re8uu3b
FLUX is incredible at prompt following, but can't do Style Transfer yet
here's a neat trick: you can use FLUX Schnell's gens as Controlnet into an SDXL workflow with an IPA + Style Ref, boom, you now both style and prompt following :)
some examples + link to a glif workflow:
here's a neat trick: you can use FLUX Schnell's gens as Controlnet into an SDXL workflow with an IPA + Style Ref, boom, you now both style and prompt following :)
some examples + link to a glif workflow:
CatVTON: A simple and most efficient virtual try-on diffusion model🤩
Lightweight (899.06M params only), Parameter-Efficient Training (49.57M parameters trainable), and Inference on less than 8G VRAM for 1024X768 resolution
Lightweight (899.06M params only), Parameter-Efficient Training (49.57M parameters trainable), and Inference on less than 8G VRAM for 1024X768 resolution
🚀 OV-DINO (Open-Vocabulary Detection with Instance-level Noise Optimization)
New approach to open-vocabulary object detection. It improves the ability of vision models to detect and identify objects in images, even objects outside training data.
🤩 SAM2 integration in Demo👇
New approach to open-vocabulary object detection. It improves the ability of vision models to detect and identify objects in images, even objects outside training data.
🤩 SAM2 integration in Demo👇
Is your summer reading list still empty? Curious if an LLM can generate a book blurb you'd enjoy and help build a KTO preference dataset at the same time?
A demo using
@huggingface
Spaces and
@gradio
to collect LLM output preferences: https://huggingface.co/spaces/davanstrien/would-you-read-it
A demo using
@huggingface
Spaces and
@gradio
to collect LLM output preferences: https://huggingface.co/spaces/davanstrien/would-you-read-it