Share and discover more about AI with social posts from the community.huggingface/OpenAi
New Flux Lora - Flat Colour Anime v3 : r/StableDiffusion
Hi all. I just released a new version of my Flat Color Anime lora updated for Flux Dev.

https://civitai.com/models/180891/flat-color-anime?modelVersionId=724241

This was something I worked on for SDXL a while ago to try and get a clean anime style similar to some of the 1.5 checkpoints I liked. I got the inspiration with Flux to have a go at updating it and I'm loving how the results are coming out so thought I'd share. I'll be refining it over the next days/weeks so any feedback is appreciated. Enjoy! https://www.reddit.com/r/StableDiffusion/comments/1esrzf1/new_flux_lora_flat_colour_anime_v3/https://preview.redd.it/new-flux-lora-flat-colour-anime-v3-v0-mm7fsj1c4tid1.png?width=832&format=png&auto=webp&s=7022ef6a073cfa9a7367f9ca47c7ca64cec72424 Flat Color Anime v3.1 - v3.1 Flux Dev | Stable Diffusion LoRA | Civitai
Online FLUX! fal.ai integrates ControlNet, providing online LoRA training
Exciting news for drawing enthusiasts! "FLUX Online Edition" at fal.ai has introduced a series of powerful new features, including important modules like ControlNet and LoRA. The best part? These features are ready to use straight out of the box, without the need for complex configurations.

Feature Highlights

ControlNet: This new feature makes it effortless to modify character expressions in images and even control objects within the scene.

LoRA Model Online Training: Users can train their own LoRA models online and share them via links, allowing creativity to spread widely.
The robust capabilities of FLUX, combined with third-party developed ControlNet and LoRA, make drawing more flexible and personalized. Netizens have already started creating a variety of styles with these tools, from cyberpunk to Chinese traditional, limited only by imagination.

The ease of use of FLUX Online Edition is one of the key reasons for its popularity. Users do not need to deploy models or set up workflows themselves; they can simply use these advanced drawing tools through the fal.ai platform.

fal.ai is one of the official online inference platforms cooperated with the FLUX development team, Black Forest Lab. Team member Jonathan Fischoff actively shares works created by users on the platform on Twitter, showcasing the diverse applications of FLUX and LoRA models.

Netizens have reacted enthusiastically to the new features of FLUX Online Edition, looking forward to combining ControlNet and LoRA with image-to-image functions to create more diverse visual effects.

Fischoff demonstrated images generated by FLUX using ControlNet contour control, and how to quickly generate new images with different styles by modifying prompt words.
The Imp project aims to provide a family of a strong multimodal small language models (MSLMs). Our imp-v1-3b is a strong MSLM with only 3B parameters, which is build upon a small yet powerful SLM Phi-2 (2.7B) and a powerful visual encoder SigLIP (0.4B), and trained on the LLaVA-v1.5 training set.

As shown in the image below, imp-v1-3b significantly outperforms the counterparts of similar model sizes, and even achieves slightly better performance than the strong LLaVA-7B model on various multimodal benchmarks.
https://huggingface.co/MILVLG/imp-v1-3b MILVLG/imp-v1-3b · Hugging Face
alvdansen
/
flux-koda
#StableDiffusion #Flux #AI #ComfyUI
Model description
Koda captures the nostalgic essence of early 1990s photography, evoking memories of disposable cameras and carefree travels. It specializes in creating images with a distinct vintage quality, characterized by slightly washed-out colors, soft focus, and the occasional light leak or film grain. The model excels at producing slice-of-life scenes that feel spontaneous and candid, as if plucked from a family photo album or a backpacker's travel diary.

Words that can highlight interesting nuances within the model:

kodachrome, blurry, realistic, still life, depth of field, scenery, no humans, monochrome, greyscale, traditional media, horizon, looking at viewer, light particles, shadow

https://cdn-uploads.huggingface.co/production/uploads/635dd6cd4fabde0df74aeae6/7CqMzFOlH6yoM-NpQdYDs.png
ostris / flux-dev-lora-trainer

Train ostris/flux-dev-lora-trainer
Trainings for this model run on Nvidia H100 GPU hardware, which costs $0.001528 per second.

If you haven’t yet trained a model on Replicate, we recommend you read one of the following guides.

Fine-tune an image model
Fine-tune SDXL with your own images
Create training
The easiest way to train ostris/flux-dev-lora-trainer is to use the form below. Upon creation, you will be redirected to the training detail page where you can monitor your training's progress, and eventually download the weights and run the trained model.https://replicate.com/ostris/flux-dev-lora-trainer/train ostris/flux-dev-lora-trainer – Replicate
The flux-lora-collection is a series of LoRA training checkpoints for the FLUX.1-dev model released by the XLabs AI team. This collection supports the generation of images in various styles and themes, such as anthropomorphism, anime, and Disney styles, offering high customizability and innovation.
https://huggingface.co/XLabs-AI/flux-lora-collection XLabs-AI/flux-lora-collection · Hugging Face
🚀 OV-DINO (Open-Vocabulary Detection with Instance-level Noise Optimization)

New approach to open-vocabulary object detection. It improves the ability of vision models to detect and identify objects in images, even objects outside training data.

🤩 SAM2 integration in Demo👇
VGGHeads

- Gradio demo: https://huggingface.co/spaces/okupyn/vgg_heads
- 5 visualization: 'Full', 'Head Boxes', 'Face Landmarks', 'Head Mesh', 'Head Pose'
- Model predicts various aspects of head detection & reconstruction perfectly
- VR, gaming, & animation uses
- 2D image inputs for 3D head outputs
📣 LLM-DetectAIve: Fine grained detection of machine-generated text

Classifies any text into 4 categories: Human-written, Machine-generated, Machine-written machine-humanized, and Human-written machine-polished💡😀

More nuanced compared to the current binary classification sota
Text-to-Single-ID Generation
🚀🚀🚀Quick start:
1. Enter a text prompt (Chinese or English), Upload an image with a face, and Click the Run button.
2. (Optional) You can also upload an image as the style reference for the results. 🤗
💡💡💡Tips:
1. Try to avoid creating too small faces, as this may lead to some artifacts. (Currently, the short side length of the generated image is limited to 512)
2. It's a good idea to upload multiple reference photos of your face to improve the prompt and ID consistency. Additional references can be uploaded in the "ID supplements".
3. The appropriate values of "Face ID Scale" and "Face Structure Scale" are important for balancing the ID and text alignment. We recommend using "Face ID Scale" (0.5~0.7) and "Face Structure Scale" (0.0~0.4). https://huggingface.co/spaces/Junjie96/UniPortrait UniPortrait - a Hugging Face Space by Junjie96
Introducing FalconMamba 7B: An attention-free 7B model which is pretty strong!

🤯Can process unlimited sequence lengths, outperforms traditional models, and fit on a single 24GB GPU.

Open-source and available on HF🤗. FalconMamba-7b Gradio Demo:https://huggingface.co/spaces/tiiuae/falcon-mamba-playground Falcon Mamba Playground - a Hugging Face Space by tiiuae
VGGHeads

🤯 Model performs simultaneous head detections and head meshes reconstruction (for multiple faces!) from a single image in a single step.

Runs on a CPU!🚀 More+Link👇https://pbs.twimg.com/media/GUyFx9WWgAAorFY?format=jpg&name=small
Online demos for BiRefNet on
@huggingface
Spaces!

Is this the best background removal model out there? 🤯
MIT licensed. 5.5G GPU memory needed for inference for 1024x1024 images.🤩
BiRefNet

🔥 Gradio Demo 1 with ImageSlider output: https://huggingface.co/spaces/not-lain/background-removal

Gradio demo 2 by the author 🙌
https://huggingface.co/spaces/ZhengPeng7/BiRefNet_demo Background Removal - a Hugging Face Space by not-lain
NEW and Hot: AuraSR Upscaler

- 600M parameter
- Based on GigaGAN paper from Adobe
- GANs are much faster than diffusion upscaling
- Upscaling to 1024px in 1/4th of a second

Model and Demo are up on Huggingface Hub. Great work by Fal AI and Gokay Aydogan, respectively.
🔥 AuraSR is a GAN-based Super-Res upscaler for generated images, a variation of the GigaGAN paper for image-conditioned upscaling.

Demo by @NONDA30: https://huggingface.co/spaces/gokaygokay/AuraSR

🔥 Torch implementation is based on the unofficial lucidrains/gigagan-pytorch repository: https://github.com/lucidrains/gigagan-pytorch?ref=blog.fal.ai AuraSR-v2 - a Hugging Face Space by gokaygokay
How to run Yi chat models with an API
Posted November 23, 2023 by @nateraw

The Yi series models are large language models trained from scratch by developers at 01.AI. Today, they’ve released two new models: Yi-6B-Chat and Yi-34B-Chat. These models extend the base models, Yi-6B and Yi-34B, and are fine-tuned for chat completion.

Yi-34B currently holds the state-of-the-art for most benchmarks, beating larger models like Llama-70B..
Run Code Llama 70B with an API
Posted January 30, 2024 by @cbh123

Code Llama is a code generation model built on top of Llama 2. It can generate code and natural language about code in many programming languages, including Python, JavaScript, TypeScript, C++, Java, PHP, C#, Bash and more.

Today, Meta announced a more powerful new version of Code Llama with 70 billion parameters. It’s one of the highest performing open models. Meta reports a 67.8 on HumanEval, which beats zero-shot GPT-4.

With Replicate, you can run Code Llama 70B in the cloud with one line of code.

Contents
Contents
Code Llama 70B variants
Run Code Llama 70B with JavaScript
Run Code Llama 70B with Python
Run Code Llama 70B with cURL
Keep up to speed
Code Llama 70B variants
There are three variants of Code Llama 70B. The code snippets in this guide use codellama-70b-instruct, but all three variants are available on Replicate:

Code Llama 70B Base is the foundation model.
Code Llama 70B Python is trained on Python code.
Code Llama 70B Instruct is fine-tuned for understanding natural language instructions.