Share and discover more about AI with social posts from the community.huggingface/OpenAi
Maybe The Best LLM with Small Parameters under 34B
Peach-9B-8k-Roleplay
Peach-9B-8k-Roleplay is a chat large language model obtained by finetuning 01-ai/Yi-1.5-9B model on more than 100K conversations created through our data synthesis approach.
How to start
The version of Transformers we are using is as follows, but a newer version may be available.

torch==1.13.1
gradio==3.50.2
transformers==4.37.2https://huggingface.co/ClosedCharacter/Peach-9B-8k-Roleplay ClosedCharacter/Peach-9B-8k-Roleplay · Hugging Face
XLabs AI is a part of an international company, a product laboratory where we strive to become leaders in machine learning and neural networks. The company develops and implements revolutionary solutions, setting new standards and inspiring to achieve the impossible in the field of information technology. Our team is an open, energized, and young collective that welcomes innovative ideas and supports the initiative and creativity of our employees.https://huggingface.co/XLabs-AI XLabs-AI (XLabs AI)
FLUX.1 Merged Models - Several different Schnell:Dev ratios
This repository includes merged models from black-forest-labs/FLUX.1-dev and black-forest-labs/FLUX.1-schnell, in different ratios. The licenses of both models apply to these merged models.

Inspired by: https://huggingface.co/sayakpaul/FLUX.1-merged

Motivation
The goal is to create modeld which balance generation speed - allowing near-Dev generations in more like 4-16 generations.

Results
FLUX.1 [dev] is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. For more information, please read our blog post.

Key Features
Cutting-edge output quality, second only to our state-of-the-art model FLUX.1 [pro].
Competitive prompt following, matching the performance of closed source alternatives .
Trained using guidance distillation, making FLUX.1 [dev] more efficient.
Open weights to drive new scientific research, and empower artists to develop innovative workflows.
Generated outputs can be used for personal, scientific, and commercial purposes as described in the flux-1-dev-non-commercial-license.
SAM2 Video Predictor
This is a simple demo for video segmentation with SAM2.

Instructions: (read the instructions)

Upload your video [MP4-24fps]
With 'include' point type selected, Click on the object to mask on first frame
Switch to 'exclude' point type if you want to specify an area to avoid
Get Mask !
Check Propagation every 15 frames
Add point on corresponding frame number if any mask needs to be refined
If propagation seems ok on every 15 frames, propagate with "render" to render final masked video !
Hit Reset button if you want to refresh and start again.
Input video will be processed over 10 seconds only for demo purpose :)https://huggingface.co/spaces/fffiloni/SAM2-Video-Predictor SAM2 Video Predictor - a Hugging Face Space by fffiloni
MiniCPM-V and OmniLMM are a family of open-source large multimodal models (LMMs) adept at vision & language modeling. The models process images and text inputs and deliver high-quality text outputs. We release two featured versions that are targeted at strong performance and efficient deployment:

MiniCPM-V 2.8B: State-of-the-art end-side large multimodal models. Our latest MiniCPM-V 2.0 can accept 1.8 million pixels (e.g., 1344x1344) images at any aspect ratio, and is adept at OCR capability. It achieves comparable performance with Gemini Pro in understanding scene-text and matches GPT-4V in preventing hallucinations.

OmniLMM 12B: The most capable version with leading performance among comparable-sized models on multiple benchmarks. The model also achieves state-of-the-art performance in trustworthy behaviors, with even less hallucination than GPT-4V.https://github.com/OpenBMB/MiniCPM-V/blob/8a1f766b85595a8095651eed9a44a83a965b305b/README_en.md#minicpm-v- MiniCPM-V/README_en.md at 8a1f766b85595a8095651eed9a44a83a965b305b · OpenBMB/MiniCPM-V
MiniCPM-V 2.8B is a strong multimodal large language model for efficient end-side deployment. The model is built based on SigLip-400M and MiniCPM-2.4B, connected by a perceiver resampler. Our latest version, MiniCPM-V 2.0 has several notable features.

🔥 State-of-the-art Performance.

MiniCPM-V 2.0 achieves state-of-the-art performance on multiple benchmarks (including OCRBench, TextVQA, MME, MMB, MathVista, etc) among models under 7B parameters. It even outperforms strong Qwen-VL-Chat 9.6B, CogVLM-Chat 17.4B, and Yi-VL 34B on OpenCompass, a comprehensive evaluation over 11 popular benchmarks. Notably, MiniCPM-V 2.0 shows strong OCR capability, achieving comparable performance to Gemini Pro in scene-text understanding, and state-of-the-art performance on OCRBench among open-source models.
A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
GitHub | Demo

MiniCPM-V 2.6
MiniCPM-V 2.6 is the latest and most capable model in the MiniCPM-V series. The model is built on SigLip-400M and Qwen2-7B with a total of 8B parameters. It exhibits a significant performance improvement over MiniCPM-Llama3-V 2.5, and introduces new features for multi-image and video understanding. Notable features of MiniCPM-V 2.6 include:https://github.com/OpenBMB/MiniCPM-V GitHub - OpenBMB/MiniCPM-V: MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
If you are interested in Knowledge Graphs, I invented all of this a year ago. It is a Encoder/Decoder that works with Knowledge Graphs. I am glad the world finally realizes this is useful a year later. I tried to tell you. I have not licensed any of the math. I own all of it. I do not have any plans to ever enforce the licensing but I like holding onto it.

https://huggingface.co/blog/TuringsSolutions/pfafresearch Probabilistic Fractal Activation Function (P-FAF) and Its Advantages Over Traditional Word Vectorization
FalconMamba 7B - a new model from TII (Technology Innovation Institute) is out !

- Blogpost: https://huggingface.co/blog/falconmamba
- Link to collection:
tiiuae/falconmamba-7b-66b9a580324dd1598b0f6d4a

- Link to playground:
tiiuae/falcon-mamba-playground
Announcements and New Features
- 🥳 We just launched the Open Source survey: https://hf.co/oss-survey. Feel free to provide feedback and shape the future of our Open Source ecosystem. You can even get a special GitHub badge!
- New 🤗 Transformers documentation! It has dark mode, new style, quick search and more https://hf.co/docs/transformers
- 🤖 RL at HF!? Last week we published Snowball Fight, a Unity ML-Agents DRL environment hosted in the Hub! Try it here: https://bit.ly/3FYhchD
- we also have two new channels ⁠无访问权限 and ⁠无访问权限! 🎉
- and an accompanying blog post https://huggingface.co/blog/snowball-fight
- New blog post! Getting Started with Hugging Face Transformers for IPUs with Optimum https://hf.co/blog/graphcore-getting-started

Events
Many events coming soon!
- Dec 8 On Sentiments and Biases @Merve Noyan and Vincent from Rasa will talk about various challenges in NLP.
https://discord.gg/2ajRMS9N?event=917383152144637982
- Dec 8 & 9: HF will be at The AI Summit in NY. If you're around you should visit! https://newyork.theaisummit.com/
- Dec 10: @lewtun and @Merve Noyan will be answering your questions in a Question Answering task special workshop! https://discord.gg/aKQTbg8d?event=917368738616057896
-🔊 Dec 14: An audio study group is forming at ⁠audio-discuss! The first meetup will happen next week! Join if you want to learn about Automatic Speech Recognition, TTS, and more.
Announcements and New Features
- Training's energy consumption and CO2 emissions can now be included in model repos https://twitter.com/julien_c/status/1461701986886336516 🌏
- New improved token system https://twitter.com/SimonBrandeis/status/1461389412621819908?s=20 🔒
- Summary of talks from the course event https://huggingface.co/course/event/1?fw=pt with some very nice visuals 🎨
- If you missed it, FaceBook XLS-R is an audio model pretrained on 128 spoken languages. We wrote a guide about fine-tuning it for multi-lingual ASR https://huggingface.co/blog/fine-tune-xlsr-wav2vec2 🎙
- New 3 part video series on how to train a vision transformer with SageMaker https://twitter.com/julsimon/status/1463934344293236741
- Building a dataset from images stored in S3 https://youtu.be/jalopOoBL5M
- Train with Transformers https://youtu.be/iiw9dNG7JcU
- Train with PyTorch Lightning https://youtu.be/rjYV0kKHjBA
- New blog post about accelerating distributed training with Intel technologies 🔥 https://huggingface.co/blog/accelerating-pytorch

Upcoming Events: 🥳
- Nov 30: Implementing DietNeRF with JAX and Flax. Learn about NeRF, JAX+Flax, 3D reconstruction, HF Spaces, and more. See you at ⁠无访问权限 https://www.youtube.com/watch?v=A9iefUXkvQU
- Dec 4: We have a nice talk about Spaces at PyCon Indonesia https://pycon.id/.
- Dec 8: On Sentiments & Biases with @Merve and Vincent Warmerdam from Rasa. https://twitter.com/mervenoyann/status/1464162219357360135
- Dec 8: Accelerating Transformers Down to 1ms - To Infinity and Beyond! at The AI Summit https://newyork.theaisummit.com/. Julien Chaumond
Our course community event is going full speed right now! Check out ⁠无访问权限 and related channels to see some very cool projects going on 😎

Announcements and New Features
- New Activity Feed! If you log in, your home screen will show an activity feed and trending repos. Check it out 🔥
- Facebook/Meta just released XLS-R, a model pretrained on 128 spoken languages 🌍. Try out "All-to-All" Speech Translation https://huggingface.co/spaces/facebook/XLS-R-2B-22-16 and read about fine-tuning it in https://huggingface.co/blog/fine-tune-xlsr-wav2vec2.
- Korean GPT is now in the Hub https://huggingface.co/kakaobrain/kogpt 🇰🇷

Upcoming Events: 🥳
- Nov 30: Implementing DietNeRF with JAX and Flax www.youtube.com/watch?v=A9iefUXkvQU

Previous events:
- All course talks playlist: https://www.youtube.com/playlist?list=PLo2EIpI_JMQvcXKx5RFReyg6Qd2UICAif
- Search Like You Mean It: Semantic Search with NLP and a Vector Database https://hubs.ly/H0_r91Q0 XLS-R All-to-All 2B - a Hugging Face Space by facebook
Releasing HQQ Llama-3.1-70b 4-bit quantized version! Check it out at
mobiuslabsgmbh/Llama-3.1-70b-instruct_4bitgs64_hqq
.

Achieves 99% of the base model performance across various benchmarks! Details in the model card.
a new reward modelling dataset:
Avelina/UltraSteer-v0


UltraSteer-V0 is a massive collection of single- and multi-turn dialogue with fine-grained reward labels produced by Nvidia's
nvidia/Llama2-13B-SteerLM-RM
reward model. We have a total of 2.3M labelled sequences taken from high quality datasets with a total of 2.8M labelled turns each containing 9 attributes produced as is from the reward model.

This is still very much an early version of the dataset (but it's fully usable!) and an updated version will be on the way with a full paper.

I would really appreciate if people could take a look at the dataset and suggest any improvements (e.g. more data sources, different cleaning approaches, different label schema, etc) in the community section.
I'm excited to announce that Transformers.js V3 is finally available on NPM! 🔥 State-of-the-art Machine Learning for the web, now with WebGPU support! 🤯⚡️

Install it from NPM with:
𝚗𝚙𝚖 𝚒 @𝚑𝚞𝚐𝚐𝚒𝚗𝚐𝚏𝚊𝚌𝚎/𝚝𝚛𝚊𝚗𝚜𝚏𝚘𝚛𝚖𝚎𝚛𝚜

or via CDN, for example: https://v2.scrimba.com/s0lmm0qh1q

Segment Anything demo:
webml-community/segment-anything-webgpu