Share and discover more about AI with social posts from the community.huggingface/OpenAi
HubSpot - Turn AI Into Your Personal Assistant
Ready to master the art of AI delegation?

Learn how to turn AI into your personal productivity powerhouse with HubSpot’s highly anticipated AI Task Delegation Playbook. It’s time to optimize your workflow like never before!

Unlock time-saving efficiencies.

Elevate your overall productivity.

Seamlessly streamline your workflow.

Get ready to save time and boost efficiency with easy-to-use templates and calculators.

Don’t miss out—grab your copy of HubSpot’s Playbook and start transforming your workday!

Check out HubSpot’s Playbook—-https://offers.hubspot.com/ai-delegation Free AI Task Delegation Playbook [Download Now]
Breaking News
The latest developments in AI

🚀 Google - A new tiny AI model, Gemma 2 2B, has been released challenging tech giants and even outperforming many of them. Alongside Gemma 2 2B, Google released ShieldGemma, a suite of safety content classifiers, and Gemma Scope, a model interpretability tool.

🎙 OpenAI - ChatGPT's new advanced voice mode impresses early users with diverse capabilities. Demos show it telling stories as an airline pilot with appropriate audio effects. While some accents aren't perfectly native, it handles interruptions well and can laugh or cry during conversations.

🎥 Runway - Gen-3 Alpha Turbo, a faster version of Runway's AI video model, has been unveiled. It's claimed to be 7x faster than the original while maintaining quality. The move is aimed at reducing costs, encouraging increased usage, and staying competitive in the AI video generation market.
Access expert-level advice about any topic

Step 1:

First, head over to Claude AI.

Note: You can use ChatGPT if you prefer.

Once you’re there, keep reading…

Step 2:

Next, you’ll need to create an account and/or log in.

Start a new chat and you’re ready to go!

One of the most common techniques for getting expert-like advice from AI is by asking it to “Act like [insert expert]”.

But how do you know who the right expert is?

The first thing to do is to give the chatbot a little background on what it’ll be helping you with, then add the final sentence as seen in this prompt:

I'm a German citizen living in the UK and working for a US company. Since I'm not an employee of the US company I work for and only freelance for them, I'm unsure how I should file my taxes and also how I would handle VAT. List the expert professionals best suited to deal with this issue.
You should get a list of 5-10 experts who would best be able to provide advice for your question.
Researchers at IISc have come up with a clever fix for AI's diversity problem in image generation. Their "attribute-aware" method is like a gentle nudge for biased algorithms, helping them create more inclusive pictures without a complete system overhaul. In tests, the AI adjusted gender ratios in professional images with impressive finesse. While it's not quite ready for the big leagues, this breakthrough could pave the way for fairer AI art down the road. It's a small step for code, but potentially a giant leap for AI representation. Looks like India's AI hub is making waves in the world of ethical AI! 🧔🏻‍♂️👩🏽https://timesofindia.indiatimes.com/
AI Tools of the week
1.📚 Opal - Transform study sessions with intelligent notes, flashcards, and quizzes that help absorb knowledge like a sponge. https://learnopal.com/

2.🎓 Thesify - Sharpen academic writing with AI tools that hunt references and distill articles, all while preserving scholarly integrity.

3.💡 BrainyBear - Build chatbots that absorb your business knowledge, turning your website into a 24/7 customer support hub.

4.🗨 Meta AI Studio - Create custom AI characters that chat like you, entertain audiences, or give form to your wildest imaginings.

5.💌 InstaBotGPT - Deploy an AI email assistant that mimics and replies in your voice, slashing response times from hours to minutes.

6.📊 ExcelDashboard AI - Turn Excel files into comprehensive dashboards and chat with your data for deeper insights.

7.🎨 AI Color Master - Elevate your designs with color palettes that capture emotions and optimize usability.

8.🔧 Yugo - Boost your web services by seamlessly connecting APIs with cutting-edge AI features.

9.🌐 AITranslator - Break language barriers with swift, expert AI translations for globally-minded businesses.

10.🧑‍💼 QSourcer - Uncover top-tier talent across LinkedIn, GitHub, and StackOverflow with AI-powered Boolean searches.

11.🎥 Video AI One - Streamline video creation by unifying multiple platforms to craft content from scripts, images, or extensions.

12.🏡 RapidREI - Accelerate real estate deals with AI that spots motivated sellers, crunches numbers, and generates smart offers.

13.🎞 Pix AI Video - Breathe life into stories with AI-crafted videos, blending visuals and rich audio to captivate your audience.

14.📖 AI Bedtime Stories for Kids - Craft custom bedtime stories for little ones, turning lights-out into an enchanting adventure.

15.🎨 CannyArt - Awaken your inner manga artist with AI doodles that blossom into anime art, zero artistic skills required. Opal - The #1 AI powered study tool
Guidde - Create how-to video guides with AI
Tired of explaining the same thing over and over again to your colleagues?

It’s time to delegate that work to AI. Guidde is a GPT-powered tool that helps you explain the most complex tasks in seconds with AI-generated documentation.

Share or embed your guide anywhere

Turn boring documentation into stunning visual guides

Save valuable time by creating video documentation 11x faster

Simply click capture on the browser extension and the app will automatically generate step-by-step video guides, complete with visuals, voiceover, and call to actions.

The best part? The extension is 100% free. https://www.guidde.com/
Running Gemma 2 2B at 41.66 tokens/s on my MacBook 💻🚀

- MLX Community's swift conversion
- One-line download from the Hub
- Small yet powerful on-device model

Try it yourself:
mlx-community/google-gemma2-667dca89bc9abbfa34080066


#GemmaAI #OnDeviceAI #MachineLearning https://cdn-uploads.huggingface.co/production/uploads/647f36a8454af0237bd49574/0Hx-zcve_R0aaM01b11BG.mp4
I’ve always wondered why holography hasn’t had much progress since its inception. Imagine what being able to harness and manipulate light with your bare hands in meaningful ways would be like: 3D photorealistic calls, truly immersive workspace. Given that it’s depicted in every futuristic scifi movie, one could not help but vision a future as such. This paper gives a clear overview why:

Turns out it’s incredibly difficult to compute and render photorealistic 3D data in real-time. The author claims immense computational power is needed for high data transmission rates, and compute of large number of phase pixels required for realistic 3D holography. The latest significant breakthrough in holography was 9years ago published in this paper - wherein they were able to achieve mid-air touchable/interactive 3D holography using a Femtosecond laser system considered safer than nanosecond lasers. Quite astounding work: arxiv.org/pdf/1506.06668

Realizing this breakthrough at scale is an unavoidably tempting research endeavor, super exciting especially with recent developments in machine learning and neural network algorithms demonstrating that computer-generated holograms can approach real-time processing. https://cdn-uploads.huggingface.co/production/uploads/6438a9027de34e8ea7e4b257/gjxRB-y9qSITf1iWIxU5F.qt
I can solve the Traveling Salesman Problem using the same methods the scientists used to solve it with 1 qubit, except I do not need quantum computers to do it. I am kind of tired of screaming this from the rooftops at this point. I can create an imaginary probability space, then I can put a bunch of imaginary agents in the imaginary box, and solve real problems in seconds. Problems that would take minutes, hours, or years to solve via other algorithms. Here is a demo of me solving the Traveling Salesman problem using 50 agents to probabilistically sample at once: https://colab.research.google.com/drive/1XplG72nQDO_-2h4DUllERLp0Dr2pI2J2?usp=sharing
Pic Smaller is a super easy-to-use online image compression tool. It's intuitive, mobile friendly, and supports compression configuration. At the same time, because of purely local compression without any server-side logic, it is completely safe.

https://github.com/joye61/pic-smaller
Revolutionizing Business Operations with Vidau.AI
Introducing Vidau.AI's Products
Vidau.AI is a leading provider of AI-powered products and services that are revolutionizing the way businesses operate. From advanced data analytics tools to machine learning solutions, Vidau.AI offers a comprehensive suite of products designed to drive efficiency and innovation in the digital age.

Operation Methods of Vidau.AI
The operation methods of Vidau.AI are designed to be user-centric and intuitive, ensuring a seamless experience for all users. With a focus on simplicity and functionality, Vidau.AI's platform allows businesses to harness the power of AI without unnecessary complexity.

Market Recognition of Vidau.AI's AI Solutions
Vidau.AI's AI solutions have garnered significant recognition in the market for their cutting-edge technology, reliability, and ability to drive tangible results for businesses. Industry experts and customers alike praise Vidau.AI for its commitment to pushing the boundaries of AI innovation.
https://www.saasinfopro.com/blog/VidauAI
Request for Quotation (RFQ)
Interested in leveraging Vidau.AI's AI solutions to transform your business operations? Submit a request for quotation today and unlock the full potential of AI for your organization.

Praise for Vidau.AI's Innovation
Vidau.AI's innovative AI solutions stand out for their ability to streamline processes, enhance decision-making, and drive business growth. Experience the future of business operations with Vidau.AI and stay ahead of the competition.
ComfyUI
The most powerful and modular stable diffusion GUI and backend.
ComfyUI Screenshot

This ui will let you design and execute advanced stable diffusion pipelines using a graph/nodes/flowchart based interface. For some workflow examples and see what ComfyUI can do you can check out:

ComfyUI Examples https://comfyanonymous.github.io/ComfyUI_examples/
Installing ComfyUI https://github.com/comfyanonymous/ComfyUI#installing
Usage from python:

from flux.api import ImageRequest

# this will create an api request directly but not block until the generation is finished
request = ImageRequest("A beautiful beach")
# or: request = ImageRequest("A beautiful beach", api_key="your_key_here")

# any of the following will block until the generation is finished
request.url
# -> https:<...>/sample.jpg
request.bytes
# -> b"..." bytes for the generated image
request.save("outputs/api.jpg")
# saves the sample to local storage
request.image
# -> a PIL image
Usage from the command line:

$ python -m flux.api --prompt="A beautiful beach" url
https:<...>/sample.jpg

# generate and save the result
$ python -m flux.api --prompt="A beautiful beach" save outputs/api

# open the image directly
$ python -m flux.api --prompt="A beautiful beach" image show black-forest-labs/flux-pro – Replicate
This repo contains minimal inference code to run text-to-image and image-to-image with our Flux latent rectified flow transformers.

Inference partners
We are happy to partner with Replicate and FAL. You can sample our models using their services. Below we list relevant links.

Replicate:

https://replicate.com/collections/flux
https://replicate.com/black-forest-labs/flux-pro
https://replicate.com/black-forest-labs/flux-dev
https://replicate.com/black-forest-labs/flux-schnell
FAL:

https://fal.ai/models/fal-ai/flux-pro
https://fal.ai/models/fal-ai/flux/dev
https://fal.ai/models/fal-ai/flux/schnell
Local installation
cd $HOME && git clone https://github.com/black-forest-labs/flux
cd $HOME/flux
python3.10 -m venv .venv
source .venv/bin/activate
pip install -e '.[all]'
Models
We are offering three models:

FLUX.1 [pro] the base model, available via API
FLUX.1 [dev] guidance-distilled variant
FLUX.1 [schnell] guidance and step-distilled variant
Name HuggingFace repo License md5sum
FLUX.1 [schnell] https://huggingface.co/black-forest-labs/FLUX.1-schnell apache-2.0 a9e1e277b9b16add186f38e3f5a34044
FLUX.1 [dev] https://huggingface.co/black-forest-labs/FLUX.1-dev FLUX.1-dev Non-Commercial License a6bd8c16dfc23db6aee2f63a2eba78c0
FLUX.1 [pro] Only available in our API.
The weights of the autoencoder are also released under apache-2.0 and can be found in either of the two HuggingFace repos above. They are the same for both models.

Usage
The weights will be downloaded automatically from HuggingFace once you start one of the demos. To download FLUX.1 [dev], you will need to be logged in, see here. If you have downloaded the model weights manually, you can specify the downloaded paths via environment-variables:

export FLUX_SCHNELL=<path_to_flux_schnell_sft_file>
export FLUX_DEV=<path_to_flux_dev_sft_file>
export AE=<path_to_ae_sft_file>
For interactive sampling run

python -m flux --name <name> --loop
Or to generate a single sample run

python -m flux --name <name> \
--height <height> --width <width> \
--prompt "<prompt>"
We also provide a streamlit demo that does both text-to-image and image-to-image. The demo can be run via

streamlit run demo_st.py
We also offer a Gradio-based demo for an interactive experience. To run the Gradio demo:

python demo_gr.py --name flux-schnell --device cuda
Options:

--name: Choose the model to use (options: "flux-schnell", "flux-dev")
--device: Specify the device to use (default: "cuda" if available, otherwise "cpu")
--offload: Offload model to CPU when not in use
--share: Create a public link to your demo
To run the demo with the dev model and create a public link:

python -m demo_gr.py --name flux-dev --share
Diffusers integration
FLUX.1 [schnell] and FLUX.1 [dev] are integrated with the 🧨 diffusers library. To use it with diffusers, install it:

pip install git+https://github.com/huggingface/diffusers.git
Then you can use FluxPipeline to run the model

import torch
from diffusers import FluxPipeline

model_id = "black-forest-labs/FLUX.1-schnell" #you can also use black-forest-labs/FLUX.1-dev

pipe = FluxPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", torch_dtype=torch.bfloat16)
pipe.enable_model_cpu_offload() #save some VRAM by offloading the model to CPU. Remove this if you have enough GPU power

prompt = "A cat holding a sign that says hello world"
seed = 42
image = pipe(
prompt,
output_type="pil",
num_inference_steps=4, #use a larger number if you are using [dev]
generator=torch.Generator("cpu").manual_seed(seed)
).images[0]
image.save("flux-schnell.png")
To learn more check out the diffusers documentation

API usage
Our API offers access to the pro model. It is documented here: docs.bfl.ml.

In this repository we also offer an easy python interface. To use this, you first need to register with the API on api.bfl.ml, and create a new API key.

To use the API key either run export BFL_API_KEY=<your_key_here> or provide it via the api_key=<your_key_here> parameter. Is is also expected that you have installed the package as above.
https://huggingface.co/mmhamdySo what are those "thinking tokens"?! Nothing fancy, they are just special tokens '<T>' that you insert after each word in a sentence whenever a complex problem is encountered. That's it!

👉 The main idea is to "buy" the model "some time" to think about the problem with these additional computations before answering. Using this method they observed an improved (a little bit) perplexity.

👉 Before getting excited note that: They have added these tokens manually, and they have used an RNN language model. From the paper:

"As a proof of concept, we have added N ’thinking tokens’ (< T >) after each observed word in a dataset. Our vision is that this basic concept can be extended to a self-adjusting model, which will be able to decide itself if and how many ’thinking tokens’ will be used for a specific problem, where N could also vary throughout the sentence. This would allow us to reduce the computational time, which would not increase N times."
mmhamdy
posted an update
May 15
Post
1272

💡 Thinking Tokens For Language Models!

How much is 56 times 37? Can you answer that right away?

In a short paper, David Herel and Tomas Mikolov propose a simple method to improve the reasoning of language models when performing complex calculations.

📌 They note that, although language models are not that good with difficult calculations, humans also cannot perform these calculations immediately and require a considerable amount of time to come up with an answer.

Inspired by this, they introduce 💡Thinking Tokens💡
Great work by Finegrain: Erase any object from your image just by naming it. Shadows or reflections will also be adjusted accordingly! https://huggingface.co/spaces/finegrain/finegrain-object-eraser