ByteDance Seed công bố mô hình Stable-DiffCoder-8B-Instruct, mở đường cho AI tạo văn bản/lập trình bằng kỹ thuật diffusion. Mô hình đã được đăng tải trên Hugging Face và nhận nhiều sự chú ý từ cộng đồng.
#AI #DeepLearning #MáyHọc #ByteDance #HuggingFace #CodeAI #DiffusionModel #CôngNghệ

https://www.reddit.com/r/LocalLLaMA/comments/1qpm48y/bytedanceseedstablediffcoder8binstruct_hugging/

Finally got SDXL running!

I looked into the error
"modules.devices.NansException: A tensor with NaNs was produced in VAE"
and here’s what it means:

Briefly, how an image is generated with a diffusion model: The text encoder interprets the prompt, the UNet "dreams" iteratively in the latent space from noise into an image structure, and the VAE translates this latent vision into visible pixels.

A tensor is simply a multi-dimensional array of numbers, basically the data structure where the model stores all its calculations, like colors, intensities, and intermediate results of the image.

In this case, the VAE experienced a numerical instability: the latent tensor contained invalid values (NaNs), so the dreamed image could not be decoded correctly. In short: the model was still dreaming in the latent space, but the numbers “exploded” along the way (e.g., division by zero, overflow, or undefined operations).

#StableDiffusion #SDXL #AIArt #DiffusionModel #VAE #LatentSpace #Tensor #DigitalArt #CinematicArt #Kunst #KI #AI #DigitalIllustration #StilizedRealism #UrbanFantasy #Motion #DramaticLighting #FilmStill #AICommunity

🚀 LLaDA2.0 được ra mắt! Phiên bản flash: kiến trúc MoE 100B, mini: MoE 16B, cả hai đều được fine‑tune cho ứng dụng thực tế. Hỗ trợ llama.cpp đang được phát triển, phiên bản trước đã có sẵn. #AI #LLM #LLaDA2 #DiffusionModel #trí_tiện #mô_hình_ngôn_ngữ #công_nghệ

https://www.reddit.com/r/LocalLLaMA/comments/1p6gsjh/llada20_103b16b_has_been_released/

Ah yes, the future is here! 🌟 A "commercial-scale diffusion language model" that can't even handle a basic website prompt without a condescending "Just a moment..." 🙄. Truly groundbreaking tech that requires you to sacrifice #JavaScript and #cookies at the altar of web browsing functionality. 🍪💻
https://www.inceptionlabs.ai/introducing-mercury #futuretech #diffusionmodel #webdevelopment #HackerNews #ngated

AI Face Anonymizer Masks Human Identity in Images

We're all pretty familiar with AI's ability to create realistic-looking images of people that don't exist, but here's an unusual implementation of using that technology for a different purpose: masking people's identity without altering the substance of the image itself. The result is the photo's content and "purpose" (for lack of a better term) of the image remains unchanged, while at the same time becoming impossible to identify the actual person in it. This invites some interesting privacy-related applications.

Originals on left, anonymized versions on the right. The substance of the images has not changed.

The paper for Face Anonymization Made Simple has all the details, but the method boils down to using diffusion models to take an input image, automatically pick out identity-related features, and alter them in a way that looks more or less natural. For this purpose, identity-related features essentially means key parts of a human face. Other elements of the photo (background, expression, pose, clothing) are left unchanged. As a concept it's been explored before, but researchers show that this versatile method is both simpler and better-performing than others.

Diffusion models are the essence of AI image generators like Stable Diffusion. The fact that they can be run locally on personal hardware has opened the doors to all kinds of interesting experimentation, like this haunted mirror and other interactive experiments. Forget tweaking dull sliders like "brightness" and "contrast" for an image. How about altering the level of "moss", "fire", or "cookie" instead?

#artificialintelligence #ai #aiimagegenerator #anonymity #anonymizer #diffusionmodel

AI Face Anonymizer Masks Human Identity In Images

We’re all pretty familiar with AI’s ability to create realistic-looking images of people that don’t exist, but here’s an unusual implementation of using that technology for …

Hackaday
If I had the time, energy, and education to pull it off, I'd do some scholarship and writing elaborating on this juxtaposition:

- Statistics, as a field of study, gained significant energy and support from eugenicists with the purpose of "scientizing" their prejudices. Some of the major early thinkers in modern statistics, like Galton, Pearson, and Fisher, were eugenicists out loud; see https://nautil.us/how-eugenics-shaped-statistics-238014/
- Large language models and diffusion models rely on certain kinds of statistical methods, but discard any notion of confidence interval or validation that's grounded in reality. For instance, the LLM inside GPT outputs a probability distribution over the tokens (words) that could follow the input prompt. However, there is no way to even make sense of a probability distribution like this in real-world terms, let alone measure anything about how well it matches reality. See for instance https://aclanthology.org/2020.acl-main.463.pdf and Michael Reddy's The conduit metaphor: A case of frame conflict in our language about language

Early on in this latest AI hype cycle I wrote a note to myself that this style of AI is necessarily biased. In other words, the bias coming out isn't primarily a function of biased input data (though of course that's a problem too). That'd be a kind of contingent bias that could be addressed. Rather, the bias these systems exhibit is a function of how the things are structured at their core, and no amount of data curating can overcome it. I can't prove this, so let's call it a hypothesis, but I believe it.

#AI #GenAI #GenerativeAI #ChatGPT #GPT #Gemini #Claude #Llama #StableDiffusion #Midjourney #DallE #LLM #DiffusionModel #linguistics #NLP
How Eugenics Shaped Statistics

Exposing the damned lies of three science pioneers.

Nautilus

#Platforms throw away the diffusion on innovations idea that you need 'knowledge' to be an innovator or early adopter.
#AI gets incorporated into your workflow whether you have knowledge, are persuaded, or not.

#rogers #diffusionmodel #innovation

This is the next stage: W.A.L.T, a #DiffusionModel for #photorealistic #VideoGeneration 🤖🎬 Developed by Gupta et al. (2023), based on a #transformer that is trained on image and video generation in a shared #LatentSpace.

🌍 https://walt-video-diffusion.github.io (including many sample movies)
📔 https://arxiv.org/abs/2312.06662

#WALT

Photorealistic Video Generation with Diffusion Models

Remove the poodle で犬が画像から取り除かれてる。こういうのは他の画像生成AIでもできてると思うけど、どうなんですか。
#Meta #AI #Emu #diffusionModel #video
Meta、テキストからの動画生成モデル「Emu Video」とマルチタスク画像編集モデル「Emu Edit」を発表
https://atmarkit.itmedia.co.jp/ait/articles/2311/24/news050.html
#Meta #AI #Emu #diffusionModel #video
Meta、テキストからの動画生成モデル「Emu Video」とマルチタスク画像編集モデル「Emu Edit」を発表:画像生成基盤モデル「Emu」がベース

Metaは、最近発表した画像生成のための基盤モデル「Emu」に基づく2つの研究成果を発表した。拡散モデルに基づいてテキストから動画を生成するモデル「Emu Video」と、マルチタスク画像編集モデル「Emu Edit」だ。

@IT