Interesting thread concluding that Samsung is using an AI model to replace users’ blurry, super-zoomed photos of the Moon with slightly higher res textures.

How did they discover this?

1. They downloaded a high res image of the Moon
2. Downscaled it to a blurry 170x170px image on their computer monitor
3. Took a photo of it on the Samsung with the room lights out

…and they got a magically higher-res Moon! Clever.

https://reddit.com/r/Android/comments/11nzrb0/samsung_space_zoom_moon_shots_are_fake_and_here/

Samsung "space zoom" moon shots are fake, and here is the proof

**This post has been updated with several additional experiments in newer posts, which address most comments and clarify what exactly is going...

reddit

AI image enhancement and upscaling isn’t new or exclusive to Samsung, of course. Take a blurry image of some grass and an algorithm can fill in the gaps based on what it thinks “grass” looks like.

People usually don’t notice or care; with the Moon it’s weird (and weirdly effective) because it looks the same for everyone, and everyone knows what it looks like.

Is it bad? I’d need to think more about it!

Smartphone photography is already incredibly processed, lightyears from film photography of old. People should probably be more aware that the image their phone shows them is increasingly distant from “reality” – though most of the time, people don’t want reality!

@adrianhon realized this, but now i wonder, how many consumer devices even give the owner actual sensor data
@hiimmrdave @adrianhon raw mode is available in a lot of smartphone cameras