Advantech SOM-6820 COM Express module is powered by Qualcomm Snapdragon X Elite Arm SoC

Advantech SOM-6820 is a COM Express Type 6 Compact Computer-on-Module powered by a Qualcomm Snapdragon X Elite SoC with up to twelve 64-bit Arm Oryon cores, instead of an x86 processor from Intel or AMD, more commonly found on COM Express modules. The COM also features up to 64GB LPDDR5 memory, two MIPI CSI camera connectors, an RTL8153B USB 3.0 Gigabit Ethernet controller, and a TPM 2.0 security chip. All I/Os are exposed through two standard 220-pin B2B connectors, including DisplayPort and LVDS/eDP interfaces for up to four 4K displays, up to four SATA III for storage, twelve USB 3.0/2.0 interfaces, multiple PCIe Gen4/3 interfaces, and more. With up to 45 TOPS of AI performance, the SOM-6820 is especially well-suited to medical imaging and machine vision applications as well as mission-critical systems and humanoid robots. Advancec SOM-6820 specifications: Snapdragon X Elite SoC variants (one or the other) X1E-00-1DE 12-core up

CNX Software - Embedded Systems News

IEEE Spectrum: TraffickCam Uses Computer Vision to Counter Human Trafficking . “Abby Stylianou built an app that asks its users to upload photos of hotel rooms they stay in when they travel. It may seem like a simple act, but the resulting database of hotel room images helps Stylianou and her colleagues assist victims of human trafficking.”

https://rbfirehose.com/2025/11/29/ieee-spectrum-traffickcam-uses-computer-vision-to-counter-human-trafficking/

IEEE Spectrum: TraffickCam Uses Computer Vision to Counter Human Trafficking | ResearchBuzz: Firehose

ResearchBuzz: Firehose | Individual posts from ResearchBuzz

⛐ Bridging Vision, Language, and Mathematics: Pictographic Character Reconstruction with Bézier Curves

https://arxiv.org/abs/2511.00076

#cs #graphics #text #characters #cg #béziercurves #llm #ai #vision #machinevision

Bridging Vision, Language, and Mathematics: Pictographic Character Reconstruction with Bézier Curves

While Vision-language Models (VLMs) have demonstrated strong semantic capabilities, their ability to interpret the underlying geometric structure of visual information is less explored. Pictographic characters, which combine visual form with symbolic structure, provide an ideal test case for this capability. We formulate this visual recognition challenge in the mathematical domain, where each character is represented by an executable program of geometric primitives. This is framed as a program synthesis task, training a VLM to decompile raster images into programs composed of Bézier curves. Our model, acting as a "visual decompiler", demonstrates performance superior to strong zero-shot baselines, including GPT-4o. The most significant finding is that when trained solely on modern Chinese characters, the model is able to reconstruct ancient Oracle Bone Script in a zero-shot context. This generalization provides strong evidence that the model acquires an abstract and transferable geometric grammar, moving beyond pixel-level pattern recognition to a more structured form of visual understanding.

arXiv.org

Kyocera triple-lens AI depth camera could help robots stop fumbling tiny objects

https://fed.brid.gy/r/https://nerds.xyz/2025/11/kyocera-triple-lens-ai-depth-camera/

VisionWave khẳng định vị thế tiên phong về thị giác máy và tự động hóa công nghiệp tại Việt Nam. Sau Diễn đàn Số VN–Hàn Quốc 2025, công ty định vị là cầu nối công nghệ Hàn Quốc với sản xuất thông minh Việt Nam, mang đến giải pháp tiên tiến cho nhà máy. #VisionWave #MachineVision #IndustrialAutomation #KoreanTech #SmartManufacturing #CôngNghệHànQuốc #SảnXuấtThôngMinh #ThịGiácMáy #TựĐộngHóaCôngNghiệp #MadeInVietnam #InnovativeManufacturing #VietnamTech #VTCNews

https://vtcnews.vn/visionwave-cau-n

Nature: Fair human-centric image dataset for ethical AI benchmarking. “…we introduce the Fair Human-Centric Image Benchmark (FHIBE, pronounced ‘Feebee’), a publicly available human image dataset implementing best practices for consent, privacy, compensation, safety, diversity and utility. FHIBE can be used responsibly as a fairness evaluation dataset for many human-centric computer vision […]

https://rbfirehose.com/2025/11/06/nature-fair-human-centric-image-dataset-for-ethical-ai-benchmarking/

Nature: Fair human-centric image dataset for ethical AI benchmarking | ResearchBuzz: Firehose

ResearchBuzz: Firehose | Individual posts from ResearchBuzz

💔 Are Foundation Models Ready for Industrial Defect Recognition? A Reality Check on Real-World Data

https://arxiv.org/abs/2509.20479

#computing #cs #machinevision #llm #ai #ml #manufacturing #defects #automation

Are Foundation Models Ready for Industrial Defect Recognition? A Reality Check on Real-World Data

Foundation Models (FMs) have shown impressive performance on various text and image processing tasks. They can generalize across domains and datasets in a zero-shot setting. This could make them suitable for automated quality inspection during series manufacturing, where various types of images are being evaluated for many different products. Replacing tedious labeling tasks with a simple text prompt to describe anomalies and utilizing the same models across many products would save significant efforts during model setup and implementation. This is a strong advantage over supervised Artificial Intelligence (AI) models, which are trained for individual applications and require labeled training data. We test multiple recent FMs on both custom real-world industrial image data and public image data. We show that all of those models fail on our real-world data, while the very same models perform well on public benchmark datasets.

arXiv.org

My brain after debugging for hours: needs spatial-temporal reasoning. General Intuition just raised $134M to teach AI agents exactly that, leveraging video game data (the same kind OpenAI wanted!). We're talking AI that understands movement through space and time. What's one mundane task you'd immediately hand over to an AI with perfect spatial reasoning?

#AI #Robotics #MachineVision #TechCrunch #FutureTech https://techcrunch.com/2025/10/16/general-intuition-lands-134m-seed-to-teach-agents-spatial-reasoning-using-video-game-clips/

General Intuition lands $134M seed to teach agents spatial reasoning using video game clips | TechCrunch

Late last year, OpenAI reportedly tried to buy Medal and its vast trove of video game data for $500M. Today, the company spun out a frontier research lab that's using that data to build AI agents that understand how they move through space and time, a concept called spatial-temporal reasoning.

TechCrunch

🔍 Novo no blog: Fundamentos de Visão Computacional para Sistemas Industriais

Descubra no post como a visão computacional substitui inspeções manuais com câmeras e algoritmos. Abordo conceitos de captura de imagem, resolução / pixels / intensidade, além de aplicações como localização, medição, inspeção e identificação. Tudo isso com foco em sistemas industriais que exigem robustez e confiabilidade.
Confira: https://danieltak.com.br/posts/vision/computer-vision-fundamentos/

#VisãoComputacional #MachineVision #danieltak #ComputerVision

Sistemas de Visão Industrial - Fundamentos Básicos

Fundamentos Básicos para a Visão Computacional na Indústria.

danieltak

The next Nerves meetup is Wednesday, October 29! Vittoria will showcase the capability of running ML inference pipelines on a Raspberry Pi 5 with the Hailo HAT using Nerves. You'll learn about writing pre or post-processing code with Nx, on low-cost hardware, without sending data over the internet.

RSVP here: https://www.meetup.com/nerves/events/305303640

#MachineVision #AI #ElixirLang #RaspberryPi

×