François REMY

@FremyCompany
173 Followers
126 Following
212 Posts
François REMY is a young Belgian researcher (NLP, Web Platform, Machine Learning, and European Politics). #NLP #AI #Healthcare #OpenWeb #W3C #HTML #CSS #JavaScript #Python
Websitehttp://fremycompany.com
GitHubhttps://github.com/FremyCompany
HuggingFacehttps://huggingface.co/FremyCompany

TIL there are sites that include

<meta name="rating" content="adult"> and this prevents some searches in "safe mode" (at least in Google) from displaying those in results?

“Why don't we have a label for cloud providers that would redistribute a part of their revenue to the #OpenSource projects that they deploy, and make that label required for 🇪🇺 procurement? There are these companies that make a shitload of money off the work of others.” — @robin
#NGIForum2025

The Braille Institute created a font designed to make reading easier for people with low vision called Atkinson Hyperlegible in 2019.

It just released an update (Atkinson Hyperlegible Next) and monospace version with enhanced characters, 7 weights, and variable weight.

They’re free for personal and commercial use.

https://www.brailleinstitute.org/freefont/?ref=activitypub&utm_source=activitypub&utm_medium=social

#a11y

Atkinson Hyperlegible Font - Braille Institute

Read easier with Atkinson Hyperlegible Font, crafted for low-vision readers. Download for free and enjoy clear letters and numbers on your computer!

Braille Institute

New dithering method dropped

I call it Surface-Stable Fractal Dithering and I've released it as open source along with this explainer video of how it works.

Explainer video:
https://www.youtube.com/watch?v=HPqGaIMVuLs

Source repository:
https://github.com/runevision/Dither3D

#gamedev #vfx

Surface-Stable Fractal Dithering Explained

YouTube

I usually don't post about my AI experiments here, but folks you've got to see what I implemented without any JavaScript!

https://huggingface.co/spaces/Parallia/Fairly-Multilingual-ModernBERT-Token-Alignment

I felt like a mini @anatudor 😁

Fairly Multilingual ModernBERT Token Alignment - a Hugging Face Space by Parallia

Aligns the tokens of two sentences

Selling your mattress? If it’s memory foam make sure to reformat it so an unscrupulous buyer can’t access your old dreams

Today is my first day at NVIDIA 🎉 I will be collaborating with the amazing team at Orsi Academy to make the life of robot-assisted surgeons easier 😊

Excited for this incredible opportunity to put my NLP & AI knowledge in application for a societal cause!

How ICL 𝘦𝘮𝘦𝘳𝘨𝘦𝘴 from unsupervised data?
𝘐𝘵 𝘭𝘦𝘢𝘳𝘯𝘴 𝘧𝘳𝘰𝘮 parallel phrases

After deleting parallel parts the ICL ability was reduced by 51% deleting random words - only 2%

🧵

https://arxiv.org/abs/2402.12530
#ICL #prompt #ML #NLP #NLProc #machinelearning #data #pretraining #LLM

Parallel Structures in Pre-training Data Yield In-Context Learning

Pre-trained language models (LMs) are capable of in-context learning (ICL): they can adapt to a task with only a few examples given in the prompt without any parameter update. However, it is unclear where this capability comes from as there is a stark distribution shift between pre-training text and ICL prompts. In this work, we study what patterns of the pre-training data contribute to ICL. We find that LMs' ICL ability depends on $\textit{parallel structures}$ in the pre-training data -- pairs of phrases following similar templates in the same context window. Specifically, we detect parallel structures by checking whether training on one phrase improves prediction of the other, and conduct ablation experiments to study their effect on ICL. We show that removing parallel structures in the pre-training data reduces LMs' ICL accuracy by 51% (vs 2% from random ablation). This drop persists even when excluding common patterns such as n-gram repetitions and long-range dependency, showing the diversity and generality of parallel structures. A closer look at the detected parallel structures indicates that they cover diverse linguistic tasks and span long distances in the data.

arXiv.org
Regarding AI *in* Firefox, this has brought us a built-in translator that does not outsource the work to the cloud. That's amazingly good work. I'd love to see the browser do more of that kind of thing. https://www.mozilla.org/en-US/firefox/features/translate/
Translate a webpage with Firefox

Firefox Translations is a built-in translation feature that allows you to easily browse the web in your preferred language. Learn more about how this feature in Firefox works, and how Mozilla helps keep what you translate private.

Mozilla

Realization: CSS Nesting also allows you to basically do "else" clauses in selectors.

complex-selector {
if-styles;

:not(&) {
else-styles
}
}

(if you’re wondering what this code is for, it’s for a bookmarklet to show element boxes for educational reasons)