Ankit Sharma

98 Followers
493 Following
154 Posts
ML/AI Engineer by day 🐤 l Hackerman by night 🐦 | Weird photographer 📷 | Hobbyist  developer
#emacs #macOS #cpp #rust #python #iOS #swift #machinelearning #deeplearning #softwareengineering #mlops #nlp #llm
Websitehttps://nezubn.com
Twitterhttps://twitter.com/nezubn
GitHubhttps://github.com/ankitsharma07
Linkedinhttps://www.linkedin.com/in/ankitkumar1107

Welcome to a new browser!

> Ladybird uses a brand new engine based on web standards, without borrowing any code from other browsers. It started as a humble HTML viewer for the SerenityOS hobby project, but since then it's grown into a full cross-platform browser project supporting Linux, macOS, and other Unix-like systems. — https://ladybird.org/announcement.html

This is good for the Web!

Announcing the Ladybird Browser Initiative

We've created a US non-profit to develop Ladybird into a truly independent web browser...

This intro to Linear Algebra is what you really needed but never had.

Link: https://pabloinsente.github.io/intro-linear-algebra

Introduction to Linear Algebra for Applied Machine Learning with Python

the github feed used to be so good.. used to discover a lot of alpha

wtf they changed that into.. now I get a couple of weird updates and then nothing, tried filtering but nothing works out well

📚 Pretty nervous about this, but here goes. Here's a side project I've been working on for the past few months. It's called Viberary and it's a semantic search engine. It gives you book recommendations based on ✨vibes. ✨ You enter a search query like "funny scifi" and it returns a list of (hopefully!) good recommendations.

https://viberary.pizza/

There is an about page that explains the data, model, etc. It's still pretty early stages but it's been a labor of love for me.

Viberary

Find your book vibe semantically!

@zhenyi hey, I was going through the Sessions app which you created last year and I was wondering how did you pull the videos there. Is there an api ??

Fast Distributed Inference Serving for Large Language Models

FastServe improves the average and tail job completion time by up to 5.1x and 6.4x, respectively, compared to the SotA solution Orca.

https://arxiv.org/abs/2305.05920

Fast Distributed Inference Serving for Large Language Models

Large language models (LLMs) power a new generation of interactive AI applications exemplified by ChatGPT. The interactive nature of these applications demands low latency for LLM inference. Existing LLM serving systems use run-to-completion processing for inference jobs, which suffers from head-of-line blocking and long latency. We present FastServe, a distributed inference serving system for LLMs. FastServe exploits the autoregressive pattern of LLM inference to enable preemption at the granularity of each output token. FastServe uses preemptive scheduling to minimize latency with a novel skip-join Multi-Level Feedback Queue scheduler. Based on the new semi-information-agnostic setting of LLM inference, the scheduler leverages the input length information to assign an appropriate initial queue for each arrival job to join. The higher priority queues than the joined queue are skipped to reduce demotions. We design an efficient GPU memory management mechanism that proactively offloads and uploads intermediate state between GPU memory and host memory for LLM inference. We build a system prototype of FastServe and experimental results show that compared to the state-of-the-art solution vLLM, FastServe improves the throughput by up to 31.4x and 17.9x under the same average and tail latency requirements, respectively.

arXiv.org

Easy to miss in Mastodon's latest update, but the most-requested features are indeed coming "soon:"

"...quote posts, improved content and profile search, and groups. We’re also continuously working on improving content and profile discovery, onboarding, and of course our extensive set of moderation tools, as well as removing friction from decentralized features. Keep a lookout for these updates soon."

https://blog.joinmastodon.org/2023/05/a-new-onboarding-experience-on-mastodon/

@spreadmastodon @fediversenews

A new onboarding experience on Mastodon

Today we’re making signing up on Mastodon easier than ever before. We understand that deciding which Mastodon service provider to kick off your experience with can be confusing. We know this is a completely new concept for many people, since traditionally the platform and the service provider are one and the same. This choice is what makes Mastodon different from existing social networks, but it also presents a unique onboarding challenge. To make this step easier, we now have a default sign-up option that works with a server we operate. If you wish to leave or join a different server, you can do so at any time.

Mastodon Blog

Bluesky

Yet another Social Media

Waiting eagerly for June 5 to come. #WWDC23

If you haven't tried running an open-ish alternative to ChatGPT on your own computer yet gpt4all is the new easiest way to do so: checkout the repo, download a 3.9GB model file and you can run the compiled binaries in the chat/ folder directly on Intel/M1 Mac, Windows or Linux: https://github.com/nomic-ai/gpt4all

It's not too bad for something that runs locally!

GitHub - nomic-ai/gpt4all: GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use.

GPT4All: Run Local LLMs on Any Device. Open-source and available for commercial use. - nomic-ai/gpt4all

GitHub