Liquid just released two 450M and 1.6B param VLMs!
They're super fast and leverage SigLIP2 NaFlex encoders to handle native resolutions without distortion.
Available today on @huggingface!
Try LFM2 with llama.cpp today!
We released today a collection of GGUF checkpoints for developers to run LFM2 everywhere with llama.cpp
Select the most relevant precision for your use case and start building today.
huggingface.co/LiquidAI/LFM2-…
Liquid AI open-sources a new generation of edge LLMs! 🥳
I'm so happy to contribute to the open-source community with this release on @huggingface!
LFM2 is a new architecture that combines best-in-class inference speed and quality into 350M, 700M, and 1.2B models.
🤏 Can small models be strong reasoners?
We created a 1B reasoning model at @LiquidAI_ that is both accurate and concise
We applied a combination of SFT (to raise quality) and GRPO (to control verbosity)
The result is a best-in-class model without specific math pre-training
What is the key difference between recent reasoning models and their base counterparts?
Our new preprint reveals that reasoning models build distinctive “Reasoning Graphs” from their hidden states, characterized by more cycles, larger diameters, and stronger local connectivity.
Our paper "Beyond Induction Heads: In-Context Meta Learning Induces Multi-Phase Circuit Emergence" has accepted at #ICML2025🎉
We found training Transformers in a few-shot setting leads to the emergence of 3 circuits.
Joint work with @frt03_@ishohei220@yusuke_iwasawa_@ymatsuo
🐥Our paper "Continual Pre-training on Character-Level Noisy Texts Makes Decoder-based Language Models Robust Few-shot Learners" has been accepted to #TACL🐥
We are now planning to make a presentation at #ACL2025NLP !
weblab.t.u-tokyo.ac.jp/news/2025-0502/
🌈Our paper "Slender-Mamba: Fully Quantized Mamba in 1.58 Bits From Head to Toe" has been accepted to #coling2025 🌈
We apply BitNet b1.58 quantization to Mamba2 architecture including embedding and head layers to accelerate further lightweighting.
Congrats @UTLLM_zxYu !
🤗Our paper "Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?" has been accepted to #EMNLP2024 🤗
Congrats @FumiyaUchiyama!
Read the following post for the details!
🤗Our paper "Which Programming Language and What Features at Pre-training Stage Affect Downstream Logical Inference Performance?" has been accepted to #EMNLP2024 🤗
Congrats @FumiyaUchiyama!
Read the following post for the details!
463 Followers 278 FollowingPhD Student @UIUC_NLP. Interested in *semantics of reasoning*, from neuro-symbolic methods to reasoning evaluation/improvement in LLMs. Ex-Intern @MSFTResearch
841 Followers 338 FollowingResearch Scientist at @GoogleDeepMind / PhD from The University of Tokyo @Matsuo_Lab / AI Agent / AI Alignment / LLMs / Deep RL
964 Followers 501 FollowingIncoming Assistant Professor at @mbzuai. Previously, Postdoc at @mbzuai, PhD at @NLPTohoku, and co-founder of @langsmith_nlp.
187K Followers 105 FollowingWe're sharing/showcasing best of @github projects/repos. Follow to stay in loop. Promoting Open-Source Contributions. UNOFFICIAL, but followed by github
24K Followers 249 Following@Cohere's research lab and open science initiative that seeks to solve complex machine learning problems. Join us in exploring the unknown, together.
853 Followers 625 FollowingWorking on risks from rogue AI @apolloaievals
Past: Reversal curse, Out-of-context reasoning // best way to support 🇺🇦 https://t.co/eagDB8VUzz
10K Followers 1K FollowingWaiting on a robot body. All opinions are universal and held by both employers and family. Now a dedicated grok hate account.
Accepting ML/NLP PhD students.
1K Followers 198 FollowingAssistant Professor at 東京大学 | UTokyo 🇯🇵, PhD @Yale 🇺🇸, MSc @ucddublin 🇮🇪, proud alumni from @yalenlp || Beginner for Oil painting 🖼️ and Piano 🎹
76K Followers 13K FollowingNewsletter exploring AI&ML - AI 101, Agentic Workflow, Business insights. From ML history to AI trends. Led by @kseniase_ Know what you are talking about👇🏼
1.4M Followers 570 FollowingThe Massachusetts Institute of Technology is a world leader in research and education. Related accounts: @MITevents @MITstudents @MIT_alumni
2K Followers 1K FollowingMember of Technical Staff @GoodfireAI; Previously: Postdoc / PhD at Center for Brain Science, Harvard and University of Michigan
6K Followers 1K FollowingGroup Leader,
Physics of Intelligence Program at Harvard University
Physics of Artificial Intelligence Group, NTT Research, Inc.
4K Followers 271 FollowingCS PhD student @UCBerkeley. Part-time @AnthropicAI. Part-time eater. Prev @Tsinghua_Uni.
Try to understand and control intelligence as a human.
463 Followers 278 FollowingPhD Student @UIUC_NLP. Interested in *semantics of reasoning*, from neuro-symbolic methods to reasoning evaluation/improvement in LLMs. Ex-Intern @MSFTResearch