MIT’s LFM2VL: Multimodal AI That Fits in Your Pocket
When we think of advanced AI models, we usually picture sprawling data centers, massive GPUs, and a constant connection to the cloud. But researchers at the Massachusetts Institute of Technology (MIT) are challenging that assumption with LFM2VL, a multimodal model that processes both vision and language—and can run locally on a mobile device. The implications […]