Toggle light / dark theme

Google presents Reuse Your Rewards.

Reward model transfer for zero-shot cross-lingual alignment.

Aligning language models (LMs) based on human-annotated preference data is a crucial step in obtaining practical and performant LM-based systems.


Join the discussion on this paper page.

From upenn BLINK multimodal large language models can see but not perceive.

From UPenn.

Multimodal large language models can see but not perceive.

We introduce Blink, a new benchmark for multimodal language models (LLMs) that focuses on core visual perception abilities not found in other evaluations.