Add and adapt natural language cues for subsequent CLIP generalization
Large pre-trained vision and language models, such as CLIP, have shown promising generalization ability, but may struggle in specialized domains ...
Large pre-trained vision and language models, such as CLIP, have shown promising generalization ability, but may struggle in specialized domains ...
The scaling of ai means greater spending on infrastructure. Massive, multidisciplinary research puts economic pressure on institutions, as high-performance computing ...
Reinforcement learning from human feedback (RLHF) is an effective approach to align language models with human preferences. Fundamental to RLHF ...
Deep learning has made significant advances in artificial intelligence, particularly in natural language processing and computer vision. However, even the ...
Reinforcement learning practitioners typically avoid hierarchical policies, especially in image-based observation spaces. Typically, the performance improvement on a single task ...
Graph machine learning remains a popular research direction, especially with the wave of AI4Science driving increasingly diverse applications of graph ...
Recently, there has been increasing interest in improving the generalization of deep networks by regulating the sharpness of the loss ...
Existing models of vision and language exhibit strong generalization across a variety of visual domains and tasks. However, these models ...
An effective method to improve LLMs' reasoning skills is to employ supervised fine-tuning (SFT) with chain-of-thought (CoT) annotations. However, this ...
There are still important differences between our current empirical setup and the fundamental problem of aligning superhuman models. For example, ...