Results 1 to 10 of about 9,460,007 (367)
Scaling Instruction-Finetuned Language Models [PDF]
Finetuning language models on a collection of datasets phrased as instructions has been shown to improve model performance and generalization to unseen tasks.
Hyung Won Chung +31 more
semanticscholar +1 more source
Swin Transformer V2: Scaling Up Capacity and Resolution [PDF]
We present techniques for scaling Swin Transformer [35] up to 3 billion parameters and making it capable of training with images of up to 1,536x1,536 resolution.
Ze Liu +11 more
semanticscholar +1 more source
Scaling Autoregressive Models for Content-Rich Text-to-Image Generation [PDF]
We present the Pathways Autoregressive Text-to-Image (Parti) model, which generates high-fidelity photorealistic images and supports content-rich synthesis involving complex compositions and world knowledge.
Jiahui Yu +16 more
semanticscholar +1 more source
Scaling Vision Transformers to 22 Billion Parameters [PDF]
The scaling of Transformers has driven breakthrough capabilities for language models. At present, the largest large language models (LLMs) contain upwards of 100B parameters.
Mostafa Dehghani +41 more
semanticscholar +1 more source
Scaling Vision Transformers [PDF]
Attention-based neural networks such as the Vision Transformer (ViT) have recently attained state-of-the-art results on many computer vision benchmarks.
Xiaohua Zhai +3 more
semanticscholar +1 more source
Reproducible Scaling Laws for Contrastive Language-Image Learning [PDF]
Scaling up neural networks has led to remarkable performance across a wide range of tasks. Moreover, performance often follows reliable scaling laws as a function of training set size, model size, and compute, which offers valuable guidance as large ...
Mehdi Cherti +8 more
semanticscholar +1 more source
Emergence of Scaling in Random Networks [PDF]
Systems as diverse as genetic networks or the World Wide Web are best described as networks with complex topology. A common property of many large networks is that the vertex connectivities follow a scale-free power-law distribution.
B. McInnes +4 more
semanticscholar +2 more sources
Scaling Data-Constrained Language Models [PDF]
The current trend of scaling language models involves increasing both parameter count and training dataset size. Extrapolating this trend suggests that training dataset size may soon be limited by the amount of text data available on the internet ...
Niklas Muennighoff +8 more
semanticscholar +1 more source
Scaling Back on Scales with a Scale of Scales [PDF]
An ever-increasing number of articles are published introducing clinical scales to describe neurovascular diseases. Unfortunately, unless you are some kind of idiot savant, there are now too many scales to remember.
H J, Cloft, D F, Kallmes
openaire +2 more sources
Hybrid perovskites are a novel type of semiconductors that show great potential for solution-processed optoelectronic devices. For all applications, the device performance is determined by the quality of the solution-processed perovskite thin films ...
Oleksandra Shargaieva +6 more
doaj +1 more source

