User profiles for Andrea Steiner

Andreas Peter Steiner

- Verified email at google.com - Cited by 4214

Andrea Karin Steiner

- Verified email at uni-graz.at - Cited by 4093

Pali: A jointly-scaled multilingual language-image model

…, C Jia, BK Ayan, C Riquelme, A Steiner… - arXiv preprint arXiv …, 2022 - arxiv.org
Effective scaling and a flexible task interface enable large language models to excel at many
tasks. We present PaLI (Pathways Language and Image model), a model that extends this …

Lit: Zero-shot transfer with locked-image text tuning

X Zhai, X Wang, B Mustafa, A Steiner… - Proceedings of the …, 2022 - openaccess.thecvf.com
This paper presents contrastive-tuning, a simple method employing contrastive training to
align image and text models while still taking advantage of their pre-training. In our empirical …

How to train your vit? data, augmentation, and regularization in vision transformers

A Steiner, A Kolesnikov, X Zhai, R Wightman… - arXiv preprint arXiv …, 2021 - arxiv.org
Vision Transformers (ViT) have been shown to attain highly competitive performance for a
wide range of vision applications, such as image classification, object detection and semantic …

[HTML][HTML] A trial of annual in-home comprehensive geriatric assessments for elderly people living in the community

AE Stuck, HU Aronow, A Steiner… - … England Journal of …, 1995 - Mass Medical Soc
Background and Methods The prevention of disability in elderly people poses a challenge for
health care and social services. We conducted a three-year, randomized, controlled trial of …

Mlp-mixer: An all-mlp architecture for vision

…, T Unterthiner, J Yung, A Steiner… - Advances in neural …, 2021 - proceedings.neurips.cc
Convolutional Neural Networks (CNNs) are the go-to model for computer vision. Recently,
attention-based networks, such as the Vision Transformer, have also become popular. In this …

Image captioners are scalable vision learners too

M Tschannen, M Kumar, A Steiner… - Advances in …, 2024 - proceedings.neurips.cc
Contrastive pretraining on image-text pairs from the web is one of the most popular large-scale
pretraining strategies for vision backbones, especially in the context of large multimodal …

Scaling vision transformers to 22 billion parameters

…, J Heek, J Gilmer, AP Steiner… - International …, 2023 - proceedings.mlr.press
The scaling of Transformers has driven breakthrough capabilities for language models. At
present, the largest large language models (LLMs) contain upwards of 100B parameters. …

Inappropriate medication use in community-residing older persons

AE Stuck, MH Beers, A Steiner… - Archives of internal …, 1994 - jamanetwork.com
Background: Elderly patients taking inappropriate drugs are at increased risk for adverse
outcomes. We investigated the prevalence of inappropriate drug use and its predisposing …

[HTML][HTML] Heat stored in the earth system: Where does the energy go? The GCOS earth heat inventory team

…, A Shepherd, DA Slater, AK Steiner… - Earth System …, 2020 - essd.copernicus.org
Human-induced atmospheric composition changes cause a radiative imbalance at the top
of the atmosphere which is driving global warming. This Earth energy imbalance (EEI) is the …

Pali-x: On scaling up a multilingual vision and language model

…, L Beyer, J Amelot, K Lee, AP Steiner… - arXiv preprint arXiv …, 2023 - arxiv.org
We present the training recipe and results of scaling up PaLI-X, a multilingual vision and
language model, both in terms of size of the components and the breadth of its training task …