%0 Conference Proceedings %B International Conference on Learning Representations (ICLR) %D 2022 %T Finding Biological Plausibility for Adversarially Robust Features via Metameric Tasks %A Anne Harrington %A Arturo Deza %K Adversarial Robustness %K Metamerism %K Perceptual Invariance %K Peripheral Computation %K Psychophysics %K Texture %B International Conference on Learning Representations (ICLR) %G eng %U https://openreview.net/forum?id=yeP_zx9vqNm %0 Conference Paper %B BrainScore Workshop at COSYNE %D 2022 %T Joint rotational invariance and adversarial training of a dual-stream Transformer yields state of the art Brain-Score for Area V4 %A William Berrios %A Arturo Deza %K adversarial training %K Brain-Score competition %K rotation invariance %K Vision Transformer %B BrainScore Workshop at COSYNE %G eng %U https://openreview.net/pdf?id=SOulrWP-Xb5 %0 Generic %D 2021 %T The Effects of Image Distribution and Task on Adversarial Robustness %A Owen Kunhardt %A Arturo Deza %A Tomaso Poggio %X

In this paper, we propose an adaptation to the area under the curve (AUC) metric to measure the adversarial robustness of a model over a particular ε-interval [ε0, ε1] (interval of adversarial perturbation strengths) that facilitates unbiased comparisons across models when they have different initial ε0 performance. This can be used to determine how adversarially robust a model is to different image distributions or task (or some other variable); and/or to measure how robust a model is comparatively to other models. We used this adversarial robustness metric on models of an MNIST, CIFAR-10, and a Fusion dataset (CIFAR-10 + MNIST) where trained models performed either a digit or object recognition task using a LeNet, ResNet50, or a fully connected network (FullyConnectedNet) architecture and found the following: 1) CIFAR-10 models are inherently less adversarially robust than MNIST models; 2) Both the image distribution and task that a model is trained on can affect the adversarial robustness of the resultant model. 3) Pretraining with a different image distribution and task sometimes carries over the adversarial robustness induced by that image distribution and task in the resultant model; Collectively, our results imply non-trivial differences of the learned representation space of one perceptual system over another given its exposure to different image statistics or tasks (mainly objects vs digits). Moreover, these results hold even when model systems are equalized to have the same level of performance, or when exposed to approximately matched image statistics of fusion images but with different tasks.

%8 02/2021 %2

https://hdl.handle.net/1721.1/129813

%0 Generic %D 2021 %T Evaluating the Adversarial Robustness of a Foveated Texture Transform Module in a CNN %A Jonathan Gant %A Andrzej Banburski %A Arturo Deza %A Tomaso Poggio %B NeurIPS 2021 %8 12/2021 %U https://nips.cc/Conferences/2021/Schedule?showEvent=21868 %0 Conference Paper %B Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS %D 2021 %T On the use of Cortical Magnification and Saccades as Biological Proxies for Data Augmentation %A Binxu Wang %A David Mayo %A Arturo Deza %A Andrei Barbu %A Colin Conwell %K Active Perception %K Data-Augmentation %K Foveation %K Self-Supervised Learning %B Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS %G eng %U https://openreview.net/forum?id=Rpazl253IHb %0 Conference Paper %B Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS %D 2021 %T What Matters In Branch Specialization? Using a Toy Task to Make Predictions %A Chenguang Li %A Arturo Deza %K branch specialization %K computational vision %K curriculum learning %B Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS %G eng %U https://openreview.net/forum?id=0kPS1i6wict %0 Conference Paper %B Shared Visual Representations in Human and Machine Intelligence (SVRHM) workshop at NeurIPS 2020 %D 2020 %T CUDA-Optimized real-time rendering of a Foveated Visual System %A Elian Malkin %A Arturo Deza %A Tomaso Poggio %X

The spatially-varying field of the human visual system has recently received a resurgence of interest with the development of virtual reality (VR) and neural networks. The computational demands of high resolution rendering desired for VR can be offset by savings in the periphery [16], while neural networks trained with foveated input have shown perceptual gains in i.i.d and o.o.d generalization [25, 6]. In this paper, we present a technique that exploits the CUDA GPU architecture to efficiently generate Gaussian-based foveated images at high definition (1920px × 1080px) in real-time (165 Hz), with a larger number of pooling regions than previous Gaussian-based foveation algorithms by several orders of magnitude [10, 25], producing a smoothly foveated image that requires no further blending or stitching, and that can be well fit for any contrast sensitivity function. The approach described can be adapted from Gaussian blurring to any eccentricity-dependent image processing and our algorithm can meet demand for experimentation to evaluate the role of spatially-varying processing across biological and artificial agents, so that foveation can be added easily on top of existing systems rather than forcing their redesign (“emulated foveated renderer” [22]). Altogether, this paper demonstrates how a GPU, with a CUDA block-wise architecture, can be employed for radially-variant rendering, with opportunities for more complex post-processing to ensure a metameric foveation scheme [33].

%B Shared Visual Representations in Human and Machine Intelligence (SVRHM) workshop at NeurIPS 2020 %8 12/2020 %G eng %U https://arxiv.org/abs/2012.08655 %0 Generic %D 2020 %T Hierarchically Local Tasks and Deep Convolutional Networks %A Arturo Deza %A Qianli Liao %A Andrzej Banburski %A Tomaso Poggio %K Compositionality %K Inductive Bias %K perception %K Theory of Deep Learning %X

The main success stories of deep learning, starting with ImageNet, depend on convolutional networks, which on certain tasks perform significantly better than traditional shallow classifiers, such as support vector machines. Is there something special about deep convolutional networks that other learning machines do not possess? Recent results in approximation theory have shown that there is an exponential advantage of deep convolutional-like networks in approximating functions with hierarchical locality in their compositional structure. These mathematical results, however, do not say which tasks are expected to have input-output functions with hierarchical locality. Among all the possible hierarchically local tasks in vision, text and speech we explore a few of them experimentally by studying how they are affected by disrupting locality in the input images. We also discuss a taxonomy of tasks ranging from local, to hierarchically local, to global and make predictions about the type of networks required to perform  efficiently on these different types of tasks.

%8 06/2020 %1

https://arxiv.org/abs/2006.13915

%2

https://hdl.handle.net/1721.1/125980