Publication

Found 912 results
Author Title [ Type(Asc)] Year
Conference Paper
Xia, F., Wang, P., Chen, L. -chieh & Yuille, A. Zoom Better to See Clearer: Human Part Segmentation with Auto Zoom Net. ECCV (2016).
Xia, F., Wang, P., Chen, L. -chieh & Yuille, A. Zoom better to see clearer: Human and object parsing with hierarchical auto-zoom net. ECCV (2016).PDF icon auto-zoom_net.pdf (5.77 MB)
Tejwani, R. et al. Zero-shot linear combinations of grounded social interactions with Linear Social MDPs. Proceedings of the 37th AAAI Conference on Artificial Intelligence (AAAI) (2023).
Dillon, M. R. & Spelke, E. S. Young children’s automatic and alternating use of scene and object information in spatial symbols. Budapest CEU Conference on Cognitive Development (2015).
Voinea, S., Zhang, C., Evangelopoulos, G., Rosasco, L. & Poggio, T. Word-level Invariant Representations From Acoustic Waveforms. INTERSPEECH 2014 - 15th Annual Conf. of the International Speech Communication Association (International Speech Communication Association (ISCA), 2014). at <http://www.isca-speech.org/archive/interspeech_2014/i14_2385.html>
Dobs, K., Kell, A. J. E., Martinez-Trujillo, J., Cohen, M. & Kanwisher, N. Why Are Face and Object Processing Segregated in the Human Brain? Testing Computational Hypotheses with Deep Convolutional Neural Networks . Conference on Cognitive Computational Neuroscience (2020).
Li, C. & Deza, A. What Matters In Branch Specialization? Using a Toy Task to Make Predictions. Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS (2021). at <https://openreview.net/forum?id=0kPS1i6wict>
Ben-Yosef, G., Kreiman, G. & Ullman, S. What can human minimal videos tell us about dynamic recognition models?. International Conference on Learning Representations (ICLR 2020) (2020). at <https://baicsworkshop.github.io/pdf/BAICS_1.pdf>PDF icon Authors' final version (516.09 KB)
Banburski, A. et al. Weight and Batch Normalization implement Classical Generalization Bounds . ICML (2019).
Owens, A. et al. Visually indicated sounds. Conference on Computer Vision and Pattern Recognition (2016).PDF icon Owens_etal_2016_visually_indicated_sounds_CVPR.pdf (7.57 MB)
Zarco, W. & Freiwald, W. A. Visual Features for Invariant Coding by Face Selective Neurons . 2019 Conference on Cognitive Computational Neuroscience (CCN) (2019).
Rosenfeld, A. & Ullman, S. Visual Concept Recognition and Localization via Iterative Introspection. . Asian Conference on Computer Vision (2016).PDF icon Focusing on parts of interest  (910.14 KB)
Phillips-Jones, T., Coronel, S. Otero, Sani, I. & Freiwald, W. A. A Virtual Reality Experimental Approach for Studying How the Brain Implements Attentive Behaviors. Tri-Institute 2019 Gateways to the Laboratory Summer Program (2019).
Dobs, K., Kell, A. J. E., Martinez-Trujillo, J., Cohen, M. & Kanwisher, N. Using task-optimized neural networks to understand why brains have specialized processing for faces . Computational and Systems Neurosciences (2020).
Subramaniam, V. et al. Using Multimodal DNNs to Study Vision-Language Integration in the Brain. ICLR 2023 (2023). at <https://openreview.net/pdf?id=OQQ1p0pFP4>
Wang, B., Mayo, D., Deza, A., Barbu, A. & Conwell, C. On the use of Cortical Magnification and Saccades as Biological Proxies for Data Augmentation. Shared Visual Representations in Human and Machine Intelligence (SVRHM) Workshop at NeurIPS (2021). at <https://openreview.net/forum?id=Rpazl253IHb>
Lotter, W., Kreiman, G. & Cox, D. Unsupervised Learning of Visual Structure using Predictive Generative Networks. International Conference on Learning Representations (ICLR) (2016). at <http://arxiv.org/pdf/1511.06380v2.pdf>
Du, Y., Smith, K. A., Ullman, T., Tenenbaum, J. B. & Wu, J. Unsupervised Discovery of 3D Physical Objects. International Conference on Learning Representations (2021). at <https://openreview.net/forum?id=lf7st0bJIA5>
Kuo, Y. - L. et al. Trajectory Prediction with Linguistic Representations. 2022 IEEE International Conference on Robotics and Automation (ICRA) (2022). doi:10.1109/ICRA46639.2022.9811928
Mao, J., Xu, J., Jing, Y. & Yuille, A. Training and Evaluating Multimodal Word Embeddings with Large-scale Web Annotated Images. NIPS 2016 (2016).PDF icon 6590-training-and-evaluating-multimodal-word-embeddings-with-large-scale-web-annotated-images.pdf (1.57 MB)
Tacchetti, A., Voinea, S. & Evangelopoulos, G. Trading robust representations for sample complexity through self-supervised visual experience. Advances in Neural Information Processing Systems 31 (Bengio, S. et al.) 9640–9650 (Curran Associates, Inc., 2018). at <http://papers.nips.cc/paper/8170-trading-robust-representations-for-sample-complexity-through-self-supervised-visual-experience.pdf>PDF icon trading-robust-representations-for-sample-complexity-through-self-supervised-visual-experience.pdf (3.32 MB)PDF icon NeurIPS2018_Poster.pdf (6.12 MB)
Eisape, T., Levy, R., Tenenbaum, J. B. & Zaslavsky, N. Toward human-like object naming in artificial neural systems . International Conference on Learning Representations (ICLR 2020), Bridging AI and Cognitive Science workshop (2020).
Jozwik, K. M., Schrimpf, M., Kanwisher, N. & DiCarlo, J. J. To find better neural network models of human vision, find better neural network models of primate vision. BioRxiv (2019). at <https://www.biorxiv.org/content/10.1101/688390v1.full>
Schrimpf, M., Sato, F., Sanghavi, S. & DiCarlo, J. J. Temporal information for action recognition only needs to be integrated at a choice level in neural networks and primates . COSYNE (2020).
Paul, R., Barbu, A., Felshin, S., Katz, B. & Roy, N. Temporal Grounding Graphs for Language Understanding with Accrued Visual-Linguistic Context. Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI 2017) (2017). at <c>

Pages