Publication

Export 56 results:
Filters: Author is Katz, Boris  [Clear All Filters]
2020
Kuo, Y. - L., Katz, B. & Barbu, A. Deep compositional robotic planners that follow natural language commands . International Conference on Robotics and Automation (ICRA) (2020).
Kuo, Y. - L., Katz, B. & Barbu, A. Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas. 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2020). doi:10.1109/IROS45743.2020.9341325
Kuo, Y. - L., Katz, B. & Barbu, A. Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas. (2020).PDF icon CBMM-Memo-125.pdf (2.12 MB)
Wang, C., Ross, C., Kuo, Y. - L., Katz, B. & Barbu, A. Learning a natural-language to LTL executable semantic parser for grounded robotics. (2020). doi:https://doi.org/10.48550/arXiv.2008.03277PDF icon CBMM-Memo-122.pdf (1.03 MB)
Wang, C., Ross, C., Kuo, Y. - L., Katz, B. & Barbu, A. Learning a Natural-language to LTL Executable Semantic Parser for Grounded Robotics. (Proceedings of Conference on Robot Learning (CoRL-2020), 2020). at <https://corlconf.github.io/paper_385/>
Netanyahu, A., Shu, T., Katz, B., Barbu, A. & Tenenbaum, J. B. PHASE: PHysically-grounded Abstract Social Eventsfor Machine Social Perception. Shared Visual Representations in Human and Machine Intelligence (SVRHM) workshop at NeurIPS 2020 (2020). at <https://openreview.net/forum?id=_bokm801zhx>PDF icon phase_physically_grounded_abstract_social_events_for_machine_social_perception.pdf (2.49 MB)
2019
Kuo, Y. - L., Katz, B. & Barbu, A. Deep Compositional Robotic Planners that Follow Natural Language Commands. Workshop on Visually Grounded Interaction and Language (ViGIL) at the Thirty-third Annual Conference on Neural Information Processing Systems (NeurIPS), (2019). at <https://vigilworkshop.github.io/>
Barbu, A., Banda, D. & Katz, B. Deep video-to-video transformations for accessibility with an application to photosensitivity. Pattern Recognition Letters (2019). doi:10.1016/j.patrec.2019.01.019
Idiart, M. A. P., Villavicencio, A., Katz, B., Rennó-Costa, C. & Lisman, J. How Does the Brain Represents Language and Answers Questions? Using an AI System to Understand the Underlying Neurobiological Mechanisms. Frontiers in Computational Neuroscience 13, (2019).
Ross, C., Berzak, Y., Katz, B. & Barbu, A. Learning Language from Vision. Workshop on Visually Grounded Interaction and Language (ViGIL) at the Thirty-third Annual Conference on Neural Information Processing Systems (NeurIPS) (2019).
Barbu, A. et al. ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models. Neural Information Processing Systems (NeurIPS 2019) (2019).PDF icon 9142-objectnet-a-large-scale-bias-controlled-dataset-for-pushing-the-limits-of-object-recognition-models.pdf (16.31 MB)
2018
Berzak, Y., Katz, B. & Levy, R. Assessing Language Proficiency from Eye Movements in Reading. 16th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (2018). at <http://naacl2018.org/>PDF icon 1804.07329.pdf (350.43 KB)
Kuo, Y. - L., Barbu, A. & Katz, B. Deep sequential models for sampling-based planning. The IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2018) (2018). doi:10.1109/IROS.2018.8593947PDF icon kuo2018planning.pdf (637.67 KB)
Ross, C., Barbu, A., Berzak, Y., Myanganbayar, B. & Katz, B. Grounding language acquisition by training semantic parsersusing captioned videos. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP 2018), (2018). at <http://aclweb.org/anthology/D18-1285>PDF icon Ross-et-al_ACL2018_Grounding language acquisition by training semantic parsing using caption videos.pdf (3.5 MB)
Katz, B., Borchardt, G., Felshin, S. & Mora, F. The Wiley Handbook of Human Computer Interaction 2, 539-559 (John Wiley & Sons , 2018).
Myanganbayar, B. et al. Partially Occluded Hands: A challenging new dataset for single-image hand pose estimation. The 14th Asian Conference on Computer Vision (ACCV 2018) (2018). at <http://accv2018.net/>PDF icon partially-occluded-hands-6.pdf (8.29 MB)
Myanganbayar, B. et al. Partially Occluded Hands: A challenging new dataset for single-image hand pose estimation. (2018).PDF icon CBMM-Memo-097.pdf (8.53 MB)
2017
Berzak, Y., Nakamura, C., Flynn, S. & Katz, B. Predicting Native Language from Gaze. Annual Meeting of the Association for Computational Linguistics (ACL 2017) (2017).
Paul, R., Barbu, A., Felshin, S., Katz, B. & Roy, N. Temporal Grounding Graphs for Language Understanding with Accrued Visual-Linguistic Context. Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI 2017) (2017). at <c>

Pages