Help
RSS
API
Feed
Maltego
Contact
Domain > www.xinleic.xyz
×
More information on this domain is in
AlienVault OTX
Is this malicious?
Yes
No
DNS Resolutions
Date
IP Address
2025-11-28
173.236.245.235
(
ClassC
)
Port 80
HTTP/1.1 301 Moved PermanentlyDate: Fri, 28 Nov 2025 10:28:54 GMTServer: ApacheLocation: https://xinleic.xyz/Content-Length: 228Content-Type: text/html; charsetiso-8859-1 !DOCTYPE HTML PUBLIC -//IETF//DTD HTML 2.0//EN>html>head>title>301 Moved Permanently/title>/head>body>h1>Moved Permanently/h1>p>The document has moved a hrefhttps://xinleic.xyz/>here/a>./p>/body>/html>
Port 443
HTTP/1.1 200 OKDate: Fri, 28 Nov 2025 10:28:54 GMTServer: ApacheUpgrade: h2Connection: UpgradeLast-Modified: Sun, 22 Jun 2025 03:42:56 GMTETag: 9e23-63820e4067000Accept-Ranges: bytesContent-Length: 40483Cache-Control: max-age600Expires: Fri, 28 Nov 2025 10:38:54 GMTVary: Accept-Encoding,User-AgentContent-Type: text/html !DOCTYPE html>html langen>head> meta charsetutf-8> title>Xinlei Chen/title> meta nameviewport contentwidthdevice-width, initial-scale1> meta namedescription content> meta nameauthor contentXinlei Chen, xinleic@meta.com> link relstylesheet typetext/css hrefcss/bootstrap.min.css mediascreen> !-- Global site tag (gtag.js) - Google Analytics --> script async srchttps://www.googletagmanager.com/gtag/js?idUA-45111154-1>/script> script> window.dataLayer window.dataLayer || ; function gtag(){dataLayer.push(arguments);} gtag(js, new Date()); gtag(config, UA-45111154-1); /script>/head>body> div classnavbar navbar-default navbar-fixed-top> div classnavbar-header> button typebutton classnavbar-toggle data-togglecollapse data-target.navbar-responsive-collapse> span classicon-bar>/span> span classicon-bar>/span> span classicon-bar>/span> /button> a classnavbar-brand hrefindex.html>Home/a> /div> div classnavbar-collapse collapse navbar-responsive-collapse> ul classnav navbar-nav> li classdropdown> a href# classdropdown-toggle data-toggledropdown>Projectsb classcaret>/b>/a> ul classdropdown-menu> li classdropdown-header>@Carnegie Mellon/li> li>a hrefweb.html>Web ConvNet/a>/li> li>a hrefpoly.html>Sense Discovery/a>/li> li>a hrefseg.html>Object Discovery and Segmentation/a>/li> li>a hrefhttp://www.neil-kb.com>Never Ending Image Learner/a>/li> li classdivider>/li> li classdropdown-header>@Zhejiang University/li> li>a hrefhttp://www.cad.zju.edu.cn/home/dengcai/Data/ReproduceExp.html#LSC>Scalable Spectral Clustering/a>/li> /ul> /li> li>a hrefindex.html#publication>Publications/a>/li> li>a hrefhttps://github.com/endernewton>Code/a>/li> /ul> /div> /div> div classcontainer> div classpage-header stylemargin-top: 120px;> div classrow stylepadding-bottom: 5px;> div classcol-xs-8 styletext-align:left;>img srcimages/eng.png stylemax-height: 100%; max-width: 100%;/>!--h1 stylefont-size: 44px; font-family: Audiowide;margin-bottom: 0px;>Xinlei Chen/h1>-->/div> div classcol-xs-4 styletext-align:right;>img srcimages/name.png stylemax-height: 100%; max-width: 100%;/>/div> /div> /div> a namebio stylevisibility: hidden;>/a>!-- h1>Brief Bio/h1>--> div classrow> div classcol-xs-12 col-md-2 styletext-align: center; vertical-align: middle;> img srcimages/image_2018.jpg stylemax-height: 100%; max-width: 100%; /> /div> div classcol-xs-12 col-md-7 stylevertical-align: middle;> p> Hey! I am right now at a hrefhttps://x.ai/>xAI/a>, pushing its multi-modal frontier. p> I was a Research Scientist at a hrefhttps://ai.facebook.com/>Meta/a> FAIR, working on on pre-training and understanding visual representations. /p> p> I got my PhD from the a hrefhttp://lti.cs.cmu.edu/>Language Technology Institute/a>, a hrefhttp://www.cs.cmu.edu/>Carnegie Mellon University/a>, while working at the a hrefhttp://ri.cmu.edu/>Robotics Institute/a>. I graduated with a bachelors degree in computer science from a hrefhttp://www.zju.edu.cn/english/>Zhejiang University/a>, China. /p>!-- p stylecolor:red;> We are always looking for research interns at Meta AI, feel free to shoot an email if interested! /p> --> /div> div classcol-xs-12 col-md-3> h3 stylemargin-top: auto;>Meta Info/h3> ul> li>E-mail: xlc at x dot ai/li> li>Google Scholar: a hrefhttps://scholar.google.com/citations?userbSU7LYoAAAAJ&hlen>link/a>/li> li>Github: a hrefhttps://github.com/endernewton/>link/a>/li> /ul> /div> /div> a namepublication stylevisibility: hidden;>/a> h1>Publications small>i>with Extra Materials/i>/small>/h1> p> For better maintenance and more clear purpose of this page, I am only listing publications with i>extra materials/i> such as code link below in reverse chronological order. /p> p> For a complete, up-to-date list of publications, please check out a hrefhttps://scholar.google.com/citations?userbSU7LYoAAAAJ&hlen>Google Scholar/a> or a hrefhttps://arxiv.org/search/?queryXinlei+Chen&searchtypeauthor>arXiv/a> instead. /p> br /> table classtable table-striped table-hover > thead> tr> th classcol-xs-2>/th> th>Information/th> th classcol-xs-2>Links/th> /tr> /thead> tbody> tr> td>img srcimages/vitok.png stylemax-height: 100%; max-width: 100%; />/td> td>Philippe Hansen-Estruch, David Yan, Ching-Yao Chung, Orr Zohar, Jialiang Wang, Tingbo Hou, Tao Xu, Sriram Vishwanath, Peter Vajda, b>Xinlei Chen/b>. Learnings from Scaling Visual Tokenizers for Reconstruction and Generation. ICML, 2025. /td> td>a hrefhttps://arxiv.org/abs/2501.09755>Link/a> /td> /tr> tr> td>img srcimages/before_after.svg stylemax-height: 100%; max-width: 100%; />/td> td>Jiachen Zhu, b>Xinlei Chen/b>, Kaiming He, Yann LeCun and Zhuang Liu. Transformers without Normalization. arXiv, 2025. /td> td>a hrefhttps://arxiv.org/abs/2503.10622>Link/a> a hrefhttps://github.com/jiachenzhu/DyT>Code/a> /td> /tr> tr> td>img srcimages/attn.png stylemax-height: 100%; max-width: 100%; />/td> td>Alexander C. Li, Yuandong Tian, Beidi Chen, Deepak Pathak, b>Xinlei Chen/b>. On the Surprising Effectiveness of Attention Transfer for Vision Transformers. NeurIPS, 2024. /td> td>a hrefhttps://arxiv.org/abs/2411.09702>Link/a> a hrefhttps://github.com/alexlioralexli/attention-transfer>Code/a> /td> /tr> tr> td>img srcimages/novideo.png stylemax-height: 100%; max-width: 100%; />/td> td>Xueyang Yu, b>Xinlei Chen/b>, Yossi Gandelsman. Learning Video Representations without Natural Videos. ArXiv, 2024. /td> td>a hrefhttps://arxiv.org/abs/2410.24213>Link/a> a hrefhttps://unicorn53547.github.io/video_syn_rep/>Project/a> /td> /tr> tr> td>img srcimages/hpt.gif stylemax-height: 100%; max-width: 100%; />/td> td>Lirui Wang, b>Xinlei Chen/b>, Jialiang Zhao, Kaiming He. Scaling Proprioceptive-Visual Learning with Heterogeneous Pre-trained Transformers. NeurIPS, 2024. b>Spotlight/b>. /td> td>a hrefhttps://arxiv.org/abs/2409.20537>Link/a> a hrefhttps://liruiw.github.io/hpt/>Project/a> a hrefhttps://github.com/liruiw/HPT>Code/a> /td> /tr> tr> td>img srcimages/mttt.png stylemax-height: 100%; max-width: 100%; />/td> td>Yu Sun*, Xinhao Li*, Karan Dalal*, Jiarui Xu, Arjun Vikram, Genghan Zhang, Yann Dubois, b>Xinlei Chen/b>†, Xiaolong Wang†, Sanmi Koyejo†, Tatsunori Hashimoto†, Carlos Guestrin†. Learning to (Learn at Test Time): RNNs with Expressive Hidden States. ArXiv, 2024. /td> td>a hrefhttps://arxiv.org/abs/2407.04620>Link/a> a hrefhttps://github.com/test-time-training/ttt-lm-pytorch>Code(PyTorch)/a> a hrefhttps://github.com/test-time-training/ttt-lm-jax>Code(Jax)/a> /td> /tr> tr> td>img srcimages/pit.png stylemax-height: 100%; max-width: 100%; />/td> td>Duy-Kien Nguyen, Mahmoud Assran, Unnat Jain, Martin Oswald, Cees Snoek, b>Xinlei Chen/b>. An Image is Worth More Than 16x16 Patches: Exploring Transformers on Individual Pixels. ArXiv, 2024. /td> td>a hrefhttps://arxiv.org/abs/2406.09415>Link/a> /td> /tr> tr> td>img srcimages/ma_act.png stylemax-height: 100%; max-width: 100%; />/td> td>Mingjie Sun, b>Xinlei Chen/b>, Zico Kolter, Zhuang Liu. Massive Activations in Large Language Models. CoLM, 2024. /td> td>a hrefhttps://arxiv.org/abs/2402.17762>Link/a> a hrefhttps://eric-mingjie.github.io/massive-activations/index.html>Project/a> a hrefhttps://github.com/locuslab/massive-activations>Code/a> /td> /tr> tr> td>img srcimages/v-jepa.png stylemax-height: 100%; max-width: 100%; />/td> td>Adrien Bardes, Quentin Garrido, Jean Ponce, b>Xinlei Chen/b>, Michael Rabbat, Yann LeCun, Mahmoud Assran†, Nicolas Ballas†. Revisiting Feature Prediction for Learning Visual Representations from Video. Meta, 2024. /td> td>a hrefhttps://ai.meta.com/research/publications/revisiting-feature-prediction-for-learning-visual-representations-from-video/>Link/a> a hrefhttps://ai.meta.com/blog/v-jepa-yann-lecun-ai-model-video-joint-embedding-predictive-architecture/>Link/a> a hrefhttps://github.com/facebookresearch/jepa>Code/a> /td> /tr> tr> td>img srcimages/l-dae.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Zhuang Liu, Saining Xie, Kaiming He. Deconstructing Denoising Diffusion Models for Self-Supervised Learning. ArXiv, 2024. /td> td>a hrefhttps://arxiv.org/abs/2401.14404>Link/a> /td> /tr> tr> td>img srcimages/rmae.png stylemax-height: 100%; max-width: 100%; />/td> td>Duy-Kien Nguyen, Vaibhav Aggarwal, Yanghao Li, Martin Oswald, Alexander Kirillov, Cees Snoek, b>Xinlei Chen/b>. R-MAE: Regions Meet Masked Autoencoders. The 12th International Conference on Learning Representations (ICLR), 2024. /td> td>a hrefhttps://arxiv.org/abs/2306.05411>Link/a> a hrefhttps://github.com/facebookresearch/r-mae>Code/a> /td> /tr> tr> td>img srcimages/convnextv2.png stylemax-height: 100%; max-width: 100%; />/td> td>Sanghyun Woo, Shoubhik Debnath, Ronghang Hu, b>Xinlei Chen/b>, Zhuang Liu, In So Kweon, Saining Xie. ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders. The 36th Conference on Computer Vision and Pattern Recognition (CVPR), 2023. /td> td>a hrefhttps://arxiv.org/abs/2301.00808>Link/a> a hrefhttps://github.com/facebookresearch/ConvNeXt-V2>Code/a> /td> /tr> tr> td>img srcimages/long-seq.png stylemax-height: 100%; max-width: 100%; />/td> td>Ronghang Hu, Shoubhik Debnath, Saining Xie, b>Xinlei Chen/b>. Exploring Long-Sequence Masked Autoencoders. ArXiv, 2022. /td> td>a hrefhttps://arxiv.org/abs/2210.07224>Link/a> a hrefhttps://github.com/facebookresearch/long_seq_mae>Code/a> /td> /tr> tr> td>img srcimages/ttt_mae.png stylemax-height: 100%; max-width: 100%; />/td> td>Yossi Gandelsman*, Yu Sun*, b>Xinlei Chen/b>, Alexei Efros. Test-Time Training with Masked Autoencoders. The 36th Conference on Neural Information Processing Systems (NeurIPS), 2022. /td> td>a hrefhttps://arxiv.org/abs/2209.07522>Link/a> a hrefhttps://yossigandelsman.github.io/ttt_mae/index.html>Project/a> a hrefhttps://github.com/yossigandelsman/test_time_training_mae>Code/a> /td> /tr> tr> td>img srcimages/asym.png stylemax-height: 100%; max-width: 100%; />/td> td>Xiao Wang*, Haoqi Fan*, Yuandong Tian, Daisuke Kihara, b>Xinlei Chen/b>. On the Importance of Asymmetry for Siamese Representation Learning. The 35th Conference on Computer Vision and Pattern Recognition (CVPR), 2022. /td> td>a hrefhttps://arxiv.org/abs/2204.00613>Link/a> a hrefhttps://github.com/facebookresearch/asym-siam>Code/a> /td> /tr> tr> td>img srcimages/point-cont.png stylemax-height: 100%; max-width: 100%; />/td> td>Yutong Bai, b>Xinlei Chen/b>, Alexander Kirillov, Alan Yuile, Alex Berg. Point-Level Region Contrast for Object Detection Pre-Training. The 35th Conference on Computer Vision and Pattern Recognition (CVPR), 2022. b>Oral, Best Paper Finalist/b>. /td> td>a hrefhttps://arxiv.org/abs/2202.04639>Link/a> a hrefhttps://github.com/facebookresearch/PLRC>Code/a> /td> /tr> tr> td>img srcimages/mae.png stylemax-height: 100%; max-width: 100%; />/td> td>Kaiming He*†, b>Xinlei Chen/b>*, Saining Xie, Yanghao Li, Piotr Dollár, Ross Girshick. Masked Autoencoders Are Scalable Vision Learners. The 35th Conference on Computer Vision and Pattern Recognition (CVPR), 2022. b>Oral, Best Paper Finalist/b>. /td> td>a hrefhttps://arxiv.org/abs/2111.06377>Link/a> a hrefhttps://github.com/facebookresearch/mae>Code/a> /td> /tr> tr> td>img srcimages/vit-det.png stylemax-height: 100%; max-width: 100%; />/td> td>Yanghao Li, Saining Xie, b>Xinlei Chen/b>, Piotr Dollár, Kaiming He, Ross Girshick. Benchmarking Detection Transfer Learning with Vision Transformers. ArXiv, 2021/td> td>a hrefhttps://arxiv.org/abs/2111.11429>Link/a> a hrefhttps://github.com/facebookresearch/detectron2/tree/main/projects/ViTDet>Code/a> /td> /tr> tr> td>img srcimages/mocov3.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>*, Saining Xie*, Kaiming He. An Empirical Study of Training Self-Supervised Vision Transformers. The 18th International Conference on Computer Vision (ICCV), 2021. b>Oral/b>. /td> td>a hrefhttps://arxiv.org/abs/2104.02057>Link/a> a hrefhttps://github.com/facebookresearch/moco-v3>Code/a> /td> /tr> tr> td>img srcimages/ussl_nc.png stylemax-height: 100%; max-width: 100%; />/td> td>Yuandong Tian, b>Xinlei Chen/b>, Surya Ganguli. Understanding Self-supervised Learning Dynamics without Contrastive Pairs. The 38th International Conference on Machine Learning (ICML), 2021. b>Long Oral, Outstanding Paper Honorable Mention/b>. /td> td>a hrefhttps://arxiv.org/abs/2102.06810>Link/a> a hrefhttps://yuandong-tian.com/understand_byol2.mp4>Video/a> a hrefhttps://github.com/facebookresearch/luckmatters/tree/master/ssl>Code/a> /td> /tr> tr> td>img srcimages/simsiam.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Kaiming He. Exploring Simple Siamese Representation Learning. The 34th Conference on Computer Vision and Pattern Recognition (CVPR), 2021. b>Oral, Best Paper Honorable Mention/b>. /td> td>a hrefhttps://arxiv.org/abs/2011.10566>Link/a> a hrefhttps://github.com/facebookresearch/simsiam>Code/a> /td> /tr> tr> td>img srcimages/krisp.png stylemax-height: 100%; max-width: 100%; />/td> td>Kenneth Marino, b>Xinlei Chen/b>, Devi Parikh, Abhinav Gupta, Marcus Rohrbach. KRISP: Integrating Implicit and Symbolic Knowledge for Open-Domain Knowledge-Based VQA. The 34th Conference on Computer Vision and Pattern Recognition (CVPR), 2021. /td> td>a hrefhttps://arxiv.org/abs/2012.11014>Link/a> a hrefhttps://github.com/facebookresearch/mmf/tree/master/projects/krisp>Code/a> /td> /tr> tr> td>img srcimages/ussl.png stylemax-height: 100%; max-width: 100%; />/td> td>Yuandong Tian, Lantao Yu, b>Xinlei Chen/b>, Surya Ganguli. Understanding Self-supervised Learning with Dual Deep Networks. ArXiv, 2020. /td> td>a hrefhttps://arxiv.org/abs/2010.00578>Link/a> a hrefhttps://yuandong-tian.com/understand_simclr2.mp4>Video/a> /td> /tr> tr> td>img srcimages/count-mech.png stylemax-height: 100%; max-width: 100%; />/td> td>Corentin Dancette*, Remi Cadene*, b>Xinlei Chen/b>, Matthieu Cord. Overcoming Statistical Shortcuts for Open-ended Visual Counting. ArXiv, 2020. /td> td>a hrefhttps://arxiv.org/abs/2006.10079>Link/a> a hrefhttps://github.com/cdancette/spatial-counting-network>Code/a> /td> /tr> tr> td>img srcimages/grid-count.png stylemax-height: 100%; max-width: 100%; />/td> td>Duy-Kien Nguyen, Vedanuj Goswami, b>Xinlei Chen/b>. Revisiting Modulated Convolutions for Visual Counting and Beyond. The 9th International Conference on Learning Representations (ICLR), 2021. b>VQA 2020 Challenge Winner/b>. /td> td>a hrefhttps://arxiv.org/abs/2004.11883>Link/a> a hrefhttps://github.com/facebookresearch/mmf/tree/master/projects/movie_mcan>Code/a> /td> /tr>!-- tr> td>img srcimages/roomnav.png stylemax-height: 100%; max-width: 100%; />/td> td>Medhini Narasimhan, Erik Wijmans, b>Xinlei Chen/b>, Trevor Darrell, Dhruv Batra, Devi Parikh, Amanpreet Singh. Seeing the Un-Scene: Learning Amodal Semantic Maps for Room Navigation. The 16th European Conference on Computer Vision (ECCV), 2020. /td> td>a hrefhttps://arxiv.org/abs/2007.09841>Link/a> /td> /tr> --> tr> td>img srcimages/mocov2.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Haoqi Fan, Ross Girshick, Kaiming He. Improved Baselines with Momentum Contrastive Learning. ArXiv, 2020. /td> td>a hrefhttps://arxiv.org/abs/2003.04297>Link/a> a hrefhttps://github.com/facebookresearch/moco>Code/a> /td> /tr> tr> td>img srcimages/imvote.png stylemax-height: 100%; max-width: 100%; />/td> td>Charles R. Qi*, b>Xinlei Chen/b>*, Or Litany, Leonidas J. Guibas. ImVoteNet: Boosting 3D Object Detection in Point Clouds with Image Votes. The 33rd Conference on Computer Vision and Pattern Recognition (CVPR), 2020. /td> td>a hrefhttps://arxiv.org/abs/2001.10692>Link/a> a hrefhttps://github.com/facebookresearch/imvotenet>Code/a> /td> /tr> tr> td>img srcimages/vqa-grid.png stylemax-height: 100%; max-width: 100%; />/td> td>Huaizu Jiang, Ishan Misra, Marcus Rohrbach, Erik Learned-Miller, b>Xinlei Chen/b>. In Defense of Grid Features for Visual Question Answering. The 33rd Conference on Computer Vision and Pattern Recognition (CVPR), 2020. /td> td>a hrefhttps://arxiv.org/abs/2001.03615>Link/a> a hrefhttps://github.com/facebookresearch/grid-feats-vqa>Code/a> /td> /tr> tr> td>img srcimages/tmask.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Ross Girshick, Kaiming He, Piotr Dollár. TensorMask: A Foundation for Dense Object Segmentation. The 17th International Conference on Computer Vision (ICCV), 2019. /td> td>a hrefhttps://arxiv.org/abs/1903.12174>Link/a> a hrefhttps://github.com/facebookresearch/detectron2/tree/master/projects/TensorMask>Code/a> /td> /tr> tr> td>img srcimages/minecraft.png stylemax-height: 100%; max-width: 100%; />/td> td>Zhuoyuan Chen*, Demi Guo*, Tong Xiao*, Saining Xie, b>Xinlei Chen/b>, Haonan Yu, Jonathan Gray, Kavya Srinet, Haoqi Fan, Jerry Ma, Charles R Qi, Shubham Tulsiani, Arthur Szlam, C Lawrence Zitnick. Order-Aware Generative Modeling Using the 3D-Craft Dataset. The 17th International Conference on Computer Vision (ICCV), 2019. /td> td>a hrefhttp://openaccess.thecvf.com/content_ICCV_2019/papers/Chen_Order-Aware_Generative_Modeling_Using_the_3D-Craft_Dataset_ICCV_2019_paper.pdf>Link/a> a hrefhttps://github.com/facebookresearch/voxelcnn>Code/a> a hrefhttps://github.com/facebookresearch/craftassist>Data/a> /td> /tr> tr> td>img srcimages/nocaps.png stylemax-height: 100%; max-width: 100%; />/td> td>Harsh Agrawal*, Karan Desai*, Yufei Wang, b>Xinlei Chen/b>, Rishabh Jain, Mark Johnson, Dhruv Batra, Devi Parikh, Stefan Lee, Peter Anderson. nocaps: novel object captioning at scale. The 17th International Conference on Computer Vision (ICCV), 2019. /td> td>a hrefhttps://arxiv.org/abs/1812.08658>Link/a> a hrefhttps://nocaps.org/>Project/a> /td> /tr> tr> td>img srcimages/cycle.png stylemax-height: 100%; max-width: 100%; />/td> td>Meet Shah, b>Xinlei Chen/b>, Marcus Rohrbach, Devi Parikh. Cycle-Consistency for Robust Visual Question Answering. The 32nd Conference on Computer Vision and Pattern Recognition (CVPR), 2019. b>Oral/b>. /td> td>a hrefhttps://arxiv.org/abs/1902.05660>Link/a> a hrefhttps://facebookresearch.github.io/VQA-Rephrasings/>Project/a> a hrefhttps://github.com/facebookresearch/mmf/tree/project/cycle-consistency>Code/a> /td> /tr> tr> td>img srcimages/ground.png stylemax-height: 100%; max-width: 100%; />/td> td>Luowei Zhou, Yannis Kalantidis, b>Xinlei Chen/b>, Jason J. Corso, Marcus Rohrbach. Grounded Video Description. The 32nd Conference on Computer Vision and Pattern Recognition (CVPR), 2019. b>Oral/b>. /td> td>a hrefhttps://arxiv.org/abs/1812.06587>Link/a> a hrefhttps://github.com/facebookresearch/ActivityNet-Entities>Code(Data)/a> /td> /tr> tr> td>img srcimages/mteqa.png stylemax-height: 100%; max-width: 100%; />/td> td>Licheng Yu, b>Xinlei Chen/b>, Georgia Gkioxari, Mohit Bansal, Tamara L. Berg, Dhruv Batra. Multi-Target Embodied Question Answering. The 32nd Conference on Computer Vision and Pattern Recognition (CVPR), 2019. /td> td>a hrefhttps://arxiv.org/abs/1904.04686>Link/a> a hrefhttps://youtu.be/pK5gYk9OgjE>Video/a> a hrefhttps://github.com/facebookresearch/MT-EQA>Code/a> /td> /tr> tr> td>img srcimages/textvqa.png stylemax-height: 100%; max-width: 100%; />/td> td>Amanpreet Singh, Vivek Natarajan, Meet Shah, Yu Jiang, b>Xinlei Chen/b>, Dhruv Batra, Devi Parikh, Marcus Rohrbach. Towards VQA Models That Can Read. The 32nd Conference on Computer Vision and Pattern Recognition (CVPR), 2019. /td> td>a hrefhttps://arxiv.org/abs/1904.08920>Link/a> a hrefhttps://textvqa.org/>Project/a> a hrefhttps://github.com/facebookresearch/mmf>Code/a> /td> /tr> tr> td>img srcimages/codraw.png stylemax-height: 100%; max-width: 100%; />/td> td>Jin-Hwa Kim, Nikita Kitaev, b>Xinlei Chen/b>, Marcus Rohrbach, Yuandong Tian, Dhruv Batra, Devi Parikh. CoDraw: Collaborative Drawing as a Testbed for Grounded Goal-driven Communication. The 57th Annual Meeting of the Association for Computational Linguistics (ACL), 2019. /td> td>a hrefhttps://arxiv.org/abs/1712.05558>Link/a> a hrefhttps://github.com/facebookresearch/CoDraw>Code(Data)/a> /td> /tr> tr> td>img srcimages/pythia.png stylemax-height: 100%; max-width: 100%; />/td> td>Yu Jiang*, Vivek Natarajan*, b>Xinlei Chen/b>*, Marcus Rohrbach, Dhruv Batra, Devi Parikh. Pythia v0.1: the Winning Entry to the VQA Challenge 2018. ArXiv, 2018. b>VQA 2018 Challenge Winner/b>. /td> td>a hrefhttps://arxiv.org/abs/1807.09956>Link/a> a hrefhttps://github.com/facebookresearch/mmf>Code/a> /td> /tr> tr> td>img srcimages/iter.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Li-Jia Li, Li Fei-Fei, Abhinav Gupta. Iterative Visual Reasoning Beyond Convolutions. The 31st Conference on Computer Vision and Pattern Recognition (CVPR), 2018. b>Spotlight/b>. /td> td>a hrefhttps://arxiv.org/abs/1803.11189>Link/a> a hrefpapers/cvpr18.pdf>PDF/a> a hrefhttps://github.com/endernewton/iter-reason>Code/a> /td> /tr> tr> td>img srcimages/thesis.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>. Visual Knowledge Learning. Doctoral Dissertation, CMU-LTI-18-001. /td> td>a hrefpapers/thesis.pdf>PDF/a> a hrefdocs/thesis/presentation.pdf>Slides/a> /td> /tr> tr> td>img srcimages/smn.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Abhinav Gupta. Spatial Memory for Context Reasoning in Object Detection. The 16th International Conference on Computer Vision (ICCV), 2017. /td> td>a hrefhttps://arxiv.org/abs/1704.04224>Link/a> a hrefpapers/iccv17.pdf>PDF/a> a hrefdocs/smn/poster.pdf>Poster/a>/td> /tr> tr> td>img srcimages/imp.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Abhinav Gupta. An Implementation of Faster RCNN with Study for Region Sampling. ArXiv, 2017. /td> td>a hrefhttps://arxiv.org/abs/1702.02138>Link/a> a hrefhttps://github.com/endernewton/tf-faster-rcnn>Code/a>/td> /tr> tr> td>img srcimages/pixel.png stylemax-height: 100%; max-width: 100%; />/td> td>Aayush Bansal, b>Xinlei Chen/b>, Bryan Russell, Abhinav Gupta, Deva Ramanan. PixelNet: Representation of the pixels, by the pixels, and for the pixels. ArXiv, 2017. /td> td>a hrefhttps://arxiv.org/abs/1702.06506>Link/a> a hrefhttp://www.cs.cmu.edu/~aayushb/pixelNet>Project/a> a hrefhttps://github.com/aayushbansal/PixelNet>Code/a> a hrefhttps://github.com/endernewton/PixelNet>Code(Old)/a> /td> /tr> tr> td>img srcimages/story.png stylemax-height: 100%; max-width: 100%; />/td> td>Gunnar A. Sigurdsson, b>Xinlei Chen/b>, Abhinav Gupta. Learning Visual Storylines with Skipping Recurrent Neural Networks. The 14th European Conference on Computer Vision (ECCV), 2016. /td> td>a hrefhttps://arxiv.org/abs/1604.04279>Link/a> a hrefpapers/eccv16.pdf>PDF/a> a hrefdocs/story/P-3A-26.pdf>Poster/a> a hrefhttps://github.com/gsig/srnn>Code/a>/td> /tr> tr> td>img srcimages/vis.png stylemax-height: 100%; max-width: 100%; />/td> td>Jiwei Li, b>Xinlei Chen/b>, Eduard Hovy, Dan Jurafsky. Visualizing and Understanding Neural Models in NLP. Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 2016. /td> td>a hrefhttp://arxiv.org/abs/1506.01066>Link/a> a hrefpapers/naacl16.pdf>PDF/a> a hrefhttps://github.com/jiweil/Visualizing-and-Understanding-Neural-Models-in-NLP>Code/a>/td> /tr> tr> td>img srcimages/web.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Abhinav Gupta. Webly Supervised Learning of Convolutional Networks. The 15th International Conference on Computer Vision (ICCV), 2015. b>Oral/b>. /td> td>a hrefhttp://arxiv.org/abs/1505.01554>Link/a> a hrefpapers/iccv15.pdf>PDF/a> a hrefdocs/deil/web_cnn_v7.pptx>Slides/a> a hrefdocs/deil/poster.pdf>Poster/a> a hrefhttps://www.youtube.com/watch?v8urSe7hkfYM>Video/a> a hrefhttp://videolectures.net/iccv2015_chen_supervised_learning/>Talk/a> a hrefhttps://github.com/endernewton/webly-supervised>Code/a> a hrefweb.html>Project/a>/td> /td> /tr> tr> td>img srcimages/coco.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollar, C. Lawrence Zitnick. Microsoft COCO Captions: Data Collection and Evaluation Server. Arxiv Preprint, 2015. /td> td>a hrefhttp://arxiv.org/abs/1504.00325>Link/a> a hrefhttps://github.com/tylin/coco-caption>Code/a>/td> /tr> tr> td>img srcimages/rnn.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, C. Lawrence Zitnick. Minds Eye: A Recurrent Visual Representation for Image Caption Generation. The 28th Conference on Computer Vision and Pattern Recognition (CVPR), 2015. /td> td>a hrefpapers/cvpr15_rnn.pdf>PDF/a> a hrefdocs/rvr/RVR-1.pdf>Poster/a>/td> /tr> tr> td> img srcimages/sense.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Alan Ritter, Abhinav Gupta, Tom Mitchell. Sense Discovery via Co-Clustering on Images and Text. The 28th Conference on Computer Vision and Pattern Recognition (CVPR), 2015. /td> td>a hrefpapers/cvpr15_sense.pdf>PDF/a> a hrefdocs/neil-poly/Sense-4.pdf>Poster/a> a hrefpoly.html>Project/a>/td> /tr>!-- tr> td>img srcimages/nel.png stylemax-height: 100%; max-width: 100%; />/td> td>Tom M. Mitchell, William Cohen, Estevam Hruschka, Partha Talukdar, Justin Betteridge, Andrew Carlson, Bhavana Dalvi, Matt Gardner, Bryan Kisiel, Jayant Krishnamurthy, Ni Lao, Kathryn Mazaitis, Thahir Mohamed, Ndapa Nakashole, Emmanouil Platanios, Alan Ritter, Mehdi Samadi, Burr Settles, Richard Wang, Derry Wijaya, Abhinav Gupta, b>Xinlei Chen/b>, Abulhair Saparov, Malcom Greaves and Joel Welling. Never-Ending Learning. The 29th AAAI Conference on Artificial Intelligence (AAAI), 2015. /td> td>a hrefpapers/aaai15.pdf>PDF/a>/td> /tr> --> !-- tr> td>img srcimages/fmri.png stylemax-height: 100%; max-width: 100%; />/td> td>Elissa M. Aminoff, Mariya Toneva, Abhinav Shrivastava, b>Xinlei Chen/b>, Ishan Misra, Abhinav Gupta, Michael Tarr. Applying artificial vision models to human scene understanding. Frontiers in Computational Neuroscience, 2015. /td> td>a hrefpapers/fncom.pdf>PDF/a>/td> /tr> --> tr> td> img srcimages/neil-seg.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Abhinav Shrivastava, Abhinav Gupta. Enriching Visual Knowledge Bases via Object Discovery and Segmentation. The 27th Conference on Computer Vision and Pattern Recognition (CVPR), 2014. /td> td>a hrefpapers/cvpr14.pdf>PDF/a>a hrefdocs/neil-seg/poster_final2.pdf>Poster/a>a hrefseg.html>Project/a>a hrefhttps://github.com/endernewton/subdiscover>Code/a>/td> /tr> tr> td> img srcimages/neil.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Abhinav Shrivastava, Abhinav Gupta. NEIL: Extracting Visual Knowledge from Web Data. The 14th International Conference on Computer Vision (ICCV), 2013. b>Oral/b>./td> td>a hrefpapers/iccv13.pdf>PDF/a> a hrefdocs/neil/neil_iccv_final.pptx>Slides/a> a hrefdocs/neil/NEIL_poster.pdf>Poster/a> a hrefhttp://www.neil-kb.com>Web/a> a hrefhttp://techtalks.tv/talks/neil-extracting-visual-knowledge-from-web-data/59408/>Talk/a> a hrefhttps://github.com/endernewton/subdiscover>Code/a> a hrefhttps://github.com/endernewton/neil-test>Test Code/a>/td> /tr> tr> td> img srcimages/lsc.png stylemax-height: 100%; max-width: 100%; />/td> td>b>Xinlei Chen/b>, Deng Cai. Large Scale Spectral Clustering with Landmark-based Representation. The 26th AAAI Conference on Artificial Intelligence (AAAI), 2011. /td> td>a hrefpapers/aaai11.pdf>PDF/a>a hrefhttp://www.cad.zju.edu.cn/home/dengcai/Data/ReproduceExp.html#LSC>Code/a>/td> /tr> tr> td> img srcimages/mesh.bmp stylemax-height: 100%; max-width: 100%; />/td> td>Jiajun Lv, b>Xinlei Chen/b>, Jin Huang, Hujun Bao. Semi-supervised Mesh Segmentation and Labeling. The 20th Pacific Conference on Computer Graphics and Applications (PG), 2012. /td> td>a hrefpapers/pg12.pdf>PDF/a>a hrefhttp://www.jiajunlu.com/SMSL.htm>Project/a>/td> /tr> /tbody> /table> footer> div classrow> ul classlist-unstyled> li classpull-right>a href#top>Back to top/a>/li> /ul> /div> div styletext-align: center> span>© script>document.write(new Date().getFullYear())/script> Xinlei Chen, /span> span>Website based on a hrefhttp://getbootstrap.com relnofollow>Bootstrap/a> and a hrefhttp://bootswatch.com/ relnofollow>Bootswatch/a>./span> /div> /footer> /div> script typetext/javascript srcjs/jquery-1.11.1.min.js>/script> script typetext/javascript srcjs/bootstrap.min.js>/script>/body>/html>
View on OTX
|
View on ThreatMiner
Please enable JavaScript to view the
comments powered by Disqus.
Data with thanks to
AlienVault OTX
,
VirusTotal
,
Malwr
and
others
. [
Sitemap
]