{"work":{"id":"c19e1b2b-ef3b-4af7-b0ee-6aae3c27fd7a","openalex_id":null,"doi":null,"arxiv_id":"2601.22060","raw_key":null,"title":"Vision-deepresearch: Incentivizing deepresearch capability in multimodal large language models.arXiv preprint arXiv:2601.22060, 2026","authors":null,"authors_text":"Wenxuan Huang, Yu Zeng, Qiuchen Wang, Zhen Fang, Shaosheng Cao, Zheng Chu, Qingyu Yin, Shuang Chen, Zhenfei Yin, Lin Chen, et al","year":2026,"venue":null,"abstract":null,"external_url":"https://arxiv.org/abs/2601.22060","cited_by_count":null,"metadata_source":"arxiv_reference","metadata_fetched_at":"2026-05-15T05:55:05.142748+00:00","pith_arxiv_id":null,"created_at":"2026-05-10T06:21:27.174628+00:00","updated_at":"2026-05-15T05:55:05.142748+00:00","title_quality_ok":true,"display_title":"Vision-deepresearch: Incentivizing deepresearch capability in multimodal large language models","render_title":"Vision-deepresearch: Incentivizing deepresearch capability in multimodal large language models"},"hub":{"state":{"work_id":"c19e1b2b-ef3b-4af7-b0ee-6aae3c27fd7a","tier":"hub","tier_reason":"10+ Pith inbound or 1,000+ external citations","pith_inbound_count":10,"external_cited_by_count":null,"distinct_field_count":4,"first_pith_cited_at":"2026-03-30T17:59:56+00:00","last_pith_cited_at":"2026-05-13T05:39:38+00:00","author_build_status":"not_needed","summary_status":"needed","contexts_status":"needed","graph_status":"needed","ask_index_status":"not_needed","reader_status":"not_needed","recognition_status":"not_needed","updated_at":"2026-05-17T00:39:19.102025+00:00","tier_text":"hub"},"tier":"hub","role_counts":[{"context_role":"background","n":4}],"polarity_counts":[{"context_polarity":"background","n":4}],"runs":{},"summary":{},"graph":{},"authors":[]}}