I am Jiazheng Xu, a second-year PhD student in Tsinghua University.
- 🔭 Interested in Machine Learning, Multimodal Generative Models.
- 🌱 Find my up-to-date publication list in Google Scholar! Some of my proud leading works:
- ImageReward (NeurIPS'23): the first general-purpose text-to-image human preference reward model (RM) for RLHF, outperforming CLIP/BLIP/Aesthetic by 30% in terms of human preference prediction.
- CogVLM (NeurIPS'24): a powerful open-source visual language model (VLM), which achieves state-of-the-art performance on 10 classic cross-modal benchmarks.
- CogAgent (CVPR'24): a visual agent being able to return a plan, next action, and specific operations with coordinates for any given task on any GUI screenshot, enhancing GUI-related question-answering capabilities.
- 💬 Feel free to drop me an email for:
- Any form of collaboration
- Any issue about my works or code
- Interesting ideas to discuss or just chatting