-
Nanyang Technological University
- 新加坡
- https://jiaxiaojunqaq.github.io/
Stars
Safety at Scale: A Comprehensive Survey of Large Model Safety
Code for "TrustRAG: Enhancing Robustness and Trustworthiness in RAG"
AriaUI / Awesome-GUI-Agent
Forked from showlab/Awesome-GUI-Agent💻 A curated list of papers and resources for multi-modal Graphical User Interface (GUI) agents.
A GUI Agent application based on UI-TARS(Vision-Lanuage Model) that allows you to control your computer using natural language.
This is the official repository of [[2408.10848\] Perception-guided Jailbreak against Text-to-Image Models](https://arxiv.org/abs/2408.10848) The paper is accepted by AAAI 2025.
Fine-tuning base models to build robust task-specific models
Every practical and proposed defense against prompt injection.
[CVPR2024]Focus on Hiders: Exploring Hidden Threats for Enhancing Adversarial Training
Agent Laboratory is an end-to-end autonomous research workflow meant to assist you as the human researcher toward implementing your research ideas
Open-sourced, Fast and Context-aware Action Grounding from GUI Instructions for GUI/Computer-use Agents
Code&Data for the paper "Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based Agents" [NeurIPS 2024]
Code and data to go with the Zhu et al. paper "An Objective for Nuanced LLM Jailbreaks"
Aligning LMMs with Factually Augmented RLHF
The reinforcement learning codes for dataset SPA-VL
Official Code and data for ACL 2024 finding, "An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language Models"
This is the code repository of our submission: Understanding the Dark Side of LLMs’ Intrinsic Self-Correction.
[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate
A collection of resources that investigate social agents.
Code repository for the paper "Heuristic Induced Multimodal Risk Distribution Jailbreak Attack for Multimodal Large Language Models"
Collection of evals for Inspect AI