Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Baseline Defenses for Adversarial Attacks Against Aligned Language Models (2309.00614v2)

Published 1 Sep 2023 in cs.LG, cs.CL, and cs.CR

Abstract: As LLMs quickly become ubiquitous, it becomes critical to understand their security vulnerabilities. Recent work shows that text optimizers can produce jailbreaking prompts that bypass moderation and alignment. Drawing from the rich body of work on adversarial machine learning, we approach these attacks with three questions: What threat models are practically useful in this domain? How do baseline defense techniques perform in this new domain? How does LLM security differ from computer vision? We evaluate several baseline defense strategies against leading adversarial attacks on LLMs, discussing the various settings in which each is feasible and effective. Particularly, we look at three types of defenses: detection (perplexity based), input preprocessing (paraphrase and retokenization), and adversarial training. We discuss white-box and gray-box settings and discuss the robustness-performance trade-off for each of the defenses considered. We find that the weakness of existing discrete optimizers for text, combined with the relatively high costs of optimization, makes standard adaptive attacks more challenging for LLMs. Future research will be needed to uncover whether more powerful optimizers can be developed, or whether the strength of filtering and preprocessing defenses is greater in the LLMs domain than it has been in computer vision.

Overview of ICLR 2024 Conference Submission Formatting Guidelines

This document delineates the formatting instructions for submissions to the ICLR 2024 conference. It serves as a comprehensive guide crafted for authors navigating the specifics of preparing their papers in compliance with the prescribed guidelines.

Key Aspects of the Submission Process

The paper emphasizes the electronic submission process via the OpenReview platform, aligning with the NeurIPS format. Precision in adherence to these guidelines is crucial, as deviations can lead to rejection. The initiation of the submission process necessitates the use of specific \LaTeX{} style files, which ensure consistency across submissions.

General Formatting Specifications

The document outlines explicit parameters for text layout, including the restriction of text within a 5.5-inch wide and 9-inch long rectangle, with a left margin of 1.5 inches. This consistency is vital for maintaining uniformity across all accepted papers. The recommended font is Times New Roman at a 10-point size, which is standard within the scientific community for conference papers.

Heading Structures and Document Elements

A hierarchical structure for headings is mandated, with three levels detailed. Each level has distinct formatting rules involving text alignment, capitalization, and spacing, ensuring systematic organization of content. Furthermore, critical instructions concerning citations, figures, tables, and references are provided to ensure clarity and ease of reading.

Key Considerations and Constraints

One prominent stipulation is the restriction of the main text to a maximum of 9 pages, with unrestricted space for citations. This stipulation emphasizes the importance of concise writing. The paper also details formatting commands and suggestions for common LaTeX issues, which are critical for authors who may be less experienced with typesetting in LaTeX.

Implications and Future Considerations

The provision of such detailed formatting guidelines serves several purposes: fostering fairness in the evaluation process by standardizing submissions, ensuring accessibility and readability, and facilitating the comparison of academic work. As conferences continue to evolve alongside technological advances, it is anticipated that formatting guidelines will similarly adapt, potentially offering more automated formatting solutions.

Overall, this document offers an indispensable resource for researchers preparing submissions for ICLR 2024, underscoring the paramount importance of adherence to prescribed formatting criteria within academic publishing.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Neel Jain (13 papers)
  2. Avi Schwarzschild (35 papers)
  3. Yuxin Wen (33 papers)
  4. Gowthami Somepalli (20 papers)
  5. John Kirchenbauer (21 papers)
  6. Ping-yeh Chiang (16 papers)
  7. Micah Goldblum (96 papers)
  8. Aniruddha Saha (19 papers)
  9. Jonas Geiping (73 papers)
  10. Tom Goldstein (226 papers)
Citations (253)
X Twitter Logo Streamline Icon: https://streamlinehq.com
Reddit Logo Streamline Icon: https://streamlinehq.com