- Видео 97
- Просмотров 37 673
CAMLIS
Добавлен 7 май 2018
Видео
AdapterSwap: Continuous Training of LLMs with Data Removal and Access-Control Guarantees
Просмотров 10921 день назад
CAMLIS 2024, William Fleshman
Defending Large Language Models Against Attacks With Residual Stream Activation Analysis
Просмотров 11021 день назад
CAMLIS 2024, Amelia Kawasaki, Andrew Davis, and Houssam Abbas
LLM Backdoor Activations Stick Together
Просмотров 9021 день назад
CAMLIS 2024, Tamás Vörös, Ben Gelman, Sean Bergeron, and Adarsh Kyadige
Curl: Private LLMs through Wavelet-Encoded Look-Up Tables
Просмотров 4421 день назад
CAMLIS 2024, Dimitris Mouris, Manuel dos Santos, Mehmet Ugurbil, Stanislaw Jarecki, José Reis, Shubho Sengupta, and Miguel de Vega
Hamm-Grams: Mining Common Regular Expressions via Locality Sensitive Hashing
Просмотров 3921 день назад
CAMLIS 2024, Derek Everett, Edward Raff, and James Holt
Let’s Make it Personal: Customizing Threat Intelligence with Metric Learning
Просмотров 9421 день назад
CAMLIS 2024, Christopher Galbraith
Defending Against Indirect Prompt Injection Attacks With Spotlighting
Просмотров 10521 день назад
CAMLIS 2024, Keegan Hines
Keynote - Acting to Ensure AI Benefits Cyber Defense in a Decade of Technological Surprise
Просмотров 16321 день назад
CAMLIS 2024, Joshua Saxe
Is F1 Score Suboptimal for Cybersecurity Models? Introducing Cscore
Просмотров 9221 день назад
CAMLIS 2024, Manish Marwah, Asad Narayanan, Stephan Jou, Martin Arlitt, and Maria Pospelova
Structure and Semantics-Aware Malware Classification with Vision Transformers
Просмотров 9521 день назад
CAMLIS 2024, David Krisiloff, and Scott Coull
PEVuln: A Benchmark Dataset for Using Machine Learning to Detect Vulnerabilities in PE Malware
Просмотров 7721 день назад
CAMLIS 2024, Nathan Ross, Oluwafemi Olukoya, Jesus Martinez del Rincon, and Domhnall Carlin
DIP-ECOD: Improving Anomaly Detection in Multimodal Distributions
Просмотров 4021 день назад
CAMLIS 2024, Kaixi Yang, Paul Miller, and Jesus Martinez del Rincon
LLM Agents for Vulnerability Identification and Verification of CVEs
Просмотров 7921 день назад
CAMLIS 2024, Rodrigo Bersa, Tadesse Zemicheal, Shawn Davis, and Hsin Chen
End-to-End Framework using LLMs for Technique Identification and Threat-Actor Attribution
Просмотров 25221 день назад
CAMLIS 2024, Kyla Guru, Robert Moss, and Mykel Kochenderfer
Towards Autonomous Cyber-Defence: Using Co-Operative Decision Making for Cybersecurity
Просмотров 7521 день назад
Towards Autonomous Cyber-Defence: Using Co-Operative Decision Making for Cybersecurity
PyRIT: A Framework for Security Risk Identification and Red Teaming in Generative AI Systems
Просмотров 22521 день назад
PyRIT: A Framework for Security Risk Identification and Red Teaming in Generative AI Systems
Cybersecurity and Infrastructure Security Agency (CISA)
Просмотров 31721 день назад
Cybersecurity and Infrastructure Security Agency (CISA)
Graph-Based User-Entity Behavior Analytics for Enterprise Insider Threat Detection
Просмотров 274Год назад
Graph-Based User-Entity Behavior Analytics for Enterprise Insider Threat Detection
Don’t you forget NLP: prompt injection using repeated sequences in ChatGPT
Просмотров 319Год назад
Don’t you forget NLP: prompt injection using repeated sequences in ChatGPT
Playing Defense: Benchmarking Cybersecurity Capabilities of Large Language Models
Просмотров 257Год назад
Playing Defense: Benchmarking Cybersecurity Capabilities of Large Language Models
LLM Prompt Injection: Attacks and Defenses
Просмотров 2 тыс.Год назад
LLM Prompt Injection: Attacks and Defenses
Model Leeching: An Extraction Attack Targeting LLMs
Просмотров 345Год назад
Model Leeching: An Extraction Attack Targeting LLMs
Anomaly Detection of Command Shell Sessions based on DistilBERT: Unsupervised and Supervised Appro
Просмотров 162Год назад
Anomaly Detection of Command Shell Sessions based on DistilBERT: Unsupervised and Supervised Appro
Razing to the Ground Machine Learning Phishing Webpage Detectors with Query Efficient Adversarial HT
Просмотров 88Год назад
Razing to the Ground Machine Learning Phishing Webpage Detectors with Query Efficient Adversarial HT
Web content filtering through knowledge distillation of Large Language Models
Просмотров 119Год назад
Web content filtering through knowledge distillation of Large Language Models
Compilation as a Defense: Enhancing DL Model Attack Robustness via Tensor Optimization
Просмотров 60Год назад
Compilation as a Defense: Enhancing DL Model Attack Robustness via Tensor Optimization
Multi-Agent Reinforcement Learning for Maritime Operational Technology Cyber Security
Просмотров 213Год назад
Multi-Agent Reinforcement Learning for Maritime Operational Technology Cyber Security
Enhancing Exfiltration Path Analysis Using Reinforcement Learning
Просмотров 91Год назад
Enhancing Exfiltration Path Analysis Using Reinforcement Learning
Keynote - Security Issues in Generative AI
Просмотров 242Год назад
Keynote - Security Issues in Generative AI
Enjoyed the high quality of this presentation ❤
Shopa chittagong panch laish khaleq dadu shohal bepul upstairs Akbar dadu pervin fufu mother's name same mine baba Azizul Haq next at Dhaka elephant road please let me know how is your parents and your other two sisters
❤
Amazing editing and story-telling - outstanding speaker. Thank you for the video!
How did you draw the loss surface?
🎯 Key Takeaways for quick navigation: 00:00 🎤 *The speaker, Mark Brighton Boach, introduces himself as a security engineer at Dropbox and discusses the topic of an AI attack called "repeated sequences."* 01:09 📦 *Dropbox is interested in AI to improve its services, including AI-powered search and question answering across Dropbox files.* 03:26 💻 *The AI/ML team at Dropbox collaborated with the security team to investigate a novel AI attack involving repeated sequences in prompts.* 05:57 🔍 *Backspaces, control characters, and meta-characters were found to produce unexpected behaviors in Transformer-based chat models.* 08:56 🕵️ *Dropbox engaged with OpenAI to address the issue but had to work on their own. They planned to open-source their findings.* 11:13 🧐 *Spaces, backslashes, and other character sequences were identified as risky inputs that could disrupt AI models.* 15:18 ⏱️ *A moderator framework was proposed to detect and block risky inputs, such as repeated sequences, to ensure model safety.* 17:32 🧰 *The repeated sequence moderator analyzes prompts for suspicious sequences and flags them based on repeat count and score.* 21:05 ❓ *The speaker addresses questions about mixing and matching special characters and discusses the potential effects.* 23:00 🤖 *The speaker considers the model's training and behavior regarding certain tokens and suggests further exploration.* 24:10 🌐 *The speaker discusses the use of UTF-8 byte order marks and Unicode characters in experiments involving repeated sequences but didn't explore Unicode bomb orderings in their work.* 25:31 📞 *A question is raised about whether prompt injection attacks are fundamentally broken and similar to phone freaking, and the speaker acknowledges the complexity of the issue without a clear solution.* Made with HARPA AIwha
🙂 *Promo sm*
So humor and excellent talk! Now I'm the fan of Nicolas Carlini
Great talk! About to head out and read the paper now. Thank you very much!
Great talk! Thanks for sharing!
Happy this guy is on our team...great advantage to not have a genius like this on the other side.
Very informative and well-organized presentation. Thank you!!
Great talk! His comparison of current state of the robustness of ML models with the crypto in 1920's is spot on.
Great talk. It's such a difficult problem that is at the heart of generalisation. The google brain paper "Adversarial Examples that Fool both Computer Vision and Time-Limited Humans" is worth a read.
Damn, this talk is epic! [ Although I disagree with his answer to the question in 39:20 'the human is the proof-of-concept that adversarial robustness exists'; humans have their own set of singular errors constrained by their visual system which gives rise to visual/optical illusions, so perhaps the goal should not be to avoid adversarial examples altogether, but to somehow align/evaluate the mistakes the machines makes with that of the human. Similar to: www.pnas.org/content/pnas/117/47/29330.full.pdf ]
An excellent talk, thank you!
only 421 views?
ruclips.net/video/otmt0-cewbc/видео.html
ruclips.net/video/MjhDkNmtjy0/видео.html
ruclips.net/video/h1BQPV-iCkU/видео.html
Activism, Human Nature, & American Repentance: Noam Chomsky interviewed by Shmuly Yanklowitz ruclips.net/video/EA7Dj-dmzEE/видео.html
Psychological Slavery Full Episode | American Black Journal ruclips.net/video/vKh0H3sg1Uo/видео.html
President John F. Kennedy's Civil Rights Address ruclips.net/video/7BEhKgoA86U/видео.html
How does this talk only has 300 views and two comments? This man's talk is absolutely brilliant.
very useful. Thank you!
great talk