Tag: large language model
-
Schneier on Security: Prompt Injection Defenses Against LLM Cyberattacks
Source URL: https://www.schneier.com/blog/archives/2024/11/prompt-injection-defenses-against-llm-cyberattacks.html Source: Schneier on Security Title: Prompt Injection Defenses Against LLM Cyberattacks Feedly Summary: Interesting research: “Hacking Back the AI-Hacker: Prompt Injection as a Defense Against LLM-driven Cyberattacks“: Large language models (LLMs) are increasingly being harnessed to automate cyberattacks, making sophisticated exploits more accessible and scalable. In response, we propose a new defense…
-
Schneier on Security: Subverting LLM Coders
Source URL: https://www.schneier.com/blog/archives/2024/11/subverting-llm-coders.html Source: Schneier on Security Title: Subverting LLM Coders Feedly Summary: Really interesting research: “An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities against Strong Detection“: Abstract: Large Language Models (LLMs) have transformed code com- pletion tasks, providing context-based suggestions to boost developer productivity in software engineering. As users often…
-
Simon Willison’s Weblog: yet-another-applied-llm-benchmark
Source URL: https://simonwillison.net/2024/Nov/6/yet-another-applied-llm-benchmark/#atom-everything Source: Simon Willison’s Weblog Title: yet-another-applied-llm-benchmark Feedly Summary: yet-another-applied-llm-benchmark Nicholas Carlini introduced this personal LLM benchmark suite back in February as a collection of over 100 automated tests he runs against new LLM models to evaluate their performance against the kinds of tasks he uses them for. There are two defining features…
-
The Register: Staff can’t code? No prob. Singapore superapp’s LLM whips up apps for them
Source URL: https://www.theregister.com/2024/11/06/grab_coding_llm/ Source: The Register Title: Staff can’t code? No prob. Singapore superapp’s LLM whips up apps for them Feedly Summary: NP-hard to NP at all Southeast Asia’s Uber-esque superapp, Grab, has developed a tool that allows its employees to build large language model (LLM) apps without coding.… AI Summary and Description: Yes Summary:…
-
Hacker News: WebRL: Training LLM Web Agents via Self-Evolving Online Reinforcement Learning
Source URL: https://arxiv.org/abs/2411.02337 Source: Hacker News Title: WebRL: Training LLM Web Agents via Self-Evolving Online Reinforcement Learning Feedly Summary: Comments AI Summary and Description: Yes Summary: The paper introduces WebRL, a novel framework that employs self-evolving online curriculum reinforcement learning to enhance the training of large language models (LLMs) as web agents. This development is…
-
Slashdot: Google’s Big Sleep LLM Agent Discovers Exploitable Bug In SQLite
Source URL: https://tech.slashdot.org/story/24/11/05/1532207/googles-big-sleep-llm-agent-discovers-exploitable-bug-in-sqlite?utm_source=rss1.0mainlinkanon&utm_medium=feed Source: Slashdot Title: Google’s Big Sleep LLM Agent Discovers Exploitable Bug In SQLite Feedly Summary: AI Summary and Description: Yes **Summary:** Google has leveraged a large language model (LLM) agent, “Big Sleep,” to identify a previously undiscovered memory vulnerability in SQLite, marking a significant advancement in automated vulnerability discovery. This initiative showcases…
-
The Register: Google claims Big Sleep ‘first’ AI to spot freshly committed security bug that fuzzing missed
Source URL: https://www.theregister.com/2024/11/05/google_ai_vulnerability_hunting/ Source: The Register Title: Google claims Big Sleep ‘first’ AI to spot freshly committed security bug that fuzzing missed Feedly Summary: You snooze, you lose, er, win Google claims one of its AI models is the first of its kind to spot a memory safety vulnerability in the wild – specifically an…