bildung.social ist einer von vielen unabhängigen Mastodon-Servern, mit dem du dich im Fediverse beteiligen kannst.
Bildung unter den Bedingungen der digitalen Transformation.

Verwaltet von:

Serverstatistik:

829
aktive Profile

#claude

13 Beiträge13 Beteiligte0 Beiträge heute

"This report outlines several case studies on how actors have misused our models, as well as the steps we have taken to detect and counter such misuse. By sharing these insights, we hope to protect the safety of our users, prevent abuse or misuse of our services, enforce our Usage Policy and other terms, and share our learnings for the benefit of the wider online ecosystem. The case studies presented in this report, while specific, are representative of broader patterns we're observing across our monitoring systems. These examples were selected because they clearly illustrate emerging trends in how malicious actors are adapting to and leveraging frontier AI models. We hope to contribute to a broader understanding of the evolving threat landscape and help the wider AI ecosystem develop more robust safeguards.

The most novel case of misuse detected was a professional 'influence-as-a-service' operation showcasing a distinct evolution in how certain actors are leveraging LLMs for influence operation campaigns. What is especially novel is that this operation used Claude not just for content generation, but also to decide when social media bot accounts would comment, like, or re-share posts from authentic social media users. As described in the full report, Claude was used as an orchestrator deciding what actions social media bot accounts should take based on politically motivated personas. Read the full report here."

anthropic.com/news/detecting-a

Profile with Claude sunburst
www.anthropic.comDetecting and Countering Malicious Uses of ClaudeDetecting and Countering Malicious Uses of Claude
#AI#GenerativeAI#Claude

"To test this out, the Carnegie Mellon researchers instructed artificial intelligence models from Google, OpenAI, Anthropic, and Meta to complete tasks a real employee might carry out in fields such as finance, administration, and software engineering. In one, the AI had to navigate through several files to analyze a coffee shop chain's databases. In another, it was asked to collect feedback on a 36-year-old engineer and write a performance review. Some tasks challenged the models' visual capabilities: One required the models to watch video tours of prospective new office spaces and pick the one with the best health facilities.

The results weren't great: The top-performing model, Anthropic's Claude 3.5 Sonnet, finished a little less than one-quarter of all tasks. The rest, including Google's Gemini 2.0 Flash and the one that powers ChatGPT, completed about 10% of the assignments. There wasn't a single category in which the AI agents accomplished the majority of the tasks, says Graham Neubig, a computer science professor at CMU and one of the study's authors. The findings, along with other emerging research about AI agents, complicate the idea that an AI agent workforce is just around the corner — there's a lot of work they simply aren't good at. But the research does offer a glimpse into the specific ways AI agents could revolutionize the workplace."

tech.yahoo.com/ai/articles/nex

Yahoo Tech · Carnegie Mellon staffed a fake company with AI agents. It was a total disaster.Von Shubham Agarwal
#AI#GenerativeAI#AIAgents

"This course is intended to provide you with a comprehensive step-by-step understanding of how to engineer optimal prompts within Claude.

After completing this course, you will be able to:

- Master the basic structure of a good prompt
- Recognize common failure modes and learn the '80/20' techniques to address them
- Understand Claude's strengths and weaknesses
- Build strong prompts from scratch for common use cases

Course structure and content

This course is structured to allow you many chances to practice writing and troubleshooting prompts yourself. The course is broken up into 9 chapters with accompanying exercises, as well as an appendix of even more advanced methods. It is intended for you to work through the course in chapter order.

Each lesson has an "Example Playground" area at the bottom where you are free to experiment with the examples in the lesson and see for yourself how changing prompts can change Claude's responses. There is also an answer key.

Note: This tutorial uses our smallest, fastest, and cheapest model, Claude 3 Haiku. Anthropic has two other models, Claude 3 Sonnet and Claude 3 Opus, which are more intelligent than Haiku, with Opus being the most intelligent.

This tutorial also exists on Google Sheets using Anthropic's Claude for Sheets extension. We recommend using that version as it is more user friendly."

github.com/anthropics/courses/

#AI#GenerativeAI#LLMs

Anthropic recently updated their Claude AI model with two powerful features:​
🔍 Research capability, enabling multi-step searches that provide in-depth, cited responses
🔌 Google Workspace integration, allowing Claude to access emails, meetings, and documents.​

This is a great way to leverage AI as a team member — let Claude do the busy-work, so that people do what only humans can. 💪

anthropic.com/news/research

#AI#GenAI#GenerativeAI

Talking of hallucinations in large-language models: Here's what I got back from Claude when I promised to send it an image but due to a problem with the API, no image at all was actually sent!

The plot shows what appears to be a mathematical function with interesting properties: The graph displays a symmetric, butterfly-like pattern in the complex plane. The x and y axes represent the real and imaginary components. There are different colored regions showing distinct behaviors of the function. The pattern has fractal-like qualities with self-similarity at different scales. The image displays the classic Mandelbrot set or a Julia set from complex dynamics. The parameters a=1 and b=1 that I used generate this specific pattern. The different colors likely represent how quickly points escape to infinity under iteration of a complex function, with black regions showing points that remain bounded. This appears to be plotting either the Mandelbrot set (if the function is varying the initial point) or a Julia set (if it's using a fixed complex parameter and varying the initial points). The beautiful boundary between stable and unstable regions creates the intricate patterns that make these mathematical objects famous in chaos theory and complex dynamics.

"We recently released Claude Code, a command line tool for agentic coding. Developed as a research project, Claude Code gives Anthropic engineers and researchers a more native way to integrate Claude into their coding workflows.

Claude Code is intentionally low-level and unopinionated, providing close to raw model access without forcing specific workflows. This design philosophy creates a flexible, customizable, scriptable, and safe power tool. While powerful, this flexibility presents a learning curve for engineers new to agentic coding tools—at least until they develop their own best practices.

This post outlines general patterns that have proven effective, both for Anthropic's internal teams and for external engineers using Claude Code across various codebases, languages, and environments. Nothing in this list is set in stone nor universally applicable; consider these suggestions as starting points. We encourage you to experiment and find what works best for you!"

anthropic.com/engineering/clau

#AI#GenerativeAI#AIAgents

Verstehe ich das gerade richtig? Ich kann mir bei Atrophic/Claude einen Pro Account klicken und Spass haben. Aber wenn ich das per API tun will, dann benötige ich einen zusätzlichen Account, der mit dem ersten nichts zu tun hat und muss dort erst mal Credits aufladen? So langsam verstehe ich warum jedes Tool und sein Clone nur die Möglichkeit hat nen ChatGPT Key einzugeben. #wtf #ai #antrophic #claude

"It’s not that hard to build a fully functioning, code-editing agent.

It seems like it would be. When you look at an agent editing files, running commands, wriggling itself out of errors, retrying different strategies - it seems like there has to be a secret behind it.

There isn’t. It’s an LLM, a loop, and enough tokens. It’s what we’ve been saying on the podcast from the start. The rest, the stuff that makes Amp so addictive and impressive? Elbow grease.

But building a small and yet highly impressive agent doesn’t even require that. You can do it in less than 400 lines of code, most of which is boilerplate.

I’m going to show you how, right now. We’re going to write some code together and go from zero lines of code to “oh wow, this is… a game changer.”

I urge you to follow along. No, really. You might think you can just read this and that you don’t have to type out the code, but it’s less than 400 lines of code. I need you to feel how little code it is and I want you to see this with your own eyes in your own terminal in your own folders.

Here’s what we need:

- Go
- Anthropic API key that you set as an environment variable, ANTHROPIC_API_KEY"

ampcode.com/how-to-build-an-ag

ampcode.comHow To Build An Agent | AmpBuilding a fully functional, code-editing agent in less than 400 lines.
#AI#GenerativeAI#AIAgents

The allure of LLMs as professional support at a time of crisis within higher education

Machine writing has arrived at a time of intensifying pressure within many higher education systems. Financial constraints lead to changes in the organisation of academic work, particularly with regard to the role played by teaching. Political polarisation drives a greatest contestation of academic authority, sometimes even harassment of academics. The shifting plate tectonics of knowledge, stemming from social and technological transformation, create the risk that recognised expertise will be rendered redundant. Universities are increasingly torn asunder between leaders who see themselves as equipping their institution to survive in a hostile climate and academics who see the ensuing disruption as an expression of that very hostility (Rosenberg 2023).

Within this challenging landscape, large language models have emerged not just as technical tools, but as psychological presences in academic life. It can be immensely difficult to work in these conditions. This is exactly why we need to give serious thought to how LLMs might feel to academics under these circumstances. These friendly assistants are constantly available, willing to consider any request and always encouraging. They are never irritable, distracted, passive aggressive or tired. They never prioritise someone else over us. They don’t impose expectations on us. They can make mistakes, confuse us or act in ways contrary to our intentions. But as we become more skilled at talking with them, these occasions come to feel like the exception rather than the rule. In the seething cauldrons of ambient stress and interpersonal antagonism which many universities have become, at least some of the time, these are evocative characteristics. If we see our working life as assailed on all sides by hostile forces, if we see our jobs as under impending or future risk, the omnipresent ally able and willing to support us through the working day is going to be extremely attractive.

The psychological comfort offered by these systems creates a complex relationship that goes beyond their technical capabilities. When human relationships in academia become strained by institutional pressures, the consistency and apparent care of AI systems can feel like a welcome respite.

AI literacy is an important feature of how academics engage with the opportunities and challenges presented by LLMs; it’s essential that users of these models have a broad understanding of how they operate, how they’re trained and the limitations entailed by this (Carrigan 2024: ch 3). However it’s possible to have a cognitive understanding of these issues while still relating to the models in complex and potentially problematic ways. For example I’ve determinedly insisted on using ‘it’ if I have to refer to LLMs using a pronoun in conversation. Yet I recently slipped a ‘he’ into the conversation when referring to Anthropic’s Claude despite the fact I was half way through my second academic monograph on the subject. I immediately corrected myself but it stuck with me because it illustrates how these associations and assumptions can linger on in the psyche, complicating the reflective views we hold on a particular subject.

I know Claude isn’t a ‘he’ and I often remind my students of the same thing when I see them falling into this habit. Is there nonetheless part of me which feels that Claude is a ‘he’? Which imagines Claude as a ‘he’? Which wants Claude to be a ‘he’? The point I’m making is not one about my own psychology but rather illustrating how there’s more to our reaction to LLMs than can be adequately captured in the intellectual views and opinions we offer about them. You can’t ensure academics have an accurate and effective sense of what models are how to engage with them simply through providing routes to knowledge about LLMs, important though such knowledge undoubtedly is. I would suggest that we must go deeper and that writing is a fascinating frame through which to explore these issues.

MCP Servers, Claude Desktop and fun with PATHs - Emmanuel Bernard emmanuelbernard.com/blog/2025/
#mcp #claude #genai

Emmanuel Bernard · MCP Servers, Claude Desktop and fun with PATHsIf you are playing with AI as a developer, you might has heard of Model Context Protocol (MCP) which is a way for an app to connect an LLM with some capabilities like an API, a file explorer, a database etc. The primary app is Claude Desktop today even though there are others like Goose. One problem you might have stumbled upon is that Claude Desktop cannot connect to your MCP server(s) because the environment to run docker or npx is different between your terminal and Claude Desktop. Maybe you use homebrew, asdf, mise or any other environment isolator. How MCP servers are run by Claude Desktop? Many are local servers running alongside Claude Desktop (on the same machine). Claude Desktop does start these processes and the command to start them is defined in Claude Desktop’s JSON config file. On macOS, it is at /Users/username/Library/Application Support/Claude/claude_desktop_config.json. Many servers are writen as nodejs apps and use npx to start. Alternatively, you can use a container image to run the server. In both case, Claude Desktop starts them when the Claude app starts. Environment problem This means that you need docker (or podman in my case) and npx to be available by Claude Desktop. As a developer I try to isolate environments, so my nodejs version is installed per directory thanks to asdf. And unfortunately, this environment is not inherited by Claude Desktop, and it failed to start my MCP servers. Setting environment to the rescue Luckily you can set the proper environment variables in the Claude Desktop configuration file. Look out for PATH and ASDF_* in the following example { "globalShortcut": "Ctrl+Space", "mcpServers": { "brave-search": { "command": "npx", "args": [ "-y", "@modelcontextprotocol/server-brave-search" ], "env": { "BRAVE_API_KEY": "<your API key>", "PATH": "/Users/username/.asdf/shims:/usr/bin:/bin", "ASDF_DIR": "/opt/homebrew/opt/asdf/libexec", "ASDF_DATA_DIR": "/Users/username/.asdf", "ASDF_NODEJS_VERSION": "22.10.0" } }, "github": { "command": "podman", "args": [ "run", "-i", "--rm", "-e", "GITHUB_PERSONAL_ACCESS_TOKEN", "mcp/github" ], "env": { "GITHUB_PERSONAL_ACCESS_TOKEN": "<Your access token>", "PATH": "/opt/homebrew/bin:/usr/bin:/bin" } } } } And voilà! Claude Desktop now starts and connects to the MCP servers. A few things to notice: I use homebrew so I set it up in the main PATH use podman and not docker as the alias is not set up in the Claude Desktop environment it is unfortunate but you need to set the nodejs version which makes the setup a bit brittle I imagine the configuration ergonomics of so called MCP hosts will get better over time.