Quantcast
Channel: Artificial Intelligence News, Analysis and Resources - The New Stack
Viewing all articles
Browse latest Browse all 782

AI Agents Must Learn From ChatGPT’s Data Wrongs

$
0
0

Large language models (LLMs) set a dangerous data precedent, entering the age of artificial intelligence (AI). ChatGPT and other generative platforms train on information without user consent or compensation, creating significant copyright and ownership issues. The output is “new,” but the input has been copied and pasted from undisclosed sources.

This is a data rights issue that we must address on the eve of AI agents. Despite promising to become superhuman helpers in personal and professional tasks, agents won’t be worthy of trust if we continue to create black boxes with little regard for intellectual property.

Instead, especially in these early days, we must prefer infrastructure that tracks information, recognizes input, and rewards contributors. This is how we learn from ChatGPT’s data wrongs and enable the new wave of agents to operate with verification, permission, and privacy at their core.

LLMs Set a Dangerous Data Precedent

Even when the content generated by Claude, ChatGPT, or Gemini feels original, it’s actually scraped from billions of data points without explicit permission or subsequent compensation to the owners. These platforms essentially take copyrighted materials, move ahead without consent, and fail to attribute sourcing.

To make matters worse, we usually don’t know how these models make decisions. They’re closed source with data going in, commands coming out, and zero transparency about what happens in between. This black box approach creates both ethical and practical problems.

I’ve previously compared models to humans in that they are what they eat. If we only eat junk food, we’re slow and sluggish. If agents only consume copyright and second-hand material, they’re inaccurate, unreliable, and general rather than specific. Their data “diet” determines performance and we can’t expect quality outputs from systems built on problematic inputs.

A new era needs a new approach. AI agents have a chance to bake in data rights from day one by leveraging blockchain to track information and strong data infrastructure to dictate its use. By building data provenance and respect into the foundations, we can arm agents with consented information and bring users in on the value it generates.

Building Data Guardrails With Infrastructure

The good news is that it’s not too late to change the data status quo. Three technical guardrails are emerging to ensure agent behavior moves toward transparent infrastructure and improved data rights.

First, we need clear pipelines that track attribution. Another much-hyped and much-misunderstood technology, blockchain, is helping with this. Blockchain-based data frameworks create immutable records of what information agents access. Unlike today’s opaque sourcing, we can build accountability into the infrastructure with verifiable credentials systems and decentralized identifiers. For example, Kite AI is building a modular layer-1 blockchain that tracks proof of attributed intelligence. This way, developers can configure incentives, coordinate collaboration across subnets, and build ideal AI tech stacks across customized data, models, and agents.

Second, privacy-preserving computation technologies allow data processing without exposure. Zero-knowledge proofs, homomorphic encryption, and secure multiparty computation create foundations of consent and keep data safe. These technologies, implemented in various blockchain systems and Trusted Execution Environment (TEE) computing platforms, enable computation over sensitive data without revealing it.

Third, we need to return confidence to these systems with proper credit. If user information or copyrighted material is used, agents and their respective models must reward rather than ignore attribution. Story Protocol is another web3 project furthering this concept — using blockchain to allow creators to establish ownership of their work, setting rules for how it can be used, and ensuring they get paid when their content is utilized. CARV ID achieves something similar — tying online identities to a single source of truth where users decide if their information is available and payable for model training. Not only does this bring users in on the AI revolution, but it goes some way to restoring trust in the overall system.

Trust Makes or Breaks Agent Adoption

Agents will stumble at the first hurdle if they don’t engender trust. Remember, these are platforms that promise autonomous, intelligent assistance across contexts. Mainstream acceptance is unlikely at home or work if our industry continues to rip off data.

We have already seen the dangers of this with Samsung meeting notes and source code leaking after use in ChatGPT. This kind of data handling isn’t good enough for corporate proprietary data and organizations bound by GDPR or HIPAA. Users and enterprises alike need to know their information isn’t only safe with these models but that it’s using top-of-the-line privacy standards and backend guardrails.

There are also operational benefits to proper data use. AI agents built on data sovereignty infrastructure produce more reliable results by accessing higher-quality, properly attributed information. They can also be audited for biases and inaccuracies. And perhaps most importantly, they earn trust by operating transparently rather than as inscrutable black boxes.

Agent success depends on creating a system where data isn’t just abundant and accredited but enriched with actionable metrics and validated through trustless consensus. When agents access high-quality, verified information — and only with explicit user consent and attribution mechanisms — they can deliver on their lofty goals. Now’s the time to learn from ChatGPT’s data wrongs and chart a new path toward safe and validated agentic infrastructure.

The post AI Agents Must Learn From ChatGPT’s Data Wrongs appeared first on The New Stack.

AI agents must move away from opaque models and embrace transparent, permission-based systems.

Viewing all articles
Browse latest Browse all 782

Latest Images

Trending Articles



Latest Images