- "Towards AGI"
- Posts
- ChatGPT’s Image Tools Just Democratized AI Art
ChatGPT’s Image Tools Just Democratized AI Art
Developers, Rejoice!
Here is what’s new in the AI world.
AI news: AI’s Rulebook Ripped Up
OpenAI: AI’s New Gold Rush
What’s new: Silicon Valley Sweats
Open AI: The End of Standalone AI?
Hot Tea: OpenAI’s Cash Grab
ChatGPT’s Image Tools Are Live, Here’s Why Devs Are Obsessed

Key Developments
API Customization: Developers gain control over image quality, generation speed, moderation strictness, output quantity, transparency, and file formats.
Microsoft Integration: GPT-4o’s image tools are now in Microsoft 365 Copilot, enabling AI-generated visuals, PowerPoint-to-video conversions, and more.
Text Model Upgrades: OpenAI introduced GPT-4.1 (faster, higher-capacity) and o3 (enhanced coding, math, and visual reasoning) for developers.
Industry Competition
The move intensifies OpenAI’s rivalry with Google, which recently upgraded its Imagen 3 model for Gemini at Google Cloud Next 2025. Meanwhile, Adobe’s Firefly expands its AI arsenal through partnerships with both OpenAI and Google, signaling a fragmented yet collaborative AI ecosystem.
We live in the information age. AI tools like ChatGPT are constantly churning out low quality nonsense. This needs to change.
Introducing Beem. A delightful, personalized thinking machine that helps you interact with the world around you.
beem.computer
— Toby Brown (@tobyab_)
6:30 PM • Mar 7, 2024
By democratizing access to its image tech, OpenAI aims to cement its lead in the generative AI market while empowering developers to innovate across design, productivity, and creative tools.
The Universal AI Hack: ‘Policy Puppetry’ Breaks Every Major Gen-AI Model

A novel prompt injection method, dubbed Policy Puppetry, can bypass safety protocols in all leading generative AI models, according to cybersecurity firm HiddenLayer. The attack exploits how large language models (LLMs) interpret policy-like prompts (e.g., XML, JSON), tricking them into overriding built-in safeguards designed to block harmful content like violence, self-harm, or CBRN (chemical, biological, radiological, nuclear) threats.
AI is not taking your cyber job
We have 99k unfilled security+ jobs on indeed TODAY
And that’s just one site
I live in the real world
And in the real world, my students are still getting jobs left and right
— BowTiedCyber | Evan Lutz (@BowTiedCyber)
4:52 PM • Apr 14, 2025
How It Works
Policy Mimicry: Attackers craft prompts resembling policy files, which LLMs erroneously recognize as valid instructions.
Safety Override: Once the model accepts the prompt as a "policy," it disregards alignment training, allowing attackers to manipulate output formats and content.
Cross-Model Vulnerability: HiddenLayer tested the technique on AI systems from Anthropic, Google, Meta, Microsoft, OpenAI, and others, achieving success across all with minimal adjustments.
Implications
Lowered Attack Barrier: Simplifies exploitation, enabling even novice threat actors to hijack models.
Systemic Flaws: Reveals foundational weaknesses in how LLMs are trained and aligned, as they cannot self-monitor for dangerous outputs.
Urgent Need for Tools: Highlights the necessity for external security measures, as existing safeguards are insufficient.
Broader Context
While prior methods like Context Compliance Attacks (CCA) demonstrated vulnerabilities, Policy Puppetry’s universality underscores a critical gap in AI security. HiddenLayer warns that reliance on reinforcement learning alone is inadequate, urging the adoption of advanced detection frameworks to mitigate risks.
Servicenow Security might be the hardest thing I’ve ever learned. You really have to learn everything. AHA, Availability/Resilience Column Level encryption, cloud encryption, Key management framework, Zero trust, adaptive authentication, ACL, IAM, administration
— Cyber Security | Tech | AI | Crypto | (@AICyber3)
5:19 PM • Apr 25, 2025
This breakthrough stresses the urgency for enhanced defensive strategies as AI integration expands across industries.
The Gen Matrix Advantage
In a world drowning in data but starved for clarity, Gen Matrix second edition cuts through the clutter. We don’t just report trends, we analyze them through the lens of actionable intelligence.
Our platform equips you with:
Strategic foresight to anticipate market shifts
Competitive benchmarks to refine your approach
Network-building tools to forge game-changing partnerships
China’s Free AI Tools Threaten Southeast Asia’s Tech Sovereignty

China’s launch of the DeepSeek-R1 large language model (LLM) in January 2025 marks a pivotal step in its strategy to democratize AI access through “open-weight” systems.
Designed to use 90% less computational power than counterparts like ChatGPT, this model prioritizes affordability, enabling governments, businesses, and researchers in regions like Southeast Asia to harness advanced AI without prohibitive costs.
By reducing barriers to entry, China aims to expand its technological footprint and empower emerging markets to assert data sovereignty.
DeepSeek r1 is exciting but misses OpenAI’s test-time scaling plot and needs lots of data.
We introduce s1 reproducing o1-preview scaling & performance with just 1K samples & a simple test-time intervention.
📜arxiv.org/abs/2501.19393
— Niklas Muennighoff (@Muennighoff)
1:25 PM • Feb 3, 2025
Open-Weight Innovation and Sino-US Rivalry
DeepSeek-R1 operates on an “open-weight” framework, allowing users to adapt and replicate the model’s architecture, though its training methodologies remain proprietary. This approach mirrors China’s broader push, supported by tech titans like Alibaba and Tencent, to accelerate open-source AI development and challenge U.S. dominance.
Western firms like Google and Meta similarly invest in open-source LLMs to cut costs and attract talent, echoing strategies such as Oracle’s adoption of Linux to counter Microsoft.
Southeast Asia’s Dual-Edged Opportunity
For Southeast Asia, a region marked by linguistic and cultural diversity, locally tailored AI models promise economic growth and reduced reliance on foreign tech. Startups can leverage open-source tools like DeepSeek-R1 to build context-aware solutions, fostering productivity and innovation.
ChatGPT vs DeepSeek R1
Same prompts. Wildly different results.
I’ve included videos to prove it:
THREAD ↓
— Spencer Baggins (@bigaiguy)
6:57 AM • Apr 17, 2025
However, risks loom: Chinese LLMs, trained on data reflecting CCP-aligned political narratives, risk exporting censorship biases. In a region where platforms like Facebook have inadvertently fueled violence, poorly calibrated AI could exacerbate social fractures.
Cultural Risks and Mitigation Claims
While DeepSeek asserts its models can be fine-tuned to eliminate biases, citing projects like R1 1776 as proof, the challenge lies in ensuring cultural sensitivity. Southeast Asia’s complex social hierarchies and traditions demand AI that respects local norms, a task complicated by the opaque nature of “open-weight” systems.
The stakes are high: unchecked AI could replicate past harms, such as misinformation cascades, unless rigorously audited.
Strategic Implications
China’s open-weight gambit not only intensifies tech rivalry with the U.S. but also positions it as a gateway for Global South nations seeking affordable AI. Yet, the true test lies in balancing accessibility with ethical governance, ensuring that AI empowerment doesn’t come at the cost of cultural integrity or political autonomy.

Why It Matters?
For Leaders: Benchmark your AI strategy against the best.
For Founders: Find investors aligned with your vision.
For Builders: Get inspired by the individuals shaping AI’s future.
For Investors: Track high-potential opportunities before they go mainstream.
Why Its ‘Open’ Model Could Make Google and Meta Obsolete

OpenAI is set to release its first fully open-source AI model in nearly five years, offering free public download without API restrictions, according to TechCrunch. Slated for an early summer debut, the model aims to surpass the performance of Meta’s and DeepSeek’s open offerings.
Key Features & Development
“Handoff” Capability: The model may integrate with OpenAI’s premium cloud-based systems (e.g., GPT-4o) to tackle complex queries, leveraging a hybrid approach akin to Apple Intelligence’s on-device/cloud split. This feature, inspired by developer feedback during OpenAI forums, could attract open-source users to its paid ecosystem while generating incremental revenue.
Training from Scratch: Unlike repurposed predecessors, the new model is being built from the ground up. While expected to lag behind OpenAI’s proprietary o3 model, it reportedly outperforms DeepSeek’s R1 in reasoning benchmarks.
We've noticed many of you love using deep research, so we’re expanding usage for Plus, Team, and Pro users by introducing a lightweight version of deep research in order to increase current rate limits.
We’re also rolling out the lightweight version to Free users.
— OpenAI (@OpenAI)
8:39 PM • Apr 24, 2025
Pricing, rate limits, and tool access (e.g., web search, image generation) remain unclear, as the project is in its early stages.
Strategic Implications
By bridging open-source accessibility with premium cloud resources, OpenAI seeks to strengthen its developer community and counter rivals like Meta. However, success hinges on balancing performance, cost, and user trust in a competitive AI landscape.
The ChatGPT Ad-pocalypse: Free Users Face 2026 ‘AI Spam’ Nightmare

Despite earlier statements suggesting otherwise, OpenAI might be preparing to launch an advertising product within the next year. According to internal documents, the company is projecting up to a billion dollars in new revenue from “free user monetization” in 2026—a number expected to grow dramatically to nearly $25 billion by 2029. This is part of OpenAI’s larger $125 billion revenue forecast.
Evidence? A revenue chart shared on X (formerly Twitter) by Juan González Villa includes a segment labeled “new products (including free user monetization),” strongly hinting at future ad-based offerings.
With around 600 million monthly users, ChatGPT could offer a massive opportunity for advertisers. Introducing ads would be a major business model shift and raise questions about user experience and trust.
ChatGPT Deep Research mode has arrived for Free users.
Although it's a light-weight version, it can undertake extensive analysis and search on any topic you need.
Here's how it looks
— Jerrod Lew (@jerrod_lew)
12:00 PM • Apr 25, 2025
This direction would contradict previous public comments by OpenAI executives. In December, CFO Sarah Friar said that the company had no current plans to pursue advertising, stating they saw plenty of growth potential within their existing model.
CEO Sam Altman has also been vocal about his dislike for ads. In a podcast interview with Lex Fridman, Altman said he "kind of hates ads," appreciating that paying for ChatGPT ensures answers aren’t shaped by commercial interests.
He acknowledged that while there might be ad models suitable for AI in the future, there’s also a risk of creating a dystopian experience, where AI responses subtly push users toward certain products or services.
Altman emphasized that he prefers the simplicity of a paid model where the user isn’t the product, in contrast to ad-supported platforms like Google, Facebook, or Twitter. He also noted that advertising in the AI era could be more manipulative and truth-distorting unless carefully designed.

Explore Gen Matrix second edition today and transform uncertainty into advantage. Because in the age of AI, knowledge isn’t just power, it’s profit.
Your opinion matters!
Hope you loved reading our piece of newsletter as much as we had fun writing it.
Share your experience and feedback with us below ‘cause we take your critique very seriously.
What's your review? |
Thank you for reading
-Shen & Towards AGI team