Thoughtworks recently launched a new agentic development platform called AI/works™. The marketing claims are quite bold, specifically targeting the "legacy modernization" problem space (which we all know is usually a nightmare).
According to the announcement and their technical guide, the workflow is:
Ingestion: "Blackbox" reverse-engineering of legacy binaries/code (even without full source access in some cases).
Specification: It generates a "SuperSpec" — a machine-readable functional specification enriched with regulatory/security context.
Forward Engineering: Agents use the Spec to generate new code, tests, and pipelines.
Lifecycle: It claims to support a "3-3-3 delivery model" (Idea to Production in 90 days) and includes self-healing/regenerative capabilities for maintenance.
This sounds like the "Holy Grail" of software engineering, but I am skeptical about how well this works on actual enterprise spaghetti code versus carefully curated demos. "Reverse engineering into a perfect spec" is historically where these tools fail.
I’m looking for insights from anyone who has piloted this or works at TW:
How does the "Code-to-Spec" reverse engineering actually handle heavy technical debt or undocumented business logic?
Is the "SuperSpec" truly editable/maintainable by humans, or does it become a new black box?
How much "human in the loop" is actually required for the 3-3-3 model?
Is this built on public LLMs (Claude/GPT-4) or proprietary models trained on legacy patterns (like COBOL/Mainframe data via their Mechanical Orchard partnership)?
Any details on the reality behind the marketing would be appreciated.
I also have been cycling for 2 years now everyday. 100kms average every week. Never have felt more enthusiastic and energetic at workplace. Helps concentrate on the work, beats bad traffic and gives a very good exercise overall
According to the announcement and their technical guide, the workflow is:
Ingestion: "Blackbox" reverse-engineering of legacy binaries/code (even without full source access in some cases).
Specification: It generates a "SuperSpec" — a machine-readable functional specification enriched with regulatory/security context.
Forward Engineering: Agents use the Spec to generate new code, tests, and pipelines.
Lifecycle: It claims to support a "3-3-3 delivery model" (Idea to Production in 90 days) and includes self-healing/regenerative capabilities for maintenance.
This sounds like the "Holy Grail" of software engineering, but I am skeptical about how well this works on actual enterprise spaghetti code versus carefully curated demos. "Reverse engineering into a perfect spec" is historically where these tools fail.
I’m looking for insights from anyone who has piloted this or works at TW:
How does the "Code-to-Spec" reverse engineering actually handle heavy technical debt or undocumented business logic?
Is the "SuperSpec" truly editable/maintainable by humans, or does it become a new black box?
How much "human in the loop" is actually required for the 3-3-3 model?
Is this built on public LLMs (Claude/GPT-4) or proprietary models trained on legacy patterns (like COBOL/Mainframe data via their Mechanical Orchard partnership)?
Any details on the reality behind the marketing would be appreciated.