Shopify/liquid: Performance: 53% faster parse+render, 61% fewer allocations
quality 2/10 · low quality
0 net
AI Summary
Shopify CEO Tobias Lütke used an AI-assisted autoresearch pattern with a coding agent to optimize the Liquid template engine, achieving 53% faster parse+render performance and 61% fewer allocations through 120 automated experiments across 93 commits. The effort demonstrates how robust test suites and AI agents enable effective performance optimization and enable high-level engineers to contribute meaningfully to code.
Tags
Entities
Shopify/liquid
Tobias Lütke
Andrej Karpathy
Simon Willison
Django
Pi
David Cortés
Shopify/liquid: Performance: 53% faster parse+render, 61% fewer allocations Simon Willison’s Weblog Subscribe 13th March 2026 - Link Blog Shopify/liquid: Performance: 53% faster parse+render, 61% fewer allocations ( via ) PR from Shopify CEO Tobias Lütke against Liquid, Shopify's open source Ruby template engine that was somewhat inspired by Django when Tobi first created it back in 2005 . Tobi found dozens of new performance micro-optimizations using a variant of autoresearch , Andrej Karpathy's new system for having a coding agent run hundreds of semi-autonomous experiments to find new effective techniques for training nanochat . Tobi's implementation started two days ago with this autoresearch.md prompt file and an autoresearch.sh script for the agent to run to execute the test suite and report on benchmark scores. The PR now lists 93 commits from around 120 automated experiments. The PR description lists what worked in detail - some examples: Replaced StringScanner tokenizer with String#byteindex . Single-byte byteindex searching is ~40% faster than regex-based skip_until . This alone reduced parse time by ~12%. Pure-byte parse_tag_token . Eliminated the costly StringScanner#string= reset that was called for every {% %} token (878 times). Manual byte scanning for tag name + markup extraction is faster than resetting and re-scanning via StringScanner. [...] Cached small integer to_s . Pre-computed frozen strings for 0-999 avoid 267 Integer#to_s allocations per render. This all added up to a 53% improvement on benchmarks - truly impressive for a codebase that's been tweaked by hundreds of contributors over 20 years. I think this illustrates a number of interesting ideas: Having a robust test suite - in this case 974 unit tests - is a massive unlock for working with coding agents. This kind of research effort would not be possible without first having a tried and tested suite of tests. The autoresearch pattern - where an agent brainstorms a multitude of potential improvements and then experiments with them one at a time - is really effective. If you provide an agent with a benchmarking script "make it faster" becomes an actionable goal. CEOs can code again! Tobi has always been more hands-on than most, but this is a much more significant contribution than anyone would expect from the leader of a company with 7,500+ employees. I've seen this pattern play out a lot over the past few months: coding agents make it feasible for people in high-interruption roles to productively work with code again. Here's Tobi's GitHub contribution graph for the past year, showing a significant uptick following that November 2025 inflection point when coding agents got really good. He used Pi as the coding agent and released a new pi-autoresearch plugin in collaboration with David Cortés, which maintains state in an autoresearch.jsonl file like this one . Posted 13th March 2026 at 3:44 am Recent articles Perhaps not Boring Technology after all - 9th March 2026 Can coding agents relicense open source through a “clean room” implementation of code? - 5th March 2026 Something is afoot in the land of Qwen - 4th March 2026 This is a link post by Simon Willison, posted on 13th March 2026 . django 587 performance 95 rails 110 ruby 74 ai 1903 andrej-karpathy 40 generative-ai 1687 llms 1653 ai-assisted-programming 362 coding-agents 173 agentic-engineering 25 november-2025-inflection 13 tobias-lutke 5 Monthly briefing Sponsor me for $10/month and get a curated email digest of the month's most important LLM developments. Pay me to send you less! Sponsor & subscribe Disclosures Colophon © 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026