I last wrote on generative AI and its use a hair under a year ago, decrying its sloppish output, its climate and water impact, and its foundational thefts of intellectual property and creative labor.
Much has changed since. A year ago Zello had a scattering of internally influential early adopters. This worried me, but I wasn’t sure where they’d take us. In my limited and pedestrian product thinking, I couldn’t imagine what use walkie-talkie users would have for AI-labelled, headlining, highly-marketed features.
In January the company made major strategic bets on AI as a productivity multiplier for frontline workers, at the same time using the increasingly popular rhetoric: “this is the future” and its implication, “don’t be left behind.” Like many others,1 (#fn1) we also had a push from leadership to increase internal AI use within the company, from development to Sales, Engineering, Design, and Product. As (then) Director of Design, I was directed both to find ways to use generative AI in my own work and my team’s.
On the receiving end of coercive, fear-mongering language like the above, I dig my heels in. Then on a push to eek further productivity from a hard-working team,2 (#fn2) I dig my heels in further. Finally, following an employee engagement survey where I certainly wrote more than I should’ve, I found myself treading water East of Madagascar (https://www.geodatos.net/en/antipodes/united-states/austin#google_vignette) after plummeting through the Earth’s core and out the other side.
Later that month I went out for drinks with our COO, who shared necessary market perspective, and allowed me some insight on our team’s planning talks. This helped me to see the value they envisioned, and while it matters less to the frontline workers I’ve invested in, I see the business opportunity.
Over the following months, generative AI use spread further through the company, its tendrils snaking in both where prominent and remote. Walking from my desk on one end of the office to the kitchen on the other, it’s common to see ChatGPT used for code references, content summarization, and copywriting. Some use it for research and brainstorming, and many of the early adopters use it in lieu of Google’s search engine.
My early objections aside,3 (#fn3) I worry about ChatGPT use for the basic tasks expected of our professions. A seasoned engineer can use Cursor to scaffold and even write extensive code — they know how, so they can proof the work — but when a junior follows that example, I worry they’ve lost a learning opportunity.4 (#fn4) I then worry about the seasoned engineer’s skills atrophying. These examples extend beyond the quantifiable in code output — I hope in a few years, folks will still write their own emails.5 (#fn5)
In my own use, I’ve had to weigh principle against my security and my team. Insisting on Luddism paints a target on my back, and may on my team’s as well. A principled refusal — even one built on both a concrete fear of climate impact and a respect for the tools’ stolen labor — would’ve not only hurt my options, but hurt my team by association, and risked unchecked and unguided use of generative AI tools overlapping with our work. Early in the year we clamped down on the use of image generators for unauthorized brand assets and illustrations, and so revealed the need to preempt further adoption.
Encouraged to find uses for myself and my team, I found tools like Kami (https://kami.alexwidua.com) to expedite our prototyping, and team members started using Cursor for work on our marketing website. I’ve used it to scaffold and even draft javascript that I didn’t want to take the time to write, and to refactor my own rough code. I built out Origami systems for use with Whisper and ChatGPT, allowing us to prototype with text to speech, translations, and a conversational partner. Working with a live voice communications tool, it’s just cool to prototype a conversation, as opposed to using prerecorded audio responses. When GPT-4o‘s image generation was released, I created a GPT to build off our illustration style to see if we could use it to further our visual design output.6 (#fn6) I’ve since then used the agent to generate large illustrated slides when pressed for time on our monthly product updates, and it does so impressively. I later started using it in my UX and documentation work. Were I to have five obvious solutions to a problem, I had tried using ChatGPT as a discussion partner to see if there were a sixth. This rarely yielded meaningful output. Lastly, I’ve started using ChatGPT to format longer bodies of content in ways my colleagues find palatable. Where I previously would have sent a Wikipedia article, or pull quotes from a research paper’s abstract, I’ve used ChatGPT to format content so that it’s more quickly consumed by colleagues. As folks grow accustomed to ChatGPT’s predictably str https://wilnichols.com/ai-use-a-year-later/