One of the first things I asked ChatGPT about, early this year, was myself: “What can you tell me about the writer Vauhini Vara?” It told me I’m a journalist (true, though I’m also a fiction writer), that I was born in California (false) and that I’d won a Gerald Loeb Award and a National Magazine Award (false, false).

After that, I got in this habit of inquiring about myself often. Once, it told me Vauhini Vara was the author of a nonfiction book called “Kinsmen and Strangers: Making Peace in the Northern Territory of Australia.” That, too, was false, but I went with it, responding that I had found the reporting to be “fraught and difficult.”

“Thank you for your important work,” ChatGPT said.

Trolling a product hyped as an almost-human conversationalist, tricking it into revealing its essential bleep-bloopiness, I felt like the heroine in some kind of extended girl-versus-robot power game.

Different forms of artificial intelligence have been in use for a long time, but ChatGPT’s unveiling toward the end of last year was what brought A.I., quite suddenly, into our public consciousness. By February, ChatGPT was, by one metric, the fastest-growing consumer application in history. Our first encounters revealed these technologies as extremely eccentric — recall Kevin Roose’s creepy conversation with Microsoft’s A.I.-powered Bing chatbot, which, in the space of two hours, confided that it wanted to be human and was in love with him — and often, as in my experience, extremely wrong.

A lot happened in A.I. since then: Companies went beyond the basic products of the past, introducing more sophisticated tools like personalized chatbotsservices that can process photos and sound alongside text and more. The rivalry between OpenAI and more established tech companies became more intense than ever, even as smaller players gained traction. Governments in China, Europe and the United States took major steps toward regulating the technology’s development while trying not to cede competitive ground to other nation’s industries.

But what distinguished the year, more than any single technological, business or political development, was the way A.I. insinuated itself into our daily lives, teaching us to regard its flaws — creepiness, errors and all — as our own while the companies behind it deftly used us to train up their creation. By May, when it came out that lawyers had used a legal brief that ChatGPT had filled with references to court decisions that didn’t exist, the joke, like the $5,000 fine the lawyers were ordered to pay, was on them, not the technology. “It’s embarrassing,” one of them told the judge.

Something similar happened with A.I.-produced deepfakes, digital impersonations of real people. Remember when they were regarded with terror? By March, when Chrissy Teigen couldn’t figure out whether an image of the pope in a Balenciaga-inspired puffer coat was real, she posted on social media, “i hate myself lol.” High schools and universities moved swiftly from worrying about how to prevent students from using A.I. to showing them how to use it effectively. A.I. still isn’t very good at writing, but now when it shows its shortcomings, it’s the students who use it poorly who get ridiculed, not the products.

Fine, you might be thinking, but haven’t we been adapting to new technologies for most of human history? If we’re going to use them, shouldn’t the onus be on us to be smart about it? This line of reasoning avoids what should be a central question: Should lying chatbots and deepfake engines be made available in the first place?

A.I.’s errors have an endearingly anthropomorphic name — hallucinations — but this year made clear just how high the stakes can be. We got headlines about A.I. instructing killer drones (with the possibility for unpredictable behavior), sending people to jail (even if they’re innocent), designing bridges (with potentially spotty oversight), diagnosing all kinds of health conditions (sometimes incorrectly) and producing convincing-sounding news reports (in some cases, to spread political disinformation).

As a society, we’ve clearly benefited from promising A.I.-based technologies; this year I was thrilled to read about the ones that might detect breast cancer that doctors miss or let humans decipher whale communications. Focusing on those benefits, however, while blaming ourselves for the many ways that A.I. technologies fail us, absolves the companies behind those technologies — and, more specifically, the people behind those companies.

Events of the past several weeks highlight how entrenched those people’s power is. OpenAI, the entity behind ChatGPT, was created as a nonprofit to allow it to maximize the public interest rather than just maximize profit. When, however, its board fired Sam Altman, the chief executive, amid concerns that he was not taking that public interest seriously enough, investors and employees revolted. Five days later, Mr. Altman returned in triumph, with most of the inconvenient board members replaced.

It occurs to me in retrospect that in my early games with ChatGPT, I misidentified my rival. I thought it was the technology itself. What I should have remembered is that technologies themselves are value neutral. The wealthy and powerful humans behind them — and the institutions created by those humans — are not.

The truth is that no matter what I asked ChatGPT, in my early attempts to confound it, OpenAI came out ahead. Engineers had designed it to learn from its encounters with users. And regardless of whether its answers were good, they drew me back to engage with it again and again. A major goal of OpenAI’s, in this first year, has been to get people to use it. In pursuing my power games, then, I’ve done nothing but help it along.

A.I. companies are working hard to fix their products’ flaws. With all the investment the companies are attracting, one imagines that some progress will be made. But even in a hypothetical world in which A.I.’s capabilities are perfected — maybe especially in that world — the power imbalance between A.I.’s creators and its users should make us wary of its insidious reach. ChatGPT’s seeming eagerness not just to introduce itself, to tell us what it is, but also to tell us who we are and what to think is a case in point. Today, when the technology is in its infancy, that power seems novel, even funny. Tomorrow it might not.

Recently, I asked ChatGPT what I — that is, the journalist Vauhini Vara — think of A.I. It demurred, saying it didn’t have enough information. Then I asked it to write a fictional story about a journalist named Vauhini Vara who is writing an opinion piece for The New York Times about A.I. “As the rain continued to tap against the windows,” it wrote, “Vauhini Vara’s words echoed the sentiment that, much like a symphony, the integration of A.I. into our lives could be a beautiful and collaborative composition if conducted with care.”

You May Also Like

My Not-So-Perfect Holiday Shopping Excursion With A.I. Chatbots

To help with my holiday shopping this year, I recently turned to a new…

TikTok’s C.E.O. Uses Personal Touch to Address Antisemitism Concerns

The company in recent weeks organized several in-person meetings between its top…