May 29, 2023
ChatGPT Hobbyists
ChatGPT has understandably garnered a huge amount of attention from all corners of academia, from philosophy to economics. One of the more quixotic examples I’ve encountered recently is Robert W. McGee, and his many papers on this topic, such as “Is Chat Gpt Biased Against Conservatives? An Empirical Study”.
A professor of accounting at Fayetteville University, McGee’s biography reads as something like a Marvel Cinematic Universe version of a nerdy academic supervillain. In addition to having published 59 non-fiction books, McGee apparently holds 23 academic degrees, including 13 doctorates, as well as being a world champion in various martial arts, such as Taekwondo and Tai Chi.
Tags:
large-language-models
,
chatgpt
,
sociotechnical-systems
,
measurement
,
evaluation
Apr 13, 2023
Samsung's Encounter with ChatGPT
As ChatGPT continues to ricochet through the news cycle, media outlets are surely on the hunt for new angles they can present to the public in order to keep this story in motion. Among other threads, one that has gained some traction is the question of risks to privacy and security presented by these new systems.
Last week, a number of US outlets reported on data leaks at Samsung, in which three employees (in separate incidents) apparently entered confidential company information into ChatGPT. According to reports, in one case, an employee tried using ChatGPT to help debug code, another to optimize code, and a third to have it produce a summary of meeting notes.
Tags:
machine-learning
,
large-language-models
,
chatgpt
,
data
,
privacy
,
security
Feb 19, 2023
ChatGPT and Sociotechnical Instability
I’ve written about this before, but it’s worth remembering that almost nothing in sociotechnical systems is guaranteed to remain stable for very long. We’ve recently had two great examples of this, with the first being the changes to Twitter, and the second being ChatGPT (and, by extension, the new Bing).
In the first case, a platform which had long seemed relatively static, (especially compared to all the rest), rather suddenly changed hands, which led to major changes in what it delivered. On some level, many of the technical changes to the actual functionality of the site were relatively minor. Much bigger, however, was the impact of many people abandoning the platform for alternatives like Mastodon. Although it seems to me like people have gradually been filtering back to Twitter, most people seem to have the sense that the experience has changed. Obviously more dramatic infrastructural changes, like prioritizing tweets from paying users, could produce even more dramatic shifts. Regardless, it’s a good reminder that what we think of as “Twitter” is the product of a combination of people and software, either or both of which can shift dramatically in a short period of time.
Tags:
large-language-models
,
twitter
,
sociotechnical-systems
,
chatgpt
,
instability
,
archives
Jan 21, 2023
ChatGPT Dominance
I expect that almost anyone reading this will have heard of ChatGPT by now. Released about a month ago, ChatGPT is a system developed by OpenAI which provides text responses to text input. Although details are scarce, under the hood ChatGPT is basically a large language model, trained with some additional tricks (see Yoav Goldberg’s write up for a good summary). In other words, it is a model which maps from the text input (treated as a sequence of tokens), to a distribution over possible next tokens, and generates text by making repeated calls to this function, and sampling tokens from the predicted distributions.
Tags:
machine-learning
,
large-language-models
,
chatgpt
,
open-science
,
evaluation