Generative AI and Creative Work

Here's your daily briefing:

  • THIS LINK WILL SELF-DESTRUCT. Depending on when you read this e-mail, the below link will or will not take you to the live feed of the 2022 Stanford HAI conference on human-in-the-loop AI:

  • Here's a megathread from Daniel Eckler on 90 advances in AI over the last 90 days:

  • Very interesting (and reassuring) paper for all the developers out there worried about being displaced by AI assistants. As we've talked about previously, before AI fully replaces most jobs it will augment them. This means that we'll likely spend several years (or decades) in a world where the most successful programmers, artists, writers, etc. won't be human or AI but humans plus AI. For now, developers with an understanding of AI assistants and their vulnerabilities are probably more employable, rather than less.

Overall, we find that participants who had access to an AI assistant based on OpenAI's codex-davinci-002 model wrote significantly less secure code than those without access. Additionally, participants with access to an AI assistant were more likely to believe they wrote secure code than those without access to the AI assistant. Furthermore, we find that participants who trusted the AI less and engaged more with the language and format of their prompts (e.g. re-phrasing, adjusting temperature) provided code with fewer security vulnerabilities.

  • Here's another entry in the discussion/debate about generative AI and "the next Google" we've touched on a few times recently:

What I am trying to say is that the way search engines work is not because it’s the best way to search, but because of the best technology available in the late 1990s and we are stuck in that paradigm. Maybe there is a better way, but change is hard.

  • Yet another crazy use-case for LLMs:

This is what we got when we searched for "lit review on psychodynamic therapy" 🤓 🤯 :

It's been a minute since we've dove into a longer read with you. Since we found this Harvard Business Review piece that is smack dab up our alley (and since we finally feel like we beat the flu and can concentrate for more than 5 minutes) we decided we'd read it and let you know our takeaways.

The article begins with with 4 opportunities created by generative AI:

1. Automated content generation: Large language and image AI models can be used to automatically generate content, such as articles, blog posts, or social media posts. This can be a valuable time-saving tool for businesses and professionals who create content on a regular basis.

2. Improved content quality: AI-generated content can be of higher quality than content created by humans, due to the fact that AI models are able to learn from a large amount of data and identify patterns that humans may not be able to see. This can result in more accurate and informative content.

3. Increased content variety: AI models can generate a variety of content types, including text, images, and video. This can help businesses and professionals to create more diverse and interesting content that appeals to a wider range of people.

4. Personalized content: AI models can generate personalized content based on the preferences of individual users. This can help businesses and professionals to create content that is more likely to be of interest to their target audience, and therefore more likely to be read or shared.

And then promptly goes on to tell us that all of the above was written by GPT-3.

All they prompted GPT-3 with to get the above was the sentence, "Large language and image AI models, sometimes called generative AI or foundation models, have created a new set of opportunities for businesses and professionals that perform content creation."

We really appreciate this way of beginning the piece because of the "show, don't tell" ethos. What better way to drive home the power of generative AI than to have someone read something they think was written by a human and then tell them it wasn't?

The piece then goes on to briefly cover the basics of generative AI:

The software uses complex machine learning models to predict the next word based on previous word sequences, or the next image based on words describing previous images.

And points out what we reiterated above (for the umpteenth time): generative AI is human creativity, augmented. The tools alone are not enough to create great works of art, it is how they are wielded (and evaluated and edited) by humans that leads to the most impressive work.

To use generative AI effectively, you still need human involvement at both the beginning and the end of the process.

To start with, a human must enter a prompt into a generative model in order to have it create content. Generally speaking, creative prompts yield creative outputs. “Prompt engineer” is likely to become an established profession, at least until the next generation of even smarter AI emerges.

Then, once a model generates content, it will need to be evaluated and edited carefully by a human.

The piece goes on to discuss several specific applications for generative AI tools, starting with:

MARKETING

These generative models are potentially valuable across a number of business functions, but marketing applications are perhaps the most common. Jasper, for example, a marketing-focused version of GPT-3, can produce blogs, social media posts, web copy, sales emails, ads, and other types of customer-facing content.

They quote Kris Ruby, owner of Ruby Media Group, who uses both image and text generation tools. She reiterates several trends we've seen and covered, such as the ways these tools will affect SEO optimization, and how they'll affect (or displace) the market for stock photos.

CODE GENERATION

They then go on to cover tools like OpenAI's Codex, which "given a description of a 'snippet' or small program function, GPT-3’s Codex program — specifically trained for code generation — can produce code in a variety of different languages."

But, just like the paper we cited in today's briefing made clear, they point out how, for the time being, these tools still rely on human involvement.

The consensus on LLM-based code generation is that it works well for such snippets, although the integration of them into a larger program and the integration of the program into a particular technical environment still require human programming capabilities.

CONVERSATION

LLMs are increasingly being used at the core of conversational AI or chatbots. They potentially offer greater levels of understanding of conversation and context awareness than current conversational technologies.

As in the film Her, these chatbots are still not quite "there":

None of these LLMs is a perfect conversationalist. They are trained on past human content and have a tendency to replicate any racist, sexist, or biased language to which they were exposed in training. Although the companies that created these systems are working on filtering out hate speech, they have not yet been fully successful.

KNOWLEDGE MANAGEMENT

They briefly discuss one of the areas we've been most interested in, AI applied to the "knowledge management" space.

Some companies are exploring the idea of LLM-based knowledge management in conjunction with the leading providers of commercial LLMs. Morgan Stanley, for example, is working with OpenAI’s GPT-3 to fine-tune training on wealth management content, so that financial advisors can both search for existing knowledge within the firm and create tailored content for clients easily. It seems likely that users of such systems will need training or assistance in creating effective prompts, and that the knowledge outputs of the LLMs might still need editing or review before being applied. Assuming that such issues are addressed, however, LLMs could rekindle the field of knowledge management and allow it to scale much more effectively.

With generative AI tools like Lex and integration into existing tools like Google Docs or Notion on the horizon, we think knowledge workers are going to have some of the largest productivity "gains" of any field touched by AI.

DEEPFAKES AND ETHICAL CONCERNS

They end with a very brief foray into the obvious ethical concerns created by these tools.

Heretofore, however, the creation of deepfakes required a considerable amount of computing skill. Now, however, almost anyone will be able to create them. OpenAI has attempted to control fake images by “watermarking” each DALL-E 2 image with a distinctive symbol. More controls are likely to be required in the future, however — particularly as generative video creation becomes mainstream.

Generative AI also raises numerous questions about what constitutes original and proprietary content. Since the created text and images are not exactly like any previous content, the providers of these systems argue that they belong to their prompt creators. But they are clearly derivative of the previous text and images used to train the models. Needless to say, these technologies will provide substantial work for intellectual property attorneys in the coming years.

This latter point is one we hadn't quite considered. We speak a lot about how the best way to stop "on top of" AI is to stay engaged and curious and to learn these tools as they come out, but we hadn't considered that one of the best ways to capitalize on the generative AI revolution might be to go into copyright law!

Overall, this is a simple but concise overview of generative AI and it's implications for creative work and a good piece to send to a friend who hasn't quite engaged with the topic yet but wants to learn more.

That's it for today.

(Special thanks to the article's authors: Nitin Mittal and Thomas Davenport.)