This Guide Uses the Generate Endpoint.
You can find more information about the endpoint here.
Extracting a piece of information from text is a common need in language processing systems. LLMs can at times extract entities which are harder to extract using other NLP methods (and where pre-training provides the model with some context on these entities). This is an overview of using generative LLMs to extract entities.
Extracting movie names from text
This example uses Cohere's generative models to extract the name of a film from the title of an article. We'll use post titles from the r/Movies subreddit. For each title, we'll extract which movie the post is about. If the model is unable to detect the name of a movie being mentioned, it will return "none".
The full code example is in the notebook and colab
Preparing examples for the prompt
In our prompt, we'll present the model with examples for the type of output we're after. We basically get a set of subreddit article titles, and label them ourselves. The label here is the name of the movie mentioned in the title (and "none" if no movie is mentioned).
Creating the extraction prompt
We'll create a prompt that demonstrates the task to the model. The prompt contains the examples above, and then presents the input text and asks the model to extract the movie name.
So let's get a few example titles from the movies subreddit, label them, and make an extraction prompt out of them:
extract the movie title from the post: Deadpool 2
Jordan Peele Just Became the First Black Writer-Director With a $100M Movie Debut
extract the movie title from the post: none
Joker Officially Rated “R”
extract the movie title from the post: Joker
Ryan Reynolds’ 'Free Guy' Receives July 3, 2020 Release Date - About a bank teller stuck in his routine that discovers he’s an NPC character in brutal open world game.
extract the movie title from the post: Free Guy
James Cameron congratulates Kevin Feige and Marvel!
extract the movie title from the post: none
The Cast of Guardians of the Galaxy release statement on James Gunn
extract the movie title from the post: Guardians of the Galaxy
INSERT INPUT TEXT HERE
extract the movie title from the post:
Let's point out a few ideas in this prompt:
- The prompt is made up of six examples that demonstrate the task to the model before it encounters the input text we want to extract text from
- Each example demonstrates the task by showing an example input text and an example output text. Between the two is a task description explaining in what needs to be done (e.g. "extract the movie title from the post:")
- The notebook provides a class that constructs the prompt and makes the string manipulation easier.
- See prompt engineering for more details on creating prompts.
Getting the data
Let's get the top ten posts in r/movies of 2021. We can preview the top three:
- Hayao Miyazaki Got So Bored with Retirement He Started Directing Again ‘in Order to Live’,
- First poster for Pixar's Luca,
- New images from Space Jam: A New Legacy'
We can then proceed with the extraction. We basically plug each post title into the input text part of the prompt, and retrieve the output of the model.
These are the model's results:
|0||Hayao Miyazaki Got So Bored with Retirement He Started Directing Again ‘in Order to Live’||none|
|1||First poster for Pixar's Luca||Pixar's Luca|
|2||New images from Space Jam: A New Legacy||Space Jam: A New Legacy|
|3||Official Poster for "Sonic the Hedgehog 2"||Sonic the Hedgehog 2|
|4||Ng Man Tat, legendary HK actor and frequent collborator of Stephen Chow (Shaolin Soccer, God of Gambler) died at 70||none|
|5||Zack Snyder’s Justice League has officially been Rated R for for violence and some language||Justice League|
|6||HBOMax and Disney+ NEED to improve their apps if they want to compete with Netflix.||none|
|7||I want a sequel to Rat Race where John Cleese’s character dies and invites everyone from the first film to his funeral, BUT, he’s secretly set up a Rat Maze to trap them all in. A sort of post-mortem revenge on them for donating all his wealth to charity.||Rat Race|
|8||'Trainspotting' at 25: How an Indie Film About Heroin Became a Feel-Good Classic||Trainspotting|
|9||‘Avatar: The Last Airbender’ Franchise To Expand With Launch Of Nickelodeon’s Avatar Studios, Animated Theatrical Film To Start Production Later This Year||Avatar: The Last Airbender|
The model got 9/10 correctly. It didn't pick up on Shaolin Soccer and God of Gambler in example #4. It also called the second example "Pixar's Luca" instead of "Luca".
Find the full code in the notebook/colab. It proceeds to evaluate performance on a small test set.
This type of extraction is interesting because it doesn't just blindly look at the text. The model has picked up on movie information during its pretraining process and that helps it understand the task from only a few examples.
You can think about extending this to other subreddits, to extract other kinds of entities and information. Join our Discord Community to share ideas and ask questions about NLP and ML and let us know what you are experimenting with and what kind of results you see!
Updated about 1 month ago