Using OpenAI for batch inference
You can use OpenAI for batch inference in Curator to generate synthetic data. In this example, we will generate reannotation of wildchat dataset, but the approach can be adapted for any data generation task.
Prerequisites
Python 3.10+
Curator: Install via
pip install bespokelabs-curator
OpenAI: OpenAI API key
Steps
1. Setup environment vars
2. Create a curator.LLM subclass
Create a class that inherits from curator.LLM
. Implement two key methods:
prompt()
: Generates the prompt for the LLM.parse()
: Processes the LLM's response into your desired format.
Here’s the implementation:
3. Configure the OpenAI model
4. Generate Data
Generate the structured data and output the results as a pandas DataFrame:
Example Output
Using the above example, the output might look like this:
Write a very long, elaborate, descriptive and ...
Scene: Omelette Apocalypse\n\n**INT. DINER...
what are you?
I am a large language model, trained by OpenAI
Batch Configuration
Check out complete batch configuration
Last updated