-
Notifications
You must be signed in to change notification settings - Fork 69
/
extracts.py
221 lines (180 loc) · 8.75 KB
/
extracts.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
# Standard library imports
import sys
import json
import os
from textwrap import dedent
import logging
from pathlib import Path
import traceback
# Third party imports
from openai import OpenAI
from dotenv import load_dotenv
# Local application imports
logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s')
load_dotenv()
api_key = os.getenv("OPENAI_API_KEY")
if not api_key:
raise ValueError("API key not found. Please set the OPENAI_API_KEY environment variable.")
client = OpenAI(api_key=api_key)
def get_whisper_output():
whisper_output_dir = Path('whisper_output')
if not whisper_output_dir.exists():
logging.error(f"Directory not found: {whisper_output_dir}")
return None, None
srt_files = list(whisper_output_dir.glob('*.srt'))
txt_files = list(whisper_output_dir.glob('*.txt'))
if not srt_files or not txt_files:
logging.warning("No .srt or .txt files found in the whisper_output directory.")
return None, None
with open(txt_files[0], 'r') as file:
transcript = file.read()
with open(srt_files[0], 'r') as file:
subtitles = file.read()
return transcript, subtitles
def call_openai_api(transcript):
logging.info("STARTING call_openai_api")
prompt = dedent(f"""
You will be given a complete transcript from a video. Your task is to identify four 1-minute long clips from this video that have the highest potential to become popular on social media.
Follow these steps to complete the task:
1. Carefully read through the entire transcript, looking for the most powerful, emotionally impactful, surprising, thought-provoking, or otherwise memorable moments. Give priority to answers and speculations rather than questions.
2. For each standout moment you identify, extract a 1-minute segment of text from the transcript, centered around that moment. Ensure each segment is approximately 1 minute long when spoken (about 125 words or 10 spoken sentences).
3. From these segments, choose the top four that you believe have the highest potential to go viral on social media.
4. Rank these four clips from most to least viral potential based on your assessment.
5. Determine the word count for each of the four selected clips.
6. Format your final output as a JSON object containing an ordered list of the selected clips, each with its extracted text. The JSON object should look like this:
{{
"clips": [
{{
"rank": 1,
"text": "<extracted text for clip 1>",
"wordcount": <length of the extracted text in words>
}},
{{
"rank": 2,
"text": "<extracted text for clip 2>",
"wordcount": <length of the extracted text in words>
}},
{{
"rank": 3,
"text": "<extracted text for clip 3>",
"wordcount": <length of the extracted text in words>
}},
{{
"rank": 4,
"text": "<extracted text for clip 4>",
"wordcount": <length of the extracted text in words>
}}
]
}}
Here is the transcript:
<transcript>
{transcript}
</transcript>
Important reminders:
- Always return EXACTLY THREE clips.
- DO NOT OMIT any text
- Return nothing else but the raw content of the JSON object itself - no comments, no extra text. Just the JSON.
- Ensure that each clip is approximately 1 minute long when spoken (about 125 words or 10 spoken sentences).
- Focus on selecting clips that are powerful, emotionally impactful, surprising, thought-provoking, or otherwise memorable.
- Prioritize answers and speculations over questions when selecting clips.
""")
try:
response = client.chat.completions.create(
model="gpt-4o-2024-08-06",
messages=[
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": prompt}
],
temperature=0.8,
max_tokens=4095,
top_p=1,
frequency_penalty=0,
presence_penalty=0,
response_format={
"type": "json_schema", # Ensure this type is specified correctly
"json_schema": {
"name": "clips_response",
"strict": True,
"schema": {
"type": "object",
"properties": {
"clips": {
"type": "array",
"items": {
"type": "object",
"properties": {
"rank": {"type": "integer"},
"text": {"type": "string"},
"wordcount": {"type": "integer"}
},
"required": ["rank", "text", "wordcount"], # Include 'wordcount' here
"additionalProperties": False
}
}
},
"required": ["clips"],
"additionalProperties": False
}
}
}
)
if not response or not response.choices:
logging.error("No response or choices from OpenAI API")
return None
response_text = response.choices[0].message.content
logging.info(f"Raw API response: {response_text}")
try:
response_data = json.loads(response_text)
# Ensure there are exactly four clips
if len(response_data.get('clips', [])) != 3:
logging.error("The response does not contain exactly four clips. Generating filler content.")
while len(response_data['clips']) < 3:
response_data['clips'].append({
"rank": len(response_data['clips']) + 1,
"text": "This is filler content to ensure there are exactly four clips."
})
return response_data
except json.JSONDecodeError as e:
logging.error(f"JSON Decode Error: {str(e)}")
logging.error(f"Problematic JSON string: {response_text}")
# Log the first 500 characters of the response for debugging
logging.error(f"First 500 characters of response: {response_text[:500]}")
return None
except Exception as e:
logging.error(f"Error calling OpenAI API: {str(e)}")
logging.error(traceback.format_exc())
return None
def save_response_to_file(response, output_path):
try:
with open(output_path, 'w') as f:
json.dump(response, f, indent=4)
logging.info(f"Response saved to {output_path}")
except Exception as e:
logging.error(f"Error saving response to file: {e}")
def main():
logging.info('STARTING extracts.py')
transcript, subtitles = get_whisper_output()
if transcript is None or subtitles is None:
logging.error("Failed to get whisper output")
return None
response = call_openai_api(transcript)
if response and 'clips' in response:
output_dir = Path('crew_output')
output_dir.mkdir(exist_ok=True)
output_path = output_dir / 'api_response.json'
save_response_to_file(response, output_path)
# Extract the text from each clip to match crew.py's expectations
extracts = [clip['text'] for clip in response['clips']]
return extracts
else:
logging.error("Failed to get a valid response from OpenAI API")
if response:
logging.error(f"Unexpected response structure: {response}")
return None
if __name__ == "__main__":
main()
# TODO: Split the below tasks into separate API queries for different large language model (LLM) calls or agents to implement a divide-and-conquer approach.
# 1. Read the entire transcript carefully, identifying key moments that stand out as particularly impactful or shareable.
# 2. For each of these moments, extract a 1-minute segment of text from the transcript, centered around that moment. Ensure each segment is approximately 1 minute long when spoken (about 8 sentences).
# 3. From these segments, choose the top four that you believe have the highest potential to go viral.
# 4. Rank these four clips from most to least viral potential based on your assessment.