import os
import openai
from concurrent.futures import ThreadPoolExecutor, as_completed
import tiktoken
import logging
from tqdm import tqdm
import time
import random
# Setup the logging system
logging.basicConfig(level=logging.INFO)
# Initialize OpenAI client with the API key
api_key = os.getenv('OPENAI_KEY')
if not api_key:
raise ValueError("API key not found. Please set the OPENAI_KEY environment variable.")
client = openai.OpenAI(api_key=api_key)
def load_text(file_path):
"""Load text from a specified file."""
try:
with open(file_path, 'r') as file:
return file.read()
except Exception as e:
logging.error(f'Failed to load file {file_path}: {str(e)}')
raise
def initialize_files(output_file, log_file):
"""Initialize the output and log files by creating empty files."""
try:
open(output_file, 'w').close()
open(log_file, 'w').close()
except Exception as e:
logging.error(f'Failed to initialize files {output_file}, {log_file}: {str(e)}')
raise
def save_to_file(responses, output_file):
"""Save API responses to an output file."""
try:
with open(output_file, 'w') as file:
for response in responses:
file.write(response + '\n')
except Exception as e:
logging.error(f'Failed to save to file {output_file}: {str(e)}')
raise
def log_to_file(log_file, message):
"""Log messages to a log file."""
try:
with open(log_file, 'a') as file:
file.write(message + '\n')
except Exception as e:
logging.error(f'Failed to log to file {log_file}: {str(e)}')
raise
def call_openai_api(chunk, model, max_tokens, temperature, prompt):
"""Call the OpenAI API with retries on rate limits."""
for i in range(3):
try:
response = client.chat.completions.create(
model=model,
messages=[
{"role": "system", "content": prompt},
{"role": "user", "content": chunk},
],
max_tokens=max_tokens,
n=1,
temperature=temperature,
)
return response.choices[0].message.content.strip()
except openai.OpenAIError as e:
if 'Rate limit' in str(e):
wait_time = (2 ** i) + random.random() # Exponential backoff with jitter
logging.warning(f'Rate limit exceeded. Retrying after {wait_time} seconds.')
time.sleep(wait_time)
else:
logging.error(f'API call failed: {str(e)}')
return None
logging.error('Failed to call OpenAI API after multiple retries due to rate limiting.')
return None
def split_into_chunks(text, model, tokens=3500):
"""Split the text into smaller chunks based on token limits."""
encoding = tiktoken.encoding_for_model(model)
words = encoding.encode(text)
chunks = []
for i in range(0, len(words), tokens):
chunks.append(''.join(encoding.decode(words[i:i + tokens])))
return chunks
def process_chunks(input_file, output_file, log_file, model, chunksize, max_tokens, temperature, prompt):
"""Process text chunks and call OpenAI API for each chunk."""
initialize_files(output_file, log_file)
text = load_text(input_file)
chunks = split_into_chunks(text, model, tokens=chunksize)
nCh = len(chunks)
print(f'{nCh} chunks.')
log_to_file(log_file, f'Number of chunks: {nCh}')
with ThreadPoolExecutor() as executor:
futures = {executor.submit(call_openai_api, chunk, model, max_tokens, temperature, prompt): chunk for chunk in chunks}
responses = []
for future in tqdm(as_completed(futures), total=len(futures), desc='Processing chunks'):
response = future.result()
if response is None:
log_to_file(log_file, f'Failed to process chunk {futures[future]}')
else:
responses.append(response)
log_to_file(log_file, 'Successfully processed chunk!')
save_to_file(responses, output_file)
if __name__ == "__main__":
input_file = 'input.txt'
output_file = 'output.txt'
log_file = 'log.txt'
model = 'gpt-3.5-turbo'
chunksize = 3500
max_tokens = 4000
temperature = 0.01
prompt = '''You will be presented with a scrambled, poorly formatted BibTeX entry.
Your task is to refactor the entry to fix all syntax problems and ensure it adheres to the standard BibTeX format.
Here is an example of a properly formatted BibTeX entry:
@article{Smith2023, author = {Smith, John and Doe, Jane}, title = {The Impact of Artificial Intelligence on Society}, journal = {Journal of Artificial Intelligence Research}, year = {2023}, volume = {10}, number = {2}, pages = {123--145}, doi = {10.1234/jair.2023.10.2.123}, abstract = {This paper explores the profound impact of artificial intelligence (AI) on various aspects of society. We discuss the ethical implications, economic consequences, and potential societal benefits of AI. Our analysis highlights the need for responsible AI development and deployment to mitigate risks and maximize its positive impact.}, keywords = {Artificial Intelligence, Society, Ethics, Economics, Impact} }
Here is the scrambled BibTeX data you need to refactor: <scrambled_bibtex> </scrambled_bibtex>
To refactor the scrambled BibTeX data, follow these steps:
1. Check for missing or incorrect delimiters, such as curly braces, commas, and equals signs. Ensure that each field is properly enclosed in curly braces and that key-value pairs are separated by commas.
2. Ensure proper indentation and line breaks. Each field should start on a new line and be indented with two spaces.
3. Verify the presence and correct order of required fields, such as author, title, and year. Optional fields can be included as needed.
4. Check for proper capitalization and punctuation within fields. Titles should be capitalized appropriately, and punctuation should be consistent.
5. If any information is missing or unclear, note this in your response.
Please provide the refactored BibTeX entry inside <refactored_bibtex> tags.'''
process_chunks(input_file, output_file, log_file, model, chunksize, max_tokens, temperature, prompt)
Reference:
github.com/andrewgcodes/lightspeedGPT