0

I'm currently working with llama index trying to parse a column of my pandas dataframe as a Document object with llama index with the final goal of fitting my data into an LLM (I'm using gpt-4-32k). Does anyone know how to do this without explicitly converting to an unstructured datasource (ie. a doc) which seems counterintuitive?

    #First I save my data into an array (of strings)
text_list = concatenated_text_array = uniqueness_data['concatenated_text'].to_numpy().flatten()
#Then I try to cast each element to the Document object
documents = [Document(t) for t in text_list]


#and receive this error:
    documents = [Document(t) for t in text_list]
                 ^^^^^^^^^^^
  File "pydantic/main.py", line 332, in pydantic.main.BaseModel.__init__
TypeError: __init__() takes exactly 1 positional argument (2 given)

0 Answers0