How to input large data into python pandas using looping or parallel computing?

后端 未结 5 639
谎友^
谎友^ 2021-02-13 21:46

I have a csv file of 8gb and I am not able to run the code as it shows memory error.

file = \"./data.csv\"
df = pd.read_csv(file, sep=\"/\", header=0, dtype=str         


        
5条回答
  •  离开以前
    2021-02-13 22:25

    pandas read_csv has two argument options that you could use to do what you want to do:

    nrows : to specify the number of rows you want to read
    skiprows : to specify the first row you want to read
    

    Refer to documentation at: https://pandas.pydata.org/pandas-docs/stable/generated/pandas.read_csv.html

提交回复
热议问题