pandas学习笔记:01、数据文件的读取与写入

1、读取数据

'''
常用的读取数据函数
'''
import pandas as pd
'''
	./	代表当前目录,当前目录也可以什么都不写,直接寻找当前目录的文件
		比如:./data/ 和 data/ 都代表当前目录下的data文件夹下的文件
	../	代表上一级目录
	/	代表根目录
		Linux系统里面会用到根目录
	~	代表当前用户目录
		比如Windows用户Dongze代表的就是'C:\\Users\\Dongz'
'''
#读取CSV格式数据,返回DataFrame格式列表
data = pd.read_csv("数据目录/xxx.csv")
#还可以使用URL来读取
pd.read_csv("http://localhost/xxx.csv")
data = pd.read_excel("数据目录/xxx.xlsx")


如果数据过多,编译器会省略中间部分数据,如下图所示:
图1
我们可以设置dataframe显示中间忽略的数据


'''
    设置dataframe显示数据
'''
#显示Dateframe所有行
pd.set_option('display.max_rows',None)
#显示Dateframe所有列(参数设置为None代表显示所有行,也可以自行设置数字)
pd.set_option('display.max_columns',None)
#设置Dataframe数据的显示长度,默认为50
pd.set_option('max_colwidth',200)
#禁止Dateframe自动换行(设置为Flase不自动换行,True反之)
pd.set_option('expand_frame_repr', False)

这样就会显示出所有数据
图2
**

2、官网提供的读取文件和写入文件的API

官网提供的read_csv函数参数详解
https://pandas.pydata.org/docs/reference/api/pandas.read_csv.html

pandas.read_csv(
	#文件路径,必须要写的参数,其他参数按需要填写
	filepath_or_buffer, 
	sep=NoDefault.no_default, 
	delimiter=None, 
	header='infer', 
	names=NoDefault.no_default, 
	index_col=None, 
	usecols=None, 
	squeeze=False, 
	prefix=NoDefault.no_default, 
	mangle_dupe_cols=True, 
	dtype=None, 
	engine=None, 
	converters=None, 
	true_values=None, 
	false_values=None, 
	skipinitialspace=False, 
	skiprows=None, 
	skipfooter=0, 
	nrows=None, 
	na_values=None, 
	keep_default_na=True, 
	na_filter=True, 
	verbose=False, 
	skip_blank_lines=True, 
	parse_dates=False, 
	infer_datetime_format=False, 
	keep_date_col=False, 
	date_parser=None, 
	dayfirst=False, 
	cache_dates=True, 
	iterator=False, 
	chunksize=None, 
	compression='infer', 
	thousands=None, 
	decimal='.', 
	lineterminator=None, 
	quotechar='"', 
	quoting=0, 
	doublequote=True, 
	escapechar=None, 
	comment=None, 
	encoding=None, 
	encoding_errors='strict', 
	dialect=None, 
	error_bad_lines=None, 
	warn_bad_lines=None, 
	on_bad_lines=None, 
	delim_whitespace=False, 
	low_memory=True, 
	memory_map=False, 
	float_precision=None, 
	storage_options=None)

**

#Input/output
#Pickling
#读取pickling文件
read_pickle(filepath_or_buffer[, ...])
#Load pickled pandas object (or any object) from file.
#写入pickle文件
DataFrame.to_pickle(path[, compression, ...])
#Pickle (serialize) object to file.

#Flat file
read_table(filepath_or_buffer[, sep, ...])
#Read general delimited file into DataFrame.
read_csv(filepath_or_buffer[, sep, ...])
#Read a comma-separated values (csv) file into DataFrame.
DataFrame.to_csv([path_or_buf, sep, na_rep, ...])
#Write object to a comma-separated values (csv) file.
read_fwf(filepath_or_buffer[, colspecs, ...])
#Read a table of fixed-width formatted lines into DataFrame.

#Clipboard
read_clipboard([sep])
#Read text from clipboard and pass to read_csv.
DataFrame.to_clipboard([excel, sep])
#Copy object to the system clipboard.

#Excel
read_excel(io[, sheet_name, header, names, ...])
#Read an Excel file into a pandas DataFrame.
DataFrame.to_excel(excel_writer[, ...])
#Write object to an Excel sheet.
ExcelFile.parse([sheet_name, header, names, ...])
#Parse specified sheet(s) into a DataFrame.
Styler.to_excel(excel_writer[, sheet_name, ...])
#Write Styler to an Excel sheet.
ExcelWriter(path[, engine, date_format, ...])
#Class for writing DataFrame objects into excel sheets.

#JSON
read_json([path_or_buf, orient, typ, dtype, ...])
#Convert a JSON string to pandas object.
to_json(path_or_buf, obj[, orient, ...])
build_table_schema(data[, index, ...])
#Create a Table schema from data.

#HTML
read_html(io[, match, flavor, header, ...])
#Read HTML tables into a list of DataFrame objects.
DataFrame.to_html([buf, columns, col_space, ...])
#Render a DataFrame as an HTML table.
Styler.to_html([buf, table_uuid, ...])
#Write Styler to a file, buffer or string in HTML-CSS format.

#XML
read_xml(path_or_buffer[, xpath, ...])
#Read XML document into a DataFrame object.
DataFrame.to_xml([path_or_buffer, index, ...])
#Render a DataFrame to an XML document.

#Latex
DataFrame.to_latex([buf, columns, ...])
#Render object to a LaTeX tabular, longtable, or nested table/tabular.
Styler.to_latex([buf, column_format, ...])
#Write Styler to a file, buffer or string in LaTeX format.
HDFStore: PyTables (HDF5)
read_hdf(path_or_buf[, key, mode, errors, ...])
#Read from the store, close it if we opened it.
HDFStore.put(key, value[, format, index, ...])
#Store object in HDFStore.
HDFStore.append(key, value[, format, axes, ...])
#Append to Table in file.
HDFStore.get(key)
#Retrieve pandas object stored in file.
HDFStore.select(key[, where, start, stop, ...])
#Retrieve pandas object stored in file, optionally based on where criteria.
HDFStore.info()
#Print detailed information on the store.
HDFStore.keys([include])
#Return a list of keys corresponding to objects stored in HDFStore.
HDFStore.groups()
#Return a list of all the top-level nodes.
HDFStore.walk([where])
#Walk the pytables group hierarchy for pandas objects.
  • 0
    点赞
  • 2
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

xMathematics

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值