数据载入与初步观察

数据载入与初步观察

如何进行数据的导入,如csv,xlsx,等
  1. table

    pandas.read_table(filepath_or_buffer, sep='\t', delimiter=None, header='infer', names=None, index_col=None, usecols=None, squeeze=False, prefix=None, mangle_dupe_cols=True, dtype=None, engine=None, converters=None, true_values=None, false_values=None, skipinitialspace=False, skiprows=None, skipfooter=0, nrows=None, na_values=None, keep_default_na=True, na_filter=True, verbose=False, skip_blank_lines=True, parse_dates=False, infer_datetime_format=False, keep_date_col=False, date_parser=None, dayfirst=False, cache_dates=True, iterator=False, chunksize=None, compression='infer', thousands=None, decimal='.', lineterminator=None, quotechar='"', quoting=0, doublequote=True, escapechar=None, comment=None, encoding=None, dialect=None, error_bad_lines=True, warn_bad_lines=True, delim_whitespace=False, low_memory=True, memory_map=False, float_precision=None)[source]
    Read general delimited file into DataFrame.
    
    Also supports optionally iterating or breaking of the file into chunks.
    
    Additional help can be found in the online docs for IO Tools.
    
    Parameters
    filepath_or_buffer:str, path object or file-like object
    Any valid string path is acceptable. The string could be a URL. Valid URL schemes include http, ftp, s3, gs, and file. For file URLs, a host is expected. A local file could be: file://localhost/path/to/table.csv.
    
    If you want to pass in a path object, pandas accepts any os.PathLike.
    
    By file-like object, we refer to objects with a read() method, such as a file handler (e.g. via builtin open function) or StringIO.
    
    sep:str, default ‘\t’ (tab-stop)
    Delimiter to use. If sep is None, the C engine cannot automatically detect the separator, but the Python parsing engine can, meaning the latter will be used and automatically detect the separator by Python’s builtin sniffer tool, csv.Sniffer. In addition, separators longer than 1 character and different from '\s+' will be interpreted as regular expressions and will also force the use of the Python parsing engine. Note that regex delimiters are prone to ignoring quoted data. Regex example: '\r\t'.
    
    delimiter:str, default None
    Alias for sep.
    
    header:int, list of int, default ‘infer’
    Row number(s) to use as the column names, and the start of the data. Default behavior is to infer the column names: if no names are passed the behavior is identical to header=0 and column names are inferred from the first line of the file, if column names are passed explicitly then the behavior is identical to header=None. Explicitly pass header=0 to be able to replace existing names. The header can be a list of integers that specify row locations for a multi-index on the columns e.g. [0,1,3]. Intervening rows that are not specified will be skipped (e.g. 2 in this example is skipped). Note that this parameter ignores commented lines and empty lines if skip_blank_lines=True, so header=0 denotes the first line of data rather than the first line of the file.
    
    names:array-like, optional
    List of column names to use. If the file contains a header row, then you should explicitly pass header=0 to override the column names. Duplicates in this list are not allowed.
    
    index_col:int, str, sequence of int / str, or False, default None
    Column(s) to use as the row labels of the DataFrame, either given as string name or column index. If a sequence of int / str is given, a MultiIndex is used.
    
    Note: index_col=False can be used to force pandas to not use the first column as the index, e.g. when you have a malformed file with delimiters at the end of each line.
    
    usecols:list-like or callable, optional
    Return a subset of the columns. If list-like, all elements must either be positional (i.e. integer indices into the document columns) or strings that correspond to column names provided either by the user in names or inferred from the document header row(s). For example, a valid list-like usecols parameter would be [0, 1, 2] or ['foo', 'bar', 'baz']. Element order is ignored, so usecols=[0, 1] is the same as [1, 0]. To instantiate a DataFrame from data with element order preserved use pd.read_csv(data, usecols=['foo', 'bar'])[['foo', 'bar']] for columns in ['foo', 'bar'] order or pd.read_csv(data, usecols=['foo', 'bar'])[['bar', 'foo']] for ['bar', 'foo'] order.
    
    If callable, the callable function will be evaluated against the column names, returning names where the callable function evaluates to True. An example of a valid callable argument would be lambda x: x.upper() in ['AAA', 'BBB', 'DDD']. Using this parameter results in much faster parsing time and lower memory usage.
    
    squeeze:bool, default False
    If the parsed data only contains one column then return a Series.
    
    prefix:str, optional
    Prefix to add to column numbers when no header, e.g. ‘X’ for X0, X1, …
    
    mangle_dupe_cols:bool, default True
    Duplicate columns will be specified as ‘X’, ‘X.1, …’X.N’, rather than ‘X’…’X’. Passing in False will cause data to be overwritten if there are duplicate names in the columns.
    
    dtype:Type name or dict of column -> type, optional
    Data type for data or columns. E.g. {‘a’: np.float64, ‘b’: np.int32, ‘c’: ‘Int64’} Use str or object together with suitable na_values settings to preserve and not interpret dtype. If converters are specified, they will be applied INSTEAD of dtype conversion.
    
    engine:{‘c’, ‘python’}, optional
    Parser engine to use. The C engine is faster while the python engine is currently more feature-complete.
    
    converters:dict, optional
    Dict of functions for converting values in certain columns. Keys can either be integers or column labels.
    
    true_values:list, optional
    Values to consider as True.
    
    false_values:list, optional
    Values to consider as False.
    
    skipinitialspace:bool, default False
    Skip spaces after delimiter.
    
    skiprows:list-like, int or callable, optional
    Line numbers to skip (0-indexed) or number of lines to skip (int) at the start of the file.
    
    If callable, the callable function will be evaluated against the row indices, returning True if the row should be skipped and False otherwise. An example of a valid callable argument would be lambda x: x in [0, 2].
    
    skipfooter:int, default 0
    Number of lines at bottom of file to skip (Unsupported with engine=’c’).
    
    nrows:int, optional
    Number of rows of file to read. Useful for reading pieces of large files.
    
    na_values:scalar, str, list-like, or dict, optional
    Additional strings to recognize as NA/NaN. If dict passed, specific per-column NA values. By default the following values are interpreted as NaN: ‘’,#N/A’, ‘#N/A N/A’, ‘#NA’, ‘-1.#IND’, ‘-1.#QNAN’, ‘-NaN’, ‘-nan’, ‘1.#IND’, ‘1.#QNAN’, ‘<NA>’, ‘N/A’, ‘NA’, ‘NULL’, ‘NaN’, ‘n/a’, ‘nan’, ‘null’.
    
    keep_default_na:bool, default True
    Whether or not to include the default NaN values when parsing the data. Depending on whether na_values is passed in, the behavior is as follows:
    
    If keep_default_na is True, and na_values are specified, na_values is appended to the default NaN values used for parsing.
    
    If keep_default_na is True, and na_values are not specified, only the default NaN values are used for parsing.
    
    If keep_default_na is False, and na_values are specified, only the NaN values specified na_values are used for parsing.
    
    If keep_default_na is False, and na_values are not specified, no strings will be parsed as NaN.
    
    Note that if na_filter is passed in as False, the keep_default_na and na_values parameters will be ignored.
    
    na_filter:bool, default True
    Detect missing value markers (empty strings and the value of na_values). In data without any NAs, passing na_filter=False can improve the performance of reading a large file.
    
    verbose:bool, default False
    Indicate number of NA values placed in non-numeric columns.
    
    skip_blank_lines:bool, default True
    If True, skip over blank lines rather than interpreting as NaN values.
    
    parse_dates:bool or list of int or names or list of lists or dict, default False
    The behavior is as follows:
    
    boolean. If True -> try parsing the index.
    
    list of int or names. e.g. If [1, 2, 3] -> try parsing columns 1, 2, 3 each as a separate date column.
    
    list of lists. e.g. If [[1, 3]] -> combine columns 1 and 3 and parse as a single date column.
    
    dict, e.g. {‘foo’ : [1, 3]} -> parse columns 1, 3 as date and call result ‘foo’
    
    If a column or index cannot be represented as an array of datetimes, say because of an unparseable value or a mixture of timezones, the column or index will be returned unaltered as an object data type. For non-standard datetime parsing, use pd.to_datetime after pd.read_csv. To parse an index or column with a mixture of timezones, specify date_parser to be a partially-applied pandas.to_datetime() with utc=True. See Parsing a CSV with mixed timezones for more.
    
    Note: A fast-path exists for iso8601-formatted dates.
    
    infer_datetime_format:bool, default False
    If True and parse_dates is enabled, pandas will attempt to infer the format of the datetime strings in the columns, and if it can be inferred, switch to a faster method of parsing them. In some cases this can increase the parsing speed by 5-10x.
    
    keep_date_col:bool, default False
    If True and parse_dates specifies combining multiple columns then keep the original columns.
    
    date_parser:function, optional
    Function to use for converting a sequence of string columns to an array of datetime instances. The default uses dateutil.parser.parser to do the conversion. Pandas will try to call date_parser in three different ways, advancing to the next if an exception occurs: 1) Pass one or more arrays (as defined by parse_dates) as arguments; 2) concatenate (row-wise) the string values from the columns defined by parse_dates into a single array and pass that; and 3) call date_parser once for each row using one or more strings (corresponding to the columns defined by parse_dates) as arguments.
    
    dayfirst:bool, default False
    DD/MM format dates, international and European format.
    
    cache_dates:bool, default True
    If True, use a cache of unique, converted dates to apply the datetime conversion. May produce significant speed-up when parsing duplicate date strings, especially ones with timezone offsets.
    
    New in version 0.25.0.
    
    iterator:bool, default False
    Return TextFileReader object for iteration or getting chunks with get_chunk().
    
    chunksize:int, optional
    Return TextFileReader object for iteration. See the IO Tools docs for more information on iterator and chunksize.
    
    compression:{‘infer’, ‘gzip’, ‘bz2’,zip, ‘xz’, None}, default ‘infer’
    For on-the-fly decompression of on-disk data. If ‘infer’ and filepath_or_buffer is path-like, then detect compression from the following extensions:.gz’,.bz2’,.zip, or.xz’ (otherwise no decompression). If using ‘zip, the ZIP file must contain only one data file to be read in. Set to None for no decompression.
    
    thousands:str, optional
    Thousands separator.
    
    decimal:str, default ‘.’
    Character to recognize as decimal point (e.g. use ‘,for European data).
    
    lineterminator:str (length 1), optional
    Character to break file into lines. Only valid with C parser.
    
    quotechar:str (length 1), optional
    The character used to denote the start and end of a quoted item. Quoted items can include the delimiter and it will be ignored.
    
    quoting:int or csv.QUOTE_* instance, default 0
    Control field quoting behavior per csv.QUOTE_* constants. Use one of QUOTE_MINIMAL (0), QUOTE_ALL (1), QUOTE_NONNUMERIC (2) or QUOTE_NONE (3).
    
    doublequote:bool, default True
    When quotechar is specified and quoting is not QUOTE_NONE, indicate whether or not to interpret two consecutive quotechar elements INSIDE a field as a single quotechar element.
    
    escapechar:str (length 1), optional
    One-character string used to escape other characters.
    
    comment:str, optional
    Indicates remainder of line should not be parsed. If found at the beginning of a line, the line will be ignored altogether. This parameter must be a single character. Like empty lines (as long as skip_blank_lines=True), fully commented lines are ignored by the parameter header but not by skiprows. For example, if comment='#', parsing #empty\na,b,c\n1,2,3 with header=0 will result in ‘a,b,c’ being treated as the header.
    
    encoding:str, optional
    Encoding to use for UTF when reading/writing (ex. ‘utf-8). List of Python standard encodings .
    
    dialect:str or csv.Dialect, optional
    If provided, this parameter will override values (default or not) for the following parameters: delimiter, doublequote, escapechar, skipinitialspace, quotechar, and quoting. If it is necessary to override values, a ParserWarning will be issued. See csv.Dialect documentation for more details.
    
    error_bad_lines:bool, default True
    Lines with too many fields (e.g. a csv line with too many commas) will by default cause an exception to be raised, and no DataFrame will be returned. If False, then these “bad lines” will dropped from the DataFrame that is returned.
    
    warn_bad_lines:bool, default True
    If error_bad_lines is False, and warn_bad_lines is True, a warning for each “bad line” will be output.
    
    delim_whitespace:bool, default False
    Specifies whether or not whitespace (e.g. ' ' or '    ') will be used as the sep. Equivalent to setting sep='\s+'. If this option is set to True, nothing should be passed in for the delimiter parameter.
    
    low_memory:bool, default True
    Internally process the file in chunks, resulting in lower memory use while parsing, but possibly mixed type inference. To ensure no mixed types either set False, or specify the type with the dtype parameter. Note that the entire file is read into a single DataFrame regardless, use the chunksize or iterator parameter to return the data in chunks. (Only valid with C parser).
    
    memory_map:bool, default False
    If a filepath is provided for filepath_or_buffer, map the file object directly onto memory and access the data directly from there. Using this option can improve performance because there is no longer any I/O overhead.
    
    float_precision:str, optional
    Specifies which converter the C engine should use for floating-point values. The options are None for the ordinary converter, high for the high-precision converter, and round_trip for the round-trip converter.
    
    Returns
    DataFrame or TextParser
    A comma-separated values (csv) file is returned as two-dimensional data structure with labeled axes.
    
  2. csv

    pandas.read_csv(filepath_or_buffer, sep=',', delimiter=None, header='infer', names=None, index_col=None, usecols=None, squeeze=False, prefix=None, mangle_dupe_cols=True, dtype=None, engine=None, converters=None, true_values=None, false_values=None, skipinitialspace=False, skiprows=None, skipfooter=0, nrows=None, na_values=None, keep_default_na=True, na_filter=True, verbose=False, skip_blank_lines=True, parse_dates=False, infer_datetime_format=False, keep_date_col=False, date_parser=None, dayfirst=False, cache_dates=True, iterator=False, chunksize=None, compression='infer', thousands=None, decimal='.', lineterminator=None, quotechar='"', quoting=0, doublequote=True, escapechar=None, comment=None, encoding=None, dialect=None, error_bad_lines=True, warn_bad_lines=True, delim_whitespace=False, low_memory=True, memory_map=False, float_precision=None)[source]
    Read a comma-separated values (csv) file into DataFrame.
    
    Also supports optionally iterating or breaking of the file into chunks.
    
    Additional help can be found in the online docs for IO Tools.
    
    Parameters
    filepath_or_bufferstr, path object or file-like object
    Any valid string path is acceptable. The string could be a URL. Valid URL schemes include http, ftp, s3, gs, and file. For file URLs, a host is expected. A local file could be: file://localhost/path/to/table.csv.
    
    If you want to pass in a path object, pandas accepts any os.PathLike.
    
    By file-like object, we refer to objects with a read() method, such as a file handler (e.g. via builtin open function) or StringIO.
    
    sepstr, default ‘,’
    Delimiter to use. If sep is None, the C engine cannot automatically detect the separator, but the Python parsing engine can, meaning the latter will be used and automatically detect the separator by Python’s builtin sniffer tool, csv.Sniffer. In addition, separators longer than 1 character and different from '\s+' will be interpreted as regular expressions and will also force the use of the Python parsing engine. Note that regex delimiters are prone to ignoring quoted data. Regex example: '\r\t'.
    
    delimiterstr, default None
    Alias for sep.
    
    headerint, list of int, default ‘infer’
    Row number(s) to use as the column names, and the start of the data. Default behavior is to infer the column names: if no names are passed the behavior is identical to header=0 and column names are inferred from the first line of the file, if column names are passed explicitly then the behavior is identical to header=None. Explicitly pass header=0 to be able to replace existing names. The header can be a list of integers that specify row locations for a multi-index on the columns e.g. [0,1,3]. Intervening rows that are not specified will be skipped (e.g. 2 in this example is skipped). Note that this parameter ignores commented lines and empty lines if skip_blank_lines=True, so header=0 denotes the first line of data rather than the first line of the file.
    
    namesarray-like, optional
    List of column names to use. If the file contains a header row, then you should explicitly pass header=0 to override the column names. Duplicates in this list are not allowed.
    
    index_colint, str, sequence of int / str, or False, default None
    Column(s) to use as the row labels of the DataFrame, either given as string name or column index. If a sequence of int / str is given, a MultiIndex is used.
    
    Note: index_col=False can be used to force pandas to not use the first column as the index, e.g. when you have a malformed file with delimiters at the end of each line.
    
    usecolslist-like or callable, optional
    Return a subset of the columns. If list-like, all elements must either be positional (i.e. integer indices into the document columns) or strings that correspond to column names provided either by the user in names or inferred from the document header row(s). For example, a valid list-like usecols parameter would be [0, 1, 2] or ['foo', 'bar', 'baz']. Element order is ignored, so usecols=[0, 1] is the same as [1, 0]. To instantiate a DataFrame from data with element order preserved use pd.read_csv(data, usecols=['foo', 'bar'])[['foo', 'bar']] for columns in ['foo', 'bar'] order or pd.read_csv(data, usecols=['foo', 'bar'])[['bar', 'foo']] for ['bar', 'foo'] order.
    
    If callable, the callable function will be evaluated against the column names, returning names where the callable function evaluates to True. An example of a valid callable argument would be lambda x: x.upper() in ['AAA', 'BBB', 'DDD']. Using this parameter results in much faster parsing time and lower memory usage.
    
    squeezebool, default False
    If the parsed data only contains one column then return a Series.
    
    prefixstr, optional
    Prefix to add to column numbers when no header, e.g. ‘X’ for X0, X1, …
    
    mangle_dupe_colsbool, default True
    Duplicate columns will be specified as ‘X’, ‘X.1, …’X.N’, rather than ‘X’…’X’. Passing in False will cause data to be overwritten if there are duplicate names in the columns.
    
    dtypeType name or dict of column -> type, optional
    Data type for data or columns. E.g. {‘a’: np.float64, ‘b’: np.int32, ‘c’: ‘Int64’} Use str or object together with suitable na_values settings to preserve and not interpret dtype. If converters are specified, they will be applied INSTEAD of dtype conversion.
    
    engine{‘c’, ‘python’}, optional
    Parser engine to use. The C engine is faster while the python engine is currently more feature-complete.
    
    convertersdict, optional
    Dict of functions for converting values in certain columns. Keys can either be integers or column labels.
    
    true_valueslist, optional
    Values to consider as True.
    
    false_valueslist, optional
    Values to consider as False.
    
    skipinitialspacebool, default False
    Skip spaces after delimiter.
    
    skiprowslist-like, int or callable, optional
    Line numbers to skip (0-indexed) or number of lines to skip (int) at the start of the file.
    
    If callable, the callable function will be evaluated against the row indices, returning True if the row should be skipped and False otherwise. An example of a valid callable argument would be lambda x: x in [0, 2].
    
    skipfooterint, default 0
    Number of lines at bottom of file to skip (Unsupported with engine=’c’).
    
    nrowsint, optional
    Number of rows of file to read. Useful for reading pieces of large files.
    
    na_valuesscalar, str, list-like, or dict, optional
    Additional strings to recognize as NA/NaN. If dict passed, specific per-column NA values. By default the following values are interpreted as NaN: ‘’,#N/A’, ‘#N/A N/A’, ‘#NA’, ‘-1.#IND’, ‘-1.#QNAN’, ‘-NaN’, ‘-nan’, ‘1.#IND’, ‘1.#QNAN’, ‘<NA>’, ‘N/A’, ‘NA’, ‘NULL’, ‘NaN’, ‘n/a’, ‘nan’, ‘null’.
    
    keep_default_nabool, default True
    Whether or not to include the default NaN values when parsing the data. Depending on whether na_values is passed in, the behavior is as follows:
    
    If keep_default_na is True, and na_values are specified, na_values is appended to the default NaN values used for parsing.
    
    If keep_default_na is True, and na_values are not specified, only the default NaN values are used for parsing.
    
    If keep_default_na is False, and na_values are specified, only the NaN values specified na_values are used for parsing.
    
    If keep_default_na is False, and na_values are not specified, no strings will be parsed as NaN.
    
    Note that if na_filter is passed in as False, the keep_default_na and na_values parameters will be ignored.
    
    na_filterbool, default True
    Detect missing value markers (empty strings and the value of na_values). In data without any NAs, passing na_filter=False can improve the performance of reading a large file.
    
    verbosebool, default False
    Indicate number of NA values placed in non-numeric columns.
    
    skip_blank_linesbool, default True
    If True, skip over blank lines rather than interpreting as NaN values.
    
    parse_datesbool or list of int or names or list of lists or dict, default False
    The behavior is as follows:
    
    boolean. If True -> try parsing the index.
    
    list of int or names. e.g. If [1, 2, 3] -> try parsing columns 1, 2, 3 each as a separate date column.
    
    list of lists. e.g. If [[1, 3]] -> combine columns 1 and 3 and parse as a single date column.
    
    dict, e.g. {‘foo’ : [1, 3]} -> parse columns 1, 3 as date and call result ‘foo’
    
    If a column or index cannot be represented as an array of datetimes, say because of an unparseable value or a mixture of timezones, the column or index will be returned unaltered as an object data type. For non-standard datetime parsing, use pd.to_datetime after pd.read_csv. To parse an index or column with a mixture of timezones, specify date_parser to be a partially-applied pandas.to_datetime() with utc=True. See Parsing a CSV with mixed timezones for more.
    
    Note: A fast-path exists for iso8601-formatted dates.
    
    infer_datetime_formatbool, default False
    If True and parse_dates is enabled, pandas will attempt to infer the format of the datetime strings in the columns, and if it can be inferred, switch to a faster method of parsing them. In some cases this can increase the parsing speed by 5-10x.
    
    keep_date_colbool, default False
    If True and parse_dates specifies combining multiple columns then keep the original columns.
    
    date_parserfunction, optional
    Function to use for converting a sequence of string columns to an array of datetime instances. The default uses dateutil.parser.parser to do the conversion. Pandas will try to call date_parser in three different ways, advancing to the next if an exception occurs: 1) Pass one or more arrays (as defined by parse_dates) as arguments; 2) concatenate (row-wise) the string values from the columns defined by parse_dates into a single array and pass that; and 3) call date_parser once for each row using one or more strings (corresponding to the columns defined by parse_dates) as arguments.
    
    dayfirstbool, default False
    DD/MM format dates, international and European format.
    
    cache_datesbool, default True
    If True, use a cache of unique, converted dates to apply the datetime conversion. May produce significant speed-up when parsing duplicate date strings, especially ones with timezone offsets.
    
    New in version 0.25.0.
    
    iteratorbool, default False
    Return TextFileReader object for iteration or getting chunks with get_chunk().
    
    chunksizeint, optional
    Return TextFileReader object for iteration. See the IO Tools docs for more information on iterator and chunksize.
    
    compression{‘infer’, ‘gzip’, ‘bz2’,zip, ‘xz’, None}, default ‘infer’
    For on-the-fly decompression of on-disk data. If ‘infer’ and filepath_or_buffer is path-like, then detect compression from the following extensions:.gz’,.bz2’,.zip, or.xz’ (otherwise no decompression). If using ‘zip, the ZIP file must contain only one data file to be read in. Set to None for no decompression.
    
    thousandsstr, optional
    Thousands separator.
    
    decimalstr, default ‘.’
    Character to recognize as decimal point (e.g. use ‘,for European data).
    
    lineterminatorstr (length 1), optional
    Character to break file into lines. Only valid with C parser.
    
    quotecharstr (length 1), optional
    The character used to denote the start and end of a quoted item. Quoted items can include the delimiter and it will be ignored.
    
    quotingint or csv.QUOTE_* instance, default 0
    Control field quoting behavior per csv.QUOTE_* constants. Use one of QUOTE_MINIMAL (0), QUOTE_ALL (1), QUOTE_NONNUMERIC (2) or QUOTE_NONE (3).
    
    doublequotebool, default True
    When quotechar is specified and quoting is not QUOTE_NONE, indicate whether or not to interpret two consecutive quotechar elements INSIDE a field as a single quotechar element.
    
    escapecharstr (length 1), optional
    One-character string used to escape other characters.
    
    commentstr, optional
    Indicates remainder of line should not be parsed. If found at the beginning of a line, the line will be ignored altogether. This parameter must be a single character. Like empty lines (as long as skip_blank_lines=True), fully commented lines are ignored by the parameter header but not by skiprows. For example, if comment='#', parsing #empty\na,b,c\n1,2,3 with header=0 will result in ‘a,b,c’ being treated as the header.
    
    encodingstr, optional
    Encoding to use for UTF when reading/writing (ex. ‘utf-8). List of Python standard encodings .
    
    dialectstr or csv.Dialect, optional
    If provided, this parameter will override values (default or not) for the following parameters: delimiter, doublequote, escapechar, skipinitialspace, quotechar, and quoting. If it is necessary to override values, a ParserWarning will be issued. See csv.Dialect documentation for more details.
    
    error_bad_linesbool, default True
    Lines with too many fields (e.g. a csv line with too many commas) will by default cause an exception to be raised, and no DataFrame will be returned. If False, then these “bad lines” will dropped from the DataFrame that is returned.
    
    warn_bad_linesbool, default True
    If error_bad_lines is False, and warn_bad_lines is True, a warning for each “bad line” will be output.
    
    delim_whitespacebool, default False
    Specifies whether or not whitespace (e.g. ' ' or '    ') will be used as the sep. Equivalent to setting sep='\s+'. If this option is set to True, nothing should be passed in for the delimiter parameter.
    
    low_memorybool, default True
    Internally process the file in chunks, resulting in lower memory use while parsing, but possibly mixed type inference. To ensure no mixed types either set False, or specify the type with the dtype parameter. Note that the entire file is read into a single DataFrame regardless, use the chunksize or iterator parameter to return the data in chunks. (Only valid with C parser).
    
    memory_mapbool, default False
    If a filepath is provided for filepath_or_buffer, map the file object directly onto memory and access the data directly from there. Using this option can improve performance because there is no longer any I/O overhead.
    
    float_precisionstr, optional
    Specifies which converter the C engine should use for floating-point values. The options are None for the ordinary converter, high for the high-precision converter, and round_trip for the round-trip converter.
    
    Returns
    DataFrame or TextParser
    A comma-separated values (csv) file is returned as two-dimensional data structure with labeled axes.
    
  3. excel

    pandas.read_excel(*args, **kwargs)[source]
    Read an Excel file into a pandas DataFrame.
    
    Supports xls, xlsx, xlsm, xlsb, odf, ods and odt file extensions read from a local filesystem or URL. Supports an option to read a single sheet or a list of sheets.
    
    Parameters
    iostr, bytes, ExcelFile, xlrd.Book, path object, or file-like object
    Any valid string path is acceptable. The string could be a URL. Valid URL schemes include http, ftp, s3, and file. For file URLs, a host is expected. A local file could be: file://localhost/path/to/table.xlsx.
    
    If you want to pass in a path object, pandas accepts any os.PathLike.
    
    By file-like object, we refer to objects with a read() method, such as a file handler (e.g. via builtin open function) or StringIO.
    
    sheet_namestr, int, list, or None, default 0
    Strings are used for sheet names. Integers are used in zero-indexed sheet positions. Lists of strings/integers are used to request multiple sheets. Specify None to get all sheets.
    
    Available cases:
    
    Defaults to 0: 1st sheet as a DataFrame
    
    1: 2nd sheet as a DataFrame
    
    "Sheet1": Load sheet with name “Sheet1”
    
    [0, 1, "Sheet5"]: Load first, second and sheet named “Sheet5” as a dict of DataFrame
    
    None: All sheets.
    
    headerint, list of int, default 0
    Row (0-indexed) to use for the column labels of the parsed DataFrame. If a list of integers is passed those row positions will be combined into a MultiIndex. Use None if there is no header.
    
    namesarray-like, default None
    List of column names to use. If file contains no header row, then you should explicitly pass header=None.
    
    index_colint, list of int, default None
    Column (0-indexed) to use as the row labels of the DataFrame. Pass None if there is no such column. If a list is passed, those columns will be combined into a MultiIndex. If a subset of data is selected with usecols, index_col is based on the subset.
    
    usecolsint, str, list-like, or callable default None
    If None, then parse all columns.
    
    If str, then indicates comma separated list of Excel column letters and column ranges (e.g. “A:E” or “A,C,E:F”). Ranges are inclusive of both sides.
    
    If list of int, then indicates list of column numbers to be parsed.
    
    If list of string, then indicates list of column names to be parsed.
    
    New in version 0.24.0.
    
    If callable, then evaluate each column name against it and parse the column if the callable returns True.
    
    Returns a subset of the columns according to behavior above.
    
    New in version 0.24.0.
    
    squeezebool, default False
    If the parsed data only contains one column then return a Series.
    
    dtypeType name or dict of column -> type, default None
    Data type for data or columns. E.g. {‘a’: np.float64, ‘b’: np.int32} Use object to preserve data as stored in Excel and not interpret dtype. If converters are specified, they will be applied INSTEAD of dtype conversion.
    
    enginestr, default None
    If io is not a buffer or path, this must be set to identify io. Supported engines: “xlrd”, “openpyxl”, “odf”, “pyxlsb”, default “xlrd”. Engine compatibility : - “xlrd” supports most old/new Excel file formats. - “openpyxl” supports newer Excel file formats. - “odf” supports OpenDocument file formats (.odf, .ods, .odt). - “pyxlsb” supports Binary Excel files.
    
    convertersdict, default None
    Dict of functions for converting values in certain columns. Keys can either be integers or column labels, values are functions that take one input argument, the Excel cell content, and return the transformed content.
    
    true_valueslist, default None
    Values to consider as True.
    
    false_valueslist, default None
    Values to consider as False.
    
    skiprowslist-like
    Rows to skip at the beginning (0-indexed).
    
    nrowsint, default None
    Number of rows to parse.
    
    New in version 0.23.0.
    
    na_valuesscalar, str, list-like, or dict, default None
    Additional strings to recognize as NA/NaN. If dict passed, specific per-column NA values. By default the following values are interpreted as NaN: ‘’,#N/A’, ‘#N/A N/A’, ‘#NA’, ‘-1.#IND’, ‘-1.#QNAN’, ‘-NaN’, ‘-nan’, ‘1.#IND’, ‘1.#QNAN’, ‘<NA>’, ‘N/A’, ‘NA’, ‘NULL’, ‘NaN’, ‘n/a’, ‘nan’, ‘null’.
    
    keep_default_nabool, default True
    Whether or not to include the default NaN values when parsing the data. Depending on whether na_values is passed in, the behavior is as follows:
    
    If keep_default_na is True, and na_values are specified, na_values is appended to the default NaN values used for parsing.
    
    If keep_default_na is True, and na_values are not specified, only the default NaN values are used for parsing.
    
    If keep_default_na is False, and na_values are specified, only the NaN values specified na_values are used for parsing.
    
    If keep_default_na is False, and na_values are not specified, no strings will be parsed as NaN.
    
    Note that if na_filter is passed in as False, the keep_default_na and na_values parameters will be ignored.
    
    na_filterbool, default True
    Detect missing value markers (empty strings and the value of na_values). In data without any NAs, passing na_filter=False can improve the performance of reading a large file.
    
    verbosebool, default False
    Indicate number of NA values placed in non-numeric columns.
    
    parse_datesbool, list-like, or dict, default False
    The behavior is as follows:
    
    bool. If True -> try parsing the index.
    
    list of int or names. e.g. If [1, 2, 3] -> try parsing columns 1, 2, 3 each as a separate date column.
    
    list of lists. e.g. If [[1, 3]] -> combine columns 1 and 3 and parse as a single date column.
    
    dict, e.g. {‘foo’ : [1, 3]} -> parse columns 1, 3 as date and call result ‘foo’
    
    If a column or index contains an unparseable date, the entire column or index will be returned unaltered as an object data type. If you don`t want to parse some cells as date just change their type in Excel to “Text”. For non-standard datetime parsing, use pd.to_datetime after pd.read_excel.
    
    Note: A fast-path exists for iso8601-formatted dates.
    
    date_parserfunction, optional
    Function to use for converting a sequence of string columns to an array of datetime instances. The default uses dateutil.parser.parser to do the conversion. Pandas will try to call date_parser in three different ways, advancing to the next if an exception occurs: 1) Pass one or more arrays (as defined by parse_dates) as arguments; 2) concatenate (row-wise) the string values from the columns defined by parse_dates into a single array and pass that; and 3) call date_parser once for each row using one or more strings (corresponding to the columns defined by parse_dates) as arguments.
    
    thousandsstr, default None
    Thousands separator for parsing string columns to numeric. Note that this parameter is only necessary for columns stored as TEXT in Excel, any numeric columns will automatically be parsed, regardless of display format.
    
    commentstr, default None
    Comments out remainder of line. Pass a character or characters to this argument to indicate comments in the input file. Any data between the comment string and the end of the current line is ignored.
    
    skipfooterint, default 0
    Rows at the end to skip (0-indexed).
    
    convert_floatbool, default True
    Convert integral floats to int (i.e., 1.0> 1). If False, all numeric data will be read in as floats: Excel stores all numbers as floats internally.
    
    mangle_dupe_colsbool, default True
    Duplicate columns will be specified as ‘X’, ‘X.1, …’X.N’, rather than ‘X’…’X’. Passing in False will cause data to be overwritten if there are duplicate names in the columns.
    
    Returns
    DataFrame or dict of DataFrames
    DataFrame from the passed in Excel file. See notes in sheet_name argument for more information on when a dict of DataFrames is returned.
    
  4. sql

    pandas.read_sql(sql, con, index_col='None', coerce_float='True', params='None', parse_dates='None', columns='None', chunksize: None = 'None') → DataFrame[source]
    pandas.read_sql(sql, con, index_col='None', coerce_float='True', params='None', parse_dates='None', columns='None', chunksize: int = '1') → Iterator[DataFrame]
    Read SQL query or database table into a DataFrame.
    
    This function is a convenience wrapper around read_sql_table and read_sql_query (for backward compatibility). It will delegate to the specific function depending on the provided input. A SQL query will be routed to read_sql_query, while a database table name will be routed to read_sql_table. Note that the delegated function might have more specific notes about their functionality not listed here.
    
    Parameters
    sqlstr or SQLAlchemy Selectable (select or text object)
    SQL query to be executed or a table name.
    
    conSQLAlchemy connectable, str, or sqlite3 connection
    Using SQLAlchemy makes it possible to use any DB supported by that library. If a DBAPI2 object, only sqlite3 is supported. The user is responsible for engine disposal and connection closure for the SQLAlchemy connectable. See here.
    
    index_colstr or list of str, optional, default: None
    Column(s) to set as index(MultiIndex).
    
    coerce_floatbool, default True
    Attempts to convert values of non-string, non-numeric objects (like decimal.Decimal) to floating point, useful for SQL result sets.
    
    paramslist, tuple or dict, optional, default: None
    List of parameters to pass to execute method. The syntax used to pass parameters is database driver dependent. Check your database driver documentation for which of the five syntax styles, described in PEP 249’s paramstyle, is supported. Eg. for psycopg2, uses %(name)s so use params={‘name’ : ‘value’}.
    
    parse_dateslist or dict, default: None
    List of column names to parse as dates.
    
    Dict of {column_name: format string} where format string is strftime compatible in case of parsing string times, or is one of (D, s, ns, ms, us) in case of parsing integer timestamps.
    
    Dict of {column_name: arg dict}, where the arg dict corresponds to the keyword arguments of pandas.to_datetime() Especially useful with databases without native Datetime support, such as SQLite.
    
    columnslist, default: None
    List of column names to select from SQL table (only used when reading a table).
    
    chunksizeint, default None
    If specified, return an iterator where chunksize is the number of rows to include in each chunk.
    
    Returns
    DataFrame or Iterator[DataFrame]
    
  5. json

    pandas.read_json
    pandas.read_json(*args, **kwargs)[source]
    Convert a JSON string to pandas object.
    
    Parameters
    path_or_bufa valid JSON str, path object or file-like object
    Any valid string path is acceptable. The string could be a URL. Valid URL schemes include http, ftp, s3, and file. For file URLs, a host is expected. A local file could be: file://localhost/path/to/table.json.
    
    If you want to pass in a path object, pandas accepts any os.PathLike.
    
    By file-like object, we refer to objects with a read() method, such as a file handler (e.g. via builtin open function) or StringIO.
    
    orientstr
    Indication of expected JSON string format. Compatible JSON strings can be produced by to_json() with a corresponding orient value. The set of possible orients is:
    
    'split' : dict like {index -> [index], columns -> [columns], data -> [values]}
    
    'records' : list like [{column -> value}, ... , {column -> value}]
    
    'index' : dict like {index -> {column -> value}}
    
    'columns' : dict like {column -> {index -> value}}
    
    'values' : just the values array
    
    The allowed and default values depend on the value of the typ parameter.
    
    when typ == 'series',
    
    allowed orients are {'split','records','index'}
    
    default is 'index'
    
    The Series index must be unique for orient 'index'.
    
    when typ == 'frame',
    
    allowed orients are {'split','records','index', 'columns','values', 'table'}
    
    default is 'columns'
    
    The DataFrame index must be unique for orients 'index' and 'columns'.
    
    The DataFrame columns must be unique for orients 'index', 'columns', and 'records'.
    
    New in version 0.23.0: ‘table’ as an allowed value for the orient argument
    
    typ{‘frame’, ‘series’}, default ‘frame’
    The type of object to recover.
    
    dtypebool or dict, default None
    If True, infer dtypes; if a dict of column to dtype, then use those; if False, then don’t infer dtypes at all, applies only to the data.
    
    For all orient values except 'table', default is True.
    
    Changed in version 0.25.0: Not applicable for orient='table'.
    
    convert_axesbool, default None
    Try to convert the axes to the proper dtypes.
    
    For all orient values except 'table', default is True.
    
    Changed in version 0.25.0: Not applicable for orient='table'.
    
    convert_datesbool or list of str, default True
    If True then default datelike columns may be converted (depending on keep_default_dates). If False, no dates will be converted. If a list of column names, then those columns will be converted and default datelike columns may also be converted (depending on keep_default_dates).
    
    keep_default_datesbool, default True
    If parsing dates (convert_dates is not False), then try to parse the default datelike columns. A column label is datelike if
    
    it ends with '_at',
    
    it ends with '_time',
    
    it begins with 'timestamp',
    
    it is 'modified', or
    
    it is 'date'.
    
    numpybool, default False
    Direct decoding to numpy arrays. Supports numeric data only, but non-numeric column and index labels are supported. Note also that the JSON ordering MUST be the same for each term if numpy=True.
    
    Deprecated since version 1.0.0.
    precise_floatbool, default False
    Set to enable usage of higher precision (strtod) function when decoding string to double values. Default (False) is to use fast but less precise builtin functionality.
    
    date_unitstr, default None
    The timestamp unit to detect if converting dates. The default behaviour is to try and detect the correct precision, but if this is not desired then pass one of ‘s’, ‘ms’, ‘us’ or ‘ns’ to force parsing only seconds, milliseconds, microseconds or nanoseconds respectively.
    
    encodingstr, default is ‘utf-8’
    The encoding to use to decode py3 bytes.
    
    linesbool, default False
    Read the file as a json object per line.
    
    chunksizeint, optional
    Return JsonReader object for iteration. See the line-delimited json docs for more information on chunksize. This can only be passed if lines=True. If this is None, the file will be read into memory all at once.
    
    compression{‘infer’, ‘gzip’, ‘bz2’,zip, ‘xz’, None}, default ‘infer’
    For on-the-fly decompression of on-disk data. If ‘infer’, then use gzip, bz2, zip or xz if path_or_buf is a string ending in.gz’,.bz2’,.zip, or ‘xz’, respectively, and no decompression otherwise. If using ‘zip, the ZIP file must contain only one data file to be read in. Set to None for no decompression.
    
    nrowsint, optional
    The number of lines from the line-delimited jsonfile that has to be read. This can only be passed if lines=True. If this is None, all the rows will be returned.
    
    New in version 1.1.
    
    Returns
    Series or DataFrame
    The type returned depends on the value of typ.
    
什么是逐块读取?为什么要逐块读取呢?

使用pandas来处理文件的时候,经常会遇到大文件,而有时候我们只想要读取其中的一部分数据或对文件进行逐块处理。

1、读取文件中前部分

通过nrows参数,来设置读取文件的前多少行,nrows是一个大于等于0的整数。

2、逐块读取文件
    #设置chunksize参数,来控制每次迭代数据的大小



    chunker = pd.read_csv("data.csv",chunksize=5)



    for piece in chunker:



        print(type(piece))



        #<class 'pandas.core.frame.DataFrame'>



        print(len(piece))



        #5

如果,将上面的chunksize参数设置为8,那么第一次迭代的时候包含8条数据,第二次迭代就只能包含2条数据了。其实,每次进行迭代的时候还是一个DataFrame类型的数据结构。

所谓将表头改为中文其中一个思路是:将英文额度表头替换成中文。还有其他的方法吗?
  1. 使用rename函数更改列名
  2. 直接更改全部的列名称
对于一个数据,还可以从哪些方面来观察?找找答案,这个将对下面的数据分析有很大的帮助
  1. 分析数据得相关性
    1. groupby方法
    2. crosstab
    3. cov \corr()
还有其他的删除多余的列的方式吗?
  1. 使用del
  2. df.drop(‘columns’,axis=1)#删除不改表原始数据,可以通过重新赋值的方式赋值该数据
  3. df.drop(‘columns’,axis=1,inplace=‘True’) #改变原始数据
对比任务五和任务六,是不是使用了不一样的方法(函数),如果使用一样的函数如何完成上面的不同的要求呢?

如果想要完全的删除你的数据结构,使用inplace=True,因为使用inplace就将原数据覆盖了,所以这里没有用

reset_index()函数的作用是什么?

重新排序 .因为通过选取符合条件的行或者列 原来的行与列序号并不是顺序的,所以通过此函数可以重新顺序排列

使用pan重新排序

因为通过选取符合条件的行或者列 原来的行与列序号并不是顺序的,所以通过此函数可以重新顺序排列das提出的简单方式,你可以看看loc方法,对比整体的数据位置,你有发现什么问题吗?那么如何解决?

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值