This page was generated from pandas2.ipynb.
Pandas (continues)¶
[1]:
import pandas as pd
import numpy as np
Creation of dataframes¶
The DataFrame is essentially a two dimensional object, and it can be created in three different ways:
out of a two dimensional NumPy array
out of given columns
out of given rows
Creating DataFrames from a NumPy array¶
In the following example a DataFrame with 2 rows and 3 column is created. The row and column indices are given explicitly.
[2]:
df=pd.DataFrame(np.random.randn(2,3), columns=["First", "Second", "Third"], index=["a", "b"])
df
[2]:
First | Second | Third | |
---|---|---|---|
a | 1.273012 | -1.645268 | 0.133877 |
b | 0.742194 | -0.225893 | 2.600842 |
Note that now both the rows and columns can be accessed using the special Index
object:
[3]:
df.index # These are the "row names"
[3]:
Index(['a', 'b'], dtype='object')
[4]:
df.columns # These are the "column names"
[4]:
Index(['First', 'Second', 'Third'], dtype='object')
If either columns
or index
argument is left out, then an implicit integer index will be used:
[5]:
df2=pd.DataFrame(np.random.randn(2,3), index=["a", "b"])
df2
[5]:
0 | 1 | 2 | |
---|---|---|---|
a | 0.095937 | -0.688698 | 0.653831 |
b | 0.978782 | 0.034069 | 1.125257 |
Now the column index is an object similar to Python’s builtin range
type:
[6]:
df2.columns
[6]:
RangeIndex(start=0, stop=3, step=1)
Creating DataFrames from columns¶
A column can be specified as a list, an NumPy array, or a Pandas’ Series. The names of the columns can be given either with the columns
parameter, or if Series objects are used, then the name
attribute of each Series is used as the column name.
[7]:
s1 = pd.Series([1,2,3])
s1
[7]:
0 1
1 2
2 3
dtype: int64
[8]:
s2 = pd.Series([4,5,6], name="b")
s2
[8]:
0 4
1 5
2 6
Name: b, dtype: int64
Give the column name explicitly:
[9]:
pd.DataFrame(s1, columns=["a"])
[9]:
a | |
---|---|
0 | 1 |
1 | 2 |
2 | 3 |
Use the name
attribute of Series s2 as the column name:
[10]:
pd.DataFrame(s2)
[10]:
b | |
---|---|
0 | 4 |
1 | 5 |
2 | 6 |
If using multiple columns, then they must be given as the dictionary, whose keys give the column names and values are the actual column content.
[11]:
pd.DataFrame({"a": s1, "b": s2})
[11]:
a | b | |
---|---|---|
0 | 1 | 4 |
1 | 2 | 5 |
2 | 3 | 6 |
Creating DataFrames from rows¶
We can give a list of rows as a parameter to the DataFrame constructor. Each row is given as a dict, list, Series, or NumPy array. If we want to give names for the columns, then either the rows must be dictionaries, where the key is the column name and the values are the elements of the DataFrame on that row and column, or else the column names must be given explicitly. An example of this:
[12]:
df=pd.DataFrame([{"Wage" : 1000, "Name" : "Jack", "Age" : 21}, {"Wage" : 1500, "Name" : "John", "Age" : 29}])
df
[12]:
Age | Name | Wage | |
---|---|---|---|
0 | 21 | Jack | 1000 |
1 | 29 | John | 1500 |
Or:
[13]:
df = pd.DataFrame([[1000, "Jack", 21], [1500, "John", 29]], columns=["Wage", "Name", "Age"])
df
[13]:
Wage | Name | Age | |
---|---|---|---|
0 | 1000 | Jack | 21 |
1 | 1500 | John | 29 |
Note that the order of columns is not always the same order as they were in the parameter list. In this case you can use the columns
parameter to specify the exact order.
In the earlier case, however, where we created DataFrames from a dictionary of columns, the order of columns should be the same as in the parameter dictionary in the recent versions of Python and Pandas.
In the sense of information content the order of columns should not matter, but sometimes you want to specify a certain order to make the Frame more readable, or to make it obey some semantic meaning of column order.
Write function cities
that returns the following DataFrame of top Finnish cities by population:
Population Total area
Helsinki 643272 715.48
Espoo 279044 528.03
Tampere 231853 689.59
Vantaa 223027 240.35
Oulu 201810 3817.52
Make function powers_of_series
that takes a Series and a positive integer k
as parameters and returns a DataFrame. The resulting DataFrame should have the same index as the input Series. The first column of the dataFrame should be the input Series, the second column should contain the Series raised to power of two. The third column should contain the Series raised to the power of three, and so on until (and including) power of k
. The columns should have indices from 1 to k.
The values should be numbers, but the index can have any type. Test your function from the main
function. Example of usage:
s = pd.Series([1,2,3,4], index=list("abcd"))
print(powers_of_series(s, 3))
Should print:
1 2 3
a 1 1 1
b 2 4 8
c 3 9 27
d 4 16 64
In the main
function load a data set of municipal information from the src
folder (originally from Statistics Finland). Use the function pd.read_csv
, and note that the separator is a tabulator.
Print the shape of the DataFrame (number of rows and columns) and the column names in the following format:
Shape: r,c
Columns:
col1
col2
...
Note, sometimes file ending tsv
(tab separated values) is used instead of csv
if the separator is a tab.
Accessing columns and rows of a dataframe¶
Even though DataFrames are basically just two dimensional arrays, the way to access their elements is different from NumPy arrays. There are a couple of complications, which we will go through in this section.
Firstly, the bracket notation []
does not allow the use of an index pair to access a single element of the DataFrame. Instead only one dimension can be specified.
Well, does this dimension specify the rows of the DataFrame, like NumPy arrays if only one index is given, or does it specify the columns of the DataFrame?
It depends!
If an integer is used, then it specifies a column of the DataFrame in the case the explicit indices for the column contain that integer. In any other case an error will result. For example, with the above DataFrame, the following indexing will not work, because the explicit column index consist of the column names “Name” and “Wage” which are not integers.
[14]:
try:
df[0]
except KeyError:
import sys
print("Key error", file=sys.stderr)
Key error
The following will however work.
[15]:
df["Wage"]
[15]:
0 1000
1 1500
Name: Wage, dtype: int64
As does the fancy indexing:
[16]:
df[["Wage", "Name"]]
[16]:
Wage | Name | |
---|---|---|
0 | 1000 | Jack |
1 | 1500 | John |
If one indexes with a slice or a boolean mask, then the rows are referred to. Examples of these:
[17]:
df[0:1] # slice
[17]:
Wage | Name | Age | |
---|---|---|---|
0 | 1000 | Jack | 21 |
[18]:
df[df.Wage > 1200] # boolean mask
[18]:
Wage | Name | Age | |
---|---|---|---|
1 | 1500 | John | 29 |
If some of the above calls return a Series object, then you can chain the bracket calls to get a single value from the DataFrame:
[19]:
df["Wage"][1] # Note order of dimensions
[19]:
1500
But there is a better way to achieve this, which we will see in the next section.
Load again the municipal information DataFrame. The rows of the DataFrame correspond to various geographical areas of Finland. The first row is about Finland as a whole, then rows from Akaa to Äänekoski are municipalities of Finland in alphabetical order. After that some larger regions are listed.
Write function municipalities_of_finland
that returns a DataFrame containing only rows about municipalities. Give an appropriate argument for pd.read_csv
so that it interprets the column about region name as the (row) index. This way you can index the DataFrame with the names of the regions.
Test your function from the main
function.
Write function swedish_and_foreigners
that
Reads the municipalities data set
Takes the subset about municipalities (like in previous exercise)
Further take a subset of rows that have proportion of Swedish speaking people and proportion of foreigners both above 5 % level
From this data set take only columns about population, the proportions of Swedish speaking people and foreigners, that is three columns.
The function should return this final DataFrame.
Do you see some kind of correlation between the columns about Swedish speaking and foreign people? Do you see correlation between the columns about the population and the proportion of Swedish speaking people in this subset?
Write function growing_municipalities
that gets subset of municipalities (a DataFrame) as a parameter and returns the proportion of municipalities with increasing population in that subset.
Test your function from the main
function using some subset of the municipalities. Print the proportion as percentages using 1 decimal precision.
Example output:
Proportion of growing municipalities: 12.4%
Alternative indexing and data selection¶
If the explanation in the previous section sounded confusing or ambiguous, or if you didn’t understand a thing, you don’t have to worry.
There is another way to index Pandas DataFrames, which
allows use of index pairs to access a single element
has the same order of dimensions as NumPy: first index specifies rows, second columns
is not ambiguous about implicit or explicit indices
Pandas DataFrames have attributes loc
and iloc
that have the above qualities. You can use loc
and iloc
attributes and forget everything about the previous section. Or you can use these attributes and sometimes use the methods from the previous section as shortcuts if you understand them well.
The difference between loc
and iloc
attributes is that the former uses explicit indices and the latter uses the implicit integer indices. Examples of use:
[20]:
df.loc[1, "Wage"]
[20]:
1500
[21]:
df.iloc[-1,-1] # Right lower corner of the DataFrame
[21]:
29
[22]:
df.loc[1, ["Name", "Wage"]]
[22]:
Name John
Wage 1500
Name: 1, dtype: object
With iloc
everything works like with NumPy arrays: indexing, slicing, fancy indexing, masking and their combinations. With loc
it is the same but now the names in the explicit indices are used for specifying rows and columns. Make sure your understand why the above examples work as they do!
Write function subsetting_with_loc
that in one go takes the subset of municipalities from Akaa to Äänekoski and restricts it to columns: “Population”, “Share of Swedish-speakers of the population, %”, and “Share of foreign citizens of the population, %”. The function should return this content as a DataFrame. Use the attribute loc
.
Write function subsetting_by_positions
that does the following.
Read the data set of the top forty singles from the beginning of the year 1964 from the src
folder. Return the top 10 entries and only the columns Title
and Artist
. Get these elements by their positions, that is, by using a single call to the iloc
attribute. The function should return these as a DataFrame.
Summary statistics¶
The summary statistic methods work in a similar way as their counter parts in NumPy. By default, the aggregation is done over columns.
[23]:
wh = pd.read_csv("https://raw.githubusercontent.com/csmastersUH/data_analysis_with_python_2020/master/kumpula-weather-2017.csv")
[24]:
wh2 = wh.drop(["Year", "m", "d"], axis=1) # taking averages over these is not very interesting
wh2.mean()
[24]:
Precipitation amount (mm) 1.966301
Snow depth (cm) 0.966480
Air temperature (degC) 6.527123
dtype: float64
The describe
method of the DataFrame
object gives different summary statistics for each (numeric) column. The result is a DataFrame. This method gives a good overview of the data, and is typically used in the exploratory data analysis phase.
[25]:
wh.describe()
[25]:
Year | m | d | Precipitation amount (mm) | Snow depth (cm) | Air temperature (degC) | |
---|---|---|---|---|---|---|
count | 365.0 | 365.000000 | 365.000000 | 365.000000 | 358.000000 | 365.000000 |
mean | 2017.0 | 6.526027 | 15.720548 | 1.966301 | 0.966480 | 6.527123 |
std | 0.0 | 3.452584 | 8.808321 | 4.858423 | 3.717472 | 7.183934 |
min | 2017.0 | 1.000000 | 1.000000 | -1.000000 | -1.000000 | -17.800000 |
25% | 2017.0 | 4.000000 | 8.000000 | -1.000000 | -1.000000 | 1.200000 |
50% | 2017.0 | 7.000000 | 16.000000 | 0.200000 | -1.000000 | 4.800000 |
75% | 2017.0 | 10.000000 | 23.000000 | 2.700000 | 0.000000 | 12.900000 |
max | 2017.0 | 12.000000 | 31.000000 | 35.000000 | 15.000000 | 19.600000 |
Write function snow_depth
that reads in the weather DataFrame from the src
folder and returns the maximum amount of snow in the year 2017.
Print the result in the main
function in the following form:
Max snow depth: xx.x
Write function average_temperature
that reads the weather data set and returns the average temperature in July.
Print the result in the main
function in the following form:
Average temperature in July: xx.x
Write function below_zero
that returns the number of days when the temperature was below zero.
Print the result in the main function in the following form:
Number of days below zero: xx
Missing data¶
You may have noticed something strange in the output of the describe
method. First, the minimum value in both precipitation and snow depth fields is -1. The special value -1 means that on that day there was absolutely no snow or rain, whereas the value 0 might indicate that the value was close to zero. Secondly, the snow depth column has count 358, whereas the other columns have count 365, one measurement/value for each day of the year. How is this possible? Every field in a DataFrame should
have the same number of rows. Let’s use the unique
method of the Series object to find out, which different values are used in this column:
[26]:
wh["Snow depth (cm)"].unique()
[26]:
array([ -1., 7., 13., 10., 12., 9., 8., 5., 6., 4., 3.,
15., 14., 2., nan, 0.])
The float
type allows a special value nan
(Not A Number), in addition to normal floating point numbers. This value can represent the result from an illegal operation. For example, the operation 0/0 can either cause an exception to occur or just silently produce a nan
. In Pandas nan
can be used to represent a missing value. In the weather DataFrame the nan
value tells us that the measurement from that day is not available, possibly due to a broken measuring instrument or some
other problem.
Note that only float types allow the nan
value (in Python, NumPy or Pandas). So, if we try to create an integer series with missing values, its dtype gets promoted to float
:
[27]:
pd.Series([1,3,2])
[27]:
0 1
1 3
2 2
dtype: int64
[28]:
pd.Series([1,3,2, np.nan])
[28]:
0 1.0
1 3.0
2 2.0
3 NaN
dtype: float64
For non-numeric types the special value None
is used to denote a missing value, and the dtype is promoted to object
.
[29]:
pd.Series(["jack", "joe", None])
[29]:
0 jack
1 joe
2 None
dtype: object
Pandas excludes the missing values from the summary statistics, like we saw in the previous section. Pandas also provides some functions to handle missing values.
The missing values can be located with the isnull
method:
[30]:
wh.isnull() # returns a boolean mask DataFrame
[30]:
Year | m | d | Time | Time zone | Precipitation amount (mm) | Snow depth (cm) | Air temperature (degC) | |
---|---|---|---|---|---|---|---|---|
0 | False | False | False | False | False | False | False | False |
1 | False | False | False | False | False | False | False | False |
2 | False | False | False | False | False | False | False | False |
3 | False | False | False | False | False | False | False | False |
4 | False | False | False | False | False | False | False | False |
5 | False | False | False | False | False | False | False | False |
6 | False | False | False | False | False | False | False | False |
7 | False | False | False | False | False | False | False | False |
8 | False | False | False | False | False | False | False | False |
9 | False | False | False | False | False | False | False | False |
10 | False | False | False | False | False | False | False | False |
11 | False | False | False | False | False | False | False | False |
12 | False | False | False | False | False | False | False | False |
13 | False | False | False | False | False | False | False | False |
14 | False | False | False | False | False | False | False | False |
15 | False | False | False | False | False | False | False | False |
16 | False | False | False | False | False | False | False | False |
17 | False | False | False | False | False | False | False | False |
18 | False | False | False | False | False | False | False | False |
19 | False | False | False | False | False | False | False | False |
20 | False | False | False | False | False | False | False | False |
21 | False | False | False | False | False | False | False | False |
22 | False | False | False | False | False | False | False | False |
23 | False | False | False | False | False | False | False | False |
24 | False | False | False | False | False | False | False | False |
25 | False | False | False | False | False | False | False | False |
26 | False | False | False | False | False | False | False | False |
27 | False | False | False | False | False | False | False | False |
28 | False | False | False | False | False | False | False | False |
29 | False | False | False | False | False | False | False | False |
... | ... | ... | ... | ... | ... | ... | ... | ... |
335 | False | False | False | False | False | False | False | False |
336 | False | False | False | False | False | False | False | False |
337 | False | False | False | False | False | False | False | False |
338 | False | False | False | False | False | False | False | False |
339 | False | False | False | False | False | False | False | False |
340 | False | False | False | False | False | False | False | False |
341 | False | False | False | False | False | False | False | False |
342 | False | False | False | False | False | False | False | False |
343 | False | False | False | False | False | False | False | False |
344 | False | False | False | False | False | False | False | False |
345 | False | False | False | False | False | False | False | False |
346 | False | False | False | False | False | False | False | False |
347 | False | False | False | False | False | False | False | False |
348 | False | False | False | False | False | False | False | False |
349 | False | False | False | False | False | False | False | False |
350 | False | False | False | False | False | False | False | False |
351 | False | False | False | False | False | False | False | False |
352 | False | False | False | False | False | False | False | False |
353 | False | False | False | False | False | False | False | False |
354 | False | False | False | False | False | False | False | False |
355 | False | False | False | False | False | False | False | False |
356 | False | False | False | False | False | False | False | False |
357 | False | False | False | False | False | False | False | False |
358 | False | False | False | False | False | False | False | False |
359 | False | False | False | False | False | False | False | False |
360 | False | False | False | False | False | False | False | False |
361 | False | False | False | False | False | False | False | False |
362 | False | False | False | False | False | False | False | False |
363 | False | False | False | False | False | False | False | False |
364 | False | False | False | False | False | False | False | False |
365 rows × 8 columns
This is not very useful as we cannot directly use the mask to index the DataFrame. We can, however, combine it with the any
method to find out all the rows that contain at least one missing value:
[31]:
wh[wh.isnull().any(axis=1)]
[31]:
Year | m | d | Time | Time zone | Precipitation amount (mm) | Snow depth (cm) | Air temperature (degC) | |
---|---|---|---|---|---|---|---|---|
74 | 2017 | 3 | 16 | 00:00 | UTC | 1.8 | NaN | 3.4 |
163 | 2017 | 6 | 13 | 00:00 | UTC | 0.6 | NaN | 12.6 |
308 | 2017 | 11 | 5 | 00:00 | UTC | 0.2 | NaN | 8.4 |
309 | 2017 | 11 | 6 | 00:00 | UTC | 2.0 | NaN | 7.5 |
313 | 2017 | 11 | 10 | 00:00 | UTC | 3.6 | NaN | 7.2 |
321 | 2017 | 11 | 18 | 00:00 | UTC | 11.3 | NaN | 5.9 |
328 | 2017 | 11 | 25 | 00:00 | UTC | 8.5 | NaN | 4.2 |
The notnull
method works conversively to the isnull
method.
The dropna
method of a DataFrame drops columns or rows that contain missing values from the DataFrame, depending on the axis
parameter.
[32]:
wh.dropna().shape # Default axis is 0
[32]:
(358, 8)
[33]:
wh.dropna(axis=1).shape # Drops the columns containing missing values
[33]:
(365, 7)
The how
and thresh
parameters of the dropna
method allow one to specify how many values need to be missing in order for the row/column to be dropped.
The fillna
method allows to fill the missing values with some constant or interpolated values. The method
parameter can be:
None
: use the given positional parameter as the constant to fill missing values withffill
: use the previous value to fill the current valuebfill
: use the next value to fill the current value
For example, for the weather data we could use forward fill
[34]:
wh = wh.fillna(method='ffill')
wh[wh.isnull().any(axis=1)]
[34]:
Year | m | d | Time | Time zone | Precipitation amount (mm) | Snow depth (cm) | Air temperature (degC) |
---|
The interpolate
method, which we will not cover here, offers more elaborate ways to interpolate the missing values from their neighbouring non-missing values.
Write function cyclists
that does the following.
Load the Helsinki bicycle data set from the src
folder (https://hri.fi/data/dataset//helsingin-pyorailijamaarat). The dataset contains the number of cyclists passing by measuring points per hour. The data is gathered over about four years, and there are 20 measuring points around Helsinki. The dataset contains some empty rows at the end. Get rid of these. Also, get rid of columns that contain only missing values. Return the cleaned dataset.
Make function missing_value_types
that returns the following DataFrame. Use the State
column as the (row) index. The value types for the two other columns should be float
and object
, respectively. Replace the dashes with the appropriate missing value symbols.
State |
Year of independence |
President |
---|---|---|
United Kingdom |
||
Finland |
1917 |
Niinistö |
USA |
1776 |
Trump |
Sweden |
1523 |
|
Germany |
Steinmeier |
|
Russia |
1992 |
Putin |
Write function special_missing_values
that does the following.
Read the data set of the top forty singles from the beginning of the year 1964 from the src
folder. Return the rows whose singles’ position dropped compared to last week’s position (column LW=Last Week).
To do this you first have to convert the special values “New” and “Re” (Re-entry) to missing values (None
).
This exercise can give two points at maximum!
Write function last_week
that reads the top40 data set mentioned in the above exercise. The function should then try to reconstruct the top40 list of the previous week based on that week’s list. Try to do this as well as possible. You can fill the values that are impossible to reconstruct by missing value symbols. Your solution should work for a top40 list of any week. So don’t rely on specific features of this top40 list. The column WoC
means “Weeks on Chart”, that is, on how many weeks
this song has been on the top 40 list.
Hint. First create the last week’s top40 list of those songs that are also on this week’s list. Then add those entries that were not on this week’s list. Finally sort by position.
Hint 2. The where
method of Series and DataFrame can be useful. It can also be nested.
Hint 3. Like in NumPy, you can use with Pandas the bitwise operators &
, |
, and ~
. Remember that he bitwise operators have higher precedence than the comparison operations, so you may have to use parentheses around comparisons, if you combined result of comparisons with bitwise operators.
You get a second point, if you get the columns LW
and Peak Pos
correct.
Converting columns from one type to another¶
There are several ways of converting a column to another type. For converting single columns (a Series) one can use the pd.to_numeric
function or the map
method. For converting several columns in one go one can use the astype
method. We will give a few examples of use of these methods/functions. For more details, look from the Pandas documentation.
[35]:
pd.Series(["1","2"]).map(int) # str -> int
[35]:
0 1
1 2
dtype: int64
[36]:
pd.Series([1,2]).map(str) # int -> str
[36]:
0 1
1 2
dtype: object
[37]:
pd.to_numeric(pd.Series([1,1.0]), downcast="integer") # object -> int
[37]:
0 1
1 1
dtype: int8
[38]:
pd.to_numeric(pd.Series([1,"a"]), errors="coerce") # conversion error produces Nan
[38]:
0 1.0
1 NaN
dtype: float64
[39]:
pd.Series([1,2]).astype(str) # works for a single series
[39]:
0 1
1 2
dtype: object
[40]:
df = pd.DataFrame({"a": [1,2,3], "b" : [4,5,6], "c" : [7,8,9]})
print(df.dtypes)
print(df)
a int64
b int64
c int64
dtype: object
a b c
0 1 4 7
1 2 5 8
2 3 6 9
[41]:
df.astype(float) # Convert all columns
[41]:
a | b | c | |
---|---|---|---|
0 | 1.0 | 4.0 | 7.0 |
1 | 2.0 | 5.0 | 8.0 |
2 | 3.0 | 6.0 | 9.0 |
[42]:
df2 = df.astype({"b" : float, "c" : str}) # different types for columns
print(df2.dtypes)
print(df2)
a int64
b float64
c object
dtype: object
a b c
0 1 4.0 7
1 2 5.0 8
2 3 6.0 9
String processing¶
If the elements in a column are strings, then the vectorized versions of Python’s string processing methods are available. These are accessed through the str
attribute of a Series or a DataFrame. For example, to capitalize all the strings of a Series, we can use the str.capitalize
method:
[43]:
names = pd.Series(["donald", "theresa", "angela", "vladimir"])
names.str.capitalize()
[43]:
0 Donald
1 Theresa
2 Angela
3 Vladimir
dtype: object
One can find all the available methods by pressing the tab key after the text names.str.
in a Python prompt. Try it in below cell!
[44]:
#names.str.
We can split a column or Series into several columns using the split
method. For example:
[45]:
full_names = pd.Series(["Donald Trump", "Theresa May", "Angela Merkel", "Vladimir Putin"])
full_names.str.split()
[45]:
0 [Donald, Trump]
1 [Theresa, May]
2 [Angela, Merkel]
3 [Vladimir, Putin]
dtype: object
This is not exactly what we wanted: now each element is a list. We need to use the expand
parameter to split into columns:
[46]:
full_names.str.split(expand=True)
[46]:
0 | 1 | |
---|---|---|
0 | Donald | Trump |
1 | Theresa | May |
2 | Angela | Merkel |
3 | Vladimir | Putin |
Read again the bicycle data set from src
folder, and clean it as in the earlier exercise. Then split the Päivämäärä
column into a DataFrame with five columns with column names Weekday
, Day
, Month
, Year
, and Hour
. Note that you also need to to do some conversions. To get Hours, drop the colon and minutes. Convert field Weekday
according the following rule:
ma -> Mon
ti -> Tue
ke -> Wed
to -> Thu
pe -> Fri
la -> Sat
su -> Sun
Convert the Month
column according to the following mapping
tammi 1
helmi 2
maalis 3
huhti 4
touko 5
kesä 6
heinä 7
elo 8
syys 9
loka 10
marras 11
joulu 12
Create function split_date
that does the above and returns a DataFrame with five columns. You may want to use the map
method of Series objects.
So the first element in the Päivämäärä
column of the original data set should be converted from ke 1 tammi 2014 00:00
to Wed 1 1 2014 0
. Test your solution from the main
function.
This exercise can give two points at maximum!
The entries in the following table of US presidents are not uniformly formatted. Make function cleaning_data
that reads the table from the tsv file src/presidents.tsv
and returns the cleaned version of it. Note, you must do the edits programmatically using the string edit methods, not by creating a new DataFrame by hand. The columns should have dtype
s object
, integer
, float
, integer
, object
. The where
method of DataFrames can be helpful, likewise the string
methods of Series objects. You get an additional point, if you manage to get the columns President and Vice-president right!
President |
Start |
Last |
Seasons |
Vice-president |
---|---|---|---|---|
donald trump |
2017 Jan |
1 |
Mike pence |
|
barack obama |
2009 |
2017 |
2 |
joe Biden |
bush, george |
2001 |
2009 |
2 |
Cheney, dick |
Clinton, Bill |
1993 |
2001 |
two |
gore, Al |
Additional information¶
We covered subsetting of DataFrames with the indexers []
, .loc[]
, and .iloc[]
quite concisely. For a more verbose explanation, look at the tutorials at Dunder Data. Especially, the problems with chained indexing operators (like df["a"][1]
) are explained well there (tutorial 4), which we did not cover at all. As a rule of thumb: one should avoid chained indexing combined with assignment! See Pandas
documentation.
Summary (week 4)¶
You can create DataFrames in several ways:
By reading from a csv file
Out out two dimensional NumPy array
Out of rows
Out of columns
You know how to access rows, columns and individual elements of DataFrames
You can use the
describe
method to get a quick overview of a DataFrameYou know how missing values are represented in Series and DataFrames, and you know how to manipulate them
There are similarities between Python’s string methods and the vectorized forms of string operations in Series and DataFrames
You can do complicated text processing with the
str.replace
method combined with regular expressionsThe powerful
where
method is the vectorized form of Python’sif-else
constructWe remember that with NumPy arrays we preferred vectorized operations instead of, for instance,
for
loops. Same goes with Pandas. It may first feel that things are easier to achieve with loops, but after a while vectorized operations will feel natural.