Keeping subset of row labels in pandas DataFrame based on second index - python
Given a DataFrame with a hierarchical index containing three levels (experiment, trial, slot) and a second DataFrame with a hierarchical index containing two levels (experiment, trial), how do I drop all the rows in the first DataFrame whose (experiment, trial) are not contained in the second dataframe?
Example data:
from io import StringIO
import pandas as pd
df1_data = StringIO(u',experiment,trial,slot,token\n0,btn144a10_p_RDT,0,0,4.0\n1,btn144a10_p_RDT,0,1,14.0\n2,btn144a10_p_RDT,1,0,12.0\n3,btn144a10_p_RDT,1,1,14.0\n4,btn145a07_p_RDT,0,0,6.0\n5,btn145a07_p_RDT,0,1,19.0\n6,btn145a07_p_RDT,1,0,17.0\n7,btn145a07_p_RDT,1,1,13.0\n8,chn004b06_p_RDT,0,0,6.0\n9,chn004b06_p_RDT,0,1,8.0\n10,chn004b06_p_RDT,1,0,2.0\n11,chn004b06_p_RDT,1,1,5.0\n12,chn008a06_p_RDT,0,0,12.0\n13,chn008a06_p_RDT,0,1,14.0\n14,chn008a06_p_RDT,1,0,6.0\n15,chn008a06_p_RDT,1,1,4.0\n16,chn008b06_p_RDT,0,0,3.0\n17,chn008b06_p_RDT,0,1,13.0\n18,chn008b06_p_RDT,1,0,12.0\n19,chn008b06_p_RDT,1,1,19.0\n20,chn008c04_p_RDT,0,0,17.0\n21,chn008c04_p_RDT,0,1,2.0\n22,chn008c04_p_RDT,1,0,1.0\n23,chn008c04_p_RDT,1,1,6.0\n')
df1 = pd.DataFrame.from_csv(df1_data).set_index(['experiment', 'trial', 'slot'])
df2_data = StringIO(u',experiment,trial,target\n0,btn145a07_p_RDT,1,13\n1,chn004b06_p_RDT,1,9\n2,chn008a06_p_RDT,0,15\n3,chn008a06_p_RDT,1,15\n4,chn008b06_p_RDT,1,1\n5,chn008c04_p_RDT,1,12\n')
df2 = pd.DataFrame.from_csv(df2_data).set_index(['experiment', 'trial'])
The first dataframe looks like:
token
experiment trial slot
btn144a10_p_RDT 0 0 4
1 14
1 0 12
1 14
btn145a07_p_RDT 0 0 6
1 19
1 0 17
1 13
chn004b06_p_RDT 0 0 6
1 8
1 0 2
1 5
chn008a06_p_RDT 0 0 12
1 14
1 0 6
1 4
chn008b06_p_RDT 0 0 3
1 13
1 0 12
1 19
chn008c04_p_RDT 0 0 17
1 2
1 0 1
1 6
The second dataframe looks like:
target
experiment trial
btn145a07_p_RDT 1 13
chn004b06_p_RDT 1 9
chn008a06_p_RDT 0 15
1 15
chn008b06_p_RDT 1 1
chn008c04_p_RDT 1 12
The result I want:
token
experiment trial slot
btn145a07_p_RDT 1 0 17
1 13
chn004b06_p_RDT 1 0 2
1 5
chn008a06_p_RDT 0 0 12
1 14
1 0 6
1 4
chn008b06_p_RDT 1 0 12
1 19
chn008c04_p_RDT 1 0 1
1 6
One way to do it would by using merge
merged = pd.merge(
df2.reset_index(),
df1.reset_index(),
left_on=['experiment', 'trial'],
right_on=['experiment', 'trial'],
how='left')
You just need to reindex merged to whatever you like (I could not tell exactly from the question).
What should work is
df1.loc[df2.index]
but multi indexing still has some problems. What does work is
df1.reset_index(2).loc[df2.index].set_index('slot', append=True)
which is a bit of a hack around this problem. Note that
df1.loc[df2.index[:1]]
gives garbage while
df.loc[df2.index[0]]
gives what you would expect. So passing multiple values from a m-level index to an n-level index where n > m > 2 doesn't work, though it should.
Related
Count number of consecutive rows that are greater than current row value but less than the value from other column
Say I have the following sample dataframe (there are about 25k rows in the real dataframe) df = pd.DataFrame({'A' : [0,3,2,9,1,0,4,7,3,2], 'B': [9,8,3,5,5,5,5,8,0,4]}) df A B 0 0 9 1 3 8 2 2 3 3 9 5 4 1 5 5 0 5 6 4 5 7 7 8 8 3 0 9 2 4 For the column A I need to know how many next and previous rows are greater than current row value but less than value in column B. So my expected output is : A B next count previous count 0 9 2 0 3 8 0 0 2 3 0 1 9 5 0 0 1 5 0 0 0 5 2 1 4 5 1 0 7 8 0 0 3 0 0 2 2 4 0 0 Explanation : First row is calculated as : since 3 and 2 are greater than 0 but less than corresponding B value 8 and 3 Second row is calculated as : since next value 2 is not greater than 3 Third row is calculated as : since 9 is greater than 2 but not greater than its corresponding B value Similarly, previous count is calculated Note : I know how to solve this problem by looping using list comprehension or using the pandas apply method but still I won't mind a clear and concise apply approach. I was looking for a more pandaic approach. My Solution Here is the apply solution, which I think is inefficient. Also, as people said that there might be no vector solution for the question. So as mentioned, a more efficient apply solution will be accepted for this question. This is what I have tried. This function gets the number of previous/next rows that satisfy the condition. def get_prev_next_count(row): next_nrow = df.loc[row['index']+1:,['A', 'B']] prev_nrow = df.loc[:row['index']-1,['A', 'B']][::-1] if (next_nrow.size == 0): return 0, ((prev_nrow.A > row.A) & (prev_nrow.A < prev_nrow.B)).argmin() if (prev_nrow.size == 0): return ((next_nrow.A > row.A) & (next_nrow.A < next_nrow.B)).argmin(), 0 return (((next_nrow.A > row.A) & (next_nrow.A < next_nrow.B)).argmin(), ((prev_nrow.A > row.A) & (prev_nrow.A < prev_nrow.B)).argmin()) Generating output : df[['next count', 'previous count']] = df.reset_index().apply(get_prev_next_count, axis=1, result_type="expand") Output : This gives us the expected output df A B next count previous count 0 0 9 2 0 1 3 8 0 0 2 2 3 0 1 3 9 5 0 0 4 1 5 0 0 5 0 5 2 1 6 4 5 1 0 7 7 8 0 0 8 3 0 0 2 9 2 4 0 0
I made some optimizations: You don't need to reset_index() you can access the index with .name If you only pass df[['A']] instead of the whole frame, that may help. prev_nrow.empty is the same as (prev_nrow.size == 0) Applied different logic to get the desired value via first_false, this speeds things up significantly. def first_false(val1, val2, A): i = 0 for x, y in zip(val1, val2): if A < x < y: i += 1 else: break return i def get_prev_next_count(row): A = row['A'] next_nrow = df.loc[row.name+1:,['A', 'B']] prev_nrow = df2.loc[row.name-1:,['A', 'B']] if next_nrow.empty: return 0, first_false(prev_nrow.A, prev_nrow.B, A) if prev_nrow.empty: return first_false(next_nrow.A, next_nrow.B, A), 0 return (first_false(next_nrow.A, next_nrow.B, A), first_false(prev_nrow.A, prev_nrow.B, A)) df2 = df[::-1].copy() # Shave a tiny bit of time by only reversing it once~ df[['next count', 'previous count']] = df[['A']].apply(get_prev_next_count, axis=1, result_type='expand') print(df) Output: A B next count previous count 0 0 9 2 0 1 3 8 0 0 2 2 3 0 1 3 9 5 0 0 4 1 5 0 0 5 0 5 2 1 6 4 5 1 0 7 7 8 0 0 8 3 0 0 2 9 2 4 0 0 Timing Expanding the data: df = pd.concat([df]*(10000//4), ignore_index=True) # df.shape == (25000, 2) Original Method: Gave up at 15 minutes. New Method: 1m 20sec Throw pandarallel at it: from pandarallel import pandarallel pandarallel.initialize() df[['A']].parallel_apply(get_prev_next_count, axis=1, result_type='expand') 26sec
counting consequtive elements in a dataframe and storing them in a new column
so i have this code: import pandas as pd id_1=[0,0,0,0,0,0,2,0,4,5,6,7,1,0,5,3] exp_1=[1,2,3,4,5,6,1,7,1,1,1,1,1,8,2,1] df = pd.DataFrame(list(zip(id_1,exp_1)), columns =['Patch', 'Exploit']) df = ( df.groupby((df.Patch != df.Patch.shift(1)).cumsum()) .agg({"Patch": ("first", "count")}) .reset_index(drop=True) ) print(df) the output is: Patch first count 0 0 6 1 2 1 2 0 1 3 4 1 4 5 1 5 6 1 6 7 1 7 1 1 8 0 1 9 5 1 10 3 1 I wanted to create a data frame with a new column called count where I can store the consecutive appearance of the patch (id_1). However, the above code creates a dictionary of the patch and I don't know how to individually manipulate only the values stored in the column called count. suppose I want to remove all the 0 from id_1 and then count the consecutive appearance. or I have to find the average of the count column only then?
If you want to remove all 0 from column Patch, then you can filter the dataframe just before .groupby. For example: df = ( df[df.Patch != 0] .groupby((df.Patch != df.Patch.shift(1)).cumsum()) .agg({"Patch": ("first", "count")}) .reset_index(drop=True) ) print(df) Prints: Patch first count 0 2 1 1 4 1 2 5 1 3 6 1 4 7 1 5 1 1 6 5 1 7 3 1
pivot table to expanded dataframe in Python/Pandas
I want to build on a previous question of mine. Let's look at some Python code. import numpy as np import pandas as pd mat = np.array([[1,2,3],[4,5,6]]) df_mat = pd.DataFrame(mat) df_mat_tidy = (df_mat.stack() .rename_axis(index = ['V1','V2']) .rename('value') .reset_index() .reindex(columns = ['value','V1','V2'])) df_mat_tidy This takes me from a pivot table (mat) to a "tidy" (in the Tidyverse sense) version of the data that gives one variable as the column from which the number came, one variable as the row from which the number came, and one variable as the number in the pivot table at the row-column position. Now I want to expand on that to get the row-column pair repeated the number of times the pivot table specifies. In other words, if position 1,1 has value 3 and position 2,1 has value 4, I want the data frame to go col row 1 1 1 1 1 1 1 2 1 2 1 2 1 2 instead of col row value 1 1 3 1 2 4 I think I know how to loop over the rows of the second example and produce that, but I want something faster. Is there a way to "melt" the pivot table the way that I am describing?
Have a look at the parts of pandas' docs entitled "Reshaping and pivot tables". Both .pivot(), .pivot_table() and .melt() are all existing functions. It looks like you are reinventing some wheels.
You could just rebuild a DataFrame from a comprehension: pd.DataFrame([i for j in [[[rec['V1'], rec['V2']]] * rec['value'] for rec in df_mat_tidy.to_dict(orient='records')] for i in j], columns=['col', 'row']) It gives as expected: col row 0 0 0 1 0 1 2 0 1 3 0 2 4 0 2 5 0 2 6 1 0 7 1 0 8 1 0 9 1 0 10 1 1 11 1 1 12 1 1 13 1 1 14 1 1 15 1 2 16 1 2 17 1 2 18 1 2 19 1 2 20 1 2
Given a value or constant, I need to only output relevant rows on Pandas
This is how my data looks like: Day Price A Price B Price C 1 0 0 0 2 0 0 0 3 0 0 0 4 0 0 0 5 64503 43692 79982 6 86664 69990 53468 7 77924 62998 68911 8 66600 68830 94396 9 82664 89972 49614 10 59741 48904 49528 11 34030 98074 72993 12 74400 85547 37715 13 51031 50031 85345 14 74700 59932 73935 15 62290 98130 88818 I have a small python script that outputs a sum for each column. I need to input an n value (for number of days) and the summing will run and output the values. However, for example, given n=5 (for days), I want to output only Price A/B/C rows starting from the next day (which is day 6). Hence, the row for Day 5 should be '0'. How can I produce this logic on Pandas ? The idea I have is to use the n input value to then, truncate values on the rows corresponding to that particular (n day value). But how can I do this on code ? if dataframe['Day'] == n: dataframe['Price A'] == 0 & dataframe['Price B'] == 0 & dataframe['Price C'] == 0
You can filter rows by condition and set all columns without first by iloc[mask, 1:], for next row add Series.shift: n = 5 df.iloc[(df['Day'].shift() <= n).values, 1:] = 0 print (df) Day Price A Price B Price C 0 1 0 0 0 1 2 0 0 0 2 3 0 0 0 3 4 0 0 0 4 5 0 0 0 5 6 0 0 0 6 7 77924 62998 68911 7 8 66600 68830 94396 8 9 82664 89972 49614 9 10 59741 48904 49528 10 11 34030 98074 72993 11 12 74400 85547 37715 12 13 51031 50031 85345 13 14 74700 59932 73935 14 15 62290 98130 88818
Pseudo Code Make sure to sort by day shift columns 'A', 'B' and 'C' by n and fill in with 0 Sum accordingly All of that can be done on one line as well
It is simply dataframe.iloc[:n+1] = 0 This sets the values of all columns for the first n days to 0 # Sample output dataframe a b 0 1 2 1 2 3 2 3 4 3 4 2 4 5 3 n = 1 dataframe.iloc[:n+1] = 0 dataframe a b 0 0 0 1 0 0 2 3 4 3 4 2 4 5 3 This truncates all for all the previous days. If you want to truncate only for the nth day. dataframe.iloc[n] = 0
Python pandas join on with overwrite
I realize this question is similar to join or merge with overwrite in pandas, but the accepted answer does not work for me since I want to use the on='keys' from df.join(). I have a DataFrame df which looks like this: keys values 0 0 0.088344 1 0 0.088344 2 0 0.088344 3 0 0.088344 4 0 0.088344 5 1 0.560857 6 1 0.560857 7 1 0.560857 8 2 0.978736 9 2 0.978736 10 2 0.978736 11 2 0.978736 12 2 0.978736 13 2 0.978736 14 2 0.978736 Then I have a Series s (which is a result from some df.groupy.apply()) with the same keys: keys 0 0.183328 1 0.239322 2 0.574962 Name: new_values, dtype: float64 Basically I want to replace the 'values' in the df with the values in the Series, by keys so every keys block gets the same new value. Currently, I do it as follows: df = df.join(s, on='keys') df['values'] = df['new_values'] df = df.drop('new_values', axis=1) The obtained (and desired) result is then: keys values 0 0 0.183328 1 0 0.183328 2 0 0.183328 3 0 0.183328 4 0 0.183328 5 1 0.239322 6 1 0.239322 7 1 0.239322 8 2 0.574962 9 2 0.574962 10 2 0.574962 11 2 0.574962 12 2 0.574962 13 2 0.574962 14 2 0.574962 That is, I add it as a new column and by using on='keys' it gets the corrects shape. Then I assign values to be new_values and remove the new_values column. This of course works perfectly, the only problem being that I find it extremely ugly. Is there a better way to do this?
You could try something like: df = df[df.columns[df.columns!='values']].join(s, on='keys') Make sure s is named 'values' instead of 'new_values'. To my knowledge, pandas doesn't have the ability to join with "force overwrite" or "overwrite with warning".