Extracting Max value along each row from strings in column - python
I've a column of strings in a DataFrame which contains comma-separated numbers. I need to extract the maximum value along each row from the strings. The maximum value returned should be the max till the 13th index from the beginning.
I've tried splitting the sting using ',' as a separator to convert it into a list with expand option enabled. Then I'm using the assign method of Pandas to find the max value along the vertical axis.
sample_dt1 = sample_dt['pyt_hist'].str.split(',', expand=True).astype(float)
sample_dt = sample_dt.assign(max_value=sample_dt1.max(axis=1))
Sample Data:
index pyt_hist
0 0,0,0,0,0,0,0,0,0,0,0
1 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
2 0,0,0,360,420,392,361,330,300,269,239,208,177
3 0,0,0,0,0,0,0,0,0,0,0,0,0,0,23,0,23,0,0,56,0
Expected Result:
index pyt_hist max_value
0 0,0,0,0,0,0,0,0,0,0,0 0
1 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 0
2 0,0,0,360,420,392,361,330,300,269,239,208,177 420
3 0,0,0,0,0,0,0,0,0,0,0,0,0,0,23,0,23,0,0,56,0 0
Results obtained using my code:
index pyt_hist max_value
0 0,0,0,0,0,0,0,0,0,0,0 0.0
1 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0 0.0
2 0,0,0,360,420,392,361,330,300,269,239,208,177 420.0
3 0,0,0,0,0,0,0,0,0,0,0,0,0,0,23,0,23,0,0,56,0 56.0
You are very close, sample_dt1.iloc[:,:13] gives you the first 13 columns of sample_dt1. So you can do:
sample_dt = sample_dt.assign(max_value=sample_dt1.iloc[:,:13].max(axis=1))
df.pyt_hist.str.split(',').apply(lambda x: max([int(i) for i in x[:13] if i]))
Output
0 0
1 0
2 420
3 0
Name: pyt_hist, dtype: int64
Related
Manipulate string values in pandas
I have a pandas dataframe with different formats for one column like this Name Values First 5-9 Second 7 Third - Fourth 12-16 I need to iterate over all Values column, and if the format is like the first row 5-9 or like fourth row 12-16 replace it with the mean between the 2 numbers in string. For first row replace 5-9 to 7, or for fourth row replace 12-16 to 14. And if the format is like third row - replace it to 0 I have tried if df["Value"].str.len() > 1: df["Value"] = df["Value"].str.split('-') df["Value"] = (df["Value"][0] + df["Value"][1]) / 2 elif df["Value"].str.len() == 1: df["Value"] = df["Value"].str.replace('-', 0) Expected output Name Values First 7 Second 7 Third 0 Fourth 14
Let us split and expand the column then cast values to float and calculate mean along column axis: s = df['Values'].str.split('-', expand=True) df['Values'] = s[s != ''].astype(float).mean(1).fillna(0) Name Values 0 First 7.0 1 Second 7.0 2 Third 0.0 3 Fourth 14.0
You can use str.replace with customized replacement function mint = lambda s: int(s or 0) repl = lambda m: str(sum(map(mint, map(m.group, [1,2])))/2) df['Values'] = df['Values'].str.replace('(\d*)-(\d*)', repl, regex=True) print(df) Name Values 0 First 7.0 1 Second 7 2 Third 0.0 3 Fourth 14.0
python / pandas: How to count each cluster of unevenly distributed distinct values in each row
I am transitioning from excel to python and finding the process a little daunting. I have a pandas dataframe and cannot find how to count the total of each cluster of '1's' per row and group by each ID (example data below). ID 20-21 19-20 18-19 17-18 16-17 15-16 14-15 13-14 12-13 11-12 0 335344 0 0 1 1 1 0 0 0 0 0 1 358213 1 1 0 1 1 1 1 0 1 0 2 358249 0 0 0 0 0 0 0 0 0 0 3 365663 0 0 0 1 1 1 1 1 0 0 The result of the above in the format ID LastColumn Heading a '1' occurs: count of '1's' in that cluster would be: 335344 16-17: 3 358213 19-20: 2 14-15: 4 12-13: 1 365663 13-14: 5 There are more than 11,000 rows of data I would like to output the result to a txt file. I have been unable to find any examples of how the same values are clustered by row, with a count for each cluster, but I am probably not using the correct python terminology. I would be grateful if someone could point me in the right direction. Thanks in advance.
First step is use DataFrame.set_index with DataFrame.stack for reshape. Then create consecutive groups by compare for not equal Series.shifted values with cumulative sum by Series.cumsum to new column g. Then filter rows with only 1 and aggregate by named aggregation by GroupBy.agg with GroupBy.last and GroupBy.size: df = df.set_index('ID').stack().reset_index(name='value') df['g'] = df['value'].ne(df['value'].shift()).cumsum() df1 = (df[df['value'].eq(1)].groupby(['ID', 'g']) .agg(a=('level_1','last'), b=('level_1','size')) .reset_index(level=1, drop=True) .reset_index()) print (df1) ID a b 0 335344 16-17 3 1 358213 19-20 2 2 358213 14-15 4 3 358213 12-13 1 4 365663 13-14 5 Last for write to txt use DataFrame.to_csv: df1.to_csv('file.txt', index=False) If need your custom format in text file use: with open("file.txt","w") as f: for i, g in df1.groupby('ID'): f.write(f"{i}\n") for a, b in g[['a','b']].to_numpy(): f.write(f"\t{a}: {b}\n")
You just need to use the sum method and then specify which axis you would like to sum on. To get the sum of each row, create a new series equal to the sum of the row. # create new series equal to sum of values in the index row df['sum'] = df.sum(axis=1) # specifies index (row) axis The best method for getting the sum of each column is dependent on how you want to use that information but in general the core is just to use the sum method on the series and assign it to a variable. # sum a column and assign result to variable foo = df['20-21'].sum() # default axis=0 bar = df['16-17'].sum() # default axis=0 print(foo) # returns 1 print(bar) # returns 3 You can get the sum of each column using a for loop and add them to a dictionary. Here is a quick function I put together that should get the sum of each column and return a dictionary of the results so you know which total belongs to which column. The two inputs are 1) the dataframe 2) a list of any column names you would like to ignore def get_df_col_sum(frame: pd.DataFrame, ignore: list) -> dict: """Get the sum of each column in a dataframe in a dictionary""" # get list of headers in dataframe dfcols = frame.columns.tolist() # create a blank dictionary to store results dfsums = {} # loop through each column and append sum to list for dfcol in dfcols: if dfcol not in ignore: dfsums.update({dfcol: frame[dfcol].sum()}) return dfsums I then ran the following code # read excel to dataframe df = pd.read_excel(test_file) # ignore the ID column ignore_list = ['ID'] # get sum for each column res_dict = get_df_col_sum(df, ignore_list) print(res_dict) and got the following result. {'20-21': 1, '19-20': 1, '18-19': 1, '17-18': 3, '16-17': 3, '15-16': 2, '14-15': 2, '13-14': 1, '12-13': 1, '11-12': 0} Sources: Sum by row, Pandas Sum, Add pairs to dictionary
Groupby when given the start positional index of each group
I have one series of values that I would like to group, and another series containing the starting positional index of each group after the first (the first group is understood to begin at positional index 0). The series of values can have an arbitrary index. Is there a way to use this to produce a groupby-aggregate? Ideally empty groups would be preserved. Example: values = pd.Series(np.arange(10, 20), index=np.arange(110, 120)) group_indices = pd.Series([3, 3, 8]) Now, values.groupby(group_indices) should be grouped so that the first group is values.iloc[:3], the second is values.iloc[3:3] (an empty group), the third is values.iloc[3:8], and the fourth is values.iloc[8:], and values.groupby(group_indices).mean() would be pd.Series([11.0, NaN, 15.0, 18.5]).
Here is a easy way values.groupby(values.index.isin(group_indices).cumsum()).mean() Out[454]: 1 11.0 2 15.0 3 18.5 dtype: float64
Straightforwardly with numpy.split routine: In [1286]: values = pd.Series(np.arange(10, 20)) In [1287]: group_indices = pd.Series([0, 3, 8]) In [1288]: pd.Series([s.mean() for s in np.split(values, group_indices) if s.size]) Out[1288]: 0 11.0 1 15.0 2 18.5 dtype: float64 To account "empty" group - just remove if s.size check: In [1304]: group_indices = pd.Series([3, 3, 8]) In [1305]: pd.Series([s.mean() for s in np.split(values, group_indices)]) Out[1305]: 0 11.0 1 NaN 2 15.0 3 18.5 dtype: float64
Given your update, here's an odd way to do this with pd.merge_asof. Some care needs to be taken to deal with the first group that's from 0 to your first index in the Series. import pandas as pd import numpy as np (pd.merge_asof(values.to_frame('val'), values.iloc[np.r_[group_indices]].reset_index().reset_index().drop(columns=0), left_index=True, right_on='index', direction='backward') .fillna({'level_0': -1}) # Because your first group is 0: first index .groupby('level_0').val.mean() .reindex([-1]+[*range(len(group_indices))]) # Get 0 size groups in output ) level_0 -1 11.0 0 NaN 1 15.0 2 18.5 Name: val, dtype: float64
Let's change the group_indicies a bit, so that the group names (1,2,3) are visible, group_indices = pd.Series([1,2,3],index=[0, 3, 8]) then values.groupby(group_indices.reindex(values.index,method='ffill')).mean() would give you what you want. Note that group_indices.reindex(values.index,method='ffill') gives you 0 1 1 1 2 1 3 2 4 2 5 2 6 2 7 2 8 3 9 3 which assigns each row of values with a group number.
My solution involves keeping the inputs as they are and doing some ugly adjustments: pd.DataFrame(values).assign(group=pd.cut(pd.DataFrame(values).index, [-1,2,7,np.inf], labels=[0,1,2])).groupby('group').mean() Output 0 group 0 11.0 1 15.0 2 18.5
Thanks to all the answers, especially WeNYoBen's. The following will produce the correct groups and skip over empty groups. # First, add the final index to `group_indices` so that # we have a series of right endpoints, or interval upper bounds upper_bounds = group_indices.append(pd.Series(values.shape[0]), ignore_index=True) # Compute indices of nonempty groups lower_bounds = upper_bounds.shift(fill_value=0) nonempty_group_idxs = upper_bounds != lower_bounds # Get means indexed from 0 to n_nonempty_groups-1 means = values.groupby(pd.RangeIndex(values.shape[0]).isin(upper_bounds).cumsum()).mean() # Reassign index for the correct groups means.index = nonempty_group_idxs.index[nonempty_group_idxs] This will have a noncontinuous index, with skipped elements corresponding to empty groups in the original groupby. If you want to place NaN in those spots, you can do means = means.reindex(index=pd.RangeIndex(group_indices.shape[0]))
In a DataFrame, how could we get a list of indexes with 0's in specific columns?
We have a large dataset that needs to be modified based on specific criteria. Here is a sample of the data: Input BL.DB BL.KB MI.RO MI.RA MI.XZ MAY.BE 0 0 1 1 1 0 1 1 0 0 1 0 0 1 SampleData1 = pd.DataFrame([[0,1,1,1,1],[0,0,1,0,0]],columns = ['BL.DB', 'BL.KB', 'MI.RO', 'MI.RA', 'MI.XZ']) The fields of this data are all formatted 'family.member', and a family may have any number of members. We need to remove all rows of the dataframe which have all 0's for any family. Simply put, we want to only keep rows of the data that contain at least one member of every family. We have no reproducible code for this problem because we are unsure of where to start. We thought about using iterrows() but the documentation says: #You should **never modify** something you are iterating over. #This is not guaranteed to work in all cases. Depending on the #data types, the iterator returns a copy and not a view, and writing #to it will have no effect. Other questions on S.O. do not quite solve our problem. Here is what we want the SampleData to look like after we run it: Expected output BL.DB BL.KB MI.RO MI.RA MI.XZ MAY.BE 0 0 1 1 1 0 1 SampleData1 = pd.DataFrame([[0,1,1,1,0]],columns = ['BL.DB', 'BL.KB', 'MI.RO', 'MI.RA', 'MI.XZ']) Also, could you please explain why we should not modify a data we iterate over when we do that all the time with for loops, and what is the correct way to modify DataFrame's too, please? Thanks for the help in advance!
Start from copying df and reformatting its columns into a MultiIndex: df2 = df.copy() df2.columns = df.columns.str.split(r'\.', expand=True) The result is: BL MI DB KB RO RA XZ 0 0 1 1 1 0 1 0 0 1 0 0 To generate "family totals", i.e. sums of elements in rows over the top (0) level of column index, run: df2.groupby(level=[0], axis=1).sum() The result is: BL MI 0 1 2 1 0 1 But actually we want to count zeroes in each row of the above table, so extend the above code to: (df2.groupby(level=[0], axis=1).sum() == 0).astype(int).sum(axis=1) The result is: 0 0 1 1 dtype: int64 meaning: row with index 0 has no "family zeroes", row with index 1 has one such zero (for one family). And to print what we are looking for, run: df[(df2.groupby(level=[0], axis=1).sum() == 0)\ .astype(int).sum(axis=1) == 0] i.e. print rows from df, with indices for which the count of "family zeroes" in df2 is zero.
It's possible to group along axis=1. For each row, check that all families (grouped on the column name before '.') have at least one 1, then slice by this Boolean Series to retain these rows. m = df.groupby(df.columns.str.split('.').str[0], axis=1).any(1).all(1) df[m] # BL.DB BL.KB MI.RO MI.RA MI.XZ MAY.BE #0 0 1 1 1 0 1 As an illustration, here's what grouping along axis=1 looks like; it partitions the DataFrame by columns. for idx, gp in df.groupby(df.columns.str.split('.').str[0], axis=1): print(idx, gp, '\n') #BL BL.DB BL.KB #0 0 1 #1 0 0 #MAY MAY.BE #0 1 #1 1 #MI MI.RO MI.RA MI.XZ #0 1 1 0 #1 1 0 0 Now it's rather straightforward to find the rows where all of these groups have any single non-zero column, by using those with axis=1.
You basically want to group on families and retain rows where there is one or more member for all families in the row. One way to do this is to transpose the original dataframe and then split the index on the period, taking the first element which is the family identifier. The columns are the index values in the original dataframe. We can then group on the families (level=0) and sum the number of members in each for every record (df2.groupby(level=0).sum()). No we retain the index values with more than one member in each family (.gt(0).all()). We create a mask using these values, and apply it to a boolean index on the original dataframe to get the relevant rows. df2 = SampleData1.T df2.index = [idx.split('.')[0] for idx in df2.index] # >>> df2 # 0 1 # BL 0 0 # BL 1 0 # MI 1 1 # MI 1 0 # MI 0 0 # >>> df2.groupby(level=0).sum() # 0 1 # BL 1 0 # MI 2 1 mask = df2.groupby(level=0).sum().gt(0).all() >>> SampleData1[mask] BL.DB BL.KB MI.RO MI.RA MI.XZ 0 0 1 1 1 0
How to get columns index which meet some condition in pandas?
I have the following: x = pd.DataFrame({'a':[1,5,5], 'b':[7,0,7]}) And for every row, i want to get the index of the first column that met the condition that it's value is greater than some value, let's say greater than 4. In this example, the answer is 1, (correspond to the index of the value 7 in the first row) and 0 (correspond to the index of the value 5 in the second row), and 1(correspond to the index of the value 5 in the third row). Which means the answer is [1,0,0]. I tried it with apply method: def get_values_from_row(row, th=0.9): """Get a list of column names that meet some condition that their values are larger than a threshold. Args: row(pd.DataFrame): a row. th(float): the threshold. Returns: string. contains the columns that it's value met the condition. """ return row[row > th].index.tolist()[0] It works, but i have a large data set, and it's quite slow. What is a better alternative.
I think you need first_valid_index with get_loc: print (x[x > 4]) a b 0 NaN 7.0 1 5.0 NaN 2 7.0 5.0 print (x[x > 4].apply(lambda x: x.index.get_loc(x.first_valid_index()), axis=1)) 0 1 1 0 2 0 dtype: int64