Extract all (or replace) from ndjson not effective - python
I am reading a file with one JSON object per line (ndjson)
dfjson = pd.read_json(path_or_buf=JsonFicMain,orient='records',lines=True)
Here is an example of 2 lines of the content of the dataframe (after dropping columns)
nomCommune codeCommune numeroComplet nomVoie codePostal meilleurePosition codesParcelles
0 Ablon-sur-Seine 94001 21 Rue Robert Schumann 94480 {'type': 'parcelle', 'geometry': {'type': 'Point', 'coordinates': [2.411247955172414, 48.726054248275865]}} [94001000AG0013]
1 Ablon-sur-Seine 94001 13 Rue Robert Schumann 94480 {'type': 'parcelle', 'geometry': {'type': 'Point', 'coordinates': [2.412065866666666, 48.72614911111111]}} [94001000AG0020]
It contents million of rows, I want to extract one geo coordinate, between square brackets, in a specific colum (named meilleurePosition). The expected output is
[2.411247955172414, 48.726054248275865]
I tried to either extract the coordinate or replace all other unwanted characters
Using extractall, or extract does not match
test=dfjson['meilleurePosition'].str.extract(pat='(\d+\.\d+)')
test2=dfjson['meilleurePosition'].str.extractall(pat='(\d+\.\d+)')
Empty DataFrame
Columns: [0]
Index: []
Using replace, or str.replace does not work
test3=dfjson["meilleurePosition"].replace(to_replace=r'[^0-9.,:]',value='',regex=True)
0 {'type': 'parcelle', 'geometry': {'type': 'Point', 'coordinates': [2.411247955172414, 48.726054248275865]}}
1 {'type': 'parcelle', 'geometry': {'type': 'Point', 'coordinates': [2.412065866666666, 48.72614911111111]}}
Even none regex type does not work
test4=dfjson['meilleurePosition'].str.replace('type','whatever')
0 NaN
1 NaN
print(test)
I have tried to find why this does not work at all.
Column type is 'object' (which is apparently good as this is a
string)
Using inplace=True without copying the dataframe leads to
similar results
Why can't I manipulate this column, is it because of the special characters in it?
How can get these coordinate in the good format?
OK, after more investigation, the column contains a nested dict, that's why it is not working
This answer helped me a lot
python pandas use map with regular expressions
I did then use the following code to create a new column with the expected coordinates
def extract_coord(meilleurepositiondict):
if isinstance(meilleurepositiondict,dict) :
return meilleurepositiondict['geometry']['coordinates']
else :
return None
dfjson['meilleurePositionclean']=dfjson['meilleurePosition'].apply(lambda x: extract_coord(x))
I found the solution using the code below
dfjson['meilleurePosition']=dfjson['meilleurePosition'].apply(lambda x: extract_coord(x) if x == x else defaultmeilleurepositionvalue)
this was required because of empty rows leading to error (not trapped in function definition).
However ,i am still convinced there is much easy way to assign a dict value of a column to the column itself , still trying...
Related
loop through json file python to get specific key and values
I have a json file built like this: {"type":"FeatureCollection","features":[ {"type":"Feature","id":"010020000A0225","geometry":{"type":"Polygon","coordinates":[[[5.430767,46.0214267],[5.4310805,46.0220116],[5.4311205,46.0220864],[5.4312362,46.0223019],[5.4308994,46.0224141],[5.43087,46.0224242],[5.430774,46.0222401],[5.4304506,46.0223202],[5.4302885,46.021982],[5.4300391,46.0216054],[5.4299637,46.0216342],[5.4300862,46.0218401],[5.4299565,46.021902],[5.4298847,46.0218195],[5.4298545,46.0217829],[5.4297689,46.0216672],[5.4297523,46.0216506],[5.4297379,46.0216389],[5.4296432,46.0215854],[5.429517,46.0214509],[5.4294188,46.0213458],[5.4293757,46.0213128],[5.4291918,46.0211768],[5.4291488,46.0211448],[5.4291083,46.0211214],[5.429024,46.0210828],[5.4292965,46.0208202],[5.4294241,46.0208894],[5.4295183,46.0209623],[5.4295455,46.0209865],[5.429613,46.0210554],[5.4296428,46.0210813],[5.4298751,46.0212862],[5.429988,46.0213782],[5.430014,46.0213973],[5.4300746,46.0214318],[5.430124,46.0214542],[5.4302569,46.0215069],[5.4303111,46.0215192],[5.4303632,46.0215166],[5.4306127,46.0214642],[5.430767,46.0214267]]]},"properties":{"id":"010020000A0225","commune":"01002","prefixe":"000","section":"A","numero":"225","contenance":9440,"arpente":false,"created":"2005-06-03","updated":"2018-09-25"}}, {"type":"Feature","id":"010020000A0346","geometry":{"type":"Polygon","coordinates":[[[5.4241952,46.0255535],[5.4233594,46.0262031],[5.4232624,46.0262774],[5.4226259,46.0267733],[5.4227608,46.0268718],[5.4227712,46.0268789],[5.4226123,46.0269855],[5.422565,46.0270182],[5.4223546,46.027145],[5.4222957,46.0271794],[5.4221794,46.0272376],[5.4221383,46.0272585],[5.4221028,46.027152],[5.4220695,46.0270523],[5.4220378,46.026962],[5.4220467,46.0269265],[5.4220524,46.0268709],[5.4220563,46.0268474],[5.4222945,46.0268985],[5.4224161,46.0267746],[5.4224581,46.0267904],[5.4226286,46.02666],[5.4226811,46.02662],[5.4227313,46.0265803],[5.4227813,46.0265406],[5.4228535,46.0264868],[5.4229063,46.0264482],[5.4229741,46.0264001],[5.4234903,46.0260331],[5.4235492,46.0259893],[5.4235787,46.0259663],[5.423645,46.0259126],[5.4237552,46.0258198],[5.4237839,46.0257951],[5.4238321,46.0257547],[5.4239258,46.0256723],[5.4239632,46.0256394],[5.4241164,46.0255075],[5.4241952,46.0255535]]]},"properties":{"id":"010020000A0346","commune":"01002","prefixe":"000","section":"A","numero":"346","contenance":2800,"arpente":false,"created":"2005-06-03","updated":"2018-09-25"}}, I would like to get for each feature: properties and geometry but I think I loop badly on my json file. here is my code data = pd.read_json(json_file_path) for key, v in data.items(): print(f"{key['features']['geometry']} : {v}", f"{key['features']['properties']} : {v}")
The values you are interested in are located in a list that is itself a value of your main dictionary. If you want to be able to process these values with pandas, it would be better to build your dataframe directly from them: import json import pandas as pd data = json.loads("""{"type":"FeatureCollection","features":[ {"type":"Feature","id":"010020000A0225","geometry":{"type":"Polygon","coordinates":[[[5.430767,46.0214267],[5.4310805,46.0220116],[5.4311205,46.0220864],[5.4312362,46.0223019],[5.4308994,46.0224141],[5.43087,46.0224242],[5.430774,46.0222401],[5.4304506,46.0223202],[5.4302885,46.021982],[5.4300391,46.0216054],[5.4299637,46.0216342],[5.4300862,46.0218401],[5.4299565,46.021902],[5.4298847,46.0218195],[5.4298545,46.0217829],[5.4297689,46.0216672],[5.4297523,46.0216506],[5.4297379,46.0216389],[5.4296432,46.0215854],[5.429517,46.0214509],[5.4294188,46.0213458],[5.4293757,46.0213128],[5.4291918,46.0211768],[5.4291488,46.0211448],[5.4291083,46.0211214],[5.429024,46.0210828],[5.4292965,46.0208202],[5.4294241,46.0208894],[5.4295183,46.0209623],[5.4295455,46.0209865],[5.429613,46.0210554],[5.4296428,46.0210813],[5.4298751,46.0212862],[5.429988,46.0213782],[5.430014,46.0213973],[5.4300746,46.0214318],[5.430124,46.0214542],[5.4302569,46.0215069],[5.4303111,46.0215192],[5.4303632,46.0215166],[5.4306127,46.0214642],[5.430767,46.0214267]]]},"properties":{"id":"010020000A0225","commune":"01002","prefixe":"000","section":"A","numero":"225","contenance":9440,"arpente":false,"created":"2005-06-03","updated":"2018-09-25"}}, {"type":"Feature","id":"010020000A0346","geometry":{"type":"Polygon","coordinates":[[[5.4241952,46.0255535],[5.4233594,46.0262031],[5.4232624,46.0262774],[5.4226259,46.0267733],[5.4227608,46.0268718],[5.4227712,46.0268789],[5.4226123,46.0269855],[5.422565,46.0270182],[5.4223546,46.027145],[5.4222957,46.0271794],[5.4221794,46.0272376],[5.4221383,46.0272585],[5.4221028,46.027152],[5.4220695,46.0270523],[5.4220378,46.026962],[5.4220467,46.0269265],[5.4220524,46.0268709],[5.4220563,46.0268474],[5.4222945,46.0268985],[5.4224161,46.0267746],[5.4224581,46.0267904],[5.4226286,46.02666],[5.4226811,46.02662],[5.4227313,46.0265803],[5.4227813,46.0265406],[5.4228535,46.0264868],[5.4229063,46.0264482],[5.4229741,46.0264001],[5.4234903,46.0260331],[5.4235492,46.0259893],[5.4235787,46.0259663],[5.423645,46.0259126],[5.4237552,46.0258198],[5.4237839,46.0257951],[5.4238321,46.0257547],[5.4239258,46.0256723],[5.4239632,46.0256394],[5.4241164,46.0255075],[5.4241952,46.0255535]]]},"properties":{"id":"010020000A0346","commune":"01002","prefixe":"000","section":"A","numero":"346","contenance":2800,"arpente":false,"created":"2005-06-03","updated":"2018-09-25"}} ] } """) df = pd.DataFrame(data['features']) print(df) It'll give you the following DataFrame: type id geometry properties 0 Feature 010020000A0225 {'type': 'Polygon', 'coordinates': [[[5.430767... {'id': '010020000A0225', 'commune': '01002', '... 1 Feature 010020000A0346 {'type': 'Polygon', 'coordinates': [[[5.424195... {'id': '010020000A0346', 'commune': '01002', '... From there you can easily access the geometry and properties columns. Furthermore, if you want geometric and other properties in their own columns, you can use json_normalize: df = pd.json_normalize(data['features']) print(df) Output: type id geometry.type geometry.coordinates ... properties.contenance properties.arpente properties.created properties.updated 0 Feature 010020000A0225 Polygon [[[5.430767, 46.0214267], [5.4310805, 46.02201... ... 9440 False 2005-06-03 2018-09-25 1 Feature 010020000A0346 Polygon [[[5.4241952, 46.0255535], [5.4233594, 46.0262... ... 2800 False 2005-06-03 2018-09-25
How to get each values of a column in a data frame that contains a list of dictionaries?
As you can see I have a column that contains an array of dictionaries. I need to see if a key in any item has a certain value and return the row if it does. 0 [{'id': 473172988, 'node_id': 'MDU6TGFiZWw0NzM... 1 [{'id': 473172988, 'node_id': 'MDU6TGFiZWw0NzM... 2 [{'id': 473172988, 'node_id': 'MDU6TGFiZWw0NzM... 3 [{'id': 473173351, 'node_id': 'MDU6TGFiZWw0NzM... Is there a straightforward approach for this? The datatype of the column is an object.
You would need to give the exact format of your dictionary, but on the general principle you should loop over the elements: key = 'xxx' value = 'yyy' out = [any(d.get(key) == value for d in l) for l in df['your_column']] # slicing rows df[out]
Splitting at specific string from a dataframe column in Python
I have a dataframe with a column called "Spl" with the values below: I am trying to extract the values next to 'name': strings (some rows have multiple values) but I see the new column generated with the specific location of the memory. I used the below code to extract. Any help how to extract the values after "name:" string is much appreciated. Column values: 'name': 'Chirotherapie', 'name': 'Innen Medizin' 'name': 'Manuelle Medizin' 'name': 'Akupunktur', 'name': 'Chirotherapie', 'name': 'Innen Medizin' Code: df['Spl'] = lambda x: len(x['Spl'].str.split("'name':")) Output: <function <lambda> at 0x0000027BF8F68940>
Just simply do:- df['Spl']=df['Spl'].str.split("'name':").str.len()
Just do count df['Spl'] = df['Spl'].str.count("'name':")+1
Extract JSON | API | Pandas DataFrame
I am using the Facebook API (v2.10) to which I've extracted the data I need, 95% of which is perfect. My problem is the 'actions' metric which returns as a dictionary within a list within another dictionary. At present, all the data is in a DataFrame, however, the 'actions' column is a list of dictionaries that contain each individual action for that day. { "actions": [ { "action_type": "offsite_conversion.custom.xxxxxxxxxxx", "value": "7" }, { "action_type": "offsite_conversion.custom.xxxxxxxxxxx", "value": "3" }, { "action_type": "offsite_conversion.custom.xxxxxxxxxxx", "value": "144" }, { "action_type": "offsite_conversion.custom.xxxxxxxxxxx", "value": "34" }]} All this appears in one cell (row) within the DataFrame. What is the best way to: Get the action type, create a new column and use the Use "action_type" as the column name? List the correct value under this column It looks like JSON but when I look at the type, it's a panda series (stored as an object). For those willing to help (thank you, I greatly appreciate it) - can you either point me in the direction of the right material and I will read it and work it out on my own (I'm not entirely sure what to look for) or if you decide this is an easy problem, explain to me how and why you solved it this way. Don't just want the answer I have tried the following (with help from a friend) and it kind of works, but I have issues with this running in my script. IE: if it runs within a bigger code block, I get the following error: for i in range(df.shape[0]): line = df.loc[i, 'Conversions'] L = ast.literal_eval(line) for l in L: cid = l['action_type'] value = l['value'] df.loc[i, cid] = value If I save the DF as a csv, call it using pd.read_csv...it executes properly, but not within the script. No idea why. Error: ValueError: malformed node or string: [{'value': '1', 'action_type': 'offsite_conversion.custom.xxxxx}] Any help would be greatly appreciated. Thanks, Adrian
You can use json_normalize: In [11]: d # e.g. dict from json.load OR instead pass the json path to json_normalize Out[11]: {'actions': [{'action_type': 'offsite_conversion.custom.xxxxxxxxxxx', 'value': '7'}, {'action_type': 'offsite_conversion.custom.xxxxxxxxxxx', 'value': '3'}, {'action_type': 'offsite_conversion.custom.xxxxxxxxxxx', 'value': '144'}, {'action_type': 'offsite_conversion.custom.xxxxxxxxxxx', 'value': '34'}]} In [12]: pd.io.json.json_normalize(d, record_path="actions") Out[12]: action_type value 0 offsite_conversion.custom.xxxxxxxxxxx 7 1 offsite_conversion.custom.xxxxxxxxxxx 3 2 offsite_conversion.custom.xxxxxxxxxxx 144 3 offsite_conversion.custom.xxxxxxxxxxx 34
You can use df.join(pd.DataFrame(df['Conversions'].tolist()).pivot(columns='action_type', values='value').reset_index(drop=True)). Explanation: df['Conversions'].tolist() returns a list of dictionaries. This list is then transformed into a DataFrame using pd.DataFrame. Then, you can use the pivot function to pivot the table into the shape that you want. Lastly, you can join the table with your original DataFrame. Note that this only works if you DataFrame's index is the default (i.e., integers starting from 0). If this is not the case, you can do this instead: df2 = pd.DataFrame(df['Conversions'].tolist()).pivot(columns='action_type', values='value').reset_index(drop=True) for col in df2.columns: df[col] = df2[col]
Pandas DataFrame from Dictionary with Lists
I have an API that returns a single row of data as a Python dictionary. Most of the keys have a single value, but some of the keys have values that are lists (or even lists-of-lists or lists-of-dictionaries). When I throw the dictionary into pd.DataFrame to try to convert it to a pandas DataFrame, it throws a "Arrays must be the same length" error. This is because it cannot process the keys which have multiple values (i.e. the keys which have values of lists). How do I get pandas to treat the lists as 'single values'? As a hypothetical example: data = { 'building': 'White House', 'DC?': True, 'occupants': ['Barack', 'Michelle', 'Sasha', 'Malia'] } I want to turn it into a DataFrame like this: ix building DC? occupants 0 'White House' True ['Barack', 'Michelle', 'Sasha', 'Malia']
This works if you pass a list (of rows): In [11]: pd.DataFrame(data) Out[11]: DC? building occupants 0 True White House Barack 1 True White House Michelle 2 True White House Sasha 3 True White House Malia In [12]: pd.DataFrame([data]) Out[12]: DC? building occupants 0 True White House [Barack, Michelle, Sasha, Malia]
This turns out to be very trivial in the end data = { 'building': 'White House', 'DC?': True, 'occupants': ['Barack', 'Michelle', 'Sasha', 'Malia'] } df = pandas.DataFrame([data]) print df Which results in: DC? building occupants 0 True White House [Barack, Michelle, Sasha, Malia]
Solution to make dataframe from dictionary of lists where keys become a sorted index and column names are provided. Good for creating dataframes from scraped html tables. d = { 'B':[10,11], 'A':[20,21] } df = pd.DataFrame(d.values(),columns=['C1','C2'],index=d.keys()).sort_index() df C1 C2 A 20 21 B 10 11
Would it be acceptable if instead of having one entry with a list of occupants, you had individual entries for each occupant? If so you could just do n = len(data['occupants']) for key, val in data.items(): if key != 'occupants': data[key] = n*[val] EDIT: Actually, I'm getting this behavior in pandas (i.e. just with pd.DataFrame(data)) even without this pre-processing. What version are you using?
I had a closely related problem, but my data structure was a multi-level dictionary with lists in the second level dictionary: result = {'hamster': {'confidence': 1, 'ids': ['id1', 'id2']}, 'zombie': {'confidence': 1, 'ids': ['id3']}} When importing this with pd.DataFrame([result]), I end up with columns named hamster and zombie. The (for me) correct import would be to have these as row titles, and confidence and ids as column titles. To achieve this, I used pd.DataFrame.from_dict: In [42]: pd.DataFrame.from_dict(result, orient="index") Out[42]: confidence ids hamster 1 [id1, id2] zombie 1 [id3] This works for me with python 3.8 + pandas 1.2.3.
if you know the keys of the dictionary beforehand, why not first create an empty data frame and then keep adding rows?