Extracting data within complex functions/loops using lists/arrays - python
I have this block of code that does several things; I loops through files I have saved in a folder that are labeled 1-100. These files are all forecast files for a specific month (example june 2016). What this function does is read all the files and goes to previous files to conduct forecasts. I store all the values by different months. I want to see the totals for how accurate predictions were for "one month ago", "two months ago", etc. I am able to do this with the code, however I am having trouble extracting the exact values that contribute to this total using arrays/lists. The portion that does not have to do with arrays or lists work, but I am wondering how I can extract these specific numbers. I would want to use the append numbers (the list) for graphing purposes later, that is why I am extracting it, the # with ?'s indicate the list portion that does not seem to work
import pandas as pd
import csv
def nmonthaccuracy(basefilenumber, n):
basefileread = pd.read_csv(str(basefilenumber)+'.csv', encoding='Latin-1')
basefilevalue = basefileread.loc[basefileread['Customer'].str.contains('Customer A', na=False), 'Jun-16\nQty']
nmonthread = pd.read_csv(str(basefilenumber-n)+'.csv', encoding = 'Latin-1')
nmonthvalue = nmonthread.loc[nmonthread['Customer'].str.contains('Customer A', na=False), 'Jun-16\nQty']
return int(nmonthvalue)/int(basefilevalue)
N = 12
total_by_month = [0] * N
total_by_month_list [] * N #????
for basefilenumber in range(24,36):
for n in range(N):
total_by_month[n] += nmonthaccuracy(basefilenumber, n)
total_by_month_list[n].append(nmonthaccuracy(basefilenumber,n)) #????
onetotal = total_by_month[1]
twototal = total_by_month[2]
#etc
Try running your code by initializing total_by_month_list as
total_by_month_list = [[] for _ in range(N)]
Without your data, it's currently speculative. What I understood is that total_by_month_list should be a list of 12 sublists.
Related
How to find matches between csv files based on two columns within a range
I'm currently struggling to put together some code that will find the matches of values in two different columns in two csv files within a range. I have tried using the code below, but it doesn't output what I am trying to accomplish. Basically, I want to output a new file that contains all of the lines in the second file that have matches to the same columns in the first file, not merge them together. I've added more detailed clarification below my code. I feel like what I've done so far is probably completely wrong. What do I need to change in order for my code to produce the results I am looking for? import csv with open('F435W.csv') as csvF435: readCSV1 = csv.reader(csvF435, delimiter=',') with open("F550Mnew.csv", "w") as new_F550M: pass with open("F550Mnew.csv", "a") as new_F550M: for header in readCSV1: new_F550M.write(','.join(header)+'\n') break for l435 in readCSV1: with open('F550M.csv') as csvF550: readCSV2 = csv.reader(csvF550, delimiter=',') for l550 in readCSV2: if isfloat(l435[12]) and isfloat(l550[12]) and abs(float(l435[12])-float(l550[12])) < 0.002778: if isfloat(l435[13]) and isfloat(l550[13]) and abs(float(l435[13])-float(l550[13])) < 0.002778: new_F550M.write(','.join(l550)+'\n') For clarification, each file has an X column and a Y column so basically each row corresponds to an (X,Y) point. In addition, there are 21 other columns of data that are not necessary for finding matches, but need to be included in the final output file. I am trying to find points in the second file that match the points in the first file within a radius. This is because I know that none of my points will be exact matches. In my data, my X is column 13 and my Y is column 14. The way I have tried to accomplish this is by finding the differences between every X in the first file and every X in the second file (eg. X1-X2), and the differences between every Y in the first file and every Y in the second file (eg. Y1-Y2). Then, every row in the second file which corresponds to differences for both X and Y which are less than my radius value (0.0002778) would be considered a match to the first file. Unfortunately, my code produces a file with over 300,000 points when my original files only have 7000 points. There should be less data, not more data. It also includes many repeats of data, when there should not be any repeats at all. Thank you for your time! Sample of what the data looks like: I apologize for the length, but I am afraid they will not contain enough matches to be useful if I don't include enough of the data. F435W.csv (file 1) 1,2017.013,0.01242859,-8.2618,0,51434.12,0.3269918,-11.7781,0,0.01957931,1387.9406,541.916,49.9898514,41.5266996,8.81E+01,1.63E+03,1.44E+02,40.535,8.65,84.72,0.00061,0.00035,62.14 2,84.73392,0.01245409,-4.8201,0.0002,112.9723,0.04012135,-5.1324,0.0004,-0.002142646,150.306,146.7986,49.9942613,41.5444392,4.92E+00,5.60E+00,-2.02E-01,2.379,2.206,-74.69,0.00339,0.0029,88.88 3,215.1939,0.01242859,-5.8321,0.0001,262.2751,0.03840466,-6.0469,0.0002,-0.002961465,3248.686,52.8478,50.003155,41.5019044,4.77E+00,5.05E+00,-1.63E-01,2.263,2.166,-65.29,0.002,0.0019,-66.78 4,0.3796681,0.01240305,1.0515,0.0355,0.5823653,0.05487975,0.587,0.1023,-0.00425157,3760.344,11.113,50.0051049,41.4949256,1.93E+00,1.02E+00,-7.42E-02,1.393,1.007,-4.61,0.05461,0.03818,-6.68 5,0.9584663,0.01249223,0.0461,0.0142,1.043696,0.0175857,-0.0464,0.0183,-0.004156116,4013.2063,9.1225,50.0057256,41.4914444,1.12E+00,9.75E-01,1.09E-01,1.085,0.957,28.34,0.01934,0.01745,44.01 6,2.379565,0.01249223,-0.9412,0.0057,0.231205,0.02710035,1.59,0.1273,-0.004135321,3824.3706,9.0756,50.0052903,41.4940468,7.81E-01,6.99E-02,4.27E-02,0.885,0.26,3.42,0.01265,0.00622,15.52 7,0.3171223,0.01250492,1.2469,0.0428,0.5233852,0.05406558,0.7029,0.1122,-0.00399635,4097.3604,7.0301,50.0059585,41.4902884,9.61E-01,1.63E+00,-3.94E-01,1.346,0.883,-65.16,0.06171,0.04005,-65.05 8,0.289245,0.0125176,1.3468,0.047,0.2744479,0.02238134,1.4039,0.0886,-0.004173243,3904.7402,7.3912,50.0055069,41.4929422,7.90E-01,2.38E-01,7.13E-02,0.894,0.479,7.24,0.04501,0.02071,8.29 9,0.3543034,0.01247953,1.1266,0.0383,0.7666836,0.06376094,0.2885,0.0903,-0.004009248,4107.0684,3.259,50.0060503,41.4901611,3.53E+00,1.28E+00,-4.60E-01,1.903,1.09,-11.12,0.06873,0.03955,-11.22 10,1.308331,0.01250492,-0.2918,0.0104,-0.005209296,0.004877397,99,99,-0.004193406,3933.9834,6,50.0056001,41.4925416,5.78E-01,8.33E-02,0.00E+00,0.76,0.289,0,0.01272,0.00424,0 11,3.995717,0.01250492,-1.504,0.0034,0.1589517,0.007450347,1.9968,0.0509,-0.003990021,4069.0469,3.0234,50.0059668,41.4906855,8.03E-01,2.29E-02,1.02E-02,0.896,0.151,0.75,0.00888,0.00361,5.59 12,1.067634,0.01250492,-0.0711,0.0127,0.1260926,0.02787585,2.2483,0.2401,-0.004042602,4048.9148,4,50.0059023,41.4909612,7.40E-01,8.33E-02,0.00E+00,0.86,0.289,0,0.02449,0.00576,0 13,0.2808423,0.01162418,1.3788,0.0449,0.4633991,0.02235104,0.8351,0.0524,-0.004015559,4114.6655,2.0641,50.0060898,41.4900585,9.65E-01,5.88E-01,-9.47E-02,0.994,0.752,-13.34,0.05405,0.03814,-15.13 14,1.067291,0.01245409,-0.0707,0.0127,1.081617,0.01516444,-0.0852,0.0152,-0.004168633,3960.8787,18.0524,50.0054405,41.4921501,6.84E-01,8.29E-01,-6.18E-02,0.923,0.813,-69.77,0.01468,0.01229,-78.83 15,0.5216251,0.0125176,0.7066,0.0261,0.584776,0.01824955,0.5825,0.0339,-0.003026338,2661.6533,58.4563,50.0016952,41.5099844,8.51E-01,1.17E+00,-7.27E-02,1.089,0.914,-77.72,0.03244,0.02498,-81.68 16,0.6062042,0.01249223,0.5435,0.0224,0.8726375,0.05509822,0.1479,0.0686,-0.003950399,4149.8169,31.0127,50.0056384,41.489524,9.30E-01,3.48E+00,2.03E-01,1.87,0.956,85.48,0.05307,0.0241,86.01 17,0.1324067,0.01242859,2.1952,0.1019,0.1208224,0.01290438,2.2946,0.116,-0.004166729,3911.6807,12.661,50.005426,41.4928374,2.17E-01,2.24E-01,-1.08E-01,0.574,0.335,-45.89,0.0721,0.04162,-44.98 18,0.2136006,0.01247953,1.676,0.0634,0.3511444,0.02471001,1.1363,0.0764,-0.003978713,4096.9111,15.6285,50.0057993,41.4902797,1.00E+00,4.37E-01,2.85E-01,1.058,0.564,22.64,0.07548,0.03957,23.17 19,0.1470979,0.01244135,2.081,0.0919,0.1216703,0.0168958,2.287,0.1508,-0.004147241,3695.311,13.7044,50.004907,41.4958173,2.14E-01,2.08E-01,9.20E-02,0.551,0.345,44.05,0.07073,0.04115,45.12 20,0.5434682,0.01250492,0.6621,0.025,0.5819249,0.01592951,0.5878,0.0297,-0.004136056,3866.6416,24.8316,50.0050981,41.493437,8.34E-01,9.96E-01,2.74E-01,1.096,0.793,53.22,0.02966,0.02055,58.08 21,0.2259093,0.01249223,1.6152,0.0601,0.2848583,0.01867901,1.3634,0.0712,-0.00409535,3645.521,20.0162,50.0046759,41.4964926,5.71E-01,4.26E-01,-1.11E-02,0.756,0.652,-4.34,0.03735,0.0305,0.08 22,0.9499883,0.01247953,0.0557,0.0143,0.9711754,0.01891141,0.0318,0.0211,-0.003134006,3378.7927,19.5305,50.0040686,41.5001691,8.66E-01,4.09E-01,3.57E-03,0.931,0.639,0.45,0.01623,0.01142,-1.19 23,1.125635,0.01240305,-0.1285,0.012,1.050538,0.02402694,-0.0535,0.0248,-0.003295973,3132.9458,24.9024,50.0034018,41.5035477,9.65E-01,7.83E-01,-1.44E-01,1.022,0.839,-28.88,0.01702,0.01288,-21 24,0.168302,0.01249223,1.9348,0.0806,0.2447732,0.01930529,1.5281,0.0857,-0.004140488,3904.7268,27.0386,50.0051454,41.4929084,4.47E-01,4.56E-01,-1.28E-02,0.682,0.662,-54.61,0.04399,0.04068,89.66 25,0.0542859,0.01244135,3.1633,0.2489,0.08799078,0.007964755,2.6389,0.0983,-0.003241792,3454.2612,25.2749,50.0041373,41.4991191,1.93E-01,1.99E-01,-7.18E-02,0.518,0.353,-46.27,0.06408,0.03839,-44.76 26,0.4379335,0.01242859,0.8965,0.0308,0.4661828,0.01542368,0.8286,0.0359,-0.00336337,3478.7058,32.3355,50.0040639,41.4987701,6.15E-01,8.96E-01,-2.91E-02,0.948,0.782,-84.15,0.02891,0.02521,-70.04 27,0.1515608,0.01249223,2.0485,0.0895,0.1935181,0.01712885,1.7832,0.0961,-0.002904789,2982.0017,29.9904,50.0029594,41.505619,3.46E-01,3.61E-01,1.55E-05,0.601,0.588,89.94,0.05241,0.05241,-80.48 28,0.6658883,0.01250492,0.4415,0.0204,0.718064,0.01780974,0.3596,0.0269,-0.00324104,3408.0103,36.2539,50.0038284,41.4997375,9.45E-01,1.11E+00,1.98E-01,1.115,0.902,56.45,0.02706,0.02147,51.52 29,0.7244126,0.01244135,0.35,0.0187,1.030102,0.02744665,-0.0322,0.0289,-0.00280412,3259.0889,37.3165,50.0034648,41.5017879,8.65E-01,1.01E+00,5.85E-02,1.017,0.919,70.87,0.02225,0.02011,55.79 30,0.1651701,0.01247953,1.9552,0.0821,0.163293,0.01641976,1.9676,0.1092,-0.003909466,3595.4846,31.9761,50.0043403,41.4971614,2.50E-01,4.42E-01,2.21E-01,0.766,0.324,56.75,0.08087,0.03087,58.28 F550M.csv (file 2) 2,1921.566,0.01258874,-8.2091,0,37128.06,0.2618096,-11.4243,0,0.01455503,4617.5225,554.576,49.9887896,41.5264699,6.09E+01,8.09E+02,1.78E+01,28.459,7.779,88.63,0.00054,0.00036,77.04 3,1.055918,0.01256313,-0.0591,0.0129,9.834856,0.1109255,-2.4819,0.0122,-0.002955142,3936.4946,85.3255,49.9949149,41.5370016,3.98E+01,1.23E+01,1.54E+01,6.83,2.336,24.13,0.06362,0.01965,23.98 4,151.2355,0.01260153,-5.4491,0.0001,184.0693,0.03634057,-5.6625,0.0002,-0.002626019,3409.2642,76.9891,49.9931935,41.5442109,4.02E+00,4.35E+00,-1.47E-03,2.086,2.005,-89.75,0.00227,0.00198,66.61 5,0.3506025,0.01258874,1.138,0.039,0.3466277,0.01300407,1.1503,0.0407,-0.002441164,3351.9893,8.9147,49.9942299,41.5451727,4.97E-01,5.07E-01,7.21E-03,0.715,0.702,62.75,0.02,0.01989,82.88 6,1.166133,0.01257594,-0.1669,0.0117,0.005819145,0.009692424,5.5879,1.8089,-0.003201006,3476.9932,10,49.9946543,41.5434658,5.88E-01,8.33E-02,0.00E+00,0.767,0.289,0,0.01497,0.00499,0 7,0.1372164,0.0125503,2.1565,0.0993,0.1238123,0.02608246,2.2681,0.2288,-0.003556473,3535.5281,13.4586,49.9947993,41.5426587,2.49E-01,2.48E-01,-7.69E-03,0.506,0.491,-43.27,0.05264,0.05237,-55.87 8,0.6174777,0.01260153,0.5234,0.0222,0.6206718,0.01300407,0.5178,0.0228,-0.002441164,3357.0044,20.0487,49.9940449,41.5450748,5.10E-01,5.22E-01,-6.28E-03,0.724,0.712,-66.7,0.01194,0.01192,84.82 9,1.46848,0.01260153,-0.4172,0.0093,0.001897994,0.009688255,6.8043,5.5435,-0.003612399,3584.0171,16,49.9949252,41.5419909,5.87E-01,8.33E-02,0.00E+00,0.766,0.289,0,0.01175,0.00392,0 10,1.452348,0.01258874,-0.4052,0.0094,3.124427,0.04807406,-1.2369,0.0167,-0.003148756,3805.6069,39.5791,49.9952831,41.5389075,2.25E+00,3.87E+00,-6.77E-01,2.03,1.416,-70.08,0.0302,0.01891,-67.61 11,0.1548658,0.01260153,2.0251,0.0884,0.1777253,0.01630147,1.8756,0.0996,-0.002919044,3459.7681,25.6248,49.9943085,41.5436591,4.64E-01,2.34E-01,8.40E-02,0.701,0.455,18.09,0.05739,0.03321,18.33 12,0.5046132,0.01253746,0.7426,0.027,0.7798272,0.04462456,0.27,0.0621,-0.00261193,3418.9119,65.5326,49.9934365,41.5441099,6.87E-01,2.77E+00,-2.92E-01,1.678,0.804,-82.19,0.05363,0.02182,-83.28 13,0.380733,0.01260153,1.0484,0.0359,0.4313257,0.01605258,0.913,0.0404,-0.003497544,3548.8484,34.5602,49.9944623,41.542421,8.27E-01,8.51E-01,8.92E-02,0.964,0.865,48.75,0.03776,0.03252,30.61 14,0.1643925,0.01258874,1.9603,0.0832,0.2181225,0.01839054,1.6532,0.0916,-0.003121084,3710.6785,33.3215,49.9950598,41.5402182,2.18E-01,2.18E-01,1.03E-01,0.567,0.339,45,0.0757,0.04376,45 15,0.3959635,0.01260153,1.0059,0.0346,0.9984215,0.0763398,0.0017,0.083,-0.003106286,3805.9988,48.3363,49.995125,41.5388789,1.87E+00,3.12E+00,4.86E-01,1.813,1.304,71.09,0.0559,0.04105,67.61 16,0.1625628,0.01260153,1.9724,0.0842,0.3490304,0.02234424,1.1428,0.0695,-0.002472953,3410.77,38.0388,49.9939083,41.544294,1.77E-01,4.75E-01,8.92E-03,0.689,0.421,88.29,0.0769,0.04707,89.86 17,0.1725209,0.01260153,1.9079,0.0793,0.2965718,0.02357189,1.3197,0.0863,-0.003454017,3629.0247,40.9706,49.9946304,41.541311,3.73E-01,7.91E-01,-3.73E-01,1.004,0.393,-59.65,0.09781,0.03734,-58.27 18,0.3034717,0.01260153,1.2947,0.0451,0.5031242,0.02774418,0.7458,0.0599,-0.003073985,4079.0825,42,49.9962105,41.5351731,6.68E-01,8.33E-02,0.00E+00,0.818,0.289,0,0.06348,0.02106,0 19,1.593927,0.01260153,-0.5062,0.0086,1.860803,0.0219809,-0.6743,0.0128,-0.003038161,4065.9434,58.3703,49.9958657,41.5353087,1.75E+00,1.41E+00,-7.15E-03,1.323,1.188,-1.21,0.01697,0.01464,-0.43 20,0.5464995,0.01258874,0.656,0.025,0.5661472,0.0144696,0.6177,0.0278,-0.003053429,4045.0474,54.439,49.9958631,41.535604,5.43E-01,8.46E-01,-1.22E-03,0.92,0.737,-89.77,0.02257,0.01649,-89.72 21,1.303251,0.01253746,-0.2876,0.0104,1.296672,0.01418861,-0.2821,0.0119,-0.00259741,4240.1406,55.2714,49.9965409,41.5329423,6.05E-01,6.81E-01,7.89E-03,0.826,0.777,84.15,0.00892,0.00852,69.62 22,0.5174786,0.01260153,0.7153,0.0264,0.5260691,0.01390194,0.6974,0.0287,-0.003019847,3828.95,55.19,49.9950817,41.5385478,5.18E-01,7.56E-01,-6.34E-02,0.879,0.709,-75.96,0.0236,0.01643,-75.02 23,0.1551826,0.01260153,2.0229,0.0882,0.166565,0.01726119,1.946,0.1125,-0.003271136,3504.7439,52.7386,49.9939745,41.5429739,1.91E-01,6.86E-01,1.89E-01,0.866,0.356,71.33,0.10376,0.04235,71.56 24,0.2214222,0.01260153,1.6369,0.0618,0.2389908,0.01360924,1.554,0.0618,-0.00285033,3750.3167,54.0027,49.994824,41.5396229,4.32E-01,5.51E-01,1.68E-03,0.742,0.657,89.18,0.04862,0.04505,89.94 25,0.1336059,0.01253746,2.1854,0.1019,0.1320868,0.009830156,2.1979,0.0808,-0.002921393,3459.6851,51.7091,49.9938331,41.5435908,2.16E-01,2.06E-01,-9.16E-02,0.55,0.345,-43.52,0.06231,0.03626,-45.19 26,0.1703959,0.01260153,1.9214,0.0803,0.1577456,0.0152816,2.0051,0.1052,-0.002779523,3446.95,49,49.9938372,41.5437717,7.29E-01,8.33E-02,0.00E+00,0.854,0.289,0,0.11183,0.03721,0 27,1.896325,0.01258874,-0.6948,0.0072,1.941203,0.0152816,-0.7202,0.0085,-0.00306097,3809.6836,57.8143,49.9949655,41.5388035,7.38E-01,6.80E-01,7.46E-03,0.86,0.824,7.18,0.00713,0.00678,59.71 28,0.6522877,0.01260153,0.4639,0.021,0.1713469,0.01312423,1.9153,0.0832,-0.002447558,4271.9614,52,49.9967135,41.5325172,5.92E-01,8.33E-02,0.00E+00,0.77,0.289,0,0.0274,0.00913,0 29,0.1370073,0.0125503,2.1581,0.0995,0.101415,0.02614047,2.4847,0.2799,-0.002207851,4324.667,55.3374,49.99684,41.5317898,2.22E-01,2.24E-01,1.12E-01,0.579,0.332,45.18,0.07753,0.04476,45 30,0.2240251,0.01253746,1.6243,0.0608,0.2254432,0.01360924,1.6174,0.0656,-0.003037372,3960.3042,58.9024,49.9954807,41.5367473,4.18E-01,4.81E-01,-1.07E-02,0.695,0.645,-80.65,0.03802,0.03492,-88.86
You are complicating the program by nesting all the loops and conditionals. Break it down into simple steps. Do the following. 1. Read both the csv files and convert them into 2d lists. 2. Compare the columns/values of the lists within a loop based on the given index, add the rows from second list to a new output list. 3. Write the output list to a csv file. def read_file(filepath): with open(filepath,'r') as f: x = csv.reader(f) l = list(x) return l l435 = read_file('F435W.csv') l550 = read_file('F550M.csv') new_F550M = [] r = 0.002778 for i in l550: for j in l435: # I did't exactly get your if condition, so I am putting it down based on what I understood, so if it is wrong, modify it accordingly. if isfloat(i[12]) and isfloat(j[12]) and abs(float(i[12]) float(j[12])) < r: if isfloat(i[13]) and isfloat(j[13]) and abs(float(i[13]) float(j[13])) < r: new_F550M.append(i) with open('new_F550M.csv','w') as f: out = csv.writer(f) out.writerows(new_F550M)
Python Basics - First Project/Challenge
I'm extremely new to Python (and software programming/development in general). I decided to use the scenario below as my first project. The project includes 5 main personal challenges. Some of the challenges I have been able to complete (although probably not the most effecient way), and others I'm struggling with. Any feedback you have on my approach and recommendations for improvement is GREATLY appreciated. Project Scenario = "If I doubled my money each day for 100 days, how much would I end up with at day #100? My starting amount on Day #1 is $1.00" 1.) Challenge 1 - What is the net TOTAL after day 100 - (COMPLETED, I think, please correct me if I'm wrong) days = 100 compound_rate = 2 print('compound_rate ** days) # 2 raised to the 100th #==Result=== 1267650600228229401496703205376 2.) Challenge 2 - Print to screen the DAYS in the first column, and corresponding Daily Total in the second column. - (COMPLETED, I think, please correct me if I'm wrong) compound_rate = 2 days_range = list(range(101)) for x in days_range: print (str(x),(compound_rate ** int(x))) # ===EXAMPLE Results # 0 1 # 1 2 # 2 4 # 3 8 # 4 16 # 5 32 # 6 64 # 100 1267650600228229401496703205376 3.) Challenge 3 - Write TOTAL result (after the 100 days) to an external txt file - (COMPLETED, I think, please correct me if I'm wrong) compound_rate = 2 days_range = list(range(101)) hundred_days = (compound_rate ** 100) textFile = open("calctest.txt", "w") textFile.write(str(hundred_days)) textFile.close() #===Result==== string of 1267650600228229401496703205376 --> written to my file 'calctest.txt' 4.) Challenge 4 - Write the Calculated running DAILY Totals to an external txt file. Column 1 will be the Day, and Column 2 will be the Amount. So just like Challenge #2 but to an external file instead of screen NEED HELP, I can't seem to figure this one out. 5.) Challenge 5 - Somehow plot or chart the Daily Results (based on #4) - NEED GUIDANCE. I appreciate everyone's feedback as I start on my personal Python journey!
challenge 2 This will work fine, but there's no need to write list(range(101)), you can just write range(101). In fact, there's no need even to create a variable to store that, you can just do this: for x in range(101): print("whatever you want to go here") challenge 3 Again, this will work fine, but when writing to a file, it is normally best to use a with statement, this means that you don't need to close the file at the end, as python will take care of that. For example: with open("calctest.txt", "w") as f: write(str(hundred_days)) challenge 4 Use a for loop as you did with challenge 2. Use "\n" to write a new line. Again do everything inside a with statement. e.g. with open("calctest.txt", "w") as f: for x in range(101): f.write("something here \n"). (would write a file with 'something here ' written 101 times) challenge 5 There is a python library called matplotlib, which I have never used, but I would suggest that would be where to go to in order to solve this task. I hope this is of some help :)
You can use what you did in challenge 3 to open and close the ouput file. In between, you have to do what you did in challenge 2 to compute the data for each day. In stead of writing the daily result to the stream, you will have to combine it into a string. After that, you can write that string to the file, exactly like you did in challenge 3.
Challenge One: This is the correct way. days = 100 compound_rate = 2 print("Result after 100 days" + (compound_rate ** days)) Challenge Two This is corrected. compound_rate = 2 days_range = list(range(101)) for x in days_range: print(x + (compound_rate ** x)) Challenge Three This one is close but you didn't need to cast the result of hundred_days to a string as you can write the integer to a file and python doesn't care most of the time. Explicit casts need only to be worried about when using the data in some way other than simply printing it. compound_rate = 2 days_range = list(range(101)) hundred_days = (compound_rate ** 100) textFile = open("calctest.txt", "w") textFile.write(hundred_days) textFile.close() Challenge Four For this challenge, you will want to look into the python CSV module. You can write the data in two rows separated by commas very simply with this module. Challenge Five For this challenge, you will want to look into the python library matplotlib. This library will give you tools to work with the data in a graphical way.
Answer for challenge 1 is as follows: l = [] for a in range(0,100): b = 2 ** a l.append(b) print("Total after 100 days", sum(l))
import os, sys import datetime import time #to get the current work directory, we use below os.getcwd() print(os.getcwd()) #to get the list of files and folders in a path, we use os.listdir print(os.listdir()) #to know the files inside a folder using path spath = (r'C:\Users\7char') l = spath print(os.listdir(l)) #converting a file format to other, ex: txt to py path = r'C:\Users\7char' print(os.listdir(path)) # after looking at the list of files, we choose to change 'rough.py' 'rough.txt' os.chdir(path) os.rename('rough.py','rough.txt') #check whether the file has changed to new format print(os.listdir(path)) #yes now the file is changed to new format print(os.stat('rough.txt').st_size) # by using os.stat function we can see the size of file (os.stat(file).sst_size) path = r"C:\Users\7char\rough.txt" datetime = os.path.getmtime(path) moddatetime = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(datetime)) print("Last Modified Time : ", moddatetime) #differentiating b/w files and folders using - os.path.splitext import os path = r"C:\Users\7char\rough.txt" dir(os.path) files = os.listdir() for file in files: print(os.path.splitext(file)) #moving a file from one folder to other (including moving with folders of a path or moving into subforlders) import os char_7 = r"C:\Users\7char" cleardata = r"C:\Users\clearadata" operating = os.listdir(r"C:\Users\7char") print(operating) for i in operating: movefrom = os.path.join(char_7,i) moveto = os.path.join(cleardata,i) print(movefrom,moveto) os.rename(movefrom,moveto) #now moving files based on length of individual charecter (even / odd) to a specified path (even or odd). import os origin_path = r"C:\Users\movefilehere" fivechar_path= r"C:\Users\5char" sevenchar_path = r"C:\Users\7char" origin_path = os.listdir(origin_path) for file_name in origin_pathlist: l = len(file_name) if l % 2 == 0: evenfilepath = os.path.join(origin_path,file_name) newevenfilepath = os.path.join(fivechar_path,file_name) print(evenfilepath,newevenfilepath) os.rename(evenfilepath,newevenfilepath) else: oddfilepath = os.path.join(origin_path,file_name) newoddfilepath = os.path.join(sevenchar_path,file_name) print(oddfilepath,newoddfilepath) os.rename(oddfilepath,newoddfilepath) #finding the extension in a folder using isdir import os path = r"C:\Users\7char" print(os.path.isdir(path)) #how a many files .py and .txt (any files) in a folder import os from os.path import join, splitext from glob import glob from collections import Counter path = r"C:\Users\7char" c = Counter([splitext(i)[1][1:] for i in glob(join(path, '*'))]) for ext, count in c.most_common(): print(ext, count) #looking at the files and extensions, including the total of extensions. import os from os.path import join, splitext from collections import defaultdict path = r"C:\Users\7char" c = defaultdict(int) files = os.listdir(path) for filenames in files: extension = os.path.splitext(filenames)[-1] c[extension]+=1 print(os.path.splitext(filenames)) print(c,extension) #getting list from range list(range(4)) #break and continue statements and else clauses on loops for n in range(2,10): for x in range(2,n): if n%x == 0: print(n,'equals',x, '*', n//x) break else: print(n, 'is a prime number') #Dictionaries #the dict() constructer builds dictionaries directly from sequences of key-value pairs dict([('ad', 1212),('dasd', 2323),('grsfd',43324)]) #loop over two or more sequences at the same time, the entries can be paired with the zip() function. questions = ['name', 'quest', 'favorite color'] answers = ['lancelot', 'the holy grail', 'blue'] for q, a in zip(questions, answers): print('What is your {0}? It is {1}.'.format(q, a)) #Using set() basket = ['apple', 'orange', 'apple', 'pear', 'orange', 'banana'] for f in sorted(set(basket)): print(f)
Why does my association model find subgroups in a dataset when there shouldn't any?
I give a lot of information on the methods that I used to write my code. If you just want to read my question, skip to the quotes at the end. I'm working on a project that has a goal of detecting sub populations in a group of patients. I thought this sounded like the perfect opportunity to use association rule mining as I'm currently taking a class on the subject. I there are 42 variables in total. Of those, 20 are continuous and had to be discretized. For each variable, I used the Freedman-Diaconis rule to determine how many categories to divide a group into. def Freedman_Diaconis(column_values): #sort the list first column_values[1].sort() first_quartile = int(len(column_values[1]) * .25) third_quartile = int(len(column_values[1]) * .75) fq_value = column_values[1][first_quartile] tq_value = column_values[1][third_quartile] iqr = tq_value - fq_value n_to_pow = len(column_values[1])**(-1/3) h = 2 * iqr * n_to_pow retval = (column_values[1][-1] - column_values[1][1])/h test = int(retval+1) return test From there I used min-max normalization def min_max_transform(column_of_data, num_bins): min_max_normalizer = preprocessing.MinMaxScaler(feature_range=(1, num_bins)) data_min_max = min_max_normalizer.fit_transform(column_of_data[1]) data_min_max_ints = take_int(data_min_max) return data_min_max_ints to transform my data and then I simply took the interger portion to get the final categorization. def take_int(list_of_float): ints = [] for flt in list_of_float: asint = int(flt) ints.append(asint) return ints I then also wrote a function that I used to combine this value with the variable name. def string_transform(prefix, column, index): transformed_list = [] transformed = "" if index < 4: for entry in column[1]: transformed = prefix+str(entry) transformed_list.append(transformed) else: prefix_num = prefix.split('x') for entry in column[1]: transformed = str(prefix_num[1])+'x'+str(entry) transformed_list.append(transformed) return transformed_list This was done to differentiate variables that have the same value, but appear in different columns. For example, having a value of 1 for variable x14 means something different from getting a value of 1 in variable x20. The string transform function would create 14x1 and 20x1 for the previously mentioned examples. After this, I wrote everything to a file in basket format def create_basket(list_of_lists, headers): #for filename in os.listdir("."): # if filename.e if not os.path.exists('baskets'): os.makedirs('baskets') down_length = len(list_of_lists[0]) with open('baskets/dataset.basket', 'w') as basketfile: basket_writer = csv.DictWriter(basketfile, fieldnames=headers) for i in range(0, down_length): basket_writer.writerow({"trt": list_of_lists[0][i], "y": list_of_lists[1][i], "x1": list_of_lists[2][i], "x2": list_of_lists[3][i], "x3": list_of_lists[4][i], "x4": list_of_lists[5][i], "x5": list_of_lists[6][i], "x6": list_of_lists[7][i], "x7": list_of_lists[8][i], "x8": list_of_lists[9][i], "x9": list_of_lists[10][i], "x10": list_of_lists[11][i], "x11": list_of_lists[12][i], "x12":list_of_lists[13][i], "x13": list_of_lists[14][i], "x14": list_of_lists[15][i], "x15": list_of_lists[16][i], "x16": list_of_lists[17][i], "x17": list_of_lists[18][i], "x18": list_of_lists[19][i], "x19": list_of_lists[20][i], "x20": list_of_lists[21][i], "x21": list_of_lists[22][i], "x22": list_of_lists[23][i], "x23": list_of_lists[24][i], "x24": list_of_lists[25][i], "x25": list_of_lists[26][i], "x26": list_of_lists[27][i], "x27": list_of_lists[28][i], "x28": list_of_lists[29][i], "x29": list_of_lists[30][i], "x30": list_of_lists[31][i], "x31": list_of_lists[32][i], "x32": list_of_lists[33][i], "x33": list_of_lists[34][i], "x34": list_of_lists[35][i], "x35": list_of_lists[36][i], "x36": list_of_lists[37][i], "x37": list_of_lists[38][i], "x38": list_of_lists[39][i], "x39": list_of_lists[40][i], "x40": list_of_lists[41][i]}) and I used the apriori package in Orange to see if there were any association rules. rules = Orange.associate.AssociationRulesSparseInducer(patient_basket, support=0.3, confidence=0.3) print "%4s %4s %s" % ("Supp", "Conf", "Rule") for r in rules: my_rule = str(r) split_rule = my_rule.split("->") if 'trt' in split_rule[1]: print 'treatment rule' print "%4.1f %4.1f %s" % (r.support, r.confidence, r) Using this, technique I found quite a few association rules with my testing data. THIS IS WHERE I HAVE A PROBLEM When I read the notes for the training data, there is this note ...That is, the only reason for the differences among observed responses to the same treatment across patients is random noise. Hence, there is NO meaningful subgroup for this dataset... My question is, why do I get multiple association rules that would imply that there are subgroups, when according to the notes I shouldn't see anything? I'm getting lift numbers that are above 2 as opposed to the 1 that you should expect if everything was random like the notes state. Supp Conf Rule 0.3 0.7 6x0 -> trt1 Even though my code runs, I'm not getting results anywhere close to what should be expected. This leads me to believe that I messed something up, but I'm not sure what it is.
After some research, I realized that my sample size is too small for the number of variables that I have. I would need a way larger sample size in order to really use the method that I was using. In fact, the method that I tried to use was developed with the assumption that it would be run on databases with hundreds of thousands or millions of rows.
Efficiently Find Partial String Match --> Values Starting From List of Values in 5 GB file with Python
I have a 5GB file of businesses and I'm trying to extract all the businesses that whose business type codes (SNACODE) start with the SNACODE corresponding to grocery stores. For example, SNACODEs for some businesses could be 42443013, 44511003, 44419041, 44512001, 44522004 and I want all businesses whose codes start with my list of grocery SNACODES codes = [4451,4452,447,772,45299,45291,45212]. In this case, I'd get the rows for 44511003, 44512001, and 44522004 Based on what I googled, the most efficient way to read in the file seemed to be one row at a time (if not the SQL route). I then used a for loop and checked if my SNACODE column started with any of my codes (which probably was a bad idea but the only way I could get to work). I have no idea how many rows are in the file, but there are 84 columns. My computer was running for so long that I asked a friend who said it should only take 10-20 min to complete this task. My friend edited the code but I think he misunderstood what I was trying to do because his result returns nothing. I am now trying to find a more efficient method than re-doing my 9.5 hours and having my laptop run for an unknown amount of time. The closest thing I've been able to find is most efficient way to find partial string matches in large file of strings (python), but it doesn't seem like what I was looking for. Questions: What's the best way to do this? How long should this take? Is there any way that I can start where I stopped? (I have no idea how many rows of my 5gb file I read, but I have the last saved line of data--is there a fast/easy way to find the line corresponding to a unique ID in the file without having to read each line?) This is what I tried -- in 9.5 hours it outputted a 72MB file (200k+ rows) of grocery stores codes = [4451,4452,447,772,45299,45291,45212] #codes for grocery stores for df in pd.read_csv('infogroup_bus_2010.csv',sep=',', chunksize=1): data = np.asarray(df) data = pd.DataFrame(data, columns = headers) for code in codes: if np.char.startswith(str(data["SNACODE"][0]), str(code)): with open("grocery.csv", "a") as myfile: data.to_csv(myfile, header = False) print code break #break code for loop if match grocery.to_csv("grocery.csv", sep = '\t') This is what my friend edited it to. I'm pretty sure the x = df[df.SNACODE.isin(codes)] is only matching perfect matches, and thus returning nothing. codes = [4451,4452,447,772,45299,45291,45212] matched = [] for df in pd.read_csv('infogroup_bus_2010.csv',sep=',', chunksize=1024*1024, dtype = str, low_memory=False): x = df[df.SNACODE.isin(codes)] if len(x): matched.append(x) print "Processed chunk and found {} matches".format(len(x)) output = pd.concat(matched, axis=0) output.to_csv("grocery.csv", index = False) Thanks!
To increase speed you could pre-build a single regexp matching the lines you need and the read the raw file lines (no csv parsing) and check them with the regexp... codes = [4451,4452,447,772,45299,45291,45212] col_number = 4 # Column number of SNACODE expr = re.compile("[^,]*," * col_num + "|".join(map(str, codes)) + ".*") for L in open('infogroup_bus_2010.csv'): if expr.match(L): print L Note that this is just a simple sketch as no escaping is considered... if the SNACODE column is not the first one and preceding fields may contain a comma you need a more sophisticated regexp like: ... '([^"][^,]*,|"([^"]|"")*",)' * col_num + ... that ignores commas inside double-quotes
You can probably make your pandas solution much faster: codes = [4451, 4452, 447, 772, 45299, 45291, 45212] codes = [str(code) for code in codes] sna = pd.read_csv('infogroup_bus_2010.csv', usecols=['SNACODE'], chunksize=int(1e6), dtype={'SNACODE': str}) with open('grocery.csv', 'w') as fout: for chunk in sna: for code in chunk['SNACODE']: for target_code in codes: if code.startswith(target_code): fout.write('{}\n'.format(code)) Read only the needed column with usecols=['SNACODE']. You can adjust the chunk size with chunksize=int(1e6). Depending on your RAM you can likely make it much bigger.
Data analysis for inconsistent string formatting
I have this task that I've been working on, but am having extreme misgivings about my methodology. So the problem is that I have a ton of excel files that are formatted strangely (and not consistently) and I need to extract certain fields for each entry. An example data set is My original approach was this: Export to csv Separate into counties Separate into districts Analyze each district individually, pull out values write to output.csv The problem I've run into is that the format (seemingly well organized) is almost random across files. Each line contains the same fields, but in a different order, spacing, and wording. I wrote a script to correctly process one file, but it doesn't work on any other files. So my question is, is there a more robust method of approaching this problem rather than simple string processing? What I had in mind was more of a fuzzy logic approach for trying to pin which field an item was, which could handle the inputs being a little arbitrary. How would you approach this problem? If it helps clear up the problem, here is the script I wrote: # This file takes a tax CSV file as input # and separates it into counties # then appends each county's entries onto # the end of the master out.csv # which will contain everything including # taxes, bonds, etc from all years #import the data csv import sys import re import csv def cleancommas(x): toggle=False for i,j in enumerate(x): if j=="\"": toggle=not toggle if toggle==True: if j==",": x=x[:i]+" "+x[i+1:] return x def districtatize(x): #list indexes of entries starting with "for" or "to" of length >5 indices=[1] for i,j in enumerate(x): if len(j)>2: if j[:2]=="to": indices.append(i) if len(j)>3: if j[:3]==" to" or j[:3]=="for": indices.append(i) if len(j)>5: if j[:5]==" \"for" or j[:5]==" \'for": indices.append(i) if len(j)>4: if j[:4]==" \"to" or j[:4]==" \'to" or j[:4]==" for": indices.append(i) if len(indices)==1: return [x[0],x[1:len(x)-1]] new=[x[0],x[1:indices[1]+1]] z=1 while z<len(indices)-1: new.append(x[indices[z]+1:indices[z+1]+1]) z+=1 return new #should return a list of lists. First entry will be county #each successive element in list will be list by district def splitforstos(string): for itemind,item in enumerate(string): # take all exception cases that didn't get processed splitfor=re.split('(?<=\d)\s\s(?=for)',item) # correctly and split them up so that the for begins splitto=re.split('(?<=\d)\s\s(?=to)',item) # a cell if len(splitfor)>1: print "\n\n\nfor detected\n\n" string.remove(item) string.insert(itemind,splitfor[0]) string.insert(itemind+1,splitfor[1]) elif len(splitto)>1: print "\n\n\nto detected\n\n" string.remove(item) string.insert(itemind,splitto[0]) string.insert(itemind+1,splitto[1]) def analyze(x): #input should be a string of content #target values are nomills,levytype,term,yearcom,yeardue clean=cleancommas(x) countylist=clean.split(',') emptystrip=filter(lambda a: a != '',countylist) empt2strip=filter(lambda a: a != ' ', emptystrip) singstrip=filter(lambda a: a != '\' \'',empt2strip) quotestrip=filter(lambda a: a !='\" \"',singstrip) splitforstos(quotestrip) distd=districtatize(quotestrip) print '\n\ndistrictized\n\n',distd county = distd[0] for x in distd[1:]: if len(x)>8: district=x[0] vote1=x[1] votemil=x[2] spaceindex=[m.start() for m in re.finditer(' ', votemil)][-1] vote2=votemil[:spaceindex] mills=votemil[spaceindex+1:] votetype=x[4] numyears=x[6] yearcom=x[8] yeardue=x[10] reason=x[11] data = [filename,county,district, vote1, vote2, mills, votetype, numyears, yearcom, yeardue, reason] print "data",data else: print "x\n\n",x district=x[0] vote1=x[1] votemil=x[2] spaceindex=[m.start() for m in re.finditer(' ', votemil)][-1] vote2=votemil[:spaceindex] mills=votemil[spaceindex+1:] votetype=x[4] special=x[5] splitspec=special.split(' ') try: forind=[i for i,j in enumerate(splitspec) if j=='for'][0] numyears=splitspec[forind+1] yearcom=splitspec[forind+6] except: forind=[i for i,j in enumerate(splitspec) if j=='commencing'][0] numyears=None yearcom=splitspec[forind+2] yeardue=str(x[6])[-4:] reason=x[7] data = [filename,county,district,vote1,vote2,mills,votetype,numyears,yearcom,yeardue,reason] print "data other", data openfile=csv.writer(open('out.csv','a'),delimiter=',', quotechar='|',quoting=csv.QUOTE_MINIMAL) openfile.writerow(data) # call the file like so: python tax.py 2007May8Tax.csv filename = sys.argv[1] #the file is the first argument f=open(filename,'r') contents=f.read() #entire csv as string #find index of every instance of the word county separators=[m.start() for m in re.finditer('\w+\sCOUNTY',contents)] #alternative implementation in regex # split contents into sections by county # analyze each section and append to out.csv for x,y in enumerate(separators): try: data = contents[y:separators[x+1]] except: data = contents[y:] analyze(data)
is there a more robust method of approaching this problem rather than simple string processing? Not really. What I had in mind was more of a fuzzy logic approach for trying to pin which field an item was, which could handle the inputs being a little arbitrary. How would you approach this problem? After a ton of analysis and programming, it won't be significantly better than what you've got. Reading stuff prepared by people requires -- sadly -- people-like brains. You can mess with NLTK to try and do a better job, but it doesn't work out terribly well either. You don't need a radically new approach. You need to streamline the approach you have. For example. district=x[0] vote1=x[1] votemil=x[2] spaceindex=[m.start() for m in re.finditer(' ', votemil)][-1] vote2=votemil[:spaceindex] mills=votemil[spaceindex+1:] votetype=x[4] numyears=x[6] yearcom=x[8] yeardue=x[10] reason=x[11] data = [filename,county,district, vote1, vote2, mills, votetype, numyears, yearcom, yeardue, reason] print "data",data Might be improved by using a named tuple. Then build something like this. data = SomeSensibleName( district= x[0], vote1=x[1], ... etc. ) So that you're not creating a lot of intermediate (and largely uninformative) loose variables. Also, keep looking at your analyze function (and any other function) to pull out the various "pattern matching" rules. The idea is that you'll examine a county's data, step through a bunch of functions until one matches the pattern; this will also create the named tuple. You want something like this. for p in ( some, list, of, functions ): match= p(data) if match: return match Each function either returns a named tuple (because it liked the row) or None (because it didn't like the row).