Handle ValueError while creating date in pd - python-3.x
I'm reading a csv file with p, day, month, and put it in a df. The goal is to create a date from day, month, current year, and I run into this error for 29th of Feb:
ValueError: cannot assemble the datetimes: day is out of range for month
I would like when this error occurs, to replace the day by the day before. How can we do that? Below are few lines of my pd and datex at the end is what I would like to get
p day month year datex
0 p1 29 02 2021 28Feb-2021
1 p2 18 07 2021 18Jul-2021
2 p3 12 09 2021 12Sep-2021
Right now, my code for the date is only the below, so I have nan where the date doesn't exist.
df['datex'] = pd.to_datetime(df[['year', 'month', 'day']], errors='coerce')
You could try something like this :
df['datex'] = pd.to_datetime(df[['year', 'month', 'day']], errors='coerce')
Indeed, you get NA :
p day year month datex
0 p1 29 2021 2 NaT
1 p2 18 2021 7 2021-07-18
2 p3 12 2021 9 2021-09-12
You could then make a particular case for these NA :
df.loc[df.datex.isnull(), 'previous_day'] = df.day -1
p day year month datex previous_day
0 p1 29 2021 2 NaT 28.0
1 p2 18 2021 7 2021-07-18 NaN
2 p3 12 2021 9 2021-09-12 NaN
df.loc[df.datex.isnull(), 'datex'] = pd.to_datetime(df[['previous_day', 'year', 'month']].rename(columns={'previous_day': 'day'}))
p day year month datex previous_day
0 p1 29 2021 2 2021-02-28 28.0
1 p2 18 2021 7 2021-07-18 NaN
2 p3 12 2021 9 2021-09-12 NaN
You have to create a new day column if you want to keep day = 29 in the day column.
Related
Pandas : Finding correct time window
I have a pandas dataframe which gets updated every hour with latest hourly data. I have to filter out IDs based upon a threshold, i.e. PR_Rate > 50 and CNT_12571 < 30 for 3 consecutive hours from a lookback period of 5 hours. I was using the below statements to accomplish this: df_thld=df[(df['Date'] > df['Date'].max() - pd.Timedelta(hours=5))& (df.PR_Rate>50) & (df.CNT_12571 < 30)] df_thld.loc[:,'HR_CNT'] = df_thld.groupby('ID')['Date'].nunique().to_frame('HR_CNT').reset_index() df_thld[(df_thld['HR_CNT'] >3] The problem with this approach is that since lookback period requirement is 5 hours, so, this HR_CNT can count any non consecutive hours breaching this critieria. MY Dataset is as below: DataFrame Date IDs CT_12571 PR_Rate 16/06/2021 10:00 A1 15 50.487 16/06/2021 11:00 A1 31 40.806 16/06/2021 12:00 A1 25 52.302 16/06/2021 13:00 A1 13 61.45 16/06/2021 14:00 A1 7 73.805 In the above Dataframe, threshold was not breached at 1100 hrs, but while counting the hours, 10,12 and 13 as the hours that breached the threshold instead of 12,13,14 as required. Each id may or may not have this critieria breached in a single day. Any idea, How can I fix this issue?
Please excuse me, if I have misinterpreted your problem. As I understand the issues you have a dataframe which is updated hourly. An example of this dataframe is illustrated below as df. From this dataframe, you want to filter only those rows which satisfy the following two conditions: PR_Rate > 50 and CNT_12571 < 30 If and only if the threshold is surpassed for three consecutive hours Given these assumptions, I would proceed as follows: df: Date IDs CT_1257 PR_Rate 0 2021-06-16 10:00:00 A1 15 50.487 1 2021-06-16 12:00:00 A1 31 40.806 2 2021-06-16 14:00:00 A1 25 52.302 3 2021-06-16 15:00:00 A1 13 61.450 4 2021-06-16 16:00:00 A1 7 73.805 Note in this dataframe, the only time fr5ame which satisfies the above conditions is the entries for the of 14:00, 15:00 and 16:00. def filterFrame(df, dur, pr_threshold, ct_threshold): ff = df[(df['CT_1257']< ct_threshold) & (df['PR_Rate'] >pr_threshold) ].reset_index() ml = list(ff.rolling(f'{dur}h', on='Date').count()['IDs']) r = len(ml)- 1 rows= [] while r >= 0: end = r start = None if int(ml[r]) < dur: r -= 1 else: k = int(ml[r]) for i in range(k): rows.append(r-i) r -= k rows = rows[::-1] return ff.filter(items= rows, axis = 0).reset_index() running filterFrame(df, 3, 50, 30) yields: level_0 index Date IDs CT_1257 PR_Rate 0 1 2 2021-06-16 14:00:00 A1 25 52.302 1 2 3 2021-06-16 15:00:00 A1 13 61.450 2 3 4 2021-06-16 16:00:00 A1 7 73.805
How to sum by month in timestamp Data Frame?
i have dataframe like this : trx_date trx_amount 2013-02-11 35 2014-03-10 26 2011-02-9 10 2013-02-12 5 2013-01-11 21 how do i filter that into month and year? so that i can sum the trx_amount example expected output : trx_monthly trx_sum 2013-02 40 2013-01 21 2014-02 35
You can convert values to month periods by Series.dt.to_period and then aggregate sum: df['trx_date'] = pd.to_datetime(df['trx_date']) df1 = (df.groupby(df['trx_date'].dt.to_period('m').rename('trx_monthly'))['trx_amount'] .sum() .reset_index(name='trx_sum')) print (df1) trx_monthly trx_sum 0 2011-02 10 1 2013-01 21 2 2013-02 40 3 2014-03 26 Or convert datetimes to strings in format YYYY-MM by Series.dt.strftime: df2 = (df.groupby(df['trx_date'].dt.strftime('%Y-%m').rename('trx_monthly'))['trx_amount'] .sum() .reset_index(name='trx_sum')) print (df2) trx_monthly trx_sum 0 2011-02 10 1 2013-01 21 2 2013-02 40 3 2014-03 26 Or convert to month and years, then output is different - 3 columns: df2 = (df.groupby([df['trx_date'].dt.year.rename('year'), df['trx_date'].dt.month.rename('month')])['trx_amount'] .sum() .reset_index(name='trx_sum')) print (df2) year month trx_sum 0 2011 2 10 1 2013 1 21 2 2013 2 40 3 2014 3 26
You can try this - df['trx_month'] = df['trx_date'].dt.month df_agg = df.groupby('trx_month')['trx_sum'].sum()
Find earliest date within daterange
I have the following market data: data = pd.DataFrame({'year': [2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020], 'month': [10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11], 'day': [1,2,5,6,7,8,9,12,13,14,15,16,19,20,21,22,23,26,27,28,29,30,2,3,5,6,9,10,11,12,13,16,17,18,19,20,23,24,25,26,27,30]}) data['date'] = pd.to_datetime(data) data['spot'] = [77.3438,78.192,78.1044,78.4357,78.0285,77.3507,76.78,77.13,77.0417,77.6525,78.0906,77.91,77.6602,77.3568,76.7243,76.5872,76.1374,76.4435,77.2906,79.2239,78.8993,79.5305,80.5313,79.3615,77.0156,77.4226,76.288,76.5648,77.1171,77.3568,77.374,76.1758,76.2325,76.0401,76.0529,76.1992,76.1648,75.474,75.551,75.7018,75.8639,76.3944] data = data.set_index('date') I'm trying to find the spot value for the first day of the month in the date column. I can find the first business day with below: def get_month_beg(d): month_beg = (d.index + pd.offsets.BMonthEnd(0) - pd.offsets.MonthBegin(normalize=True)) return month_beg data['month_beg'] = get_month_beg(data) However, due to data issues, sometimes the earliest date from my data does not match up with the first business day of the month. We'll call the earliest spot value of each month the "strike", which is what I'm trying to find. So for October, the spot value would be 77.3438 (10/1/21) and in Nov it would be 80.5313 (which is on 11/2/21 NOT 11/1/21). I tried below, which only works if my data's earliest date matches up with the first business date of the month (eg it works in Oct, but not in Nov) data['strike'] = data.month_beg.map(data.spot) As you can see, I get NaN in Nov because the first business day in my data is 11/2 (spot rate 80.5313) not 11/1. Does anyone know how to find the earliest date within a date range (in this case the earliest date of each month)? I was hoping the final df would like like below: data = pd.DataFrame({'year': [2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020], 'month': [10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11], 'day': [1,2,5,6,7,8,9,12,13,14,15,16,19,20,21,22,23,26,27,28,29,30,2,3,5,6,9,10,11,12,13,16,17,18,19,20,23,24,25,26,27,30]}) data['date'] = pd.to_datetime(data) data['spot'] = [77.3438,78.192,78.1044,78.4357,78.0285,77.3507,76.78,77.13,77.0417,77.6525,78.0906,77.91,77.6602,77.3568,76.7243,76.5872,76.1374,76.4435,77.2906,79.2239,78.8993,79.5305,80.5313,79.3615,77.0156,77.4226,76.288,76.5648,77.1171,77.3568,77.374,76.1758,76.2325,76.0401,76.0529,76.1992,76.1648,75.474,75.551,75.7018,75.8639,76.3944] data['strike'] = [77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,77.3438,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313,80.5313] data = data.set_index('date')
I Believe, We can get the first() for every year and month combination and later on join that with main data. data2=data.groupby(['year','month']).first().reset_index() #join data 2 with data based on month and year later on year month day spot 0 2020 10 1 77.3438 1 2020 11 2 80.5313 Based on the question, What i have understood is that we need to take every month's first day and respective 'SPOT' column value. Correct me if i have understood it wrong.
Strike = Spot value from first day of each month To do this, we need to do the following: Step 1. Get the Year/Month value from the Date column. Alternate, we can use Year and Month columns you already have in the DataFrame. Step 2: We need to groupby Year and Month. That will give all the records by Year+Month. From this, we need to get the first record (which will be the earliest date of the month). The earliest date can either be 1st or 2nd or 3rd of the month depending on the data in the column. Step 3: By using transform in Groupby, pandas will send back the results to match the dataframe length. So for each record, it will send the same result. In this example, we have only 2 months (Oct & Nov). However, we have 42 rows. Transform will send us back 42 rows. The code: groupby('[year','month'])['date'].transform('first') will give first day of month. Use This: data['dy'] = data.groupby(['year','month'])['date'].transform('first') or: data['dx'] = data.date.dt.to_period('M') #to get yyyy-mm value Step 4: Using transform, we can also get the Spot value. This can be assigned to Strike giving us the desired result. Instead of getting first day of the month, we can change it to return Spot value. The code will be: groupby('date')['spot'].transform('first') Use this: data['strike'] = data.groupby(['year','month'])['spot'].transform('first') or data['strike'] = data.groupby('dx')['spot'].transform('first') Putting all this together The full code to get Strike Price using Spot Price from first day of month import pandas as pd import numpy as np data = pd.DataFrame({'year': [2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020], 'month': [10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11], 'day': [1,2,5,6,7,8,9,12,13,14,15,16,19,20,21,22,23,26,27,28,29,30,2,3,5,6,9,10,11,12,13,16,17,18,19,20,23,24,25,26,27,30]}) data['date'] = pd.to_datetime(data) data['spot'] = [77.3438,78.192,78.1044,78.4357,78.0285,77.3507,76.78,77.13,77.0417,77.6525,78.0906,77.91,77.6602,77.3568,76.7243,76.5872,76.1374,76.4435,77.2906,79.2239,78.8993,79.5305,80.5313,79.3615,77.0156,77.4226,76.288,76.5648,77.1171,77.3568,77.374,76.1758,76.2325,76.0401,76.0529,76.1992,76.1648,75.474,75.551,75.7018,75.8639,76.3944] #Pick the first day of month Spot price as the Strike price data['strike'] = data.groupby(['year','month'])['spot'].transform('first') #This will give you the first row of each month print (data) The output of this will be: year month day date spot strike 0 2020 10 1 2020-10-01 77.3438 77.3438 1 2020 10 2 2020-10-02 78.1920 77.3438 2 2020 10 5 2020-10-05 78.1044 77.3438 3 2020 10 6 2020-10-06 78.4357 77.3438 4 2020 10 7 2020-10-07 78.0285 77.3438 5 2020 10 8 2020-10-08 77.3507 77.3438 6 2020 10 9 2020-10-09 76.7800 77.3438 7 2020 10 12 2020-10-12 77.1300 77.3438 8 2020 10 13 2020-10-13 77.0417 77.3438 9 2020 10 14 2020-10-14 77.6525 77.3438 10 2020 10 15 2020-10-15 78.0906 77.3438 11 2020 10 16 2020-10-16 77.9100 77.3438 12 2020 10 19 2020-10-19 77.6602 77.3438 13 2020 10 20 2020-10-20 77.3568 77.3438 14 2020 10 21 2020-10-21 76.7243 77.3438 15 2020 10 22 2020-10-22 76.5872 77.3438 16 2020 10 23 2020-10-23 76.1374 77.3438 17 2020 10 26 2020-10-26 76.4435 77.3438 18 2020 10 27 2020-10-27 77.2906 77.3438 19 2020 10 28 2020-10-28 79.2239 77.3438 20 2020 10 29 2020-10-29 78.8993 77.3438 21 2020 10 30 2020-10-30 79.5305 77.3438 22 2020 11 2 2020-11-02 80.5313 80.5313 23 2020 11 3 2020-11-03 79.3615 80.5313 24 2020 11 5 2020-11-05 77.0156 80.5313 25 2020 11 6 2020-11-06 77.4226 80.5313 26 2020 11 9 2020-11-09 76.2880 80.5313 27 2020 11 10 2020-11-10 76.5648 80.5313 28 2020 11 11 2020-11-11 77.1171 80.5313 29 2020 11 12 2020-11-12 77.3568 80.5313 30 2020 11 13 2020-11-13 77.3740 80.5313 31 2020 11 16 2020-11-16 76.1758 80.5313 32 2020 11 17 2020-11-17 76.2325 80.5313 33 2020 11 18 2020-11-18 76.0401 80.5313 34 2020 11 19 2020-11-19 76.0529 80.5313 35 2020 11 20 2020-11-20 76.1992 80.5313 36 2020 11 23 2020-11-23 76.1648 80.5313 37 2020 11 24 2020-11-24 75.4740 80.5313 38 2020 11 25 2020-11-25 75.5510 80.5313 39 2020 11 26 2020-11-26 75.7018 80.5313 40 2020 11 27 2020-11-27 75.8639 80.5313 41 2020 11 30 2020-11-30 76.3944 80.5313 Previous Answer to get the first day of each month (within the column data) One way to do it is to create a dummy column to store the first day of each month. Then use drop_duplicates() and retain only the first row. Key assumption: The assumption with this logic is that we have at least 2 rows for each month. If there is only one row for a month, then it will not be part of the duplicates and you will NOT get that month's data. That will give you the first day of each month. import pandas as pd import numpy as np data = pd.DataFrame({'year': [2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020,2020], 'month': [10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,10,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11,11], 'day': [1,2,5,6,7,8,9,12,13,14,15,16,19,20,21,22,23,26,27,28,29,30,2,3,5,6,9,10,11,12,13,16,17,18,19,20,23,24,25,26,27,30]}) data['date'] = pd.to_datetime(data) data['spot'] = [77.3438,78.192,78.1044,78.4357,78.0285,77.3507,76.78,77.13,77.0417,77.6525,78.0906,77.91,77.6602,77.3568,76.7243,76.5872,76.1374,76.4435,77.2906,79.2239,78.8993,79.5305,80.5313,79.3615,77.0156,77.4226,76.288,76.5648,77.1171,77.3568,77.374,76.1758,76.2325,76.0401,76.0529,76.1992,76.1648,75.474,75.551,75.7018,75.8639,76.3944] #create a dummy column to store the first day of the month data['dx'] = data.date.dt.to_period('M') #drop duplicates while retaining only the first row of each month dx = data.drop_duplicates('dx',keep='first') #This will give you the first row of each month print (dx) The output of this will be: year month day date spot dx 0 2020 10 1 2020-10-01 77.3438 2020-10 22 2020 11 2 2020-11-02 80.5313 2020-11 If there is only one row for a given month, then you can use groupby the month and take the first record. data.groupby(['dx']).first() This will give you: year month day date spot dx 2020-10 2020 10 1 2020-10-01 77.3438 2020-11 2020 11 2 2020-11-02 80.5313
data['strike']=data.groupby(['year','month'])['spot'].transform('first') I guess this can be achieved by this without creating any other dataframe.
Pandas: How to ctrate DateTime index
There is Pandas Dataframe as: year month count 0 2014 Jan 12 1 2014 Feb 10 2 2015 Jan 12 3 2015 Feb 10 How to create DateTime index from 'year' and 'month',so result would be : count 2014.01.31 12 2014.02.28 10 2015.01.31 12 2015.02.28 10
Use to_datetime with DataFrame.pop for use and remove columns and add offsets.MonthEnd: dates = pd.to_datetime(df.pop('year').astype(str) + df.pop('month'), format='%Y%b') df.index = dates + pd.offsets.MonthEnd() print (df) count 2014-01-31 12 2014-02-28 10 2015-01-31 12 2015-02-28 10 Or: dates = pd.to_datetime(df.pop('year').astype(str) + df.pop('month'), format='%Y%b') df.index = dates + pd.to_timedelta(dates.dt.daysinmonth - 1, unit='d') print (df) count 2014-01-31 12 2014-02-28 10 2015-01-31 12 2015-02-28 10
how to Get week number from specified year date in python?
I have a time-series data and i want to get the week number from the initial date date 20180401 20180402 20180902 20190130 20190401 Things Tried Code df["date"]= pd.to_datetime(df.date,format='%Y%m%d') df["week_no"]= df.date.dt.week But the week getting reset in 2019 results in getting a common week number of 2018. is there anything we can do in it ??
You can use this function that will calculate the difference between two days in weeks: def Wdiff(fromdate, todate): d = pd.to_datetime(todate) - pd.to_datetime(fromdate) return int(d / np.timedelta64(1, 'W'))
You can create a datetime object with the specified date, then retrieve the week number using the isocalendar method: import datetime myDate = datetime.date(2018, 4, 1) week = myDate.isocalendar()[1] print(week) You could then calculate the total number of remaining weeks in 2018, then add the total number of weeks in each year in between, and finally add the week number of the current date. For example, this code would print the number of weeks from the 1st of April 2018 to the 6th May 2020: import datetime myDate = datetime.date(2018, 4, 1) currentDate = datetime.date(2020, 5, 6) weeks = datetime.date(myDate.year, 12, 28).isocalendar()[1] - myDate.isocalendar()[1] for i in range(myDate.year, currentDate.year): weeks += datetime.date(i, 12, 28).isocalendar()[1] weeks += currentDate.isocalendar()[1] print(weeks) Note that because of the way isocalendar works, the 28th of December will always be in the last week of the given year. The ISO year consists of 52 or 53 full weeks, and where a week starts on a Monday and ends on a Sunday. The first week of an ISO year is the first (Gregorian) calendar week of a year containing a Thursday. This is called week number 1, and the ISO year of that Thursday is the same as its Gregorian year. You can get more information about isocalendar here: https://docs.python.org/3/library/datetime.html
To get the week number, but as a 2-digit string (with leading zero), you can run: df['week_no'] = df.date.dt.strftime('%W') The result, for slightly extended source data is: date week_no 0 2018-04-01 13 1 2018-04-02 14 2 2018-09-02 35 3 2018-12-30 52 4 2018-12-31 53 5 2019-01-01 00 6 2019-01-02 00 7 2019-01-03 00 8 2019-01-04 00 9 2019-01-05 00 10 2019-01-06 00 11 2019-01-07 01 12 2019-01-30 04 13 2019-04-01 13 Note that the last day of 2018 (monday) has week No == 53 and "initial" days in 2019 (up to 2019-01-06 - Sunday) have week No == 00. If you want this column as int, append .astype(int) to the above code.