-
Hello guys! Firstly, i want to really thank you for your work!. I have a big dataset which looks like that: I apply filtering at certain LAT and LON to choose the area (geo window) which i am interested. The filtering is done using dictionary like this: list_of_values = [23. , 23.05, 23.1 , 23.15, 23.2 , 23.25, 23.3 , 23.35, 23.4 , vaex_df_sparse_crete = vaex_df_sparse[vaex_df_sparse.LON.isin(list_of_values)] list_of_values = [34. , 34.05, 34.1 , 34.15, 34.2 , 34.25, 34.3 , 34.35, 34.4 , vaex_df_sparse_crete = vaex_df_sparse_crete[vaex_df_sparse_crete.LAT.isin(list_of_values)] After the filtering is implemented i am seeing that some dates (hourly sampling) are missing which is pretty weird because the data are provided by a software. Do you see anything wrong in the way i am treating the dataset;; Is there a better way to apply filtering;; Do you think that maybe after saving the filtered dataset an error is happening; Kind Regards, GK |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
It looks like you could keep your filtering simpler? min_lat = 34.0
max_lat = 36.5
min_lon = 23.0
max_lon = 27.5
filtered = df[
(df.LAT >= min_lat) &
(df.LAT <= max_lat) &
(df.LON >= min_lon) &
(df.LON <= max_lon) &
] Why don't you do the filtering, and from there you can see which rows indices are looking abnormal, and then you can go back to the original unfiltered data and find those rows and see if they look weird before? |
Beta Was this translation helpful? Give feedback.
It looks like you could keep your filtering simpler?
Why don't you do the filtering, and from there you can see which rows indices are looking abnormal, and then you can go back to the original unfiltered data and find those rows and see if they look weird before?