python /home/admin/mtr/script_for_cron.py -j python_test3 -m 12 -a ' --short_python3 -v ' -s python_test3 -M 0 -S 0 -U 100,100,120 import MySQLdb succeeded Import error (python version) python version = 3 warning , we can't find thcl infos in json_data warning , we can't find pdt infos in json_data list_job_run_as_list : ['mask_detection', 'datou', 'CacheModelData_queries', 'CachePhotoData_queries', 'test_fork', 'prepare_maskdata', 'portfolio_queries', 'sla_mensuel'] python version used : 3 liste_fichiers : [('tests/mask_test', True, 'Test mask-detection ', 'mask_detection'), ('tests/datou_test', True, 'Datou All Test', 'datou', 'all'), ('mtr/database_queries/CacheModelData_queries', True, 'Test Cache Model Data', 'CacheModelData_queries'), ('tests/cache_photo_data_test', True, 'Test local_cache_photo ', 'CachePhotoData_queries'), ('mtr/mask_rcnn/prepare_maskdata', True, 'test prepare mask data', 'prepare_maskdata', 'all'), ('mtr/database_queries/portfolio_queries', True, 'test portfolio queries', 'portfolio_queries'), ('prod/memo/memo', True, 'SLA Mensuel', 'sla_mensuel', 'all')] #&_# BEGIN OF TEST : tests/mask_test #&_# /home/admin/workarea/git/Velours/python/tests/mask_test.py Test mask-detection python version used : 3 ############################### TEST memory used ################################ free memory at begining : begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 7218 run mask_detect Inside batchDatouExec : verbose : False # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! List Step Type Loaded in datou : mask_detect list_input_json : [] origin BFwe have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 time to download the photos : 0.12707877159118652 About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : False number of steps : 1 step1:mask_detect Fri May 30 03:35:27 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Beginning of datou step mask_detect ! save_polygon : True begin detect begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 7218 max_wait_temp : 1 max_wait : 0 gpu_flag : 0 /home/admin/workarea/git/Velours/python/tests/python_tests.py:11: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses import imp 2025-05-30 03:35:30.316378: I tensorflow/core/platform/cpu_feature_guard.cc:143] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA 2025-05-30 03:35:30.343049: I tensorflow/core/platform/profile_utils/cpu_utils.cc:102] CPU Frequency: 3493065000 Hz 2025-05-30 03:35:30.345093: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7fbc40000b60 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:35:30.345151: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2025-05-30 03:35:30.348620: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcuda.so.1 2025-05-30 03:35:30.586204: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x1ba6ff00 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:35:30.586240: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA GeForce RTX 2080 Ti, Compute Capability 7.5 2025-05-30 03:35:30.587117: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:30.587346: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:30.589051: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:30.590648: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:30.590922: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:30.592765: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:30.593753: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:30.597514: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:30.598665: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:30.598720: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:30.599380: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:35:30.599396: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:35:30.599405: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:35:30.600574: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 6642 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) WARNING:tensorflow:From /home/admin/workarea/git/Velours/python/mtr/mask_rcnn/mask_detection.py:69: The name tf.keras.backend.set_session is deprecated. Please use tf.compat.v1.keras.backend.set_session instead. Inside mask_sub_process Inside mask_detect About to load cache.load_thcl_param To do loadFromThcl(), then load ParamDescType : thcl454 thcls : [{'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}] thcl {'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'} Update svm_hashtag_type_desc : 3473 FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3473, 'mask_coco_origin', 16384, 25088, 'mask_coco_origin', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2018, 3, 19, 10, 42, 21), datetime.datetime(2018, 3, 19, 10, 42, 21)) {'thcl': {'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}, 'list_hashtags': ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'], 'list_hashtags_csv': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'svm_hashtag_type_desc': 3473, 'photo_desc_type': 3473, 'pb_hashtag_id_or_classifier': 0} list_class_names : ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] Configurations: BACKBONE resnet101 BACKBONE_SHAPES [[160 160] [ 80 80] [ 40 40] [ 20 20] [ 10 10]] BACKBONE_STRIDES [4, 8, 16, 32, 64] BATCH_SIZE 1 BBOX_STD_DEV [0.1 0.1 0.2 0.2] DETECTION_MAX_INSTANCES 100 DETECTION_MIN_CONFIDENCE 0.3 DETECTION_NMS_THRESHOLD 0.3 GPU_COUNT 1 IMAGES_PER_GPU 1 IMAGE_MAX_DIM 640 IMAGE_MIN_DIM 640 IMAGE_PADDING True IMAGE_SHAPE [640 640 3] LEARNING_MOMENTUM 0.9 LEARNING_RATE 0.001 LOSS_WEIGHTS {'rpn_class_loss': 1.0, 'rpn_bbox_loss': 1.0, 'mrcnn_class_loss': 1.0, 'mrcnn_bbox_loss': 1.0, 'mrcnn_mask_loss': 1.0} MASK_POOL_SIZE 14 MASK_SHAPE [28, 28] MAX_GT_INSTANCES 100 MEAN_PIXEL [123.7 116.8 103.9] MINI_MASK_SHAPE (56, 56) NAME mask_coco_origin NUM_CLASSES 81 POOL_SIZE 7 POST_NMS_ROIS_INFERENCE 1000 POST_NMS_ROIS_TRAINING 2000 ROI_POSITIVE_RATIO 0.33 RPN_ANCHOR_RATIOS [0.5, 1, 2] RPN_ANCHOR_SCALES (16, 32, 64, 128, 256) RPN_ANCHOR_STRIDE 1 2025-05-30 03:35:31.255279: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:31.255355: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:31.255376: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:31.255394: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:31.255412: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:31.255430: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:31.255460: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:31.255479: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:31.256782: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:31.257898: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:31.257932: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:31.257951: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:31.257968: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:31.257985: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:31.258002: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:31.258019: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:31.258037: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:31.259348: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:31.259381: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:35:31.259391: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:35:31.259400: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:35:31.260732: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 6642 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) Using TensorFlow backend. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:396: calling crop_and_resize_v1 (from tensorflow.python.ops.image_ops_impl) with box_ind is deprecated and will be removed in a future version. Instructions for updating: box_ind is deprecated, use box_indices instead WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:703: to_int32 (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:729: to_float (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. RPN_BBOX_STD_DEV [0.1 0.1 0.2 0.2] RPN_NMS_THRESHOLD 0.7 RPN_TRAIN_ANCHORS_PER_IMAGE 256 STEPS_PER_EPOCH 1000 TRAIN_ROIS_PER_IMAGE 200 USE_MINI_MASK True USE_RPN_ROIS True VALIDATION_STEPS 50 WEIGHT_DECAY 0.0001 model_param file didn't exist model_name : mask_coco_origin model_type : mask_rcnn list file need : ['mask_model.h5'] file exist in s3 : ['mask_model.h5'] file manque in s3 : [] 2025-05-30 03:35:38.088918: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:38.275515: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 local folder : /data/models_weight/mask_coco_origin /data/models_weight/mask_coco_origin/mask_model.h5 size_local : 257557808 size in s3 : 257557808 create time local : 2021-08-09 05:27:17 create time in s3 : 2021-08-06 19:45:17 mask_model.h5 already exist and didn't need to update list_images length : 1 NEW PHOTO Processing 1 images image shape: (480, 640, 3) min: 0.00000 max: 255.00000 molded_images shape: (1, 640, 640, 3) min: -123.70000 max: 151.10000 image_metas shape: (1, 89) min: 0.00000 max: 640.00000 nb d'objets trouves : 5 Detection mask done ! Trying to reset tf kernel 1113232 begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 1929 tf kernel not reseted sub process len(results) : 1 len(list_Values) 0 None max_time_sub_proc : 3600 parent process len(results) : 1 len(list_Values) 0 process is alive finish correctly or not : True after detect begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 7218 list_Values should be empty [] To do loadFromThcl(), then load ParamDescType : thcl454 Catched exception ! Connect or reconnect ! thcls : [{'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}] thcl {'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'} Update svm_hashtag_type_desc : 3473 ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] time for calcul the mask position with numpy : 0.00057220458984375 nb_pixel_total : 15551 time to create 1 rle with old method : 0.0174405574798584 length of segment : 256 time for calcul the mask position with numpy : 0.0026721954345703125 nb_pixel_total : 146494 time to create 1 rle with old method : 0.16016745567321777 length of segment : 374 time for calcul the mask position with numpy : 0.00023365020751953125 nb_pixel_total : 14254 time to create 1 rle with old method : 0.015879392623901367 length of segment : 151 time for calcul the mask position with numpy : 0.00011944770812988281 nb_pixel_total : 5613 time to create 1 rle with old method : 0.006912708282470703 length of segment : 48 time for calcul the mask position with numpy : 5.888938903808594e-05 nb_pixel_total : 1824 time to create 1 rle with old method : 0.002398967742919922 length of segment : 39 time spent for convertir_results : 0.9902560710906982 time spend for datou_step_exec : 17.212612867355347 time spend to save output : 6.413459777832031e-05 total time spend for step 1 : 17.212677001953125 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : False eke 12-6-18 : saveMask need to be cleaned for new output ! Number saved : None batch 1 Loaded 3336 chid ids of type : 445 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++Number RLEs to save : 0 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 time used for this insertion : 0.017175674438476562 save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {'957285035': [[(957285035, 492601069, 445, 0, 186, 22, 282, 0.99549, [(140, 26, 6), (135, 27, 15), (133, 28, 18), (131, 29, 22), (127, 30, 27), (10, 31, 1), (120, 31, 35), (8, 32, 13), (27, 32, 3), (115, 32, 41), (7, 33, 52), (109, 33, 48), (6, 34, 70), (103, 34, 55), (5, 35, 154), (4, 36, 155), (3, 37, 156), (3, 38, 156), (3, 39, 156), (2, 40, 157), (2, 41, 157), (2, 42, 157), (2, 43, 157), (2, 44, 157), (2, 45, 157), (1, 46, 158), (1, 47, 158), (1, 48, 158), (1, 49, 157), (1, 50, 157), (1, 51, 156), (1, 52, 156), (1, 53, 155), (1, 54, 154), (1, 55, 152), (1, 56, 149), (1, 57, 145), (1, 58, 141), (1, 59, 136), (1, 60, 133), (1, 61, 130), (1, 62, 127), (1, 63, 126), (1, 64, 124), (1, 65, 123), (1, 66, 121), (1, 67, 120), (1, 68, 118), (1, 69, 117), (1, 70, 116), (1, 71, 115), (1, 72, 114), (1, 73, 113), (1, 74, 112), (1, 75, 111), (1, 76, 110), (1, 77, 108), (1, 78, 108), (1, 79, 107), (1, 80, 106), (1, 81, 105), (2, 82, 104), (2, 83, 103), (2, 84, 103), (2, 85, 102), (2, 86, 102), (2, 87, 101), (2, 88, 100), (2, 89, 99), (2, 90, 99), (2, 91, 98), (2, 92, 97), (2, 93, 96), (2, 94, 95), (2, 95, 93), (2, 96, 91), (2, 97, 90), (2, 98, 89), (2, 99, 87), (2, 100, 86), (2, 101, 86), (2, 102, 85), (2, 103, 84), (2, 104, 83), (2, 105, 83), (2, 106, 82), (2, 107, 81), (2, 108, 80), (2, 109, 80), (2, 110, 79), (2, 111, 78), (2, 112, 77), (2, 113, 76), (1, 114, 76), (1, 115, 75), (1, 116, 74), (1, 117, 73), (1, 118, 72), (1, 119, 71), (1, 120, 71), (1, 121, 70), (1, 122, 69), (1, 123, 69), (1, 124, 68), (1, 125, 68), (1, 126, 67), (1, 127, 67), (1, 128, 66), (1, 129, 66), (1, 130, 66), (1, 131, 65), (1, 132, 65), (1, 133, 64), (1, 134, 63), (1, 135, 63), (1, 136, 62), (1, 137, 61), (1, 138, 60), (1, 139, 60), (1, 140, 59), (1, 141, 58), (1, 142, 58), (1, 143, 57), (1, 144, 56), (1, 145, 56), (1, 146, 55), (1, 147, 54), (1, 148, 54), (1, 149, 53), (1, 150, 52), (1, 151, 52), (1, 152, 51), (1, 153, 50), (1, 154, 49), (1, 155, 48), (1, 156, 47), (1, 157, 46), (1, 158, 45), (1, 159, 45), (1, 160, 44), (1, 161, 43), (1, 162, 42), (1, 163, 41), (1, 164, 41), (1, 165, 40), (1, 166, 40), (1, 167, 39), (1, 168, 38), (1, 169, 37), (1, 170, 36), (1, 171, 35), (1, 172, 34), (1, 173, 34), (1, 174, 33), (1, 175, 33), (1, 176, 32), (1, 177, 32), (1, 178, 32), (1, 179, 32), (1, 180, 31), (1, 181, 31), (1, 182, 31), (1, 183, 30), (1, 184, 30), (1, 185, 30), (1, 186, 29), (1, 187, 29), (1, 188, 29), (1, 189, 28), (1, 190, 28), (1, 191, 27), (1, 192, 27), (1, 193, 26), (1, 194, 26), (1, 195, 26), (1, 196, 26), (1, 197, 26), (1, 198, 26), (1, 199, 26), (1, 200, 25), (1, 201, 25), (1, 202, 25), (1, 203, 25), (1, 204, 25), (1, 205, 25), (1, 206, 25), (1, 207, 25), (1, 208, 25), (1, 209, 25), (1, 210, 25), (1, 211, 25), (1, 212, 25), (1, 213, 25), (1, 214, 25), (1, 215, 25), (1, 216, 25), (1, 217, 25), (1, 218, 25), (1, 219, 25), (1, 220, 24), (1, 221, 24), (1, 222, 24), (1, 223, 24), (1, 224, 24), (1, 225, 24), (1, 226, 25), (1, 227, 25), (1, 228, 25), (2, 229, 24), (2, 230, 24), (2, 231, 24), (2, 232, 23), (2, 233, 23), (2, 234, 23), (2, 235, 23), (2, 236, 23), (2, 237, 23), (2, 238, 23), (2, 239, 23), (2, 240, 23), (2, 241, 23), (2, 242, 23), (2, 243, 23), (2, 244, 23), (2, 245, 23), (2, 246, 23), (2, 247, 23), (2, 248, 23), (2, 249, 24), (2, 250, 24), (2, 251, 23), (2, 252, 23), (2, 253, 23), (2, 254, 23), (2, 255, 23), (2, 256, 23), (2, 257, 23), (2, 258, 23), (2, 259, 23), (2, 260, 23), (2, 261, 23), (3, 262, 22), (3, 263, 22), (3, 264, 22), (3, 265, 22), (4, 266, 21), (4, 267, 21), (5, 268, 20), (5, 269, 20), (6, 270, 19), (7, 271, 17), (8, 272, 16), (8, 273, 16), (9, 274, 13), (11, 275, 9), (15, 276, 2)], ['16,276,8,273,2,261,2,229,1,228,1,114,2,113,2,82,1,81,1,46,3,37,8,32,20,32,21,33,58,33,59,34,75,34,76,35,102,35,114,33,120,31,130,30,135,27,145,26,152,29,158,35,158,48,154,54,141,58,128,61,119,67,105,81,103,86,96,94,89,98,81,109,71,119,65,132,60,138,52,151,45,158,40,166,34,172,29,188,26,193,25,200,25,219,24,232,24,270,23,273']), (957285035, 492601069, 445, 28, 582, 19, 423, 0.99172103, [(302, 35, 37), (266, 36, 101), (250, 37, 129), (241, 38, 143), (200, 39, 189), (190, 40, 203), (186, 41, 217), (182, 42, 233), (178, 43, 243), (175, 44, 252), (172, 45, 262), (168, 46, 275), (164, 47, 289), (158, 48, 301), (155, 49, 310), (153, 50, 317), (151, 51, 323), (150, 52, 326), (148, 53, 329), (146, 54, 333), (145, 55, 335), (144, 56, 337), (143, 57, 339), (141, 58, 342), (140, 59, 345), (138, 60, 348), (136, 61, 352), (134, 62, 355), (132, 63, 359), (130, 64, 363), (128, 65, 366), (126, 66, 370), (124, 67, 373), (123, 68, 375), (121, 69, 378), (119, 70, 380), (117, 71, 383), (116, 72, 385), (115, 73, 387), (114, 74, 389), (112, 75, 392), (111, 76, 394), (110, 77, 396), (109, 78, 398), (109, 79, 399), (108, 80, 401), (107, 81, 403), (106, 82, 405), (105, 83, 407), (105, 84, 408), (104, 85, 410), (103, 86, 413), (102, 87, 415), (101, 88, 417), (100, 89, 419), (99, 90, 421), (98, 91, 422), (97, 92, 424), (95, 93, 427), (94, 94, 429), (93, 95, 431), (92, 96, 432), (91, 97, 434), (90, 98, 436), (90, 99, 436), (90, 100, 437), (90, 101, 438), (90, 102, 440), (91, 103, 440), (91, 104, 441), (91, 105, 443), (91, 106, 444), (91, 107, 445), (91, 108, 446), (92, 109, 446), (92, 110, 447), (92, 111, 448), (92, 112, 448), (92, 113, 449), (91, 114, 450), (90, 115, 452), (90, 116, 452), (89, 117, 454), (89, 118, 454), (88, 119, 456), (87, 120, 457), (87, 121, 457), (86, 122, 459), (85, 123, 461), (85, 124, 461), (84, 125, 463), (84, 126, 463), (83, 127, 465), (82, 128, 467), (81, 129, 469), (79, 130, 472), (78, 131, 474), (77, 132, 476), (75, 133, 479), (73, 134, 482), (72, 135, 484), (70, 136, 487), (69, 137, 489), (68, 138, 491), (66, 139, 494), (65, 140, 495), (64, 141, 497), (63, 142, 498), (62, 143, 500), (61, 144, 501), (60, 145, 502), (59, 146, 503), (58, 147, 505), (57, 148, 506), (57, 149, 506), (56, 150, 508), (56, 151, 508), (55, 152, 510), (55, 153, 510), (55, 154, 510), (54, 155, 512), (54, 156, 512), (53, 157, 513), (53, 158, 513), (52, 159, 514), (52, 160, 514), (51, 161, 516), (51, 162, 516), (50, 163, 517), (50, 164, 517), (49, 165, 518), (49, 166, 519), (48, 167, 520), (48, 168, 520), (47, 169, 521), (47, 170, 521), (47, 171, 521), (46, 172, 522), (46, 173, 522), (46, 174, 522), (45, 175, 522), (45, 176, 522), (45, 177, 522), (44, 178, 523), (44, 179, 523), (43, 180, 524), (43, 181, 524), (43, 182, 524), (42, 183, 524), (42, 184, 524), (42, 185, 524), (41, 186, 525), (40, 187, 525), (40, 188, 525), (39, 189, 526), (38, 190, 526), (38, 191, 526), (37, 192, 527), (37, 193, 526), (36, 194, 527), (36, 195, 527), (35, 196, 527), (35, 197, 527), (34, 198, 528), (34, 199, 528), (34, 200, 527), (34, 201, 527), (34, 202, 526), (33, 203, 527), (33, 204, 526), (33, 205, 525), (33, 206, 525), (33, 207, 524), (33, 208, 523), (32, 209, 524), (32, 210, 523), (32, 211, 522), (32, 212, 521), (32, 213, 520), (32, 214, 519), (32, 215, 518), (32, 216, 517), (32, 217, 516), (32, 218, 515), (32, 219, 514), (32, 220, 513), (32, 221, 513), (32, 222, 512), (31, 223, 512), (31, 224, 511), (31, 225, 511), (31, 226, 510), (31, 227, 509), (31, 228, 509), (31, 229, 508), (31, 230, 507), (31, 231, 505), (31, 232, 504), (31, 233, 502), (31, 234, 501), (31, 235, 499), (31, 236, 497), (31, 237, 495), (31, 238, 494), (31, 239, 493), (31, 240, 492), (31, 241, 491), (30, 242, 491), (30, 243, 490), (31, 244, 487), (31, 245, 486), (31, 246, 485), (31, 247, 483), (31, 248, 481), (31, 249, 479), (31, 250, 477), (31, 251, 475), (31, 252, 473), (31, 253, 472), (31, 254, 471), (31, 255, 469), (31, 256, 468), (31, 257, 467), (31, 258, 466), (31, 259, 465), (31, 260, 463), (31, 261, 461), (31, 262, 459), (31, 263, 457), (31, 264, 456), (31, 265, 454), (31, 266, 453), (31, 267, 452), (31, 268, 451), (31, 269, 450), (30, 270, 450), (30, 271, 449), (31, 272, 447), (31, 273, 446), (31, 274, 444), (31, 275, 443), (31, 276, 441), (32, 277, 438), (32, 278, 436), (32, 279, 434), (32, 280, 433), (33, 281, 430), (33, 282, 429), (33, 283, 427), (33, 284, 426), (34, 285, 424), (34, 286, 423), (34, 287, 421), (35, 288, 419), (35, 289, 417), (36, 290, 414), (36, 291, 412), (37, 292, 410), (37, 293, 408), (38, 294, 406), (38, 295, 405), (39, 296, 402), (40, 297, 400), (40, 298, 399), (41, 299, 397), (42, 300, 395), (42, 301, 394), (43, 302, 391), (44, 303, 389), (45, 304, 386), (46, 305, 384), (47, 306, 381), (48, 307, 379), (48, 308, 377), (49, 309, 375), (50, 310, 373), (51, 311, 371), (52, 312, 369), (53, 313, 367), (54, 314, 365), (55, 315, 363), (56, 316, 361), (58, 317, 358), (60, 318, 355), (63, 319, 350), (66, 320, 346), (69, 321, 342), (72, 322, 338), (75, 323, 334), (78, 324, 331), (80, 325, 328), (83, 326, 324), (84, 327, 323), (86, 328, 320), (88, 329, 318), (90, 330, 315), (92, 331, 313), (94, 332, 311), (97, 333, 307), (100, 334, 304), (103, 335, 301), (107, 336, 297), (110, 337, 293), (113, 338, 290), (116, 339, 287), (118, 340, 284), (120, 341, 282), (122, 342, 280), (123, 343, 279), (124, 344, 277), (126, 345, 275), (128, 346, 272), (130, 347, 270), (132, 348, 267), (134, 349, 265), (137, 350, 261), (141, 351, 257), (144, 352, 253), (147, 353, 250), (149, 354, 247), (151, 355, 244), (153, 356, 242), (155, 357, 239), (157, 358, 236), (158, 359, 235), (160, 360, 232), (162, 361, 230), (163, 362, 228), (165, 363, 226), (167, 364, 223), (169, 365, 221), (170, 366, 219), (172, 367, 217), (174, 368, 215), (176, 369, 212), (178, 370, 210), (179, 371, 209), (181, 372, 206), (182, 373, 205), (184, 374, 203), (186, 375, 200), (188, 376, 198), (190, 377, 196), (193, 378, 192), (195, 379, 190), (198, 380, 187), (200, 381, 184), (203, 382, 181), (205, 383, 179), (206, 384, 178), (208, 385, 175), (209, 386, 174), (211, 387, 172), (212, 388, 170), (214, 389, 167), (216, 390, 165), (219, 391, 161), (223, 392, 156), (227, 393, 152), (231, 394, 147), (235, 395, 142), (239, 396, 137), (243, 397, 132), (247, 398, 126), (250, 399, 122), (253, 400, 118), (256, 401, 114), (261, 402, 107), (265, 403, 102), (271, 404, 94), (279, 405, 84), (292, 406, 68), (303, 407, 54), (313, 408, 39)], ['351,408,313,408,260,401,212,388,190,377,146,352,134,349,115,338,94,332,56,316,42,301,31,276,31,223,34,198,42,185,47,169,57,148,84,126,92,113,90,98,117,71,163,48,190,40,240,39,266,36,338,35,378,37,393,41,414,42,434,46,473,51,495,66,539,111,546,126,559,139,567,166,566,182,561,199,555,209,537,230,515,246,485,264,476,273,445,292,425,307,408,323,401,343,388,366,382,387,366,403']), (957285035, 492601069, 445, 485, 636, 23, 174, 0.97112954, [(540, 24, 21), (626, 24, 3), (531, 25, 49), (594, 25, 40), (527, 26, 107), (523, 27, 111), (520, 28, 114), (517, 29, 118), (516, 30, 119), (515, 31, 120), (513, 32, 122), (512, 33, 123), (510, 34, 125), (509, 35, 126), (507, 36, 128), (506, 37, 129), (504, 38, 131), (503, 39, 132), (501, 40, 134), (500, 41, 135), (499, 42, 136), (498, 43, 137), (497, 44, 138), (496, 45, 139), (496, 46, 139), (495, 47, 140), (495, 48, 140), (494, 49, 141), (493, 50, 142), (492, 51, 143), (491, 52, 144), (491, 53, 144), (490, 54, 145), (490, 55, 145), (490, 56, 145), (490, 57, 146), (490, 58, 146), (490, 59, 146), (491, 60, 145), (491, 61, 145), (491, 62, 145), (492, 63, 144), (493, 64, 143), (494, 65, 142), (495, 66, 141), (496, 67, 139), (497, 68, 138), (498, 69, 138), (499, 70, 137), (500, 71, 136), (501, 72, 135), (503, 73, 133), (503, 74, 133), (505, 75, 131), (506, 76, 130), (507, 77, 129), (508, 78, 128), (509, 79, 127), (510, 80, 126), (511, 81, 125), (512, 82, 124), (513, 83, 123), (514, 84, 122), (515, 85, 121), (516, 86, 120), (517, 87, 119), (518, 88, 118), (519, 89, 117), (521, 90, 115), (521, 91, 115), (522, 92, 114), (523, 93, 113), (524, 94, 112), (525, 95, 111), (526, 96, 110), (527, 97, 109), (529, 98, 107), (530, 99, 106), (532, 100, 104), (533, 101, 103), (534, 102, 102), (535, 103, 101), (536, 104, 100), (538, 105, 98), (540, 106, 96), (541, 107, 95), (543, 108, 93), (546, 109, 90), (548, 110, 88), (549, 111, 87), (551, 112, 84), (552, 113, 83), (553, 114, 82), (555, 115, 80), (556, 116, 79), (556, 117, 79), (557, 118, 78), (558, 119, 77), (559, 120, 76), (560, 121, 75), (560, 122, 75), (561, 123, 74), (561, 124, 74), (561, 125, 74), (562, 126, 73), (562, 127, 73), (563, 128, 72), (563, 129, 72), (564, 130, 70), (564, 131, 70), (565, 132, 69), (565, 133, 68), (565, 134, 68), (565, 135, 67), (566, 136, 65), (566, 137, 64), (566, 138, 64), (566, 139, 62), (566, 140, 61), (566, 141, 59), (566, 142, 57), (566, 143, 56), (566, 144, 55), (566, 145, 54), (567, 146, 53), (567, 147, 52), (567, 148, 51), (568, 149, 50), (568, 150, 49), (568, 151, 48), (568, 152, 47), (569, 153, 45), (569, 154, 44), (570, 155, 42), (570, 156, 42), (570, 157, 41), (571, 158, 39), (571, 159, 39), (572, 160, 37), (572, 161, 37), (573, 162, 35), (573, 163, 34), (573, 164, 34), (574, 165, 32), (575, 166, 30), (577, 167, 28), (578, 168, 26), (581, 169, 22), (584, 170, 19), (587, 171, 15), (591, 172, 8)], ['598,172,591,172,586,170,578,168,573,164,573,162,568,152,568,149,566,145,566,136,565,132,561,125,560,121,556,116,547,109,543,108,536,104,531,99,527,97,491,62,490,54,495,48,496,45,501,40,514,32,517,29,531,25,539,25,540,24,560,24,561,25,579,25,580,26,593,26,594,25,633,25,634,29,634,56,635,57,635,111,634,112,634,129,632,134,629,138,623,141,619,145,617,149,611,155,608,161,604,166']), (957285035, 492601069, 445, 280, 481, 2, 55, 0.82996476, [(292, 3, 128), (284, 4, 146), (282, 5, 151), (281, 6, 154), (281, 7, 156), (281, 8, 157), (281, 9, 158), (281, 10, 160), (281, 11, 162), (281, 12, 165), (281, 13, 167), (281, 14, 169), (281, 15, 171), (281, 16, 173), (281, 17, 174), (281, 18, 175), (281, 19, 177), (281, 20, 178), (281, 21, 179), (281, 22, 180), (281, 23, 181), (281, 24, 182), (281, 25, 183), (281, 26, 184), (281, 27, 185), (281, 28, 185), (281, 29, 185), (282, 30, 185), (283, 31, 27), (337, 31, 131), (371, 32, 97), (401, 33, 68), (409, 34, 61), (419, 35, 52), (424, 36, 48), (429, 37, 44), (432, 38, 41), (434, 39, 40), (436, 40, 39), (438, 41, 37), (441, 42, 35), (444, 43, 32), (448, 44, 29), (452, 45, 25), (454, 46, 23), (459, 47, 17), (463, 48, 12), (468, 49, 5)], ['472,49,468,49,467,48,459,47,458,46,454,46,451,44,448,44,447,43,444,43,440,41,438,41,428,36,424,36,423,35,419,35,418,34,409,34,408,33,401,33,400,32,371,32,370,31,337,31,336,30,283,31,281,29,281,6,284,4,291,4,292,3,419,3,420,4,429,4,430,5,432,5,436,7,441,11,445,12,453,16,456,19,457,19,465,27,465,29,472,37,476,44,476,46']), (957285035, 492601069, 445, 456, 547, 6, 45, 0.7402554, [(482, 8, 19), (464, 9, 3), (481, 9, 44), (457, 10, 12), (479, 10, 50), (457, 11, 13), (476, 11, 56), (457, 12, 15), (475, 12, 65), (457, 13, 84), (457, 14, 85), (457, 15, 89), (457, 16, 89), (458, 17, 88), (459, 18, 87), (460, 19, 86), (461, 20, 80), (464, 21, 71), (466, 22, 63), (467, 23, 59), (468, 24, 55), (469, 25, 52), (469, 26, 51), (470, 27, 48), (471, 28, 46), (471, 29, 44), (472, 30, 42), (473, 31, 39), (473, 32, 38), (474, 33, 36), (475, 34, 33), (475, 35, 32), (476, 36, 30), (476, 37, 29), (477, 38, 26), (478, 39, 23), (479, 40, 20), (480, 41, 17), (488, 42, 5)], ['492,42,488,42,487,41,480,41,476,37,475,34,473,32,469,25,465,21,461,20,457,16,457,10,463,10,464,9,466,9,470,12,474,13,476,11,480,10,482,8,500,8,501,9,524,9,525,10,528,10,532,12,539,12,542,15,545,15,545,19,535,20,534,21,529,21,525,23,523,23,513,30,512,30,504,37,496,41,493,41'])], 'temp/1748568927_1112947_957285035_a42482e51c93c8025d243dd179aee85b.jpg']} free memory after detection : begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 6929 ############################### TEST detect object ################################ run mask_detect Inside batchDatouExec : verbose : False # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! List Step Type Loaded in datou : mask_detect list_input_json : [] origin BFwe have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 time to download the photos : 0.17648959159851074 About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : False number of steps : 1 step1:mask_detect Fri May 30 03:35:47 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Beginning of datou step mask_detect ! save_polygon : True begin detect begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 6675 max_wait_temp : 1 max_wait : 0 gpu_flag : 0 2025-05-30 03:35:49.943381: I tensorflow/core/platform/cpu_feature_guard.cc:143] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA 2025-05-30 03:35:49.971143: I tensorflow/core/platform/profile_utils/cpu_utils.cc:102] CPU Frequency: 3493065000 Hz 2025-05-30 03:35:49.973421: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7fbc40000b60 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:35:49.973485: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2025-05-30 03:35:49.977819: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcuda.so.1 2025-05-30 03:35:50.098267: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x1cb383d0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:35:50.098355: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA GeForce RTX 2080 Ti, Compute Capability 7.5 2025-05-30 03:35:50.099938: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:50.100485: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:50.103735: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:50.106848: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:50.107376: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:50.110951: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:50.112371: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:50.117043: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:50.118280: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:50.118354: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:50.119014: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:35:50.119031: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:35:50.119040: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:35:50.120167: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 5891 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) WARNING:tensorflow:From /home/admin/workarea/git/Velours/python/mtr/mask_rcnn/mask_detection.py:69: The name tf.keras.backend.set_session is deprecated. Please use tf.compat.v1.keras.backend.set_session instead. 2025-05-30 03:35:50.200604: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:50.200757: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:50.200786: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:50.200814: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:50.200840: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:50.200865: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:50.200891: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:50.200936: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:50.202300: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:50.203699: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:35:50.203745: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:35:50.203769: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:50.203795: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:35:50.203817: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:35:50.203840: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:35:50.203862: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:35:50.203885: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:50.204846: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:35:50.204881: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:35:50.204889: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:35:50.204896: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:35:50.205884: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 5891 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) Using TensorFlow backend. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:396: calling crop_and_resize_v1 (from tensorflow.python.ops.image_ops_impl) with box_ind is deprecated and will be removed in a future version. Instructions for updating: box_ind is deprecated, use box_indices instead WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:703: to_int32 (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:729: to_float (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. Inside mask_sub_process Inside mask_detect About to load cache.load_thcl_param FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3473, 'mask_coco_origin', 16384, 25088, 'mask_coco_origin', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2018, 3, 19, 10, 42, 21), datetime.datetime(2018, 3, 19, 10, 42, 21)) {'thcl': {'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}, 'list_hashtags': ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'], 'list_hashtags_csv': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'svm_hashtag_type_desc': 3473, 'photo_desc_type': 3473, 'pb_hashtag_id_or_classifier': 0} list_class_names : ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] Configurations: BACKBONE resnet101 BACKBONE_SHAPES [[160 160] [ 80 80] [ 40 40] [ 20 20] [ 10 10]] BACKBONE_STRIDES [4, 8, 16, 32, 64] BATCH_SIZE 1 BBOX_STD_DEV [0.1 0.1 0.2 0.2] DETECTION_MAX_INSTANCES 100 DETECTION_MIN_CONFIDENCE 0.3 DETECTION_NMS_THRESHOLD 0.3 GPU_COUNT 1 IMAGES_PER_GPU 1 IMAGE_MAX_DIM 640 IMAGE_MIN_DIM 640 IMAGE_PADDING True IMAGE_SHAPE [640 640 3] LEARNING_MOMENTUM 0.9 LEARNING_RATE 0.001 LOSS_WEIGHTS {'rpn_class_loss': 1.0, 'rpn_bbox_loss': 1.0, 'mrcnn_class_loss': 1.0, 'mrcnn_bbox_loss': 1.0, 'mrcnn_mask_loss': 1.0} MASK_POOL_SIZE 14 MASK_SHAPE [28, 28] MAX_GT_INSTANCES 100 MEAN_PIXEL [123.7 116.8 103.9] MINI_MASK_SHAPE (56, 56) NAME mask_coco_origin NUM_CLASSES 81 POOL_SIZE 7 POST_NMS_ROIS_INFERENCE 1000 POST_NMS_ROIS_TRAINING 2000 ROI_POSITIVE_RATIO 0.33 RPN_ANCHOR_RATIOS [0.5, 1, 2] RPN_ANCHOR_SCALES (16, 32, 64, 128, 256) RPN_ANCHOR_STRIDE 1 RPN_BBOX_STD_DEV [0.1 0.1 0.2 0.2] RPN_NMS_THRESHOLD 0.7 RPN_TRAIN_ANCHORS_PER_IMAGE 256 STEPS_PER_EPOCH 1000 TRAIN_ROIS_PER_IMAGE 200 USE_MINI_MASK True USE_RPN_ROIS True VALIDATION_STEPS 50 WEIGHT_DECAY 0.0001 model_param file didn't exist model_name : mask_coco_origin model_type : mask_rcnn list file need : ['mask_model.h5'] file exist in s3 : ['mask_model.h5'] file manque in s3 : [] 2025-05-30 03:35:57.790802: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:35:58.045556: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:35:59.607106: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.607854: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 3.60G (3865470464 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.608488: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 3.24G (3478923264 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.609117: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 2.92G (3131030784 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.609756: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 2.62G (2817927680 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.610385: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 2.36G (2536134912 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.611052: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 2.12G (2282521344 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.611093: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.611758: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.611777: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.630066: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.630088: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.630576: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.630599: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.661654: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.661676: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 466.56MiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.662166: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.662180: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 466.56MiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.716869: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.716957: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.717512: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.717534: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.723125: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.723164: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 243.25MiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.723718: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.723740: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 243.25MiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:35:59.751699: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.752237: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.753830: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.754352: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.790336: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.790889: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.792760: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.793266: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.799648: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.800176: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.804180: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.804688: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.814946: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.815475: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.816917: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.817408: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.822770: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.823318: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.824900: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.825394: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.830882: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.831444: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.832929: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.833456: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.898313: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.898814: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.899333: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.899858: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.903355: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.903883: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.930956: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.931489: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.932013: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.932536: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.954599: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.955138: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.955663: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.956186: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.960384: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.960894: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.965569: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.966069: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.978730: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.979287: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.983510: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.984040: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.984562: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:35:59.985087: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.046401: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.046951: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.047494: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.048036: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.048559: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.049080: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.049603: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.050124: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.059603: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.060132: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.067323: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.067855: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.106605: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.106695: W tensorflow/core/kernels/gpu_utils.cc:49] Failed to allocate memory for convolution redzone checking; skipping this check. This is benign and only means that we won't check cudnn for out-of-bounds reads and writes. This message will only be printed once. 2025-05-30 03:36:00.107632: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.108548: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.115943: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.116839: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.117771: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.118704: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.126759: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.127324: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.143148: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.144028: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.144886: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.145728: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.149867: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.150405: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.150962: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.151491: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.152503: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.162767: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.163331: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.173801: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.174342: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.174883: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.175430: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.175962: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:00.176484: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 4.00G (4294967296 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory local folder : /data/models_weight/mask_coco_origin /data/models_weight/mask_coco_origin/mask_model.h5 size_local : 257557808 size in s3 : 257557808 create time local : 2021-08-09 05:27:17 create time in s3 : 2021-08-06 19:45:17 mask_model.h5 already exist and didn't need to update list_images length : 1 NEW PHOTO Processing 1 images image shape: (720, 1280, 3) min: 0.00000 max: 255.00000 molded_images shape: (1, 640, 640, 3) min: -123.70000 max: 151.10000 image_metas shape: (1, 89) min: 0.00000 max: 1280.00000 nb d'objets trouves : 4 Detection mask done ! Trying to reset tf kernel 1114191 begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 736 tf kernel not reseted sub process len(results) : 1 len(list_Values) 0 None max_time_sub_proc : 3600 parent process len(results) : 1 len(list_Values) 0 process is alive finish correctly or not : True after detect begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 1929 list_Values should be empty [] ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] time for calcul the mask position with numpy : 0.0006952285766601562 nb_pixel_total : 16902 time to create 1 rle with old method : 0.024600744247436523 length of segment : 107 time for calcul the mask position with numpy : 0.31412410736083984 nb_pixel_total : 480752 time to create 1 rle with new method : 0.03649020195007324 length of segment : 632 time for calcul the mask position with numpy : 0.0005257129669189453 nb_pixel_total : 36642 time to create 1 rle with old method : 0.04146075248718262 length of segment : 133 time for calcul the mask position with numpy : 9.632110595703125e-05 nb_pixel_total : 4791 time to create 1 rle with old method : 0.005820512771606445 length of segment : 51 time spent for convertir_results : 0.605231523513794 time spend for datou_step_exec : 16.823044538497925 time spend to save output : 7.867813110351562e-05 total time spend for step 1 : 16.82312321662903 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : False eke 12-6-18 : saveMask need to be cleaned for new output ! Catched exception ! Connect or reconnect ! Number saved : None batch 1 Loaded 428 chid ids of type : 445 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++Number RLEs to save : 0 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 time used for this insertion : 0.01536107063293457 save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {'917855882': [[(917855882, 492601069, 445, 1092, 1280, 0, 108, 0.9988373, [(1205, 1, 58), (1165, 2, 105), (1159, 3, 113), (1149, 4, 124), (1113, 5, 161), (1100, 6, 174), (1097, 7, 177), (1095, 8, 179), (1095, 9, 179), (1095, 10, 179), (1095, 11, 179), (1095, 12, 179), (1095, 13, 179), (1095, 14, 178), (1095, 15, 178), (1095, 16, 178), (1095, 17, 178), (1095, 18, 177), (1095, 19, 177), (1095, 20, 177), (1095, 21, 177), (1095, 22, 177), (1095, 23, 178), (1095, 24, 178), (1095, 25, 178), (1095, 26, 179), (1095, 27, 179), (1095, 28, 180), (1095, 29, 181), (1095, 30, 182), (1095, 31, 183), (1095, 32, 183), (1095, 33, 184), (1095, 34, 184), (1096, 35, 183), (1096, 36, 183), (1096, 37, 184), (1097, 38, 183), (1097, 39, 183), (1097, 40, 183), (1098, 41, 182), (1098, 42, 182), (1098, 43, 182), (1099, 44, 181), (1099, 45, 181), (1099, 46, 181), (1100, 47, 180), (1100, 48, 180), (1101, 49, 179), (1101, 50, 179), (1102, 51, 178), (1102, 52, 178), (1103, 53, 177), (1103, 54, 177), (1104, 55, 176), (1104, 56, 176), (1104, 57, 176), (1104, 58, 176), (1105, 59, 175), (1105, 60, 175), (1105, 61, 175), (1105, 62, 175), (1105, 63, 175), (1106, 64, 174), (1106, 65, 174), (1106, 66, 174), (1106, 67, 174), (1106, 68, 174), (1106, 69, 174), (1106, 70, 174), (1106, 71, 174), (1106, 72, 174), (1106, 73, 174), (1107, 74, 173), (1107, 75, 173), (1107, 76, 173), (1107, 77, 173), (1107, 78, 173), (1107, 79, 173), (1108, 80, 172), (1108, 81, 172), (1109, 82, 171), (1110, 83, 170), (1110, 84, 170), (1111, 85, 169), (1112, 86, 168), (1113, 87, 166), (1114, 88, 165), (1115, 89, 164), (1117, 90, 162), (1120, 91, 159), (1138, 92, 141), (1146, 93, 133), (1154, 94, 125), (1167, 95, 112), (1177, 96, 102), (1183, 97, 95), (1185, 98, 93), (1187, 99, 90), (1188, 100, 55), (1264, 100, 12), (1190, 101, 50), (1191, 102, 46), (1194, 103, 40), (1197, 104, 34), (1202, 105, 25), (1207, 106, 16)], ['1222,106,1207,106,1206,105,1197,104,1191,102,1182,96,1176,95,1167,95,1166,94,1154,94,1153,93,1146,93,1145,92,1137,91,1120,91,1115,89,1110,84,1107,79,1106,73,1106,64,1104,55,1099,46,1095,34,1095,8,1100,6,1112,6,1113,5,1148,5,1149,4,1158,4,1165,2,1204,2,1205,1,1262,1,1269,2,1273,5,1273,13,1271,18,1271,22,1273,27,1277,31,1279,37,1279,86,1278,87,1278,96,1275,100,1264,100,1263,99,1243,99,1230,104']), (917855882, 492601069, 445, 52, 1128, 16, 668, 0.9977483, [(710, 22, 23), (925, 22, 47), (608, 23, 146), (893, 23, 104), (598, 24, 234), (849, 24, 159), (590, 25, 427), (582, 26, 444), (575, 27, 458), (569, 28, 466), (565, 29, 472), (560, 30, 480), (556, 31, 486), (550, 32, 495), (544, 33, 503), (538, 34, 512), (532, 35, 520), (527, 36, 527), (523, 37, 534), (518, 38, 541), (514, 39, 548), (510, 40, 554), (506, 41, 561), (503, 42, 566), (499, 43, 572), (496, 44, 577), (493, 45, 582), (491, 46, 585), (488, 47, 590), (487, 48, 592), (485, 49, 595), (483, 50, 598), (482, 51, 600), (481, 52, 602), (480, 53, 603), (479, 54, 605), (478, 55, 606), (476, 56, 608), (475, 57, 610), (474, 58, 611), (473, 59, 613), (472, 60, 614), (470, 61, 616), (469, 62, 618), (468, 63, 619), (466, 64, 621), (465, 65, 623), (464, 66, 624), (462, 67, 626), (461, 68, 628), (459, 69, 630), (458, 70, 631), (456, 71, 633), (455, 72, 635), (453, 73, 637), (452, 74, 638), (451, 75, 639), (449, 76, 641), (448, 77, 642), (447, 78, 643), (446, 79, 644), (445, 80, 645), (444, 81, 646), (442, 82, 648), (441, 83, 649), (440, 84, 650), (439, 85, 651), (438, 86, 652), (437, 87, 653), (436, 88, 654), (435, 89, 655), (434, 90, 656), (433, 91, 657), (432, 92, 658), (431, 93, 659), (430, 94, 660), (429, 95, 661), (428, 96, 662), (427, 97, 663), (425, 98, 665), (423, 99, 667), (421, 100, 669), (419, 101, 671), (417, 102, 673), (413, 103, 677), (410, 104, 680), (405, 105, 685), (401, 106, 689), (397, 107, 693), (392, 108, 698), (387, 109, 703), (382, 110, 708), (377, 111, 713), (373, 112, 717), (368, 113, 722), (365, 114, 725), (362, 115, 728), (358, 116, 732), (356, 117, 734), (353, 118, 737), (351, 119, 739), (348, 120, 742), (346, 121, 744), (344, 122, 746), (341, 123, 749), (338, 124, 752), (335, 125, 755), (331, 126, 759), (327, 127, 763), (323, 128, 767), (319, 129, 770), (314, 130, 775), (308, 131, 781), (303, 132, 786), (294, 133, 795), (287, 134, 802), (279, 135, 810), (273, 136, 816), (267, 137, 822), (262, 138, 827), (258, 139, 831), (255, 140, 834), (252, 141, 837), (250, 142, 839), (247, 143, 842), (245, 144, 844), (242, 145, 847), (240, 146, 849), (237, 147, 852), (234, 148, 855), (230, 149, 859), (226, 150, 863), (220, 151, 869), (213, 152, 876), (207, 153, 882), (200, 154, 889), (193, 155, 896), (187, 156, 902), (183, 157, 906), (181, 158, 908), (178, 159, 911), (176, 160, 913), (174, 161, 915), (172, 162, 917), (170, 163, 919), (168, 164, 921), (167, 165, 922), (165, 166, 924), (164, 167, 925), (162, 168, 927), (161, 169, 928), (159, 170, 930), (157, 171, 932), (155, 172, 934), (153, 173, 935), (151, 174, 937), (148, 175, 940), (146, 176, 942), (144, 177, 944), (142, 178, 946), (140, 179, 948), (139, 180, 949), (137, 181, 951), (136, 182, 952), (134, 183, 954), (133, 184, 955), (132, 185, 956), (131, 186, 957), (130, 187, 958), (129, 188, 959), (128, 189, 960), (127, 190, 960), (126, 191, 961), (126, 192, 961), (125, 193, 962), (124, 194, 963), (123, 195, 964), (122, 196, 965), (122, 197, 965), (121, 198, 966), (120, 199, 967), (119, 200, 968), (118, 201, 969), (117, 202, 970), (116, 203, 971), (114, 204, 973), (113, 205, 973), (112, 206, 974), (111, 207, 975), (109, 208, 977), (108, 209, 978), (107, 210, 979), (106, 211, 980), (105, 212, 981), (104, 213, 982), (103, 214, 983), (102, 215, 984), (101, 216, 985), (101, 217, 984), (100, 218, 985), (100, 219, 985), (99, 220, 986), (98, 221, 987), (98, 222, 987), (97, 223, 988), (97, 224, 987), (96, 225, 988), (96, 226, 988), (95, 227, 989), (95, 228, 989), (94, 229, 990), (94, 230, 990), (94, 231, 990), (93, 232, 990), (93, 233, 990), (92, 234, 991), (92, 235, 991), (92, 236, 991), (91, 237, 992), (91, 238, 991), (91, 239, 991), (91, 240, 991), (91, 241, 990), (90, 242, 991), (90, 243, 990), (90, 244, 990), (90, 245, 989), (90, 246, 989), (89, 247, 990), (89, 248, 989), (89, 249, 989), (89, 250, 988), (89, 251, 988), (88, 252, 988), (88, 253, 988), (88, 254, 987), (88, 255, 986), (88, 256, 986), (87, 257, 986), (87, 258, 986), (87, 259, 985), (87, 260, 984), (87, 261, 983), (86, 262, 983), (86, 263, 983), (86, 264, 982), (86, 265, 981), (85, 266, 981), (85, 267, 980), (85, 268, 980), (84, 269, 980), (84, 270, 979), (84, 271, 979), (84, 272, 978), (83, 273, 979), (83, 274, 978), (83, 275, 977), (82, 276, 978), (82, 277, 977), (82, 278, 977), (81, 279, 977), (81, 280, 977), (81, 281, 977), (80, 282, 977), (80, 283, 977), (80, 284, 976), (79, 285, 977), (79, 286, 976), (79, 287, 976), (78, 288, 976), (78, 289, 976), (78, 290, 975), (77, 291, 976), (77, 292, 975), (77, 293, 975), (76, 294, 975), (76, 295, 975), (76, 296, 974), (75, 297, 975), (75, 298, 974), (74, 299, 975), (74, 300, 974), (74, 301, 974), (73, 302, 974), (73, 303, 974), (72, 304, 974), (72, 305, 974), (71, 306, 974), (71, 307, 973), (71, 308, 972), (70, 309, 972), (70, 310, 971), (70, 311, 970), (70, 312, 968), (69, 313, 968), (69, 314, 966), (69, 315, 964), (69, 316, 962), (68, 317, 961), (68, 318, 959), (68, 319, 958), (68, 320, 956), (67, 321, 955), (67, 322, 954), (67, 323, 952), (67, 324, 951), (66, 325, 951), (66, 326, 950), (66, 327, 948), (66, 328, 947), (65, 329, 947), (65, 330, 946), (65, 331, 946), (65, 332, 945), (65, 333, 944), (65, 334, 942), (65, 335, 941), (65, 336, 940), (65, 337, 939), (65, 338, 938), (65, 339, 936), (64, 340, 936), (64, 341, 934), (64, 342, 932), (64, 343, 930), (64, 344, 928), (64, 345, 927), (64, 346, 925), (64, 347, 923), (64, 348, 922), (64, 349, 920), (64, 350, 919), (63, 351, 919), (63, 352, 918), (63, 353, 917), (63, 354, 916), (63, 355, 915), (63, 356, 914), (63, 357, 912), (63, 358, 911), (63, 359, 910), (63, 360, 909), (63, 361, 908), (63, 362, 906), (63, 363, 905), (63, 364, 904), (63, 365, 902), (63, 366, 901), (63, 367, 899), (63, 368, 898), (63, 369, 896), (62, 370, 895), (62, 371, 893), (62, 372, 891), (62, 373, 890), (62, 374, 888), (62, 375, 887), (62, 376, 886), (62, 377, 885), (62, 378, 884), (62, 379, 883), (63, 380, 880), (63, 381, 879), (63, 382, 878), (63, 383, 877), (63, 384, 876), (63, 385, 875), (63, 386, 874), (63, 387, 873), (63, 388, 872), (64, 389, 870), (64, 390, 869), (64, 391, 868), (64, 392, 867), (64, 393, 865), (64, 394, 864), (64, 395, 863), (65, 396, 861), (65, 397, 860), (65, 398, 859), (65, 399, 858), (65, 400, 857), (65, 401, 856), (65, 402, 854), (65, 403, 853), (65, 404, 851), (65, 405, 850), (65, 406, 848), (66, 407, 846), (66, 408, 844), (66, 409, 843), (66, 410, 842), (66, 411, 841), (66, 412, 840), (66, 413, 838), (66, 414, 837), (66, 415, 836), (66, 416, 835), (66, 417, 835), (66, 418, 834), (66, 419, 833), (67, 420, 831), (67, 421, 830), (67, 422, 829), (67, 423, 829), (67, 424, 828), (67, 425, 827), (67, 426, 826), (67, 427, 825), (67, 428, 824), (68, 429, 822), (68, 430, 820), (68, 431, 819), (68, 432, 818), (68, 433, 816), (68, 434, 815), (68, 435, 813), (68, 436, 811), (69, 437, 809), (69, 438, 807), (69, 439, 806), (69, 440, 804), (69, 441, 803), (69, 442, 802), (69, 443, 800), (70, 444, 798), (70, 445, 797), (70, 446, 796), (70, 447, 796), (71, 448, 794), (71, 449, 794), (72, 450, 792), (72, 451, 792), (73, 452, 790), (73, 453, 789), (74, 454, 788), (74, 455, 787), (75, 456, 786), (75, 457, 785), (76, 458, 784), (76, 459, 783), (77, 460, 782), (77, 461, 781), (77, 462, 781), (78, 463, 779), (78, 464, 779), (79, 465, 777), (79, 466, 777), (79, 467, 776), (80, 468, 775), (80, 469, 774), (80, 470, 774), (81, 471, 772), (81, 472, 771), (82, 473, 770), (82, 474, 769), (83, 475, 767), (83, 476, 766), (83, 477, 766), (84, 478, 764), (84, 479, 763), (85, 480, 761), (85, 481, 760), (85, 482, 759), (86, 483, 757), (86, 484, 755), (87, 485, 753), (87, 486, 752), (87, 487, 750), (88, 488, 748), (88, 489, 747), (88, 490, 746), (89, 491, 744), (89, 492, 743), (90, 493, 741), (90, 494, 741), (91, 495, 739), (91, 496, 738), (92, 497, 737), (93, 498, 735), (94, 499, 733), (94, 500, 733), (95, 501, 731), (96, 502, 729), (97, 503, 728), (98, 504, 726), (99, 505, 724), (99, 506, 724), (100, 507, 722), (101, 508, 721), (102, 509, 719), (104, 510, 717), (105, 511, 715), (106, 512, 714), (107, 513, 712), (108, 514, 711), (110, 515, 708), (111, 516, 707), (113, 517, 704), (114, 518, 703), (115, 519, 701), (117, 520, 698), (118, 521, 697), (119, 522, 695), (121, 523, 693), (122, 524, 691), (124, 525, 689), (125, 526, 687), (126, 527, 685), (128, 528, 683), (129, 529, 681), (131, 530, 678), (132, 531, 676), (134, 532, 673), (135, 533, 672), (137, 534, 669), (138, 535, 667), (140, 536, 664), (141, 537, 662), (143, 538, 659), (144, 539, 657), (146, 540, 654), (148, 541, 651), (149, 542, 649), (151, 543, 645), (153, 544, 642), (154, 545, 640), (156, 546, 638), (158, 547, 635), (159, 548, 633), (161, 549, 630), (162, 550, 628), (164, 551, 626), (166, 552, 623), (167, 553, 621), (169, 554, 618), (170, 555, 617), (171, 556, 615), (173, 557, 613), (174, 558, 611), (176, 559, 608), (177, 560, 607), (178, 561, 605), (180, 562, 603), (181, 563, 601), (183, 564, 599), (185, 565, 597), (186, 566, 595), (189, 567, 592), (192, 568, 589), (195, 569, 585), (198, 570, 582), (201, 571, 579), (204, 572, 575), (206, 573, 573), (209, 574, 569), (212, 575, 566), (215, 576, 563), (218, 577, 559), (221, 578, 556), (223, 579, 553), (226, 580, 550), (228, 581, 547), (230, 582, 545), (232, 583, 542), (234, 584, 540), (235, 585, 539), (237, 586, 536), (238, 587, 534), (240, 588, 531), (242, 589, 528), (243, 590, 526), (245, 591, 523), (247, 592, 520), (249, 593, 516), (251, 594, 513), (253, 595, 510), (256, 596, 505), (258, 597, 501), (261, 598, 497), (263, 599, 493), (267, 600, 488), (271, 601, 482), (274, 602, 478), (278, 603, 473), (281, 604, 468), (284, 605, 464), (287, 606, 460), (290, 607, 456), (292, 608, 453), (295, 609, 449), (298, 610, 445), (300, 611, 442), (303, 612, 438), (305, 613, 434), (308, 614, 430), (310, 615, 427), (312, 616, 423), (315, 617, 418), (317, 618, 415), (320, 619, 410), (322, 620, 406), (325, 621, 401), (327, 622, 396), (330, 623, 390), (333, 624, 384), (335, 625, 379), (338, 626, 374), (341, 627, 369), (345, 628, 362), (349, 629, 356), (353, 630, 350), (357, 631, 344), (360, 632, 340), (364, 633, 334), (368, 634, 328), (373, 635, 320), (378, 636, 313), (383, 637, 305), (389, 638, 295), (395, 639, 282), (401, 640, 270), (408, 641, 256), (416, 642, 240), (432, 643, 216), (448, 644, 193), (465, 645, 169), (480, 646, 148), (495, 647, 126), (511, 648, 104), (526, 649, 82), (565, 650, 9)], ['526,649,416,642,341,627,289,606,263,599,220,577,186,566,144,539,102,509,91,496,70,447,63,388,65,329,86,265,91,237,101,216,134,183,187,156,225,151,279,135,318,130,358,116,416,103,493,45,527,36,608,23,754,24,892,24,925,22,996,23,1032,27,1066,41,1082,52,1089,72,1088,172,1082,237,1045,305,1019,322,1002,338,950,373,920,401,865,446,851,473,822,505,810,528,786,554,773,585,740,612,683,638,607,649']), (917855882, 492601069, 445, 0, 440, 0, 116, 0.99194473, [(127, 1, 141), (94, 2, 206), (384, 2, 2), (59, 3, 273), (340, 3, 57), (22, 4, 381), (19, 5, 387), (16, 6, 392), (15, 7, 394), (14, 8, 396), (14, 9, 397), (13, 10, 399), (12, 11, 400), (12, 12, 400), (11, 13, 402), (10, 14, 403), (11, 15, 403), (11, 16, 404), (12, 17, 403), (12, 18, 404), (12, 19, 405), (12, 20, 405), (12, 21, 406), (12, 22, 406), (12, 23, 407), (12, 24, 407), (12, 25, 408), (12, 26, 408), (12, 27, 408), (12, 28, 408), (12, 29, 409), (12, 30, 409), (12, 31, 409), (12, 32, 409), (12, 33, 409), (12, 34, 410), (12, 35, 410), (12, 36, 410), (12, 37, 410), (12, 38, 410), (12, 39, 410), (12, 40, 410), (12, 41, 411), (12, 42, 411), (12, 43, 411), (12, 44, 411), (12, 45, 411), (12, 46, 410), (12, 47, 410), (12, 48, 410), (12, 49, 410), (12, 50, 410), (12, 51, 410), (12, 52, 409), (12, 53, 408), (12, 54, 408), (12, 55, 407), (12, 56, 406), (12, 57, 404), (12, 58, 403), (11, 59, 403), (11, 60, 402), (11, 61, 401), (11, 62, 400), (11, 63, 400), (11, 64, 399), (11, 65, 398), (11, 66, 397), (11, 67, 397), (11, 68, 396), (11, 69, 395), (11, 70, 395), (11, 71, 394), (11, 72, 394), (11, 73, 394), (11, 74, 393), (11, 75, 393), (11, 76, 393), (11, 77, 393), (11, 78, 393), (11, 79, 393), (11, 80, 392), (10, 81, 394), (10, 82, 394), (10, 83, 395), (9, 84, 396), (9, 85, 262), (279, 85, 126), (9, 86, 75), (98, 86, 28), (142, 86, 117), (292, 86, 112), (9, 87, 71), (152, 87, 103), (294, 87, 110), (8, 88, 68), (161, 88, 91), (296, 88, 107), (8, 89, 63), (176, 89, 73), (297, 89, 106), (7, 90, 61), (205, 90, 40), (298, 90, 104), (7, 91, 57), (299, 91, 103), (6, 92, 54), (300, 92, 102), (6, 93, 50), (301, 93, 100), (7, 94, 46), (303, 94, 97), (7, 95, 44), (306, 95, 92), (7, 96, 42), (308, 96, 89), (7, 97, 40), (310, 97, 86), (7, 98, 38), (312, 98, 83), (8, 99, 34), (314, 99, 79), (8, 100, 32), (317, 100, 75), (8, 101, 29), (319, 101, 71), (13, 102, 19), (324, 102, 63), (20, 103, 6), (330, 103, 51), (337, 104, 37), (344, 105, 22), (352, 106, 3)], ['344,105,319,101,301,93,291,85,259,85,244,90,205,90,204,89,176,89,161,88,141,85,126,85,125,86,98,86,84,85,56,92,36,101,26,102,8,101,6,92,11,80,11,59,12,58,12,17,10,14,16,6,22,4,58,4,59,3,93,3,94,2,126,2,127,1,267,1,268,2,331,3,396,3,407,6,411,10,419,25,421,34,421,51,410,62,404,71,402,80,404,85,401,92,394,98,386,102,365,105']), (917855882, 492601069, 445, 390, 550, 0, 54, 0.93909764, [(415, 0, 5), (441, 0, 60), (508, 0, 28), (402, 1, 142), (401, 2, 146), (402, 3, 145), (404, 4, 143), (406, 5, 140), (408, 6, 137), (410, 7, 134), (411, 8, 132), (412, 9, 130), (413, 10, 127), (414, 11, 125), (415, 12, 123), (415, 13, 122), (416, 14, 120), (417, 15, 117), (417, 16, 116), (418, 17, 114), (418, 18, 113), (418, 19, 111), (418, 20, 109), (419, 21, 107), (419, 22, 105), (419, 23, 103), (419, 24, 102), (420, 25, 99), (420, 26, 97), (420, 27, 95), (420, 28, 94), (421, 29, 91), (421, 30, 90), (422, 31, 88), (422, 32, 88), (422, 33, 87), (423, 34, 84), (423, 35, 82), (423, 36, 81), (424, 37, 79), (424, 38, 77), (424, 39, 75), (424, 40, 73), (424, 41, 71), (425, 42, 67), (425, 43, 66), (426, 44, 62), (426, 45, 6), (433, 45, 52), (443, 46, 30), (450, 47, 1)], ['449,46,443,46,442,45,426,45,424,41,424,37,423,36,422,31,420,28,420,25,419,24,419,21,418,20,418,17,417,15,409,6,402,3,402,1,414,1,415,0,419,0,420,1,440,1,441,0,500,0,501,1,507,1,508,0,535,0,536,1,543,1,546,2,546,4,542,8,530,18,527,19,525,21,522,22,520,24,512,28,508,33,505,34,502,37,494,41,492,41,490,43,488,43,484,45,473,45,472,46'])], 'temp/1748568947_1112947_917855882_da0fa7b7e6b5b551fe26c0ba8713276d.jpg']} ############################### TEST POLYGON ################################ Inside batchDatouExec : verbose : False # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! List Step Type Loaded in datou : mask_detect list_input_json : [] origin BFwe have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 time to download the photos : 0.17472600936889648 About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : False number of steps : 1 step1:mask_detect Fri May 30 03:36:05 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Beginning of datou step mask_detect ! save_polygon : True begin detect begin to check gpu status inside check gpu memory havn't enough memory gpu , need / 3000 l 3632 free memory gpu now : 1929 wait 20 seconds l 3637 free memory gpu now : 1929 max_wait_temp : 1 max_wait : 0 gpu_flag : 0 2025-05-30 03:36:28.800938: I tensorflow/core/platform/cpu_feature_guard.cc:143] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA 2025-05-30 03:36:28.827128: I tensorflow/core/platform/profile_utils/cpu_utils.cc:102] CPU Frequency: 3493065000 Hz 2025-05-30 03:36:28.828670: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7fbc40000b60 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:36:28.828729: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2025-05-30 03:36:28.831706: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcuda.so.1 2025-05-30 03:36:29.015706: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x1c7de250 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:36:29.015803: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA GeForce RTX 2080 Ti, Compute Capability 7.5 2025-05-30 03:36:29.016766: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:36:29.017638: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:36:29.021911: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:36:29.025102: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:36:29.026028: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:36:29.029875: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:36:29.032230: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:36:29.038689: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:36:29.039922: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:36:29.040048: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:36:29.040611: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:36:29.040629: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:36:29.040639: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:36:29.041497: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 1552 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) WARNING:tensorflow:From /home/admin/workarea/git/Velours/python/mtr/mask_rcnn/mask_detection.py:69: The name tf.keras.backend.set_session is deprecated. Please use tf.compat.v1.keras.backend.set_session instead. 2025-05-30 03:36:29.136639: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:36:29.136784: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:36:29.136805: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:36:29.136823: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:36:29.136840: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:36:29.136857: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:36:29.136874: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:36:29.136891: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:36:29.137614: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:36:29.138561: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:36:29.138598: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:36:29.138619: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:36:29.138637: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:36:29.138656: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:36:29.138674: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:36:29.138693: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:36:29.138711: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:36:29.139463: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:36:29.139501: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:36:29.139510: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:36:29.139518: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:36:29.140308: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 1552 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) Using TensorFlow backend. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:396: calling crop_and_resize_v1 (from tensorflow.python.ops.image_ops_impl) with box_ind is deprecated and will be removed in a future version. Instructions for updating: box_ind is deprecated, use box_indices instead WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:703: to_int32 (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. WARNING:tensorflow:From /home/admin/workarea/install/Mask_RCNN/model.py:729: to_float (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version. Instructions for updating: Use `tf.cast` instead. Inside mask_sub_process Inside mask_detect About to load cache.load_thcl_param FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3473, 'mask_coco_origin', 16384, 25088, 'mask_coco_origin', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2018, 3, 19, 10, 42, 21), datetime.datetime(2018, 3, 19, 10, 42, 21)) {'thcl': {'id': 454, 'mtr_user_id': 31, 'name': 'mask_coco_origin', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'photo_desc_type': 3473, 'type_classification': 'mask_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}, 'list_hashtags': ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'], 'list_hashtags_csv': 'backgroud,person,bicycle,car,motorcycle,airplane,bus,train,truck,boat,trafficlight,firehydrant,stopsign,parkingmeter,bench,bird,cat,dog,horse,sheep,cow,elephant,bear,zebra,giraffe,backpack,umbrella,handbag,tie,suitcase,frisbee,skis,snowboard,sportsball,kite,baseballbat,baseballglove,skateboard,surfboard,tennisracket,bottle,wineglass,cup,fork,knife,spoon,bowl,banana,apple,sandwich,orange,broccoli,carrot,hotdog,pizza,donut,cake,chair,couch,pottedplant,bed,diningtable,toilet,tv,laptop,mouse,remote,keyboard,cellphone,microwave,oven,toaster,sink,refrigerator,book,clock,vase,scissors,teddybear,hairdrier,toothbrush', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 445, 'svm_hashtag_type_desc': 3473, 'photo_desc_type': 3473, 'pb_hashtag_id_or_classifier': 0} list_class_names : ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] Configurations: BACKBONE resnet101 BACKBONE_SHAPES [[160 160] [ 80 80] [ 40 40] [ 20 20] [ 10 10]] BACKBONE_STRIDES [4, 8, 16, 32, 64] BATCH_SIZE 1 BBOX_STD_DEV [0.1 0.1 0.2 0.2] DETECTION_MAX_INSTANCES 100 DETECTION_MIN_CONFIDENCE 0.3 DETECTION_NMS_THRESHOLD 0.3 GPU_COUNT 1 IMAGES_PER_GPU 1 IMAGE_MAX_DIM 640 IMAGE_MIN_DIM 640 IMAGE_PADDING True IMAGE_SHAPE [640 640 3] LEARNING_MOMENTUM 0.9 LEARNING_RATE 0.001 LOSS_WEIGHTS {'rpn_class_loss': 1.0, 'rpn_bbox_loss': 1.0, 'mrcnn_class_loss': 1.0, 'mrcnn_bbox_loss': 1.0, 'mrcnn_mask_loss': 1.0} MASK_POOL_SIZE 14 MASK_SHAPE [28, 28] MAX_GT_INSTANCES 100 MEAN_PIXEL [123.7 116.8 103.9] MINI_MASK_SHAPE (56, 56) NAME mask_coco_origin NUM_CLASSES 81 POOL_SIZE 7 POST_NMS_ROIS_INFERENCE 1000 POST_NMS_ROIS_TRAINING 2000 ROI_POSITIVE_RATIO 0.33 RPN_ANCHOR_RATIOS [0.5, 1, 2] RPN_ANCHOR_SCALES (16, 32, 64, 128, 256) RPN_ANCHOR_STRIDE 1 RPN_BBOX_STD_DEV [0.1 0.1 0.2 0.2] RPN_NMS_THRESHOLD 0.7 RPN_TRAIN_ANCHORS_PER_IMAGE 256 STEPS_PER_EPOCH 1000 TRAIN_ROIS_PER_IMAGE 200 USE_MINI_MASK True USE_RPN_ROIS True VALIDATION_STEPS 50 WEIGHT_DECAY 0.0001 model_param file didn't exist model_name : mask_coco_origin model_type : mask_rcnn list file need : ['mask_model.h5'] file exist in s3 : ['mask_model.h5'] file manque in s3 : [] 2025-05-30 03:36:37.978547: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:36:38.195285: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:36:39.679915: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.680016: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.686007: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.686031: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.732360: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.732454: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.06GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.812984: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.09GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.813076: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.09GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.885435: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.15GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.885500: W tensorflow/core/common_runtime/bfc_allocator.cc:245] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.15GiB with freed_by_count=0. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available. 2025-05-30 03:36:39.911257: W tensorflow/core/common_runtime/bfc_allocator.cc:311] Garbage collection: deallocate free memory regions (i.e., allocations) so that we can re-allocate a larger region to avoid OOM due to memory fragmentation. If you see this message frequently, you are running near the threshold of the available device memory and re-allocation may incur great performance overhead. You may try smaller batch sizes to observe the performance impact. Set TF_ENABLE_GPU_GARBAGE_COLLECTION=false if you'd like to disable this feature. 2025-05-30 03:36:39.928685: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.00G (1073741824 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:39.929632: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 921.60M (966367744 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:39.930565: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 829.44M (869731072 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:39.931558: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 746.50M (782758144 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.201664: W tensorflow/core/common_runtime/bfc_allocator.cc:311] Garbage collection: deallocate free memory regions (i.e., allocations) so that we can re-allocate a larger region to avoid OOM due to memory fragmentation. If you see this message frequently, you are running near the threshold of the available device memory and re-allocation may incur great performance overhead. You may try smaller batch sizes to observe the performance impact. Set TF_ENABLE_GPU_GARBAGE_COLLECTION=false if you'd like to disable this feature. 2025-05-30 03:36:40.220809: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.221781: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.230611: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.231310: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.235867: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.236367: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.248430: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.248959: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.253047: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.253556: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.275239: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.275759: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.276279: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.276782: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.277284: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.277785: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.338122: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.338207: W tensorflow/core/kernels/gpu_utils.cc:49] Failed to allocate memory for convolution redzone checking; skipping this check. This is benign and only means that we won't check cudnn for out-of-bounds reads and writes. This message will only be printed once. 2025-05-30 03:36:40.339249: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.340076: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.347593: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.348349: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.360226: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.360775: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.396940: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.397474: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.397985: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.398485: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.402565: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.403092: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.403635: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.404154: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.405060: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.415676: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.416181: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.426634: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.427148: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.427657: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.428155: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.428660: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory 2025-05-30 03:36:40.429158: I tensorflow/stream_executor/cuda/cuda_driver.cc:763] failed to allocate 1.02G (1092485120 bytes) from device: CUDA_ERROR_OUT_OF_MEMORY: out of memory local folder : /data/models_weight/mask_coco_origin /data/models_weight/mask_coco_origin/mask_model.h5 size_local : 257557808 size in s3 : 257557808 create time local : 2021-08-09 05:27:17 create time in s3 : 2021-08-06 19:45:17 mask_model.h5 already exist and didn't need to update list_images length : 1 NEW PHOTO Processing 1 images image shape: (2448, 2448, 3) min: 0.00000 max: 255.00000 molded_images shape: (1, 640, 640, 3) min: -123.70000 max: 151.10000 image_metas shape: (1, 89) min: 0.00000 max: 2448.00000 nb d'objets trouves : 1 Detection mask done ! Trying to reset tf kernel 1116332 begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 736 tf kernel not reseted sub process len(results) : 1 len(list_Values) 0 None max_time_sub_proc : 3600 parent process len(results) : 1 len(list_Values) 0 process is alive finish correctly or not : True after detect begin to check gpu status inside check gpu memory l 3610 free memory gpu now : 1929 list_Values should be empty [] ['backgroud', 'person', 'bicycle', 'car', 'motorcycle', 'airplane', 'bus', 'train', 'truck', 'boat', 'trafficlight', 'firehydrant', 'stopsign', 'parkingmeter', 'bench', 'bird', 'cat', 'dog', 'horse', 'sheep', 'cow', 'elephant', 'bear', 'zebra', 'giraffe', 'backpack', 'umbrella', 'handbag', 'tie', 'suitcase', 'frisbee', 'skis', 'snowboard', 'sportsball', 'kite', 'baseballbat', 'baseballglove', 'skateboard', 'surfboard', 'tennisracket', 'bottle', 'wineglass', 'cup', 'fork', 'knife', 'spoon', 'bowl', 'banana', 'apple', 'sandwich', 'orange', 'broccoli', 'carrot', 'hotdog', 'pizza', 'donut', 'cake', 'chair', 'couch', 'pottedplant', 'bed', 'diningtable', 'toilet', 'tv', 'laptop', 'mouse', 'remote', 'keyboard', 'cellphone', 'microwave', 'oven', 'toaster', 'sink', 'refrigerator', 'book', 'clock', 'vase', 'scissors', 'teddybear', 'hairdrier', 'toothbrush'] time for calcul the mask position with numpy : 0.6964271068572998 nb_pixel_total : 3693278 time to create 1 rle with new method : 0.30445265769958496 length of segment : 2042 time spent for convertir_results : 1.8834445476531982 time spend for datou_step_exec : 39.92736339569092 time spend to save output : 2.8848648071289062e-05 total time spend for step 1 : 39.92739224433899 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : False eke 12-6-18 : saveMask need to be cleaned for new output ! Catched exception ! Connect or reconnect ! Number saved : None batch 1 Loaded 722 chid ids of type : 445 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++Number RLEs to save : 0 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 time used for this insertion : 0.016752243041992188 save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {'917877156': [[(917877156, 492601069, 445, 7, 2268, 118, 2241, 0.9850123, [(675, 120, 112), (520, 121, 481), (1051, 121, 380), (502, 122, 948), (486, 123, 982), (470, 124, 1015), (455, 125, 1046), (442, 126, 1092), (429, 127, 1136), (417, 128, 1168), (405, 129, 1187), (394, 130, 1205), (383, 131, 1223), (373, 132, 1239), (368, 133, 1250), (366, 134, 1258), (363, 135, 1267), (361, 136, 1274), (359, 137, 1281), (357, 138, 1288), (355, 139, 1295), (353, 140, 1302), (351, 141, 1309), (349, 142, 1315), (347, 143, 1320), (345, 144, 1326), (343, 145, 1331), (342, 146, 1335), (340, 147, 1340), (338, 148, 1345), (337, 149, 1349), (335, 150, 1354), (334, 151, 1358), (332, 152, 1363), (331, 153, 1366), (330, 154, 1370), (328, 155, 1374), (327, 156, 1378), (326, 157, 1381), (325, 158, 1385), (323, 159, 1389), (322, 160, 1393), (321, 161, 1397), (319, 162, 1402), (318, 163, 1406), (317, 164, 1410), (315, 165, 1415), (314, 166, 1419), (312, 167, 1424), (310, 168, 1429), (309, 169, 1434), (307, 170, 1439), (305, 171, 1444), (304, 172, 1448), (302, 173, 1453), (300, 174, 1458), (298, 175, 1463), (296, 176, 1469), (294, 177, 1474), (292, 178, 1480), (289, 179, 1487), (286, 180, 1493), (283, 181, 1500), (280, 182, 1508), (278, 183, 1514), (275, 184, 1521), (272, 185, 1529), (269, 186, 1536), (266, 187, 1544), (263, 188, 1552), (260, 189, 1561), (257, 190, 1569), (254, 191, 1579), (251, 192, 1588), (248, 193, 1597), (245, 194, 1606), (242, 195, 1615), (239, 196, 1624), (237, 197, 1631), (234, 198, 1640), (231, 199, 1648), (228, 200, 1657), (225, 201, 1665), (222, 202, 1673), (219, 203, 1681), (216, 204, 1689), (213, 205, 1694), (210, 206, 1699), (208, 207, 1702), (206, 208, 1706), (204, 209, 1710), (203, 210, 1712), (201, 211, 1716), (199, 212, 1719), (198, 213, 1721), (196, 214, 1725), (195, 215, 1727), (193, 216, 1730), (192, 217, 1733), (191, 218, 1735), (189, 219, 1738), (188, 220, 1740), (187, 221, 1742), (186, 222, 1744), (185, 223, 1746), (183, 224, 1749), (182, 225, 1751), (181, 226, 1753), (180, 227, 1755), (179, 228, 1757), (178, 229, 1759), (177, 230, 1761), (176, 231, 1762), (176, 232, 1763), (175, 233, 1765), (174, 234, 1767), (173, 235, 1768), (172, 236, 1770), (171, 237, 1772), (170, 238, 1774), (169, 239, 1775), (168, 240, 1777), (167, 241, 1779), (166, 242, 1781), (165, 243, 1783), (164, 244, 1785), (163, 245, 1787), (162, 246, 1789), (161, 247, 1791), (159, 248, 1794), (158, 249, 1796), (157, 250, 1798), (156, 251, 1800), (154, 252, 1803), (153, 253, 1805), (152, 254, 1807), (150, 255, 1810), (149, 256, 1812), (148, 257, 1815), (146, 258, 1818), (145, 259, 1820), (143, 260, 1823), (142, 261, 1826), (140, 262, 1829), (138, 263, 1833), (137, 264, 1835), (135, 265, 1839), (133, 266, 1842), (132, 267, 1845), (130, 268, 1849), (128, 269, 1852), (126, 270, 1856), (125, 271, 1859), (124, 272, 1862), (122, 273, 1865), (121, 274, 1868), (120, 275, 1871), (119, 276, 1873), (118, 277, 1876), (116, 278, 1879), (115, 279, 1881), (114, 280, 1884), (113, 281, 1886), (112, 282, 1888), (111, 283, 1890), (110, 284, 1892), (109, 285, 1895), (108, 286, 1897), (108, 287, 1898), (107, 288, 1900), (106, 289, 1902), (105, 290, 1904), (104, 291, 1906), (103, 292, 1908), (103, 293, 1909), (102, 294, 1910), (101, 295, 1912), (101, 296, 1913), (100, 297, 1915), (99, 298, 1917), (99, 299, 1918), (98, 300, 1919), (97, 301, 1921), (97, 302, 1922), (96, 303, 1924), (95, 304, 1925), (95, 305, 1926), (94, 306, 1928), (94, 307, 1928), (93, 308, 1930), (93, 309, 1930), (93, 310, 1931), (93, 311, 1931), (92, 312, 1933), (92, 313, 1933), (92, 314, 1934), (92, 315, 1934), (91, 316, 1936), (91, 317, 1936), (91, 318, 1937), (91, 319, 1937), (90, 320, 1939), (90, 321, 1939), (90, 322, 1940), (89, 323, 1941), (89, 324, 1942), (89, 325, 1943), (89, 326, 1943), (88, 327, 1945), (88, 328, 1945), (88, 329, 1946), (87, 330, 1948), (87, 331, 1948), (87, 332, 1949), (87, 333, 1949), (86, 334, 1951), (86, 335, 1952), (86, 336, 1952), (85, 337, 1954), (85, 338, 1955), (85, 339, 1955), (85, 340, 1956), (84, 341, 1958), (84, 342, 1959), (84, 343, 1959), (83, 344, 1961), (83, 345, 1962), (83, 346, 1963), (83, 347, 1963), (82, 348, 1965), (82, 349, 1966), (82, 350, 1967), (81, 351, 1969), (81, 352, 1970), (81, 353, 1970), (80, 354, 1972), (80, 355, 1973), (80, 356, 1974), (80, 357, 1975), (79, 358, 1977), (79, 359, 1978), (79, 360, 1979), (78, 361, 1981), (78, 362, 1982), (78, 363, 1983), (77, 364, 1985), (77, 365, 1986), (77, 366, 1987), (76, 367, 1989), (76, 368, 1990), (76, 369, 1991), (76, 370, 1992), (75, 371, 1994), (75, 372, 1995), (75, 373, 1996), (74, 374, 1998), (74, 375, 1999), (74, 376, 2000), (73, 377, 2002), (73, 378, 2003), (73, 379, 2004), (72, 380, 2005), (72, 381, 2006), (72, 382, 2007), (71, 383, 2009), (71, 384, 2009), (71, 385, 2010), (70, 386, 2012), (70, 387, 2012), (70, 388, 2013), (70, 389, 2013), (69, 390, 2015), (69, 391, 2015), (69, 392, 2016), (68, 393, 2017), (68, 394, 2018), (68, 395, 2019), (67, 396, 2020), (67, 397, 2021), (67, 398, 2021), (66, 399, 2023), (66, 400, 2023), (65, 401, 2025), (65, 402, 2025), (65, 403, 2026), (64, 404, 2027), (64, 405, 2028), (64, 406, 2028), (63, 407, 2030), (63, 408, 2030), (63, 409, 2031), (62, 410, 2032), (62, 411, 2033), (61, 412, 2034), (61, 413, 2034), (61, 414, 2035), (60, 415, 2036), (60, 416, 2037), (59, 417, 2038), (59, 418, 2039), (58, 419, 2040), (58, 420, 2041), (58, 421, 2041), (57, 422, 2042), (57, 423, 2043), (56, 424, 2044), (56, 425, 2045), (55, 426, 2046), (55, 427, 2047), (54, 428, 2048), (54, 429, 2048), (53, 430, 2050), (53, 431, 2050), (52, 432, 2052), (52, 433, 2052), (51, 434, 2053), (51, 435, 2054), (50, 436, 2055), (50, 437, 2055), (49, 438, 2057), (49, 439, 2057), (48, 440, 2059), (48, 441, 2059), (47, 442, 2060), (47, 443, 2061), (46, 444, 2062), (46, 445, 2062), (45, 446, 2064), (45, 447, 2064), (44, 448, 2065), (44, 449, 2066), (43, 450, 2067), (43, 451, 2068), (42, 452, 2069), (42, 453, 2069), (41, 454, 2071), (41, 455, 2071), (40, 456, 2072), (40, 457, 2073), (39, 458, 2074), (39, 459, 2074), (39, 460, 2074), (39, 461, 2075), (38, 462, 2076), (38, 463, 2076), (38, 464, 2077), (38, 465, 2077), (38, 466, 2077), (38, 467, 2078), (37, 468, 2079), (37, 469, 2079), (37, 470, 2080), (37, 471, 2080), (37, 472, 2080), (37, 473, 2081), (37, 474, 2081), (36, 475, 2082), (36, 476, 2083), (36, 477, 2083), (36, 478, 2083), (36, 479, 2084), (36, 480, 2084), (35, 481, 2085), (35, 482, 2086), (35, 483, 2086), (35, 484, 2087), (35, 485, 2087), (35, 486, 2087), (34, 487, 2089), (34, 488, 2089), (34, 489, 2089), (34, 490, 2090), (34, 491, 2090), (34, 492, 2091), (34, 493, 2091), (33, 494, 2093), (33, 495, 2093), (33, 496, 2093), (33, 497, 2094), (33, 498, 2094), (33, 499, 2095), (33, 500, 2095), (32, 501, 2097), (32, 502, 2097), (32, 503, 2098), (32, 504, 2098), (32, 505, 2099), (32, 506, 2099), (32, 507, 2100), (31, 508, 2101), (31, 509, 2102), (31, 510, 2102), (31, 511, 2103), (31, 512, 2103), (31, 513, 2104), (31, 514, 2104), (30, 515, 2106), (30, 516, 2106), (30, 517, 2107), (30, 518, 2108), (30, 519, 2108), (30, 520, 2109), (30, 521, 2109), (30, 522, 2110), (29, 523, 2112), (29, 524, 2112), (29, 525, 2113), (29, 526, 2114), (29, 527, 2114), (29, 528, 2115), (29, 529, 2116), (29, 530, 2116), (28, 531, 2118), (28, 532, 2119), (28, 533, 2119), (28, 534, 2120), (28, 535, 2121), (28, 536, 2121), (28, 537, 2121), (28, 538, 2122), (28, 539, 2122), (28, 540, 2122), (28, 541, 2123), (28, 542, 2123), (27, 543, 2124), (27, 544, 2125), (27, 545, 2125), (27, 546, 2125), (27, 547, 2126), (27, 548, 2126), (27, 549, 2126), (27, 550, 2127), (27, 551, 2127), (27, 552, 2127), (27, 553, 2127), (27, 554, 2128), (27, 555, 2128), (27, 556, 2128), (27, 557, 2129), (27, 558, 2129), (27, 559, 2129), (27, 560, 2130), (26, 561, 2131), (26, 562, 2131), (26, 563, 2131), (26, 564, 2132), (26, 565, 2132), (26, 566, 2132), (26, 567, 2132), (26, 568, 2133), (26, 569, 2133), (26, 570, 2133), (26, 571, 2133), (26, 572, 2134), (26, 573, 2134), (26, 574, 2134), (26, 575, 2134), (26, 576, 2135), (26, 577, 2135), (26, 578, 2135), (25, 579, 2136), (25, 580, 2137), (25, 581, 2137), (25, 582, 2137), (25, 583, 2137), (25, 584, 2138), (25, 585, 2138), (25, 586, 2138), (25, 587, 2138), (25, 588, 2139), (25, 589, 2139), (25, 590, 2139), (25, 591, 2139), (25, 592, 2140), (25, 593, 2140), (25, 594, 2140), (25, 595, 2140), (25, 596, 2140), (25, 597, 2141), (24, 598, 2142), (24, 599, 2142), (24, 600, 2142), (24, 601, 2142), (24, 602, 2143), (24, 603, 2143), (24, 604, 2143), (24, 605, 2143), (24, 606, 2143), (24, 607, 2144), (24, 608, 2144), (24, 609, 2144), (24, 610, 2144), (24, 611, 2144), (24, 612, 2144), (24, 613, 2145), (24, 614, 2145), (24, 615, 2145), (24, 616, 2145), (24, 617, 2145), (24, 618, 2145), (23, 619, 2146), (23, 620, 2146), (23, 621, 2146), (23, 622, 2146), (23, 623, 2146), (23, 624, 2146), (23, 625, 2146), (23, 626, 2146), (23, 627, 2146), (23, 628, 2146), (23, 629, 2147), (23, 630, 2147), (23, 631, 2147), (23, 632, 2147), (23, 633, 2147), (23, 634, 2147), (23, 635, 2147), (23, 636, 2147), (23, 637, 2147), (23, 638, 2147), (23, 639, 2147), (23, 640, 2147), (23, 641, 2147), (23, 642, 2147), (22, 643, 2148), (22, 644, 2148), (22, 645, 2148), (22, 646, 2149), (22, 647, 2149), (22, 648, 2149), (22, 649, 2149), (22, 650, 2149), (22, 651, 2149), (22, 652, 2149), (22, 653, 2149), (22, 654, 2149), (22, 655, 2149), (22, 656, 2149), (22, 657, 2149), (22, 658, 2149), (22, 659, 2149), (22, 660, 2149), (22, 661, 2149), (22, 662, 2150), (22, 663, 2150), (22, 664, 2150), (22, 665, 2150), (22, 666, 2150), (22, 667, 2150), (21, 668, 2151), (21, 669, 2151), (21, 670, 2151), (21, 671, 2151), (21, 672, 2151), (21, 673, 2151), (21, 674, 2151), (21, 675, 2151), (21, 676, 2151), (21, 677, 2151), (21, 678, 2151), (21, 679, 2151), (21, 680, 2152), (21, 681, 2152), (21, 682, 2152), (21, 683, 2152), (21, 684, 2152), (21, 685, 2152), (21, 686, 2152), (21, 687, 2152), (21, 688, 2152), (21, 689, 2152), (21, 690, 2152), (21, 691, 2152), (21, 692, 2152), (21, 693, 2152), (21, 694, 2152), (21, 695, 2151), (21, 696, 2151), (22, 697, 2150), (22, 698, 2150), (22, 699, 2150), (22, 700, 2150), (22, 701, 2150), (22, 702, 2150), (22, 703, 2150), (22, 704, 2150), (22, 705, 2150), (22, 706, 2150), (22, 707, 2150), (22, 708, 2150), (22, 709, 2150), (23, 710, 2149), (23, 711, 2149), (23, 712, 2149), (23, 713, 2149), (23, 714, 2149), (23, 715, 2149), (23, 716, 2149), (23, 717, 2149), (23, 718, 2148), (23, 719, 2148), (23, 720, 2148), (23, 721, 2148), (24, 722, 2147), (24, 723, 2147), (24, 724, 2147), (24, 725, 2147), (24, 726, 2147), (24, 727, 2147), (24, 728, 2147), (24, 729, 2147), (24, 730, 2147), (24, 731, 2147), (24, 732, 2147), (24, 733, 2147), (25, 734, 2146), (25, 735, 2146), (25, 736, 2146), (25, 737, 2146), (25, 738, 2146), (25, 739, 2145), (25, 740, 2145), (25, 741, 2145), (25, 742, 2145), (25, 743, 2145), (25, 744, 2145), (25, 745, 2145), (26, 746, 2144), (26, 747, 2144), (26, 748, 2144), (26, 749, 2144), (26, 750, 2144), (26, 751, 2144), (26, 752, 2144), (26, 753, 2144), (26, 754, 2144), (26, 755, 2144), (26, 756, 2144), (27, 757, 2143), (27, 758, 2143), (27, 759, 2143), (27, 760, 2143), (27, 761, 2142), (27, 762, 2142), (27, 763, 2142), (27, 764, 2142), (27, 765, 2142), (27, 766, 2142), (27, 767, 2142), (27, 768, 2142), (27, 769, 2142), (27, 770, 2142), (27, 771, 2142), (27, 772, 2142), (27, 773, 2142), (27, 774, 2142), (27, 775, 2142), (27, 776, 2142), (27, 777, 2142), (27, 778, 2142), (27, 779, 2142), (27, 780, 2141), (27, 781, 2141), (27, 782, 2141), (27, 783, 2141), (27, 784, 2141), (27, 785, 2141), (27, 786, 2141), (27, 787, 2141), (27, 788, 2141), (27, 789, 2141), (26, 790, 2142), (26, 791, 2142), (26, 792, 2142), (26, 793, 2142), (26, 794, 2142), (26, 795, 2142), (26, 796, 2142), (26, 797, 2142), (26, 798, 2141), (26, 799, 2141), (26, 800, 2141), (26, 801, 2141), (26, 802, 2141), (26, 803, 2141), (26, 804, 2141), (26, 805, 2141), (26, 806, 2141), (26, 807, 2141), (26, 808, 2141), (26, 809, 2141), (26, 810, 2141), (26, 811, 2141), (26, 812, 2141), (26, 813, 2141), (26, 814, 2141), (26, 815, 2141), (26, 816, 2140), (26, 817, 2140), (26, 818, 2140), (26, 819, 2140), (26, 820, 2140), (26, 821, 2140), (26, 822, 2140), (26, 823, 2140), (26, 824, 2140), (26, 825, 2140), (26, 826, 2140), (26, 827, 2140), (26, 828, 2140), (26, 829, 2140), (26, 830, 2140), (26, 831, 2140), (26, 832, 2140), (26, 833, 2140), (26, 834, 2139), (26, 835, 2139), (26, 836, 2139), (26, 837, 2139), (26, 838, 2139), (26, 839, 2139), (26, 840, 2138), (26, 841, 2138), (26, 842, 2138), (26, 843, 2137), (26, 844, 2137), (26, 845, 2137), (26, 846, 2136), (26, 847, 2136), (26, 848, 2135), (26, 849, 2135), (26, 850, 2135), (26, 851, 2134), (26, 852, 2134), (26, 853, 2133), (27, 854, 2132), (27, 855, 2131), (27, 856, 2131), (27, 857, 2130), (27, 858, 2130), (27, 859, 2130), (27, 860, 2129), (27, 861, 2129), (27, 862, 2128), (27, 863, 2128), (27, 864, 2127), (27, 865, 2127), (27, 866, 2126), (27, 867, 2126), (27, 868, 2125), (27, 869, 2125), (27, 870, 2124), (27, 871, 2123), (27, 872, 2123), (28, 873, 2121), (28, 874, 2121), (28, 875, 2120), (28, 876, 2120), (28, 877, 2119), (28, 878, 2118), (28, 879, 2118), (28, 880, 2117), (28, 881, 2117), (28, 882, 2116), (28, 883, 2116), (28, 884, 2115), (28, 885, 2114), (28, 886, 2114), (28, 887, 2113), (28, 888, 2113), (28, 889, 2112), (28, 890, 2112), (28, 891, 2111), (29, 892, 2110), (29, 893, 2109), (29, 894, 2109), (29, 895, 2108), (29, 896, 2108), (29, 897, 2107), (29, 898, 2107), (29, 899, 2107), (29, 900, 2106), (29, 901, 2106), (29, 902, 2105), (29, 903, 2105), (29, 904, 2104), (29, 905, 2104), (29, 906, 2104), (29, 907, 2103), (29, 908, 2103), (29, 909, 2102), (30, 910, 2101), (30, 911, 2101), (30, 912, 2100), (30, 913, 2100), (30, 914, 2099), (30, 915, 2099), (30, 916, 2099), (30, 917, 2099), (30, 918, 2098), (30, 919, 2098), (30, 920, 2098), (29, 921, 2099), (29, 922, 2098), (29, 923, 2098), (29, 924, 2098), (29, 925, 2098), (29, 926, 2097), (29, 927, 2097), (29, 928, 2097), (29, 929, 2097), (29, 930, 2097), (29, 931, 2096), (29, 932, 2096), (29, 933, 2096), (29, 934, 2096), (29, 935, 2095), (29, 936, 2095), (29, 937, 2095), (29, 938, 2095), (29, 939, 2094), (29, 940, 2094), (29, 941, 2094), (29, 942, 2094), (29, 943, 2094), (29, 944, 2093), (29, 945, 2093), (29, 946, 2093), (29, 947, 2093), (29, 948, 2093), (28, 949, 2093), (28, 950, 2093), (28, 951, 2093), (28, 952, 2093), (28, 953, 2093), (28, 954, 2092), (28, 955, 2092), (28, 956, 2092), (28, 957, 2092), (28, 958, 2092), (28, 959, 2091), (28, 960, 2091), (28, 961, 2091), (28, 962, 2091), (28, 963, 2091), (28, 964, 2090), (28, 965, 2090), (28, 966, 2090), (28, 967, 2090), (28, 968, 2090), (28, 969, 2089), (28, 970, 2089), (28, 971, 2089), (28, 972, 2089), (28, 973, 2089), (28, 974, 2089), (28, 975, 2088), (28, 976, 2088), (28, 977, 2088), (28, 978, 2088), (27, 979, 2089), (27, 980, 2089), (27, 981, 2088), (27, 982, 2088), (27, 983, 2088), (27, 984, 2088), (27, 985, 2088), (27, 986, 2088), (27, 987, 2087), (27, 988, 2087), (27, 989, 2087), (27, 990, 2087), (27, 991, 2086), (27, 992, 2086), (27, 993, 2086), (27, 994, 2086), (27, 995, 2085), (27, 996, 2085), (27, 997, 2085), (27, 998, 2084), (27, 999, 2084), (27, 1000, 2084), (28, 1001, 2082), (28, 1002, 2082), (28, 1003, 2082), (28, 1004, 2081), (28, 1005, 2081), (28, 1006, 2081), (28, 1007, 2080), (28, 1008, 2080), (28, 1009, 2080), (28, 1010, 2079), (28, 1011, 2079), (28, 1012, 2078), (28, 1013, 2078), (28, 1014, 2078), (28, 1015, 2077), (28, 1016, 2077), (28, 1017, 2077), (28, 1018, 2076), (28, 1019, 2076), (28, 1020, 2076), (28, 1021, 2075), (28, 1022, 2075), (28, 1023, 2074), (28, 1024, 2074), (28, 1025, 2074), (28, 1026, 2073), (28, 1027, 2073), (28, 1028, 2073), (28, 1029, 2072), (29, 1030, 2071), (29, 1031, 2070), (29, 1032, 2070), (29, 1033, 2069), (29, 1034, 2069), (29, 1035, 2069), (29, 1036, 2068), (29, 1037, 2068), (29, 1038, 2067), (29, 1039, 2067), (29, 1040, 2067), (29, 1041, 2066), (29, 1042, 2066), (29, 1043, 2065), (29, 1044, 2065), (29, 1045, 2064), (29, 1046, 2064), (29, 1047, 2063), (29, 1048, 2063), (29, 1049, 2063), (29, 1050, 2062), (29, 1051, 2062), (29, 1052, 2061), (29, 1053, 2061), (29, 1054, 2060), (29, 1055, 2060), (29, 1056, 2059), (29, 1057, 2059), (29, 1058, 2058), (30, 1059, 2057), (30, 1060, 2056), (30, 1061, 2056), (30, 1062, 2055), (30, 1063, 2055), (30, 1064, 2054), (30, 1065, 2054), (30, 1066, 2053), (30, 1067, 2052), (30, 1068, 2052), (30, 1069, 2051), (30, 1070, 2050), (30, 1071, 2049), (30, 1072, 2048), (30, 1073, 2048), (30, 1074, 2047), (30, 1075, 2046), (30, 1076, 2045), (30, 1077, 2044), (30, 1078, 2043), (30, 1079, 2042), (29, 1080, 2042), (29, 1081, 2041), (29, 1082, 2040), (29, 1083, 2039), (29, 1084, 2038), (29, 1085, 2037), (29, 1086, 2036), (29, 1087, 2035), (29, 1088, 2034), (29, 1089, 2033), (29, 1090, 2032), (29, 1091, 2031), (29, 1092, 2030), (29, 1093, 2029), (29, 1094, 2028), (29, 1095, 2027), (29, 1096, 2026), (29, 1097, 2026), (29, 1098, 2025), (29, 1099, 2024), (29, 1100, 2023), (29, 1101, 2022), (29, 1102, 2021), (29, 1103, 2021), (29, 1104, 2020), (29, 1105, 2019), (29, 1106, 2019), (29, 1107, 2018), (29, 1108, 2017), (29, 1109, 2016), (29, 1110, 2016), (29, 1111, 2015), (29, 1112, 2014), (29, 1113, 2014), (29, 1114, 2013), (29, 1115, 2013), (29, 1116, 2012), (29, 1117, 2011), (29, 1118, 2011), (29, 1119, 2010), (29, 1120, 2010), (29, 1121, 2009), (29, 1122, 2009), (29, 1123, 2008), (29, 1124, 2007), (29, 1125, 2007), (29, 1126, 2006), (29, 1127, 2006), (29, 1128, 2005), (29, 1129, 2005), (29, 1130, 2004), (29, 1131, 2004), (29, 1132, 2003), (29, 1133, 2003), (28, 1134, 2004), (28, 1135, 2003), (28, 1136, 2003), (28, 1137, 2002), (28, 1138, 2002), (28, 1139, 2001), (28, 1140, 2001), (28, 1141, 2001), (28, 1142, 2000), (28, 1143, 2000), (28, 1144, 2000), (28, 1145, 1999), (28, 1146, 1999), (28, 1147, 1999), (28, 1148, 1998), (28, 1149, 1998), (28, 1150, 1998), (28, 1151, 1997), (29, 1152, 1996), (29, 1153, 1996), (29, 1154, 1995), (29, 1155, 1995), (29, 1156, 1995), (29, 1157, 1994), (29, 1158, 1994), (29, 1159, 1994), (29, 1160, 1993), (29, 1161, 1993), (29, 1162, 1993), (29, 1163, 1992), (29, 1164, 1992), (29, 1165, 1991), (29, 1166, 1991), (29, 1167, 1991), (29, 1168, 1990), (29, 1169, 1990), (29, 1170, 1989), (29, 1171, 1989), (29, 1172, 1989), (29, 1173, 1988), (29, 1174, 1988), (29, 1175, 1987), (29, 1176, 1987), (29, 1177, 1987), (29, 1178, 1986), (29, 1179, 1986), (29, 1180, 1985), (29, 1181, 1985), (29, 1182, 1985), (29, 1183, 1984), (29, 1184, 1984), (29, 1185, 1983), (29, 1186, 1983), (29, 1187, 1982), (29, 1188, 1982), (29, 1189, 1981), (29, 1190, 1981), (29, 1191, 1981), (29, 1192, 1980), (29, 1193, 1980), (29, 1194, 1979), (29, 1195, 1979), (29, 1196, 1978), (29, 1197, 1978), (29, 1198, 1977), (29, 1199, 1977), (29, 1200, 1976), (29, 1201, 1976), (29, 1202, 1975), (29, 1203, 1975), (29, 1204, 1974), (29, 1205, 1974), (29, 1206, 1973), (29, 1207, 1973), (29, 1208, 1972), (29, 1209, 1971), (29, 1210, 1971), (29, 1211, 1970), (29, 1212, 1970), (29, 1213, 1969), (29, 1214, 1969), (29, 1215, 1968), (29, 1216, 1967), (29, 1217, 1967), (29, 1218, 1966), (29, 1219, 1965), (29, 1220, 1965), (29, 1221, 1964), (29, 1222, 1963), (29, 1223, 1963), (29, 1224, 1962), (29, 1225, 1961), (29, 1226, 1960), (29, 1227, 1960), (29, 1228, 1959), (29, 1229, 1958), (29, 1230, 1957), (29, 1231, 1956), (29, 1232, 1955), (29, 1233, 1955), (29, 1234, 1954), (29, 1235, 1953), (29, 1236, 1952), (29, 1237, 1951), (29, 1238, 1951), (29, 1239, 1950), (29, 1240, 1949), (29, 1241, 1948), (29, 1242, 1948), (29, 1243, 1947), (29, 1244, 1946), (30, 1245, 1945), (30, 1246, 1944), (30, 1247, 1943), (30, 1248, 1943), (30, 1249, 1942), (30, 1250, 1941), (30, 1251, 1941), (30, 1252, 1940), (30, 1253, 1940), (30, 1254, 1939), (30, 1255, 1938), (30, 1256, 1938), (30, 1257, 1937), (30, 1258, 1937), (30, 1259, 1936), (30, 1260, 1936), (30, 1261, 1935), (30, 1262, 1935), (30, 1263, 1934), (30, 1264, 1934), (30, 1265, 1933), (30, 1266, 1933), (30, 1267, 1932), (30, 1268, 1932), (30, 1269, 1931), (30, 1270, 1931), (30, 1271, 1930), (30, 1272, 1930), (30, 1273, 1929), (30, 1274, 1929), (30, 1275, 1929), (30, 1276, 1928), (30, 1277, 1928), (30, 1278, 1927), (30, 1279, 1927), (30, 1280, 1927), (30, 1281, 1926), (30, 1282, 1926), (30, 1283, 1925), (30, 1284, 1925), (30, 1285, 1925), (30, 1286, 1924), (30, 1287, 1924), (30, 1288, 1924), (30, 1289, 1923), (30, 1290, 1923), (30, 1291, 1923), (30, 1292, 1922), (30, 1293, 1922), (30, 1294, 1922), (30, 1295, 1921), (30, 1296, 1921), (30, 1297, 1921), (30, 1298, 1921), (30, 1299, 1920), (30, 1300, 1920), (30, 1301, 1920), (30, 1302, 1920), (30, 1303, 1920), (30, 1304, 1919), (30, 1305, 1919), (30, 1306, 1919), (30, 1307, 1919), (30, 1308, 1918), (30, 1309, 1918), (30, 1310, 1918), (30, 1311, 1918), (30, 1312, 1917), (31, 1313, 1916), (31, 1314, 1916), (31, 1315, 1916), (31, 1316, 1915), (31, 1317, 1915), (31, 1318, 1915), (31, 1319, 1915), (31, 1320, 1914), (31, 1321, 1914), (31, 1322, 1914), (31, 1323, 1914), (31, 1324, 1913), (31, 1325, 1913), (31, 1326, 1913), (31, 1327, 1912), (31, 1328, 1912), (31, 1329, 1912), (31, 1330, 1912), (31, 1331, 1911), (31, 1332, 1911), (31, 1333, 1911), (31, 1334, 1911), (31, 1335, 1910), (31, 1336, 1910), (31, 1337, 1910), (31, 1338, 1910), (31, 1339, 1909), (31, 1340, 1909), (32, 1341, 1908), (32, 1342, 1907), (32, 1343, 1907), (32, 1344, 1907), (32, 1345, 1907), (32, 1346, 1906), (32, 1347, 1906), (32, 1348, 1906), (32, 1349, 1905), (32, 1350, 1905), (32, 1351, 1905), (32, 1352, 1904), (32, 1353, 1904), (32, 1354, 1904), (32, 1355, 1904), (32, 1356, 1903), (32, 1357, 1903), (32, 1358, 1903), (32, 1359, 1902), (32, 1360, 1902), (32, 1361, 1902), (32, 1362, 1901), (32, 1363, 1901), (32, 1364, 1901), (32, 1365, 1900), (32, 1366, 1900), (33, 1367, 1899), (33, 1368, 1899), (33, 1369, 1898), (33, 1370, 1898), (33, 1371, 1897), (33, 1372, 1897), (33, 1373, 1896), (33, 1374, 1896), (33, 1375, 1895), (33, 1376, 1895), (33, 1377, 1894), (33, 1378, 1894), (33, 1379, 1893), (33, 1380, 1892), (33, 1381, 1892), (34, 1382, 1890), (34, 1383, 1890), (34, 1384, 1889), (34, 1385, 1888), (34, 1386, 1888), (34, 1387, 1887), (34, 1388, 1887), (34, 1389, 1886), (34, 1390, 1885), (34, 1391, 1884), (34, 1392, 1884), (34, 1393, 1883), (34, 1394, 1882), (34, 1395, 1882), (35, 1396, 1880), (35, 1397, 1879), (35, 1398, 1878), (35, 1399, 1878), (35, 1400, 1877), (35, 1401, 1876), (35, 1402, 1875), (35, 1403, 1874), (35, 1404, 1873), (35, 1405, 1872), (35, 1406, 1872), (35, 1407, 1871), (35, 1408, 1870), (35, 1409, 1869), (36, 1410, 1867), (36, 1411, 1866), (36, 1412, 1865), (36, 1413, 1864), (36, 1414, 1863), (36, 1415, 1863), (36, 1416, 1862), (36, 1417, 1861), (36, 1418, 1860), (36, 1419, 1859), (36, 1420, 1859), (36, 1421, 1858), (36, 1422, 1857), (37, 1423, 1855), (37, 1424, 1855), (37, 1425, 1854), (37, 1426, 1853), (37, 1427, 1853), (37, 1428, 1852), (37, 1429, 1852), (37, 1430, 1851), (37, 1431, 1850), (37, 1432, 1850), (37, 1433, 1849), (37, 1434, 1848), (38, 1435, 1847), (38, 1436, 1846), (38, 1437, 1846), (38, 1438, 1845), (38, 1439, 1845), (38, 1440, 1844), (38, 1441, 1844), (38, 1442, 1843), (38, 1443, 1843), (38, 1444, 1842), (38, 1445, 1842), (38, 1446, 1842), (38, 1447, 1841), (38, 1448, 1841), (38, 1449, 1841), (38, 1450, 1841), (38, 1451, 1840), (38, 1452, 1840), (38, 1453, 1840), (38, 1454, 1840), (39, 1455, 1839), (39, 1456, 1838), (39, 1457, 1838), (39, 1458, 1838), (39, 1459, 1838), (39, 1460, 1838), (39, 1461, 1837), (39, 1462, 1837), (39, 1463, 1837), (39, 1464, 1837), (39, 1465, 1836), (39, 1466, 1836), (39, 1467, 1836), (39, 1468, 1836), (39, 1469, 1836), (39, 1470, 1835), (39, 1471, 1835), (39, 1472, 1835), (39, 1473, 1835), (39, 1474, 1835), (39, 1475, 1834), (39, 1476, 1834), (39, 1477, 1834), (39, 1478, 1834), (39, 1479, 1834), (39, 1480, 1834), (39, 1481, 1833), (39, 1482, 1833), (39, 1483, 1833), (39, 1484, 1833), (39, 1485, 1833), (39, 1486, 1832), (39, 1487, 1832), (39, 1488, 1832), (39, 1489, 1832), (39, 1490, 1832), (39, 1491, 1831), (39, 1492, 1831), (39, 1493, 1831), (39, 1494, 1831), (39, 1495, 1831), (39, 1496, 1831), (40, 1497, 1829), (40, 1498, 1829), (40, 1499, 1829), (40, 1500, 1829), (40, 1501, 1829), (40, 1502, 1829), (40, 1503, 1828), (40, 1504, 1828), (40, 1505, 1828), (40, 1506, 1828), (40, 1507, 1828), (40, 1508, 1828), (40, 1509, 1827), (40, 1510, 1827), (40, 1511, 1827), (40, 1512, 1827), (40, 1513, 1827), (40, 1514, 1827), (40, 1515, 1826), (40, 1516, 1826), (40, 1517, 1826), (40, 1518, 1826), (40, 1519, 1826), (40, 1520, 1826), (40, 1521, 1825), (40, 1522, 1825), (40, 1523, 1825), (40, 1524, 1825), (40, 1525, 1825), (40, 1526, 1825), (40, 1527, 1825), (40, 1528, 1825), (40, 1529, 1824), (40, 1530, 1824), (40, 1531, 1824), (40, 1532, 1824), (40, 1533, 1824), (40, 1534, 1824), (40, 1535, 1824), (40, 1536, 1824), (40, 1537, 1823), (40, 1538, 1823), (40, 1539, 1823), (41, 1540, 1822), (41, 1541, 1822), (41, 1542, 1822), (41, 1543, 1822), (41, 1544, 1822), (41, 1545, 1821), (41, 1546, 1821), (41, 1547, 1821), (41, 1548, 1821), (41, 1549, 1821), (41, 1550, 1821), (41, 1551, 1821), (41, 1552, 1821), (41, 1553, 1820), (41, 1554, 1820), (41, 1555, 1820), (41, 1556, 1820), (41, 1557, 1820), (41, 1558, 1820), (41, 1559, 1820), (41, 1560, 1820), (41, 1561, 1819), (41, 1562, 1819), (41, 1563, 1819), (41, 1564, 1819), (41, 1565, 1819), (41, 1566, 1819), (41, 1567, 1819), (41, 1568, 1819), (41, 1569, 1818), (41, 1570, 1818), (41, 1571, 1818), (41, 1572, 1818), (41, 1573, 1818), (41, 1574, 1818), (41, 1575, 1818), (41, 1576, 1818), (41, 1577, 1817), (41, 1578, 1817), (41, 1579, 1817), (41, 1580, 1817), (41, 1581, 1817), (41, 1582, 1817), (41, 1583, 1817), (41, 1584, 1816), (42, 1585, 1815), (42, 1586, 1815), (42, 1587, 1815), (42, 1588, 1815), (42, 1589, 1815), (42, 1590, 1815), (42, 1591, 1815), (42, 1592, 1814), (42, 1593, 1814), (42, 1594, 1814), (42, 1595, 1814), (42, 1596, 1814), (42, 1597, 1814), (42, 1598, 1814), (42, 1599, 1814), (42, 1600, 1813), (42, 1601, 1813), (42, 1602, 1813), (41, 1603, 1814), (41, 1604, 1814), (41, 1605, 1814), (41, 1606, 1814), (41, 1607, 1814), (41, 1608, 1814), (41, 1609, 1813), (41, 1610, 1813), (41, 1611, 1813), (41, 1612, 1813), (41, 1613, 1813), (41, 1614, 1813), (41, 1615, 1813), (41, 1616, 1813), (41, 1617, 1813), (41, 1618, 1812), (41, 1619, 1812), (41, 1620, 1812), (41, 1621, 1812), (41, 1622, 1812), (41, 1623, 1812), (41, 1624, 1812), (41, 1625, 1812), (40, 1626, 1812), (40, 1627, 1812), (40, 1628, 1812), (40, 1629, 1812), (40, 1630, 1812), (40, 1631, 1812), (40, 1632, 1812), (40, 1633, 1812), (40, 1634, 1811), (40, 1635, 1811), (40, 1636, 1811), (40, 1637, 1811), (40, 1638, 1811), (40, 1639, 1811), (40, 1640, 1811), (40, 1641, 1811), (40, 1642, 1810), (40, 1643, 1810), (40, 1644, 1810), (40, 1645, 1810), (40, 1646, 1810), (40, 1647, 1810), (40, 1648, 1810), (40, 1649, 1809), (40, 1650, 1809), (39, 1651, 1810), (39, 1652, 1810), (39, 1653, 1810), (39, 1654, 1810), (39, 1655, 1810), (39, 1656, 1810), (39, 1657, 1809), (39, 1658, 1809), (39, 1659, 1809), (39, 1660, 1809), (39, 1661, 1809), (39, 1662, 1809), (39, 1663, 1809), (39, 1664, 1808), (39, 1665, 1808), (39, 1666, 1808), (39, 1667, 1808), (39, 1668, 1808), (39, 1669, 1808), (39, 1670, 1808), (39, 1671, 1807), (39, 1672, 1807), (39, 1673, 1807), (39, 1674, 1807), (39, 1675, 1806), (39, 1676, 1806), (39, 1677, 1806), (40, 1678, 1805), (40, 1679, 1804), (40, 1680, 1804), (40, 1681, 1804), (40, 1682, 1804), (40, 1683, 1803), (41, 1684, 1802), (41, 1685, 1802), (41, 1686, 1802), (41, 1687, 1801), (41, 1688, 1801), (41, 1689, 1801), (41, 1690, 1801), (42, 1691, 1799), (42, 1692, 1799), (42, 1693, 1799), (42, 1694, 1798), (42, 1695, 1798), (42, 1696, 1798), (43, 1697, 1797), (43, 1698, 1796), (43, 1699, 1796), (43, 1700, 1796), (43, 1701, 1795), (43, 1702, 1795), (44, 1703, 1794), (44, 1704, 1793), (44, 1705, 1793), (44, 1706, 1793), (44, 1707, 1793), (45, 1708, 1791), (45, 1709, 1791), (45, 1710, 1791), (45, 1711, 1790), (45, 1712, 1790), (45, 1713, 1790), (46, 1714, 1788), (46, 1715, 1788), (46, 1716, 1788), (46, 1717, 1787), (46, 1718, 1787), (47, 1719, 1785), (47, 1720, 1785), (47, 1721, 1785), (47, 1722, 1784), (47, 1723, 1784), (48, 1724, 1783), (48, 1725, 1782), (48, 1726, 1782), (48, 1727, 1782), (48, 1728, 1781), (49, 1729, 1780), (49, 1730, 1779), (49, 1731, 1779), (49, 1732, 1779), (49, 1733, 1778), (50, 1734, 1777), (50, 1735, 1776), (50, 1736, 1776), (50, 1737, 1776), (50, 1738, 1775), (51, 1739, 1774), (51, 1740, 1773), (51, 1741, 1773), (51, 1742, 1772), (51, 1743, 1772), (52, 1744, 1771), (52, 1745, 1770), (52, 1746, 1770), (52, 1747, 1769), (52, 1748, 1769), (52, 1749, 1768), (52, 1750, 1768), (52, 1751, 1768), (52, 1752, 1767), (52, 1753, 1767), (53, 1754, 1765), (53, 1755, 1765), (53, 1756, 1765), (53, 1757, 1764), (53, 1758, 1764), (53, 1759, 1763), (53, 1760, 1763), (53, 1761, 1763), (53, 1762, 1762), (53, 1763, 1762), (53, 1764, 1762), (53, 1765, 1761), (53, 1766, 1761), (53, 1767, 1760), (53, 1768, 1760), (53, 1769, 1760), (53, 1770, 1759), (53, 1771, 1759), (53, 1772, 1759), (53, 1773, 1758), (53, 1774, 1758), (53, 1775, 1758), (53, 1776, 1757), (53, 1777, 1757), (53, 1778, 1757), (53, 1779, 1756), (53, 1780, 1756), (53, 1781, 1756), (53, 1782, 1755), (53, 1783, 1755), (53, 1784, 1755), (53, 1785, 1754), (53, 1786, 1754), (53, 1787, 1754), (53, 1788, 1753), (53, 1789, 1753), (53, 1790, 1753), (53, 1791, 1753), (53, 1792, 1752), (53, 1793, 1752), (53, 1794, 1752), (53, 1795, 1751), (53, 1796, 1751), (53, 1797, 1751), (53, 1798, 1750), (53, 1799, 1750), (53, 1800, 1750), (53, 1801, 1750), (53, 1802, 1749), (53, 1803, 1749), (53, 1804, 1749), (53, 1805, 1748), (53, 1806, 1748), (53, 1807, 1748), (53, 1808, 1748), (53, 1809, 1747), (53, 1810, 1747), (53, 1811, 1747), (53, 1812, 1746), (53, 1813, 1746), (53, 1814, 1746), (53, 1815, 1746), (54, 1816, 1744), (54, 1817, 1744), (54, 1818, 1744), (54, 1819, 1744), (54, 1820, 1743), (54, 1821, 1743), (54, 1822, 1743), (54, 1823, 1743), (54, 1824, 1742), (54, 1825, 1742), (55, 1826, 1741), (55, 1827, 1740), (56, 1828, 1739), (56, 1829, 1739), (57, 1830, 1737), (57, 1831, 1737), (58, 1832, 1736), (58, 1833, 1735), (59, 1834, 1734), (59, 1835, 1734), (60, 1836, 1732), (60, 1837, 1732), (61, 1838, 1731), (61, 1839, 1730), (61, 1840, 1730), (62, 1841, 1729), (62, 1842, 1728), (63, 1843, 1727), (63, 1844, 1726), (64, 1845, 1725), (64, 1846, 1725), (65, 1847, 1723), (65, 1848, 1723), (66, 1849, 1722), (66, 1850, 1721), (67, 1851, 1720), (67, 1852, 1720), (67, 1853, 1719), (68, 1854, 1718), (68, 1855, 1718), (69, 1856, 1716), (69, 1857, 1716), (70, 1858, 1714), (70, 1859, 1714), (71, 1860, 1713), (71, 1861, 1712), (71, 1862, 1712), (72, 1863, 1710), (72, 1864, 1710), (73, 1865, 1709), (73, 1866, 1708), (74, 1867, 1707), (74, 1868, 1707), (75, 1869, 1705), (75, 1870, 1705), (75, 1871, 1704), (76, 1872, 1703), (76, 1873, 1703), (77, 1874, 1701), (77, 1875, 1701), (78, 1876, 1699), (78, 1877, 1699), (79, 1878, 1698), (79, 1879, 1697), (79, 1880, 1697), (80, 1881, 1695), (80, 1882, 1695), (81, 1883, 1693), (81, 1884, 1693), (81, 1885, 1693), (82, 1886, 1691), (82, 1887, 1691), (83, 1888, 1689), (83, 1889, 1689), (84, 1890, 1687), (84, 1891, 1687), (84, 1892, 1687), (85, 1893, 1685), (85, 1894, 1685), (86, 1895, 1683), (86, 1896, 1683), (87, 1897, 1681), (87, 1898, 1681), (87, 1899, 1680), (88, 1900, 1679), (88, 1901, 1678), (88, 1902, 1678), (89, 1903, 1677), (89, 1904, 1676), (90, 1905, 1675), (90, 1906, 1674), (90, 1907, 1674), (91, 1908, 1672), (91, 1909, 1672), (91, 1910, 1671), (92, 1911, 1670), (92, 1912, 1669), (93, 1913, 1667), (93, 1914, 1667), (93, 1915, 1666), (94, 1916, 1665), (94, 1917, 1664), (95, 1918, 1663), (95, 1919, 1662), (96, 1920, 1661), (96, 1921, 1660), (97, 1922, 1658), (97, 1923, 1658), (97, 1924, 1657), (98, 1925, 1656), (98, 1926, 1655), (99, 1927, 1653), (99, 1928, 1653), (100, 1929, 1651), (100, 1930, 1651), (101, 1931, 1649), (101, 1932, 1648), (102, 1933, 1647), (102, 1934, 1646), (103, 1935, 1644), (103, 1936, 1644), (104, 1937, 1642), (105, 1938, 1640), (105, 1939, 1640), (106, 1940, 1638), (106, 1941, 1637), (107, 1942, 1635), (107, 1943, 1635), (108, 1944, 1633), (109, 1945, 1631), (109, 1946, 1630), (110, 1947, 1628), (110, 1948, 1628), (111, 1949, 1626), (112, 1950, 1624), (112, 1951, 1623), (113, 1952, 1622), (114, 1953, 1620), (114, 1954, 1619), (115, 1955, 1617), (116, 1956, 1616), (116, 1957, 1615), (117, 1958, 1613), (118, 1959, 1611), (119, 1960, 1610), (119, 1961, 1609), (120, 1962, 1607), (121, 1963, 87), (209, 1963, 1517), (122, 1964, 74), (216, 1964, 1510), (123, 1965, 61), (223, 1965, 1502), (123, 1966, 51), (230, 1966, 1494), (124, 1967, 40), (237, 1967, 1487), (125, 1968, 30), (244, 1968, 1479), (126, 1969, 21), (251, 1969, 1471), (127, 1970, 12), (259, 1970, 1463), (128, 1971, 4), (266, 1971, 1455), (274, 1972, 1446), (281, 1973, 1439), (289, 1974, 1430), (294, 1975, 1424), (299, 1976, 1418), (303, 1977, 1413), (308, 1978, 1407), (314, 1979, 1400), (319, 1980, 1394), (325, 1981, 1387), (331, 1982, 1380), (337, 1983, 1373), (344, 1984, 1365), (351, 1985, 1357), (358, 1986, 1349), (366, 1987, 1339), (372, 1988, 1332), (376, 1989, 1327), (380, 1990, 1322), (384, 1991, 1317), (388, 1992, 1311), (393, 1993, 1305), (397, 1994, 1300), (401, 1995, 1294), (406, 1996, 1288), (411, 1997, 1282), (415, 1998, 1276), (420, 1999, 1270), (425, 2000, 1263), (430, 2001, 1257), (435, 2002, 1251), (440, 2003, 1244), (445, 2004, 1237), (451, 2005, 1230), (456, 2006, 1223), (461, 2007, 1217), (466, 2008, 1210), (471, 2009, 1203), (476, 2010, 1197), (481, 2011, 1190), (486, 2012, 1183), (492, 2013, 1175), (497, 2014, 1168), (503, 2015, 1160), (508, 2016, 1074), (514, 2017, 1063), (519, 2018, 1054), (525, 2019, 1044), (531, 2020, 1034), (536, 2021, 1025), (541, 2022, 1017), (546, 2023, 1008), (551, 2024, 999), (556, 2025, 990), (561, 2026, 982), (565, 2027, 974), (570, 2028, 966), (574, 2029, 958), (578, 2030, 951), (583, 2031, 942), (587, 2032, 935), (591, 2033, 928), (595, 2034, 921), (599, 2035, 913), (603, 2036, 906), (607, 2037, 899), (611, 2038, 892), (614, 2039, 884), (617, 2040, 870), (619, 2041, 858), (622, 2042, 845), (624, 2043, 833), (627, 2044, 821), (629, 2045, 809), (631, 2046, 798), (634, 2047, 786), (636, 2048, 781), (638, 2049, 775), (640, 2050, 770), (642, 2051, 765), (643, 2052, 761), (645, 2053, 756), (647, 2054, 751), (649, 2055, 746), (650, 2056, 742), (652, 2057, 737), (654, 2058, 731), (656, 2059, 726), (658, 2060, 721), (660, 2061, 715), (662, 2062, 709), (664, 2063, 704), (666, 2064, 698), (668, 2065, 692), (671, 2066, 685), (673, 2067, 679), (675, 2068, 673), (678, 2069, 666), (680, 2070, 660), (683, 2071, 652), (685, 2072, 645), (688, 2073, 636), (691, 2074, 628), (694, 2075, 619), (699, 2076, 609), (703, 2077, 599), (708, 2078, 588), (712, 2079, 579), (717, 2080, 568), (721, 2081, 558), (725, 2082, 548), (730, 2083, 536), (734, 2084, 526), (738, 2085, 516), (743, 2086, 506), (747, 2087, 497), (751, 2088, 488), (755, 2089, 479), (760, 2090, 469), (764, 2091, 460), (768, 2092, 451), (772, 2093, 443), (776, 2094, 434), (779, 2095, 427), (782, 2096, 419), (786, 2097, 411), (789, 2098, 404), (792, 2099, 397), (795, 2100, 390), (799, 2101, 382), (802, 2102, 375), (805, 2103, 370), (808, 2104, 364), (811, 2105, 359), (814, 2106, 354), (818, 2107, 347), (821, 2108, 342), (824, 2109, 337), (827, 2110, 332), (830, 2111, 327), (833, 2112, 322), (836, 2113, 317), (839, 2114, 312), (842, 2115, 307), (845, 2116, 302), (848, 2117, 297), (851, 2118, 292), (854, 2119, 288), (857, 2120, 283), (860, 2121, 278), (863, 2122, 273), (866, 2123, 269), (869, 2124, 264), (872, 2125, 259), (875, 2126, 255), (877, 2127, 251), (880, 2128, 246), (883, 2129, 242), (886, 2130, 237), (889, 2131, 232), (893, 2132, 226), (896, 2133, 221), (899, 2134, 215), (903, 2135, 209), (906, 2136, 204), (909, 2137, 199), (913, 2138, 193), (917, 2139, 186), (920, 2140, 181), (924, 2141, 174), (928, 2142, 166), (932, 2143, 154), (936, 2144, 142), (946, 2145, 124), (956, 2146, 106), (967, 2147, 87), (978, 2148, 67), (989, 2149, 48), (1001, 2150, 27), (1013, 2151, 6)], ['1001,2150,936,2144,685,2072,582,2030,365,1986,215,1963,128,1971,103,1936,53,1815,39,1677,39,1455,29,1244,27,757,21,696,27,543,39,458,93,308,126,270,210,206,291,179,363,135,520,121,1430,121,1584,128,1663,142,1768,178,1904,204,2011,293,2094,411,2148,535,2167,607,2171,717,2165,833,2128,914,2112,994,2081,1068,2031,1132,1967,1255,1931,1368,1879,1444,1846,1670,1785,1855,1719,1973,1662,2015,1582,2015,1497,2039,1420,2046,1329,2072,1177,2101,1105,2138'])], 'temp/1748568965_1112947_917877156_a9c2d4b99270c9302def4ed40606e685.jpg']} nb pixel non reg : 3692295 nb pixel common : 3691469 proportion of common points : 0.9997762908976666 [('test release memory', 'SUCCESS', True), ('test detect objet', 'SUCCESS', True), ('test polygone', 'SUCCESS', True)] res_total : True #&_# TEST SUCCEEDED #&_# : tests/mask_test #&_# /home/admin/workarea/git/Velours/python/tests/python_tests.py refs/heads/master_0b6e4cc05cd91ca7b75e9aba8c17ac40bcdc3de3 SQL :INSERT INTO MTRAdmin.monitor_sys (name, type, server, version_code, result_str, result_bool, lien , test_group ,test_name) VALUES ('python_test3','1','marlene','refs/heads/master_0b6e4cc05cd91ca7b75e9aba8c17ac40bcdc3de3','{"mask_detection": "success"}','1','http://marlene.fotonower-preprod.com/job/2025/May/30052025/python_test3//data_2/data_log/job/2025/May/30052025/python_test3/log-python3----short_python3--v--marlene-03:35:00.txt','mask_detection','unknown'); #&_# END OF TEST #&_# : tests/mask_test #&_# #&_# BEGIN OF TEST : tests/datou_test #&_# /home/admin/workarea/git/Velours/python/tests/datou_test.py Datou All Test python version used : 3 ############################### TEST sam ################################ TEST SAM Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=4573 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=4573 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 4573 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=4573 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : sam list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (1189321094) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 1189321094 download finish for photo 1189321094 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.19739985466003418 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! WARNING : we have an input that is not a photo, we should get rid of it Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:sam Fri May 30 03:36:51 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569011_1112947_1189321094_9626af7f95d010f2a4fd524688d4ea22_76896585.png': 1189321094} map_photo_id_path_extension : {1189321094: {'path': 'temp/1748569011_1112947_1189321094_9626af7f95d010f2a4fd524688d4ea22_76896585.png', 'extension': 'png'}} map_subphoto_mainphoto : {} Beginning of datou step sam ! pht : 4677 Inside sam : nb paths : 1 (640, 960, 3) ERROR in datou_step_exec, will save and exit ! CUDA out of memory. Tried to allocate 768.00 MiB (GPU 0; 10.76 GiB total capacity; 443.59 MiB already allocated; 445.88 MiB free; 498.00 MiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF File "/home/admin/workarea/git/Velours/python/mtr/datou/datou_lib.py", line 2329, in datou_exec output = datou_step_exec(sNext, args, cache, context, map_info, verbose, mtr_user_id) File "/home/admin/workarea/git/Velours/python/mtr/datou/datou_lib.py", line 2430, in datou_step_exec return lib_process.datou_step_sam(param, json_param, args, cache, context, map_info, verbose) File "/home/admin/workarea/git/Velours/python/mtr/datou/lib_step_exec/lib_step_process.py", line 396, in datou_step_sam masks = mask_generator.generate(image) File "/home/admin/.local/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "/home/admin/workarea/install/segment-anything/segment_anything/automatic_mask_generator.py", line 163, in generate mask_data = self._generate_masks(image) File "/home/admin/workarea/install/segment-anything/segment_anything/automatic_mask_generator.py", line 206, in _generate_masks crop_data = self._process_crop(image, crop_box, layer_idx, orig_size) File "/home/admin/workarea/install/segment-anything/segment_anything/automatic_mask_generator.py", line 236, in _process_crop self.predictor.set_image(cropped_im) File "/home/admin/workarea/install/segment-anything/segment_anything/predictor.py", line 60, in set_image self.set_torch_image(input_image_torch, image.shape[:2]) File "/home/admin/.local/lib/python3.8/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context return func(*args, **kwargs) File "/home/admin/workarea/install/segment-anything/segment_anything/predictor.py", line 89, in set_torch_image self.features = self.model.image_encoder(input_image) File "/home/admin/.local/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/admin/workarea/install/segment-anything/segment_anything/modeling/image_encoder.py", line 112, in forward x = blk(x) File "/home/admin/.local/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/admin/workarea/install/segment-anything/segment_anything/modeling/image_encoder.py", line 174, in forward x = self.attn(x) File "/home/admin/.local/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1130, in _call_impl return forward_call(*input, **kwargs) File "/home/admin/workarea/install/segment-anything/segment_anything/modeling/image_encoder.py", line 231, in forward attn = (q * self.scale) @ k.transpose(-2, -1) [1189321094] map_info['map_portfolio_photo'] : {} final : True mtd_id 4573 list_pids : [1189321094] begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4573', None, '1189321094', "[>, , , , , 'CUDA out of memory. Tried to allocate 768.00 MiB (GPU 0; 10.76 GiB total capacity; 443.59 MiB already allocated; 445.88 MiB free; 498.00 MiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF']", '-1', '-1.0', '501120777', '1.0', None)] time used for this insertion : 0.0171051025390625 save_final ERROR in last step sam, CUDA out of memory. Tried to allocate 768.00 MiB (GPU 0; 10.76 GiB total capacity; 443.59 MiB already allocated; 445.88 MiB free; 498.00 MiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF time spend for datou_step_exec : 6.323915004730225 time spend to save output : 0.01938652992248535 total time spend for step 0 : 6.34330153465271 need to delete datou_research and reload, so keep current state 1 need to delete datou_research and reload, so keep current state 1 need to delete datou_research and reload, so keep current state 1 caffe_path_current : About to save ! 2 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : None ERROR nb objects espect : 98 nb_objects detect : 0 ERROR sam FAILED ############################### TEST frcnn ################################ test frcnn Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=4184 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=4184 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 4184 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=4184 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : frcnn list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (917754606) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 917754606 download finish for photo 917754606 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.11984634399414062 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:frcnn Fri May 30 03:36:58 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569018_1112947_917754606_35f3c9ae49686a6be16030c6ec25c9ee.jpg': 917754606} map_photo_id_path_extension : {917754606: {'path': 'temp/1748569018_1112947_917754606_35f3c9ae49686a6be16030c6ec25c9ee.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step Faster rcnn ! classes : ['background', 'plaque'] pht : 4370 caffemodel_name (should be vgg16_immat_307 but not used because net loaded outside in the fonction) : {'id': 3375, 'mtr_user_id': 31, 'name': 'detection_plaque_valcor_010622', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'background,plaque', 'svm_portfolios_learning': '0,0', 'photo_hashtag_type': 4370, 'photo_desc_type': 5676, 'type_classification': 'caffe_faster_rcnn', 'hashtag_id_list': '0,0'} To loadFromThcl() model_param file didn't exist model_name : detection_plaque_valcor_010622 model_type : caffe_faster_rcnn list file need : ['caffemodel', 'test.prototxt'] file exist in s3 : ['caffemodel', 'test.prototxt'] file manque in s3 : [] local folder : /data/models_weight/detection_plaque_valcor_010622 /data/models_weight/detection_plaque_valcor_010622/caffemodel size_local : 349723073 size in s3 : 349723073 create time local : 2022-07-12 14:12:27 create time in s3 : 2022-06-01 15:05:56 caffemodel already exist and didn't need to update /data/models_weight/detection_plaque_valcor_010622/test.prototxt size_local : 7163 size in s3 : 7163 create time local : 2022-07-12 14:12:27 create time in s3 : 2022-06-01 15:05:55 test.prototxt already exist and didn't need to update prototxt : /data/models_weight/detection_plaque_valcor_010622/test.prototxt caffemodel : /data/models_weight/detection_plaque_valcor_010622/caffemodel Loaded network /data/models_weight/detection_plaque_valcor_010622/caffemodel About to compute detect_faster_rcnn : len(args) : 1 Inside frcnn step exec : nb paths : 1 image_path : temp/1748569018_1112947_917754606_35f3c9ae49686a6be16030c6ec25c9ee.jpg image_size (600, 800, 3) [[[ 4 6 6] [ 5 7 7] [ 6 8 8] ... [207 215 214] [206 214 213] [206 214 213]] [[ 4 6 6] [ 5 7 7] [ 6 8 8] ... [207 215 214] [206 214 213] [206 214 213]] [[ 4 6 6] [ 5 7 7] [ 6 8 8] ... [207 215 214] [206 214 213] [206 214 213]] ... [[ 14 16 16] [ 13 15 15] [ 11 13 13] ... [198 206 205] [198 206 205] [198 206 205]] [[ 16 18 18] [ 14 16 16] [ 11 13 13] ... [206 214 213] [206 214 213] [206 214 213]] [[ 13 15 15] [ 12 14 14] [ 9 11 11] ... [210 218 217] [210 218 217] [210 218 217]]] Detection took 0.078s for 300 object proposals c : plaque list_crops.shape (72, 5) proba : 0.063839585 (374.12692, 293.9194, 430.8102, 317.8087) proba : 0.052226253 (382.178, 297.18765, 552.3607, 344.65793) proba : 0.012270755 (345.35675, 272.4301, 468.85754, 320.72437) We are managing local photo_id len de result frcnn : 1 After datou_step_exec type output : time spend for datou_step_exec : 2.5490405559539795 time spend to save output : 0.00016260147094726562 total time spend for step 1 : 2.5492031574249268 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True Inside saveFrcnn : final : True verbose : True threshold to save the result : 0.1 output flattener : [(0, 493029425, 4370, 374, 430, 293, 317, 0.063839585, None), (0, 493029425, 4370, 382, 552, 297, 344, 0.052226253, None), (0, 493029425, 4370, 345, 468, 272, 320, 0.012270755, None)] Warning : no hashtag_ids to insert in the database final : True begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4184', None, '917754606', '0', 0, '0', 493061979, '0', None)] time used for this insertion : 0.01743483543395996 [917754606] map_info['map_portfolio_photo'] : {} final : True mtd_id 4184 list_pids : [917754606] Looping around the photos to save general results len do output : 1 /0 before output type Managing all output in save final without adding information in the mtr_datou_result ('4184', None, None, None, None, None, None, None, None) ('4184', None, '917754606', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4184', None, '917754606', None, None, None, None, None, None)] time used for this insertion : 0.014480352401733398 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {0: [[(0, 493029425, 4370, 374, 430, 293, 317, 0.063839585, None), (0, 493029425, 4370, 382, 552, 297, 344, 0.052226253, None), (0, 493029425, 4370, 345, 468, 272, 320, 0.012270755, None)], 'temp/1748569018_1112947_917754606_35f3c9ae49686a6be16030c6ec25c9ee.jpg']} ############################### TEST thcl ################################ TEST THCL Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=2 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=2 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 2 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=2 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 1 thcl is not linked in the step_by_step architecture ! WARNING : step 2 argmax is not linked in the step_by_step architecture ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : thcl, argmax list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (916235064) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 916235064 download finish for photo 916235064 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.15030431747436523 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 2 step1:thcl Fri May 30 03:37:00 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg': 916235064} map_photo_id_path_extension : {916235064: {'path': 'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step Thcl ! multi_thcl or not :False multi_thcl_cond or not :False dic_thcl : {'355': 1} we are using the classfication for only one thcl 355 In convert_file_to_np l 337 : 1 l343 1 l357 after caffe.io.load_image dimension du image : (3, (66, 66, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.007599830627441406 time to convert the images to numpy array : 0.0011324882507324219 total time to convert the images to numpy array : 0.009020805358886719 list photo_ids error: [] list photo_ids correct : [916235064] number of photos to traite : 1 try to delete the photos incorrect in DB tagging for thcl : 355 To do loadFromThcl(), then load ParamDescType : thcl355 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (355) thcls : [{'id': 355, 'mtr_user_id': 31, 'name': 'car_360_1027', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'c_elysee_1027_gao__port_506302,mokka_1027_gao__port_506374,captur_1027_gao__port_506399,sorento_1027_gao__port_506192,navara_1027_gao__port_506205,xc90_1027_gao__port_506350,saxo_1027_gao__port_506052,trafic_1027_gao__port_506295,punto_evo_1027_gao__port_506066,5_1027_gao__port_506117,250_1027_gao__port_506065,d_max_1027_gao__port_506125,panamera_1027_gao__port_506387,alhambra_1027_gao__port_506381,x6_1027_gao__port_506349,vitara_1027_gao__port_506328,fiesta_1027_gao__port_506377,qashqai_1027_gao__port_506286,147_1027_gao__port_506124,c5_1027_gao__port_506172,q5_1027_gao__port_506206,giulia_1027_gao__port_506178,karl_1027_gao__port_506371,mehari_1027_gao__port_506076,911_1027_gao__port_506114,508_1027_gao__port_506329,idea_1027_gao__port_506122,megane_1027_gao__port_506220,ghibli_1027_gao__port_506174,touareg_1027_gao__port_506224,i10_1027_gao__port_506232,jumper_1027_gao__port_506234,classe_clk_1027_gao__port_506173,kuga_1027_gao__port_506181,ct_1027_gao__port_506323,leon_1027_gao__port_506326,ds5_1027_gao__port_506376,cordoba_1027_gao__port_506048,classe_cla_1027_gao__port_506400,jumpy_1027_gao__port_506179,avensis_1027_gao__port_506311,juke_1027_gao__port_506325,4008_1027_gao__port_506402,190_series_1027_gao__port_506051,serie_3_1027_gao__port_506294,q7_1027_gao__port_506318,glc_1027_gao__port_506303,grand_vitara_1027_gao__port_506175,s40_1027_gao__port_506099,toledo_1027_gao__port_506061,5008_1027_gao__port_506337,continental_1027_gao__port_506250,coupe_1027_gao__port_506082,iq_1027_gao__port_506166,407_1027_gao__port_506133,touran_1027_gao__port_506308,300c_1027_gao__port_506078,classe_gl_1027_gao__port_506340,vivaro_1027_gao__port_506310,sl_1027_gao__port_506100,elise_1027_gao__port_506121,1007_1027_gao__port_506070,i40_1027_gao__port_506218,bipper_tepee_1027_gao__port_506227,focus_1027_gao__port_506272,primera_1027_gao__port_506147,r4_1027_gao__port_506160,a8_1027_gao__port_506265,boxer_1027_gao__port_506202,s5_1027_gao__port_506222,r21_1027_gao__port_506093,c3_1027_gao__port_506257,santa_fe_1027_gao__port_506208,m4_1027_gao__port_506344,safrane_1027_gao__port_506077,classe_gle_1027_gao__port_506395,0_1027_gao__port_506094,ix35_1027_gao__port_506219,carens_1027_gao__port_506298,classe_a_1027_gao__port_506339,ix20_1027_gao__port_506343,note_1027_gao__port_506365,a5_1027_gao__port_506200,sx4_1027_gao__port_506348,sandero_1027_gao__port_506198,3008_1027_gao__port_506385,q50_1027_gao__port_506239,latitude_1027_gao__port_506236,v40_1027_gao__port_506391,xsara_1027_gao__port_506087,grand_c_max_1027_gao__port_506342,swift_1027_gao__port_506149,serie_1_1027_gao__port_506184,xc70_1027_gao__port_506393,master_1027_gao__port_506203,clio_1027_gao__port_506280,duster_1027_gao__port_506216,traveller_1027_gao__port_506403,tipo_1027_gao__port_506355,rav_4_1027_gao__port_506332,coccinelle_1027_gao__port_506259,spacetourer_1027_gao__port_506401,xe_1027_gao__port_506357,ds3_1027_gao__port_506324,mx_5_1027_gao__port_506098,land_cruiser_1027_gao__port_506315,classe_b_1027_gao__port_506335,806_1027_gao__port_506088,rx_8_1027_gao__port_506046,spark_1027_gao__port_506185,6_1027_gao__port_506171,bravo_1027_gao__port_506080,nx_1027_gao__port_506345,sharan_1027_gao__port_506347,x_type_1027_gao__port_506067,jimny_1027_gao__port_506233,wrangler_1027_gao__port_506225,c_crosser_1027_gao__port_506312,v70_1027_gao__port_506278,classe_e_1027_gao__port_506300,classe_v_1027_gao__port_506258,m3_1027_gao__port_506182,abarth_500_1027_gao__port_506226,serie_6_1027_gao__port_506262,modus_1027_gao__port_506146,3_1027_gao__port_506113,405_1027_gao__port_506108,allroad_1027_gao__port_506297,auris_1027_gao__port_506322,galaxy_1027_gao__port_506143,giulietta_1027_gao__port_506363,106_1027_gao__port_506073,classe_m_1027_gao__port_506154,espace_1027_gao__port_506313,panda_1027_gao__port_506189,rcz_1027_gao__port_506197,4007_1027_gao__port_506162,classe_cl_1027_gao__port_506249,leaf_1027_gao__port_506139,octavia_1027_gao__port_506237,ds4_1027_gao__port_506336,freelander_1027_gao__port_506084,evasion_1027_gao__port_506109,punto_1027_gao__port_506106,2cv_1027_gao__port_506045,x4_1027_gao__port_506392,antara_1027_gao__port_506247,murano_1027_gao__port_506316,alto_1027_gao__port_506201,meriva_1027_gao__port_506353,orlando_1027_gao__port_506305,new_beetle_1027_gao__port_506050,306_1027_gao__port_506145,tiguan_1027_gao__port_506362,s_type_1027_gao__port_506101,c1_1027_gao__port_506128,vectra_1027_gao__port_506044,outlander_1027_gao__port_506317,307_1027_gao__port_506074,a6_s6_1027_gao__port_506134,nemo_combi_1027_gao__port_506196,berlingo_1027_gao__port_506194,partner_1027_gao__port_506285,cayenne_1027_gao__port_506177,quattroporte_1027_gao__port_506240,c_max_1027_gao__port_506282,fabia_1027_gao__port_506396,cx_3_1027_gao__port_506281,x_trail_1027_gao__port_506264,scirocco_1027_gao__port_506276,matiz_1027_gao__port_506144,tigra_1027_gao__port_506069,escort_1027_gao__port_506091,c2_1027_gao__port_506081,mini_1027_gao__port_506168,i30_1027_gao__port_506291,picanto_1027_gao__port_506238,mito_1027_gao__port_506072,impreza_1027_gao__port_506085,kangoo_1027_gao__port_506235,a4_1027_gao__port_506193,cayman_1027_gao__port_506268,sportage_1027_gao__port_506148,up_1027_gao__port_506356,optima_1027_gao__port_506386,defender_1027_gao__port_506229,serie_2_1027_gao__port_506256,edge_1027_gao__port_506187,r19_1027_gao__port_506110,jetta_1027_gao__port_506304,eos_1027_gao__port_506115,accord_1027_gao__port_506214,yaris_1027_gao__port_506334,classe_cls_1027_gao__port_506289,polo_1027_gao__port_506361,serie_4_1027_gao__port_506366,mini_cabriolet_1027_gao__port_506204,prius_1027_gao__port_506190,lodgy_1027_gao__port_506188,serie_7_1027_gao__port_506307,c15_1027_gao__port_506055,kadjar_1027_gao__port_506389,insignia_1027_gao__port_506364,308_1027_gao__port_506279,roomster_1027_gao__port_506241,80_1027_gao__port_506057,309_1027_gao__port_506063,tucson_1027_gao__port_506320,x3_1027_gao__port_506212,xf_1027_gao__port_506263,2008_1027_gao__port_506394,passat_1027_gao__port_506306,compass_1027_gao__port_506260,twingo_1027_gao__port_506309,micra_1027_gao__port_506221,golf_1027_gao__port_506155,soul_1027_gao__port_506176,rapid_1027_gao__port_506398,forester_1027_gao__port_506360,slk_1027_gao__port_506210,forfour_1027_gao__port_506341,serie_5_1027_gao__port_506209,xj_1027_gao__port_506170,pajero_1027_gao__port_506097,agila_1027_gao__port_506119,a6_1027_gao__port_506163,fox_1027_gao__port_506092,boxster_1027_gao__port_506267,altea_1027_gao__port_506246,samurai_1027_gao__port_506047,trax_1027_gao__port_506296,getz_1027_gao__port_506058,cherokee_1027_gao__port_506269,koleos_1027_gao__port_506378,z_series_1027_gao__port_506123,ecosport_1027_gao__port_506271,space_star_1027_gao__port_506277,rs3_sportback_1027_gao__port_506207,civic_1027_gao__port_506141,talisman_1027_gao__port_506390,f_pace_1027_gao__port_506314,classe_c_1027_gao__port_506299,tt_1027_gao__port_506075,pathfinder_1027_gao__port_506183,156_1027_gao__port_506157,cx_5_1027_gao__port_506228,scenic_1027_gao__port_506255,yeti_1027_gao__port_506358,mustang_1027_gao__port_506053,stilo_1027_gao__port_506060,ateca_1027_gao__port_506382,fiorino_1027_gao__port_506217,classe_glk_1027_gao__port_506290,fortwo_1027_gao__port_506230,cruze_1027_gao__port_506186,107_1027_gao__port_506213,aygo_1027_gao__port_506248,rx_1027_gao__port_506354,500_1027_gao__port_506245,bora_1027_gao__port_506104,transit_1027_gao__port_506111,pt_cruiser_1027_gao__port_506054,patrol_1027_gao__port_506068,r8_1027_gao__port_506156,xm_1027_gao__port_506102,s60_1027_gao__port_506191,aveo_1027_gao__port_506158,captiva_1027_gao__port_506159,ax_1027_gao__port_506153,rexton_1027_gao__port_506107,camaro_1027_gao__port_506056,ypsilon_1027_gao__port_506131,delta_1027_gao__port_506165,c4_1027_gao__port_506370,zx_1027_gao__port_506161,verso_1027_gao__port_506242,superb_1027_gao__port_506327,r5_1027_gao__port_506253,caddy_1027_gao__port_506330,x5_1027_gao__port_506243,f_type_1027_gao__port_506231,fusion_1027_gao__port_506096,dokker_1027_gao__port_506331,205_1027_gao__port_506062,macan_1027_gao__port_506195,tourneo_1027_gao__port_506369,108_1027_gao__port_506384,9_3_1027_gao__port_506071,mondeo_1027_gao__port_506116,cr_v_1027_gao__port_506164,c30_1027_gao__port_506090,pulsar_1027_gao__port_506397,ibiza_1027_gao__port_506273,a1_1027_gao__port_506338,matrix_1027_gao__port_506140,carnival_1027_gao__port_506136,xantia_1027_gao__port_506086,terrano_1027_gao__port_506083,q3_1027_gao__port_506275,hr_v_1027_gao__port_506283,expert_1027_gao__port_506142,multivan_1027_gao__port_506383,venga_1027_gao__port_506380,scudo_1027_gao__port_506129,laguna_1027_gao__port_506368,vel_satis_1027_gao__port_506130,b_max_1027_gao__port_506367,ignis_1027_gao__port_506292,159_1027_gao__port_506064,grande_punto_1027_gao__port_506138,logan_1027_gao__port_506167,s_max_1027_gao__port_506223,caravelle_1027_gao__port_506351,adam_1027_gao__port_506079,406_1027_gao__port_506132,q30_1027_gao__port_506293,almera_1027_gao__port_506089,corsa_1027_gao__port_506095,corolla_1027_gao__port_506120,xc60_1027_gao__port_506388,viano_1027_gao__port_506211,pro_cee_d_1027_gao__port_506274,a3_1027_gao__port_506321,v50_1027_gao__port_506150,voyager_1027_gao__port_506169,corvette_1027_gao__port_506049,rio_1027_gao__port_506379,jazz_1027_gao__port_506252,200_1027_gao__port_506112,tts_1027_gao__port_506199,zafira_1027_gao__port_506287,asx_1027_gao__port_506266,607_1027_gao__port_506118,207_1027_gao__port_506103,classe_s_1027_gao__port_506301,c6_1027_gao__port_506105,express_1027_gao__port_506137,classe_gla_1027_gao__port_506352,v60_1027_gao__port_506333,ka_1027_gao__port_506180,range_rover_1027_gao__port_506254,discovery_1027_gao__port_506375,classe_r_1027_gao__port_506270,transporter_1027_gao__port_506319,cee_d_1027_gao__port_506288,zoe_1027_gao__port_506244,i20_1027_gao__port_506284,gtv_1027_gao__port_506059,s4_avant_1027_gao__port_506261,x1_1027_gao__port_506372,autres_1027_gao__port_506127,208_1027_gao__port_506359,c8_1027_gao__port_506135,astra_1027_gao__port_506215,2_1027_gao__port_506151,doblo_1027_gao__port_506251,807_1027_gao__port_506152,206_1027_gao__port_506126,a7_1027_gao__port_506373,renegade_1027_gao__port_506346', 'svm_portfolios_learning': '506302,506374,506399,506192,506205,506350,506052,506295,506066,506117,506065,506125,506387,506381,506349,506328,506377,506286,506124,506172,506206,506178,506371,506076,506114,506329,506122,506220,506174,506224,506232,506234,506173,506181,506323,506326,506376,506048,506400,506179,506311,506325,506402,506051,506294,506318,506303,506175,506099,506061,506337,506250,506082,506166,506133,506308,506078,506340,506310,506100,506121,506070,506218,506227,506272,506147,506160,506265,506202,506222,506093,506257,506208,506344,506077,506395,506094,506219,506298,506339,506343,506365,506200,506348,506198,506385,506239,506236,506391,506087,506342,506149,506184,506393,506203,506280,506216,506403,506355,506332,506259,506401,506357,506324,506098,506315,506335,506088,506046,506185,506171,506080,506345,506347,506067,506233,506225,506312,506278,506300,506258,506182,506226,506262,506146,506113,506108,506297,506322,506143,506363,506073,506154,506313,506189,506197,506162,506249,506139,506237,506336,506084,506109,506106,506045,506392,506247,506316,506201,506353,506305,506050,506145,506362,506101,506128,506044,506317,506074,506134,506196,506194,506285,506177,506240,506282,506396,506281,506264,506276,506144,506069,506091,506081,506168,506291,506238,506072,506085,506235,506193,506268,506148,506356,506386,506229,506256,506187,506110,506304,506115,506214,506334,506289,506361,506366,506204,506190,506188,506307,506055,506389,506364,506279,506241,506057,506063,506320,506212,506263,506394,506306,506260,506309,506221,506155,506176,506398,506360,506210,506341,506209,506170,506097,506119,506163,506092,506267,506246,506047,506296,506058,506269,506378,506123,506271,506277,506207,506141,506390,506314,506299,506075,506183,506157,506228,506255,506358,506053,506060,506382,506217,506290,506230,506186,506213,506248,506354,506245,506104,506111,506054,506068,506156,506102,506191,506158,506159,506153,506107,506056,506131,506165,506370,506161,506242,506327,506253,506330,506243,506231,506096,506331,506062,506195,506369,506384,506071,506116,506164,506090,506397,506273,506338,506140,506136,506086,506083,506275,506283,506142,506383,506380,506129,506368,506130,506367,506292,506064,506138,506167,506223,506351,506079,506132,506293,506089,506095,506120,506388,506211,506274,506321,506150,506169,506049,506379,506252,506112,506199,506287,506266,506118,506103,506301,506105,506137,506352,506333,506180,506254,506375,506270,506319,506288,506244,506284,506059,506261,506372,506127,506359,506135,506215,506151,506251,506152,506126,506373,506346', 'photo_hashtag_type': 332, 'photo_desc_type': 3390, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}] thcl {'id': 355, 'mtr_user_id': 31, 'name': 'car_360_1027', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'c_elysee_1027_gao__port_506302,mokka_1027_gao__port_506374,captur_1027_gao__port_506399,sorento_1027_gao__port_506192,navara_1027_gao__port_506205,xc90_1027_gao__port_506350,saxo_1027_gao__port_506052,trafic_1027_gao__port_506295,punto_evo_1027_gao__port_506066,5_1027_gao__port_506117,250_1027_gao__port_506065,d_max_1027_gao__port_506125,panamera_1027_gao__port_506387,alhambra_1027_gao__port_506381,x6_1027_gao__port_506349,vitara_1027_gao__port_506328,fiesta_1027_gao__port_506377,qashqai_1027_gao__port_506286,147_1027_gao__port_506124,c5_1027_gao__port_506172,q5_1027_gao__port_506206,giulia_1027_gao__port_506178,karl_1027_gao__port_506371,mehari_1027_gao__port_506076,911_1027_gao__port_506114,508_1027_gao__port_506329,idea_1027_gao__port_506122,megane_1027_gao__port_506220,ghibli_1027_gao__port_506174,touareg_1027_gao__port_506224,i10_1027_gao__port_506232,jumper_1027_gao__port_506234,classe_clk_1027_gao__port_506173,kuga_1027_gao__port_506181,ct_1027_gao__port_506323,leon_1027_gao__port_506326,ds5_1027_gao__port_506376,cordoba_1027_gao__port_506048,classe_cla_1027_gao__port_506400,jumpy_1027_gao__port_506179,avensis_1027_gao__port_506311,juke_1027_gao__port_506325,4008_1027_gao__port_506402,190_series_1027_gao__port_506051,serie_3_1027_gao__port_506294,q7_1027_gao__port_506318,glc_1027_gao__port_506303,grand_vitara_1027_gao__port_506175,s40_1027_gao__port_506099,toledo_1027_gao__port_506061,5008_1027_gao__port_506337,continental_1027_gao__port_506250,coupe_1027_gao__port_506082,iq_1027_gao__port_506166,407_1027_gao__port_506133,touran_1027_gao__port_506308,300c_1027_gao__port_506078,classe_gl_1027_gao__port_506340,vivaro_1027_gao__port_506310,sl_1027_gao__port_506100,elise_1027_gao__port_506121,1007_1027_gao__port_506070,i40_1027_gao__port_506218,bipper_tepee_1027_gao__port_506227,focus_1027_gao__port_506272,primera_1027_gao__port_506147,r4_1027_gao__port_506160,a8_1027_gao__port_506265,boxer_1027_gao__port_506202,s5_1027_gao__port_506222,r21_1027_gao__port_506093,c3_1027_gao__port_506257,santa_fe_1027_gao__port_506208,m4_1027_gao__port_506344,safrane_1027_gao__port_506077,classe_gle_1027_gao__port_506395,0_1027_gao__port_506094,ix35_1027_gao__port_506219,carens_1027_gao__port_506298,classe_a_1027_gao__port_506339,ix20_1027_gao__port_506343,note_1027_gao__port_506365,a5_1027_gao__port_506200,sx4_1027_gao__port_506348,sandero_1027_gao__port_506198,3008_1027_gao__port_506385,q50_1027_gao__port_506239,latitude_1027_gao__port_506236,v40_1027_gao__port_506391,xsara_1027_gao__port_506087,grand_c_max_1027_gao__port_506342,swift_1027_gao__port_506149,serie_1_1027_gao__port_506184,xc70_1027_gao__port_506393,master_1027_gao__port_506203,clio_1027_gao__port_506280,duster_1027_gao__port_506216,traveller_1027_gao__port_506403,tipo_1027_gao__port_506355,rav_4_1027_gao__port_506332,coccinelle_1027_gao__port_506259,spacetourer_1027_gao__port_506401,xe_1027_gao__port_506357,ds3_1027_gao__port_506324,mx_5_1027_gao__port_506098,land_cruiser_1027_gao__port_506315,classe_b_1027_gao__port_506335,806_1027_gao__port_506088,rx_8_1027_gao__port_506046,spark_1027_gao__port_506185,6_1027_gao__port_506171,bravo_1027_gao__port_506080,nx_1027_gao__port_506345,sharan_1027_gao__port_506347,x_type_1027_gao__port_506067,jimny_1027_gao__port_506233,wrangler_1027_gao__port_506225,c_crosser_1027_gao__port_506312,v70_1027_gao__port_506278,classe_e_1027_gao__port_506300,classe_v_1027_gao__port_506258,m3_1027_gao__port_506182,abarth_500_1027_gao__port_506226,serie_6_1027_gao__port_506262,modus_1027_gao__port_506146,3_1027_gao__port_506113,405_1027_gao__port_506108,allroad_1027_gao__port_506297,auris_1027_gao__port_506322,galaxy_1027_gao__port_506143,giulietta_1027_gao__port_506363,106_1027_gao__port_506073,classe_m_1027_gao__port_506154,espace_1027_gao__port_506313,panda_1027_gao__port_506189,rcz_1027_gao__port_506197,4007_1027_gao__port_506162,classe_cl_1027_gao__port_506249,leaf_1027_gao__port_506139,octavia_1027_gao__port_506237,ds4_1027_gao__port_506336,freelander_1027_gao__port_506084,evasion_1027_gao__port_506109,punto_1027_gao__port_506106,2cv_1027_gao__port_506045,x4_1027_gao__port_506392,antara_1027_gao__port_506247,murano_1027_gao__port_506316,alto_1027_gao__port_506201,meriva_1027_gao__port_506353,orlando_1027_gao__port_506305,new_beetle_1027_gao__port_506050,306_1027_gao__port_506145,tiguan_1027_gao__port_506362,s_type_1027_gao__port_506101,c1_1027_gao__port_506128,vectra_1027_gao__port_506044,outlander_1027_gao__port_506317,307_1027_gao__port_506074,a6_s6_1027_gao__port_506134,nemo_combi_1027_gao__port_506196,berlingo_1027_gao__port_506194,partner_1027_gao__port_506285,cayenne_1027_gao__port_506177,quattroporte_1027_gao__port_506240,c_max_1027_gao__port_506282,fabia_1027_gao__port_506396,cx_3_1027_gao__port_506281,x_trail_1027_gao__port_506264,scirocco_1027_gao__port_506276,matiz_1027_gao__port_506144,tigra_1027_gao__port_506069,escort_1027_gao__port_506091,c2_1027_gao__port_506081,mini_1027_gao__port_506168,i30_1027_gao__port_506291,picanto_1027_gao__port_506238,mito_1027_gao__port_506072,impreza_1027_gao__port_506085,kangoo_1027_gao__port_506235,a4_1027_gao__port_506193,cayman_1027_gao__port_506268,sportage_1027_gao__port_506148,up_1027_gao__port_506356,optima_1027_gao__port_506386,defender_1027_gao__port_506229,serie_2_1027_gao__port_506256,edge_1027_gao__port_506187,r19_1027_gao__port_506110,jetta_1027_gao__port_506304,eos_1027_gao__port_506115,accord_1027_gao__port_506214,yaris_1027_gao__port_506334,classe_cls_1027_gao__port_506289,polo_1027_gao__port_506361,serie_4_1027_gao__port_506366,mini_cabriolet_1027_gao__port_506204,prius_1027_gao__port_506190,lodgy_1027_gao__port_506188,serie_7_1027_gao__port_506307,c15_1027_gao__port_506055,kadjar_1027_gao__port_506389,insignia_1027_gao__port_506364,308_1027_gao__port_506279,roomster_1027_gao__port_506241,80_1027_gao__port_506057,309_1027_gao__port_506063,tucson_1027_gao__port_506320,x3_1027_gao__port_506212,xf_1027_gao__port_506263,2008_1027_gao__port_506394,passat_1027_gao__port_506306,compass_1027_gao__port_506260,twingo_1027_gao__port_506309,micra_1027_gao__port_506221,golf_1027_gao__port_506155,soul_1027_gao__port_506176,rapid_1027_gao__port_506398,forester_1027_gao__port_506360,slk_1027_gao__port_506210,forfour_1027_gao__port_506341,serie_5_1027_gao__port_506209,xj_1027_gao__port_506170,pajero_1027_gao__port_506097,agila_1027_gao__port_506119,a6_1027_gao__port_506163,fox_1027_gao__port_506092,boxster_1027_gao__port_506267,altea_1027_gao__port_506246,samurai_1027_gao__port_506047,trax_1027_gao__port_506296,getz_1027_gao__port_506058,cherokee_1027_gao__port_506269,koleos_1027_gao__port_506378,z_series_1027_gao__port_506123,ecosport_1027_gao__port_506271,space_star_1027_gao__port_506277,rs3_sportback_1027_gao__port_506207,civic_1027_gao__port_506141,talisman_1027_gao__port_506390,f_pace_1027_gao__port_506314,classe_c_1027_gao__port_506299,tt_1027_gao__port_506075,pathfinder_1027_gao__port_506183,156_1027_gao__port_506157,cx_5_1027_gao__port_506228,scenic_1027_gao__port_506255,yeti_1027_gao__port_506358,mustang_1027_gao__port_506053,stilo_1027_gao__port_506060,ateca_1027_gao__port_506382,fiorino_1027_gao__port_506217,classe_glk_1027_gao__port_506290,fortwo_1027_gao__port_506230,cruze_1027_gao__port_506186,107_1027_gao__port_506213,aygo_1027_gao__port_506248,rx_1027_gao__port_506354,500_1027_gao__port_506245,bora_1027_gao__port_506104,transit_1027_gao__port_506111,pt_cruiser_1027_gao__port_506054,patrol_1027_gao__port_506068,r8_1027_gao__port_506156,xm_1027_gao__port_506102,s60_1027_gao__port_506191,aveo_1027_gao__port_506158,captiva_1027_gao__port_506159,ax_1027_gao__port_506153,rexton_1027_gao__port_506107,camaro_1027_gao__port_506056,ypsilon_1027_gao__port_506131,delta_1027_gao__port_506165,c4_1027_gao__port_506370,zx_1027_gao__port_506161,verso_1027_gao__port_506242,superb_1027_gao__port_506327,r5_1027_gao__port_506253,caddy_1027_gao__port_506330,x5_1027_gao__port_506243,f_type_1027_gao__port_506231,fusion_1027_gao__port_506096,dokker_1027_gao__port_506331,205_1027_gao__port_506062,macan_1027_gao__port_506195,tourneo_1027_gao__port_506369,108_1027_gao__port_506384,9_3_1027_gao__port_506071,mondeo_1027_gao__port_506116,cr_v_1027_gao__port_506164,c30_1027_gao__port_506090,pulsar_1027_gao__port_506397,ibiza_1027_gao__port_506273,a1_1027_gao__port_506338,matrix_1027_gao__port_506140,carnival_1027_gao__port_506136,xantia_1027_gao__port_506086,terrano_1027_gao__port_506083,q3_1027_gao__port_506275,hr_v_1027_gao__port_506283,expert_1027_gao__port_506142,multivan_1027_gao__port_506383,venga_1027_gao__port_506380,scudo_1027_gao__port_506129,laguna_1027_gao__port_506368,vel_satis_1027_gao__port_506130,b_max_1027_gao__port_506367,ignis_1027_gao__port_506292,159_1027_gao__port_506064,grande_punto_1027_gao__port_506138,logan_1027_gao__port_506167,s_max_1027_gao__port_506223,caravelle_1027_gao__port_506351,adam_1027_gao__port_506079,406_1027_gao__port_506132,q30_1027_gao__port_506293,almera_1027_gao__port_506089,corsa_1027_gao__port_506095,corolla_1027_gao__port_506120,xc60_1027_gao__port_506388,viano_1027_gao__port_506211,pro_cee_d_1027_gao__port_506274,a3_1027_gao__port_506321,v50_1027_gao__port_506150,voyager_1027_gao__port_506169,corvette_1027_gao__port_506049,rio_1027_gao__port_506379,jazz_1027_gao__port_506252,200_1027_gao__port_506112,tts_1027_gao__port_506199,zafira_1027_gao__port_506287,asx_1027_gao__port_506266,607_1027_gao__port_506118,207_1027_gao__port_506103,classe_s_1027_gao__port_506301,c6_1027_gao__port_506105,express_1027_gao__port_506137,classe_gla_1027_gao__port_506352,v60_1027_gao__port_506333,ka_1027_gao__port_506180,range_rover_1027_gao__port_506254,discovery_1027_gao__port_506375,classe_r_1027_gao__port_506270,transporter_1027_gao__port_506319,cee_d_1027_gao__port_506288,zoe_1027_gao__port_506244,i20_1027_gao__port_506284,gtv_1027_gao__port_506059,s4_avant_1027_gao__port_506261,x1_1027_gao__port_506372,autres_1027_gao__port_506127,208_1027_gao__port_506359,c8_1027_gao__port_506135,astra_1027_gao__port_506215,2_1027_gao__port_506151,doblo_1027_gao__port_506251,807_1027_gao__port_506152,206_1027_gao__port_506126,a7_1027_gao__port_506373,renegade_1027_gao__port_506346', 'svm_portfolios_learning': '506302,506374,506399,506192,506205,506350,506052,506295,506066,506117,506065,506125,506387,506381,506349,506328,506377,506286,506124,506172,506206,506178,506371,506076,506114,506329,506122,506220,506174,506224,506232,506234,506173,506181,506323,506326,506376,506048,506400,506179,506311,506325,506402,506051,506294,506318,506303,506175,506099,506061,506337,506250,506082,506166,506133,506308,506078,506340,506310,506100,506121,506070,506218,506227,506272,506147,506160,506265,506202,506222,506093,506257,506208,506344,506077,506395,506094,506219,506298,506339,506343,506365,506200,506348,506198,506385,506239,506236,506391,506087,506342,506149,506184,506393,506203,506280,506216,506403,506355,506332,506259,506401,506357,506324,506098,506315,506335,506088,506046,506185,506171,506080,506345,506347,506067,506233,506225,506312,506278,506300,506258,506182,506226,506262,506146,506113,506108,506297,506322,506143,506363,506073,506154,506313,506189,506197,506162,506249,506139,506237,506336,506084,506109,506106,506045,506392,506247,506316,506201,506353,506305,506050,506145,506362,506101,506128,506044,506317,506074,506134,506196,506194,506285,506177,506240,506282,506396,506281,506264,506276,506144,506069,506091,506081,506168,506291,506238,506072,506085,506235,506193,506268,506148,506356,506386,506229,506256,506187,506110,506304,506115,506214,506334,506289,506361,506366,506204,506190,506188,506307,506055,506389,506364,506279,506241,506057,506063,506320,506212,506263,506394,506306,506260,506309,506221,506155,506176,506398,506360,506210,506341,506209,506170,506097,506119,506163,506092,506267,506246,506047,506296,506058,506269,506378,506123,506271,506277,506207,506141,506390,506314,506299,506075,506183,506157,506228,506255,506358,506053,506060,506382,506217,506290,506230,506186,506213,506248,506354,506245,506104,506111,506054,506068,506156,506102,506191,506158,506159,506153,506107,506056,506131,506165,506370,506161,506242,506327,506253,506330,506243,506231,506096,506331,506062,506195,506369,506384,506071,506116,506164,506090,506397,506273,506338,506140,506136,506086,506083,506275,506283,506142,506383,506380,506129,506368,506130,506367,506292,506064,506138,506167,506223,506351,506079,506132,506293,506089,506095,506120,506388,506211,506274,506321,506150,506169,506049,506379,506252,506112,506199,506287,506266,506118,506103,506301,506105,506137,506352,506333,506180,506254,506375,506270,506319,506288,506244,506284,506059,506261,506372,506127,506359,506135,506215,506151,506251,506152,506126,506373,506346', 'photo_hashtag_type': 332, 'photo_desc_type': 3390, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'} Update svm_hashtag_type_desc : 3390 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (3390) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3390, 'car_360_1027', 16384, 25088, 'car_360_1027', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2017, 10, 28, 12, 29, 27), datetime.datetime(2017, 10, 28, 12, 29, 27)) To loadFromThcl() : net_3390 begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 5817 max_wait_temp : 1 max_wait : 0 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (3390) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3390, 'car_360_1027', 16384, 25088, 'car_360_1027', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2017, 10, 28, 12, 29, 27), datetime.datetime(2017, 10, 28, 12, 29, 27)) param : , param.caffemodel : car_360_1027 None mean_file_type : mean_file_path : prototxt_file_path : model : car_360_1027 Inside get_net Inside get_net before cache_data_model model_param file didn't exist Inside get_net before CDM.load_model_par_type model_name : car_360_1027 model_type : caffe list file need : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file exist in s3 : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file manque in s3 : [] local folder : /data/models_weight/car_360_1027 /data/models_weight/car_360_1027/caffemodel size_local : 542944640 size in s3 : 542944640 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:43 caffemodel already exist and didn't need to update /data/models_weight/car_360_1027/deploy_conv_normal.prototxt size_local : 4626 size in s3 : 4626 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:42 deploy_conv_normal.prototxt already exist and didn't need to update /data/models_weight/car_360_1027/deploy_fc.prototxt size_local : 1132 size in s3 : 1132 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:43 deploy_fc.prototxt already exist and didn't need to update /data/models_weight/car_360_1027/deploy.prototxt size_local : 5654 size in s3 : 5654 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:42 deploy.prototxt already exist and didn't need to update /data/models_weight/car_360_1027/mean.npy size_local : 1572944 size in s3 : 1572944 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:55 mean.npy already exist and didn't need to update /data/models_weight/car_360_1027/synset_words.txt size_local : 13687 size in s3 : 13687 create time local : 2021-08-09 05:28:34 create time in s3 : 2021-08-06 17:57:43 synset_words.txt already exist and didn't need to update Inside get_net after CDM.load_model_par_type After if not only_with_local_cache: /home/admin/workarea/install/caffe_cuda8_python3/python/:/home/admin/workarea/install/darknet/:/home/admin/workarea/git/Velours/python:/home/admin/workarea/install/caffe_frcnn_python3/py-faster-rcnn/caffe-fast-rcnn/python:/home/admin/mtr/.credentials:/home/admin/workarea/install/caffe/python:/home/admin/workarea/install/caffe_frcnn/py-faster-rcnn/tools/:/home/admin/workarea/git/fotonowerpip/:/home/admin/workarea/install/segment-anything:/home/admin//workarea/git/pyfvs/ Here before set mode gpu Doing nothing but we could set mode gpu after set mode gpu prototxt_filename : /data/models_weight/car_360_1027/deploy.prototxt caffemodel_filename : /data/models_weight/car_360_1027/caffemodel now we set caffe to gpu mode before predict begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 9500 max_wait_temp : 1 max_wait : 0 dict_keys(['prob', 'pool5']) time used to do the prepocess of the images : 0.01207876205444336 time used to do the prediction : 0.06095385551452637 save descriptor for thcl : 355 (1, 512, 7, 7) Got the blobs of the net to insert : [0, 0, 0, 0, 0, 0, 0, 0, 0, 0] code_as_byte_string:b'0000000000'| time to traite the descriptors : 0.051703453063964844 Testing : ['916235064'] In select_photos_meta_from_ids: SELECT photo_id, url, FROM_UNIXTIME(uploaded_at), latitude, longitude, text FROM MTRBack.photos WHERE photo_id IN (916235064) Catched exception ! Connect or reconnect ! result : {916235064: {'photo_id': 916235064, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2017/10/14/6293d1bb790dc6902450e7c572b7d10b.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': None}} list_photo_exists : [916235064] storage_type for insertDescriptorsMulti : 1 To insert : 916235064 time to insert the descriptors : 1.7028627395629883 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : False verbose : True time used to find the portfolios of the photos select button_legend_list from MTRDatou.classification_theme where id = 355 SAVE THCL, output : {'916235064': [[('916235064', 'c_elysee_1027_gao__port_506302', 0.0018814829, 332, '355'), ('916235064', 'mokka_1027_gao__port_506374', 0.0011635778, 332, '355'), ('916235064', 'captur_1027_gao__port_506399', 0.0008157971, 332, '355'), ('916235064', 'sorento_1027_gao__port_506192', 0.0011772993, 332, '355'), ('916235064', 'navara_1027_gao__port_506205', 0.0025849268, 332, '355'), ('916235064', 'xc90_1027_gao__port_506350', 0.004169867, 332, '355'), ('916235064', 'saxo_1027_gao__port_506052', 0.0034809553, 332, '355'), ('916235064', 'trafic_1027_gao__port_506295', 0.007366904, 332, '355'), ('916235064', 'punto_evo_1027_gao__port_506066', 0.0021888195, 332, '355'), ('916235064', '5_1027_gao__port_506117', 0.00057978963, 332, '355'), ('916235064', '250_1027_gao__port_506065', 0.004591475, 332, '355'), ('916235064', 'd_max_1027_gao__port_506125', 0.0031586797, 332, '355'), ('916235064', 'panamera_1027_gao__port_506387', 0.0022507673, 332, '355'), ('916235064', 'alhambra_1027_gao__port_506381', 0.005320502, 332, '355'), ('916235064', 'x6_1027_gao__port_506349', 0.0010999657, 332, '355'), ('916235064', 'vitara_1027_gao__port_506328', 0.005402428, 332, '355'), ('916235064', 'fiesta_1027_gao__port_506377', 0.0039188615, 332, '355'), ('916235064', 'qashqai_1027_gao__port_506286', 0.0014787969, 332, '355'), ('916235064', '147_1027_gao__port_506124', 0.0019778193, 332, '355'), ('916235064', 'c5_1027_gao__port_506172', 0.0012442436, 332, '355'), ('916235064', 'q5_1027_gao__port_506206', 0.0015050066, 332, '355'), ('916235064', 'giulia_1027_gao__port_506178', 0.00216944, 332, '355'), ('916235064', 'karl_1027_gao__port_506371', 0.0027081575, 332, '355'), ('916235064', 'mehari_1027_gao__port_506076', 0.0047041136, 332, '355'), ('916235064', '911_1027_gao__port_506114', 0.0019417465, 332, '355'), ('916235064', '508_1027_gao__port_506329', 0.0009585346, 332, '355'), ('916235064', 'idea_1027_gao__port_506122', 0.0007700529, 332, '355'), ('916235064', 'megane_1027_gao__port_506220', 0.0019469425, 332, '355'), ('916235064', 'ghibli_1027_gao__port_506174', 0.0013724434, 332, '355'), ('916235064', 'touareg_1027_gao__port_506224', 0.0016202505, 332, '355'), ('916235064', 'i10_1027_gao__port_506232', 0.0013924964, 332, '355'), ('916235064', 'jumper_1027_gao__port_506234', 0.010044858, 332, '355'), ('916235064', 'classe_clk_1027_gao__port_506173', 0.0010793374, 332, '355'), ('916235064', 'kuga_1027_gao__port_506181', 0.0008447343, 332, '355'), ('916235064', 'ct_1027_gao__port_506323', 0.0012521547, 332, '355'), ('916235064', 'leon_1027_gao__port_506326', 0.002584474, 332, '355'), ('916235064', 'ds5_1027_gao__port_506376', 0.0012430315, 332, '355'), ('916235064', 'cordoba_1027_gao__port_506048', 0.002865061, 332, '355'), ('916235064', 'classe_cla_1027_gao__port_506400', 0.001294973, 332, '355'), ('916235064', 'jumpy_1027_gao__port_506179', 0.010339059, 332, '355'), ('916235064', 'avensis_1027_gao__port_506311', 0.0018767802, 332, '355'), ('916235064', 'juke_1027_gao__port_506325', 0.0011343346, 332, '355'), ('916235064', '4008_1027_gao__port_506402', 0.0015757276, 332, '355'), ('916235064', '190_series_1027_gao__port_506051', 0.003980616, 332, '355'), ('916235064', 'serie_3_1027_gao__port_506294', 0.002874152, 332, '355'), ('916235064', 'q7_1027_gao__port_506318', 0.0023354774, 332, '355'), ('916235064', 'glc_1027_gao__port_506303', 0.0012106928, 332, '355'), ('916235064', 'grand_vitara_1027_gao__port_506175', 0.0011447335, 332, '355'), ('916235064', 's40_1027_gao__port_506099', 0.002233854, 332, '355'), ('916235064', 'toledo_1027_gao__port_506061', 0.0017465191, 332, '355'), ('916235064', '5008_1027_gao__port_506337', 0.004699296, 332, '355'), ('916235064', 'continental_1027_gao__port_506250', 0.002191374, 332, '355'), ('916235064', 'coupe_1027_gao__port_506082', 0.002263206, 332, '355'), ('916235064', 'iq_1027_gao__port_506166', 0.001817546, 332, '355'), ('916235064', '407_1027_gao__port_506133', 0.0009056496, 332, '355'), ('916235064', 'touran_1027_gao__port_506308', 0.0020402444, 332, '355'), ('916235064', '300c_1027_gao__port_506078', 0.0025334184, 332, '355'), ('916235064', 'classe_gl_1027_gao__port_506340', 0.004489152, 332, '355'), ('916235064', 'vivaro_1027_gao__port_506310', 0.0034252217, 332, '355'), ('916235064', 'sl_1027_gao__port_506100', 0.0031354658, 332, '355'), ('916235064', 'elise_1027_gao__port_506121', 0.0010255374, 332, '355'), ('916235064', '1007_1027_gao__port_506070', 0.0015355238, 332, '355'), ('916235064', 'i40_1027_gao__port_506218', 0.0005915156, 332, '355'), ('916235064', 'bipper_tepee_1027_gao__port_506227', 0.0040294323, 332, '355'), ('916235064', 'focus_1027_gao__port_506272', 0.0011586437, 332, '355'), ('916235064', 'primera_1027_gao__port_506147', 0.0012158152, 332, '355'), ('916235064', 'r4_1027_gao__port_506160', 0.014965757, 332, '355'), ('916235064', 'a8_1027_gao__port_506265', 0.0011321428, 332, '355'), ('916235064', 'boxer_1027_gao__port_506202', 0.010545174, 332, '355'), ('916235064', 's5_1027_gao__port_506222', 0.0011985006, 332, '355'), ('916235064', 'r21_1027_gao__port_506093', 0.0041857874, 332, '355'), ('916235064', 'c3_1027_gao__port_506257', 0.0023636455, 332, '355'), ('916235064', 'santa_fe_1027_gao__port_506208', 0.0016323163, 332, '355'), ('916235064', 'm4_1027_gao__port_506344', 0.0015568695, 332, '355'), ('916235064', 'safrane_1027_gao__port_506077', 0.0013958998, 332, '355'), ('916235064', 'classe_gle_1027_gao__port_506395', 0.002197972, 332, '355'), ('916235064', '0_1027_gao__port_506094', 0.008828087, 332, '355'), ('916235064', 'ix35_1027_gao__port_506219', 0.0014614734, 332, '355'), ('916235064', 'carens_1027_gao__port_506298', 0.0008825415, 332, '355'), ('916235064', 'classe_a_1027_gao__port_506339', 0.0024715206, 332, '355'), ('916235064', 'ix20_1027_gao__port_506343', 0.0010093137, 332, '355'), ('916235064', 'note_1027_gao__port_506365', 0.0015963409, 332, '355'), ('916235064', 'a5_1027_gao__port_506200', 0.0015330354, 332, '355'), ('916235064', 'sx4_1027_gao__port_506348', 0.0014917004, 332, '355'), ('916235064', 'sandero_1027_gao__port_506198', 0.0014586165, 332, '355'), ('916235064', '3008_1027_gao__port_506385', 0.0056458567, 332, '355'), ('916235064', 'q50_1027_gao__port_506239', 0.0011165835, 332, '355'), ('916235064', 'latitude_1027_gao__port_506236', 0.00080198894, 332, '355'), ('916235064', 'v40_1027_gao__port_506391', 0.0017145325, 332, '355'), ('916235064', 'xsara_1027_gao__port_506087', 0.0009823082, 332, '355'), ('916235064', 'grand_c_max_1027_gao__port_506342', 0.0017958744, 332, '355'), ('916235064', 'swift_1027_gao__port_506149', 0.0015019632, 332, '355'), ('916235064', 'serie_1_1027_gao__port_506184', 0.0015140551, 332, '355'), ('916235064', 'xc70_1027_gao__port_506393', 0.0036190376, 332, '355'), ('916235064', 'master_1027_gao__port_506203', 0.007957619, 332, '355'), ('916235064', 'clio_1027_gao__port_506280', 0.0029575732, 332, '355'), ('916235064', 'duster_1027_gao__port_506216', 0.0007443716, 332, '355'), ('916235064', 'traveller_1027_gao__port_506403', 0.004293686, 332, '355'), ('916235064', 'tipo_1027_gao__port_506355', 0.0010929607, 332, '355'), ('916235064', 'rav_4_1027_gao__port_506332', 0.001360382, 332, '355'), ('916235064', 'coccinelle_1027_gao__port_506259', 0.0034947079, 332, '355'), ('916235064', 'spacetourer_1027_gao__port_506401', 0.0030971877, 332, '355'), ('916235064', 'xe_1027_gao__port_506357', 0.0014470419, 332, '355'), ('916235064', 'ds3_1027_gao__port_506324', 0.0013093858, 332, '355'), ('916235064', 'mx_5_1027_gao__port_506098', 0.0025887066, 332, '355'), ('916235064', 'land_cruiser_1027_gao__port_506315', 0.009531981, 332, '355'), ('916235064', 'classe_b_1027_gao__port_506335', 0.0017215989, 332, '355'), ('916235064', '806_1027_gao__port_506088', 0.0025619462, 332, '355'), ('916235064', 'rx_8_1027_gao__port_506046', 0.0036217184, 332, '355'), ('916235064', 'spark_1027_gao__port_506185', 0.0010076985, 332, '355'), ('916235064', '6_1027_gao__port_506171', 0.0011183012, 332, '355'), ('916235064', 'bravo_1027_gao__port_506080', 0.0014649678, 332, '355'), ('916235064', 'nx_1027_gao__port_506345', 0.0013682715, 332, '355'), ('916235064', 'sharan_1027_gao__port_506347', 0.0050927107, 332, '355'), ('916235064', 'x_type_1027_gao__port_506067', 0.0007802908, 332, '355'), ('916235064', 'jimny_1027_gao__port_506233', 0.0046058036, 332, '355'), ('916235064', 'wrangler_1027_gao__port_506225', 0.0017997589, 332, '355'), ('916235064', 'c_crosser_1027_gao__port_506312', 0.0015926886, 332, '355'), ('916235064', 'v70_1027_gao__port_506278', 0.0019676902, 332, '355'), ('916235064', 'classe_e_1027_gao__port_506300', 0.0017369951, 332, '355'), ('916235064', 'classe_v_1027_gao__port_506258', 0.012731326, 332, '355'), ('916235064', 'm3_1027_gao__port_506182', 0.0023372243, 332, '355'), ('916235064', 'abarth_500_1027_gao__port_506226', 0.0040435228, 332, '355'), ('916235064', 'serie_6_1027_gao__port_506262', 0.0011314946, 332, '355'), ('916235064', 'modus_1027_gao__port_506146', 0.0018293789, 332, '355'), ('916235064', '3_1027_gao__port_506113', 0.0015082262, 332, '355'), ('916235064', '405_1027_gao__port_506108', 0.008055167, 332, '355'), ('916235064', 'allroad_1027_gao__port_506297', 0.0010597749, 332, '355'), ('916235064', 'auris_1027_gao__port_506322', 0.0011525698, 332, '355'), ('916235064', 'galaxy_1027_gao__port_506143', 0.0025150008, 332, '355'), ('916235064', 'giulietta_1027_gao__port_506363', 0.0008651339, 332, '355'), ('916235064', '106_1027_gao__port_506073', 0.008269972, 332, '355'), ('916235064', 'classe_m_1027_gao__port_506154', 0.003002006, 332, '355'), ('916235064', 'espace_1027_gao__port_506313', 0.0010646414, 332, '355'), ('916235064', 'panda_1027_gao__port_506189', 0.009029825, 332, '355'), ('916235064', 'rcz_1027_gao__port_506197', 0.0011293615, 332, '355'), ('916235064', '4007_1027_gao__port_506162', 0.00067925703, 332, '355'), ('916235064', 'classe_cl_1027_gao__port_506249', 0.0010860062, 332, '355'), ('916235064', 'leaf_1027_gao__port_506139', 0.0018037985, 332, '355'), ('916235064', 'octavia_1027_gao__port_506237', 0.0018602135, 332, '355'), ('916235064', 'ds4_1027_gao__port_506336', 0.0024159732, 332, '355'), ('916235064', 'freelander_1027_gao__port_506084', 0.0023474798, 332, '355'), ('916235064', 'evasion_1027_gao__port_506109', 0.0031143944, 332, '355'), ('916235064', 'punto_1027_gao__port_506106', 0.0019495414, 332, '355'), ('916235064', '2cv_1027_gao__port_506045', 0.007974106, 332, '355'), ('916235064', 'x4_1027_gao__port_506392', 0.0017950044, 332, '355'), ('916235064', 'antara_1027_gao__port_506247', 0.0012469033, 332, '355'), ('916235064', 'murano_1027_gao__port_506316', 0.0006088874, 332, '355'), ('916235064', 'alto_1027_gao__port_506201', 0.009230792, 332, '355'), ('916235064', 'meriva_1027_gao__port_506353', 0.0013765352, 332, '355'), ('916235064', 'orlando_1027_gao__port_506305', 0.0018461632, 332, '355'), ('916235064', 'new_beetle_1027_gao__port_506050', 0.0011637663, 332, '355'), ('916235064', '306_1027_gao__port_506145', 0.0035069708, 332, '355'), ('916235064', 'tiguan_1027_gao__port_506362', 0.0026825026, 332, '355'), ('916235064', 's_type_1027_gao__port_506101', 0.0011381682, 332, '355'), ('916235064', 'c1_1027_gao__port_506128', 0.0027514528, 332, '355'), ('916235064', 'vectra_1027_gao__port_506044', 0.001199021, 332, '355'), ('916235064', 'outlander_1027_gao__port_506317', 0.0017121343, 332, '355'), ('916235064', '307_1027_gao__port_506074', 0.0020010846, 332, '355'), ('916235064', 'a6_s6_1027_gao__port_506134', 0.001657005, 332, '355'), ('916235064', 'nemo_combi_1027_gao__port_506196', 0.0022662862, 332, '355'), ('916235064', 'berlingo_1027_gao__port_506194', 0.0046627144, 332, '355'), ('916235064', 'partner_1027_gao__port_506285', 0.003941104, 332, '355'), ('916235064', 'cayenne_1027_gao__port_506177', 0.0037977814, 332, '355'), ('916235064', 'quattroporte_1027_gao__port_506240', 0.0024441155, 332, '355'), ('916235064', 'c_max_1027_gao__port_506282', 0.0013124109, 332, '355'), ('916235064', 'fabia_1027_gao__port_506396', 0.0052995076, 332, '355'), ('916235064', 'cx_3_1027_gao__port_506281', 0.0014462798, 332, '355'), ('916235064', 'x_trail_1027_gao__port_506264', 0.0018314921, 332, '355'), ('916235064', 'scirocco_1027_gao__port_506276', 0.004790705, 332, '355'), ('916235064', 'matiz_1027_gao__port_506144', 0.0017558916, 332, '355'), ('916235064', 'tigra_1027_gao__port_506069', 0.000854213, 332, '355'), ('916235064', 'escort_1027_gao__port_506091', 0.0048400545, 332, '355'), ('916235064', 'c2_1027_gao__port_506081', 0.0014904654, 332, '355'), ('916235064', 'mini_1027_gao__port_506168', 0.0011922178, 332, '355'), ('916235064', 'i30_1027_gao__port_506291', 0.00063256203, 332, '355'), ('916235064', 'picanto_1027_gao__port_506238', 0.0029747852, 332, '355'), ('916235064', 'mito_1027_gao__port_506072', 0.0015077011, 332, '355'), ('916235064', 'impreza_1027_gao__port_506085', 0.0020192128, 332, '355'), ('916235064', 'kangoo_1027_gao__port_506235', 0.006583951, 332, '355'), ('916235064', 'a4_1027_gao__port_506193', 0.0019874813, 332, '355'), ('916235064', 'cayman_1027_gao__port_506268', 0.0018141003, 332, '355'), ('916235064', 'sportage_1027_gao__port_506148', 0.0014275493, 332, '355'), ('916235064', 'up_1027_gao__port_506356', 0.0068631307, 332, '355'), ('916235064', 'optima_1027_gao__port_506386', 0.0008917972, 332, '355'), ('916235064', 'defender_1027_gao__port_506229', 0.0067220475, 332, '355'), ('916235064', 'serie_2_1027_gao__port_506256', 0.0022275767, 332, '355'), ('916235064', 'edge_1027_gao__port_506187', 0.00087482895, 332, '355'), ('916235064', 'r19_1027_gao__port_506110', 0.0049421852, 332, '355'), ('916235064', 'jetta_1027_gao__port_506304', 0.0036194725, 332, '355'), ('916235064', 'eos_1027_gao__port_506115', 0.0038921875, 332, '355'), ('916235064', 'accord_1027_gao__port_506214', 0.0020126614, 332, '355'), ('916235064', 'yaris_1027_gao__port_506334', 0.0032336449, 332, '355'), ('916235064', 'classe_cls_1027_gao__port_506289', 0.0007852041, 332, '355'), ('916235064', 'polo_1027_gao__port_506361', 0.004310732, 332, '355'), ('916235064', 'serie_4_1027_gao__port_506366', 0.0011476774, 332, '355'), ('916235064', 'mini_cabriolet_1027_gao__port_506204', 0.0008377949, 332, '355'), ('916235064', 'prius_1027_gao__port_506190', 0.0011490105, 332, '355'), ('916235064', 'lodgy_1027_gao__port_506188', 0.0020173641, 332, '355'), ('916235064', 'serie_7_1027_gao__port_506307', 0.0012477537, 332, '355'), ('916235064', 'c15_1027_gao__port_506055', 0.01771186, 332, '355'), ('916235064', 'kadjar_1027_gao__port_506389', 0.0012505174, 332, '355'), ('916235064', 'insignia_1027_gao__port_506364', 0.0016434571, 332, '355'), ('916235064', '308_1027_gao__port_506279', 0.0021236916, 332, '355'), ('916235064', 'roomster_1027_gao__port_506241', 0.0018013252, 332, '355'), ('916235064', '80_1027_gao__port_506057', 0.0046134964, 332, '355'), ('916235064', '309_1027_gao__port_506063', 0.013524377, 332, '355'), ('916235064', 'tucson_1027_gao__port_506320', 0.0021237175, 332, '355'), ('916235064', 'x3_1027_gao__port_506212', 0.0008977029, 332, '355'), ('916235064', 'xf_1027_gao__port_506263', 0.0011165164, 332, '355'), ('916235064', '2008_1027_gao__port_506394', 0.0026409712, 332, '355'), ('916235064', 'passat_1027_gao__port_506306', 0.0014976205, 332, '355'), ('916235064', 'compass_1027_gao__port_506260', 0.0032560462, 332, '355'), ('916235064', 'twingo_1027_gao__port_506309', 0.00649872, 332, '355'), ('916235064', 'micra_1027_gao__port_506221', 0.0035858422, 332, '355'), ('916235064', 'golf_1027_gao__port_506155', 0.003193206, 332, '355'), ('916235064', 'soul_1027_gao__port_506176', 0.0012889692, 332, '355'), ('916235064', 'rapid_1027_gao__port_506398', 0.0025915848, 332, '355'), ('916235064', 'forester_1027_gao__port_506360', 0.0022763938, 332, '355'), ('916235064', 'slk_1027_gao__port_506210', 0.0015845684, 332, '355'), ('916235064', 'forfour_1027_gao__port_506341', 0.002179691, 332, '355'), ('916235064', 'serie_5_1027_gao__port_506209', 0.0013695395, 332, '355'), ('916235064', 'xj_1027_gao__port_506170', 0.002600796, 332, '355'), ('916235064', 'pajero_1027_gao__port_506097', 0.005211117, 332, '355'), ('916235064', 'agila_1027_gao__port_506119', 0.00484948, 332, '355'), ('916235064', 'a6_1027_gao__port_506163', 0.0018911689, 332, '355'), ('916235064', 'fox_1027_gao__port_506092', 0.0008465549, 332, '355'), ('916235064', 'boxster_1027_gao__port_506267', 0.0015941188, 332, '355'), ('916235064', 'altea_1027_gao__port_506246', 0.0021391409, 332, '355'), ('916235064', 'samurai_1027_gao__port_506047', 0.006252161, 332, '355'), ('916235064', 'trax_1027_gao__port_506296', 0.0019439291, 332, '355'), ('916235064', 'getz_1027_gao__port_506058', 0.001638486, 332, '355'), ('916235064', 'cherokee_1027_gao__port_506269', 0.00298086, 332, '355'), ('916235064', 'koleos_1027_gao__port_506378', 0.0015591934, 332, '355'), ('916235064', 'z_series_1027_gao__port_506123', 0.0016563105, 332, '355'), ('916235064', 'ecosport_1027_gao__port_506271', 0.0013229847, 332, '355'), ('916235064', 'space_star_1027_gao__port_506277', 0.0021143614, 332, '355'), ('916235064', 'rs3_sportback_1027_gao__port_506207', 0.0019117146, 332, '355'), ('916235064', 'civic_1027_gao__port_506141', 0.0026899758, 332, '355'), ('916235064', 'talisman_1027_gao__port_506390', 0.00076137367, 332, '355'), ('916235064', 'f_pace_1027_gao__port_506314', 0.0016165014, 332, '355'), ('916235064', 'classe_c_1027_gao__port_506299', 0.0017942552, 332, '355'), ('916235064', 'tt_1027_gao__port_506075', 0.0013935153, 332, '355'), ('916235064', 'pathfinder_1027_gao__port_506183', 0.0016514035, 332, '355'), ('916235064', '156_1027_gao__port_506157', 0.001544394, 332, '355'), ('916235064', 'cx_5_1027_gao__port_506228', 0.0014413374, 332, '355'), ('916235064', 'scenic_1027_gao__port_506255', 0.0016085693, 332, '355'), ('916235064', 'yeti_1027_gao__port_506358', 0.0020913945, 332, '355'), ('916235064', 'mustang_1027_gao__port_506053', 0.010049889, 332, '355'), ('916235064', 'stilo_1027_gao__port_506060', 0.0010832807, 332, '355'), ('916235064', 'ateca_1027_gao__port_506382', 0.001701089, 332, '355'), ('916235064', 'fiorino_1027_gao__port_506217', 0.009198187, 332, '355'), ('916235064', 'classe_glk_1027_gao__port_506290', 0.0017017497, 332, '355'), ('916235064', 'fortwo_1027_gao__port_506230', 0.0016011475, 332, '355'), ('916235064', 'cruze_1027_gao__port_506186', 0.0010052798, 332, '355'), ('916235064', '107_1027_gao__port_506213', 0.0016275812, 332, '355'), ('916235064', 'aygo_1027_gao__port_506248', 0.0032433267, 332, '355'), ('916235064', 'rx_1027_gao__port_506354', 0.0010633005, 332, '355'), ('916235064', '500_1027_gao__port_506245', 0.0016355028, 332, '355'), ('916235064', 'bora_1027_gao__port_506104', 0.0038165518, 332, '355'), ('916235064', 'transit_1027_gao__port_506111', 0.0048620095, 332, '355'), ('916235064', 'pt_cruiser_1027_gao__port_506054', 0.0019164442, 332, '355'), ('916235064', 'patrol_1027_gao__port_506068', 0.004240165, 332, '355'), ('916235064', 'r8_1027_gao__port_506156', 0.0012722577, 332, '355'), ('916235064', 'xm_1027_gao__port_506102', 0.0022679532, 332, '355'), ('916235064', 's60_1027_gao__port_506191', 0.0031991128, 332, '355'), ('916235064', 'aveo_1027_gao__port_506158', 0.0038398106, 332, '355'), ('916235064', 'captiva_1027_gao__port_506159', 0.0017188059, 332, '355'), ('916235064', 'ax_1027_gao__port_506153', 0.0068961456, 332, '355'), ('916235064', 'rexton_1027_gao__port_506107', 0.0013020452, 332, '355'), ('916235064', 'camaro_1027_gao__port_506056', 0.0024900383, 332, '355'), ('916235064', 'ypsilon_1027_gao__port_506131', 0.0019542584, 332, '355'), ('916235064', 'delta_1027_gao__port_506165', 0.001400259, 332, '355'), ('916235064', 'c4_1027_gao__port_506370', 0.0013009398, 332, '355'), ('916235064', 'zx_1027_gao__port_506161', 0.004593916, 332, '355'), ('916235064', 'verso_1027_gao__port_506242', 0.00077215425, 332, '355'), ('916235064', 'superb_1027_gao__port_506327', 0.0019942494, 332, '355'), ('916235064', 'r5_1027_gao__port_506253', 0.009545273, 332, '355'), ('916235064', 'caddy_1027_gao__port_506330', 0.013826391, 332, '355'), ('916235064', 'x5_1027_gao__port_506243', 0.0011203957, 332, '355'), ('916235064', 'f_type_1027_gao__port_506231', 0.00082991883, 332, '355'), ('916235064', 'fusion_1027_gao__port_506096', 0.0012669459, 332, '355'), ('916235064', 'dokker_1027_gao__port_506331', 0.005357574, 332, '355'), ('916235064', '205_1027_gao__port_506062', 0.006684758, 332, '355'), ('916235064', 'macan_1027_gao__port_506195', 0.0015572289, 332, '355'), ('916235064', 'tourneo_1027_gao__port_506369', 0.0064013884, 332, '355'), ('916235064', '108_1027_gao__port_506384', 0.0052643116, 332, '355'), ('916235064', '9_3_1027_gao__port_506071', 0.00083742343, 332, '355'), ('916235064', 'mondeo_1027_gao__port_506116', 0.0014396032, 332, '355'), ('916235064', 'cr_v_1027_gao__port_506164', 0.001641316, 332, '355'), ('916235064', 'c30_1027_gao__port_506090', 0.0017484791, 332, '355'), ('916235064', 'pulsar_1027_gao__port_506397', 0.0012019557, 332, '355'), ('916235064', 'ibiza_1027_gao__port_506273', 0.003723403, 332, '355'), ('916235064', 'a1_1027_gao__port_506338', 0.0012347094, 332, '355'), ('916235064', 'matrix_1027_gao__port_506140', 0.0007077519, 332, '355'), ('916235064', 'carnival_1027_gao__port_506136', 0.0022812732, 332, '355'), ('916235064', 'xantia_1027_gao__port_506086', 0.0021960465, 332, '355'), ('916235064', 'terrano_1027_gao__port_506083', 0.0020294248, 332, '355'), ('916235064', 'q3_1027_gao__port_506275', 0.0011264904, 332, '355'), ('916235064', 'hr_v_1027_gao__port_506283', 0.0017805048, 332, '355'), ('916235064', 'expert_1027_gao__port_506142', 0.007369416, 332, '355'), ('916235064', 'multivan_1027_gao__port_506383', 0.0065044495, 332, '355'), ('916235064', 'venga_1027_gao__port_506380', 0.00080044894, 332, '355'), ('916235064', 'scudo_1027_gao__port_506129', 0.0055925758, 332, '355'), ('916235064', 'laguna_1027_gao__port_506368', 0.0007135058, 332, '355'), ('916235064', 'vel_satis_1027_gao__port_506130', 0.002726597, 332, '355'), ('916235064', 'b_max_1027_gao__port_506367', 0.0017247384, 332, '355'), ('916235064', 'ignis_1027_gao__port_506292', 0.004355817, 332, '355'), ('916235064', '159_1027_gao__port_506064', 0.001078337, 332, '355'), ('916235064', 'grande_punto_1027_gao__port_506138', 0.0023633605, 332, '355'), ('916235064', 'logan_1027_gao__port_506167', 0.004397026, 332, '355'), ('916235064', 's_max_1027_gao__port_506223', 0.0012527433, 332, '355'), ('916235064', 'caravelle_1027_gao__port_506351', 0.003029407, 332, '355'), ('916235064', 'adam_1027_gao__port_506079', 0.0010538583, 332, '355'), ('916235064', '406_1027_gao__port_506132', 0.0013573936, 332, '355'), ('916235064', 'q30_1027_gao__port_506293', 0.0009715915, 332, '355'), ('916235064', 'almera_1027_gao__port_506089', 0.0010239567, 332, '355'), ('916235064', 'corsa_1027_gao__port_506095', 0.0025205594, 332, '355'), ('916235064', 'corolla_1027_gao__port_506120', 0.0026821597, 332, '355'), ('916235064', 'xc60_1027_gao__port_506388', 0.001898481, 332, '355'), ('916235064', 'viano_1027_gao__port_506211', 0.0026943116, 332, '355'), ('916235064', 'pro_cee_d_1027_gao__port_506274', 0.0008319828, 332, '355'), ('916235064', 'a3_1027_gao__port_506321', 0.003738277, 332, '355'), ('916235064', 'v50_1027_gao__port_506150', 0.00079195184, 332, '355'), ('916235064', 'voyager_1027_gao__port_506169', 0.0030525425, 332, '355'), ('916235064', 'corvette_1027_gao__port_506049', 0.0037228519, 332, '355'), ('916235064', 'rio_1027_gao__port_506379', 0.001774063, 332, '355'), ('916235064', 'jazz_1027_gao__port_506252', 0.0015305928, 332, '355'), ('916235064', '200_1027_gao__port_506112', 0.004087509, 332, '355'), ('916235064', 'tts_1027_gao__port_506199', 0.0011862651, 332, '355'), ('916235064', 'zafira_1027_gao__port_506287', 0.0026954839, 332, '355'), ('916235064', 'asx_1027_gao__port_506266', 0.001140712, 332, '355'), ('916235064', '607_1027_gao__port_506118', 0.0012529216, 332, '355'), ('916235064', '207_1027_gao__port_506103', 0.001514894, 332, '355'), ('916235064', 'classe_s_1027_gao__port_506301', 0.0031655158, 332, '355'), ('916235064', 'c6_1027_gao__port_506105', 0.0017347689, 332, '355'), ('916235064', 'express_1027_gao__port_506137', 0.01672629, 332, '355'), ('916235064', 'classe_gla_1027_gao__port_506352', 0.0018256686, 332, '355'), ('916235064', 'v60_1027_gao__port_506333', 0.0021458382, 332, '355'), ('916235064', 'ka_1027_gao__port_506180', 0.0014152136, 332, '355'), ('916235064', 'range_rover_1027_gao__port_506254', 0.0020553288, 332, '355'), ('916235064', 'discovery_1027_gao__port_506375', 0.0022964948, 332, '355'), ('916235064', 'classe_r_1027_gao__port_506270', 0.0013943783, 332, '355'), ('916235064', 'transporter_1027_gao__port_506319', 0.011968426, 332, '355'), ('916235064', 'cee_d_1027_gao__port_506288', 0.0010548136, 332, '355'), ('916235064', 'zoe_1027_gao__port_506244', 0.0020714959, 332, '355'), ('916235064', 'i20_1027_gao__port_506284', 0.0017869296, 332, '355'), ('916235064', 'gtv_1027_gao__port_506059', 0.0057224617, 332, '355'), ('916235064', 's4_avant_1027_gao__port_506261', 0.0027665582, 332, '355'), ('916235064', 'x1_1027_gao__port_506372', 0.0017145165, 332, '355'), ('916235064', 'autres_1027_gao__port_506127', 0.0048252293, 332, '355'), ('916235064', '208_1027_gao__port_506359', 0.0018687898, 332, '355'), ('916235064', 'c8_1027_gao__port_506135', 0.0012579792, 332, '355'), ('916235064', 'astra_1027_gao__port_506215', 0.0012625885, 332, '355'), ('916235064', '2_1027_gao__port_506151', 0.0009244923, 332, '355'), ('916235064', 'doblo_1027_gao__port_506251', 0.007465958, 332, '355'), ('916235064', '807_1027_gao__port_506152', 0.0007290425, 332, '355'), ('916235064', '206_1027_gao__port_506126', 0.001038612, 332, '355'), ('916235064', 'a7_1027_gao__port_506373', 0.0006911823, 332, '355'), ('916235064', 'renegade_1027_gao__port_506346', 0.002141784, 332, '355')]]} begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 0 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 5.4836273193359375e-06 save missing photos in datou_result : time spend for datou_step_exec : 7.772297620773315 time spend to save output : 1.7725918292999268 total time spend for step 1 : 9.544889450073242 step2:argmax Fri May 30 03:37:10 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg': 916235064} map_photo_id_path_extension : {916235064: {'path': 'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Argmax ! calculate argmax for thcl : 355 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : True verbose : True photo_id : 916235064 output[photo_id] : [('916235064', 'c15_1027_gao__port_506055', 0.01771186, 332, '355'), 'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg'] begin to insert list_values into photo_hahstag_ids : length of list_valuse in save_photo_hashtag_id_type : 1 insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) first line : ('916235064', '2049863950', '332') ... last line : ('916235064', '2049863950', '332') time used for this insertion : 0.013213872909545898 begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 1 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 0.01705622673034668 len list_finale : 1, len picture : 1 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('2', None, '916235064', 'c15_1027_gao__port_506055', None, None, '2049863950', '0.01771186', None)] time used for this insertion : 0.016518115997314453 saving photo_ids in datou_result photo id not in port begin to insert list_values into mtr_datou_result : length of list_values in save_final : 0 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [] time used for this insertion : 3.5762786865234375e-06 save missing photos in datou_result : time spend for datou_step_exec : 0.00020122528076171875 time spend to save output : 0.04703950881958008 total time spend for step 2 : 0.0472407341003418 caffe_path_current : About to save ! 2 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 2 output : {'916235064': [('916235064', 'c15_1027_gao__port_506055', 0.01771186, 332, '355'), 'temp/1748569020_1112947_916235064_6293d1bb790dc6902450e7c572b7d10b.jpg']} ############################### TEST tfhub2 ################################ TEST TFHUB2 ######################## test with use_multi_inputs=0 ######################## Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=4567 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=4567 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 4567 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=4567 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 12835 tfhub_classification2 is not linked in the step_by_step architecture ! WARNING : step 12836 argmax is not linked in the step_by_step architecture ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : tfhub_classification2, argmax list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (1171252784,1171252764,1171252487) Found this number of photos: 3 ##### Call download_photos : nb_thread : 5 begin to download photo : 1171252487 begin to download photo : 1171252764 begin to download photo : 1171252784 download finish for photo 1171252487 download finish for photo 1171252764 download finish for photo 1171252784 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 3 ; length of list_pids : 3 ; length of list_args : 3 ##### After load_data_input time to download the photos : 0.1566324234008789 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 2 step1:tfhub_classification2 Fri May 30 03:37:10 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg': 1171252487, 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg': 1171252764, 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg': 1171252784} map_photo_id_path_extension : {1171252487: {'path': 'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg', 'extension': 'jpg'}, 1171252764: {'path': 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg', 'extension': 'jpg'}, 1171252784: {'path': 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step TFHub with tf2 ! multi_thcl or not :False multi_thcl_cond or not :False dic_thcl : {'3609': 1} we are using the classfication for only one thcl 3609 begin to check gpu status inside check gpu memory 2025-05-30 03:37:13.504405: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcuda.so.1 2025-05-30 03:37:13.505203: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:37:13.505277: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:37:13.505323: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:37:13.507386: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:37:13.507469: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:37:13.510520: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:37:13.511633: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:37:13.517718: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:37:13.519393: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:37:13.519774: I tensorflow/core/platform/cpu_feature_guard.cc:143] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 FMA 2025-05-30 03:37:13.551190: I tensorflow/core/platform/profile_utils/cpu_utils.cc:102] CPU Frequency: 3493065000 Hz 2025-05-30 03:37:13.552892: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x7fba54000b60 initialized for platform Host (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:37:13.552922: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version 2025-05-30 03:37:13.556937: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x22fe39b0 initialized for platform CUDA (this does not guarantee that XLA will be used). Devices: 2025-05-30 03:37:13.556968: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): NVIDIA GeForce RTX 2080 Ti, Compute Capability 7.5 2025-05-30 03:37:13.558165: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1561] Found device 0 with properties: pciBusID: 0000:41:00.0 name: NVIDIA GeForce RTX 2080 Ti computeCapability: 7.5 coreClock: 1.545GHz coreCount: 68 deviceMemorySize: 10.76GiB deviceMemoryBandwidth: 573.69GiB/s 2025-05-30 03:37:13.558274: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:37:13.558305: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcublas.so.10 2025-05-30 03:37:13.558397: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcufft.so.10 2025-05-30 03:37:13.558436: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcurand.so.10 2025-05-30 03:37:13.558482: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusolver.so.10 2025-05-30 03:37:13.558533: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcusparse.so.10 2025-05-30 03:37:13.558585: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 2025-05-30 03:37:13.560555: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1703] Adding visible gpu devices: 0 2025-05-30 03:37:13.560628: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.1 2025-05-30 03:37:13.560684: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1102] Device interconnect StreamExecutor with strength 1 edge matrix: 2025-05-30 03:37:13.560700: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1108] 0 2025-05-30 03:37:13.560713: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1121] 0: N 2025-05-30 03:37:13.562722: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1247] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 3096 MB memory) -> physical GPU (device: 0, name: NVIDIA GeForce RTX 2080 Ti, pci bus id: 0000:41:00.0, compute capability: 7.5) l 3637 free memory gpu now : 9500 max_wait_temp : 1 max_wait : 5 1 Physical GPUs, 1 Logical GPUs tagging for thcl : 3609 To do loadFromThcl(), then load ParamDescType : thcl3609 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (3609) thcls : [{'id': 3609, 'mtr_user_id': 31, 'name': 'tfhub_19_06_2023', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'jrm,pcm,pcnc,pehd,tapis_vide', 'svm_portfolios_learning': '9336903,9336904,9336905,9336906,9336909', 'photo_hashtag_type': 4674, 'photo_desc_type': 5832, 'type_classification': 'tf_classification2', 'hashtag_id_list': '495916461,560181804,1284539308,628944319,2107748999'}] thcl {'id': 3609, 'mtr_user_id': 31, 'name': 'tfhub_19_06_2023', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'jrm,pcm,pcnc,pehd,tapis_vide', 'svm_portfolios_learning': '9336903,9336904,9336905,9336906,9336909', 'photo_hashtag_type': 4674, 'photo_desc_type': 5832, 'type_classification': 'tf_classification2', 'hashtag_id_list': '495916461,560181804,1284539308,628944319,2107748999'} Update svm_hashtag_type_desc : 5832 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (5832) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (5832, 'tfhub_19_06_2023', 1280, 1280, 'tfhub_19_06_2023', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 3, datetime.datetime(2023, 6, 19, 12, 55, 22), datetime.datetime(2023, 6, 19, 12, 55, 22)) model_name : tfhub_19_06_2023 model_param file didn't exist model_name : tfhub_19_06_2023 model_type : tf_classification2 list file need : ['Confusion_Matrix.png', 'Precision_Recall_jrm.jpg', 'Precision_Recall_pcm.jpg', 'Precision_Recall_pcnc.jpg', 'Precision_Recall_pehd.jpg', 'Precision_Recall_tapis_vide.jpg', 'Result_Summary.txt', 'checkpoint', 'model_checkpoint.ckpt.data-00000-of-00002', 'model_checkpoint.ckpt.data-00001-of-00002', 'model_checkpoint.ckpt.index', 'model_weights.h5'] file exist in s3 : ['Confusion_Matrix.png', 'Precision_Recall_jrm.jpg', 'Precision_Recall_pcm.jpg', 'Precision_Recall_pcnc.jpg', 'Precision_Recall_pehd.jpg', 'Precision_Recall_tapis_vide.jpg', 'Result_Summary.txt', 'checkpoint', 'model_checkpoint.ckpt.data-00000-of-00002', 'model_checkpoint.ckpt.data-00001-of-00002', 'model_checkpoint.ckpt.index', 'model_weights.h5'] file manque in s3 : [] /home/admin/workarea/install/caffe_frcnn_python3/py-faster-rcnn/caffe-fast-rcnn/python/../../tools/../lib/rpn/proposal_layer.py:28: YAMLLoadWarning: calling yaml.load() without Loader=... is deprecated, as the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full details. layer_params = yaml.load(self.param_str_) local folder : /data/models_weight/tfhub_19_06_2023 /data/models_weight/tfhub_19_06_2023/Confusion_Matrix.png size_local : 57753 size in s3 : 57753 create time local : 2023-06-22 17:09:38 create time in s3 : 2023-06-19 10:55:15 Confusion_Matrix.png already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Precision_Recall_jrm.jpg size_local : 79724 size in s3 : 79724 create time local : 2023-06-22 17:09:38 create time in s3 : 2023-06-19 10:55:20 Precision_Recall_jrm.jpg already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Precision_Recall_pcm.jpg size_local : 83556 size in s3 : 83556 create time local : 2023-06-22 17:09:38 create time in s3 : 2023-06-19 10:55:15 Precision_Recall_pcm.jpg already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Precision_Recall_pcnc.jpg size_local : 74107 size in s3 : 74107 create time local : 2023-06-22 17:09:38 create time in s3 : 2023-06-19 10:55:20 Precision_Recall_pcnc.jpg already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Precision_Recall_pehd.jpg size_local : 72705 size in s3 : 72705 create time local : 2023-06-22 17:09:39 create time in s3 : 2023-06-19 10:55:20 Precision_Recall_pehd.jpg already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Precision_Recall_tapis_vide.jpg size_local : 70874 size in s3 : 70874 create time local : 2023-06-22 17:09:39 create time in s3 : 2023-06-19 10:55:15 Precision_Recall_tapis_vide.jpg already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/Result_Summary.txt size_local : 642 size in s3 : 642 create time local : 2023-06-22 17:09:39 create time in s3 : 2023-06-19 10:55:22 Result_Summary.txt already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/checkpoint size_local : 99 size in s3 : 99 create time local : 2023-06-22 17:09:39 create time in s3 : 2023-06-19 10:55:22 checkpoint already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/model_checkpoint.ckpt.data-00000-of-00002 size_local : 216488 size in s3 : 216488 create time local : 2023-06-22 17:09:39 create time in s3 : 2023-06-19 10:55:22 model_checkpoint.ckpt.data-00000-of-00002 already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/model_checkpoint.ckpt.data-00001-of-00002 size_local : 32279708 size in s3 : 32279708 create time local : 2023-06-22 17:09:40 create time in s3 : 2023-06-19 10:55:21 model_checkpoint.ckpt.data-00001-of-00002 already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/model_checkpoint.ckpt.index size_local : 43546 size in s3 : 43546 create time local : 2023-06-22 17:09:40 create time in s3 : 2023-06-19 10:55:22 model_checkpoint.ckpt.index already exist and didn't need to update /data/models_weight/tfhub_19_06_2023/model_weights.h5 size_local : 16499144 size in s3 : 16499144 create time local : 2023-06-22 17:09:40 create time in s3 : 2023-06-19 10:55:15 model_weights.h5 already exist and didn't need to update desc size : 1280 Model: "sequential" _________________________________________________________________ Layer (type) Output Shape Param # ================================================================= module (KerasLayer) (None, 1280) 4049564 _________________________________________________________________ tfhub_19_06_2023dense (Dense (None, 5) 6405 ================================================================= Total params: 4,055,969 Trainable params: 6,405 Non-trainable params: 4,049,564 _________________________________________________________________ Loading Weights... time used to create the model : 8.38087272644043 time used to load_weights : 0.16702485084533691 0it [00:00, ?it/s] 3it [00:00, 639.54it/s]2025-05-30 03:37:24.663048: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudnn.so.7 temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg Found 3 images belonging to 1 classes. begin to do the prediction : time used to do the prediction : 3.104058027267456 ['temp/image000000000_1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg', 'temp/image000000001_1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg', 'temp/image000000002_1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg'] (3,) (3, 5) (3, 1280) shape of features : (3, 1280) shape of new features : (1, 3, 1280) save descriptor for thcl : 3609 (3, 1280) Got the blobs of the net to insert : [0, 9, 0, 0, 0, 0, 1, 0, 0, 0] code_as_byte_string:b'0009000000'| Got the blobs of the net to insert : [0, 6, 0, 1, 0, 0, 0, 1, 0, 0] code_as_byte_string:b'0006000100'| Got the blobs of the net to insert : [0, 6, 0, 0, 1, 0, 0, 1, 0, 0] code_as_byte_string:b'0006000001'| time to traite the descriptors : 0.024588584899902344 Testing : ['1171252487', '1171252764', '1171252784'] In select_photos_meta_from_ids: SELECT photo_id, url, FROM_UNIXTIME(uploaded_at), latitude, longitude, text FROM MTRBack.photos WHERE photo_id IN (1171252487,1171252764,1171252784) result : {1171252487: {'photo_id': 1171252487, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/22/5ebdd6b0a6bb39942a3808ed114806de.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_21_55_35_005998m0.jpg 0.4259977941513062 for time 6.000020980834961, id_amount 3 this amount prod time diff : 0.006000020980834961'}, 1171252764: {'photo_id': 1171252764, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/22/29d5179a892cc50aadc9d67245534b59.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_21_55_41_005998m0.jpg 0.4319977941513062 for time 6.0, id_amount 3 this amount prod time diff : 0.006'}, 1171252784: {'photo_id': 1171252784, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/22/5a3c5d3bb155a7a116f67ded51bffb59.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_21_55_47_006033m0.jpg 0.4379978291988373 for time 6.000035047531128, id_amount 4 this amount prod time diff : 0.006000035047531128'}} list_photo_exists : [1171252487, 1171252764, 1171252784] storage_type for insertDescriptorsMulti : 3 To insert : 1171252487 To insert : 1171252764 To insert : 1171252784 time to insert the descriptors : 0.8584628105163574 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : False verbose : True saveOutput not yet implemented for datou_step.type : tfhub_classification2 we use saveGeneral [1171252487, 1171252764, 1171252784] map_info['map_portfolio_photo'] : {} final : False mtd_id 4567 list_pids : [1171252487, 1171252764, 1171252784] Looping around the photos to save general results len do output : 3 /1171252487Didn't retrieve data . /1171252764Didn't retrieve data . /1171252784Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('4567', None, None, None, None, None, None, None, None) ('4567', None, '1171252487', None, None, None, None, None, None) ('4567', None, None, None, None, None, None, None, None) ('4567', None, '1171252764', None, None, None, None, None, None) ('4567', None, None, None, None, None, None, None, None) ('4567', None, '1171252784', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 6 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4567', None, '1171252487', 'None', None, None, None, None, None), ('4567', None, '1171252764', 'None', None, None, None, None, None), ('4567', None, '1171252784', 'None', None, None, None, None, None)] time used for this insertion : 0.0171658992767334 save_final save missing photos in datou_result : time spend for datou_step_exec : 17.8849880695343 time spend to save output : 0.017550230026245117 total time spend for step 1 : 17.902538299560547 step2:argmax Fri May 30 03:37:28 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg': 1171252487, 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg': 1171252764, 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg': 1171252784} map_photo_id_path_extension : {1171252487: {'path': 'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg', 'extension': 'jpg'}, 1171252764: {'path': 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg', 'extension': 'jpg'}, 1171252784: {'path': 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Argmax ! calculate argmax for thcl : 3609 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : True verbose : True photo_id : 1171252487 output[photo_id] : [(1171252487, 'jrm', 0.9262437, 4674, '3609'), 'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg'] photo_id : 1171252764 output[photo_id] : [(1171252764, 'jrm', 0.98535734, 4674, '3609'), 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg'] photo_id : 1171252784 output[photo_id] : [(1171252784, 'jrm', 0.96776813, 4674, '3609'), 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg'] begin to insert list_values into photo_hahstag_ids : length of list_valuse in save_photo_hashtag_id_type : 3 insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) first line : ('1171252487', '495916461', '4674') ... last line : ('1171252784', '495916461', '4674') time used for this insertion : 0.026948928833007812 begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 3 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 0.019032716751098633 len list_finale : 3, len picture : 3 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 3 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4567', None, '1171252487', 'jrm', None, None, '495916461', '0.9262437', None), ('4567', None, '1171252764', 'jrm', None, None, '495916461', '0.98535734', None), ('4567', None, '1171252784', 'jrm', None, None, '495916461', '0.96776813', None)] time used for this insertion : 0.013668060302734375 saving photo_ids in datou_result photo id not in port photo id not in port photo id not in port begin to insert list_values into mtr_datou_result : length of list_values in save_final : 0 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [] time used for this insertion : 3.5762786865234375e-06 save missing photos in datou_result : time spend for datou_step_exec : 0.0001800060272216797 time spend to save output : 0.06439590454101562 total time spend for step 2 : 0.0645759105682373 caffe_path_current : About to save ! 2 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 2 output : {'1171252487': [(1171252487, 'jrm', 0.9262437, 4674, '3609'), 'temp/1748569030_1112947_1171252487_5ebdd6b0a6bb39942a3808ed114806de.jpg'], '1171252764': [(1171252764, 'jrm', 0.98535734, 4674, '3609'), 'temp/1748569030_1112947_1171252764_29d5179a892cc50aadc9d67245534b59.jpg'], '1171252784': [(1171252784, 'jrm', 0.96776813, 4674, '3609'), 'temp/1748569030_1112947_1171252784_5a3c5d3bb155a7a116f67ded51bffb59.jpg']} --------------------- test with use_multi_inputs=0 is succeded ------------------- ######################## test with use_multi_inputs=1 ######################## Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=4621 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=4621 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 4621 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=4621 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 12927 tfhub_classification2 is not linked in the step_by_step architecture ! WARNING : step 12928 argmax is not linked in the step_by_step architecture ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : tfhub_classification2, argmax list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (1171291875,1171275372,1171275314) Found this number of photos: 3 ##### Call download_photos : nb_thread : 5 begin to download photo : 1171275314 begin to download photo : 1171275372 begin to download photo : 1171291875 download finish for photo 1171275314 download finish for photo 1171291875 download finish for photo 1171275372 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 3 ; length of list_pids : 3 ; length of list_args : 3 ##### After load_data_input time to download the photos : 0.24903392791748047 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 2 step1:tfhub_classification2 Fri May 30 03:37:28 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg': 1171275314, 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg': 1171291875, 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg': 1171275372} map_photo_id_path_extension : {1171275314: {'path': 'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg', 'extension': 'jpg'}, 1171291875: {'path': 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg', 'extension': 'jpg'}, 1171275372: {'path': 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step TFHub with tf2 ! multi_thcl or not :False multi_thcl_cond or not :False dic_thcl : {'3655': 1} we are using the classfication for only one thcl 3655 begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 5948 max_wait_temp : 1 max_wait : 5 1 Physical GPUs, 1 Logical GPUs tagging for thcl : 3655 To do loadFromThcl(), then load ParamDescType : thcl3655 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (3655) thcls : [{'id': 3655, 'mtr_user_id': 31, 'name': 'tfhub_18_7_2023', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'pcm,pcnc,jrm,pehd,tapis_vide', 'svm_portfolios_learning': '9336904,9336905,9336903,9336906,9336909', 'photo_hashtag_type': 4723, 'photo_desc_type': 5862, 'type_classification': 'tf_classification2', 'hashtag_id_list': '560181804,1284539308,495916461,628944319,2107748999'}] thcl {'id': 3655, 'mtr_user_id': 31, 'name': 'tfhub_18_7_2023', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'pcm,pcnc,jrm,pehd,tapis_vide', 'svm_portfolios_learning': '9336904,9336905,9336903,9336906,9336909', 'photo_hashtag_type': 4723, 'photo_desc_type': 5862, 'type_classification': 'tf_classification2', 'hashtag_id_list': '560181804,1284539308,495916461,628944319,2107748999'} Update svm_hashtag_type_desc : 5862 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (5862) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (5862, 'tfhub_18_7_2023', 1280, 1280, 'tfhub_18_7_2023', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 3, datetime.datetime(2023, 7, 18, 22, 46, 29), datetime.datetime(2023, 7, 18, 22, 46, 29)) model_name : tfhub_18_7_2023 model_param file didn't exist model_name : tfhub_18_7_2023 model_type : tf_classification2 list file need : ['Confusion_Matrix.png', 'Precision_Recall_jrm.jpg', 'Precision_Recall_pcm.jpg', 'Precision_Recall_pcnc.jpg', 'Precision_Recall_pehd.jpg', 'Precision_Recall_tapis_vide.jpg', 'Result_Summary.txt', 'checkpoint', 'model_checkpoint.ckpt.data-00000-of-00002', 'model_checkpoint.ckpt.data-00001-of-00002', 'model_checkpoint.ckpt.index', 'model_weights.h5'] file exist in s3 : ['Confusion_Matrix.png', 'Precision_Recall_jrm.jpg', 'Precision_Recall_pcm.jpg', 'Precision_Recall_pcnc.jpg', 'Precision_Recall_pehd.jpg', 'Precision_Recall_tapis_vide.jpg', 'Result_Summary.txt', 'checkpoint', 'model_checkpoint.ckpt.data-00000-of-00002', 'model_checkpoint.ckpt.data-00001-of-00002', 'model_checkpoint.ckpt.index', 'model_weights.h5'] file manque in s3 : [] local folder : /data/models_weight/tfhub_18_7_2023 /data/models_weight/tfhub_18_7_2023/Confusion_Matrix.png size_local : 54360 size in s3 : 54360 create time local : 2023-08-11 11:22:56 create time in s3 : 2023-07-18 20:46:28 Confusion_Matrix.png already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Precision_Recall_jrm.jpg size_local : 72583 size in s3 : 72583 create time local : 2023-08-11 11:22:56 create time in s3 : 2023-07-18 20:46:23 Precision_Recall_jrm.jpg already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Precision_Recall_pcm.jpg size_local : 81681 size in s3 : 81681 create time local : 2023-08-11 11:22:56 create time in s3 : 2023-07-18 20:46:17 Precision_Recall_pcm.jpg already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Precision_Recall_pcnc.jpg size_local : 79510 size in s3 : 79510 create time local : 2023-08-11 11:22:56 create time in s3 : 2023-07-18 20:46:23 Precision_Recall_pcnc.jpg already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Precision_Recall_pehd.jpg size_local : 59936 size in s3 : 59936 create time local : 2023-08-11 11:22:57 create time in s3 : 2023-07-18 20:46:23 Precision_Recall_pehd.jpg already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Precision_Recall_tapis_vide.jpg size_local : 78974 size in s3 : 78974 create time local : 2023-08-11 11:22:57 create time in s3 : 2023-07-18 20:46:17 Precision_Recall_tapis_vide.jpg already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/Result_Summary.txt size_local : 642 size in s3 : 642 create time local : 2023-08-11 11:22:57 create time in s3 : 2023-07-18 20:46:23 Result_Summary.txt already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/checkpoint size_local : 99 size in s3 : 99 create time local : 2023-08-11 11:22:57 create time in s3 : 2023-07-18 20:46:23 checkpoint already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/model_checkpoint.ckpt.data-00000-of-00002 size_local : 216529 size in s3 : 216529 create time local : 2023-08-11 11:22:57 create time in s3 : 2023-07-18 20:46:17 model_checkpoint.ckpt.data-00000-of-00002 already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/model_checkpoint.ckpt.data-00001-of-00002 size_local : 32279748 size in s3 : 32279748 create time local : 2023-08-11 11:22:58 create time in s3 : 2023-07-18 20:46:19 model_checkpoint.ckpt.data-00001-of-00002 already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/model_checkpoint.ckpt.index size_local : 43546 size in s3 : 43546 create time local : 2023-08-11 11:22:58 create time in s3 : 2023-07-18 20:46:19 model_checkpoint.ckpt.index already exist and didn't need to update /data/models_weight/tfhub_18_7_2023/model_weights.h5 size_local : 16500868 size in s3 : 16500868 create time local : 2023-08-11 11:22:58 create time in s3 : 2023-07-18 20:46:18 model_weights.h5 already exist and didn't need to update desc size : 1280 Model: "model" __________________________________________________________________________________________________ Layer (type) Output Shape Param # Connected to ================================================================================================== input_1 (InputLayer) [(None, 224, 224, 3) 0 __________________________________________________________________________________________________ input_2 (InputLayer) [(None, 1)] 0 __________________________________________________________________________________________________ module (KerasLayer) (None, 1280) 4049564 input_1[0][0] __________________________________________________________________________________________________ concatenate (Concatenate) (None, 1281) 0 input_2[0][0] module[0][0] __________________________________________________________________________________________________ tfhub_18_7_2023dense (Dense) (None, 5) 6410 concatenate[0][0] ================================================================================================== Total params: 4,055,974 Trainable params: 0 Non-trainable params: 4,055,974 __________________________________________________________________________________________________ Loading Weights... time used to create the model : 7.6462242603302 time used to load_weights : 0.13023996353149414 found 3 data found 0 labels begin to do the prediction : time used to do the prediction : 1.0508487224578857 ['temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg', 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg', 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg'] (3,) (3, 5) (3, 1280) shape of features : (3, 1280) shape of new features : (1, 3, 1280) save descriptor for thcl : 3655 (3, 1280) Got the blobs of the net to insert : [0, 0, 0, 0, 8, 0, 0, 0, 3, 0] code_as_byte_string:b'0000000008'| Got the blobs of the net to insert : [0, 1, 0, 0, 11, 0, 2, 2, 0, 0] code_as_byte_string:b'000100000b'| Got the blobs of the net to insert : [0, 0, 0, 0, 14, 0, 1, 4, 0, 0] code_as_byte_string:b'000000000e'| time to traite the descriptors : 0.04223155975341797 Testing : ['1171275314', '1171291875', '1171275372'] In select_photos_meta_from_ids: SELECT photo_id, url, FROM_UNIXTIME(uploaded_at), latitude, longitude, text FROM MTRBack.photos WHERE photo_id IN (1171275314,1171291875,1171275372) result : {1171275314: {'photo_id': 1171275314, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/23/6e0a72c8fa00d5e4b018bd689b547133.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_23_54_22_6187.jpg'}, 1171275372: {'photo_id': 1171275372, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/23/76d81364ff7df843bff095f45c07ba35.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_23_56_46_6098.jpg'}, 1171291875: {'photo_id': 1171291875, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2023/2/23/b62cd9e0d976b143f86fe82d072798c0.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22022023_23_59_04_5803.jpg'}} list_photo_exists : [1171275314, 1171275372, 1171291875] storage_type for insertDescriptorsMulti : 3 To insert : 1171275314 To insert : 1171291875 To insert : 1171275372 time to insert the descriptors : 1.1219301223754883 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : False verbose : True saveOutput not yet implemented for datou_step.type : tfhub_classification2 we use saveGeneral [1171275314, 1171291875, 1171275372] map_info['map_portfolio_photo'] : {} final : False mtd_id 4621 list_pids : [1171275314, 1171291875, 1171275372] Looping around the photos to save general results len do output : 3 /1171275314Didn't retrieve data . /1171291875Didn't retrieve data . /1171275372Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('4621', None, None, None, None, None, None, None, None) ('4621', None, '1171275314', None, None, None, None, None, None) ('4621', None, None, None, None, None, None, None, None) ('4621', None, '1171291875', None, None, None, None, None, None) ('4621', None, None, None, None, None, None, None, None) ('4621', None, '1171275372', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 6 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4621', None, '1171275314', 'None', None, None, None, None, None), ('4621', None, '1171291875', 'None', None, None, None, None, None), ('4621', None, '1171275372', 'None', None, None, None, None, None)] time used for this insertion : 0.014189004898071289 save_final save missing photos in datou_result : time spend for datou_step_exec : 13.24964427947998 time spend to save output : 0.014552831649780273 total time spend for step 1 : 13.26419711112976 step2:argmax Fri May 30 03:37:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg': 1171275314, 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg': 1171291875, 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg': 1171275372} map_photo_id_path_extension : {1171275314: {'path': 'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg', 'extension': 'jpg'}, 1171291875: {'path': 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg', 'extension': 'jpg'}, 1171275372: {'path': 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Argmax ! calculate argmax for thcl : 3655 After datou_step_exec type output : map_portfolio_photo : len 0 keys : dict_keys([]) Inside saveOutput : final : True verbose : True photo_id : 1171275314 output[photo_id] : [(1171275314, 'tapis_vide', 0.9651836, 4723, '3655'), 'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg'] photo_id : 1171291875 output[photo_id] : [(1171291875, 'tapis_vide', 0.9706325, 4723, '3655'), 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg'] photo_id : 1171275372 output[photo_id] : [(1171275372, 'tapis_vide', 0.9674696, 4723, '3655'), 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg'] begin to insert list_values into photo_hahstag_ids : length of list_valuse in save_photo_hashtag_id_type : 3 insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) first line : ('1171275314', '2107748999', '4723') ... last line : ('1171275372', '2107748999', '4723') time used for this insertion : 0.02501225471496582 begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 3 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 0.01639246940612793 len list_finale : 3, len picture : 3 begin to insert list_values into mtr_datou_result : length of list_values in save_final : 3 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('4621', None, '1171275314', 'tapis_vide', None, None, '2107748999', '0.9651836', None), ('4621', None, '1171291875', 'tapis_vide', None, None, '2107748999', '0.9706325', None), ('4621', None, '1171275372', 'tapis_vide', None, None, '2107748999', '0.9674696', None)] time used for this insertion : 0.01503753662109375 saving photo_ids in datou_result photo id not in port photo id not in port photo id not in port begin to insert list_values into mtr_datou_result : length of list_values in save_final : 0 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [] time used for this insertion : 5.245208740234375e-06 save missing photos in datou_result : time spend for datou_step_exec : 0.00017118453979492188 time spend to save output : 0.0609283447265625 total time spend for step 2 : 0.06109952926635742 caffe_path_current : About to save ! 2 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 2 output : {'1171275314': [(1171275314, 'tapis_vide', 0.9651836, 4723, '3655'), 'temp/1748569048_1112947_1171275314_6e0a72c8fa00d5e4b018bd689b547133.jpg'], '1171291875': [(1171291875, 'tapis_vide', 0.9706325, 4723, '3655'), 'temp/1748569048_1112947_1171291875_b62cd9e0d976b143f86fe82d072798c0.jpg'], '1171275372': [(1171275372, 'tapis_vide', 0.9674696, 4723, '3655'), 'temp/1748569048_1112947_1171275372_76d81364ff7df843bff095f45c07ba35.jpg']} --------------------- test with use_multi_inputs=1 is succeded ------------------- ############################### TEST ordonner ################################ To do loadFromThcl(), then load ParamDescType : thcl358 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (358) thcls : [{'id': 358, 'mtr_user_id': 31, 'name': 'car_orientation_0111', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'FirstUploadExperveo_vignette__port_505674,CAR_EXTERIEUR_Roue__port_503398,FirstUploadExperveo_carrosseriegrosplan_VIndanslamoquette__port_506486,FirstUploadExperveo_carrosseriegrosplan_siegegrosplan__port_506485,CAR_EXTERIEUR_Cote_droit_axe_avant__port_504465,CAR_EXTERIEUR_Cote_gauche_axe_arriere__port_504198,CAR_EXTERIEUR_Face_avant_axe_droit__port_504451,CAR_EXTERIEUR_angle_avant_gauche_axe_avant__port_504235,FirstUploadExperveo_vin__port_505675,CAR_EXTERIEUR_cote_droite__port_504108,CAR_INTERIEUR_avant_volant_class_6_levierdevitesse__port_506565,FirstUploadExperveo_carrosseriegrosplan_carrosserie__port_506483,CAR_EXTERIEUR_Angle_arriere_gauche_axe_arriere__port_504201,cartegrise_orientation__port_505064,CAR_EXTERIEUR_Angle_arriere_droit_axe_arriere__port_504217,CAR_INTERIEUR_avant_vue-arriere_class_1__port_506531,CAR_EXTERIEUR_Face_arriere_axe_droit__port_504218,CAR_EXTERIEUR_Cote_droit_axe_arriere__port_504214,CAR_EXTERIEUR_Angle_avant_droit__port_504087,FirstUploadExperveo_carrosseriegrosplan_morceauderoue__port_506484,CAR_INTERIEUR_avant_volant_class_6_class_2__port_506563,CAR_EXTERIEUR_Angle_arriere_droit__port_504160,CAR_EXTERIEUR_arriere__port_504184,CAR_INTERIEUR_avant_volant_class_6_boutonrond__port_506562,INTERIEUR_Compteur_kilometrique__port_503644,CAR_INTERIEUR_avant_vue_gauche_habitacle_class_1__port_506494,CAR_EXTERIEUR_Angle_arriere_gauche__port_504170,CAR_EXTERIEUR_Angle_avant_droit_axe_arriere__port_504226,CAR_EXTERIEUR_Face_arriere_axe_gauche__port_504202,CAR_EXTERIEUR_moteur__port_503704,FirstUploadExperveo_carrosseriegrosplan_class_6__port_506487,CAR_INTERIEUR_siege_arriere_class_1__port_506551,CAR_EXTERIEUR_avant__port_504146,CAR_EXTERIEUR_Angle_arriere_droit_axe_droit__port_504215,CAR_EXTERIEUR_Angle_avant_droit_axe_droit__port_504225,CAR_INTERIEUR_avant_volant_class_6_ecrangrosplan__port_506564,FirstUploadExperveo_carrosseriegrosplan_moteurgrosplanetdegat__port_506482,CAR_INTERIEUR_coffre__port_503412,FirstUploadExperveo_rouetranche__port_505677,UploadPhotoImmatBest_class_1__port_505051,CAR_INTERIEUR_avant_vue-arriere_class_2__port_506532,CAR_EXTERIEUR_angle_avant_gauche__port_504098,CAR_EXTERIEUR_face_avant_axe_gauche__port_504236,CAR_INTERIEUR_avant_vue_droite_habitacle_class_1__port_506540,CAR_EXTERIEUR_cote_gauche_axe_avant__port_504233,CAR_EXTERIEUR_roue_de_secour__port_503763,CAR_EXTERIEUR_Angle_arriere_gauche_axe_gauche__port_504199,CAR_EXTERIEUR_cote_gauche__port_504017,CAR_INTERIEUR_avant_volant_class_1__port_506503,CAR_INTERIEUR_avant_volant_class_2__port_506504,CAR_EXTERIEUR_angle_avant_gauche_axe_gauche__port_504234', 'svm_portfolios_learning': '505674,503398,506486,506485,504465,504198,504451,504235,505675,504108,506565,506483,504201,505064,504217,506531,504218,504214,504087,506484,506563,504160,504184,506562,503644,506494,504170,504226,504202,503704,506487,506551,504146,504215,504225,506564,506482,503412,505677,505051,506532,504098,504236,506540,504233,503763,504199,504017,506503,506504,504234', 'photo_hashtag_type': 337, 'photo_desc_type': 3392, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'}] thcl {'id': 358, 'mtr_user_id': 31, 'name': 'car_orientation_0111', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'FirstUploadExperveo_vignette__port_505674,CAR_EXTERIEUR_Roue__port_503398,FirstUploadExperveo_carrosseriegrosplan_VIndanslamoquette__port_506486,FirstUploadExperveo_carrosseriegrosplan_siegegrosplan__port_506485,CAR_EXTERIEUR_Cote_droit_axe_avant__port_504465,CAR_EXTERIEUR_Cote_gauche_axe_arriere__port_504198,CAR_EXTERIEUR_Face_avant_axe_droit__port_504451,CAR_EXTERIEUR_angle_avant_gauche_axe_avant__port_504235,FirstUploadExperveo_vin__port_505675,CAR_EXTERIEUR_cote_droite__port_504108,CAR_INTERIEUR_avant_volant_class_6_levierdevitesse__port_506565,FirstUploadExperveo_carrosseriegrosplan_carrosserie__port_506483,CAR_EXTERIEUR_Angle_arriere_gauche_axe_arriere__port_504201,cartegrise_orientation__port_505064,CAR_EXTERIEUR_Angle_arriere_droit_axe_arriere__port_504217,CAR_INTERIEUR_avant_vue-arriere_class_1__port_506531,CAR_EXTERIEUR_Face_arriere_axe_droit__port_504218,CAR_EXTERIEUR_Cote_droit_axe_arriere__port_504214,CAR_EXTERIEUR_Angle_avant_droit__port_504087,FirstUploadExperveo_carrosseriegrosplan_morceauderoue__port_506484,CAR_INTERIEUR_avant_volant_class_6_class_2__port_506563,CAR_EXTERIEUR_Angle_arriere_droit__port_504160,CAR_EXTERIEUR_arriere__port_504184,CAR_INTERIEUR_avant_volant_class_6_boutonrond__port_506562,INTERIEUR_Compteur_kilometrique__port_503644,CAR_INTERIEUR_avant_vue_gauche_habitacle_class_1__port_506494,CAR_EXTERIEUR_Angle_arriere_gauche__port_504170,CAR_EXTERIEUR_Angle_avant_droit_axe_arriere__port_504226,CAR_EXTERIEUR_Face_arriere_axe_gauche__port_504202,CAR_EXTERIEUR_moteur__port_503704,FirstUploadExperveo_carrosseriegrosplan_class_6__port_506487,CAR_INTERIEUR_siege_arriere_class_1__port_506551,CAR_EXTERIEUR_avant__port_504146,CAR_EXTERIEUR_Angle_arriere_droit_axe_droit__port_504215,CAR_EXTERIEUR_Angle_avant_droit_axe_droit__port_504225,CAR_INTERIEUR_avant_volant_class_6_ecrangrosplan__port_506564,FirstUploadExperveo_carrosseriegrosplan_moteurgrosplanetdegat__port_506482,CAR_INTERIEUR_coffre__port_503412,FirstUploadExperveo_rouetranche__port_505677,UploadPhotoImmatBest_class_1__port_505051,CAR_INTERIEUR_avant_vue-arriere_class_2__port_506532,CAR_EXTERIEUR_angle_avant_gauche__port_504098,CAR_EXTERIEUR_face_avant_axe_gauche__port_504236,CAR_INTERIEUR_avant_vue_droite_habitacle_class_1__port_506540,CAR_EXTERIEUR_cote_gauche_axe_avant__port_504233,CAR_EXTERIEUR_roue_de_secour__port_503763,CAR_EXTERIEUR_Angle_arriere_gauche_axe_gauche__port_504199,CAR_EXTERIEUR_cote_gauche__port_504017,CAR_INTERIEUR_avant_volant_class_1__port_506503,CAR_INTERIEUR_avant_volant_class_2__port_506504,CAR_EXTERIEUR_angle_avant_gauche_axe_gauche__port_504234', 'svm_portfolios_learning': '505674,503398,506486,506485,504465,504198,504451,504235,505675,504108,506565,506483,504201,505064,504217,506531,504218,504214,504087,506484,506563,504160,504184,506562,503644,506494,504170,504226,504202,503704,506487,506551,504146,504215,504225,506564,506482,503412,505677,505051,506532,504098,504236,506540,504233,503763,504199,504017,506503,506504,504234', 'photo_hashtag_type': 337, 'photo_desc_type': 3392, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'} Update svm_hashtag_type_desc : 3392 ['FirstUploadExperveo_vignette__port_505674', 'CAR_EXTERIEUR_Roue__port_503398', 'FirstUploadExperveo_carrosseriegrosplan_VIndanslamoquette__port_506486', 'FirstUploadExperveo_carrosseriegrosplan_siegegrosplan__port_506485', 'CAR_EXTERIEUR_Cote_droit_axe_avant__port_504465', 'CAR_EXTERIEUR_Cote_gauche_axe_arriere__port_504198', 'CAR_EXTERIEUR_Face_avant_axe_droit__port_504451', 'CAR_EXTERIEUR_angle_avant_gauche_axe_avant__port_504235', 'FirstUploadExperveo_vin__port_505675', 'CAR_EXTERIEUR_cote_droite__port_504108', 'CAR_INTERIEUR_avant_volant_class_6_levierdevitesse__port_506565', 'FirstUploadExperveo_carrosseriegrosplan_carrosserie__port_506483', 'CAR_EXTERIEUR_Angle_arriere_gauche_axe_arriere__port_504201', 'cartegrise_orientation__port_505064', 'CAR_EXTERIEUR_Angle_arriere_droit_axe_arriere__port_504217', 'CAR_INTERIEUR_avant_vue-arriere_class_1__port_506531', 'CAR_EXTERIEUR_Face_arriere_axe_droit__port_504218', 'CAR_EXTERIEUR_Cote_droit_axe_arriere__port_504214', 'CAR_EXTERIEUR_Angle_avant_droit__port_504087', 'FirstUploadExperveo_carrosseriegrosplan_morceauderoue__port_506484', 'CAR_INTERIEUR_avant_volant_class_6_class_2__port_506563', 'CAR_EXTERIEUR_Angle_arriere_droit__port_504160', 'CAR_EXTERIEUR_arriere__port_504184', 'CAR_INTERIEUR_avant_volant_class_6_boutonrond__port_506562', 'INTERIEUR_Compteur_kilometrique__port_503644', 'CAR_INTERIEUR_avant_vue_gauche_habitacle_class_1__port_506494', 'CAR_EXTERIEUR_Angle_arriere_gauche__port_504170', 'CAR_EXTERIEUR_Angle_avant_droit_axe_arriere__port_504226', 'CAR_EXTERIEUR_Face_arriere_axe_gauche__port_504202', 'CAR_EXTERIEUR_moteur__port_503704', 'FirstUploadExperveo_carrosseriegrosplan_class_6__port_506487', 'CAR_INTERIEUR_siege_arriere_class_1__port_506551', 'CAR_EXTERIEUR_avant__port_504146', 'CAR_EXTERIEUR_Angle_arriere_droit_axe_droit__port_504215', 'CAR_EXTERIEUR_Angle_avant_droit_axe_droit__port_504225', 'CAR_INTERIEUR_avant_volant_class_6_ecrangrosplan__port_506564', 'FirstUploadExperveo_carrosseriegrosplan_moteurgrosplanetdegat__port_506482', 'CAR_INTERIEUR_coffre__port_503412', 'FirstUploadExperveo_rouetranche__port_505677', 'UploadPhotoImmatBest_class_1__port_505051', 'CAR_INTERIEUR_avant_vue-arriere_class_2__port_506532', 'CAR_EXTERIEUR_angle_avant_gauche__port_504098', 'CAR_EXTERIEUR_face_avant_axe_gauche__port_504236', 'CAR_INTERIEUR_avant_vue_droite_habitacle_class_1__port_506540', 'CAR_EXTERIEUR_cote_gauche_axe_avant__port_504233', 'CAR_EXTERIEUR_roue_de_secour__port_503763', 'CAR_EXTERIEUR_Angle_arriere_gauche_axe_gauche__port_504199', 'CAR_EXTERIEUR_cote_gauche__port_504017', 'CAR_INTERIEUR_avant_volant_class_1__port_506503', 'CAR_INTERIEUR_avant_volant_class_2__port_506504', 'CAR_EXTERIEUR_angle_avant_gauche_axe_gauche__port_504234'] 51 SELECT hashtag_id,hashtag FROM MTRBack.hashtags where hashtag in ('FirstUploadExperveo_vignette__port_505674','CAR_EXTERIEUR_Roue__port_503398','FirstUploadExperveo_carrosseriegrosplan_VIndanslamoquette__port_506486','FirstUploadExperveo_carrosseriegrosplan_siegegrosplan__port_506485','CAR_EXTERIEUR_Cote_droit_axe_avant__port_504465','CAR_EXTERIEUR_Cote_gauche_axe_arriere__port_504198','CAR_EXTERIEUR_Face_avant_axe_droit__port_504451','CAR_EXTERIEUR_angle_avant_gauche_axe_avant__port_504235','FirstUploadExperveo_vin__port_505675','CAR_EXTERIEUR_cote_droite__port_504108','CAR_INTERIEUR_avant_volant_class_6_levierdevitesse__port_506565','FirstUploadExperveo_carrosseriegrosplan_carrosserie__port_506483','CAR_EXTERIEUR_Angle_arriere_gauche_axe_arriere__port_504201','cartegrise_orientation__port_505064','CAR_EXTERIEUR_Angle_arriere_droit_axe_arriere__port_504217','CAR_INTERIEUR_avant_vue-arriere_class_1__port_506531','CAR_EXTERIEUR_Face_arriere_axe_droit__port_504218','CAR_EXTERIEUR_Cote_droit_axe_arriere__port_504214','CAR_EXTERIEUR_Angle_avant_droit__port_504087','FirstUploadExperveo_carrosseriegrosplan_morceauderoue__port_506484','CAR_INTERIEUR_avant_volant_class_6_class_2__port_506563','CAR_EXTERIEUR_Angle_arriere_droit__port_504160','CAR_EXTERIEUR_arriere__port_504184','CAR_INTERIEUR_avant_volant_class_6_boutonrond__port_506562','INTERIEUR_Compteur_kilometrique__port_503644','CAR_INTERIEUR_avant_vue_gauche_habitacle_class_1__port_506494','CAR_EXTERIEUR_Angle_arriere_gauche__port_504170','CAR_EXTERIEUR_Angle_avant_droit_axe_arriere__port_504226','CAR_EXTERIEUR_Face_arriere_axe_gauche__port_504202','CAR_EXTERIEUR_moteur__port_503704','FirstUploadExperveo_carrosseriegrosplan_class_6__port_506487','CAR_INTERIEUR_siege_arriere_class_1__port_506551','CAR_EXTERIEUR_avant__port_504146','CAR_EXTERIEUR_Angle_arriere_droit_axe_droit__port_504215','CAR_EXTERIEUR_Angle_avant_droit_axe_droit__port_504225','CAR_INTERIEUR_avant_volant_class_6_ecrangrosplan__port_506564','FirstUploadExperveo_carrosseriegrosplan_moteurgrosplanetdegat__port_506482','CAR_INTERIEUR_coffre__port_503412','FirstUploadExperveo_rouetranche__port_505677','UploadPhotoImmatBest_class_1__port_505051','CAR_INTERIEUR_avant_vue-arriere_class_2__port_506532','CAR_EXTERIEUR_angle_avant_gauche__port_504098','CAR_EXTERIEUR_face_avant_axe_gauche__port_504236','CAR_INTERIEUR_avant_vue_droite_habitacle_class_1__port_506540','CAR_EXTERIEUR_cote_gauche_axe_avant__port_504233','CAR_EXTERIEUR_roue_de_secour__port_503763','CAR_EXTERIEUR_Angle_arriere_gauche_axe_gauche__port_504199','CAR_EXTERIEUR_cote_gauche__port_504017','CAR_INTERIEUR_avant_volant_class_1__port_506503','CAR_INTERIEUR_avant_volant_class_2__port_506504','CAR_EXTERIEUR_angle_avant_gauche_axe_gauche__port_504234'); 51 dict_keys(['cartegrise_orientation__port_505064', 'car_exterieur_angle_arriere_droit_axe_arriere__port_504217', 'car_exterieur_angle_arriere_droit_axe_droit__port_504215', 'car_exterieur_angle_arriere_droit__port_504160', 'car_exterieur_angle_arriere_gauche_axe_arriere__port_504201', 'car_exterieur_angle_arriere_gauche_axe_gauche__port_504199', 'car_exterieur_angle_arriere_gauche__port_504170', 'car_exterieur_angle_avant_droit_axe_arriere__port_504226', 'car_exterieur_angle_avant_droit_axe_droit__port_504225', 'car_exterieur_angle_avant_droit__port_504087', 'car_exterieur_angle_avant_gauche_axe_avant__port_504235', 'car_exterieur_angle_avant_gauche_axe_gauche__port_504234', 'car_exterieur_angle_avant_gauche__port_504098', 'car_exterieur_arriere__port_504184', 'car_exterieur_avant__port_504146', 'car_exterieur_cote_droite__port_504108', 'car_exterieur_cote_droit_axe_arriere__port_504214', 'car_exterieur_cote_droit_axe_avant__port_504465', 'car_exterieur_cote_gauche_axe_arriere__port_504198', 'car_exterieur_cote_gauche_axe_avant__port_504233', 'car_exterieur_cote_gauche__port_504017', 'car_exterieur_face_arriere_axe_droit__port_504218', 'car_exterieur_face_arriere_axe_gauche__port_504202', 'car_exterieur_face_avant_axe_droit__port_504451', 'car_exterieur_face_avant_axe_gauche__port_504236', 'car_exterieur_moteur__port_503704', 'car_exterieur_roue_de_secour__port_503763', 'car_exterieur_roue__port_503398', 'car_interieur_avant_volant_class_1__port_506503', 'car_interieur_avant_volant_class_2__port_506504', 'car_interieur_avant_volant_class_6_boutonrond__port_506562', 'car_interieur_avant_volant_class_6_class_2__port_506563', 'car_interieur_avant_volant_class_6_ecrangrosplan__port_506564', 'car_interieur_avant_volant_class_6_levierdevitesse__port_506565', 'car_interieur_avant_vue-arriere_class_1__port_506531', 'car_interieur_avant_vue-arriere_class_2__port_506532', 'car_interieur_avant_vue_droite_habitacle_class_1__port_506540', 'car_interieur_avant_vue_gauche_habitacle_class_1__port_506494', 'car_interieur_coffre__port_503412', 'car_interieur_siege_arriere_class_1__port_506551', 'firstuploadexperveo_carrosseriegrosplan_carrosserie__port_506483', 'firstuploadexperveo_carrosseriegrosplan_class_6__port_506487', 'firstuploadexperveo_carrosseriegrosplan_morceauderoue__port_506484', 'firstuploadexperveo_carrosseriegrosplan_moteurgrosplanetdegat__port_506482', 'firstuploadexperveo_carrosseriegrosplan_siegegrosplan__port_506485', 'firstuploadexperveo_carrosseriegrosplan_vindanslamoquette__port_506486', 'firstuploadexperveo_rouetranche__port_505677', 'firstuploadexperveo_vignette__port_505674', 'firstuploadexperveo_vin__port_505675', 'interieur_compteur_kilometrique__port_503644', 'uploadphotoimmatbest_class_1__port_505051']) select photo_hashtag_type from MTRDatou.classification_theme where id = 358 thcl : 358 photo_hashtag_type : 337 SELECT phi.hashtag_id , phi.photo_id FROM MTRBack.photo_hashtag_ids phi, MTRUser.mtr_portfolio_photos mp where phi.type = 337 and phi.photo_id = mp.mtr_photo_id and mp.mtr_portfolio_id =510365; {510365: [(917973295, 1), (917973297, 1), (917973302, 1), (917973293, 7), (917973296, 11), (917973300, 11), (917973286, 13), (917973289, 13), (917973301, 24), (917973285, 29), (917973290, 29), (917973299, 29), (917973304, 35), (917973287, 36), (917973298, 36), (917973305, 36), (917973292, 37), (917973291, 41), (917973303, 41), (917973294, 42), (917973288, 46)]} ############################### TEST rotate ################################ test rotate only Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=230 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=230 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 230 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=230 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : rotate list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (917849322) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 917849322 download finish for photo 917849322 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.17151355743408203 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:rotate Fri May 30 03:37:46 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg': 917849322} map_photo_id_path_extension : {917849322: {'path': 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step_rotate ! We are in a linear step without datou_depend ! rotate photos of 90,180,270 degres batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 917849322) and `type` in (0) Loaded 0 chid ids of type : 0 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in () map_chi : {} photo_id in download_rotate_and_save : 917849322 list_chi_loc : 0 Use all angle ! Rotation of photo 917849322 of 90 degree temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg [] 90 remove_crop_border : False version de PIL : 9.5.0 Needs to change image size ! [[ 6.123234e-17 1.000000e+00] [-1.000000e+00 6.123234e-17]] 90 [[ 6.123234e-17 1.000000e+00] [-1.000000e+00 6.123234e-17]] shrink_image : False image_rotate : image_path : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg path_name_rotate : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c454890.jpg image_rotate.mode : RGB Rotation of photo 917849322 of 180 degree temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg [] 180 remove_crop_border : False version de PIL : 9.5.0 Needs to change image size ! [[-1.0000000e+00 1.2246468e-16] [-1.2246468e-16 -1.0000000e+00]] 180 [[-1.0000000e+00 1.2246468e-16] [-1.2246468e-16 -1.0000000e+00]] shrink_image : False image_rotate : image_path : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg path_name_rotate : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548180.jpg image_rotate.mode : RGB Rotation of photo 917849322 of 270 degree temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg [] 270 remove_crop_border : False version de PIL : 9.5.0 Needs to change image size ! [[-1.8369702e-16 -1.0000000e+00] [ 1.0000000e+00 -1.8369702e-16]] 270 [[-1.8369702e-16 -1.0000000e+00] [ 1.0000000e+00 -1.8369702e-16]] shrink_image : False image_rotate : image_path : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg path_name_rotate : temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548270.jpg image_rotate.mode : RGB About to upload 3 photos upload in portfolio : 551782 init cache_photo without model_param we have 3 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569066_1112947 we have uploaded 3 photos in the portfolio 551782 time of upload the photos Elapsed time : 0.9839768409729004 map_filename_photo_id : 3 map_filename_photo_id : {'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c454890.jpg': 1361688556, 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548180.jpg': 1361688557, 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548270.jpg': 1361688558} Len new_chis : 3 Len list_new_chi_with_photo_id : 0 of type : 0 list_new_chi_with_photo_id : [] After datou_step_exec type output : time spend for datou_step_exec : 1.2213459014892578 time spend to save output : 6.29425048828125e-05 total time spend for step 1 : 1.2214088439941406 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : rotate we use saveGeneral [917849322] map_info['map_portfolio_photo'] : {} final : True mtd_id 230 list_pids : [917849322] Looping around the photos to save general results len do output : 3 /1361688556Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688557Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688558Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('230', None, None, None, None, None, None, None, None) ('230', None, '917849322', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 10 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('230', None, '1361688556', 'None', None, None, None, None, None), ('230', None, '1361688557', 'None', None, None, None, None, None), ('230', None, '1361688558', 'None', None, None, None, None, None), ('230', None, '917849322', None, None, None, None, None, None)] time used for this insertion : 0.01357412338256836 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {1361688556: ['917849322', 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c454890.jpg', []], 1361688557: ['917849322', 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548180.jpg', []], 1361688558: ['917849322', 'temp/1748569065_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548270.jpg', []]} test rotate only is a success ! test rotate conditionnel Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=233 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=233 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 233 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=233 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) All sons are already in current list ! All sons are already in current list ! DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : thcl, argmax, rotate list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (917849322) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 917849322 download finish for photo 917849322 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.12744379043579102 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 3 step1:thcl Fri May 30 03:37:47 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg': 917849322} map_photo_id_path_extension : {917849322: {'path': 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step Thcl ! multi_thcl or not :False multi_thcl_cond or not :False dic_thcl : {'500': 1} we are using the classfication for only one thcl 500 In convert_file_to_np l 337 : 1 l343 1 l357 after caffe.io.load_image dimension du image : (3, (2448, 3264, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.0001952648162841797 time to convert the images to numpy array : 0.9133217334747314 total time to convert the images to numpy array : 0.9138388633728027 list photo_ids error: [] list photo_ids correct : [917849322] number of photos to traite : 1 try to delete the photos incorrect in DB tagging for thcl : 500 To do loadFromThcl(), then load ParamDescType : thcl500 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (500) thcls : [{'id': 500, 'mtr_user_id': 31, 'name': 'orientation_carte_grise_all_2', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'carteGrisesVerticales__port_549774,cartegrise_90deg__port_550987,cartesGrisesEnvers__port_549765,portfolio_270deg__port_550988', 'svm_portfolios_learning': '549774,550987,549765,550988', 'photo_hashtag_type': 507, 'photo_desc_type': 3517, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0'}] thcl {'id': 500, 'mtr_user_id': 31, 'name': 'orientation_carte_grise_all_2', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'carteGrisesVerticales__port_549774,cartegrise_90deg__port_550987,cartesGrisesEnvers__port_549765,portfolio_270deg__port_550988', 'svm_portfolios_learning': '549774,550987,549765,550988', 'photo_hashtag_type': 507, 'photo_desc_type': 3517, 'type_classification': 'caffe', 'hashtag_id_list': '0,0,0,0'} Update svm_hashtag_type_desc : 3517 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (3517) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3517, 'orientation_carte_grise_all_2', 16384, 25088, 'orientation_carte_grise_all_2', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2018, 4, 18, 20, 4, 34), datetime.datetime(2018, 4, 18, 20, 4, 34)) To loadFromThcl() : net_3517 begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 5948 max_wait_temp : 1 max_wait : 0 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (3517) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (3517, 'orientation_carte_grise_all_2', 16384, 25088, 'orientation_carte_grise_all_2', 'pool5', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2018, 4, 18, 20, 4, 34), datetime.datetime(2018, 4, 18, 20, 4, 34)) param : , param.caffemodel : orientation_carte_grise_all_2 None mean_file_type : mean_file_path : prototxt_file_path : model : orientation_carte_grise_all_2 Inside get_net Inside get_net before cache_data_model model_param file didn't exist Inside get_net before CDM.load_model_par_type model_name : orientation_carte_grise_all_2 model_type : caffe list file need : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file exist in s3 : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file manque in s3 : [] local folder : /data/models_weight/orientation_carte_grise_all_2 /data/models_weight/orientation_carte_grise_all_2/caffemodel size_local : 537110520 size in s3 : 537110520 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:17 caffemodel already exist and didn't need to update /data/models_weight/orientation_carte_grise_all_2/deploy_conv_normal.prototxt size_local : 4626 size in s3 : 4626 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:16 deploy_conv_normal.prototxt already exist and didn't need to update /data/models_weight/orientation_carte_grise_all_2/deploy_fc.prototxt size_local : 1130 size in s3 : 1130 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:16 deploy_fc.prototxt already exist and didn't need to update /data/models_weight/orientation_carte_grise_all_2/deploy.prototxt size_local : 5653 size in s3 : 5653 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:16 deploy.prototxt already exist and didn't need to update /data/models_weight/orientation_carte_grise_all_2/mean.npy size_local : 1572992 size in s3 : 1572992 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:31 mean.npy already exist and didn't need to update /data/models_weight/orientation_carte_grise_all_2/synset_words.txt size_local : 159 size in s3 : 159 create time local : 2021-08-09 05:29:00 create time in s3 : 2021-08-06 20:07:16 synset_words.txt already exist and didn't need to update Inside get_net after CDM.load_model_par_type After if not only_with_local_cache: /home/admin/workarea/install/caffe_cuda8_python3/python/:/home/admin/workarea/install/darknet/:/home/admin/workarea/git/Velours/python:/home/admin/workarea/install/caffe_frcnn_python3/py-faster-rcnn/caffe-fast-rcnn/python:/home/admin/mtr/.credentials:/home/admin/workarea/install/caffe/python:/home/admin/workarea/install/caffe_frcnn/py-faster-rcnn/tools/:/home/admin/workarea/git/fotonowerpip/:/home/admin/workarea/install/segment-anything:/home/admin//workarea/git/pyfvs/ Here before set mode gpu Doing nothing but we could set mode gpu after set mode gpu prototxt_filename : /data/models_weight/orientation_carte_grise_all_2/deploy.prototxt caffemodel_filename : /data/models_weight/orientation_carte_grise_all_2/caffemodel now we set caffe to gpu mode before predict begin to check gpu status inside check gpu memory l 3637 free memory gpu now : 5795 max_wait_temp : 1 max_wait : 0 dict_keys(['prob', 'pool5']) time used to do the prepocess of the images : 1.6408703327178955 time used to do the prediction : 0.11218404769897461 save descriptor for thcl : 500 (1, 512, 7, 7) Got the blobs of the net to insert : [0, 0, 0, 0, 0, 0, 0, 0, 0, 0] code_as_byte_string:b'0000000000'| time to traite the descriptors : 0.0496525764465332 Testing : ['917849322'] In select_photos_meta_from_ids: SELECT photo_id, url, FROM_UNIXTIME(uploaded_at), latitude, longitude, text FROM MTRBack.photos WHERE photo_id IN (917849322) result : {917849322: {'photo_id': 917849322, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2022/9/13/2bd260e91e91df8378dde8bb8b8c4548.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_13092022_12_32_19_5566.jpg'}} list_photo_exists : [917849322] storage_type for insertDescriptorsMulti : 1 To insert : 917849322 time to insert the descriptors : 0.521902322769165 After datou_step_exec type output : time spend for datou_step_exec : 8.554388523101807 time spend to save output : 7.462501525878906e-05 total time spend for step 1 : 8.554463148117065 step2:argmax Fri May 30 03:37:56 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'917849322': [[('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), ('917849322', 'cartegrise_90deg__port_550987', 0.00050350867, 507, '500'), ('917849322', 'cartesGrisesEnvers__port_549765', 0.00036641024, 507, '500'), ('917849322', 'portfolio_270deg__port_550988', 0.001480374, 507, '500')]]} input_args_next_step : {'917849322': ()} output_args : {'917849322': [[('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), ('917849322', 'cartegrise_90deg__port_550987', 0.00050350867, 507, '500'), ('917849322', 'cartesGrisesEnvers__port_549765', 0.00036641024, 507, '500'), ('917849322', 'portfolio_270deg__port_550988', 0.001480374, 507, '500')]]} args : 917849322 depend.output_id : 0 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ([('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), ('917849322', 'cartegrise_90deg__port_550987', 0.00050350867, 507, '500'), ('917849322', 'cartesGrisesEnvers__port_549765', 0.00036641024, 507, '500'), ('917849322', 'portfolio_270deg__port_550988', 0.001480374, 507, '500')],) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg': 917849322} map_photo_id_path_extension : {917849322: {'path': 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Argmax ! calculate argmax for thcl : 500 After datou_step_exec type output : time spend for datou_step_exec : 0.00024890899658203125 time spend to save output : 6.222724914550781e-05 total time spend for step 2 : 0.00031113624572753906 step3:rotate Fri May 30 03:37:56 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'917849322': [('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg']} input_args_next_step : {'917849322': ()} output_args : {'917849322': [('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg']} args : 917849322 depend.output_id : 1 complete output_args for input 1 : {'917849322': [('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg']} input_args_next_step : {'917849322': ('temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg',)} output_args : {'917849322': [('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500'), 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg']} args : 917849322 depend.output_id : 0 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ('temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg', ('917849322', 'carteGrisesVerticales__port_549774', 0.9976496, 507, '500')) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg': 917849322} map_photo_id_path_extension : {917849322: {'path': 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step_rotate ! We are in a datou with depends ! angle_condi : {'carteGrisesVerticales__port_549774': 0, 'cartegrise_90deg__port_550987': 270, 'portfolio_270deg__port_550988': 90, 'cartesGrisesEnvers__port_549765': 180} rotate photos for hashtag carteGrisesVerticales__port_549774 of 0 degres 1 photos founded : [917849322] batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 917849322) and `type` in (0) Loaded 0 chid ids of type : 0 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in () map_chi : {} photo_id in download_rotate_and_save : 917849322 list_chi_loc : 0 Use all angle ! Rotation of photo 917849322 of 0 degree temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg [] 0 remove_crop_border : False version de PIL : 9.5.0 Needs to change image size ! [[ 1. 0.] [-0. 1.]] 0 [[ 1. 0.] [-0. 1.]] shrink_image : False image_rotate : image_path : temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c4548.jpg path_name_rotate : temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c45480.jpg image_rotate.mode : RGB About to upload 1 photos upload in portfolio : 551782 init cache_photo without model_param we have 1 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569076_1112947 we have uploaded 1 photos in the portfolio 551782 time of upload the photos Elapsed time : 0.6857540607452393 map_filename_photo_id : 1 map_filename_photo_id : {'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c45480.jpg': 1361688559} Len new_chis : 1 Len list_new_chi_with_photo_id : 0 of type : 0 list_new_chi_with_photo_id : [] rotate photos for hashtag cartegrise_90deg__port_550987 of 270 degres 0 photos founded : [] rotate photos for hashtag portfolio_270deg__port_550988 of 90 degres 0 photos founded : [] rotate photos for hashtag cartesGrisesEnvers__port_549765 of 180 degres 0 photos founded : [] After datou_step_exec type output : time spend for datou_step_exec : 0.7945914268493652 time spend to save output : 4.982948303222656e-05 total time spend for step 3 : 0.7946412563323975 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : rotate we use saveGeneral [917849322] map_info['map_portfolio_photo'] : {} final : True mtd_id 233 list_pids : [917849322] Looping around the photos to save general results len do output : 1 /1361688559Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('233', None, None, None, None, None, None, None, None) ('233', None, '917849322', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 4 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('233', None, '1361688559', 'None', None, None, None, None, None), ('233', None, '917849322', None, None, None, None, None, None)] time used for this insertion : 0.013103485107421875 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 3 output : {1361688559: ['917849322', 'temp/1748569067_1112947_917849322_2bd260e91e91df8378dde8bb8b8c45480.jpg', []]} ############################### TEST data_augmentation_ellipse_varroa_tile_rotate ################################ SELECT id FROM MTRPhoto.crop_hashtag_ids WHERE photo_id=937852786 AND `type`=520 DELETE FROM MTRPhoto.crop_hashtag_ids WHERE id IN (3815275204,3815275205,3815275416,3815275417) # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) All sons are already in current list ! DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 316 crop is not linked in the step_by_step architecture ! Step 318 rotate have less inputs used (1) than in the step definition (2) : maybe we manage optionnal inputs ! Step 318 rotate have less outputs used (0) than in the step definition (3) : some outputs may be not used ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! Unexpected type seems boolean for variable list_input_json ERROR or WARNING : can't parse json string Expecting value: line 1 column 1 (char 0) Tried to parse : DATA AUGMENTATION ELLIPSE VARROA TILE ROTATE Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=243 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=243 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 243 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=243 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) All sons are already in current list ! DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 316 crop is not linked in the step_by_step architecture ! Step 318 rotate have less inputs used (1) than in the step definition (2) : maybe we manage optionnal inputs ! Step 318 rotate have less outputs used (0) than in the step definition (3) : some outputs may be not used ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : crop, tile, rotate list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (937852786) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 937852786 download finish for photo 937852786 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.18542027473449707 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 3 step1:crop Fri May 30 03:37:57 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg': 937852786} map_photo_id_path_extension : {937852786: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Crop ! param_json : {'hashtag_id_ellipse': 2087736828, 'photo_hashtag_type_from_ellipse': 520, 'token': '78d09a0790ec6ecbf119343125a81fdc', 'portfolio_name': 'crop_detect_varroa', 'photo_hashtag_type': 407, 'feed_id_new_photos_not_used': 549103, 'host': 'www.fotonower.com', 'margin': 8, 'upload_type': 'python'} margin_type : margin margin_value : [8, 8, 8, 8] Loading chi in step crop with photo_hashtag_type : 407 Loading chi in step crop for list_pids : 1 ! batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 937852786) and `type` in (407) Loaded 4 chid ids of type : 407 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (8165075,8165076,8165077,8165078) +WARNING : Unexpected points, we should remove this data for chi_id : 8165075, for now we just ignore these empty polygon points +WARNING : Unexpected points, we should remove this data for chi_id : 8165076, for now we just ignore these empty polygon points +WARNING : Unexpected points, we should remove this data for chi_id : 8165077, for now we just ignore these empty polygon points +WARNING : Unexpected points, we should remove this data for chi_id : 8165078, for now we just ignore these empty polygon points SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (8165075,8165076,8165077,8165078) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (8165075,8165076,8165077,8165078) select photo_id, sub_photo_id, x0, x1, y0, y1, resize_coeff_x, resize_coeff_y, crop_type, id from MTRPhoto.photo_sub_photos where photo_id in ( 937852786) WARNING : margin is only used for type bib ! type of cropped photo chosen : we resize croppped photo by 1 on x axis and by 1 on y axis new_file_path_bib_crop : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg new_file_path_bib_crop : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg new_file_path_bib_crop : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg new_file_path_bib_crop : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg map_result returned by crop_photo_return_map_crop : length : 4 map_result after crop : {8165075: {'crop': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg', 'photo_id': 937852786, 'sub_photo_id': -1, 'coordonates': (426, 467, 312, 347), 'sub_photo_infos': (418, 475, 304, 355, 1, 1), 'same_chi': False}, 8165076: {'crop': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg', 'photo_id': 937852786, 'sub_photo_id': -1, 'coordonates': (411, 445, 443, 480), 'sub_photo_infos': (403, 453, 435, 480, 1, 1), 'same_chi': False}, 8165077: {'crop': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg', 'photo_id': 937852786, 'sub_photo_id': -1, 'coordonates': (103, 138, 358, 396), 'sub_photo_infos': (95, 146, 350, 404, 1, 1), 'same_chi': False}, 8165078: {'crop': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg', 'photo_id': 937852786, 'sub_photo_id': -1, 'coordonates': (104, 131, 256, 292), 'sub_photo_infos': (96, 139, 248, 300, 1, 1), 'same_chi': False}} Here we crop with rles About to insert : list_path_to_insert length 4 new photo from crops ! About to upload 4 photos https://marlene.fotonower.com/api/v1/secured/portfolio/new?name=crop_detect_varroa&access_token=78d09a0790ec6ecbf119343125a81fdc upload in portfolio : 23444765 init cache_photo without model_param we have 4 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569079_1112947 INSERT ignore into MTRUser.mtr_portfolio_photos (`mtr_portfolio_id`, `mtr_photo_id`, `mtr_user_id`, `created_at`) VALUES (23444765, 1361688560, 0, NOW()),(23444765, 1361688561, 0, NOW()),(23444765, 1361688562, 0, NOW()),(23444765, 1361688563, 0, NOW()) 4 we have uploaded 4 photos in the portfolio 23444765 time of upload the photos Elapsed time : 3.2430458068847656 {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg': 1361688560, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg': 1361688561, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg': 1361688562, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg': 1361688563} list_errors : [] map_result_insert : {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg': 1361688560, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg': 1361688561, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg': 1361688562, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg': 1361688563} Now we prepare data that will be used for ellipse search ! chi_id found to be used 8165075 path of cropped varroa found to be used to match on an ellipse temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg sub_photo_id found to be used 1361688560 chi_id found to be used 8165076 path of cropped varroa found to be used to match on an ellipse temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg sub_photo_id found to be used 1361688561 chi_id found to be used 8165077 path of cropped varroa found to be used to match on an ellipse temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg sub_photo_id found to be used 1361688562 chi_id found to be used 8165078 path of cropped varroa found to be used to match on an ellipse temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg sub_photo_id found to be used 1361688563 insert ignore into MTRPhoto.crop_sub_photo_ids (crop_hashtag_id, sub_photo_id, mtr_user_id) VALUES (%s,%s,%s) : [(8165075, '1361688560', 31), (8165076, '1361688561', 31), (8165077, '1361688562', 31), (8165078, '1361688563', 31)] map of cropped photos with some data : {'1361688560': ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg', (426, 467, 312, 347)], '1361688561': ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg', (411, 445, 443, 480)], '1361688562': ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg', (103, 138, 358, 396)], '1361688563': ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg', (104, 131, 256, 292)]} About to compute ellipse and record with type : 520 (54, 57) (51, 57) [54, 57] (54, 57) score : 5120 strategy_opt : 5| strategy_opt : [('excentricity', [0.5, 2.0, 0.05]), ('angle', [-90.0, 90.0, 5.0]), ('xc', [14.25, 42.75, 1.78125]), ('yc', [12.75, 38.25, 1.59375]), ('radius', [14.25, 42.75, 1.78125])] {0.5: 18351, 0.55: 16004, 0.6000000000000001: 14127, 0.65: 12575, 0.7000000000000001: 11300, 0.75: 10158, 0.8: 9208, 0.8500000000000001: 8418, 0.9: 7642, 0.9500000000000001: 6925, 1.0: 6364, 1.05: 5843, 1.1: 5353, 1.1500000000000001: 4868, 1.2000000000000002: 4553, 1.25: 4250, 1.3: 3947, 1.35: 3698, 1.4000000000000001: 3438, 1.4500000000000002: 3230, 1.5: 3012, 1.55: 2829, 1.6: 2636, 1.6500000000000001: 2535, 1.7000000000000002: 2460, 1.75: 2388, 1.8: 2372, 1.85: 2354, 1.9000000000000001: 2332, 1.9500000000000002: 2311} [(1.9500000000000002, 2311), (1.9000000000000001, 2332), (1.85, 2354), (1.8, 2372), (1.75, 2388), (1.7000000000000002, 2460), (1.6500000000000001, 2535), (1.6, 2636), (1.55, 2829), (1.5, 3012), (1.4500000000000002, 3230), (1.4000000000000001, 3438), (1.35, 3698), (1.3, 3947), (1.25, 4250), (1.2000000000000002, 4553), (1.1500000000000001, 4868), (1.1, 5353), (1.05, 5843), (1.0, 6364), (0.9500000000000001, 6925), (0.9, 7642), (0.8500000000000001, 8418), (0.8, 9208), (0.75, 10158), (0.7000000000000001, 11300), (0.65, 12575), (0.6000000000000001, 14127), (0.55, 16004), (0.5, 18351)] arg_min reach at : 1.9500000000000002 with value = 2311 | arg_min : 1.9500000000000002 min_score : 2311{-90.0: 3359, -85.0: 3372, -80.0: 3395, -75.0: 3380, -70.0: 3304, -65.0: 3180, -60.0: 2991, -55.0: 2793, -50.0: 2601, -45.0: 2377, -40.0: 2183, -35.0: 2078, -30.0: 1968, -25.0: 1968, -20.0: 2021, -15.0: 2036, -10.0: 2102, -5.0: 2188, 0.0: 2208, 5.0: 2265, 10.0: 2311, 15.0: 2333, 20.0: 2362, 25.0: 2375, 30.0: 2375, 35.0: 2397, 40.0: 2370, 45.0: 2421, 50.0: 2491, 55.0: 2661, 60.0: 2826, 65.0: 2993, 70.0: 3106, 75.0: 3182, 80.0: 3263, 85.0: 3306} [(-30.0, 1968), (-25.0, 1968), (-20.0, 2021), (-15.0, 2036), (-35.0, 2078), (-10.0, 2102), (-40.0, 2183), (-5.0, 2188), (0.0, 2208), (5.0, 2265), (10.0, 2311), (15.0, 2333), (20.0, 2362), (40.0, 2370), (25.0, 2375), (30.0, 2375), (-45.0, 2377), (35.0, 2397), (45.0, 2421), (50.0, 2491), (-50.0, 2601), (55.0, 2661), (-55.0, 2793), (60.0, 2826), (-60.0, 2991), (65.0, 2993), (70.0, 3106), (-65.0, 3180), (75.0, 3182), (80.0, 3263), (-70.0, 3304), (85.0, 3306), (-90.0, 3359), (-85.0, 3372), (-75.0, 3380), (-80.0, 3395)] arg_min reach at : -30.0 with value = 1968 | arg_min : -30.0 min_score : 1968{14.25: 1703, 16.03125: 1654, 17.8125: 1614, 19.59375: 1679, 21.375: 1702, 23.15625: 1712, 24.9375: 1730, 26.71875: 1847, 28.5: 1968, 30.28125: 2177, 32.0625: 2456, 33.84375: 2827, 35.625: 3294, 37.40625: 3723, 39.1875: 4161, 40.96875: 4643} [(17.8125, 1614), (16.03125, 1654), (19.59375, 1679), (21.375, 1702), (14.25, 1703), (23.15625, 1712), (24.9375, 1730), (26.71875, 1847), (28.5, 1968), (30.28125, 2177), (32.0625, 2456), (33.84375, 2827), (35.625, 3294), (37.40625, 3723), (39.1875, 4161), (40.96875, 4643)] arg_min reach at : 17.8125 with value = 1614 | arg_min : 17.8125 min_score : 1614{12.75: 5534, 14.34375: 5352, 15.9375: 5116, 17.53125: 4723, 19.125: 4111, 20.71875: 3506, 22.3125: 2799, 23.90625: 2194, 25.5: 1614, 27.09375: 1339, 28.6875: 1206, 30.28125: 1137, 31.875: 1105, 33.46875: 1339, 35.0625: 1659, 36.65625: 2022} [(31.875, 1105), (30.28125, 1137), (28.6875, 1206), (27.09375, 1339), (33.46875, 1339), (25.5, 1614), (35.0625, 1659), (36.65625, 2022), (23.90625, 2194), (22.3125, 2799), (20.71875, 3506), (19.125, 4111), (17.53125, 4723), (15.9375, 5116), (14.34375, 5352), (12.75, 5534)] arg_min reach at : 31.875 with value = 1105 | arg_min : 31.875 min_score : 1105{14.25: 1942, 16.03125: 1853, 17.8125: 1756, 19.59375: 1651, 21.375: 1533, 23.15625: 1424, 24.9375: 1309, 26.71875: 1199, 28.5: 1105, 30.28125: 1219, 32.0625: 1332, 33.84375: 1547, 35.625: 1881, 37.40625: 2378, 39.1875: 3042, 40.96875: 3824} [(28.5, 1105), (26.71875, 1199), (30.28125, 1219), (24.9375, 1309), (32.0625, 1332), (23.15625, 1424), (21.375, 1533), (33.84375, 1547), (19.59375, 1651), (17.8125, 1756), (16.03125, 1853), (35.625, 1881), (14.25, 1942), (37.40625, 2378), (39.1875, 3042), (40.96875, 3824)] arg_min reach at : 28.5 with value = 1105 | arg_min : 28.5 min_score : 1105{0.5: 17397, 0.55: 15138, 0.6000000000000001: 13352, 0.65: 11859, 0.7000000000000001: 10523, 0.75: 9463, 0.8: 8469, 0.8500000000000001: 7565, 0.9: 6551, 0.9500000000000001: 5592, 1.0: 4856, 1.05: 4291, 1.1: 3654, 1.1500000000000001: 3099, 1.2000000000000002: 2647, 1.25: 2309, 1.3: 1983, 1.35: 1755, 1.4000000000000001: 1594, 1.4500000000000002: 1447, 1.5: 1358, 1.55: 1323, 1.6: 1219, 1.6500000000000001: 1215, 1.7000000000000002: 1201, 1.75: 1131, 1.8: 1134, 1.85: 1142, 1.9000000000000001: 1137, 1.9500000000000002: 1105} [(1.9500000000000002, 1105), (1.75, 1131), (1.8, 1134), (1.9000000000000001, 1137), (1.85, 1142), (1.7000000000000002, 1201), (1.6500000000000001, 1215), (1.6, 1219), (1.55, 1323), (1.5, 1358), (1.4500000000000002, 1447), (1.4000000000000001, 1594), (1.35, 1755), (1.3, 1983), (1.25, 2309), (1.2000000000000002, 2647), (1.1500000000000001, 3099), (1.1, 3654), (1.05, 4291), (1.0, 4856), (0.9500000000000001, 5592), (0.9, 6551), (0.8500000000000001, 7565), (0.8, 8469), (0.75, 9463), (0.7000000000000001, 10523), (0.65, 11859), (0.6000000000000001, 13352), (0.55, 15138), (0.5, 17397)] arg_min reach at : 1.9500000000000002 with value = 1105 arg_min : 1.9500000000000002 min_score : 1105{-90.0: 3514, -85.0: 3396, -80.0: 3238, -75.0: 3017, -70.0: 2764, -65.0: 2549, -60.0: 2339, -55.0: 2105, -50.0: 1912, -45.0: 1701, -40.0: 1467, -35.0: 1266, -30.0: 1105, -25.0: 1110, -20.0: 1114, -15.0: 1111, -10.0: 1110, -5.0: 1121, 0.0: 1120, 5.0: 1118, 10.0: 1109, 15.0: 1109, 20.0: 1099, 25.0: 1088, 30.0: 1108, 35.0: 1355, 40.0: 1649, 45.0: 1925, 50.0: 2257, 55.0: 2542, 60.0: 2845, 65.0: 3137, 70.0: 3370, 75.0: 3500, 80.0: 3578, 85.0: 3540} [(25.0, 1088), (20.0, 1099), (-30.0, 1105), (30.0, 1108), (10.0, 1109), (15.0, 1109), (-25.0, 1110), (-10.0, 1110), (-15.0, 1111), (-20.0, 1114), (5.0, 1118), (0.0, 1120), (-5.0, 1121), (-35.0, 1266), (35.0, 1355), (-40.0, 1467), (40.0, 1649), (-45.0, 1701), (-50.0, 1912), (45.0, 1925), (-55.0, 2105), (50.0, 2257), (-60.0, 2339), (55.0, 2542), (-65.0, 2549), (-70.0, 2764), (60.0, 2845), (-75.0, 3017), (65.0, 3137), (-80.0, 3238), (70.0, 3370), (-85.0, 3396), (75.0, 3500), (-90.0, 3514), (85.0, 3540), (80.0, 3578)] arg_min reach at : 25.0 with value = 1088 arg_min : 25.0 min_score : 1088{14.25: 1174, 16.03125: 1133, 17.8125: 1088, 19.59375: 1051, 21.375: 1026, 23.15625: 997, 24.9375: 979, 26.71875: 1139, 28.5: 1344, 30.28125: 1564, 32.0625: 1949, 33.84375: 2422, 35.625: 2930, 37.40625: 3453, 39.1875: 4006, 40.96875: 4592} [(24.9375, 979), (23.15625, 997), (21.375, 1026), (19.59375, 1051), (17.8125, 1088), (16.03125, 1133), (26.71875, 1139), (14.25, 1174), (28.5, 1344), (30.28125, 1564), (32.0625, 1949), (33.84375, 2422), (35.625, 2930), (37.40625, 3453), (39.1875, 4006), (40.96875, 4592)] arg_min reach at : 24.9375 with value = 979 arg_min : 24.9375 min_score : 979{12.75: 5539, 14.34375: 5239, 15.9375: 4911, 17.53125: 4425, 19.125: 3749, 20.71875: 3181, 22.3125: 2575, 23.90625: 2124, 25.5: 1808, 27.09375: 1478, 28.6875: 1290, 30.28125: 1104, 31.875: 979, 33.46875: 1070, 35.0625: 1368, 36.65625: 1872} [(31.875, 979), (33.46875, 1070), (30.28125, 1104), (28.6875, 1290), (35.0625, 1368), (27.09375, 1478), (25.5, 1808), (36.65625, 1872), (23.90625, 2124), (22.3125, 2575), (20.71875, 3181), (19.125, 3749), (17.53125, 4425), (15.9375, 4911), (14.34375, 5239), (12.75, 5539)] arg_min reach at : 31.875 with value = 979 arg_min : 31.875 min_score : 979{14.25: 1939, 16.03125: 1852, 17.8125: 1755, 19.59375: 1643, 21.375: 1531, 23.15625: 1397, 24.9375: 1264, 26.71875: 1119, 28.5: 979, 30.28125: 1023, 32.0625: 1319, 33.84375: 1771, 35.625: 2432, 37.40625: 3324, 39.1875: 4299, 40.96875: 5405} [(28.5, 979), (30.28125, 1023), (26.71875, 1119), (24.9375, 1264), (32.0625, 1319), (23.15625, 1397), (21.375, 1531), (19.59375, 1643), (17.8125, 1755), (33.84375, 1771), (16.03125, 1852), (14.25, 1939), (35.625, 2432), (37.40625, 3324), (39.1875, 4299), (40.96875, 5405)] arg_min reach at : 28.5 with value = 979 arg_min : 28.5 min_score : 979 yc : 31.875 xc : 24.9375 angle : 25.0 radius : 28.5 excentricity : 1.9500000000000002 yc : 31.875 xc : 24.9375 angle : 25.0 radius : 28.5 excentricity : 1.9500000000000002 x0 : 426 y1 : 347 width : 41, height : 35, area : 1435, score : 1.0 x0 : 432 y1 : 355 width : 35, height : 52, area : 1820, score : 1.0 Now saving polygons points : 1| batch 1 Loaded 1 chid ids of type : 520 CHI and polygons saved ! (47, 50) (45, 50) [47, 50] (47, 50) score : 5362 strategy_opt : 5| strategy_opt : [('excentricity', [0.5, 2.0, 0.05]), ('angle', [-90.0, 90.0, 5.0]), ('xc', [12.5, 37.5, 1.5625]), ('yc', [11.25, 33.75, 1.40625]), ('radius', [12.5, 37.5, 1.5625])] {0.5: 15776, 0.55: 13889, 0.6000000000000001: 12446, 0.65: 11147, 0.7000000000000001: 10181, 0.75: 9228, 0.8: 8441, 0.8500000000000001: 7765, 0.9: 7258, 0.9500000000000001: 6684, 1.0: 6195, 1.05: 5920, 1.1: 5487, 1.1500000000000001: 5099, 1.2000000000000002: 4795, 1.25: 4511, 1.3: 4245, 1.35: 4042, 1.4000000000000001: 3828, 1.4500000000000002: 3629, 1.5: 3467, 1.55: 3235, 1.6: 3126, 1.6500000000000001: 2946, 1.7000000000000002: 2826, 1.75: 2700, 1.8: 2549, 1.85: 2472, 1.9000000000000001: 2381, 1.9500000000000002: 2281} [(1.9500000000000002, 2281), (1.9000000000000001, 2381), (1.85, 2472), (1.8, 2549), (1.75, 2700), (1.7000000000000002, 2826), (1.6500000000000001, 2946), (1.6, 3126), (1.55, 3235), (1.5, 3467), (1.4500000000000002, 3629), (1.4000000000000001, 3828), (1.35, 4042), (1.3, 4245), (1.25, 4511), (1.2000000000000002, 4795), (1.1500000000000001, 5099), (1.1, 5487), (1.05, 5920), (1.0, 6195), (0.9500000000000001, 6684), (0.9, 7258), (0.8500000000000001, 7765), (0.8, 8441), (0.75, 9228), (0.7000000000000001, 10181), (0.65, 11147), (0.6000000000000001, 12446), (0.55, 13889), (0.5, 15776)] arg_min reach at : 1.9500000000000002 with value = 2281 | arg_min : 1.9500000000000002 min_score : 2281{-90.0: 3295, -85.0: 3240, -80.0: 3290, -75.0: 3269, -70.0: 3287, -65.0: 3223, -60.0: 3095, -55.0: 3043, -50.0: 2985, -45.0: 2906, -40.0: 2796, -35.0: 2724, -30.0: 2625, -25.0: 2489, -20.0: 2330, -15.0: 2220, -10.0: 2127, -5.0: 2127, 0.0: 2143, 5.0: 2182, 10.0: 2281, 15.0: 2385, 20.0: 2506, 25.0: 2665, 30.0: 2768, 35.0: 2900, 40.0: 2939, 45.0: 2994, 50.0: 3007, 55.0: 3076, 60.0: 3095, 65.0: 3201, 70.0: 3243, 75.0: 3247, 80.0: 3279, 85.0: 3240} [(-10.0, 2127), (-5.0, 2127), (0.0, 2143), (5.0, 2182), (-15.0, 2220), (10.0, 2281), (-20.0, 2330), (15.0, 2385), (-25.0, 2489), (20.0, 2506), (-30.0, 2625), (25.0, 2665), (-35.0, 2724), (30.0, 2768), (-40.0, 2796), (35.0, 2900), (-45.0, 2906), (40.0, 2939), (-50.0, 2985), (45.0, 2994), (50.0, 3007), (-55.0, 3043), (55.0, 3076), (-60.0, 3095), (60.0, 3095), (65.0, 3201), (-65.0, 3223), (-85.0, 3240), (85.0, 3240), (70.0, 3243), (75.0, 3247), (-75.0, 3269), (80.0, 3279), (-70.0, 3287), (-80.0, 3290), (-90.0, 3295)] arg_min reach at : -10.0 with value = 2127 | arg_min : -10.0 min_score : 2127{12.5: 2171, 14.0625: 2288, 15.625: 2314, 17.1875: 2360, 18.75: 2347, 20.3125: 2300, 21.875: 2249, 23.4375: 2188, 25.0: 2127, 26.5625: 2163, 28.125: 2315, 29.6875: 2476, 31.25: 2756, 32.8125: 2992, 34.375: 3319, 35.9375: 3672} [(25.0, 2127), (26.5625, 2163), (12.5, 2171), (23.4375, 2188), (21.875, 2249), (14.0625, 2288), (20.3125, 2300), (15.625, 2314), (28.125, 2315), (18.75, 2347), (17.1875, 2360), (29.6875, 2476), (31.25, 2756), (32.8125, 2992), (34.375, 3319), (35.9375, 3672)] arg_min reach at : 25.0 with value = 2127 | arg_min : 25.0 min_score : 2127{11.25: 6815, 12.65625: 6505, 14.0625: 5928, 15.46875: 5326, 16.875: 4772, 18.28125: 4095, 19.6875: 3413, 21.09375: 2728, 22.5: 2127, 23.90625: 1694, 25.3125: 1345, 26.71875: 1048, 28.125: 823, 29.53125: 728, 30.9375: 714, 32.34375: 1027} [(30.9375, 714), (29.53125, 728), (28.125, 823), (32.34375, 1027), (26.71875, 1048), (25.3125, 1345), (23.90625, 1694), (22.5, 2127), (21.09375, 2728), (19.6875, 3413), (18.28125, 4095), (16.875, 4772), (15.46875, 5326), (14.0625, 5928), (12.65625, 6505), (11.25, 6815)] arg_min reach at : 30.9375 with value = 714 | arg_min : 30.9375 min_score : 714{12.5: 1275, 14.0625: 1207, 15.625: 1132, 17.1875: 1049, 18.75: 966, 20.3125: 866, 21.875: 784, 23.4375: 729, 25.0: 714, 26.5625: 976, 28.125: 1546, 29.6875: 2147, 31.25: 2981, 32.8125: 3767, 34.375: 4761, 35.9375: 5732} [(25.0, 714), (23.4375, 729), (21.875, 784), (20.3125, 866), (18.75, 966), (26.5625, 976), (17.1875, 1049), (15.625, 1132), (14.0625, 1207), (12.5, 1275), (28.125, 1546), (29.6875, 2147), (31.25, 2981), (32.8125, 3767), (34.375, 4761), (35.9375, 5732)] arg_min reach at : 25.0 with value = 714 | arg_min : 25.0 min_score : 714{0.5: 20107, 0.55: 18155, 0.6000000000000001: 16503, 0.65: 15058, 0.7000000000000001: 13673, 0.75: 12474, 0.8: 11100, 0.8500000000000001: 9667, 0.9: 8374, 0.9500000000000001: 7344, 1.0: 6274, 1.05: 5570, 1.1: 4846, 1.1500000000000001: 4170, 1.2000000000000002: 3623, 1.25: 3211, 1.3: 2832, 1.35: 2475, 1.4000000000000001: 2182, 1.4500000000000002: 1949, 1.5: 1655, 1.55: 1513, 1.6: 1333, 1.6500000000000001: 1182, 1.7000000000000002: 1038, 1.75: 921, 1.8: 841, 1.85: 732, 1.9000000000000001: 733, 1.9500000000000002: 714} [(1.9500000000000002, 714), (1.85, 732), (1.9000000000000001, 733), (1.8, 841), (1.75, 921), (1.7000000000000002, 1038), (1.6500000000000001, 1182), (1.6, 1333), (1.55, 1513), (1.5, 1655), (1.4500000000000002, 1949), (1.4000000000000001, 2182), (1.35, 2475), (1.3, 2832), (1.25, 3211), (1.2000000000000002, 3623), (1.1500000000000001, 4170), (1.1, 4846), (1.05, 5570), (1.0, 6274), (0.9500000000000001, 7344), (0.9, 8374), (0.8500000000000001, 9667), (0.8, 11100), (0.75, 12474), (0.7000000000000001, 13673), (0.65, 15058), (0.6000000000000001, 16503), (0.55, 18155), (0.5, 20107)] arg_min reach at : 1.9500000000000002 with value = 714 arg_min : 1.9500000000000002 min_score : 714{-90.0: 2866, -85.0: 3006, -80.0: 3063, -75.0: 3168, -70.0: 3193, -65.0: 3193, -60.0: 3149, -55.0: 3022, -50.0: 2885, -45.0: 2656, -40.0: 2413, -35.0: 2119, -30.0: 1819, -25.0: 1500, -20.0: 1171, -15.0: 865, -10.0: 714, -5.0: 668, 0.0: 689, 5.0: 734, 10.0: 824, 15.0: 898, 20.0: 1116, 25.0: 1368, 30.0: 1610, 35.0: 1866, 40.0: 2105, 45.0: 2304, 50.0: 2467, 55.0: 2604, 60.0: 2731, 65.0: 2753, 70.0: 2753, 75.0: 2761, 80.0: 2722, 85.0: 2786} [(-5.0, 668), (0.0, 689), (-10.0, 714), (5.0, 734), (10.0, 824), (-15.0, 865), (15.0, 898), (20.0, 1116), (-20.0, 1171), (25.0, 1368), (-25.0, 1500), (30.0, 1610), (-30.0, 1819), (35.0, 1866), (40.0, 2105), (-35.0, 2119), (45.0, 2304), (-40.0, 2413), (50.0, 2467), (55.0, 2604), (-45.0, 2656), (80.0, 2722), (60.0, 2731), (65.0, 2753), (70.0, 2753), (75.0, 2761), (85.0, 2786), (-90.0, 2866), (-50.0, 2885), (-85.0, 3006), (-55.0, 3022), (-80.0, 3063), (-60.0, 3149), (-75.0, 3168), (-70.0, 3193), (-65.0, 3193)] arg_min reach at : -5.0 with value = 668 arg_min : -5.0 min_score : 668{12.5: 1164, 14.0625: 1082, 15.625: 1019, 17.1875: 933, 18.75: 866, 20.3125: 765, 21.875: 713, 23.4375: 655, 25.0: 668, 26.5625: 845, 28.125: 1107, 29.6875: 1374, 31.25: 1691, 32.8125: 2036, 34.375: 2375, 35.9375: 2731} [(23.4375, 655), (25.0, 668), (21.875, 713), (20.3125, 765), (26.5625, 845), (18.75, 866), (17.1875, 933), (15.625, 1019), (14.0625, 1082), (28.125, 1107), (12.5, 1164), (29.6875, 1374), (31.25, 1691), (32.8125, 2036), (34.375, 2375), (35.9375, 2731)] arg_min reach at : 23.4375 with value = 655 arg_min : 23.4375 min_score : 655{11.25: 7209, 12.65625: 6826, 14.0625: 6195, 15.46875: 5638, 16.875: 4941, 18.28125: 4177, 19.6875: 3484, 21.09375: 2750, 22.5: 2183, 23.90625: 1709, 25.3125: 1327, 26.71875: 1008, 28.125: 779, 29.53125: 631, 30.9375: 655, 32.34375: 920} [(29.53125, 631), (30.9375, 655), (28.125, 779), (32.34375, 920), (26.71875, 1008), (25.3125, 1327), (23.90625, 1709), (22.5, 2183), (21.09375, 2750), (19.6875, 3484), (18.28125, 4177), (16.875, 4941), (15.46875, 5638), (14.0625, 6195), (12.65625, 6826), (11.25, 7209)] arg_min reach at : 29.53125 with value = 631 arg_min : 29.53125 min_score : 631{12.5: 1282, 14.0625: 1211, 15.625: 1130, 17.1875: 1050, 18.75: 960, 20.3125: 871, 21.875: 772, 23.4375: 672, 25.0: 631, 26.5625: 666, 28.125: 947, 29.6875: 1472, 31.25: 2184, 32.8125: 3032, 34.375: 3932, 35.9375: 5005} [(25.0, 631), (26.5625, 666), (23.4375, 672), (21.875, 772), (20.3125, 871), (28.125, 947), (18.75, 960), (17.1875, 1050), (15.625, 1130), (14.0625, 1211), (12.5, 1282), (29.6875, 1472), (31.25, 2184), (32.8125, 3032), (34.375, 3932), (35.9375, 5005)] arg_min reach at : 25.0 with value = 631 arg_min : 25.0 min_score : 631 yc : 29.53125 xc : 23.4375 angle : -5.0 radius : 25.0 excentricity : 1.9500000000000002 yc : 29.53125 xc : 23.4375 angle : -5.0 radius : 25.0 excentricity : 1.9500000000000002 x0 : 411 y1 : 480 width : 34, height : 37, area : 1258, score : 1.0 x0 : 419 y1 : 483 width : 26, height : 49, area : 1274, score : 1.0 Now saving polygons points : 1| batch 1 Loaded 2 chid ids of type : 520 + CHI and polygons saved ! (55, 54) (54, 51) [55, 54] (55, 54) score : 4603 strategy_opt : 5| strategy_opt : [('excentricity', [0.5, 2.0, 0.05]), ('angle', [-90.0, 90.0, 5.0]), ('xc', [12.75, 38.25, 1.59375]), ('yc', [13.5, 40.5, 1.6875]), ('radius', [13.5, 40.5, 1.6875])] {0.5: 14831, 0.55: 12756, 0.6000000000000001: 10989, 0.65: 9469, 0.7000000000000001: 8158, 0.75: 7251, 0.8: 6254, 0.8500000000000001: 5559, 0.9: 5119, 0.9500000000000001: 4564, 1.0: 4072, 1.05: 3783, 1.1: 3496, 1.1500000000000001: 3342, 1.2000000000000002: 3228, 1.25: 3161, 1.3: 3127, 1.35: 3087, 1.4000000000000001: 3047, 1.4500000000000002: 3050, 1.5: 2990, 1.55: 2992, 1.6: 2992, 1.6500000000000001: 2985, 1.7000000000000002: 2988, 1.75: 2999, 1.8: 2994, 1.85: 2981, 1.9000000000000001: 3002, 1.9500000000000002: 2995} [(1.85, 2981), (1.6500000000000001, 2985), (1.7000000000000002, 2988), (1.5, 2990), (1.55, 2992), (1.6, 2992), (1.8, 2994), (1.9500000000000002, 2995), (1.75, 2999), (1.9000000000000001, 3002), (1.4000000000000001, 3047), (1.4500000000000002, 3050), (1.35, 3087), (1.3, 3127), (1.25, 3161), (1.2000000000000002, 3228), (1.1500000000000001, 3342), (1.1, 3496), (1.05, 3783), (1.0, 4072), (0.9500000000000001, 4564), (0.9, 5119), (0.8500000000000001, 5559), (0.8, 6254), (0.75, 7251), (0.7000000000000001, 8158), (0.65, 9469), (0.6000000000000001, 10989), (0.55, 12756), (0.5, 14831)] arg_min reach at : 1.85 with value = 2981 | arg_min : 1.85 min_score : 2981{-90.0: 1442, -85.0: 1411, -80.0: 1411, -75.0: 1423, -70.0: 1409, -65.0: 1428, -60.0: 1417, -55.0: 1385, -50.0: 1356, -45.0: 1374, -40.0: 1401, -35.0: 1501, -30.0: 1635, -25.0: 1772, -20.0: 1988, -15.0: 2183, -10.0: 2376, -5.0: 2552, 0.0: 2691, 5.0: 2860, 10.0: 2981, 15.0: 3063, 20.0: 3044, 25.0: 2938, 30.0: 2889, 35.0: 2865, 40.0: 2787, 45.0: 2694, 50.0: 2566, 55.0: 2452, 60.0: 2275, 65.0: 2055, 70.0: 1849, 75.0: 1731, 80.0: 1532, 85.0: 1466} [(-50.0, 1356), (-45.0, 1374), (-55.0, 1385), (-40.0, 1401), (-70.0, 1409), (-85.0, 1411), (-80.0, 1411), (-60.0, 1417), (-75.0, 1423), (-65.0, 1428), (-90.0, 1442), (85.0, 1466), (-35.0, 1501), (80.0, 1532), (-30.0, 1635), (75.0, 1731), (-25.0, 1772), (70.0, 1849), (-20.0, 1988), (65.0, 2055), (-15.0, 2183), (60.0, 2275), (-10.0, 2376), (55.0, 2452), (-5.0, 2552), (50.0, 2566), (0.0, 2691), (45.0, 2694), (40.0, 2787), (5.0, 2860), (35.0, 2865), (30.0, 2889), (25.0, 2938), (10.0, 2981), (20.0, 3044), (15.0, 3063)] arg_min reach at : -50.0 with value = 1356 | arg_min : -50.0 min_score : 1356{12.75: 3483, 14.34375: 3154, 15.9375: 2818, 17.53125: 2499, 19.125: 2212, 20.71875: 1992, 22.3125: 1750, 23.90625: 1554, 25.5: 1356, 27.09375: 1213, 28.6875: 1123, 30.28125: 1079, 31.875: 1312, 33.46875: 1586, 35.0625: 1971, 36.65625: 2464} [(30.28125, 1079), (28.6875, 1123), (27.09375, 1213), (31.875, 1312), (25.5, 1356), (23.90625, 1554), (33.46875, 1586), (22.3125, 1750), (35.0625, 1971), (20.71875, 1992), (19.125, 2212), (36.65625, 2464), (17.53125, 2499), (15.9375, 2818), (14.34375, 3154), (12.75, 3483)] arg_min reach at : 30.28125 with value = 1079 | arg_min : 30.28125 min_score : 1079{13.5: 1299, 15.1875: 1190, 16.875: 1129, 18.5625: 1073, 20.25: 1041, 21.9375: 1025, 23.625: 995, 25.3125: 1002, 27.0: 1079, 28.6875: 1280, 30.375: 1506, 32.0625: 1818, 33.75: 2218, 35.4375: 2606, 37.125: 3013, 38.8125: 3545} [(23.625, 995), (25.3125, 1002), (21.9375, 1025), (20.25, 1041), (18.5625, 1073), (27.0, 1079), (16.875, 1129), (15.1875, 1190), (28.6875, 1280), (13.5, 1299), (30.375, 1506), (32.0625, 1818), (33.75, 2218), (35.4375, 2606), (37.125, 3013), (38.8125, 3545)] arg_min reach at : 23.625 with value = 995 | arg_min : 23.625 min_score : 995{13.5: 1907, 15.1875: 1827, 16.875: 1734, 18.5625: 1635, 20.25: 1520, 21.9375: 1400, 23.625: 1268, 25.3125: 1139, 27.0: 995, 28.6875: 1081, 30.375: 1376, 32.0625: 1837, 33.75: 2365, 35.4375: 3034, 37.125: 3644, 38.8125: 4482} [(27.0, 995), (28.6875, 1081), (25.3125, 1139), (23.625, 1268), (30.375, 1376), (21.9375, 1400), (20.25, 1520), (18.5625, 1635), (16.875, 1734), (15.1875, 1827), (32.0625, 1837), (13.5, 1907), (33.75, 2365), (35.4375, 3034), (37.125, 3644), (38.8125, 4482)] arg_min reach at : 27.0 with value = 995 | arg_min : 27.0 min_score : 995{0.5: 16358, 0.55: 14291, 0.6000000000000001: 12544, 0.65: 11117, 0.7000000000000001: 9800, 0.75: 8586, 0.8: 7533, 0.8500000000000001: 6488, 0.9: 5599, 0.9500000000000001: 4779, 1.0: 4075, 1.05: 3520, 1.1: 3021, 1.1500000000000001: 2582, 1.2000000000000002: 2150, 1.25: 1820, 1.3: 1534, 1.35: 1347, 1.4000000000000001: 1219, 1.4500000000000002: 1130, 1.5: 1073, 1.55: 1037, 1.6: 998, 1.6500000000000001: 961, 1.7000000000000002: 981, 1.75: 984, 1.8: 995, 1.85: 995, 1.9000000000000001: 1023, 1.9500000000000002: 1058} [(1.6500000000000001, 961), (1.7000000000000002, 981), (1.75, 984), (1.8, 995), (1.85, 995), (1.6, 998), (1.9000000000000001, 1023), (1.55, 1037), (1.9500000000000002, 1058), (1.5, 1073), (1.4500000000000002, 1130), (1.4000000000000001, 1219), (1.35, 1347), (1.3, 1534), (1.25, 1820), (1.2000000000000002, 2150), (1.1500000000000001, 2582), (1.1, 3021), (1.05, 3520), (1.0, 4075), (0.9500000000000001, 4779), (0.9, 5599), (0.8500000000000001, 6488), (0.8, 7533), (0.75, 8586), (0.7000000000000001, 9800), (0.65, 11117), (0.6000000000000001, 12544), (0.55, 14291), (0.5, 16358)] arg_min reach at : 1.6500000000000001 with value = 961 arg_min : 1.6500000000000001 min_score : 961{-90.0: 906, -85.0: 871, -80.0: 858, -75.0: 866, -70.0: 852, -65.0: 864, -60.0: 855, -55.0: 853, -50.0: 961, -45.0: 1180, -40.0: 1412, -35.0: 1690, -30.0: 1959, -25.0: 2243, -20.0: 2484, -15.0: 2677, -10.0: 2820, -5.0: 2976, 0.0: 3039, 5.0: 3049, 10.0: 3098, 15.0: 3122, 20.0: 3061, 25.0: 2954, 30.0: 2845, 35.0: 2631, 40.0: 2417, 45.0: 2130, 50.0: 1863, 55.0: 1564, 60.0: 1439, 65.0: 1326, 70.0: 1220, 75.0: 1142, 80.0: 1056, 85.0: 986} [(-70.0, 852), (-55.0, 853), (-60.0, 855), (-80.0, 858), (-65.0, 864), (-75.0, 866), (-85.0, 871), (-90.0, 906), (-50.0, 961), (85.0, 986), (80.0, 1056), (75.0, 1142), (-45.0, 1180), (70.0, 1220), (65.0, 1326), (-40.0, 1412), (60.0, 1439), (55.0, 1564), (-35.0, 1690), (50.0, 1863), (-30.0, 1959), (45.0, 2130), (-25.0, 2243), (40.0, 2417), (-20.0, 2484), (35.0, 2631), (-15.0, 2677), (-10.0, 2820), (30.0, 2845), (25.0, 2954), (-5.0, 2976), (0.0, 3039), (5.0, 3049), (20.0, 3061), (10.0, 3098), (15.0, 3122)] arg_min reach at : -70.0 with value = 852 arg_min : -70.0 min_score : 852{12.75: 4966, 14.34375: 4695, 15.9375: 4364, 17.53125: 3945, 19.125: 3465, 20.71875: 2958, 22.3125: 2415, 23.90625: 1841, 25.5: 1295, 27.09375: 936, 28.6875: 847, 30.28125: 852, 31.875: 883, 33.46875: 997, 35.0625: 1322, 36.65625: 1859} [(28.6875, 847), (30.28125, 852), (31.875, 883), (27.09375, 936), (33.46875, 997), (25.5, 1295), (35.0625, 1322), (23.90625, 1841), (36.65625, 1859), (22.3125, 2415), (20.71875, 2958), (19.125, 3465), (17.53125, 3945), (15.9375, 4364), (14.34375, 4695), (12.75, 4966)] arg_min reach at : 28.6875 with value = 847 arg_min : 28.6875 min_score : 847{13.5: 1489, 15.1875: 1374, 16.875: 1244, 18.5625: 1126, 20.25: 996, 21.9375: 893, 23.625: 847, 25.3125: 933, 27.0: 1153, 28.6875: 1446, 30.375: 1827, 32.0625: 2217, 33.75: 2658, 35.4375: 3137, 37.125: 3653, 38.8125: 4142} [(23.625, 847), (21.9375, 893), (25.3125, 933), (20.25, 996), (18.5625, 1126), (27.0, 1153), (16.875, 1244), (15.1875, 1374), (28.6875, 1446), (13.5, 1489), (30.375, 1827), (32.0625, 2217), (33.75, 2658), (35.4375, 3137), (37.125, 3653), (38.8125, 4142)] arg_min reach at : 23.625 with value = 847 arg_min : 23.625 min_score : 847{13.5: 1870, 15.1875: 1783, 16.875: 1676, 18.5625: 1564, 20.25: 1442, 21.9375: 1302, 23.625: 1153, 25.3125: 997, 27.0: 847, 28.6875: 864, 30.375: 1163, 32.0625: 1613, 33.75: 2300, 35.4375: 3191, 37.125: 4251, 38.8125: 5356} [(27.0, 847), (28.6875, 864), (25.3125, 997), (23.625, 1153), (30.375, 1163), (21.9375, 1302), (20.25, 1442), (18.5625, 1564), (32.0625, 1613), (16.875, 1676), (15.1875, 1783), (13.5, 1870), (33.75, 2300), (35.4375, 3191), (37.125, 4251), (38.8125, 5356)] arg_min reach at : 27.0 with value = 847 arg_min : 27.0 min_score : 847 yc : 23.625 xc : 28.6875 angle : -70.0 radius : 27.0 excentricity : 1.6500000000000001 yc : 23.625 xc : 28.6875 angle : -70.0 radius : 27.0 excentricity : 1.6500000000000001 x0 : 103 y1 : 396 width : 35, height : 38, area : 1330, score : 1.0 x0 : 93 y1 : 396 width : 51, height : 36, area : 1836, score : 1.0 Now saving polygons points : 1| batch 1 Loaded 3 chid ids of type : 520 ++ CHI and polygons saved ! (57, 52) (52, 43) [57, 52] (57, 52) score : 7970 strategy_opt : 5| strategy_opt : [('excentricity', [0.5, 2.0, 0.05]), ('angle', [-90.0, 90.0, 5.0]), ('xc', [10.75, 32.25, 1.34375]), ('yc', [13.0, 39.0, 1.625]), ('radius', [13.0, 39.0, 1.625])] {0.5: 16167, 0.55: 14430, 0.6000000000000001: 12864, 0.65: 11529, 0.7000000000000001: 10346, 0.75: 9256, 0.8: 8418, 0.8500000000000001: 7609, 0.9: 6862, 0.9500000000000001: 6277, 1.0: 5482, 1.05: 4840, 1.1: 4264, 1.1500000000000001: 3773, 1.2000000000000002: 3275, 1.25: 2948, 1.3: 2727, 1.35: 2514, 1.4000000000000001: 2254, 1.4500000000000002: 2155, 1.5: 1986, 1.55: 1892, 1.6: 1846, 1.6500000000000001: 1782, 1.7000000000000002: 1717, 1.75: 1656, 1.8: 1641, 1.85: 1602, 1.9000000000000001: 1587, 1.9500000000000002: 1576} [(1.9500000000000002, 1576), (1.9000000000000001, 1587), (1.85, 1602), (1.8, 1641), (1.75, 1656), (1.7000000000000002, 1717), (1.6500000000000001, 1782), (1.6, 1846), (1.55, 1892), (1.5, 1986), (1.4500000000000002, 2155), (1.4000000000000001, 2254), (1.35, 2514), (1.3, 2727), (1.25, 2948), (1.2000000000000002, 3275), (1.1500000000000001, 3773), (1.1, 4264), (1.05, 4840), (1.0, 5482), (0.9500000000000001, 6277), (0.9, 6862), (0.8500000000000001, 7609), (0.8, 8418), (0.75, 9256), (0.7000000000000001, 10346), (0.65, 11529), (0.6000000000000001, 12864), (0.55, 14430), (0.5, 16167)] arg_min reach at : 1.9500000000000002 with value = 1576 | arg_min : 1.9500000000000002 min_score : 1576{-90.0: 2291, -85.0: 2369, -80.0: 2390, -75.0: 2401, -70.0: 2341, -65.0: 2390, -60.0: 2418, -55.0: 2445, -50.0: 2599, -45.0: 2701, -40.0: 2755, -35.0: 2787, -30.0: 2753, -25.0: 2678, -20.0: 2611, -15.0: 2478, -10.0: 2313, -5.0: 2098, 0.0: 1870, 5.0: 1702, 10.0: 1576, 15.0: 1422, 20.0: 1258, 25.0: 1061, 30.0: 872, 35.0: 752, 40.0: 632, 45.0: 677, 50.0: 872, 55.0: 1070, 60.0: 1274, 65.0: 1477, 70.0: 1637, 75.0: 1895, 80.0: 2049, 85.0: 2193} [(40.0, 632), (45.0, 677), (35.0, 752), (30.0, 872), (50.0, 872), (25.0, 1061), (55.0, 1070), (20.0, 1258), (60.0, 1274), (15.0, 1422), (65.0, 1477), (10.0, 1576), (70.0, 1637), (5.0, 1702), (0.0, 1870), (75.0, 1895), (80.0, 2049), (-5.0, 2098), (85.0, 2193), (-90.0, 2291), (-10.0, 2313), (-70.0, 2341), (-85.0, 2369), (-80.0, 2390), (-65.0, 2390), (-75.0, 2401), (-60.0, 2418), (-55.0, 2445), (-15.0, 2478), (-50.0, 2599), (-20.0, 2611), (-25.0, 2678), (-45.0, 2701), (-30.0, 2753), (-40.0, 2755), (-35.0, 2787)] arg_min reach at : 40.0 with value = 632 | arg_min : 40.0 min_score : 632{10.75: 982, 12.09375: 841, 13.4375: 751, 14.78125: 672, 16.125: 624, 17.46875: 599, 18.8125: 575, 20.15625: 561, 21.5: 632, 22.84375: 895, 24.1875: 1257, 25.53125: 1708, 26.875: 2192, 28.21875: 2686, 29.5625: 3202, 30.90625: 3701} [(20.15625, 561), (18.8125, 575), (17.46875, 599), (16.125, 624), (21.5, 632), (14.78125, 672), (13.4375, 751), (12.09375, 841), (22.84375, 895), (10.75, 982), (24.1875, 1257), (25.53125, 1708), (26.875, 2192), (28.21875, 2686), (29.5625, 3202), (30.90625, 3701)] arg_min reach at : 20.15625 with value = 561 | arg_min : 20.15625 min_score : 561{13.0: 3254, 14.625: 2871, 16.25: 2436, 17.875: 1977, 19.5: 1474, 21.125: 1071, 22.75: 835, 24.375: 629, 26.0: 561, 27.625: 716, 29.25: 937, 30.875: 1286, 32.5: 1738, 34.125: 2226, 35.75: 2782, 37.375: 3291} [(26.0, 561), (24.375, 629), (27.625, 716), (22.75, 835), (29.25, 937), (21.125, 1071), (30.875, 1286), (19.5, 1474), (32.5, 1738), (17.875, 1977), (34.125, 2226), (16.25, 2436), (35.75, 2782), (14.625, 2871), (13.0, 3254), (37.375, 3291)] arg_min reach at : 26.0 with value = 561 | arg_min : 26.0 min_score : 561{13.0: 1371, 14.625: 1301, 16.25: 1217, 17.875: 1127, 19.5: 1028, 21.125: 926, 22.75: 811, 24.375: 683, 26.0: 561, 27.625: 624, 29.25: 1026, 30.875: 1508, 32.5: 2147, 34.125: 2981, 35.75: 3949, 37.375: 5069} [(26.0, 561), (27.625, 624), (24.375, 683), (22.75, 811), (21.125, 926), (29.25, 1026), (19.5, 1028), (17.875, 1127), (16.25, 1217), (14.625, 1301), (13.0, 1371), (30.875, 1508), (32.5, 2147), (34.125, 2981), (35.75, 3949), (37.375, 5069)] arg_min reach at : 26.0 with value = 561 | arg_min : 26.0 min_score : 561{0.5: 13327, 0.55: 12420, 0.6000000000000001: 11540, 0.65: 10560, 0.7000000000000001: 9652, 0.75: 8782, 0.8: 7834, 0.8500000000000001: 7061, 0.9: 6294, 0.9500000000000001: 5633, 1.0: 4916, 1.05: 4424, 1.1: 3865, 1.1500000000000001: 3323, 1.2000000000000002: 2817, 1.25: 2401, 1.3: 1935, 1.35: 1600, 1.4000000000000001: 1218, 1.4500000000000002: 1019, 1.5: 834, 1.55: 693, 1.6: 614, 1.6500000000000001: 568, 1.7000000000000002: 549, 1.75: 529, 1.8: 520, 1.85: 521, 1.9000000000000001: 534, 1.9500000000000002: 561} [(1.8, 520), (1.85, 521), (1.75, 529), (1.9000000000000001, 534), (1.7000000000000002, 549), (1.9500000000000002, 561), (1.6500000000000001, 568), (1.6, 614), (1.55, 693), (1.5, 834), (1.4500000000000002, 1019), (1.4000000000000001, 1218), (1.35, 1600), (1.3, 1935), (1.25, 2401), (1.2000000000000002, 2817), (1.1500000000000001, 3323), (1.1, 3865), (1.05, 4424), (1.0, 4916), (0.9500000000000001, 5633), (0.9, 6294), (0.8500000000000001, 7061), (0.8, 7834), (0.75, 8782), (0.7000000000000001, 9652), (0.65, 10560), (0.6000000000000001, 11540), (0.55, 12420), (0.5, 13327)] arg_min reach at : 1.8 with value = 520 arg_min : 1.8 min_score : 520{-90.0: 2114, -85.0: 2268, -80.0: 2343, -75.0: 2405, -70.0: 2409, -65.0: 2435, -60.0: 2467, -55.0: 2514, -50.0: 2590, -45.0: 2654, -40.0: 2632, -35.0: 2610, -30.0: 2565, -25.0: 2526, -20.0: 2435, -15.0: 2330, -10.0: 2181, -5.0: 2013, 0.0: 1792, 5.0: 1584, 10.0: 1389, 15.0: 1208, 20.0: 1038, 25.0: 843, 30.0: 673, 35.0: 564, 40.0: 520, 45.0: 597, 50.0: 764, 55.0: 963, 60.0: 1158, 65.0: 1390, 70.0: 1573, 75.0: 1800, 80.0: 1969, 85.0: 2070} [(40.0, 520), (35.0, 564), (45.0, 597), (30.0, 673), (50.0, 764), (25.0, 843), (55.0, 963), (20.0, 1038), (60.0, 1158), (15.0, 1208), (10.0, 1389), (65.0, 1390), (70.0, 1573), (5.0, 1584), (0.0, 1792), (75.0, 1800), (80.0, 1969), (-5.0, 2013), (85.0, 2070), (-90.0, 2114), (-10.0, 2181), (-85.0, 2268), (-15.0, 2330), (-80.0, 2343), (-75.0, 2405), (-70.0, 2409), (-65.0, 2435), (-20.0, 2435), (-60.0, 2467), (-55.0, 2514), (-25.0, 2526), (-30.0, 2565), (-50.0, 2590), (-35.0, 2610), (-40.0, 2632), (-45.0, 2654)] arg_min reach at : 40.0 with value = 520 arg_min : 40.0 min_score : 520{10.75: 1004, 12.09375: 902, 13.4375: 775, 14.78125: 655, 16.125: 580, 17.46875: 515, 18.8125: 494, 20.15625: 520, 21.5: 692, 22.84375: 1028, 24.1875: 1466, 25.53125: 2001, 26.875: 2490, 28.21875: 3027, 29.5625: 3629, 30.90625: 4163} [(18.8125, 494), (17.46875, 515), (20.15625, 520), (16.125, 580), (14.78125, 655), (21.5, 692), (13.4375, 775), (12.09375, 902), (10.75, 1004), (22.84375, 1028), (24.1875, 1466), (25.53125, 2001), (26.875, 2490), (28.21875, 3027), (29.5625, 3629), (30.90625, 4163)] arg_min reach at : 18.8125 with value = 494 arg_min : 18.8125 min_score : 494{13.0: 3121, 14.625: 2763, 16.25: 2397, 17.875: 1962, 19.5: 1409, 21.125: 1056, 22.75: 809, 24.375: 589, 26.0: 494, 27.625: 635, 29.25: 987, 30.875: 1406, 32.5: 1937, 34.125: 2442, 35.75: 2987, 37.375: 3537} [(26.0, 494), (24.375, 589), (27.625, 635), (22.75, 809), (29.25, 987), (21.125, 1056), (30.875, 1406), (19.5, 1409), (32.5, 1937), (17.875, 1962), (16.25, 2397), (34.125, 2442), (14.625, 2763), (35.75, 2987), (13.0, 3121), (37.375, 3537)] arg_min reach at : 26.0 with value = 494 arg_min : 26.0 min_score : 494{13.0: 1347, 14.625: 1266, 16.25: 1182, 17.875: 1086, 19.5: 979, 21.125: 865, 22.75: 737, 24.375: 616, 26.0: 494, 27.625: 559, 29.25: 954, 30.875: 1524, 32.5: 2347, 34.125: 3350, 35.75: 4468, 37.375: 5635} [(26.0, 494), (27.625, 559), (24.375, 616), (22.75, 737), (21.125, 865), (29.25, 954), (19.5, 979), (17.875, 1086), (16.25, 1182), (14.625, 1266), (13.0, 1347), (30.875, 1524), (32.5, 2347), (34.125, 3350), (35.75, 4468), (37.375, 5635)] arg_min reach at : 26.0 with value = 494 arg_min : 26.0 min_score : 494 yc : 26.0 xc : 18.8125 angle : 40.0 radius : 26.0 excentricity : 1.8 yc : 26.0 xc : 18.8125 angle : 40.0 radius : 26.0 excentricity : 1.8 x0 : 104 y1 : 292 width : 27, height : 36, area : 972, score : 1.0 x0 : 102 y1 : 288 width : 39, height : 43, area : 1677, score : 1.0 Now saving polygons points : 1| batch 1 Loaded 4 chid ids of type : 520 +++ CHI and polygons saved ! ['temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0_varroa_with_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0_varroa_with_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0_varroa_with_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0_ellipsebest.jpg', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0_varroa_with_ellipsebest.jpg'] About to upload 8 photos https://marlene.fotonower.com/api/v1/secured/portfolio/new?access_token=78d09a0790ec6ecbf119343125a81fdc upload in portfolio : 23444766 Result OK ! uploaded one batch 0 Elapsed time : 20.149830102920532 After datou_step_exec type output : time spend for datou_step_exec : 24.967286348342896 time spend to save output : 2.8848648071289062e-05 total time spend for step 1 : 24.967315196990967 step2:tile Fri May 30 03:38:22 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : ['temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg'] We expect there is only one output and this part is used while all output are not tuple or array We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : [('temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg',)] After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg': 937852786, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg': 1361688560, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg': 1361688561, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg': 1361688562, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg': 1361688563} map_photo_id_path_extension : {937852786: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg', 'extension': 'jpg'}, 1361688560: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg'}, 1361688561: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg'}, 1361688562: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg'}, 1361688563: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg'}} map_subphoto_mainphoto : {1361688560: 937852786, 1361688561: 937852786, 1361688562: 937852786, 1361688563: 937852786} verbose : True param_json : {'photo_tile_type': 17, 'whiten': True, 'remove_crop_border': True, 'minimal_size_crop_border': 900, 'stride': 240, 'crop_hashtag_type_tiled': 521, 'ETA': 86400, 'new_width': 480, 'new_height': 480, 'token': '78d09a0790ec6ecbf119343125a81fdc', 'portfolio_name': 'tile_taggage_varroa', 'crop_hashtag_type': 520, 'host': 'www.fotonower.com', 'arg_aux_upload': {'type_upload': 'python'}} type(crop_hashtag_type) : type(crop_hashtag_type_tiled) : We consider crop_hashtag_type is an integer ! map_chi_type_to_chi_type_cropped : {520: 521} TO DEPRECATE VR 14-6-18 map_filenames : {937852786: 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg'} list_pids : 1 list_pids : 2 list_subpids to replace list_pids : 0 batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 937852786,937852786) and `type` in (520) Loaded 4 chid ids of type : 520 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (3815451602,3815451603,3815451604,3815451821) ++++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (3815451602,3815451603,3815451604,3815451821) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (3815451602,3815451603,3815451604,3815451821) https://marlene.fotonower.com/api/v1/secured/portfolio/new?name=tile_taggage_varroa&access_token=78d09a0790ec6ecbf119343125a81fdc created feed_id_new_photos : 23444767 with name tile_taggage_varroa feed_id_new_photos : 23444767 filename : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg photo_id : 937852786 height_image_input : 480 width_image_input : 480 new_width : 480 new_height : 480 stride : 240 stride_relative : 0.1 chi to copy from the main photo to the tiled photo input_chi_for_this_image_as_chi : 4 list_bib_to_crops : 1 [(0, 480, 0, 480, 0)] calcul des nouveaux crops pour le tile x0:0,x1:480,y0:0,y1:480 calcul avec la methode originale calcul avec la methode originale calcul avec la methode originale calcul avec la methode originale chi selectionnes : [, , , ] new_crops_tiles : 1 crop_transformed : 4 insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) [(937852786, 2090988864, 17, 0, 480, 0, 480, 1.0)] list_photo_ids_cropped : [937852786] batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 937852786) and `type` in (17) Loaded 1 chid ids of type : 17 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (8165084) SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (8165084) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (8165084) treat the image : temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg , 0 before upload mediasElapsed time : 0.01006007194519043 on upload les photos avec python init cache_photo without model_param we have 1 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569109_1112947 INSERT ignore into MTRUser.mtr_portfolio_photos (`mtr_portfolio_id`, `mtr_photo_id`, `mtr_user_id`, `created_at`) VALUES (23444767, 1361688576, 0, NOW()) 1 we have uploaded 1 photos in the portfolio 23444767 Importing ! upload mediasElapsed time : 0.6165158748626709 , 0insert ignore into MTRPhoto.crop_sub_photo_ids (crop_hashtag_id, sub_photo_id, mtr_user_id) VALUES (%s,%s,%s) : [(8165084, 1361688576, 0)] Saving 4 CHIs. list_chi_tile : [": {'photo_id': 1361688576, 'hashtag_id': 2087736828, 'type': 521, 'x0': 432, 'x1': 467, 'y0': 303, 'y1': 355, 'score': 1.0, 'id': 0, 'points': ['463,352,452,353,439,350,426,342,418,333,417,325,422,319,433,318,446,321,459,328,467,338,469,346', '463,352,452,353,439,350,426,342,418,333,417,325,422,319,433,318,446,321,459,328,467,338,469,346'], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', 'sum_segment': 0}", ": {'photo_id': 1361688576, 'hashtag_id': 2087736828, 'type': 521, 'x0': 419, 'x1': 445, 'y0': 434, 'y1': 480, 'score': 1.0, 'id': 0, 'points': ['451,461,449,467,441,473,429,477,416,477,406,473,402,467,403,461,411,455,423,451,435,452,446,455', '451,461,449,467,441,473,429,477,416,477,406,473,402,467,403,461,411,455,423,451,435,452,446,455'], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', 'sum_segment': 0}", ": {'photo_id': 1361688576, 'hashtag_id': 2087736828, 'type': 521, 'x0': 93, 'x1': 144, 'y0': 360, 'y1': 396, 'score': 1.0, 'id': 0, 'points': ['112,359,120,350,130,348,137,351,141,361,140,374,134,387,126,396,117,399,109,395,105,385,106,372', '112,359,120,350,130,348,137,351,141,361,140,374,134,387,126,396,117,399,109,395,105,385,106,372'], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', 'sum_segment': 0}", ": {'photo_id': 1361688576, 'hashtag_id': 2087736828, 'type': 521, 'x0': 102, 'x1': 141, 'y0': 245, 'y1': 288, 'score': 1.0, 'id': 0, 'points': ['124,293,112,290,102,281,95,271,93,262,96,255,105,254,116,257,127,266,134,276,136,285,132,292', '124,293,112,290,102,281,95,271,93,262,96,255,105,254,116,257,127,266,134,276,136,285,132,292'], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', 'sum_segment': 0}"] insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) batch 1 Loaded 4 chid ids of type : 521 INSERT IGNORE INTO MTRPhoto.crop_polygon_points (`crop_hashtag_id`, `points`) VALUES (%s, %s) Number RLEs to save : 0 INSERT IGNORE INTO MTRPhoto.crop_sum_segments (`crop_hashtag_id`, `sum_segments`) VALUES (%s, %s) TO DO : save crop sub photo not yet done ! end of tileElapsed time : 0.691401481628418 map_pid_results : {'1361688576': ['temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg']} After datou_step_exec type output : time spend for datou_step_exec : 7.487670660018921 time spend to save output : 9.5367431640625e-05 total time spend for step 2 : 7.4877660274505615 step3:rotate Fri May 30 03:38:29 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'1361688576': ['temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg']} input_args_next_step : {'1361688576': ()} output_args : {'1361688576': ['temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg']} args : 1361688576 depend.output_id : 0 We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ('temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg',) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg': 937852786, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg': 1361688560, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg': 1361688561, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg': 1361688562, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg': 1361688563, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg': 1361688576} map_photo_id_path_extension : {937852786: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67.jpg', 'extension': 'jpg'}, 1361688560: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165075_0.jpg'}, 1361688561: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165076_0.jpg'}, 1361688562: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165077_0.jpg'}, 1361688563: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_bib_crop_8165078_0.jpg'}, 1361688576: {'path': 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg'}} map_subphoto_mainphoto : {1361688560: 937852786, 1361688561: 937852786, 1361688562: 937852786, 1361688563: 937852786, 1361688576: 937852786} Beginning of datou_step_rotate ! Warning, new_feed_id is empty ! We are in a datou with depends ! rotate photos of 0,15,30,45,60,75,90,105,120,135,150,165,180,195,210,225,240,255,270,285,300,315,330,345 degres batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 1361688576) and `type` in (521) Loaded 4 chid ids of type : 521 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (3815453402,3815453403,3815453401,3815453400) ++WARNING : duplicated polygon, we should remove this data for chi_id : 3815453400. Ignored now ++WARNING : duplicated polygon, we should remove this data for chi_id : 3815453401. Ignored now ++WARNING : duplicated polygon, we should remove this data for chi_id : 3815453402. Ignored now ++WARNING : duplicated polygon, we should remove this data for chi_id : 3815453403. Ignored now SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (3815453402,3815453403,3815453401,3815453400) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (3815453402,3815453403,3815453401,3815453400) map_chi : {1361688576: [, , , ]} https://marlene.fotonower.com/api/v1/secured/portfolio/new?name=rotate_data_augmentation_varroa_480_ellipse_320&access_token=78d09a0790ec6ecbf119343125a81fdc feed_id_new_photos : 23444768 photo_id in download_rotate_and_save : 1361688576 list_chi_loc : 4 Use all angle ! Rotation of photo 1361688576 of 0 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 0 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 1. 0.] [-0. 1.]] 0 [[ 1. 0.] [-0. 1.]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00038313865661621094 nb_pixel_total : 1389 time to create 1 rle with old method : 0.0016710758209228516 .time for calcul the mask position with numpy : 0.0003685951232910156 nb_pixel_total : 1157 time to create 1 rle with old method : 0.0015528202056884766 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 2 list_crops_rotate : : {'photo_id': -1, 'hashtag_id': 2087736828, 'type': 529, 'x0': 25, 'x1': 61, 'y0': 268, 'y1': 319, 'score': 1.0, 'id': None, 'points': ['32,279,40,270,50,268,57,271,61,281,60,294,54,307,46,316,37,319,29,315,25,305,26,292'], 'sub_photo_id': 0, 'rles': [(-1, 48, 268, 4), (-1, 43, 269, 11), (-1, 40, 270, 16), (-1, 39, 271, 19), (-1, 38, 272, 20), (-1, 37, 273, 22), (-1, 36, 274, 23), (-1, 36, 275, 24), (-1, 35, 276, 25), (-1, 34, 277, 26), (-1, 33, 278, 28), (-1, 32, 279, 29), (-1, 32, 280, 30), (-1, 31, 281, 31), (-1, 31, 282, 31), (-1, 30, 283, 32), (-1, 30, 284, 32), (-1, 29, 285, 33), (-1, 29, 286, 33), (-1, 28, 287, 34), (-1, 28, 288, 33), (-1, 27, 289, 34), (-1, 27, 290, 34), (-1, 26, 291, 35), (-1, 26, 292, 35), (-1, 26, 293, 35), (-1, 26, 294, 35), (-1, 26, 295, 35), (-1, 26, 296, 34), (-1, 26, 297, 34), (-1, 26, 298, 33), (-1, 25, 299, 34), (-1, 25, 300, 33), (-1, 25, 301, 33), (-1, 25, 302, 32), (-1, 25, 303, 32), (-1, 25, 304, 31), (-1, 25, 305, 31), (-1, 25, 306, 30), (-1, 26, 307, 29), (-1, 26, 308, 28), (-1, 27, 309, 26), (-1, 27, 310, 25), (-1, 27, 311, 24), (-1, 28, 312, 23), (-1, 28, 313, 22), (-1, 29, 314, 20), (-1, 29, 315, 19), (-1, 31, 316, 16), (-1, 33, 317, 12), (-1, 35, 318, 7), (-1, 37, 319, 2)], 'hashtag': '', 'sum_segment': 0},: {'photo_id': -1, 'hashtag_id': 2087736828, 'type': 529, 'x0': 13, 'x1': 56, 'y0': 174, 'y1': 213, 'score': 1.0, 'id': None, 'points': ['44,213,32,210,22,201,15,191,13,182,16,175,25,174,36,177,47,186,54,196,56,205,52,212'], 'sub_photo_id': 0, 'rles': [(-1, 21, 174, 6), (-1, 16, 175, 15), (-1, 16, 176, 19), (-1, 15, 177, 22), (-1, 15, 178, 23), (-1, 14, 179, 26), (-1, 14, 180, 27), (-1, 13, 181, 29), (-1, 13, 182, 30), (-1, 13, 183, 31), (-1, 13, 184, 33), (-1, 14, 185, 33), (-1, 14, 186, 34), (-1, 14, 187, 35), (-1, 14, 188, 35), (-1, 15, 189, 35), (-1, 15, 190, 36), (-1, 15, 191, 36), (-1, 16, 192, 36), (-1, 16, 193, 37), (-1, 17, 194, 37), (-1, 18, 195, 36), (-1, 18, 196, 37), (-1, 19, 197, 36), (-1, 20, 198, 35), (-1, 21, 199, 35), (-1, 21, 200, 35), (-1, 22, 201, 34), (-1, 23, 202, 33), (-1, 24, 203, 33), (-1, 25, 204, 32), (-1, 26, 205, 31), (-1, 28, 206, 28), (-1, 29, 207, 27), (-1, 30, 208, 25), (-1, 31, 209, 24), (-1, 32, 210, 22), (-1, 35, 211, 19), (-1, 39, 212, 14), (-1, 43, 213, 6)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 15 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 15 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.96592583 0.25881905] [-0.25881905 0.96592583]] 15 [[ 0.96592583 0.25881905] [-0.25881905 0.96592583]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0004150867462158203 nb_pixel_total : 694 time to create 1 rle with old method : 0.0011675357818603516 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00042700767517089844 nb_pixel_total : 1162 time to create 1 rle with old method : 0.0023527145385742188 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -2, 'hashtag_id': 2087736828, 'type': 529, 'x0': 24, 'x1': 72, 'y0': 209, 'y1': 242, 'score': 1.0, 'id': None, 'points': ['62,241,50,242,38,236,28,228,24,220,25,212,34,209,45,209,58,215,67,222,72,231,70,238'], 'sub_photo_id': 0, 'rles': [(-1, 33, 209, 3), (-1, 37, 209, 9), (-1, 30, 210, 18), (-1, 49, 210, 1), (-1, 30, 211, 21), (-1, 26, 212, 27), (-1, 26, 213, 29), (-1, 26, 214, 32), (-1, 25, 215, 35), (-1, 25, 216, 36), (-1, 25, 217, 37), (-1, 25, 218, 38), (-1, 24, 219, 40), (-1, 25, 220, 40), (-1, 25, 221, 42), (-1, 25, 222, 43), (-1, 26, 223, 43), (-1, 27, 224, 42), (-1, 27, 225, 43), (-1, 28, 226, 43), (-1, 29, 227, 42), (-1, 29, 228, 42), (-1, 30, 229, 43), (-1, 30, 230, 43), (-1, 32, 231, 41), (-1, 33, 232, 39), (-1, 34, 233, 39), (-1, 36, 234, 36), (-1, 37, 235, 35), (-1, 38, 236, 34), (-1, 40, 237, 32), (-1, 42, 238, 30), (-1, 43, 239, 28), (-1, 46, 240, 22), (-1, 48, 241, 19), (-1, 50, 242, 15)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 30 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 30 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.8660254 0.5 ] [-0.5 0.8660254]] 30 [[ 0.8660254 0.5 ] [-0.5 0.8660254]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00035262107849121094 nb_pixel_total : 221 time to create 1 rle with old method : 0.000392913818359375 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00034427642822265625 nb_pixel_total : 1155 time to create 1 rle with old method : 0.001524209976196289 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -3, 'hashtag_id': 2087736828, 'type': 529, 'x0': 44, 'x1': 93, 'y0': 238, 'y1': 268, 'score': 1.0, 'id': None, 'points': ['86,264,74,268,61,265,50,260,44,253,43,245,50,240,61,237,75,239,87,245,93,251,93,259'], 'sub_photo_id': 0, 'rles': [(-1, 59, 238, 9), (-1, 55, 239, 17), (-1, 73, 239, 1), (-1, 51, 240, 27), (-1, 49, 241, 30), (-1, 48, 242, 34), (-1, 48, 243, 36), (-1, 46, 244, 40), (-1, 45, 245, 43), (-1, 44, 246, 45), (-1, 44, 247, 46), (-1, 44, 248, 47), (-1, 44, 249, 48), (-1, 44, 250, 49), (-1, 44, 251, 50), (-1, 44, 252, 50), (-1, 44, 253, 50), (-1, 45, 254, 49), (-1, 45, 255, 49), (-1, 47, 256, 47), (-1, 48, 257, 46), (-1, 48, 258, 46), (-1, 50, 259, 44), (-1, 51, 260, 43), (-1, 52, 261, 41), (-1, 54, 262, 37), (-1, 56, 263, 35), (-1, 59, 264, 31), (-1, 60, 265, 28), (-1, 63, 266, 21), (-1, 67, 267, 2), (-1, 70, 267, 11), (-1, 74, 268, 3)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 45 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 45 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.70710678 0.70710678] [-0.70710678 0.70710678]] 45 [[ 0.70710678 0.70710678] [-0.70710678 0.70710678]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00036334991455078125 nb_pixel_total : 143 time to create 1 rle with old method : 0.00025081634521484375 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003380775451660156 nb_pixel_total : 1161 time to create 1 rle with old method : 0.0015208721160888672 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -4, 'hashtag_id': 2087736828, 'type': 529, 'x0': 69, 'x1': 121, 'y0': 258, 'y1': 286, 'score': 1.0, 'id': None, 'points': ['115,279,105,285,91,286,79,284,72,279,69,272,74,265,84,259,98,258,110,260,118,265,120,273'], 'sub_photo_id': 0, 'rles': [(-1, 96, 258, 5), (-1, 88, 259, 17), (-1, 84, 260, 26), (-1, 111, 260, 1), (-1, 82, 261, 31), (-1, 81, 262, 34), (-1, 78, 263, 38), (-1, 77, 264, 42), (-1, 75, 265, 45), (-1, 74, 266, 46), (-1, 73, 267, 47), (-1, 72, 268, 48), (-1, 72, 269, 49), (-1, 71, 270, 50), (-1, 70, 271, 51), (-1, 69, 272, 53), (-1, 70, 273, 52), (-1, 70, 274, 51), (-1, 70, 275, 50), (-1, 71, 276, 48), (-1, 71, 277, 49), (-1, 72, 278, 47), (-1, 71, 279, 47), (-1, 72, 280, 45), (-1, 73, 281, 41), (-1, 76, 282, 37), (-1, 77, 283, 34), (-1, 79, 284, 31), (-1, 82, 285, 25), (-1, 85, 286, 21)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 60 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 60 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.5 0.8660254] [-0.8660254 0.5 ]] 60 [[ 0.5 0.8660254] [-0.8660254 0.5 ]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003705024719238281 nb_pixel_total : 414 time to create 1 rle with old method : 0.0006012916564941406 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00033974647521972656 nb_pixel_total : 1159 time to create 1 rle with old method : 0.0014903545379638672 . crop are not in the shrunk photo ! On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -5, 'hashtag_id': 2087736828, 'type': 529, 'x0': 102, 'x1': 152, 'y0': 269, 'y1': 300, 'score': 1.0, 'id': None, 'points': ['148,286,140,295,127,299,115,300,106,297,101,291,105,283,113,275,126,270,139,268,147,271,151,278'], 'sub_photo_id': 0, 'rles': [(-1, 132, 269, 1), (-1, 135, 269, 6), (-1, 126, 270, 18), (-1, 124, 271, 24), (-1, 121, 272, 28), (-1, 118, 273, 32), (-1, 116, 274, 34), (-1, 113, 275, 38), (-1, 112, 276, 39), (-1, 111, 277, 41), (-1, 111, 278, 42), (-1, 109, 279, 44), (-1, 108, 280, 44), (-1, 108, 281, 44), (-1, 106, 282, 45), (-1, 105, 283, 47), (-1, 105, 284, 46), (-1, 104, 285, 46), (-1, 104, 286, 46), (-1, 104, 287, 45), (-1, 104, 288, 44), (-1, 103, 289, 44), (-1, 103, 290, 43), (-1, 102, 291, 43), (-1, 102, 292, 42), (-1, 103, 293, 41), (-1, 104, 294, 38), (-1, 104, 295, 37), (-1, 105, 296, 33), (-1, 106, 297, 28), (-1, 107, 298, 27), (-1, 111, 299, 19), (-1, 112, 300, 1), (-1, 114, 300, 9)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 75 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 75 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.25881905 0.96592583] [-0.96592583 0.25881905]] 75 [[ 0.25881905 0.96592583] [-0.96592583 0.25881905]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003941059112548828 nb_pixel_total : 1204 time to create 1 rle with old method : 0.0015342235565185547 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003345012664794922 nb_pixel_total : 1157 time to create 1 rle with old method : 0.0015308856964111328 . crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.0003273487091064453 nb_pixel_total : 264 time to create 1 rle with old method : 0.00046062469482421875 On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -6, 'hashtag_id': 2087736828, 'type': 529, 'x0': 138, 'x1': 183, 'y0': 271, 'y1': 308, 'score': 1.0, 'id': None, 'points': ['182,285,176,295,164,303,153,307,144,306,138,302,139,293,145,283,156,275,168,270,177,271,183,277'], 'sub_photo_id': 0, 'rles': [(-1, 168, 271, 10), (-1, 165, 272, 14), (-1, 162, 273, 18), (-1, 160, 274, 22), (-1, 157, 275, 25), (-1, 155, 276, 28), (-1, 154, 277, 30), (-1, 153, 278, 31), (-1, 152, 279, 32), (-1, 150, 280, 33), (-1, 148, 281, 36), (-1, 148, 282, 36), (-1, 146, 283, 38), (-1, 145, 284, 38), (-1, 144, 285, 39), (-1, 144, 286, 39), (-1, 143, 287, 39), (-1, 143, 288, 38), (-1, 142, 289, 39), (-1, 141, 290, 40), (-1, 141, 291, 38), (-1, 140, 292, 39), (-1, 140, 293, 39), (-1, 139, 294, 39), (-1, 139, 295, 38), (-1, 139, 296, 38), (-1, 139, 297, 36), (-1, 139, 298, 35), (-1, 139, 299, 33), (-1, 139, 300, 32), (-1, 138, 301, 32), (-1, 138, 302, 30), (-1, 139, 303, 27), (-1, 140, 304, 25), (-1, 142, 305, 19), (-1, 143, 306, 17), (-1, 143, 307, 14), (-1, 150, 308, 1)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 90 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 90 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 6.123234e-17 1.000000e+00] [-1.000000e+00 6.123234e-17]] 90 [[ 6.123234e-17 1.000000e+00] [-1.000000e+00 6.123234e-17]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.000392913818359375 nb_pixel_total : 1389 time to create 1 rle with old method : 0.0017969608306884766 .time for calcul the mask position with numpy : 0.0003459453582763672 nb_pixel_total : 1157 time to create 1 rle with old method : 0.001497507095336914 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 2 list_crops_rotate : : {'photo_id': -7, 'hashtag_id': 2087736828, 'type': 529, 'x0': 268, 'x1': 319, 'y0': 258, 'y1': 294, 'score': 1.0, 'id': None, 'points': ['279,287,270,279,268,269,271,262,281,258,294,259,307,265,316,273,319,282,315,290,305,294,292,293'], 'sub_photo_id': 0, 'rles': [(-1, 280, 258, 8), (-1, 278, 259, 18), (-1, 275, 260, 23), (-1, 273, 261, 27), (-1, 271, 262, 31), (-1, 271, 263, 33), (-1, 270, 264, 36), (-1, 270, 265, 38), (-1, 269, 266, 40), (-1, 269, 267, 41), (-1, 268, 268, 43), (-1, 268, 269, 45), (-1, 268, 270, 46), (-1, 268, 271, 47), (-1, 269, 272, 47), (-1, 269, 273, 48), (-1, 269, 274, 48), (-1, 269, 275, 49), (-1, 269, 276, 49), (-1, 270, 277, 48), (-1, 270, 278, 49), (-1, 270, 279, 49), (-1, 271, 280, 48), (-1, 272, 281, 48), (-1, 273, 282, 47), (-1, 274, 283, 45), (-1, 276, 284, 43), (-1, 277, 285, 41), (-1, 278, 286, 40), (-1, 279, 287, 38), (-1, 281, 288, 36), (-1, 283, 289, 33), (-1, 285, 290, 31), (-1, 287, 291, 27), (-1, 289, 292, 23), (-1, 291, 293, 18), (-1, 299, 294, 8)], 'hashtag': '', 'sum_segment': 0},: {'photo_id': -7, 'hashtag_id': 2087736828, 'type': 529, 'x0': 174, 'x1': 213, 'y0': 263, 'y1': 306, 'score': 1.0, 'id': None, 'points': ['213,275,210,287,201,297,191,304,182,306,175,303,174,294,177,283,186,272,196,265,205,263,212,267'], 'sub_photo_id': 0, 'rles': [(-1, 203, 263, 3), (-1, 199, 264, 9), (-1, 196, 265, 14), (-1, 194, 266, 18), (-1, 193, 267, 20), (-1, 192, 268, 21), (-1, 190, 269, 23), (-1, 189, 270, 24), (-1, 187, 271, 27), (-1, 186, 272, 28), (-1, 185, 273, 29), (-1, 184, 274, 30), (-1, 184, 275, 30), (-1, 183, 276, 31), (-1, 182, 277, 31), (-1, 181, 278, 32), (-1, 180, 279, 33), (-1, 179, 280, 34), (-1, 179, 281, 33), (-1, 178, 282, 34), (-1, 177, 283, 35), (-1, 177, 284, 35), (-1, 176, 285, 35), (-1, 176, 286, 35), (-1, 176, 287, 35), (-1, 176, 288, 34), (-1, 175, 289, 34), (-1, 175, 290, 33), (-1, 175, 291, 32), (-1, 175, 292, 31), (-1, 174, 293, 32), (-1, 174, 294, 31), (-1, 174, 295, 30), (-1, 174, 296, 29), (-1, 174, 297, 28), (-1, 174, 298, 27), (-1, 175, 299, 24), (-1, 175, 300, 23), (-1, 175, 301, 22), (-1, 175, 302, 20), (-1, 175, 303, 19), (-1, 177, 304, 15), (-1, 179, 305, 10), (-1, 181, 306, 4)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 105 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 105 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.25881905 0.96592583] [-0.96592583 -0.25881905]] 105 [[-0.25881905 0.96592583] [-0.96592583 -0.25881905]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003731250762939453 nb_pixel_total : 694 time to create 1 rle with old method : 0.0009322166442871094 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003497600555419922 nb_pixel_total : 1162 time to create 1 rle with old method : 0.0014910697937011719 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -8, 'hashtag_id': 2087736828, 'type': 529, 'x0': 209, 'x1': 242, 'y0': 248, 'y1': 296, 'score': 1.0, 'id': None, 'points': ['241,257,242,269,236,281,228,291,220,295,212,294,209,285,209,274,215,261,222,252,231,247,238,249'], 'sub_photo_id': 0, 'rles': [(-1, 229, 248, 3), (-1, 233, 248, 1), (-1, 229, 249, 10), (-1, 226, 250, 14), (-1, 225, 251, 15), (-1, 223, 252, 17), (-1, 222, 253, 19), (-1, 221, 254, 21), (-1, 221, 255, 21), (-1, 220, 256, 23), (-1, 219, 257, 24), (-1, 218, 258, 25), (-1, 217, 259, 26), (-1, 216, 260, 27), (-1, 215, 261, 28), (-1, 215, 262, 28), (-1, 214, 263, 29), (-1, 214, 264, 29), (-1, 214, 265, 29), (-1, 213, 266, 30), (-1, 213, 267, 30), (-1, 212, 268, 31), (-1, 212, 269, 31), (-1, 211, 270, 32), (-1, 210, 271, 32), (-1, 211, 272, 31), (-1, 210, 273, 31), (-1, 210, 274, 31), (-1, 209, 275, 31), (-1, 209, 276, 31), (-1, 209, 277, 31), (-1, 209, 278, 30), (-1, 209, 279, 29), (-1, 209, 280, 29), (-1, 209, 281, 28), (-1, 209, 282, 28), (-1, 209, 283, 27), (-1, 210, 284, 25), (-1, 209, 285, 25), (-1, 209, 286, 25), (-1, 209, 287, 24), (-1, 210, 288, 22), (-1, 210, 289, 21), (-1, 210, 290, 21), (-1, 212, 291, 17), (-1, 212, 292, 15), (-1, 212, 293, 14), (-1, 212, 294, 12), (-1, 215, 295, 8), (-1, 219, 296, 1)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 120 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 120 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.5 0.8660254] [-0.8660254 -0.5 ]] 120 [[-0.5 0.8660254] [-0.8660254 -0.5 ]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00037288665771484375 nb_pixel_total : 221 time to create 1 rle with old method : 0.00043582916259765625 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00033974647521972656 nb_pixel_total : 1155 time to create 1 rle with old method : 0.0015149116516113281 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -9, 'hashtag_id': 2087736828, 'type': 529, 'x0': 238, 'x1': 268, 'y0': 227, 'y1': 276, 'score': 1.0, 'id': None, 'points': ['264,233,268,245,265,258,260,269,253,275,245,276,240,269,237,258,239,244,245,232,251,226,259,226'], 'sub_photo_id': 0, 'rles': [(-1, 251, 227, 10), (-1, 250, 228, 12), (-1, 249, 229, 13), (-1, 248, 230, 16), (-1, 247, 231, 18), (-1, 246, 232, 19), (-1, 245, 233, 21), (-1, 245, 234, 21), (-1, 244, 235, 22), (-1, 244, 236, 22), (-1, 243, 237, 24), (-1, 243, 238, 24), (-1, 242, 239, 25), (-1, 242, 240, 26), (-1, 242, 241, 26), (-1, 241, 242, 27), (-1, 240, 243, 28), (-1, 240, 244, 29), (-1, 240, 245, 29), (-1, 240, 246, 29), (-1, 239, 247, 29), (-1, 240, 248, 28), (-1, 239, 249, 29), (-1, 239, 250, 29), (-1, 239, 251, 28), (-1, 239, 252, 29), (-1, 238, 253, 30), (-1, 238, 254, 29), (-1, 238, 255, 29), (-1, 238, 256, 29), (-1, 238, 257, 29), (-1, 238, 258, 28), (-1, 238, 259, 28), (-1, 238, 260, 28), (-1, 238, 261, 27), (-1, 239, 262, 25), (-1, 239, 263, 25), (-1, 239, 264, 25), (-1, 239, 265, 24), (-1, 240, 266, 23), (-1, 240, 267, 22), (-1, 240, 268, 22), (-1, 240, 269, 21), (-1, 241, 270, 19), (-1, 241, 271, 18), (-1, 242, 272, 17), (-1, 244, 273, 13), (-1, 244, 274, 12), (-1, 245, 275, 11), (-1, 246, 276, 8)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 135 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 135 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.70710678 0.70710678] [-0.70710678 -0.70710678]] 135 [[-0.70710678 0.70710678] [-0.70710678 -0.70710678]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003647804260253906 nb_pixel_total : 143 time to create 1 rle with old method : 0.00032138824462890625 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003428459167480469 nb_pixel_total : 1160 time to create 1 rle with old method : 0.0014786720275878906 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -10, 'hashtag_id': 2087736828, 'type': 529, 'x0': 258, 'x1': 286, 'y0': 199, 'y1': 251, 'score': 1.0, 'id': None, 'points': ['279,204,285,214,286,228,284,240,279,247,272,250,265,245,259,235,258,221,260,209,265,201,273,199'], 'sub_photo_id': 0, 'rles': [(-1, 272, 199, 2), (-1, 269, 200, 6), (-1, 265, 201, 1), (-1, 267, 201, 9), (-1, 277, 201, 1), (-1, 264, 202, 15), (-1, 264, 203, 16), (-1, 264, 204, 17), (-1, 263, 205, 18), (-1, 262, 206, 19), (-1, 262, 207, 20), (-1, 261, 208, 22), (-1, 260, 209, 23), (-1, 261, 210, 23), (-1, 260, 211, 25), (-1, 260, 212, 25), (-1, 260, 213, 25), (-1, 260, 214, 26), (-1, 260, 215, 27), (-1, 259, 216, 28), (-1, 259, 217, 28), (-1, 259, 218, 28), (-1, 259, 219, 28), (-1, 258, 220, 29), (-1, 258, 221, 29), (-1, 258, 222, 29), (-1, 258, 223, 29), (-1, 258, 224, 29), (-1, 259, 225, 28), (-1, 259, 226, 28), (-1, 259, 227, 28), (-1, 259, 228, 28), (-1, 259, 229, 28), (-1, 259, 230, 28), (-1, 259, 231, 28), (-1, 259, 232, 28), (-1, 260, 233, 27), (-1, 260, 234, 27), (-1, 260, 235, 27), (-1, 260, 236, 26), (-1, 261, 237, 25), (-1, 261, 238, 25), (-1, 262, 239, 23), (-1, 263, 240, 22), (-1, 263, 241, 22), (-1, 263, 242, 21), (-1, 264, 243, 20), (-1, 265, 244, 18), (-1, 265, 245, 17), (-1, 266, 246, 16), (-1, 267, 247, 15), (-1, 268, 248, 13), (-1, 270, 249, 8), (-1, 279, 249, 1), (-1, 271, 250, 5), (-1, 272, 251, 1)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 150 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 150 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.8660254 0.5 ] [-0.5 -0.8660254]] 150 [[-0.8660254 0.5 ] [-0.5 -0.8660254]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00033354759216308594 nb_pixel_total : 414 time to create 1 rle with old method : 0.0006346702575683594 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003476142883300781 nb_pixel_total : 1159 time to create 1 rle with old method : 0.0014812946319580078 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.00032591819763183594 nb_pixel_total : 1 time to create 1 rle with old method : 1.9788742065429688e-05 len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -11, 'hashtag_id': 2087736828, 'type': 529, 'x0': 269, 'x1': 300, 'y0': 168, 'y1': 218, 'score': 1.0, 'id': None, 'points': ['286,171,295,179,299,192,300,204,297,213,291,218,283,214,275,206,270,193,268,180,271,172,278,168'], 'sub_photo_id': 0, 'rles': [(-1, 278, 168, 2), (-1, 277, 169, 5), (-1, 283, 169, 1), (-1, 275, 170, 10), (-1, 273, 171, 14), (-1, 272, 172, 16), (-1, 271, 173, 18), (-1, 271, 174, 19), (-1, 271, 175, 20), (-1, 271, 176, 21), (-1, 270, 177, 24), (-1, 270, 178, 24), (-1, 270, 179, 25), (-1, 269, 180, 27), (-1, 269, 181, 27), (-1, 269, 182, 27), (-1, 269, 183, 28), (-1, 269, 184, 28), (-1, 269, 185, 28), (-1, 270, 186, 27), (-1, 270, 187, 29), (-1, 269, 188, 30), (-1, 270, 189, 29), (-1, 270, 190, 29), (-1, 270, 191, 30), (-1, 270, 192, 30), (-1, 270, 193, 30), (-1, 270, 194, 30), (-1, 271, 195, 29), (-1, 271, 196, 29), (-1, 272, 197, 28), (-1, 272, 198, 29), (-1, 272, 199, 29), (-1, 273, 200, 28), (-1, 273, 201, 28), (-1, 273, 202, 28), (-1, 274, 203, 27), (-1, 274, 204, 27), (-1, 275, 205, 26), (-1, 275, 206, 26), (-1, 275, 207, 25), (-1, 276, 208, 25), (-1, 277, 209, 23), (-1, 279, 210, 20), (-1, 279, 211, 20), (-1, 280, 212, 19), (-1, 282, 213, 17), (-1, 282, 214, 16), (-1, 283, 215, 14), (-1, 285, 216, 11), (-1, 289, 217, 5), (-1, 291, 218, 2)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 165 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 165 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.96592583 0.25881905] [-0.25881905 -0.96592583]] 165 [[-0.96592583 0.25881905] [-0.25881905 -0.96592583]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00038313865661621094 nb_pixel_total : 1204 time to create 1 rle with old method : 0.0015146732330322266 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00033211708068847656 nb_pixel_total : 1158 time to create 1 rle with old method : 0.0014853477478027344 . crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.0003235340118408203 nb_pixel_total : 264 time to create 1 rle with old method : 0.00047326087951660156 On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -12, 'hashtag_id': 2087736828, 'type': 529, 'x0': 271, 'x1': 308, 'y0': 137, 'y1': 182, 'score': 1.0, 'id': None, 'points': ['285,137,295,143,303,155,307,166,306,175,302,181,293,180,283,174,275,163,270,151,271,142,277,136'], 'sub_photo_id': 0, 'rles': [(-1, 276, 137, 4), (-1, 281, 137, 3), (-1, 276, 138, 11), (-1, 274, 139, 14), (-1, 274, 140, 17), (-1, 273, 141, 18), (-1, 272, 142, 22), (-1, 271, 143, 24), (-1, 271, 144, 26), (-1, 271, 145, 26), (-1, 271, 146, 27), (-1, 271, 147, 28), (-1, 271, 148, 28), (-1, 271, 149, 29), (-1, 271, 150, 30), (-1, 271, 151, 31), (-1, 271, 152, 31), (-1, 272, 153, 31), (-1, 272, 154, 31), (-1, 272, 155, 32), (-1, 273, 156, 32), (-1, 273, 157, 32), (-1, 273, 158, 32), (-1, 274, 159, 31), (-1, 274, 160, 32), (-1, 275, 161, 32), (-1, 275, 162, 32), (-1, 275, 163, 32), (-1, 276, 164, 32), (-1, 276, 165, 32), (-1, 277, 166, 31), (-1, 278, 167, 30), (-1, 279, 168, 29), (-1, 280, 169, 28), (-1, 280, 170, 29), (-1, 281, 171, 27), (-1, 281, 172, 27), (-1, 283, 173, 25), (-1, 283, 174, 25), (-1, 284, 175, 24), (-1, 285, 176, 23), (-1, 287, 177, 21), (-1, 289, 178, 17), (-1, 290, 179, 15), (-1, 292, 180, 13), (-1, 294, 181, 10), (-1, 301, 182, 2)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 180 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 180 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-1.0000000e+00 1.2246468e-16] [-1.2246468e-16 -1.0000000e+00]] 180 [[-1.0000000e+00 1.2246468e-16] [-1.2246468e-16 -1.0000000e+00]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00038313865661621094 nb_pixel_total : 1389 time to create 1 rle with old method : 0.0017387866973876953 .time for calcul the mask position with numpy : 0.0003376007080078125 nb_pixel_total : 1157 time to create 1 rle with old method : 0.0015463829040527344 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 2 list_crops_rotate : : {'photo_id': -13, 'hashtag_id': 2087736828, 'type': 529, 'x0': 258, 'x1': 294, 'y0': 1, 'y1': 52, 'score': 1.0, 'id': None, 'points': ['287,41,279,50,269,52,262,49,258,39,259,26,265,13,273,4,282,1,290,5,294,15,293,28'], 'sub_photo_id': 0, 'rles': [(-1, 281, 1, 2), (-1, 278, 2, 7), (-1, 275, 3, 12), (-1, 273, 4, 16), (-1, 272, 5, 19), (-1, 271, 6, 20), (-1, 270, 7, 22), (-1, 269, 8, 23), (-1, 269, 9, 24), (-1, 268, 10, 25), (-1, 267, 11, 26), (-1, 266, 12, 28), (-1, 265, 13, 29), (-1, 265, 14, 30), (-1, 264, 15, 31), (-1, 264, 16, 31), (-1, 263, 17, 32), (-1, 263, 18, 32), (-1, 262, 19, 33), (-1, 262, 20, 33), (-1, 261, 21, 34), (-1, 261, 22, 33), (-1, 260, 23, 34), (-1, 260, 24, 34), (-1, 259, 25, 35), (-1, 259, 26, 35), (-1, 259, 27, 35), (-1, 259, 28, 35), (-1, 259, 29, 35), (-1, 259, 30, 34), (-1, 259, 31, 34), (-1, 259, 32, 33), (-1, 258, 33, 34), (-1, 258, 34, 33), (-1, 258, 35, 33), (-1, 258, 36, 32), (-1, 258, 37, 32), (-1, 258, 38, 31), (-1, 258, 39, 31), (-1, 258, 40, 30), (-1, 259, 41, 29), (-1, 259, 42, 28), (-1, 260, 43, 26), (-1, 260, 44, 25), (-1, 260, 45, 24), (-1, 261, 46, 23), (-1, 261, 47, 22), (-1, 262, 48, 20), (-1, 262, 49, 19), (-1, 264, 50, 16), (-1, 266, 51, 11), (-1, 268, 52, 4)], 'hashtag': '', 'sum_segment': 0},: {'photo_id': -13, 'hashtag_id': 2087736828, 'type': 529, 'x0': 263, 'x1': 306, 'y0': 107, 'y1': 146, 'score': 1.0, 'id': None, 'points': ['275,107,287,110,297,119,304,129,306,138,303,145,294,146,283,143,272,134,265,124,263,115,267,108'], 'sub_photo_id': 0, 'rles': [(-1, 271, 107, 6), (-1, 267, 108, 14), (-1, 266, 109, 19), (-1, 266, 110, 22), (-1, 265, 111, 24), (-1, 265, 112, 25), (-1, 264, 113, 27), (-1, 264, 114, 28), (-1, 263, 115, 31), (-1, 263, 116, 32), (-1, 263, 117, 33), (-1, 264, 118, 33), (-1, 264, 119, 34), (-1, 264, 120, 35), (-1, 264, 121, 35), (-1, 265, 122, 35), (-1, 265, 123, 36), (-1, 265, 124, 37), (-1, 266, 125, 36), (-1, 266, 126, 37), (-1, 267, 127, 37), (-1, 268, 128, 36), (-1, 269, 129, 36), (-1, 269, 130, 36), (-1, 270, 131, 35), (-1, 271, 132, 35), (-1, 271, 133, 35), (-1, 272, 134, 34), (-1, 273, 135, 33), (-1, 274, 136, 33), (-1, 276, 137, 31), (-1, 277, 138, 30), (-1, 278, 139, 29), (-1, 279, 140, 27), (-1, 280, 141, 26), (-1, 282, 142, 23), (-1, 283, 143, 22), (-1, 285, 144, 19), (-1, 289, 145, 15), (-1, 293, 146, 6)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 195 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 195 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.96592583 -0.25881905] [ 0.25881905 -0.96592583]] 195 [[-0.96592583 -0.25881905] [ 0.25881905 -0.96592583]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00038242340087890625 nb_pixel_total : 727 time to create 1 rle with old method : 0.0009789466857910156 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003426074981689453 nb_pixel_total : 1162 time to create 1 rle with old method : 0.001516103744506836 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -14, 'hashtag_id': 2087736828, 'type': 529, 'x0': 248, 'x1': 296, 'y0': 78, 'y1': 111, 'score': 1.0, 'id': None, 'points': ['257,78,269,77,281,83,291,91,295,99,294,107,285,110,274,110,261,104,252,97,247,88,249,81'], 'sub_photo_id': 0, 'rles': [(-1, 256, 78, 15), (-1, 254, 79, 19), (-1, 253, 80, 22), (-1, 250, 81, 28), (-1, 249, 82, 30), (-1, 249, 83, 32), (-1, 249, 84, 34), (-1, 249, 85, 35), (-1, 249, 86, 36), (-1, 248, 87, 39), (-1, 249, 88, 39), (-1, 248, 89, 41), (-1, 248, 90, 43), (-1, 248, 91, 43), (-1, 250, 92, 42), (-1, 250, 93, 42), (-1, 250, 94, 43), (-1, 251, 95, 43), (-1, 252, 96, 42), (-1, 252, 97, 43), (-1, 253, 98, 43), (-1, 254, 99, 42), (-1, 256, 100, 40), (-1, 257, 101, 40), (-1, 258, 102, 38), (-1, 259, 103, 37), (-1, 260, 104, 36), (-1, 261, 105, 35), (-1, 263, 106, 32), (-1, 266, 107, 29), (-1, 268, 108, 27), (-1, 270, 109, 21), (-1, 271, 110, 1), (-1, 273, 110, 18), (-1, 275, 111, 9), (-1, 285, 111, 3)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 210 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 210 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.8660254 -0.5 ] [ 0.5 -0.8660254]] 210 [[-0.8660254 -0.5 ] [ 0.5 -0.8660254]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003509521484375 nb_pixel_total : 250 time to create 1 rle with old method : 0.00042510032653808594 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00034308433532714844 nb_pixel_total : 1155 time to create 1 rle with old method : 0.0015227794647216797 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -15, 'hashtag_id': 2087736828, 'type': 529, 'x0': 227, 'x1': 276, 'y0': 52, 'y1': 82, 'score': 1.0, 'id': None, 'points': ['233,55,245,51,258,54,269,59,275,66,276,74,269,79,258,82,244,80,232,74,226,68,226,60'], 'sub_photo_id': 0, 'rles': [(-1, 244, 52, 3), (-1, 240, 53, 11), (-1, 252, 53, 2), (-1, 237, 54, 21), (-1, 233, 55, 28), (-1, 231, 56, 31), (-1, 230, 57, 35), (-1, 230, 58, 37), (-1, 228, 59, 41), (-1, 227, 60, 43), (-1, 227, 61, 44), (-1, 227, 62, 46), (-1, 227, 63, 46), (-1, 227, 64, 47), (-1, 227, 65, 49), (-1, 227, 66, 49), (-1, 227, 67, 50), (-1, 227, 68, 50), (-1, 227, 69, 50), (-1, 228, 70, 49), (-1, 229, 71, 48), (-1, 230, 72, 47), (-1, 231, 73, 46), (-1, 232, 74, 45), (-1, 233, 75, 43), (-1, 235, 76, 40), (-1, 237, 77, 36), (-1, 239, 78, 34), (-1, 242, 79, 30), (-1, 243, 80, 27), (-1, 247, 81, 1), (-1, 249, 81, 17), (-1, 253, 82, 9)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 225 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 225 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.70710678 -0.70710678] [ 0.70710678 -0.70710678]] 225 [[-0.70710678 -0.70710678] [ 0.70710678 -0.70710678]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00037217140197753906 nb_pixel_total : 169 time to create 1 rle with old method : 0.00032210350036621094 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003476142883300781 nb_pixel_total : 1161 time to create 1 rle with old method : 0.0015063285827636719 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -16, 'hashtag_id': 2087736828, 'type': 529, 'x0': 199, 'x1': 251, 'y0': 34, 'y1': 62, 'score': 1.0, 'id': None, 'points': ['204,40,214,34,228,33,240,35,247,40,250,47,245,54,235,60,221,61,209,59,201,54,199,46'], 'sub_photo_id': 0, 'rles': [(-1, 215, 34, 21), (-1, 214, 35, 25), (-1, 211, 36, 31), (-1, 210, 37, 34), (-1, 208, 38, 37), (-1, 207, 39, 41), (-1, 204, 40, 45), (-1, 203, 41, 47), (-1, 202, 42, 47), (-1, 201, 43, 49), (-1, 202, 44, 48), (-1, 201, 45, 50), (-1, 200, 46, 51), (-1, 199, 47, 52), (-1, 199, 48, 53), (-1, 200, 49, 51), (-1, 200, 50, 50), (-1, 200, 51, 49), (-1, 201, 52, 48), (-1, 201, 53, 47), (-1, 201, 54, 46), (-1, 201, 55, 45), (-1, 202, 56, 42), (-1, 205, 57, 38), (-1, 206, 58, 34), (-1, 208, 59, 31), (-1, 209, 60, 1), (-1, 211, 60, 26), (-1, 216, 61, 17), (-1, 220, 62, 5)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 240 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 240 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.5 -0.8660254] [ 0.8660254 -0.5 ]] 240 [[-0.5 -0.8660254] [ 0.8660254 -0.5 ]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00037479400634765625 nb_pixel_total : 450 time to create 1 rle with old method : 0.0006463527679443359 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003407001495361328 nb_pixel_total : 1159 time to create 1 rle with old method : 0.001491546630859375 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.0003402233123779297 nb_pixel_total : 1 time to create 1 rle with old method : 2.9087066650390625e-05 len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -17, 'hashtag_id': 2087736828, 'type': 529, 'x0': 168, 'x1': 218, 'y0': 20, 'y1': 51, 'score': 1.0, 'id': None, 'points': ['171,33,179,24,192,20,204,19,213,22,218,28,214,36,206,44,193,49,180,51,172,48,168,41'], 'sub_photo_id': 0, 'rles': [(-1, 198, 20, 9), (-1, 208, 20, 1), (-1, 191, 21, 19), (-1, 187, 22, 27), (-1, 187, 23, 28), (-1, 183, 24, 33), (-1, 180, 25, 37), (-1, 179, 26, 38), (-1, 177, 27, 41), (-1, 177, 28, 42), (-1, 176, 29, 43), (-1, 175, 30, 43), (-1, 174, 31, 44), (-1, 173, 32, 44), (-1, 172, 33, 45), (-1, 171, 34, 46), (-1, 171, 35, 46), (-1, 170, 36, 46), (-1, 169, 37, 47), (-1, 170, 38, 45), (-1, 169, 39, 44), (-1, 169, 40, 44), (-1, 168, 41, 44), (-1, 168, 42, 42), (-1, 169, 43, 41), (-1, 170, 44, 39), (-1, 170, 45, 38), (-1, 171, 46, 34), (-1, 171, 47, 32), (-1, 172, 48, 28), (-1, 173, 49, 24), (-1, 177, 50, 18), (-1, 180, 51, 6), (-1, 188, 51, 1)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 255 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 255 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-0.25881905 -0.96592583] [ 0.96592583 -0.25881905]] 255 [[-0.25881905 -0.96592583] [ 0.96592583 -0.25881905]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0004010200500488281 nb_pixel_total : 1237 time to create 1 rle with old method : 0.0015938282012939453 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00035762786865234375 nb_pixel_total : 1158 time to create 1 rle with old method : 0.0014638900756835938 . crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.00032782554626464844 nb_pixel_total : 234 time to create 1 rle with old method : 0.00040340423583984375 On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -18, 'hashtag_id': 2087736828, 'type': 529, 'x0': 137, 'x1': 182, 'y0': 12, 'y1': 49, 'score': 1.0, 'id': None, 'points': ['137,34,143,24,155,16,166,12,175,13,181,17,180,26,174,36,163,44,151,49,142,48,136,42'], 'sub_photo_id': 0, 'rles': [(-1, 170, 12, 1), (-1, 164, 13, 14), (-1, 161, 14, 17), (-1, 160, 15, 19), (-1, 156, 16, 25), (-1, 155, 17, 27), (-1, 153, 18, 30), (-1, 151, 19, 32), (-1, 150, 20, 32), (-1, 149, 21, 33), (-1, 147, 22, 35), (-1, 146, 23, 36), (-1, 144, 24, 38), (-1, 144, 25, 38), (-1, 143, 26, 39), (-1, 142, 27, 39), (-1, 142, 28, 39), (-1, 142, 29, 38), (-1, 140, 30, 40), (-1, 140, 31, 39), (-1, 140, 32, 38), (-1, 139, 33, 39), (-1, 138, 34, 39), (-1, 138, 35, 39), (-1, 138, 36, 38), (-1, 137, 37, 38), (-1, 137, 38, 36), (-1, 137, 39, 36), (-1, 138, 40, 33), (-1, 137, 41, 32), (-1, 137, 42, 31), (-1, 137, 43, 30), (-1, 137, 44, 29), (-1, 139, 45, 25), (-1, 139, 46, 22), (-1, 141, 47, 18), (-1, 142, 48, 14), (-1, 143, 49, 10)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 270 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 270 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[-1.8369702e-16 -1.0000000e+00] [ 1.0000000e+00 -1.8369702e-16]] 270 [[-1.8369702e-16 -1.0000000e+00] [ 1.0000000e+00 -1.8369702e-16]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003712177276611328 nb_pixel_total : 1389 time to create 1 rle with old method : 0.0017790794372558594 .time for calcul the mask position with numpy : 0.0003349781036376953 nb_pixel_total : 1157 time to create 1 rle with old method : 0.001535654067993164 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 2 list_crops_rotate : : {'photo_id': -19, 'hashtag_id': 2087736828, 'type': 529, 'x0': 1, 'x1': 52, 'y0': 26, 'y1': 62, 'score': 1.0, 'id': None, 'points': ['41,32,50,40,52,50,49,57,39,61,26,60,13,54,4,46,1,37,5,29,15,25,28,26'], 'sub_photo_id': 0, 'rles': [(-1, 14, 26, 8), (-1, 12, 27, 18), (-1, 9, 28, 23), (-1, 7, 29, 27), (-1, 5, 30, 31), (-1, 5, 31, 33), (-1, 4, 32, 36), (-1, 4, 33, 38), (-1, 3, 34, 40), (-1, 3, 35, 41), (-1, 2, 36, 43), (-1, 2, 37, 45), (-1, 1, 38, 47), (-1, 1, 39, 48), (-1, 2, 40, 48), (-1, 2, 41, 49), (-1, 2, 42, 49), (-1, 3, 43, 48), (-1, 3, 44, 49), (-1, 3, 45, 49), (-1, 4, 46, 48), (-1, 4, 47, 48), (-1, 5, 48, 47), (-1, 6, 49, 47), (-1, 7, 50, 46), (-1, 8, 51, 45), (-1, 10, 52, 43), (-1, 11, 53, 41), (-1, 12, 54, 40), (-1, 13, 55, 38), (-1, 15, 56, 36), (-1, 17, 57, 33), (-1, 19, 58, 31), (-1, 21, 59, 27), (-1, 23, 60, 23), (-1, 25, 61, 18), (-1, 33, 62, 8)], 'hashtag': '', 'sum_segment': 0},: {'photo_id': -19, 'hashtag_id': 2087736828, 'type': 529, 'x0': 107, 'x1': 146, 'y0': 14, 'y1': 57, 'score': 1.0, 'id': None, 'points': ['107,44,110,32,119,22,129,15,138,13,144,16,145,25,143,36,134,47,124,54,115,56,108,52'], 'sub_photo_id': 0, 'rles': [(-1, 136, 14, 4), (-1, 132, 15, 10), (-1, 129, 16, 15), (-1, 127, 17, 19), (-1, 126, 18, 20), (-1, 124, 19, 22), (-1, 123, 20, 23), (-1, 122, 21, 24), (-1, 120, 22, 27), (-1, 119, 23, 28), (-1, 118, 24, 29), (-1, 117, 25, 30), (-1, 116, 26, 31), (-1, 115, 27, 32), (-1, 115, 28, 31), (-1, 114, 29, 32), (-1, 113, 30, 33), (-1, 112, 31, 34), (-1, 111, 32, 34), (-1, 110, 33, 35), (-1, 110, 34, 35), (-1, 110, 35, 35), (-1, 109, 36, 35), (-1, 109, 37, 35), (-1, 109, 38, 34), (-1, 109, 39, 33), (-1, 108, 40, 34), (-1, 108, 41, 33), (-1, 108, 42, 32), (-1, 108, 43, 31), (-1, 107, 44, 31), (-1, 107, 45, 30), (-1, 107, 46, 30), (-1, 107, 47, 29), (-1, 107, 48, 28), (-1, 107, 49, 27), (-1, 108, 50, 24), (-1, 108, 51, 23), (-1, 108, 52, 21), (-1, 108, 53, 20), (-1, 109, 54, 18), (-1, 111, 55, 14), (-1, 113, 56, 9), (-1, 115, 57, 3)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 285 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 285 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.25881905 -0.96592583] [ 0.96592583 0.25881905]] 285 [[ 0.25881905 -0.96592583] [ 0.96592583 0.25881905]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003876686096191406 nb_pixel_total : 727 time to create 1 rle with old method : 0.0009860992431640625 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003387928009033203 nb_pixel_total : 1162 time to create 1 rle with old method : 0.0015180110931396484 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -20, 'hashtag_id': 2087736828, 'type': 529, 'x0': 78, 'x1': 111, 'y0': 24, 'y1': 72, 'score': 1.0, 'id': None, 'points': ['78,62,77,50,83,38,91,28,99,24,107,25,110,34,110,45,104,58,97,67,88,72,81,70'], 'sub_photo_id': 0, 'rles': [(-1, 101, 24, 1), (-1, 98, 25, 8), (-1, 97, 26, 12), (-1, 95, 27, 14), (-1, 94, 28, 15), (-1, 92, 29, 17), (-1, 90, 30, 21), (-1, 90, 31, 21), (-1, 89, 32, 22), (-1, 88, 33, 24), (-1, 87, 34, 25), (-1, 87, 35, 25), (-1, 86, 36, 25), (-1, 85, 37, 27), (-1, 84, 38, 28), (-1, 84, 39, 28), (-1, 83, 40, 29), (-1, 83, 41, 29), (-1, 82, 42, 30), (-1, 81, 43, 31), (-1, 81, 44, 31), (-1, 81, 45, 31), (-1, 80, 46, 31), (-1, 80, 47, 31), (-1, 79, 48, 31), (-1, 79, 49, 32), (-1, 78, 50, 32), (-1, 78, 51, 31), (-1, 78, 52, 31), (-1, 78, 53, 30), (-1, 78, 54, 30), (-1, 78, 55, 29), (-1, 78, 56, 29), (-1, 78, 57, 29), (-1, 78, 58, 28), (-1, 78, 59, 28), (-1, 78, 60, 27), (-1, 78, 61, 26), (-1, 78, 62, 25), (-1, 78, 63, 24), (-1, 78, 64, 23), (-1, 79, 65, 21), (-1, 79, 66, 21), (-1, 80, 67, 19), (-1, 81, 68, 17), (-1, 81, 69, 15), (-1, 81, 70, 14), (-1, 82, 71, 10), (-1, 87, 72, 1), (-1, 89, 72, 3)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 300 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 300 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.5 -0.8660254] [ 0.8660254 0.5 ]] 300 [[ 0.5 -0.8660254] [ 0.8660254 0.5 ]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003294944763183594 nb_pixel_total : 250 time to create 1 rle with old method : 0.00042128562927246094 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.00033926963806152344 nb_pixel_total : 1155 time to create 1 rle with old method : 0.0014853477478027344 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -21, 'hashtag_id': 2087736828, 'type': 529, 'x0': 52, 'x1': 82, 'y0': 44, 'y1': 93, 'score': 1.0, 'id': None, 'points': ['55,86,51,74,54,61,59,50,66,44,74,43,79,50,82,61,80,75,74,87,68,93,60,93'], 'sub_photo_id': 0, 'rles': [(-1, 67, 44, 8), (-1, 65, 45, 11), (-1, 65, 46, 12), (-1, 64, 47, 13), (-1, 62, 48, 17), (-1, 62, 49, 18), (-1, 61, 50, 19), (-1, 60, 51, 21), (-1, 59, 52, 22), (-1, 59, 53, 22), (-1, 58, 54, 23), (-1, 58, 55, 24), (-1, 57, 56, 25), (-1, 57, 57, 25), (-1, 57, 58, 25), (-1, 56, 59, 27), (-1, 55, 60, 28), (-1, 55, 61, 28), (-1, 55, 62, 28), (-1, 54, 63, 29), (-1, 54, 64, 29), (-1, 54, 65, 29), (-1, 54, 66, 29), (-1, 53, 67, 30), (-1, 53, 68, 29), (-1, 54, 69, 28), (-1, 53, 70, 29), (-1, 53, 71, 29), (-1, 53, 72, 28), (-1, 53, 73, 29), (-1, 52, 74, 29), (-1, 52, 75, 29), (-1, 52, 76, 29), (-1, 53, 77, 28), (-1, 53, 78, 27), (-1, 53, 79, 26), (-1, 53, 80, 26), (-1, 54, 81, 25), (-1, 54, 82, 24), (-1, 54, 83, 24), (-1, 55, 84, 22), (-1, 55, 85, 22), (-1, 55, 86, 21), (-1, 55, 87, 21), (-1, 56, 88, 19), (-1, 56, 89, 18), (-1, 57, 90, 16), (-1, 59, 91, 13), (-1, 59, 92, 12), (-1, 60, 93, 10)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 315 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 315 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.70710678 -0.70710678] [ 0.70710678 0.70710678]] 315 [[ 0.70710678 -0.70710678] [ 0.70710678 0.70710678]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.0003685951232910156 nb_pixel_total : 169 time to create 1 rle with old method : 0.00029540061950683594 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003368854522705078 nb_pixel_total : 1161 time to create 1 rle with old method : 0.0019059181213378906 . crop are not in the shrunk photo ! crop are not in the shrunk photo ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -22, 'hashtag_id': 2087736828, 'type': 529, 'x0': 34, 'x1': 62, 'y0': 69, 'y1': 121, 'score': 1.0, 'id': None, 'points': ['40,115,34,105,33,91,35,79,40,72,47,69,54,74,60,84,61,98,59,110,54,118,46,120'], 'sub_photo_id': 0, 'rles': [(-1, 48, 69, 1), (-1, 45, 70, 5), (-1, 41, 71, 1), (-1, 43, 71, 8), (-1, 40, 72, 13), (-1, 39, 73, 15), (-1, 39, 74, 16), (-1, 39, 75, 17), (-1, 38, 76, 18), (-1, 37, 77, 20), (-1, 37, 78, 21), (-1, 36, 79, 22), (-1, 36, 80, 22), (-1, 36, 81, 23), (-1, 35, 82, 25), (-1, 35, 83, 25), (-1, 35, 84, 26), (-1, 34, 85, 27), (-1, 34, 86, 27), (-1, 34, 87, 27), (-1, 34, 88, 28), (-1, 34, 89, 28), (-1, 34, 90, 28), (-1, 34, 91, 28), (-1, 34, 92, 28), (-1, 34, 93, 28), (-1, 34, 94, 28), (-1, 34, 95, 28), (-1, 34, 96, 29), (-1, 34, 97, 29), (-1, 34, 98, 29), (-1, 34, 99, 29), (-1, 34, 100, 29), (-1, 34, 101, 28), (-1, 34, 102, 28), (-1, 34, 103, 28), (-1, 34, 104, 28), (-1, 34, 105, 27), (-1, 35, 106, 26), (-1, 36, 107, 25), (-1, 36, 108, 25), (-1, 36, 109, 25), (-1, 37, 110, 23), (-1, 38, 111, 23), (-1, 38, 112, 22), (-1, 39, 113, 20), (-1, 40, 114, 19), (-1, 40, 115, 18), (-1, 40, 116, 17), (-1, 41, 117, 16), (-1, 42, 118, 15), (-1, 43, 119, 1), (-1, 45, 119, 11), (-1, 46, 120, 6), (-1, 47, 121, 2)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 330 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 330 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.8660254 -0.5 ] [ 0.5 0.8660254]] 330 [[ 0.8660254 -0.5 ] [ 0.5 0.8660254]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.00039458274841308594 nb_pixel_total : 450 time to create 1 rle with old method : 0.0006349086761474609 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003535747528076172 nb_pixel_total : 1159 time to create 1 rle with old method : 0.0019044876098632812 . crop are not in the shrunk photo ! On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -23, 'hashtag_id': 2087736828, 'type': 529, 'x0': 20, 'x1': 51, 'y0': 102, 'y1': 152, 'score': 1.0, 'id': None, 'points': ['33,148,24,140,20,127,19,115,22,106,28,101,36,105,44,113,49,126,51,139,48,147,41,151'], 'sub_photo_id': 0, 'rles': [(-1, 28, 102, 2), (-1, 27, 103, 5), (-1, 25, 104, 11), (-1, 24, 105, 14), (-1, 23, 106, 16), (-1, 22, 107, 17), (-1, 22, 108, 19), (-1, 22, 109, 20), (-1, 22, 110, 20), (-1, 21, 111, 23), (-1, 20, 112, 25), (-1, 21, 113, 25), (-1, 20, 114, 26), (-1, 20, 115, 26), (-1, 20, 116, 27), (-1, 20, 117, 27), (-1, 20, 118, 28), (-1, 20, 119, 28), (-1, 20, 120, 28), (-1, 20, 121, 29), (-1, 20, 122, 29), (-1, 21, 123, 28), (-1, 21, 124, 29), (-1, 21, 125, 29), (-1, 21, 126, 30), (-1, 21, 127, 30), (-1, 21, 128, 30), (-1, 21, 129, 30), (-1, 22, 130, 29), (-1, 22, 131, 29), (-1, 22, 132, 30), (-1, 22, 133, 29), (-1, 24, 134, 27), (-1, 24, 135, 28), (-1, 24, 136, 28), (-1, 24, 137, 28), (-1, 25, 138, 27), (-1, 25, 139, 27), (-1, 25, 140, 27), (-1, 26, 141, 25), (-1, 27, 142, 24), (-1, 27, 143, 24), (-1, 29, 144, 21), (-1, 30, 145, 20), (-1, 31, 146, 19), (-1, 32, 147, 18), (-1, 33, 148, 16), (-1, 34, 149, 14), (-1, 36, 150, 10), (-1, 37, 151, 1), (-1, 39, 151, 5), (-1, 41, 152, 2)], 'hashtag': '', 'sum_segment': 0} Rotation of photo 1361688576 of 345 degree temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0.jpg [, , , ] 345 remove_crop_border : True version de PIL : 9.5.0 Needs to change image size ! [[ 0.96592583 -0.25881905] [ 0.25881905 0.96592583]] 345 [[ 0.96592583 -0.25881905] [ 0.25881905 0.96592583]] shrink_image : True len(list_crops) : 4 time for calcul the mask position with numpy : 0.000392913818359375 nb_pixel_total : 1237 time to create 1 rle with old method : 0.0015621185302734375 On the border Smaller than minimal size ! time for calcul the mask position with numpy : 0.0003285408020019531 nb_pixel_total : 1157 time to create 1 rle with old method : 0.01791071891784668 . crop are not in the shrunk photo ! time for calcul the mask position with numpy : 0.000331878662109375 nb_pixel_total : 234 time to create 1 rle with old method : 0.00040531158447265625 On the border Smaller than minimal size ! len(list_crops_rotate) : 1 list_crops_rotate : : {'photo_id': -24, 'hashtag_id': 2087736828, 'type': 529, 'x0': 12, 'x1': 49, 'y0': 138, 'y1': 183, 'score': 1.0, 'id': None, 'points': ['34,182,24,176,16,164,12,153,13,144,17,138,26,139,36,145,44,156,49,168,48,177,42,183'], 'sub_photo_id': 0, 'rles': [(-1, 18, 138, 2), (-1, 17, 139, 10), (-1, 16, 140, 13), (-1, 16, 141, 15), (-1, 15, 142, 17), (-1, 13, 143, 21), (-1, 13, 144, 23), (-1, 13, 145, 24), (-1, 13, 146, 25), (-1, 13, 147, 25), (-1, 13, 148, 27), (-1, 13, 149, 27), (-1, 12, 150, 29), (-1, 13, 151, 28), (-1, 13, 152, 29), (-1, 13, 153, 30), (-1, 13, 154, 31), (-1, 13, 155, 32), (-1, 13, 156, 32), (-1, 14, 157, 32), (-1, 14, 158, 32), (-1, 14, 159, 32), (-1, 15, 160, 32), (-1, 16, 161, 31), (-1, 16, 162, 32), (-1, 16, 163, 32), (-1, 16, 164, 32), (-1, 17, 165, 32), (-1, 18, 166, 31), (-1, 18, 167, 31), (-1, 19, 168, 31), (-1, 19, 169, 31), (-1, 20, 170, 30), (-1, 21, 171, 29), (-1, 22, 172, 28), (-1, 22, 173, 28), (-1, 23, 174, 27), (-1, 24, 175, 26), (-1, 24, 176, 26), (-1, 26, 177, 24), (-1, 27, 178, 22), (-1, 30, 179, 18), (-1, 30, 180, 17), (-1, 33, 181, 14), (-1, 34, 182, 11), (-1, 37, 183, 3), (-1, 41, 183, 3)], 'hashtag': '', 'sum_segment': 0} About to upload 24 photos upload in portfolio : 23444768 init cache_photo without model_param we have 24 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569112_1112947 we have uploaded 24 photos in the portfolio 23444768 time of upload the photos Elapsed time : 6.74091362953186 map_filename_photo_id : 24 map_filename_photo_id : {'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_00.jpg': 1361688583, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_015.jpg': 1361688584, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_030.jpg': 1361688585, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_045.jpg': 1361688586, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_060.jpg': 1361688587, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_075.jpg': 1361688588, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_090.jpg': 1361688589, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0105.jpg': 1361688590, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0120.jpg': 1361688591, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0135.jpg': 1361688592, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0150.jpg': 1361688593, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0165.jpg': 1361688595, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0180.jpg': 1361688596, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0195.jpg': 1361688597, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0210.jpg': 1361688598, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0225.jpg': 1361688599, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0240.jpg': 1361688600, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0255.jpg': 1361688601, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0270.jpg': 1361688602, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0285.jpg': 1361688603, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0300.jpg': 1361688604, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0315.jpg': 1361688605, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0330.jpg': 1361688606, 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0345.jpg': 1361688607} Len new_chis : 24 Len list_new_chi_with_photo_id : 28 of type : 529 list_new_chi_with_photo_id : [, , , , , , , , , , , , , , , , , , , , , , , , , , , ] batch 1 Loaded 28 chid ids of type : 529 Number RLEs to save : 1197 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! batch 1 Loaded 28 chid ids of type : 529 ++++++++++++++++++++++++++++Number RLEs to save : 0 TO DO : save crop sub photo not yet done ! After datou_step_exec type output : time spend for datou_step_exec : 10.270458936691284 time spend to save output : 7.534027099609375e-05 total time spend for step 3 : 10.27053427696228 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : rotate we use saveGeneral [937852786, 937852786, '1361688576'] map_info['map_portfolio_photo'] : {} final : True mtd_id 243 list_pids : [937852786, 937852786, '1361688576'] Looping around the photos to save general results len do output : 24 /1361688583Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688584Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688585Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688586Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688587Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688588Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688589Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688590Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688591Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688592Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688593Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688595Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688596Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688597Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688598Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688599Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688600Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688601Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688602Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688603Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688604Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688605Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688606Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688607Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('243', None, None, None, None, None, None, None, None) ('243', None, '937852786', None, None, None, None, None, None) ('243', None, None, None, None, None, None, None, None) ('243', None, '937852786', None, None, None, None, None, None) ('243', None, None, None, None, None, None, None, None) ('243', None, '1361688576', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 75 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('243', None, '1361688583', 'None', None, None, None, None, None), ('243', None, '1361688584', 'None', None, None, None, None, None), ('243', None, '1361688585', 'None', None, None, None, None, None), ('243', None, '1361688586', 'None', None, None, None, None, None), ('243', None, '1361688587', 'None', None, None, None, None, None), ('243', None, '1361688588', 'None', None, None, None, None, None), ('243', None, '1361688589', 'None', None, None, None, None, None), ('243', None, '1361688590', 'None', None, None, None, None, None), ('243', None, '1361688591', 'None', None, None, None, None, None), ('243', None, '1361688592', 'None', None, None, None, None, None), ('243', None, '1361688593', 'None', None, None, None, None, None), ('243', None, '1361688595', 'None', None, None, None, None, None), ('243', None, '1361688596', 'None', None, None, None, None, None), ('243', None, '1361688597', 'None', None, None, None, None, None), ('243', None, '1361688598', 'None', None, None, None, None, None), ('243', None, '1361688599', 'None', None, None, None, None, None), ('243', None, '1361688600', 'None', None, None, None, None, None), ('243', None, '1361688601', 'None', None, None, None, None, None), ('243', None, '1361688602', 'None', None, None, None, None, None), ('243', None, '1361688603', 'None', None, None, None, None, None), ('243', None, '1361688604', 'None', None, None, None, None, None), ('243', None, '1361688605', 'None', None, None, None, None, None), ('243', None, '1361688606', 'None', None, None, None, None, None), ('243', None, '1361688607', 'None', None, None, None, None, None), ('243', None, '937852786', None, None, None, None, None, None), ('243', None, '1361688576', None, None, None, None, None, None)] time used for this insertion : 0.023141145706176758 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 3 output : {1361688583: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_00.jpg', [, ]], 1361688584: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_015.jpg', []], 1361688585: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_030.jpg', []], 1361688586: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_045.jpg', []], 1361688587: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_060.jpg', []], 1361688588: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_075.jpg', []], 1361688589: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_090.jpg', [, ]], 1361688590: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0105.jpg', []], 1361688591: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0120.jpg', []], 1361688592: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0135.jpg', []], 1361688593: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0150.jpg', []], 1361688595: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0165.jpg', []], 1361688596: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0180.jpg', [, ]], 1361688597: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0195.jpg', []], 1361688598: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0210.jpg', []], 1361688599: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0225.jpg', []], 1361688600: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0240.jpg', []], 1361688601: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0255.jpg', []], 1361688602: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0270.jpg', [, ]], 1361688603: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0285.jpg', []], 1361688604: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0300.jpg', []], 1361688605: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0315.jpg', []], 1361688606: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0330.jpg', []], 1361688607: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0345.jpg', []]} ret_da : {1361688583: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_00.jpg', [, ]], 1361688584: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_015.jpg', []], 1361688585: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_030.jpg', []], 1361688586: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_045.jpg', []], 1361688587: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_060.jpg', []], 1361688588: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_075.jpg', []], 1361688589: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_090.jpg', [, ]], 1361688590: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0105.jpg', []], 1361688591: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0120.jpg', []], 1361688592: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0135.jpg', []], 1361688593: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0150.jpg', []], 1361688595: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0165.jpg', []], 1361688596: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0180.jpg', [, ]], 1361688597: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0195.jpg', []], 1361688598: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0210.jpg', []], 1361688599: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0225.jpg', []], 1361688600: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0240.jpg', []], 1361688601: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0255.jpg', []], 1361688602: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0270.jpg', [, ]], 1361688603: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0285.jpg', []], 1361688604: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0300.jpg', []], 1361688605: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0315.jpg', []], 1361688606: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0330.jpg', []], 1361688607: ['937852786', 'temp/1748569077_1112947_937852786_7d9a231a08a1c63d0868e56a5361bf67_0345.jpg', []]} list chi : [[, ], [], [], [], [], [], [, ], [], [], [], [], [], [, ], [], [], [], [], [], [, ], [], [], [], [], []] ############################### TEST flip ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=571 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=571 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 571 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=571 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : flip list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (911785586) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 911785586 download finish for photo 911785586 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.14235186576843262 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:flip Fri May 30 03:38:41 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big.jpg': 911785586} map_photo_id_path_extension : {911785586: {'path': 'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step_flip ! We are in a linear step without datou_depend ! batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 911785586) and `type` in (741) Loaded 6 chid ids of type : 741 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (18344206,18344211,18344210,18344209,18344208,18344207) +++++WARNING : Unexpected points, we should remove this data for chi_id : 18344210, for now we just ignore these empty polygon points +SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (18344206,18344211,18344210,18344209,18344208,18344207) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (18344206,18344211,18344210,18344209,18344208,18344207) map_chi_objs : {911785586: [, , , , , ]} photo_id in download_rotate_and_save : 911785586 list_chi_loc : 6 Vertical flip of photo 911785586 version de PIL : 9.5.0 vertically flipped image is saved in temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_vert.jpg Horizontal flip of photo 911785586 version de PIL : 9.5.0 horizontally flipped image is saved in temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_hori.jpg About to upload 2 photos upload in portfolio : 1090565 init cache_photo without model_param we have 2 photo to upload uploaded to storage server : ovh folder_temporaire : temp/1748569122_1112947 we have uploaded 2 photos in the portfolio 1090565 time of upload the photos Elapsed time : 0.7974135875701904 map_filename_photo_id : 2 map_filename_photo_id : {'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_vert.jpg': 1361688610, 'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_hori.jpg': 1361688611} Len new_chis : 12 Len list_new_chi_with_photo_id : 12 of type : 741 list_new_chi_with_photo_id : [, , , , , , , , , , , ] insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) batch 1 Loaded 12 chid ids of type : 741 INSERT IGNORE INTO MTRPhoto.crop_polygon_points (`crop_hashtag_id`, `points`) VALUES (%s, %s) Number RLEs to save : 0 INSERT IGNORE INTO MTRPhoto.crop_sum_segments (`crop_hashtag_id`, `sum_segments`) VALUES (%s, %s) TO DO : save crop sub photo not yet done ! After datou_step_exec type output : time spend for datou_step_exec : 0.9339425563812256 time spend to save output : 6.937980651855469e-05 total time spend for step 1 : 0.9340119361877441 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : flip we use saveGeneral [911785586] map_info['map_portfolio_photo'] : {} final : True mtd_id 571 list_pids : [911785586] Looping around the photos to save general results len do output : 2 /1361688610 /1361688611 before output type Managing all output in save final without adding information in the mtr_datou_result ('571', None, None, None, None, None, None, None, None) ('571', None, '911785586', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 1 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('571', None, '911785586', None, None, None, None, None, None)] time used for this insertion : 0.015514612197875977 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {'1361688610': ['911785586', 'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_vert.jpg', [, , , , , ]], '1361688611': ['911785586', 'temp/1748569121_1112947_911785586_d8582feabcd359151ff718b5832248c7-big_flip_hori.jpg', [, , , , , ]]} ############################### TEST crop_rles ################################ # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! Unexpected type seems boolean for variable list_input_json ERROR or WARNING : can't parse json string Expecting value: line 1 column 1 (char 0) Tried to parse : TEST CROP RLES Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=686 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=686 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 686 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=686 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : crop list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (950103132) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 950103132 download finish for photo 950103132 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.14583826065063477 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:crop Fri May 30 03:38:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00.jpg': 950103132} map_photo_id_path_extension : {950103132: {'path': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Crop ! param_json : {'photo_hashtag_type': 755, 'token': '78d09a0790ec6ecbf119343125a81fdc', 'feed_id_new_photos': 0, 'host': 'www.fotonower.com', 'crop_type': 'rle', 'margin_relative': 0.1, 'min_score': 0.3, 'upload,type': 'python'} margin_type : margin_relative margin_value : [0.1, 0.1, 0.1, 0.1] Loading chi in step crop with photo_hashtag_type : 755 Loading chi in step crop for list_pids : 1 ! batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 950103132) and `type` in (755) and score>0.3 Loaded 8 chid ids of type : 755 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1947670931,1947670932,1947670933,1947670934,1947670935,1947670936,1947670937,1947670938) ++++++++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947670931,1947670932,1947670933,1947670934,1947670935,1947670936,1947670937,1947670938) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1947670931,1947670932,1947670933,1947670934,1947670935,1947670936,1947670937,1947670938) select photo_id, sub_photo_id, x0, x1, y0, y1, resize_coeff_x, resize_coeff_y, crop_type, id from MTRPhoto.photo_sub_photos where photo_id in ( 950103132) WARNING : margin is only used for type bib ! type of cropped photo chosen : rle we resize croppped photo by 1 on x axis and by 1 on y axis we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! we have both polygon and rles Here we manage rles ! map_result returned by crop_photo_return_map_crop : length : 8 map_result after crop : {1947670931: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670931_0.jpg', 'coordonates': (183, 199, 15, 41), 'sub_photo_id': -1, 'same_chi': False}, 1947670932: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670932_0.jpg', 'coordonates': (38, 85, 113, 140), 'sub_photo_id': -1, 'same_chi': False}, 1947670933: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670933_0.jpg', 'coordonates': (168, 194, 141, 151), 'sub_photo_id': -1, 'same_chi': False}, 1947670934: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670934_0.jpg', 'coordonates': (47, 101, 16, 110), 'sub_photo_id': -1, 'same_chi': False}, 1947670935: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670935_0.jpg', 'coordonates': (175, 199, 104, 111), 'sub_photo_id': -1, 'same_chi': False}, 1947670936: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670936_0.jpg', 'coordonates': (86, 130, 184, 196), 'sub_photo_id': -1, 'same_chi': False}, 1947670937: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670937_0.jpg', 'coordonates': (79, 195, 0, 61), 'sub_photo_id': -1, 'same_chi': False}, 1947670938: {'crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg', 'photo_id': 950103132, 'bib_crop': 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_bib_crop_1947670938_0.jpg', 'coordonates': (131, 155, 181, 195), 'sub_photo_id': -1, 'same_chi': False}} Here we crop with rles About to insert : list_path_to_insert length 8 new photo from crops ! About to upload 8 photos https://marlene.fotonower.com/api/v1/secured/portfolio/new?access_token=78d09a0790ec6ecbf119343125a81fdc upload in portfolio : 23444769 in upload media Upload medias : ['temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg'] : url : https://marlene.fotonower.com/api/v1/secured/photo/upload?token=78d09a0790ec6ecbf119343125a81fdc&datou=0 temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg after data_to_send, before sending request after request b'{"photo_ids":["1361688617","1361688620","1361688613","1361688616","1361688619","1361688621","1361688615","1361688618"],"photo_ids_order":["1361688613","1361688615","1361688616","1361688617","1361688618","1361688619","1361688620","1361688621"],"photo_detail":[{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/decc4cfa999feba4489c4bf3fa6164f5.jpg","text":"TemporaryFile(/tmp/multipartBody5931535568324031045asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/27bf95c1abb2060ad2655ca8c106ee03.jpg","text":"TemporaryFile(/tmp/multipartBody2811704202164692600asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/e9c2c011e181f18527bf6e49b16efe2e.jpg","text":"TemporaryFile(/tmp/multipartBody7622246482048089256asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/a5104915d35773c940f95725d1086c3f.jpg","text":"TemporaryFile(/tmp/multipartBody197600969379431526asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/806c6c1fcaee188f98d179042821ddaf.jpg","text":"TemporaryFile(/tmp/multipartBody9085186986937906668asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/42ee573c2685b8a53143d8e2493bf01f.jpg","text":"TemporaryFile(/tmp/multipartBody6486161091875575619asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/295c51f489894724a503d7cf6c9ef0fa.jpg","text":"TemporaryFile(/tmp/multipartBody7044657031443549396asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg","height":0,"width":0},{"mtr_user_id":440,"url":"https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2025/5/30/1379681188b84e229ab41da9960db376.jpg","text":"TemporaryFile(/tmp/multipartBody8739915455538050335asTemporaryFile)","latitude":0.0,"longitude":0.0,"uploaded_at":1748569124558,"filename":"1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg","height":0,"width":0}],"map_files_photo_id":{"file2":"1361688616","file6":"1361688620","file1":"1361688615","file7":"1361688621","file0":"1361688613","file4":"1361688618","file5":"1361688619","file3":"1361688617"},"map_files_photo_id_array":[{"photo_id":"1361688619","filename":"file5"},{"photo_id":"1361688616","filename":"file2"},{"photo_id":"1361688618","filename":"file4"},{"photo_id":"1361688621","filename":"file7"},{"photo_id":"1361688615","filename":"file1"},{"photo_id":"1361688613","filename":"file0"},{"photo_id":"1361688617","filename":"file3"},{"photo_id":"1361688620","filename":"file6"}],"portfolio_id":23444769,"hashtag_by_photo_ids":[{"1361688617":["hashtag1","hashtag2"]},{"1361688620":["hashtag1","hashtag2"]},{"1361688613":["hashtag1","hashtag2"]},{"1361688616":["hashtag1","hashtag2"]},{"1361688619":["hashtag1","hashtag2"]},{"1361688621":["hashtag1","hashtag2"]},{"1361688615":["hashtag1","hashtag2"]},{"1361688618":["hashtag1","hashtag2"]}],"comms":"Portfolio 23444769 used, photo_id : ArrayBuffer(1361688617, 1361688620, 1361688613, 1361688616, 1361688619, 1361688621, 1361688615, 1361688618)","result":[],"list_datou_current":[]}' Result OK ! uploaded one batch 0 Elapsed time : 19.052467584609985 map_result_insert : {'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg': 1361688616, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg': 1361688620, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg': 1361688615, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg': 1361688621, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg': 1361688613, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg': 1361688618, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg': 1361688619, 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg': 1361688617} Now we prepare data that will be used for ellipse search ! chi_id found to be used 1947670931 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg sub_photo_id found to be used 1361688613 chi_id found to be used 1947670932 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg sub_photo_id found to be used 1361688615 chi_id found to be used 1947670933 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg sub_photo_id found to be used 1361688616 chi_id found to be used 1947670934 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg sub_photo_id found to be used 1361688617 chi_id found to be used 1947670935 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg sub_photo_id found to be used 1361688618 chi_id found to be used 1947670936 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg sub_photo_id found to be used 1361688619 chi_id found to be used 1947670937 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg sub_photo_id found to be used 1361688620 chi_id found to be used 1947670938 path of cropped varroa found to be used to match on an ellipse temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg sub_photo_id found to be used 1361688621 insert ignore into MTRPhoto.crop_sub_photo_ids (crop_hashtag_id, sub_photo_id, mtr_user_id) VALUES (%s,%s,%s) : [(1947670931, '1361688613', 31), (1947670932, '1361688615', 31), (1947670933, '1361688616', 31), (1947670934, '1361688617', 31), (1947670935, '1361688618', 31), (1947670936, '1361688619', 31), (1947670937, '1361688620', 31), (1947670938, '1361688621', 31)] map of cropped photos with some data : {'1361688613': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg', (183, 199, 15, 41)], '1361688615': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg', (38, 85, 113, 140)], '1361688616': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg', (168, 194, 141, 151)], '1361688617': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg', (47, 101, 16, 110)], '1361688618': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg', (175, 199, 104, 111)], '1361688619': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg', (86, 130, 184, 196)], '1361688620': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg', (79, 195, 0, 61)], '1361688621': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg', (131, 155, 181, 195)]} After datou_step_exec type output : time spend for datou_step_exec : 19.127057552337646 time spend to save output : 7.319450378417969e-05 total time spend for step 1 : 19.12713074684143 caffe_path_current : About to save ! 1 Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : crop we use saveGeneral [950103132] map_info['map_portfolio_photo'] : {} final : True mtd_id 686 list_pids : [950103132] Looping around the photos to save general results len do output : 8 /1361688613Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688615Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688616Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688617Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688618Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688619Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688620Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . /1361688621Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('686', None, None, None, None, None, None, None, None) ('686', None, '950103132', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 25 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('686', None, '1361688613', 'None', None, None, None, None, None), ('686', None, '1361688615', 'None', None, None, None, None, None), ('686', None, '1361688616', 'None', None, None, None, None, None), ('686', None, '1361688617', 'None', None, None, None, None, None), ('686', None, '1361688618', 'None', None, None, None, None, None), ('686', None, '1361688619', 'None', None, None, None, None, None), ('686', None, '1361688620', 'None', None, None, None, None, None), ('686', None, '1361688621', 'None', None, None, None, None, None), ('686', None, '950103132', None, None, None, None, None, None)] time used for this insertion : 0.014973878860473633 save_final save missing photos in datou_result : After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {'1361688613': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg', (183, 199, 15, 41)], '1361688615': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg', (38, 85, 113, 140)], '1361688616': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg', (168, 194, 141, 151)], '1361688617': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg', (47, 101, 16, 110)], '1361688618': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg', (175, 199, 104, 111)], '1361688619': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg', (86, 130, 184, 196)], '1361688620': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg', (79, 195, 0, 61)], '1361688621': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg', (131, 155, 181, 195)]} ret_da : {'1361688613': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670931_0.jpg', (183, 199, 15, 41)], '1361688615': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg', (38, 85, 113, 140)], '1361688616': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670933_0.jpg', (168, 194, 141, 151)], '1361688617': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670934_0.jpg', (47, 101, 16, 110)], '1361688618': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670935_0.jpg', (175, 199, 104, 111)], '1361688619': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670936_0.jpg', (86, 130, 184, 196)], '1361688620': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670937_0.jpg', (79, 195, 0, 61)], '1361688621': ['950103132', 'temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670938_0.jpg', (131, 155, 181, 195)]} 8 Found filename_to_hash : temp/1748569122_1112947_950103132_4f47bd527301396b0a701a1b4183ba00_rle_crop_1947670932_0.jpg ############################### TEST angular_coeff ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=852 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=852 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 852 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=852 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : angular_coeff list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (932296368) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 932296368 download finish for photo 932296368 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.141371488571167 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:angular_coeff Fri May 30 03:39:02 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569142_1112947_932296368_97c5e7b0f2830e550e2d6eeb248d8006.jpg': 932296368} map_photo_id_path_extension : {932296368: {'path': 'temp/1748569142_1112947_932296368_97c5e7b0f2830e550e2d6eeb248d8006.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} beginning of step detection filter param_json : {'input_type': 846, 'output_type': -1, 'orientation_type': 872, 'ref_crop_type': 846, 'condition_crop': 'car', 'criteria_crop': 'center_rect', 'crops_coeffs': {'CAR_EXTERIEUR_angle_avant_droit.*': {'aile-avant': [[15, 0.0], [240, 0.0], [285, 1.0], [345, 1.0]], 'capot': [[45, 1.0], [60, 0.5], [270, 0.0], [315, 1.0], [360, 1.0]]}}} angular_coefficients_to_crops batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 932296368) and `type` in (846) Loaded 19 chid ids of type : 846 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (769189713,769189714,769189715,769189716,769189717,769189718,769189721,769189723,769189724,769189725,769189727,769189729,769189730,769189732,769189733,769189734,769189737,769189738,769189739) SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (769189713,769189714,769189715,769189716,769189717,769189718,769189721,769189723,769189724,769189725,769189727,769189729,769189730,769189732,769189733,769189734,769189737,769189738,769189739) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (769189713,769189714,769189715,769189716,769189717,769189718,769189721,769189723,769189724,769189725,769189727,769189729,769189730,769189732,769189733,769189734,769189737,769189738,769189739) select distinct hashtag_id from MTRBack.photo_hashtag_ids where photo_id in (932296368) and type=872 treating photo 932296368 select distinct hashtag_id from MTRBack.photo_hashtag_ids where photo_id in (932296368) and type=872 After datou_step_exec type output : time spend for datou_step_exec : 0.10008120536804199 time spend to save output : 0.002645730972290039 total time spend for step 1 : 0.10272693634033203 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {932296368: ([(932296368, 2106233860, 846, 1066, 1277, 93, 340, 0.31964028378983567, 0, []), (932296368, 2106233860, 846, 434, 690, 218, 498, 0.7170410105787726, 0, []), (932296368, 503548896, 846, 902, 1111, 466, 576, 0.31724966, 769189715, []), (932296368, 599722655, 846, 523, 1100, 152, 337, 0.98039776, 0, []), (932296368, 492601069, 846, 143, 1190, 90, 695, 0.9696157, 769189717, []), (932296368, 492601069, 846, 0, 408, 246, 719, 0.9431181, 769189718, []), (932296368, 2096875722, 846, 567, 964, 162, 215, 0.55490255, 769189721, []), (932296368, 2096875709, 846, 437, 939, 24, 198, 0.9983077, 769189723, []), (932296368, 2096875709, 846, 1004, 1263, 28, 144, 0.9485744, 769189724, []), (932296368, 624624117, 846, 595, 1122, 331, 640, 0.99100167, 769189725, []), (932296368, 492624020, 846, 585, 874, 308, 393, 0.78697366, 769189727, []), (932296368, 2096875719, 846, 943, 1100, 428, 547, 0.96733797, 769189729, []), (932296368, 492654799, 846, 253, 467, 35, 441, 0.99621326, 769189730, []), (932296368, 492689227, 846, 1118, 1264, 270, 438, 0.9901647, 769189732, []), (932296368, 492689227, 846, 486, 671, 378, 690, 0.98789483, 769189733, []), (932296368, 492689227, 846, 161, 255, 229, 409, 0.70801014, 769189734, []), (932296368, 492925064, 846, 261, 421, 27, 193, 0.92215157, 769189737, []), (932296368, 492925064, 846, 873, 1045, 46, 156, 0.7535122, 769189738, []), (932296368, 492925064, 846, 1090, 1279, 20, 107, 0.45259848, 769189739, [])],)} test angular coeff is a success ! ############################### TEST detection_filter_by_crop ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=708 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=708 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 708 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=708 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : detection_filter_by_crop list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (946711423) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 946711423 download finish for photo 946711423 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.12580633163452148 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:detection_filter_by_crop Fri May 30 03:39:02 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569142_1112947_946711423_b4bef6b5c6c4b6ffae23f8718c42183c.jpg': 946711423} map_photo_id_path_extension : {946711423: {'path': 'temp/1748569142_1112947_946711423_b4bef6b5c6c4b6ffae23f8718c42183c.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} beginning of step detection filter param_json : {'input_type': 631, 'output_type': -1, 'condition_type': 445, 'condition_crop': 'car', 'criteria_crop': 'center_rect', 'min_surface_ratio': 0.7} conditional_crop_copy batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 946711423) and `type` in (445) Loaded 3 chid ids of type : 445 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1947734477,18345275,18345276) +++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947734477,18345275,18345276) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1947734477,18345275,18345276) batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 946711423) and `type` in (631) Loaded 35 chid ids of type : 631 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 946711423) and `type` in (445) Loaded 3 chid ids of type : 445 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1947734477,18345275,18345276) +++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947734477,18345275,18345276) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1947734477,18345275,18345276) treating photo 946711423 SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 624624117, 'type': 631, 'x0': 226, 'x1': 569, 'y0': 252, 'y1': 425, 'score': 0.99812776, 'id': 1947740368, 'points': ['395,419,341,419,340,418,316,418,315,417,306,417, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 162, 'x1': 245, 'y0': 233, 'y1': 396, 'score': 0.99702626, 'id': 1947740369, 'points': ['215,393,206,393,202,390,200,390,192,383,191,380, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 96, 'x1': 172, 'y0': 39, 'y1': 261, 'score': 0.9928518, 'id': 1947740370, 'points': ['143,252,143,249,141,246,140,246,138,248,138,251,137 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 545, 'x1': 612, 'y0': 186, 'y1': 276, 'score': 0.9876676, 'id': 1947740371, 'points': ['584,267,583,266,578,266,574,262,574,259,573,258,5 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875719, 'type': 631, 'x0': 468, 'x1': 555, 'y0': 292, 'y1': 365, 'score': 0.9830025, 'id': 1947740372, 'points': ['491,350,489,350,488,349,487,350,483,350,480,348, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 599722655, 'type': 631, 'x0': 176, 'x1': 535, 'y0': 138, 'y1': 264, 'score': 0.9818268, 'id': 1947740373, 'points': ['453,253,413,253,412,252,387,252,386,250,386,248,3 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 53, 'x1': 87, 'y0': 127, 'y1': 212, 'score': 0.9786105, 'id': 1947740374, 'points': ['74,201,69,201,67,199,66,199,65,198,62,192,62,190,61 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492844413, 'type': 631, 'x0': 89, 'x1': 163, 'y0': 93, 'y1': 144, 'score': 0.9772748, 'id': 1947740375, 'points': ['159,142,153,141,151,139,148,138,145,135,141,133,139 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 418, 'x1': 522, 'y0': 69, 'y1': 136, 'score': 0.97407305, 'id': 1947740376, 'points': ['510,121,507,121,505,119,501,120,500,119,500,113,4 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875709, 'type': 631, 'x0': 185, 'x1': 431, 'y0': 39, 'y1': 136, 'score': 0.97171515, 'id': 1947740377, 'points': ['331,134,287,134,286,133,284,133,283,134,272,134, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875722, 'type': 631, 'x0': 198, 'x1': 395, 'y0': 118, 'y1': 142, 'score': 0.9699756, 'id': 1947740378, 'points': ['328,137,251,137,250,136,249,137,241,137,240,136, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 499500794, 'type': 631, 'x0': 93, 'x1': 107, 'y0': 127, 'y1': 146, 'score': 0.9574813, 'id': 1947740379, 'points': ['101,143,98,143,95,139,95,131,97,129,100,129,101,13 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 71, 'x1': 125, 'y0': 36, 'y1': 95, 'score': 0.95296955, 'id': 1947740380, 'points': ['104,92,96,92,93,90,91,90,86,86,83,85,83,84,81,82,80 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 101, 'x1': 167, 'y0': 38, 'y1': 127, 'score': 0.9508439, 'id': 1947740381, 'points': ['154,117,152,115,152,112,150,110,148,106,148,104,14 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492624020, 'type': 631, 'x0': 249, 'x1': 400, 'y0': 219, 'y1': 316, 'score': 0.8792459, 'id': 1947740382, 'points': ['395,313,390,313,386,311,384,312,381,312,376,309,3 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 540, 'x1': 625, 'y0': 78, 'y1': 221, 'score': 0.87864035, 'id': 1947740383, 'points': ['567,127,566,127,565,126,564,109,562,106,560,106,5 surface aoutside conditional crop crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 547, 'x1': 640, 'y0': 79, 'y1': 129, 'score': 0.8165246, 'id': 1947740384, 'points': ['630,96,627,96,627,94,628,92,629,92,631,94,631,95', surface aoutside conditional crop crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 360, 'x1': 434, 'y0': 62, 'y1': 116, 'score': 0.74684095, 'id': 1947740385, 'points': ['415,103,413,103,411,101,408,101,405,99,403,99,401 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 503548896, 'type': 631, 'x0': 302, 'x1': 540, 'y0': 339, 'y1': 403, 'score': 0.7406652, 'id': 1947740386, 'points': ['442,401,372,401,372,397,370,395,369,392,366,390,3 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 53, 'x1': 85, 'y0': 75, 'y1': 182, 'score': 0.73015845, 'id': 1947740387, 'points': ['70,147,68,145,65,139,65,137,62,132,61,128,57,126,5 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875717, 'type': 631, 'x0': 477, 'x1': 510, 'y0': 220, 'y1': 243, 'score': 0.69028217, 'id': 1947740388, 'points': ['501,241,493,241,489,239,488,237,487,237,480,232 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 61, 'x1': 115, 'y0': 42, 'y1': 188, 'score': 0.6900027, 'id': 1947740389, 'points': ['92,47,91,45,92,44,96,45,94,45', '73,141,73,136,72,1 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875712, 'type': 631, 'x0': 309, 'x1': 326, 'y0': 382, 'y1': 404, 'score': 0.6633776, 'id': 1947740390, 'points': ['309,383,309,382,311,382', '325,385,324,383,319,3 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875719, 'type': 631, 'x0': 427, 'x1': 553, 'y0': 258, 'y1': 315, 'score': 0.6446218, 'id': 1947740391, 'points': ['531,284,526,284,525,283,525,281,523,279,522,280, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233861, 'type': 631, 'x0': 144, 'x1': 267, 'y0': 181, 'y1': 307, 'score': 0.63958377, 'id': 1947740392, 'points': ['212,251,209,251,208,250,203,251,201,250,201,249 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875712, 'type': 631, 'x0': 285, 'x1': 433, 'y0': 343, 'y1': 377, 'score': 0.61493844, 'id': 1947740393, 'points': ['431,376,286,376,285,375,285,368,286,367,286,362 crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 146, 'x1': 287, 'y0': 140, 'y1': 311, 'score': 0.54784286, 'id': 1947740394, 'points': ['234,254,227,254,221,251,219,248,215,253,212,253 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 496, 'x1': 624, 'y0': 141, 'y1': 245, 'score': 0.46262404, 'id': 1947740395, 'points': ['603,176,599,173,595,176,593,176,590,174,589,174 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 495920967, 'type': 631, 'x0': 202, 'x1': 524, 'y0': 112, 'y1': 333, 'score': 0.45109355, 'id': 1947740396, 'points': ['483,289,483,286,482,285,482,283,480,279,480,274, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 2096875722, 'type': 631, 'x0': 433, 'x1': 558, 'y0': 248, 'y1': 286, 'score': 0.44133398, 'id': 1947740397, 'points': ['492,272,474,272,473,271,468,271,465,269,460,269 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233861, 'type': 631, 'x0': 535, 'x1': 630, 'y0': 138, 'y1': 231, 'score': 0.42747068, 'id': 1947740398, 'points': ['590,171,589,170,585,170,584,171,581,169,579,169 surface aoutside conditional crop crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 399, 'x1': 569, 'y0': 68, 'y1': 251, 'score': 0.41876298, 'id': 1947740399, 'points': [], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', ' crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 492624020, 'type': 631, 'x0': 420, 'x1': 552, 'y0': 244, 'y1': 293, 'score': 0.35962066, 'id': 1947740400, 'points': ['474,289,453,289,452,288,439,288,437,286,431,286, crop duplicated : : {'photo_id': 946711423, 'hashtag_id': 503548896, 'type': 631, 'x0': 301, 'x1': 540, 'y0': 339, 'y1': 403, 'score': 0.740756, 'id': 3140491551, 'points': ['442,401,371,401,371,397,366,390,365,386,356,386,35 crop not duplicated : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 496, 'x1': 624, 'y0': 140, 'y1': 245, 'score': 0.4627206, 'id': 3140491552, 'points': ['595,176,593,176,589,173,586,176,583,176,580,174, surface aoutside conditional crop After datou_step_exec type output : time spend for datou_step_exec : 0.12784767150878906 time spend to save output : 5.316734313964844e-05 total time spend for step 1 : 0.1279008388519287 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {946711423: ([(946711423, 624624117, 631, 226, 569, 252, 425, 0.99812776, 1947740368, ['395,419,341,419,340,418,316,418,315,417,306,417,305,416,293,415,290,413,284,412,283,411,280,411,272,407,264,405,258,400,254,398,250,394,244,391,242,389,242,386,239,380,240,368,239,367,239,347,238,346,238,331,237,330,237,327,238,326,237,314,239,311,239,308,237,304,238,302,243,298,244,296,244,292,246,291,250,291,251,290,259,290,260,289,264,289,265,288,269,288,271,290,273,294,278,299,280,300,285,300,286,301,293,301,294,302,302,304,305,307,309,308,312,310,314,310,317,312,335,312,336,313,343,313,344,314,370,314,371,315,381,315,382,314,389,313,393,311,405,309,406,308,408,308,412,306,414,304,417,304,421,307,426,308,427,309,433,309,434,310,464,309,467,306,471,304,476,304,477,303,489,303,490,302,494,302,495,301,500,301,501,300,515,300,516,299,519,298,522,292,525,290,533,290,534,291,540,291,541,290,543,290,547,288,550,285,550,285,552,289,552,291,553,292,553,313,552,314,552,324,550,328,550,333,549,334,549,336,544,346,543,353,539,361,532,368,531,368,527,372,519,374,509,379,503,384,499,385,498,386,496,386,492,388,490,390,486,392,484,392,479,396,475,397,474,398,472,398,471,399,469,399,462,403,460,403,459,404,457,404,456,405,454,405,450,407,448,407,443,410,425,413,424,414,422,414,416,417,404,417,403,418,396,418']), (946711423, 492689227, 631, 162, 245, 233, 396, 0.99702626, 1947740369, ['215,393,206,393,202,390,200,390,192,383,191,380,187,375,184,369,184,367,180,360,180,358,179,357,177,349,175,347,174,339,172,336,171,330,170,329,169,324,168,323,168,313,167,312,167,304,166,303,166,298,165,297,165,288,164,287,165,286,165,272,166,271,166,268,167,267,167,263,168,262,169,254,173,249,177,247,178,247,181,251,184,251,184,252,187,255,189,255,193,259,193,261,195,263,195,264,201,270,203,278,207,282,208,289,211,293,211,296,213,299,214,304,215,305,216,312,219,316,219,319,220,320,220,325,222,329,222,335,223,336,223,338,225,342,225,349,226,350,226,359,227,360,227,366,228,367,228,371,231,375,231,382,227,385,226,388,225,389,223,388,219,392,216,392']), (946711423, 492654799, 631, 96, 172, 39, 261, 0.9928518, 1947740370, ['143,252,143,249,141,246,140,246,138,248,138,251,137,250,137,248,135,246,134,246,132,248,127,244,124,244,122,241,122,236,121,235,121,232,118,229,117,225,116,224,116,212,113,209,115,207,116,201,111,194,110,184,106,178,107,154,108,152,112,148,113,144,112,143,112,138,110,136,108,136,107,135,103,128,103,124,102,123,102,121,103,120,103,118,106,115,106,106,107,105,110,104,113,101,117,93,117,71,114,65,116,61,116,59,117,58,117,55,118,54,119,49,122,45,122,44,124,42,150,42,151,43,153,43,153,47,152,48,152,50,154,52,155,56,156,57,156,85,155,86,155,95,154,96,154,98,155,99,155,105,156,106,155,107,155,116,157,120,159,121,159,123,156,127,156,134,157,135,157,138,156,139,156,141,154,145,152,147,150,151,149,159,148,160,148,164,149,165,149,174,148,175,148,197,149,198,149,215,150,216,150,241,149,242,149,245,148,247,146,245,144,247', '122,147,121,138,120,141,119,142,119,144,118,145,121,148']), (946711423, 2096875719, 631, 468, 555, 292, 365, 0.9830025, 1947740372, ['491,350,489,350,488,349,487,350,483,350,480,348,480,341,482,339,482,337,485,334,487,334,491,330,494,330,495,328,498,326,501,326,503,324,507,325,509,323,514,321,516,319,518,321,520,321,521,319,522,319,524,321,527,321,530,317,530,315,531,314,535,313,540,309,543,310,544,311,542,313,542,314,544,316,541,318,541,322,536,322,535,323,533,323,532,322,528,322,527,321,524,321,522,323,518,322,516,324,517,327,516,328,512,327,510,329,512,332,513,332,515,330,516,331,516,333,514,332,511,333,511,336,514,337,516,336,516,339,515,339,513,338,511,340,512,341,512,342,510,343,507,343,502,347,500,347,497,349,492,349', '514,325,515,324,513,322,512,322,511,325,512,326', '522,327,521,327,521,326,522,325']), (946711423, 599722655, 631, 176, 535, 138, 264, 0.9818268, 1947740373, ['453,253,413,253,412,252,387,252,386,250,386,248,383,246,379,245,376,243,361,243,361,240,362,239,359,238,358,237,356,237,355,236,352,236,351,235,333,235,332,234,329,234,329,233,331,231,331,229,329,228,328,224,330,222,330,221,324,218,308,219,307,218,302,218,298,216,288,217,287,218,285,218,283,220,283,221,287,224,295,225,295,225,294,226,289,226,288,227,283,227,282,228,273,228,272,229,271,228,259,228,258,227,254,227,253,226,247,225,247,225,251,221,248,218,243,216,247,213,248,213,249,212,248,211,246,211,245,210,241,210,240,209,237,209,236,208,231,207,230,206,228,202,224,201,223,200,221,200,220,199,214,198,213,195,211,193,208,193,203,189,203,184,201,181,201,176,198,171,199,170,199,158,203,154,205,153,205,151,206,149,209,149,210,148,225,148,226,147,283,147,284,148,287,148,288,147,305,147,306,148,312,148,313,147,354,147,355,146,428,146,429,147,433,147,434,148,437,148,438,149,451,149,457,156,459,162,462,165,464,166,471,166,472,165,477,165,480,167,480,171,486,175,488,175,489,176,502,176,503,178,503,180,509,185,509,189,512,193,512,199,513,200,513,203,514,204,514,210,513,211,514,217,512,221,513,222,513,225,510,229,510,235,507,237,504,238,502,243,490,243,489,244,485,244,484,245,480,245,479,246,463,246,462,247,460,247,458,249,457,252,454,252', '528,212,528,207,526,206,524,203,526,203,527,202,528,202', '299,215,302,212,299,211,298,210,291,210,290,211,281,212,286,215,290,215,291,216', '375,242,376,240,375,238,363,239,368,242,371,242,372,243']), (946711423, 492844413, 631, 89, 163, 93, 144, 0.9772748, 1947740375, ['159,142,153,141,151,139,148,138,145,135,141,133,139,133,138,132,131,132,130,131,125,131,124,130,121,130,120,129,116,129,115,128,112,128,108,126,106,126,100,123,98,121,94,113,94,104,97,101,103,98,105,98,106,97,110,97,111,96,116,96,117,95,132,95,133,96,139,97,141,99,144,100,149,105,150,107,154,108,155,113,157,115,158,115,160,118,160,120,161,121,161,133,160,134,160,140']), (946711423, 2096875709, 631, 185, 431, 39, 136, 0.97171515, 1947740377, ['331,134,287,134,286,133,284,133,283,134,272,134,271,133,264,133,263,134,258,134,257,133,254,133,253,132,236,132,235,131,225,131,224,132,223,131,213,131,212,130,208,130,207,129,204,129,203,128,199,127,193,121,192,117,189,113,189,110,188,109,187,93,186,92,187,91,187,89,186,88,186,65,185,64,186,63,186,61,185,60,185,48,186,47,186,42,187,40,232,40,233,41,248,41,249,42,281,43,282,44,290,44,291,45,300,45,301,46,308,46,309,47,314,47,315,48,322,49,328,53,334,54,336,56,339,57,344,62,349,64,351,66,353,67,356,67,358,69,359,72,363,76,367,78,369,80,379,91,380,93,383,94,390,100,393,101,395,103,396,106,399,109,402,110,406,115,408,115,410,117,410,120,412,123,411,127,409,129,399,129,398,130,395,130,394,131,378,131,377,132,368,132,367,131,346,131,345,132,342,132,341,133,332,133']), (946711423, 2096875722, 631, 198, 395, 118, 142, 0.9699756, 1947740378, ['328,137,251,137,250,136,249,137,241,137,240,136,219,136,218,135,213,135,212,134,206,133,205,132,201,131,200,130,200,122,201,121,205,121,206,122,222,122,226,124,239,124,240,125,369,125,370,124,371,125,389,125,391,127,391,133,390,134,386,134,385,135,380,135,379,134,375,134,374,135,341,135,340,136,329,136']), (946711423, 499500794, 631, 93, 107, 127, 146, 0.9574813, 1947740379, ['101,143,98,143,95,139,95,131,97,129,100,129,101,133,102,134,102,136,103,137,103,140']), (946711423, 492925064, 631, 71, 125, 36, 95, 0.95296955, 1947740380, ['104,92,96,92,93,90,91,90,86,86,83,85,83,84,81,82,80,82,75,77,75,75,74,74,74,66,75,65,75,62,77,60,77,58,80,55,80,54,83,51,83,50,88,45,94,44,95,43,99,43,100,42,113,42,117,45,117,47,116,48,116,51,115,52,114,59,113,60,112,65,111,66,111,69,110,70,110,75,109,76,109,83,108,84,108,86,109,87,108,89']), (946711423, 492925064, 631, 101, 167, 38, 127, 0.9508439, 1947740381, ['154,117,152,115,152,112,150,110,148,106,148,104,145,101,143,100,138,100,137,99,135,99,133,95,131,95,126,93,126,91,128,88,128,83,129,82,129,70,127,68,127,66,128,65,125,61,127,59,127,56,129,52,129,49,130,47,135,42,144,42,148,45,151,49,151,60,152,61,152,75,153,76,153,80,155,83,155,87,156,88,156,105,155,106,156,107,156,110,154,112,156,116', '109,100,108,100,107,99,109,97']), (946711423, 492624020, 631, 249, 400, 219, 316, 0.8792459, 1947740382, ['395,313,390,313,386,311,384,312,381,312,376,309,358,308,357,307,354,307,353,306,350,306,349,307,345,305,343,303,341,304,337,304,334,302,325,302,324,301,315,300,313,298,313,297,310,295,304,295,300,293,295,293,291,288,289,287,283,287,281,285,281,283,278,280,274,280,272,279,272,276,270,273,270,270,269,268,266,265,265,265,264,264,264,262,261,260,260,258,260,256,261,255,259,252,259,248,258,246,255,244,256,241,252,239,251,238,251,226,265,226,266,227,268,227,272,232,276,232,277,233,279,233,281,234,283,237,285,238,290,239,293,241,296,241,304,246,312,247,316,251,318,252,320,252,326,255,328,255,337,260,342,260,343,261,345,261,349,264,351,264,355,266,357,268,364,271,366,273,370,275,374,275,376,276,377,279,379,281,383,282,384,283,386,283,387,286,390,289,390,290,394,294,396,294,398,296,398,308,397,309,397,311']), (946711423, 503548896, 631, 302, 540, 339, 403, 0.7406652, 1947740386, ['442,401,372,401,372,397,370,395,369,392,366,390,367,389,366,386,357,386,354,384,350,384,349,383,320,383,319,382,320,378,318,376,318,374,314,370,309,370,308,369,306,369,305,363,305,357,306,356,306,353,307,353,308,354,313,354,314,355,315,354,320,354,321,353,331,353,332,354,335,354,336,355,339,355,340,356,379,356,380,357,406,357,407,356,409,356,410,357,474,357,475,356,482,356,484,357,485,356,488,356,493,353,501,354,502,353,506,353,507,352,517,352,518,351,522,351,525,347,527,346,530,347,530,349,533,351,530,355,528,355,527,356,515,356,509,359,508,361,505,362,503,365,497,368,494,372,490,373,489,374,492,376,495,376,493,377,488,378,490,380,495,380,497,381,497,381,487,382,485,385,476,387,469,392,466,392,465,393,460,393,456,396,453,397,451,399,443,400', '519,353,518,352,517,353,518,354']), (946711423, 2106233860, 631, 53, 85, 75, 182, 0.73015845, 1947740387, ['70,147,68,145,65,139,65,137,62,132,61,128,57,126,56,124,56,121,54,119,54,110,56,108,56,103,59,100,60,101,61,100,61,96,63,93,63,89,66,84,65,83,65,80,66,78,68,78,68,79,70,80,70,83,74,87,74,90,75,91,75,100,77,102,77,105,75,106,75,125,76,126,77,125,77,125,77,128,76,129,77,131,77,136,75,139,75,143,78,145,76,146,71,146', '61,107,60,106,59,107,60,108', '77,134,76,131,75,134,76,135']), (946711423, 2096875717, 631, 477, 510, 220, 243, 0.69028217, 1947740388, ['501,241,493,241,489,239,488,237,487,237,480,232,479,230,479,226,484,222,487,222,488,223,492,224,496,228,497,228,497,229,502,234,502,235,504,236,504,240,502,240']), (946711423, 2096875712, 631, 309, 326, 382, 404, 0.6633776, 1947740390, ['309,383,309,382,311,382', '325,385,324,383,319,383,318,382,325,382', '320,400,311,400,309,398,309,385,310,386,311,385,315,385,316,384,318,384,319,385,322,385,325,387,325,398,323,398']), (946711423, 2096875719, 631, 427, 553, 258, 315, 0.6446218, 1947740391, ['531,284,526,284,525,283,525,281,523,279,522,280,519,281,521,282,521,283,519,284,518,283,519,281,515,279,516,277,515,276,513,276,512,277,513,279,511,280,507,279,505,276,504,279,497,279,496,278,495,279,485,279,484,280,480,280,479,281,481,283,482,283,482,283,469,283,468,284,438,284,436,283,440,279,446,279,447,278,456,278,458,276,457,275,457,275,467,275,468,274,490,274,491,273,494,273,496,271,496,269,499,268,501,266,503,265,506,265,510,263,514,263,516,261,520,259,544,259,544,259,543,260,545,262,548,262,550,263,550,276,549,277,547,277,540,282,538,282,537,283,532,283']), (946711423, 2106233861, 631, 144, 267, 181, 307, 0.63958377, 1947740392, ['212,251,209,251,208,250,203,251,201,250,201,249,195,243,189,242,188,241,185,241,184,240,182,236,180,236,179,235,173,235,172,234,170,235,164,235,163,234,163,232,162,231,163,217,166,217,168,218,170,215,171,210,172,209,173,209,176,212,178,210,178,208,181,203,186,203,188,201,193,201,194,202,195,201,201,201,202,202,204,202,205,201,209,201,210,202,212,202,215,200,217,200,220,202,221,201,227,201,231,205,231,206,234,209,235,209,235,210,238,213,238,224,234,228,235,232,234,233,228,234,225,237,224,241,222,242,216,242,209,246,211,248,212,248,213,250', '221,228,220,227,219,228,220,229', '224,238,224,237,221,235,217,237,219,239']), (946711423, 2096875712, 631, 285, 433, 343, 377, 0.61493844, 1947740393, ['431,376,286,376,285,375,285,368,286,367,286,362,287,361,287,359,291,359,297,362,306,363,307,364,312,364,313,365,322,366,323,367,331,366,332,368,334,368,335,369,338,368,337,366,336,366,337,365,336,364,327,364,325,361,319,361,317,357,317,356,318,355,325,355,326,353,331,353,333,351,332,350,330,350,328,348,326,348,325,347,319,347,315,345,306,345,305,344,297,344,299,344,300,343,431,343,432,344,432,353,431,354,431,358,430,359,430,365,427,365,425,363,424,363,422,364,421,366,418,366,413,369,404,370,409,371,409,371,399,371,398,372,395,373,419,374,420,373,426,372,428,370,428,367,429,367,430,368,429,369,430,370,430,373,431,374', '381,373,378,372,377,371,356,371,356,371,359,370,347,369,345,367,343,367,342,368,343,369,341,370,354,371,354,371,352,372,353,373,359,373,360,374']), (946711423, 2106233860, 631, 146, 287, 140, 311, 0.54784286, 1947740394, ['234,254,227,254,221,251,219,248,215,253,212,253,210,252,206,247,203,247,198,243,197,243,194,239,189,238,186,236,182,236,181,235,167,235,164,233,164,228,159,227,158,226,158,219,159,218,159,213,162,207,162,205,169,192,169,186,170,185,172,185,177,179,175,175,173,173,177,171,181,171,182,170,184,170,187,167,187,164,188,163,188,161,199,161,202,164,205,165,207,167,209,167,212,165,215,165,216,168,218,170,219,170,221,168,221,164,220,163,220,161,222,161,223,160,230,160,231,159,242,159,244,158,247,161,248,161,247,162,246,168,248,172,248,174,253,176,254,180,253,182,249,182,247,185,249,188,253,188,254,189,254,194,249,194,247,196,247,198,249,200,252,200,253,199,255,202,255,205,254,206,254,208,250,207,249,206,246,209,246,210,249,214,252,212,254,212,254,214,255,215,255,217,254,218,254,221,252,221,249,219,247,221,247,225,249,228,250,228,252,226,253,224,253,224,253,229,252,229,251,228,249,228,247,230,247,233,246,234,246,237,245,238,245,240,243,244,243,247,239,251,237,251', '230,167,229,166,227,167,228,168']), (946711423, 495920967, 631, 202, 524, 112, 333, 0.45109355, 1947740396, ['483,289,483,286,482,285,482,283,480,279,480,274,476,270,472,268,465,268,464,269,459,269,458,268,454,268,453,267,437,267,436,268,428,268,427,269,418,269,417,270,414,270,410,266,410,265,416,262,418,262,421,260,423,260,425,259,426,257,424,255,422,255,419,253,417,253,416,252,412,251,410,250,410,249,412,249,413,248,415,248,416,247,422,246,428,243,429,242,428,241,424,240,423,239,420,239,419,238,390,238,389,237,386,237,385,236,369,236,368,235,363,234,363,233,364,232,364,230,366,226,365,225,357,220,344,220,341,218,339,218,339,218,342,212,342,210,336,207,327,207,326,206,319,206,318,205,314,205,313,204,297,204,291,207,288,210,288,212,291,217,290,220,288,222,284,224,282,224,278,227,273,228,271,230,270,235,265,239,262,236,261,232,263,228,266,226,261,224,256,219,256,210,249,206,242,205,237,202,234,195,226,186,227,184,227,180,228,179,225,175,225,174,222,171,225,165,227,163,229,158,230,157,232,156,235,156,236,155,239,155,240,154,245,154,246,155,254,155,255,156,258,156,259,157,268,157,269,156,272,156,273,155,280,155,281,156,298,156,300,155,301,156,307,156,308,157,311,157,318,152,322,151,323,150,333,150,338,146,339,146,342,143,343,143,346,140,357,140,362,136,366,134,368,134,369,133,373,133,374,132,377,132,378,131,388,131,389,130,410,130,411,131,417,131,428,140,432,142,434,142,435,143,443,145,446,147,448,147,451,154,453,156,457,158,462,159,463,160,466,160,467,161,472,162,474,163,474,164,481,171,489,175,491,175,492,176,494,176,495,177,499,178,500,179,502,184,507,189,517,194,518,195,514,201,514,203,518,207,519,209,518,214,515,218,517,227,515,229,515,231,514,232,515,236,518,239,518,252,519,253,519,263,518,264,518,267,517,269,514,272,512,273,512,274,506,280,500,277,498,273,496,272,493,272,491,274,491,278,490,279,490,281', '312,179,311,178,308,179,309,180', '268,269,264,269,259,266,259,262,261,258,261,250,265,245,269,250,270,257,274,260,278,265,275,267,269,268', '414,281,401,281,414,281']), (946711423, 2096875722, 631, 433, 558, 248, 286, 0.44133398, 1947740397, ['492,272,474,272,473,271,468,271,465,269,460,269,460,268,465,266,467,266,468,265,470,265,471,264,475,264,476,263,479,263,480,262,486,262,487,261,491,261,492,260,495,260,496,259,502,259,506,257,510,257,514,255,517,255,518,254,530,253,531,252,535,252,536,251,538,251,539,252,543,252,544,253,547,253,549,251,553,251,555,253,555,267,552,270,550,270,550,269,548,267,547,267,547,267,548,266,547,265,545,266,540,266,539,264,530,264,529,263,524,263,519,266,513,266,510,268,507,268,506,269,499,270,498,271,493,271', '438,279,435,279,435,273,436,272,448,271,449,272,448,274,443,274,440,277,440,278']), (946711423, 492654799, 631, 399, 569, 68, 251, 0.41876298, 1947740399, []), (946711423, 492624020, 631, 420, 552, 244, 293, 0.35962066, 1947740400, ['474,289,453,289,452,288,439,288,437,286,431,286,427,284,423,284,422,283,422,275,427,275,428,273,430,272,435,272,436,271,438,271,442,269,447,269,450,267,454,267,460,264,464,264,467,262,483,261,484,260,488,260,489,259,494,259,495,258,502,258,503,257,505,257,509,255,512,255,516,252,520,252,521,251,526,250,530,248,534,248,535,247,546,247,547,248,549,248,549,250,550,251,550,266,551,267,551,275,550,276,550,278,549,279,549,281,537,282,535,284,528,284,527,285,504,285,503,286,495,286,492,288,488,287,487,288,475,288']), (946711423, 503548896, 631, 301, 540, 339, 403, 0.740756, 3140491551, ['442,401,371,401,371,397,366,390,365,386,356,386,353,384,348,383,319,383,319,378,314,370,310,370,305,368,304,357,305,353,330,353,339,356,378,356,379,357,474,357,475,356,488,356,493,353,501,354,507,352,517,352,522,351,527,346,530,347,533,351,530,355,527,356,515,356,505,362,503,365,497,368,494,372,489,374,492,376,488,378,490,380,495,380,487,382,485,385,476,387,469,392,461,393,456,395,451,399,447,399', '519,353,518,352,517,353,518,354'])],)} test detection filter by crop is a success ! ############################### TEST detection_filter_by_classif ################################ t SELECT id FROM MTRPhoto.crop_hashtag_ids WHERE photo_id=946711423 AND `type`=816 DELETE FROM MTRPhoto.crop_hashtag_ids WHERE id IN (3815277467,3815277466,3815277465,3815277474,3815277473,3815277472,3815277471,3815277470,3815277479,3815277482,3815277468,3815277469,3815277478,3815277477,3815277483,3815277484,3815277485,3815277487,3815277475,3815277481,3815277480,3815277486,3815277476) Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=672 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=672 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 672 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=672 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : detection_filter_by_classif list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (946711423) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos ##### After load_data_input time to download the photos : 0.004637479782104492 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:detection_filter_by_classif Fri May 30 03:39:02 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {} map_photo_id_path_extension : {} map_subphoto_mainphoto : {} beginning of step detection filter with classification results param_json : {'input_type': 631, 'output_type': 816, 'condition_type': 872, 'crops_ok': {'CAR_DOCUMENT.*': {}, 'CAR_INTERIEUR.*': {}, 'CAR_EXTERIEUR_angle_avant_droit.*': {'Retroviseur': 2, 'Roue': 2, 'Capot': 1, 'Pare-brise': 1, 'vitre': 10, 'phare': 2, 'Feu-antibrouillard': 2, 'poignee': 2, 'porte': 2, 'calandre': 1, 'logo-marque': 1, 'Plaque-immatriculation': 1, 'Essuie-glace': 1, 'pare-choc': 1, 'toit': 1, 'logo-roue': 1, 'aile-avant': 1}}, 'separation': {'CAR_EXTERIEUR_avant.*': {'pare-choc': ['pare-chocs-avant'], 'phare': ['phare-gauche', 'a-droite-de', 'phare-droit']}, 'CAR_EXTERIEUR_angle_avant_droit.*': {'pare-choc': ['pare-chocs-avant'], 'phare': ['phare-droite', 'a-gauche-de', 'phare-gauche'], 'porte': ['porte-avant', 'a-droite-de', 'porte-arriere']}}} conditional_crop_by_classif_copy select distinct hashtag_id from MTRBack.photo_hashtag_ids where photo_id in (946711423) and type=872 batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 946711423) and `type` in (631) Loaded 35 chid ids of type : 631 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1947740368,1947740369,1947740370,1947740371,1947740372,1947740373,1947740374,1947740375,1947740376,1947740377,1947740378,1947740379,1947740380,1947740381,1947740382,1947740383,1947740384,1947740385,1947740386,1947740387,1947740388,1947740389,1947740390,1947740391,1947740392,1947740393,1947740394,1947740395,1947740396,1947740397,1947740398,1947740399,1947740400,3140491551,3140491552) treating photo 946711423 select distinct hashtag_id from MTRBack.photo_hashtag_ids where photo_id in (946711423) and type=872 list of crops kept {'retroviseur': 2, 'roue': 2, 'capot': 1, 'pare-brise': 1, 'vitre': 10, 'phare': 2, 'feu-antibrouillard': 2, 'poignee': 2, 'porte': 2, 'calandre': 1, 'logo-marque': 1, 'plaque-immatriculation': 1, 'essuie-glace': 1, 'pare-choc': 1, 'toit': 1, 'logo-roue': 1, 'aile-avant': 1} for hahstag car_exterieur_angle_avant_droit_merge__port_551052 crop not duplicated for hashtag aile-arriere : : {'photo_id': 946711423, 'hashtag_id': 2106233861, 'type': 631, 'x0': 144, 'x1': 267, 'y0': 181, 'y1': 307, 'score': 0.63958377, 'id': 1947740392, 'points': ['212,251,209,251,208,250,203,251,201,250,201,249 crop not duplicated for hashtag coffre : : {'photo_id': 946711423, 'hashtag_id': 495920967, 'type': 631, 'x0': 202, 'x1': 524, 'y0': 112, 'y1': 333, 'score': 0.45109355, 'id': 1947740396, 'points': ['483,289,483,286,482,285,482,283,480,279,480,274, crop not duplicated for hashtag aile-arriere : : {'photo_id': 946711423, 'hashtag_id': 2106233861, 'type': 631, 'x0': 535, 'x1': 630, 'y0': 138, 'y1': 231, 'score': 0.42747068, 'id': 1947740398, 'points': ['590,171,589,170,585,170,584,171,581,169,579,169 crop duplicated for hashtag retroviseur : : {'photo_id': 946711423, 'hashtag_id': 492844413, 'type': 631, 'x0': 89, 'x1': 163, 'y0': 93, 'y1': 144, 'score': 0.9772748, 'id': 1947740375, 'points': ['159,142,153,141,151,139,148,138,145,135,141,133,139 crop duplicated for hashtag roue : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 162, 'x1': 245, 'y0': 233, 'y1': 396, 'score': 0.99702626, 'id': 1947740369, 'points': ['215,393,206,393,202,390,200,390,192,383,191,380, crop duplicated for hashtag roue : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 545, 'x1': 612, 'y0': 186, 'y1': 276, 'score': 0.9876676, 'id': 1947740371, 'points': ['584,267,583,266,578,266,574,262,574,259,573,258,5 crop not duplicated for hashtag roue : : {'photo_id': 946711423, 'hashtag_id': 492689227, 'type': 631, 'x0': 53, 'x1': 87, 'y0': 127, 'y1': 212, 'score': 0.9786105, 'id': 1947740374, 'points': ['74,201,69,201,67,199,66,199,65,198,62,192,62,190,61 crop duplicated for hashtag capot : : {'photo_id': 946711423, 'hashtag_id': 599722655, 'type': 631, 'x0': 176, 'x1': 535, 'y0': 138, 'y1': 264, 'score': 0.9818268, 'id': 1947740373, 'points': ['453,253,413,253,412,252,387,252,386,250,386,248,3 crop duplicated for hashtag pare-brise : : {'photo_id': 946711423, 'hashtag_id': 2096875709, 'type': 631, 'x0': 185, 'x1': 431, 'y0': 39, 'y1': 136, 'score': 0.97171515, 'id': 1947740377, 'points': ['331,134,287,134,286,133,284,133,283,134,272,134, crop duplicated for hashtag vitre : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 418, 'x1': 522, 'y0': 69, 'y1': 136, 'score': 0.97407305, 'id': 1947740376, 'points': ['510,121,507,121,505,119,501,120,500,119,500,113,4 crop duplicated for hashtag vitre : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 71, 'x1': 125, 'y0': 36, 'y1': 95, 'score': 0.95296955, 'id': 1947740380, 'points': ['104,92,96,92,93,90,91,90,86,86,83,85,83,84,81,82,80 crop duplicated for hashtag vitre : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 101, 'x1': 167, 'y0': 38, 'y1': 127, 'score': 0.9508439, 'id': 1947740381, 'points': ['154,117,152,115,152,112,150,110,148,106,148,104,14 crop duplicated for hashtag vitre : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 547, 'x1': 640, 'y0': 79, 'y1': 129, 'score': 0.8165246, 'id': 1947740384, 'points': ['630,96,627,96,627,94,628,92,629,92,631,94,631,95', crop duplicated for hashtag vitre : : {'photo_id': 946711423, 'hashtag_id': 492925064, 'type': 631, 'x0': 360, 'x1': 434, 'y0': 62, 'y1': 116, 'score': 0.74684095, 'id': 1947740385, 'points': ['415,103,413,103,411,101,408,101,405,99,403,99,401 crop duplicated for hashtag phare : : {'photo_id': 946711423, 'hashtag_id': 492624020, 'type': 631, 'x0': 249, 'x1': 400, 'y0': 219, 'y1': 316, 'score': 0.8792459, 'id': 1947740382, 'points': ['395,313,390,313,386,311,384,312,381,312,376,309,3 crop duplicated for hashtag phare : : {'photo_id': 946711423, 'hashtag_id': 492624020, 'type': 631, 'x0': 420, 'x1': 552, 'y0': 244, 'y1': 293, 'score': 0.35962066, 'id': 1947740400, 'points': ['474,289,453,289,452,288,439,288,437,286,431,286, crop duplicated for hashtag feu-antibrouillard : : {'photo_id': 946711423, 'hashtag_id': 2096875712, 'type': 631, 'x0': 309, 'x1': 326, 'y0': 382, 'y1': 404, 'score': 0.6633776, 'id': 1947740390, 'points': ['309,383,309,382,311,382', '325,385,324,383,319,3 crop duplicated for hashtag feu-antibrouillard : : {'photo_id': 946711423, 'hashtag_id': 2096875712, 'type': 631, 'x0': 285, 'x1': 433, 'y0': 343, 'y1': 377, 'score': 0.61493844, 'id': 1947740393, 'points': ['431,376,286,376,285,375,285,368,286,367,286,362 crop duplicated for hashtag poignee : : {'photo_id': 946711423, 'hashtag_id': 499500794, 'type': 631, 'x0': 93, 'x1': 107, 'y0': 127, 'y1': 146, 'score': 0.9574813, 'id': 1947740379, 'points': ['101,143,98,143,95,139,95,131,97,129,100,129,101,13 crop duplicated for hashtag porte : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 96, 'x1': 172, 'y0': 39, 'y1': 261, 'score': 0.9928518, 'id': 1947740370, 'points': ['143,252,143,249,141,246,140,246,138,248,138,251,137 crop duplicated for hashtag porte : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 540, 'x1': 625, 'y0': 78, 'y1': 221, 'score': 0.87864035, 'id': 1947740383, 'points': ['567,127,566,127,565,126,564,109,562,106,560,106,5 crop not duplicated for hashtag porte : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 61, 'x1': 115, 'y0': 42, 'y1': 188, 'score': 0.6900027, 'id': 1947740389, 'points': ['92,47,91,45,92,44,96,45,94,45', '73,141,73,136,72,1 crop not duplicated for hashtag porte : : {'photo_id': 946711423, 'hashtag_id': 492654799, 'type': 631, 'x0': 399, 'x1': 569, 'y0': 68, 'y1': 251, 'score': 0.41876298, 'id': 1947740399, 'points': [], 'sub_photo_id': 0, 'rles': [], 'hashtag': '', ' crop duplicated for hashtag calandre : : {'photo_id': 946711423, 'hashtag_id': 503548896, 'type': 631, 'x0': 301, 'x1': 540, 'y0': 339, 'y1': 403, 'score': 0.740756, 'id': 3140491551, 'points': ['442,401,371,401,371,397,366,390,365,386,356,386,35 crop not duplicated for hashtag calandre : : {'photo_id': 946711423, 'hashtag_id': 503548896, 'type': 631, 'x0': 302, 'x1': 540, 'y0': 339, 'y1': 403, 'score': 0.7406652, 'id': 1947740386, 'points': ['442,401,372,401,372,397,370,395,369,392,366,390,3 crop duplicated for hashtag logo-marque : : {'photo_id': 946711423, 'hashtag_id': 2096875717, 'type': 631, 'x0': 477, 'x1': 510, 'y0': 220, 'y1': 243, 'score': 0.69028217, 'id': 1947740388, 'points': ['501,241,493,241,489,239,488,237,487,237,480,232 crop duplicated for hashtag plaque-immatriculation : : {'photo_id': 946711423, 'hashtag_id': 2096875719, 'type': 631, 'x0': 468, 'x1': 555, 'y0': 292, 'y1': 365, 'score': 0.9830025, 'id': 1947740372, 'points': ['491,350,489,350,488,349,487,350,483,350,480,348, crop not duplicated for hashtag plaque-immatriculation : : {'photo_id': 946711423, 'hashtag_id': 2096875719, 'type': 631, 'x0': 427, 'x1': 553, 'y0': 258, 'y1': 315, 'score': 0.6446218, 'id': 1947740391, 'points': ['531,284,526,284,525,283,525,281,523,279,522,280, crop duplicated for hashtag essuie-glace : : {'photo_id': 946711423, 'hashtag_id': 2096875722, 'type': 631, 'x0': 198, 'x1': 395, 'y0': 118, 'y1': 142, 'score': 0.9699756, 'id': 1947740378, 'points': ['328,137,251,137,250,136,249,137,241,137,240,136, crop not duplicated for hashtag essuie-glace : : {'photo_id': 946711423, 'hashtag_id': 2096875722, 'type': 631, 'x0': 433, 'x1': 558, 'y0': 248, 'y1': 286, 'score': 0.44133398, 'id': 1947740397, 'points': ['492,272,474,272,473,271,468,271,465,269,460,269 crop duplicated for hashtag pare-choc : : {'photo_id': 946711423, 'hashtag_id': 624624117, 'type': 631, 'x0': 226, 'x1': 569, 'y0': 252, 'y1': 425, 'score': 0.99812776, 'id': 1947740368, 'points': ['395,419,341,419,340,418,316,418,315,417,306,417, crop duplicated for hashtag aile-avant : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 53, 'x1': 85, 'y0': 75, 'y1': 182, 'score': 0.73015845, 'id': 1947740387, 'points': ['70,147,68,145,65,139,65,137,62,132,61,128,57,126,5 crop not duplicated for hashtag aile-avant : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 146, 'x1': 287, 'y0': 140, 'y1': 311, 'score': 0.54784286, 'id': 1947740394, 'points': ['234,254,227,254,221,251,219,248,215,253,212,253 crop not duplicated for hashtag aile-avant : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 496, 'x1': 624, 'y0': 140, 'y1': 245, 'score': 0.4627206, 'id': 3140491552, 'points': ['595,176,593,176,589,173,586,176,583,176,580,174, crop not duplicated for hashtag aile-avant : : {'photo_id': 946711423, 'hashtag_id': 2106233860, 'type': 631, 'x0': 496, 'x1': 624, 'y0': 141, 'y1': 245, 'score': 0.46262404, 'id': 1947740395, 'points': ['603,176,599,173,595,176,593,176,590,174,589,174 list of crops kept {'pare-choc': ('pare-chocs-avant',), 'phare': ('phare-droite', 'a-gauche-de', 'phare-gauche'), 'porte': ('porte-avant', 'a-droite-de', 'porte-arriere')} for hahstag car_exterieur_angle_avant_droit_merge__port_551052 batch 1 Loaded 0 chid ids of type : 0 insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) batch 1 Loaded 23 chid ids of type : 816 INSERT IGNORE INTO MTRPhoto.crop_polygon_points (`crop_hashtag_id`, `points`) VALUES (%s, %s) Number RLEs to save : 1600 INSERT IGNORE INTO MTRPhoto.crop_segments (`crop_hashtag_id`, `x0`, `y0`, `length`) VALUES (%s, %s, %s , %s) first line : ('3815455012', '117', '95', '16') ... last line : ('3815455034', '70', '147', '1') INSERT IGNORE INTO MTRPhoto.crop_sum_segments (`crop_hashtag_id`, `sum_segments`) VALUES (%s, %s) TO DO : save crop sub photo not yet done ! After datou_step_exec type output : time spend for datou_step_exec : 0.3154714107513428 time spend to save output : 0.00015234947204589844 total time spend for step 1 : 0.31562376022338867 caffe_path_current : About to save ! 0 After save, about to update current ! test detection filter by classif is a success ! ############################### TEST blur_detection ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=1243 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=1243 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 1243 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=1243 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : blur_detection list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (930729675) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 930729675 download finish for photo 930729675 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.14591383934020996 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:blur_detection Fri May 30 03:39:03 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569143_1112947_930729675_b2d2beaaee733d521cbb0c9800a29073.jpg': 930729675} map_photo_id_path_extension : {930729675: {'path': 'temp/1748569143_1112947_930729675_b2d2beaaee733d521cbb0c9800a29073.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} inside step blur_detection methode: ratio et variance treat image : temp/1748569143_1112947_930729675_b2d2beaaee733d521cbb0c9800a29073.jpg resize: (600, 800) 930729675 12.961859636534896 score_blur_detection : {930729675: [(930729675, 12.961859636534896, 492688767)]} After datou_step_exec type output : time spend for datou_step_exec : 0.2008354663848877 time spend to save output : 3.814697265625e-05 total time spend for step 1 : 0.20087361335754395 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {930729675: [(930729675, 12.961859636534896, 492688767)]} {930729675: [(930729675, 12.961859636534896, 492688767)]} ############################### TEST detect_point_224x224 ################################ test_detect_point_224x224 Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=1908 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=1908 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 1908 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=1908 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 4589 thcl is not linked in the step_by_step architecture ! WARNING : step 4590 argmax is not linked in the step_by_step architecture ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : thcl, argmax list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (987515175,987515176,987515177,987515178,987515179,987515180,987515181,987515182,987515183,987515184,987515185,987515186,987515187,987515188,987515189,987515190,987515192,987515193,987515195,987515196,987515198,987515200,987515201,987515202,987515204,987515205,987515207,987515208,987515209,987515211,987515212,987515213,987515215,987515216,987515217,987515219,987515220,987515222,987515223,987515224,987515226,987515227,987515228,987515230,987515231,987515232,987515233,987515234,987515235,987515236,987515237,987515238,987515239,987515240,987515241,987515242,987515243,987515244,987515245,987515246,987515247,987515248,987515249,987515250) Found this number of photos: 64 ##### Call download_photos : nb_thread : 5 begin to download photo : 987515175 begin to download photo : 987515188 begin to download photo : 987515207 begin to download photo : 987515224 begin to download photo : 987515239 download finish for photo 987515239 begin to download photo : 987515240 download finish for photo 987515224 begin to download photo : 987515226 download finish for photo 987515207 begin to download photo : 987515208 download finish for photo 987515175 begin to download photo : 987515176 download finish for photo 987515188 begin to download photo : 987515189 download finish for photo 987515189 begin to download photo : 987515190 download finish for photo 987515240 begin to download photo : 987515241 download finish for photo 987515226 begin to download photo : 987515227 download finish for photo 987515208 begin to download photo : 987515209 download finish for photo 987515176 begin to download photo : 987515177 download finish for photo 987515190 begin to download photo : 987515192 download finish for photo 987515227 begin to download photo : 987515228 download finish for photo 987515241 download finish for photo 987515209 begin to download photo : 987515242 begin to download photo : 987515211 download finish for photo 987515228 begin to download photo : 987515230 download finish for photo 987515192 begin to download photo : 987515193 download finish for photo 987515177 begin to download photo : 987515178 download finish for photo 987515211 begin to download photo : 987515212 download finish for photo 987515242 begin to download photo : 987515243 download finish for photo 987515230 begin to download photo : 987515231 download finish for photo 987515193 begin to download photo : 987515195 download finish for photo 987515178 begin to download photo : 987515179 download finish for photo 987515243 begin to download photo : 987515244 download finish for photo 987515212 begin to download photo : 987515213 download finish for photo 987515244 begin to download photo : 987515245 download finish for photo 987515195 begin to download photo : 987515196 download finish for photo 987515179 begin to download photo : 987515180 download finish for photo 987515231 begin to download photo : 987515232 download finish for photo 987515196 begin to download photo : 987515198 download finish for photo 987515232 begin to download photo : 987515233 download finish for photo 987515213 begin to download photo : 987515215 download finish for photo 987515245 begin to download photo : 987515246 download finish for photo 987515198 begin to download photo : 987515200 download finish for photo 987515233 begin to download photo : 987515234 download finish for photo 987515246 begin to download photo : 987515247 download finish for photo 987515215 begin to download photo : 987515216 download finish for photo 987515234 begin to download photo : 987515235 download finish for photo 987515180 begin to download photo : 987515181 download finish for photo 987515216 begin to download photo : 987515217 download finish for photo 987515247 begin to download photo : 987515248 download finish for photo 987515200 begin to download photo : 987515201 download finish for photo 987515248 begin to download photo : 987515249 download finish for photo 987515181 begin to download photo : 987515182 download finish for photo 987515235 begin to download photo : 987515236 download finish for photo 987515201 begin to download photo : 987515202 download finish for photo 987515217 begin to download photo : 987515219 download finish for photo 987515236 begin to download photo : 987515237 download finish for photo 987515249 begin to download photo : 987515250 download finish for photo 987515182 begin to download photo : 987515183 download finish for photo 987515219 begin to download photo : 987515220 download finish for photo 987515202 begin to download photo : 987515204 download finish for photo 987515250 download finish for photo 987515237 begin to download photo : 987515238 download finish for photo 987515204 begin to download photo : 987515205 download finish for photo 987515220 begin to download photo : 987515222 download finish for photo 987515183 begin to download photo : 987515184 download finish for photo 987515238 download finish for photo 987515222 begin to download photo : 987515223 download finish for photo 987515205 download finish for photo 987515184 begin to download photo : 987515185 download finish for photo 987515223 download finish for photo 987515185 begin to download photo : 987515186 download finish for photo 987515186 begin to download photo : 987515187 download finish for photo 987515187 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 64 ; length of list_pids : 64 ; length of list_args : 64 ##### After load_data_input time to download the photos : 1.7604339122772217 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 2 step1:thcl Fri May 30 03:39:05 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg': 987515239, 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg': 987515240, 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg': 987515241, 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg': 987515242, 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg': 987515243, 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg': 987515244, 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg': 987515245, 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg': 987515246, 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg': 987515247, 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg': 987515248, 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg': 987515249, 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg': 987515250, 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg': 987515224, 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg': 987515226, 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg': 987515227, 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg': 987515228, 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg': 987515230, 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg': 987515231, 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg': 987515232, 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg': 987515233, 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg': 987515234, 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg': 987515235, 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg': 987515236, 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg': 987515237, 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg': 987515238, 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg': 987515188, 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg': 987515189, 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg': 987515190, 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg': 987515192, 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg': 987515193, 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg': 987515195, 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg': 987515196, 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg': 987515198, 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg': 987515200, 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg': 987515201, 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg': 987515202, 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg': 987515204, 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg': 987515205, 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg': 987515207, 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg': 987515208, 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg': 987515209, 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg': 987515211, 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg': 987515212, 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg': 987515213, 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg': 987515215, 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg': 987515216, 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg': 987515217, 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg': 987515219, 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg': 987515220, 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg': 987515222, 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg': 987515223, 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg': 987515175, 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg': 987515176, 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg': 987515177, 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg': 987515178, 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg': 987515179, 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg': 987515180, 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg': 987515181, 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg': 987515182, 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg': 987515183, 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg': 987515184, 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg': 987515185, 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg': 987515186, 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg': 987515187} map_photo_id_path_extension : {987515239: {'path': 'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg', 'extension': 'jpg'}, 987515240: {'path': 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg', 'extension': 'jpg'}, 987515241: {'path': 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg', 'extension': 'jpg'}, 987515242: {'path': 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg', 'extension': 'jpg'}, 987515243: {'path': 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg', 'extension': 'jpg'}, 987515244: {'path': 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg', 'extension': 'jpg'}, 987515245: {'path': 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg', 'extension': 'jpg'}, 987515246: {'path': 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg', 'extension': 'jpg'}, 987515247: {'path': 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg', 'extension': 'jpg'}, 987515248: {'path': 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg', 'extension': 'jpg'}, 987515249: {'path': 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg', 'extension': 'jpg'}, 987515250: {'path': 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg', 'extension': 'jpg'}, 987515224: {'path': 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg', 'extension': 'jpg'}, 987515226: {'path': 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg', 'extension': 'jpg'}, 987515227: {'path': 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg', 'extension': 'jpg'}, 987515228: {'path': 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg', 'extension': 'jpg'}, 987515230: {'path': 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg', 'extension': 'jpg'}, 987515231: {'path': 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg', 'extension': 'jpg'}, 987515232: {'path': 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg', 'extension': 'jpg'}, 987515233: {'path': 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg', 'extension': 'jpg'}, 987515234: {'path': 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg', 'extension': 'jpg'}, 987515235: {'path': 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg', 'extension': 'jpg'}, 987515236: {'path': 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg', 'extension': 'jpg'}, 987515237: {'path': 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg', 'extension': 'jpg'}, 987515238: {'path': 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg', 'extension': 'jpg'}, 987515188: {'path': 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg', 'extension': 'jpg'}, 987515189: {'path': 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg', 'extension': 'jpg'}, 987515190: {'path': 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg', 'extension': 'jpg'}, 987515192: {'path': 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg', 'extension': 'jpg'}, 987515193: {'path': 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg', 'extension': 'jpg'}, 987515195: {'path': 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg', 'extension': 'jpg'}, 987515196: {'path': 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg', 'extension': 'jpg'}, 987515198: {'path': 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg', 'extension': 'jpg'}, 987515200: {'path': 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg', 'extension': 'jpg'}, 987515201: {'path': 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg', 'extension': 'jpg'}, 987515202: {'path': 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg', 'extension': 'jpg'}, 987515204: {'path': 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg', 'extension': 'jpg'}, 987515205: {'path': 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg', 'extension': 'jpg'}, 987515207: {'path': 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg', 'extension': 'jpg'}, 987515208: {'path': 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg', 'extension': 'jpg'}, 987515209: {'path': 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg', 'extension': 'jpg'}, 987515211: {'path': 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg', 'extension': 'jpg'}, 987515212: {'path': 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'extension': 'jpg'}, 987515213: {'path': 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'extension': 'jpg'}, 987515215: {'path': 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg', 'extension': 'jpg'}, 987515216: {'path': 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg', 'extension': 'jpg'}, 987515217: {'path': 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg', 'extension': 'jpg'}, 987515219: {'path': 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg', 'extension': 'jpg'}, 987515220: {'path': 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg', 'extension': 'jpg'}, 987515222: {'path': 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg', 'extension': 'jpg'}, 987515223: {'path': 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg', 'extension': 'jpg'}, 987515175: {'path': 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg', 'extension': 'jpg'}, 987515176: {'path': 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg', 'extension': 'jpg'}, 987515177: {'path': 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg', 'extension': 'jpg'}, 987515178: {'path': 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg', 'extension': 'jpg'}, 987515179: {'path': 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg', 'extension': 'jpg'}, 987515180: {'path': 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg', 'extension': 'jpg'}, 987515181: {'path': 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg', 'extension': 'jpg'}, 987515182: {'path': 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg', 'extension': 'jpg'}, 987515183: {'path': 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg', 'extension': 'jpg'}, 987515184: {'path': 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg', 'extension': 'jpg'}, 987515185: {'path': 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg', 'extension': 'jpg'}, 987515186: {'path': 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg', 'extension': 'jpg'}, 987515187: {'path': 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step Thcl ! multi_thcl or not :False multi_thcl_cond or not :False dic_thcl : {'1528': 1} we are using the classfication for only one thcl 1528 In convert_file_to_np l 337 : 1 l343 1 In convert_file_to_np l 337 : 7 l343 7 In convert_file_to_np l 337 : 7 l343 7 In convert_file_to_np l 337 : 7 l343 7 In convert_file_to_np l 337 : 7 l343 7 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.0009925365447998047 time to convert the images to numpy array : 0.005064964294433594 In convert_file_to_np l 337 : 7 l343 7 In convert_file_to_np l 337 : 7 l343 7 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed !In convert_file_to_np l 337 : 7 l343 7 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! In convert_file_to_np l 337 : 7 l343 7 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! In convert_file_to_np l 337 : 7 l343 7 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.006537675857543945 time to convert the images to numpy array : 0.03518104553222656 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.006522655487060547 time to convert the images to numpy array : 0.038722991943359375 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.006957292556762695 time to convert the images to numpy array : 0.03734779357910156 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.00745081901550293 time to convert the images to numpy array : 0.03821206092834473 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.007935047149658203 time to convert the images to numpy array : 0.038150787353515625 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.009185552597045898 time to convert the images to numpy array : 0.03850722312927246 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.009784936904907227 time to convert the images to numpy array : 0.0383143424987793 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.009737730026245117 time to convert the images to numpy array : 0.03890419006347656 l357 after caffe.io.load_image dimension du image : (3, (224, 224, 3)) dimension displayed ! time to import caffe and check if the image exist : 0.016573429107666016 time to convert the images to numpy array : 0.03280806541442871 total time to convert the images to numpy array : 0.05211782455444336 list photo_ids error: [] list photo_ids correct : [987515187, 987515246, 987515247, 987515248, 987515249, 987515250, 987515224, 987515226, 987515239, 987515240, 987515241, 987515242, 987515243, 987515244, 987515245, 987515235, 987515236, 987515237, 987515238, 987515188, 987515189, 987515190, 987515227, 987515228, 987515230, 987515231, 987515232, 987515233, 987515234, 987515202, 987515204, 987515205, 987515207, 987515208, 987515209, 987515211, 987515192, 987515193, 987515195, 987515196, 987515198, 987515200, 987515201, 987515222, 987515223, 987515175, 987515176, 987515177, 987515178, 987515179, 987515212, 987515213, 987515215, 987515216, 987515217, 987515219, 987515220, 987515180, 987515181, 987515182, 987515183, 987515184, 987515185, 987515186] number of photos to traite : 64 try to delete the photos incorrect in DB tagging for thcl : 1528 To do loadFromThcl(), then load ParamDescType : thcl1528 get_desc_type_from_thcl : type of cat SELECT id, mtr_user_id, name, pb_hashtag_id, hashtag_id_list, button_legend_list, portfolio_id_lists, photo_hashtag_type, photo_desc_type, svm_limit, limit_tagging, is_public, live, created_at, updated_at, type_classification FROM MTRDatou.classification_theme WHERE `id` IN (1528) thcls : [{'id': 1528, 'mtr_user_id': 31, 'name': 'learn_refus_upm_blanches_1924', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'Autre_Environement,Carton,Kraft,Lointain_Papier_Magazine,Metal,Papier_Magazine,Plastique,Sol_Environement,Teint_Dans_La_Masse,autre_refus', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 1927, 'photo_desc_type': 4421, 'type_classification': 'caffe', 'hashtag_id_list': '2107752388,492774966,493202403,2107752389,492628673,2107752386,492725882,2107752387,2107752385,2107752406'}] thcl {'id': 1528, 'mtr_user_id': 31, 'name': 'learn_refus_upm_blanches_1924', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'Autre_Environement,Carton,Kraft,Lointain_Papier_Magazine,Metal,Papier_Magazine,Plastique,Sol_Environement,Teint_Dans_La_Masse,autre_refus', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 1927, 'photo_desc_type': 4421, 'type_classification': 'caffe', 'hashtag_id_list': '2107752388,492774966,493202403,2107752389,492628673,2107752386,492725882,2107752387,2107752385,2107752406'} Update svm_hashtag_type_desc : 4421 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (4421) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (4421, 'learn_refus_upm_blanches_1924', 16384, 25088, 'learn_refus_upm_blanches_1924', 'res5b', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2019, 10, 22, 17, 39, 25), datetime.datetime(2019, 10, 22, 17, 39, 25)) To loadFromThcl() : net_4421 begin to check gpu status inside check gpu memory havn't enough memory gpu , need / 2500 l 3632 free memory gpu now : 2171 wait 20 seconds l 3637 free memory gpu now : 2171 max_wait_temp : 1 max_wait : 0 SELECT * FROM MTRDatou.photo_desc_type_params WHERE id in (4421) FOUND : 1 Here is data_from_sql_as_vec to set the ParamDescriptorType : (4421, 'learn_refus_upm_blanches_1924', 16384, 25088, 'learn_refus_upm_blanches_1924', 'res5b', 10.0, None, None, 256, None, 0, None, 8, None, None, -1000.0, 1, datetime.datetime(2019, 10, 22, 17, 39, 25), datetime.datetime(2019, 10, 22, 17, 39, 25)) param : , param.caffemodel : learn_refus_upm_blanches_1924 None mean_file_type : mean_file_path : prototxt_file_path : model : learn_refus_upm_blanches_1924 Inside get_net Inside get_net before cache_data_model model_param file didn't exist Inside get_net before CDM.load_model_par_type model_name : learn_refus_upm_blanches_1924 model_type : caffe list file need : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file exist in s3 : ['caffemodel', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file manque in s3 : ['deploy_conv_normal.prototxt', 'deploy_fc.prototxt'] local folder : /data/models_weight/learn_refus_upm_blanches_1924 /data/models_weight/learn_refus_upm_blanches_1924/caffemodel size_local : 45774543 size in s3 : 45774543 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 caffemodel already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/deploy.prototxt size_local : 17312 size in s3 : 17312 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:03 deploy.prototxt already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/mean.npy size_local : 1572992 size in s3 : 1572992 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:05 mean.npy already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/synset_words.txt size_local : 218 size in s3 : 218 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 synset_words.txt already exist and didn't need to update Inside get_net after CDM.load_model_par_type After if not only_with_local_cache: /home/admin/workarea/install/caffe_cuda8_python3/python/:/home/admin/workarea/install/darknet/:/home/admin/workarea/git/Velours/python:/home/admin/workarea/install/caffe_frcnn_python3/py-faster-rcnn/caffe-fast-rcnn/python:/home/admin/mtr/.credentials:/home/admin/workarea/install/caffe/python:/home/admin/workarea/install/caffe_frcnn/py-faster-rcnn/tools/:/home/admin/workarea/git/fotonowerpip/:/home/admin/workarea/install/segment-anything:/home/admin//workarea/git/pyfvs/ Here before set mode gpu Doing nothing but we could set mode gpu after set mode gpu prototxt_filename : /data/models_weight/learn_refus_upm_blanches_1924/deploy.prototxt caffemodel_filename : /data/models_weight/learn_refus_upm_blanches_1924/caffemodel now we set caffe to gpu mode before predict begin to check gpu status inside check gpu memory havn't enough memory gpu , need / 2500 l 3632 free memory gpu now : 2171 wait 20 seconds l 3637 free memory gpu now : 2171 max_wait_temp : 1 max_wait : 0 dict_keys(['prob', 'res5b']) time used to do the prepocess of the images : 0.05671072006225586 time used to do the prediction : 0.2579522132873535 save descriptor for thcl : 1528 (64, 512, 7, 7) Got the blobs of the net to insert : [3, 1, 1, 1, 1, 1, 2, 1, 1, 2] code_as_byte_string:b'0301010101'| Got the blobs of the net to insert : [0, 0, 0, 1, 0, 0, 0, 0, 1, 3] code_as_byte_string:b'0000000100'| Got the blobs of the net to insert : [0, 0, 0, 1, 2, 0, 1, 4, 0, 2] code_as_byte_string:b'0000000102'| Got the blobs of the net to insert : [4, 1, 3, 4, 7, 3, 4, 2, 1, 0] code_as_byte_string:b'0401030407'| Got the blobs of the net to insert : [4, 1, 3, 4, 7, 3, 4, 2, 1, 0] code_as_byte_string:b'0401030407'| Got the blobs of the net to insert : [3, 3, 1, 4, 7, 7, 9, 6, 2, 1] code_as_byte_string:b'0303010407'| Got the blobs of the net to insert : [7, 7, 3, 8, 9, 6, 7, 10, 11, 6] code_as_byte_string:b'0707030809'| Got the blobs of the net to insert : [3, 3, 3, 1, 6, 7, 10, 5, 3, 5] code_as_byte_string:b'0303030106'| Got the blobs of the net to insert : [1, 1, 0, 0, 0, 0, 0, 3, 5, 1] code_as_byte_string:b'0101000000'| Got the blobs of the net to insert : [0, 3, 0, 1, 1, 6, 7, 5, 5, 3] code_as_byte_string:b'0003000101'| Got the blobs of the net to insert : [6, 6, 3, 3, 8, 8, 6, 6, 2, 0] code_as_byte_string:b'0606030308'| Got the blobs of the net to insert : [5, 5, 2, 4, 6, 5, 9, 9, 4, 2] code_as_byte_string:b'0505020406'| Got the blobs of the net to insert : [2, 1, 3, 5, 7, 5, 3, 4, 1, 3] code_as_byte_string:b'0201030507'| Got the blobs of the net to insert : [0, 0, 0, 1, 1, 3, 2, 1, 0, 2] code_as_byte_string:b'0000000101'| Got the blobs of the net to insert : [0, 0, 0, 0, 0, 0, 1, 3, 1, 0] code_as_byte_string:b'0000000000'| Got the blobs of the net to insert : [1, 2, 1, 5, 7, 10, 8, 2, 2, 1] code_as_byte_string:b'0102010507'| Got the blobs of the net to insert : [0, 0, 0, 1, 5, 4, 6, 3, 1, 3] code_as_byte_string:b'0000000105'| Got the blobs of the net to insert : [0, 2, 1, 0, 0, 0, 0, 0, 4, 2] code_as_byte_string:b'0002010000'| Got the blobs of the net to insert : [0, 2, 0, 1, 1, 0, 0, 0, 1, 3] code_as_byte_string:b'0002000101'| Got the blobs of the net to insert : [1, 1, 1, 2, 0, 3, 4, 3, 4, 5] code_as_byte_string:b'0101010200'| Got the blobs of the net to insert : [3, 2, 1, 4, 5, 7, 6, 5, 6, 5] code_as_byte_string:b'0302010405'| Got the blobs of the net to insert : [3, 4, 5, 5, 9, 9, 9, 9, 11, 7] code_as_byte_string:b'0304050509'| Got the blobs of the net to insert : [3, 2, 2, 1, 1, 2, 2, 3, 2, 5] code_as_byte_string:b'0302020101'| Got the blobs of the net to insert : [1, 0, 3, 1, 0, 0, 0, 0, 0, 5] code_as_byte_string:b'0100030100'| Got the blobs of the net to insert : [0, 0, 2, 2, 0, 0, 0, 0, 0, 3] code_as_byte_string:b'0000020200'| Got the blobs of the net to insert : [1, 1, 2, 1, 0, 0, 0, 3, 6, 7] code_as_byte_string:b'0101020100'| Got the blobs of the net to insert : [1, 3, 3, 3, 0, 1, 5, 6, 7, 4] code_as_byte_string:b'0103030300'| Got the blobs of the net to insert : [5, 7, 3, 2, 5, 9, 10, 3, 5, 0] code_as_byte_string:b'0507030205'| Got the blobs of the net to insert : [6, 8, 3, 6, 6, 6, 5, 10, 10, 3] code_as_byte_string:b'0608030606'| Got the blobs of the net to insert : [2, 5, 4, 6, 9, 9, 10, 9, 11, 5] code_as_byte_string:b'0205040609'| Got the blobs of the net to insert : [5, 2, 3, 7, 5, 7, 6, 2, 4, 3] code_as_byte_string:b'0502030705'| Got the blobs of the net to insert : [5, 7, 3, 2, 1, 2, 2, 2, 0, 0] code_as_byte_string:b'0507030201'| Got the blobs of the net to insert : [4, 2, 1, 2, 3, 1, 0, 1, 0, 1] code_as_byte_string:b'0402010203'| Got the blobs of the net to insert : [0, 0, 0, 1, 2, 2, 0, 1, 0, 0] code_as_byte_string:b'0000000102'| Got the blobs of the net to insert : [1, 0, 0, 1, 1, 1, 0, 1, 1, 3] code_as_byte_string:b'0100000101'| Got the blobs of the net to insert : [1, 1, 2, 3, 4, 4, 1, 2, 8, 7] code_as_byte_string:b'0101020304'| Got the blobs of the net to insert : [5, 2, 3, 6, 8, 12, 9, 9, 3, 5] code_as_byte_string:b'0502030608'| Got the blobs of the net to insert : [6, 3, 3, 0, 1, 2, 2, 5, 2, 4] code_as_byte_string:b'0603030001'| Got the blobs of the net to insert : [0, 0, 0, 0, 0, 2, 5, 1, 0, 0] code_as_byte_string:b'0000000000'| Got the blobs of the net to insert : [0, 0, 0, 0, 0, 2, 5, 1, 0, 0] code_as_byte_string:b'0000000000'| Got the blobs of the net to insert : [1, 1, 0, 0, 1, 2, 0, 0, 0, 1] code_as_byte_string:b'0101000001'| Got the blobs of the net to insert : [0, 1, 0, 1, 3, 4, 2, 1, 3, 5] code_as_byte_string:b'0001000103'| Got the blobs of the net to insert : [3, 3, 3, 8, 7, 5, 5, 4, 5, 3] code_as_byte_string:b'0303030807'| Got the blobs of the net to insert : [0, 1, 1, 4, 2, 1, 3, 7, 9, 9] code_as_byte_string:b'0001010402'| Got the blobs of the net to insert : [7, 10, 10, 2, 4, 7, 8, 5, 5, 2] code_as_byte_string:b'070a0a0204'| Got the blobs of the net to insert : [13, 9, 9, 8, 11, 13, 6, 14, 9, 18] code_as_byte_string:b'0d0909080b'| Got the blobs of the net to insert : [13, 9, 9, 8, 11, 13, 6, 14, 9, 18] code_as_byte_string:b'0d0909080b'| Got the blobs of the net to insert : [2, 2, 6, 7, 8, 6, 4, 3, 1, 8] code_as_byte_string:b'0202060708'| Got the blobs of the net to insert : [1, 2, 2, 1, 0, 1, 1, 2, 1, 1] code_as_byte_string:b'0102020100'| Got the blobs of the net to insert : [2, 2, 1, 3, 2, 3, 2, 0, 0, 1] code_as_byte_string:b'0202010302'| Got the blobs of the net to insert : [6, 8, 5, 7, 7, 8, 10, 12, 12, 6] code_as_byte_string:b'0608050707'| Got the blobs of the net to insert : [6, 8, 5, 7, 7, 8, 10, 12, 12, 6] code_as_byte_string:b'0608050707'| Got the blobs of the net to insert : [5, 5, 4, 3, 5, 6, 3, 4, 3, 3] code_as_byte_string:b'0505040305'| Got the blobs of the net to insert : [5, 3, 3, 1, 1, 1, 2, 2, 2, 4] code_as_byte_string:b'0503030101'| Got the blobs of the net to insert : [2, 1, 2, 2, 2, 0, 0, 1, 0, 3] code_as_byte_string:b'0201020202'| Got the blobs of the net to insert : [0, 0, 2, 3, 3, 1, 0, 0, 0, 0] code_as_byte_string:b'0000020303'| Got the blobs of the net to insert : [1, 0, 0, 1, 0, 0, 0, 3, 4, 5] code_as_byte_string:b'0100000100'| Got the blobs of the net to insert : [1, 0, 1, 2, 1, 0, 3, 3, 3, 8] code_as_byte_string:b'0100010201'| Got the blobs of the net to insert : [5, 5, 6, 4, 3, 6, 9, 7, 7, 7] code_as_byte_string:b'0505060403'| Got the blobs of the net to insert : [4, 6, 6, 4, 7, 8, 8, 8, 12, 8] code_as_byte_string:b'0406060407'| Got the blobs of the net to insert : [11, 8, 5, 9, 12, 14, 13, 14, 12, 6] code_as_byte_string:b'0b0805090c'| Got the blobs of the net to insert : [8, 7, 6, 4, 2, 1, 2, 4, 4, 4] code_as_byte_string:b'0807060402'| Got the blobs of the net to insert : [2, 3, 5, 5, 2, 2, 3, 0, 1, 4] code_as_byte_string:b'0203050502'| Got the blobs of the net to insert : [0, 0, 0, 1, 1, 1, 3, 3, 0, 0] code_as_byte_string:b'0000000101'| time to traite the descriptors : 3.014061212539673 Testing : ['987515187', '987515246', '987515247', '987515248', '987515249', '987515250', '987515224', '987515226', '987515239', '987515240', '987515241', '987515242', '987515243', '987515244', '987515245', '987515235', '987515236', '987515237', '987515238', '987515188', '987515189', '987515190', '987515227', '987515228', '987515230', '987515231', '987515232', '987515233', '987515234', '987515202', '987515204', '987515205', '987515207', '987515208', '987515209', '987515211', '987515192', '987515193', '987515195', '987515196', '987515198', '987515200', '987515201', '987515222', '987515223', '987515175', '987515176', '987515177', '987515178', '987515179', '987515212', '987515213', '987515215', '987515216', '987515217', '987515219', '987515220', '987515180', '987515181', '987515182', '987515183', '987515184', '987515185', '987515186'] In select_photos_meta_from_ids: SELECT photo_id, url, FROM_UNIXTIME(uploaded_at), latitude, longitude, text FROM MTRBack.photos WHERE photo_id IN (987515187,987515246,987515247,987515248,987515249,987515250,987515224,987515226,987515239,987515240,987515241,987515242,987515243,987515244,987515245,987515235,987515236,987515237,987515238,987515188,987515189,987515190,987515227,987515228,987515230,987515231,987515232,987515233,987515234,987515202,987515204,987515205,987515207,987515208,987515209,987515211,987515192,987515193,987515195,987515196,987515198,987515200,987515201,987515222,987515223,987515175,987515176,987515177,987515178,987515179,987515212,987515213,987515215,987515216,987515217,987515219,987515220,987515180,987515181,987515182,987515183,987515184,987515185,987515186) result : {987515175: {'photo_id': 987515175, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/8b398cba2f448622cd9657f5eb3f9796.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22062023_14_16_02_694514_0001.jpg'}, 987515176: {'photo_id': 987515176, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/8b398cba2f448622cd9657f5eb3f9796.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_144.jpg'}, 987515177: {'photo_id': 987515177, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/4a54e9967227806219ddf45d256539d8.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_176.jpg'}, 987515178: {'photo_id': 987515178, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/298b3d2bfe0fda6787b59a78e2e68867.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_208.jpg'}, 987515179: {'photo_id': 987515179, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/f7d4d1757a470f4c96dc3541eac88b9e.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_240.jpg'}, 987515180: {'photo_id': 987515180, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/776a5d7d8486ee2961bbe3a0d90f95b5.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_272.jpg'}, 987515181: {'photo_id': 987515181, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/1738c2798fb31152809ecb443ac286d6.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_304.jpg'}, 987515182: {'photo_id': 987515182, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/fe7f29bf6d13e08c3e985f91b5232178.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_112_y_336.jpg'}, 987515183: {'photo_id': 987515183, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/6aab9ca0421398b4899892c10c2594c6.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_112.jpg'}, 987515184: {'photo_id': 987515184, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/19c8c2177209a285df6014d95fe53f2c.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_144.jpg'}, 987515185: {'photo_id': 987515185, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e172d54457cabee9d7f02ee1300f3ae9.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_176.jpg'}, 987515186: {'photo_id': 987515186, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/797def426440b544aa80dbd63a19234a.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_208.jpg'}, 987515187: {'photo_id': 987515187, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/9f62f98efd3caca0b9c17d27f5c70440.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_240.jpg'}, 987515188: {'photo_id': 987515188, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/4116f9906657a69bb76c2fda982037b9.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_272.jpg'}, 987515189: {'photo_id': 987515189, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/8e8590a26f72249d4c2116dffd0cf668.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_304.jpg'}, 987515190: {'photo_id': 987515190, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/d56932bfc6ba2a8c974c691108755017.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_144_y_336.jpg'}, 987515192: {'photo_id': 987515192, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b661073b218f5f056833d6af1c617153.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_112.jpg'}, 987515193: {'photo_id': 987515193, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/1a97fceb4dcbf5821d783b2e00b52fe6.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_144.jpg'}, 987515195: {'photo_id': 987515195, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/30ccb89dfe410c445878a7f2819ddc36.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22062023_17_37_58_622227.jpg'}, 987515196: {'photo_id': 987515196, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/30ccb89dfe410c445878a7f2819ddc36.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_208.jpg'}, 987515198: {'photo_id': 987515198, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/599e80f444c876f407e94b533c89360b.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_240.jpg'}, 987515200: {'photo_id': 987515200, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/978964436b5d5fb0eeda17e3bfafe889.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_272.jpg'}, 987515201: {'photo_id': 987515201, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b224d2acdc7fa2bbb134c09db6bca7ce.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_304.jpg'}, 987515202: {'photo_id': 987515202, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/3314bd90d1404f31b827d8925abf2d62.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_176_y_336.jpg'}, 987515204: {'photo_id': 987515204, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/9779c4f9d44360a9c80499e3b01e8a09.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_112.jpg'}, 987515205: {'photo_id': 987515205, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/fd4b136d0b3a9a1a347942d7191f6fea.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_144.jpg'}, 987515207: {'photo_id': 987515207, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/de216ddb041e249524b0fb2b949064a5.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_176.jpg'}, 987515208: {'photo_id': 987515208, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/a2b90cb74908aa64bbc4aae58f0c5ae8.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_208.jpg'}, 987515209: {'photo_id': 987515209, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/02dfe1ae39f51994652f4a8538844aea.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_240.jpg'}, 987515211: {'photo_id': 987515211, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/72cc7664d45bd40477351b9b764f1500.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_272.jpg'}, 987515212: {'photo_id': 987515212, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22062023_09_32_14_525625.jpg'}, 987515213: {'photo_id': 987515213, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_208_y_336.jpg'}, 987515215: {'photo_id': 987515215, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/902ef348a7eebb9a8b87f42927347936.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_112.jpg'}, 987515216: {'photo_id': 987515216, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/4f7dc21f1d2cd3fcabadc4a6755921e1.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_144.jpg'}, 987515217: {'photo_id': 987515217, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/78877bb2c5760be28518d17f77d1c609.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_176.jpg'}, 987515219: {'photo_id': 987515219, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/c2d417a5ba6ccf7c84527636f8d5eef9.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_208.jpg'}, 987515220: {'photo_id': 987515220, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e729f316c4c3b32049adfbaaa336d95c.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_240.jpg'}, 987515222: {'photo_id': 987515222, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/067a027bc7402f969b6277d0dcb47eaa.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_272.jpg'}, 987515223: {'photo_id': 987515223, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/ebb57f09941cd11d7ee45a9368a883c1.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_304.jpg'}, 987515224: {'photo_id': 987515224, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e8747b400e713ecbd08d5b75db4d7568.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_240_y_336.jpg'}, 987515226: {'photo_id': 987515226, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/a18048dca1a77ae086b62cf07759f704.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_112.jpg'}, 987515227: {'photo_id': 987515227, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e9c45a0e576ec9e44c1379c3fc5fec7c.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_144.jpg'}, 987515228: {'photo_id': 987515228, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/9f1759f20c9e603bccb9f9879d2f0d54.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_176.jpg'}, 987515230: {'photo_id': 987515230, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/846ad925884264181565c81d152a2e94.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_208.jpg'}, 987515231: {'photo_id': 987515231, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/dbf4cafa71b6db4771c5c8f0c25e9cda.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_240.jpg'}, 987515232: {'photo_id': 987515232, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/38db7950cdb3c674ee0ad65915b021f3.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_272.jpg'}, 987515233: {'photo_id': 987515233, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/a92514bed0e8c5724f2d032d3ab1e2ad.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_304.jpg'}, 987515234: {'photo_id': 987515234, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/2eca3480aed0f8b876242675ad99b666.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_272_y_336.jpg'}, 987515235: {'photo_id': 987515235, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/87075955a2f76b3948b47ffe1825ecd9.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_112.jpg'}, 987515236: {'photo_id': 987515236, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/8b44a98b1aceadad73ed000d65836a9a.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_144.jpg'}, 987515237: {'photo_id': 987515237, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/1183dfa371a457f11ce2b622c7cf9467.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_176.jpg'}, 987515238: {'photo_id': 987515238, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e6292cb81e05894cfeb4b99f21a1d3f8.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_208.jpg'}, 987515239: {'photo_id': 987515239, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b3fa6f29636080b5138c8d8c33fea309.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_240.jpg'}, 987515240: {'photo_id': 987515240, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_272.jpg'}, 987515241: {'photo_id': 987515241, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/073420d938f5f010ffd5b4353c064e09.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_304.jpg'}, 987515242: {'photo_id': 987515242, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/327abb5215d6fd1f0aad51f53ed8c324.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_304_y_336.jpg'}, 987515243: {'photo_id': 987515243, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/4375283f3bc5cdaa431c2fc6f17f53a4.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_112.jpg'}, 987515244: {'photo_id': 987515244, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/419530eaef5ef868f75c758b94eea4b4.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_144.jpg'}, 987515245: {'photo_id': 987515245, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/757d9d208d5bd4375c5f21f68b699148.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_176.jpg'}, 987515246: {'photo_id': 987515246, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/671a708f67f2efa19004b8257fc7b9c8.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_208.jpg'}, 987515247: {'photo_id': 987515247, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/e47b65403df916ba909bc9c439b0af73.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_240.jpg'}, 987515248: {'photo_id': 987515248, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/a70ad88462a22fb62a120721a42b2d42.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'image_22062023_14_16_02_694514_0002.jpg'}, 987515249: {'photo_id': 987515249, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/a70ad88462a22fb62a120721a42b2d42.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_304.jpg'}, 987515250: {'photo_id': 987515250, 'url': 'https://storage.sbg.cloud.ovh.net/v1/AUTH_3b171620e76e4af496c5fd050759c9f0/media.fotonower.com/2019/11/22/b2827c9639df69656f23abcc7f2f82d9.jpg', 'latitude': 0.0, 'longitude': 0.0, 'text': 'photo_origin_x_336_y_336.jpg'}} list_photo_exists : [987515175, 987515176, 987515177, 987515178, 987515179, 987515180, 987515181, 987515182, 987515183, 987515184, 987515185, 987515186, 987515187, 987515188, 987515189, 987515190, 987515192, 987515193, 987515195, 987515196, 987515198, 987515200, 987515201, 987515202, 987515204, 987515205, 987515207, 987515208, 987515209, 987515211, 987515212, 987515213, 987515215, 987515216, 987515217, 987515219, 987515220, 987515222, 987515223, 987515224, 987515226, 987515227, 987515228, 987515230, 987515231, 987515232, 987515233, 987515234, 987515235, 987515236, 987515237, 987515238, 987515239, 987515240, 987515241, 987515242, 987515243, 987515244, 987515245, 987515246, 987515247, 987515248, 987515249, 987515250] storage_type for insertDescriptorsMulti : 1 To insert : 987515187 To insert : 987515246 To insert : 987515247 To insert : 987515248 To insert : 987515249 To insert : 987515250 To insert : 987515224 To insert : 987515226 To insert : 987515239 To insert : 987515240 To insert : 987515241 To insert : 987515242 To insert : 987515243 To insert : 987515244 To insert : 987515245 To insert : 987515235 To insert : 987515236 To insert : 987515237 To insert : 987515238 To insert : 987515188 To insert : 987515189 To insert : 987515190 To insert : 987515227 To insert : 987515228 To insert : 987515230 To insert : 987515231 To insert : 987515232 To insert : 987515233 To insert : 987515234 To insert : 987515202 To insert : 987515204 To insert : 987515205 To insert : 987515207 To insert : 987515208 To insert : 987515209 To insert : 987515211 To insert : 987515192 To insert : 987515193 To insert : 987515195 To insert : 987515196 To insert : 987515198 To insert : 987515200 To insert : 987515201 To insert : 987515222 To insert : 987515223 To insert : 987515175 To insert : 987515176 To insert : 987515177 To insert : 987515178 To insert : 987515179 To insert : 987515212 To insert : 987515213 To insert : 987515215 To insert : 987515216 To insert : 987515217 To insert : 987515219 To insert : 987515220 To insert : 987515180 To insert : 987515181 To insert : 987515182 To insert : 987515183 To insert : 987515184 To insert : 987515185 To insert : 987515186 time to insert the descriptors : 18.349255323410034 After datou_step_exec type output : time spend for datou_step_exec : 65.63354706764221 time spend to save output : 0.00012159347534179688 total time spend for step 1 : 65.63366866111755 step2:argmax Fri May 30 03:40:10 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg': 987515239, 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg': 987515240, 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg': 987515241, 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg': 987515242, 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg': 987515243, 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg': 987515244, 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg': 987515245, 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg': 987515246, 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg': 987515247, 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg': 987515248, 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg': 987515249, 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg': 987515250, 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg': 987515224, 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg': 987515226, 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg': 987515227, 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg': 987515228, 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg': 987515230, 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg': 987515231, 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg': 987515232, 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg': 987515233, 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg': 987515234, 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg': 987515235, 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg': 987515236, 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg': 987515237, 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg': 987515238, 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg': 987515188, 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg': 987515189, 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg': 987515190, 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg': 987515192, 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg': 987515193, 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg': 987515195, 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg': 987515196, 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg': 987515198, 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg': 987515200, 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg': 987515201, 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg': 987515202, 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg': 987515204, 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg': 987515205, 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg': 987515207, 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg': 987515208, 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg': 987515209, 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg': 987515211, 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg': 987515212, 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg': 987515213, 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg': 987515215, 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg': 987515216, 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg': 987515217, 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg': 987515219, 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg': 987515220, 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg': 987515222, 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg': 987515223, 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg': 987515175, 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg': 987515176, 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg': 987515177, 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg': 987515178, 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg': 987515179, 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg': 987515180, 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg': 987515181, 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg': 987515182, 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg': 987515183, 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg': 987515184, 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg': 987515185, 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg': 987515186, 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg': 987515187} map_photo_id_path_extension : {987515239: {'path': 'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg', 'extension': 'jpg'}, 987515240: {'path': 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg', 'extension': 'jpg'}, 987515241: {'path': 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg', 'extension': 'jpg'}, 987515242: {'path': 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg', 'extension': 'jpg'}, 987515243: {'path': 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg', 'extension': 'jpg'}, 987515244: {'path': 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg', 'extension': 'jpg'}, 987515245: {'path': 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg', 'extension': 'jpg'}, 987515246: {'path': 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg', 'extension': 'jpg'}, 987515247: {'path': 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg', 'extension': 'jpg'}, 987515248: {'path': 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg', 'extension': 'jpg'}, 987515249: {'path': 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg', 'extension': 'jpg'}, 987515250: {'path': 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg', 'extension': 'jpg'}, 987515224: {'path': 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg', 'extension': 'jpg'}, 987515226: {'path': 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg', 'extension': 'jpg'}, 987515227: {'path': 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg', 'extension': 'jpg'}, 987515228: {'path': 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg', 'extension': 'jpg'}, 987515230: {'path': 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg', 'extension': 'jpg'}, 987515231: {'path': 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg', 'extension': 'jpg'}, 987515232: {'path': 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg', 'extension': 'jpg'}, 987515233: {'path': 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg', 'extension': 'jpg'}, 987515234: {'path': 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg', 'extension': 'jpg'}, 987515235: {'path': 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg', 'extension': 'jpg'}, 987515236: {'path': 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg', 'extension': 'jpg'}, 987515237: {'path': 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg', 'extension': 'jpg'}, 987515238: {'path': 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg', 'extension': 'jpg'}, 987515188: {'path': 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg', 'extension': 'jpg'}, 987515189: {'path': 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg', 'extension': 'jpg'}, 987515190: {'path': 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg', 'extension': 'jpg'}, 987515192: {'path': 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg', 'extension': 'jpg'}, 987515193: {'path': 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg', 'extension': 'jpg'}, 987515195: {'path': 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg', 'extension': 'jpg'}, 987515196: {'path': 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg', 'extension': 'jpg'}, 987515198: {'path': 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg', 'extension': 'jpg'}, 987515200: {'path': 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg', 'extension': 'jpg'}, 987515201: {'path': 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg', 'extension': 'jpg'}, 987515202: {'path': 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg', 'extension': 'jpg'}, 987515204: {'path': 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg', 'extension': 'jpg'}, 987515205: {'path': 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg', 'extension': 'jpg'}, 987515207: {'path': 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg', 'extension': 'jpg'}, 987515208: {'path': 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg', 'extension': 'jpg'}, 987515209: {'path': 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg', 'extension': 'jpg'}, 987515211: {'path': 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg', 'extension': 'jpg'}, 987515212: {'path': 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'extension': 'jpg'}, 987515213: {'path': 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg', 'extension': 'jpg'}, 987515215: {'path': 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg', 'extension': 'jpg'}, 987515216: {'path': 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg', 'extension': 'jpg'}, 987515217: {'path': 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg', 'extension': 'jpg'}, 987515219: {'path': 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg', 'extension': 'jpg'}, 987515220: {'path': 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg', 'extension': 'jpg'}, 987515222: {'path': 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg', 'extension': 'jpg'}, 987515223: {'path': 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg', 'extension': 'jpg'}, 987515175: {'path': 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg', 'extension': 'jpg'}, 987515176: {'path': 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg', 'extension': 'jpg'}, 987515177: {'path': 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg', 'extension': 'jpg'}, 987515178: {'path': 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg', 'extension': 'jpg'}, 987515179: {'path': 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg', 'extension': 'jpg'}, 987515180: {'path': 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg', 'extension': 'jpg'}, 987515181: {'path': 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg', 'extension': 'jpg'}, 987515182: {'path': 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg', 'extension': 'jpg'}, 987515183: {'path': 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg', 'extension': 'jpg'}, 987515184: {'path': 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg', 'extension': 'jpg'}, 987515185: {'path': 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg', 'extension': 'jpg'}, 987515186: {'path': 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg', 'extension': 'jpg'}, 987515187: {'path': 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou_step Argmax ! calculate argmax for thcl : 1528 After datou_step_exec type output : time spend for datou_step_exec : 0.0007529258728027344 time spend to save output : 7.414817810058594e-05 total time spend for step 2 : 0.0008270740509033203 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 2 output : {'987515187': [('987515187', 'Carton', 0.9811443, 1927, '1528'), 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg'], '987515246': [('987515246', 'Carton', 0.9992336, 1927, '1528'), 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg'], '987515247': [('987515247', 'Carton', 0.9996687, 1927, '1528'), 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg'], '987515248': [('987515248', 'Carton', 0.98129404, 1927, '1528'), 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg'], '987515249': [('987515249', 'Carton', 0.98130006, 1927, '1528'), 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg'], '987515250': [('987515250', 'Carton', 0.98082167, 1927, '1528'), 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg'], '987515224': [('987515224', 'Carton', 0.90838104, 1927, '1528'), 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg'], '987515226': [('987515226', 'Papier_Magazine', 0.9869912, 1927, '1528'), 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg'], '987515239': [('987515239', 'Carton', 0.99978286, 1927, '1528'), 'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg'], '987515240': [('987515240', 'Carton', 0.9995203, 1927, '1528'), 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg'], '987515241': [('987515241', 'Carton', 0.9821817, 1927, '1528'), 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg'], '987515242': [('987515242', 'Carton', 0.93587774, 1927, '1528'), 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg'], '987515243': [('987515243', 'Papier_Magazine', 0.87416136, 1927, '1528'), 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg'], '987515244': [('987515244', 'Papier_Magazine', 0.81743693, 1927, '1528'), 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg'], '987515245': [('987515245', 'Carton', 0.8658532, 1927, '1528'), 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg'], '987515235': [('987515235', 'Papier_Magazine', 0.8919676, 1927, '1528'), 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg'], '987515236': [('987515236', 'Papier_Magazine', 0.53694963, 1927, '1528'), 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg'], '987515237': [('987515237', 'Carton', 0.770268, 1927, '1528'), 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg'], '987515238': [('987515238', 'Carton', 0.99957377, 1927, '1528'), 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg'], '987515188': [('987515188', 'Carton', 0.9956579, 1927, '1528'), 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg'], '987515189': [('987515189', 'Carton', 0.99780136, 1927, '1528'), 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg'], '987515190': [('987515190', 'Carton', 0.9762947, 1927, '1528'), 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg'], '987515227': [('987515227', 'Papier_Magazine', 0.9008627, 1927, '1528'), 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg'], '987515228': [('987515228', 'Papier_Magazine', 0.52163786, 1927, '1528'), 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg'], '987515230': [('987515230', 'Carton', 0.99940526, 1927, '1528'), 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg'], '987515231': [('987515231', 'Carton', 0.9994222, 1927, '1528'), 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg'], '987515232': [('987515232', 'Carton', 0.9992473, 1927, '1528'), 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg'], '987515233': [('987515233', 'Carton', 0.98351353, 1927, '1528'), 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg'], '987515234': [('987515234', 'Carton', 0.94495124, 1927, '1528'), 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg'], '987515202': [('987515202', 'Carton', 0.9911075, 1927, '1528'), 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg'], '987515204': [('987515204', 'Papier_Magazine', 0.9950631, 1927, '1528'), 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg'], '987515205': [('987515205', 'Papier_Magazine', 0.99087125, 1927, '1528'), 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg'], '987515207': [('987515207', 'Papier_Magazine', 0.87419236, 1927, '1528'), 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg'], '987515208': [('987515208', 'Carton', 0.99174726, 1927, '1528'), 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg'], '987515209': [('987515209', 'Carton', 0.96787965, 1927, '1528'), 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg'], '987515211': [('987515211', 'Carton', 0.9733934, 1927, '1528'), 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg'], '987515192': [('987515192', 'Papier_Magazine', 0.9999114, 1927, '1528'), 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg'], '987515193': [('987515193', 'Papier_Magazine', 0.9993962, 1927, '1528'), 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg'], '987515195': [('987515195', 'Carton', 0.98459876, 1927, '1528'), 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg'], '987515196': [('987515196', 'Carton', 0.9846154, 1927, '1528'), 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg'], '987515198': [('987515198', 'Carton', 0.9661438, 1927, '1528'), 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg'], '987515200': [('987515200', 'Carton', 0.9859518, 1927, '1528'), 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg'], '987515201': [('987515201', 'Carton', 0.9954549, 1927, '1528'), 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg'], '987515222': [('987515222', 'Carton', 0.99747145, 1927, '1528'), 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg'], '987515223': [('987515223', 'Carton', 0.99210656, 1927, '1528'), 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg'], '987515175': [('987515175', 'Papier_Magazine', 0.99981445, 1927, '1528'), 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg'], '987515176': [('987515176', 'Papier_Magazine', 0.9998135, 1927, '1528'), 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg'], '987515177': [('987515177', 'Papier_Magazine', 0.9771309, 1927, '1528'), 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg'], '987515178': [('987515178', 'Carton', 0.857373, 1927, '1528'), 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg'], '987515179': [('987515179', 'Carton', 0.9270931, 1927, '1528'), 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg'], '987515212': [('987515212', 'Carton', 0.98691887, 1927, '1528'), 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg'], '987515213': [('987515213', 'Carton', 0.98693603, 1927, '1528'), 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg'], '987515215': [('987515215', 'Papier_Magazine', 0.9939308, 1927, '1528'), 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg'], '987515216': [('987515216', 'Papier_Magazine', 0.97749305, 1927, '1528'), 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg'], '987515217': [('987515217', 'Carton', 0.52847713, 1927, '1528'), 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg'], '987515219': [('987515219', 'Carton', 0.9993692, 1927, '1528'), 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg'], '987515220': [('987515220', 'Carton', 0.99638546, 1927, '1528'), 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg'], '987515180': [('987515180', 'Carton', 0.98998404, 1927, '1528'), 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg'], '987515181': [('987515181', 'Carton', 0.99777883, 1927, '1528'), 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg'], '987515182': [('987515182', 'Carton', 0.99242836, 1927, '1528'), 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg'], '987515183': [('987515183', 'Papier_Magazine', 0.99999213, 1927, '1528'), 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg'], '987515184': [('987515184', 'Papier_Magazine', 0.99973184, 1927, '1528'), 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg'], '987515185': [('987515185', 'Papier_Magazine', 0.7978397, 1927, '1528'), 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg'], '987515186': [('987515186', 'Carton', 0.9847239, 1927, '1528'), 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg']} Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=1879 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=1879 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 1879 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=1879 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : detect_points list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (987515173) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 987515173 download finish for photo 987515173 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.11803865432739258 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:detect_points Fri May 30 03:40:10 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569210_1112947_987515173_91fa471b1a04f95b356afdbaf021f623.jpg': 987515173} map_photo_id_path_extension : {987515173: {'path': 'temp/1748569210_1112947_987515173_91fa471b1a04f95b356afdbaf021f623.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step predict points ! Inside try reload ! classes : ['Autre_Environement', 'Carton', 'Kraft', 'Lointain_Papier_Magazine', 'Metal', 'Papier_Magazine', 'Plastique', 'Sol_Environement', 'Teint_Dans_La_Masse', 'autre_refus'] pht : 1927 model_name : learn_refus_upm_blanches_1924 {'id': 1528, 'mtr_user_id': 31, 'name': 'learn_refus_upm_blanches_1924', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'Autre_Environement,Carton,Kraft,Lointain_Papier_Magazine,Metal,Papier_Magazine,Plastique,Sol_Environement,Teint_Dans_La_Masse,autre_refus', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 1927, 'photo_desc_type': 4421, 'type_classification': 'caffe', 'hashtag_id_list': '2107752388,492774966,493202403,2107752389,492628673,2107752386,492725882,2107752387,2107752385,2107752406'} gpu_mode in detect_points : 1 To load net FromThcl() model_param file didn't exist model_name : learn_refus_upm_blanches_1924 model_type : caffe list file need : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file exist in s3 : ['caffemodel', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file manque in s3 : ['deploy_conv_normal.prototxt', 'deploy_fc.prototxt'] local folder : /data/models_weight/learn_refus_upm_blanches_1924 /data/models_weight/learn_refus_upm_blanches_1924/caffemodel size_local : 45774543 size in s3 : 45774543 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 caffemodel already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/deploy.prototxt size_local : 17312 size in s3 : 17312 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:03 deploy.prototxt already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/mean.npy size_local : 1572992 size in s3 : 1572992 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:05 mean.npy already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/synset_words.txt size_local : 218 size in s3 : 218 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 synset_words.txt already exist and didn't need to update reshape net's input to : (224, 224) origin shape : (10, 3, 224, 224) after reshape : (1, 3, 224, 224) [('data', (1, 3, 224, 224)), ('conv1', (1, 64, 112, 112)), ('pool1', (1, 64, 56, 56)), ('pool1_pool1_0_split_0', (1, 64, 56, 56)), ('pool1_pool1_0_split_1', (1, 64, 56, 56)), ('res2a_branch1', (1, 64, 56, 56)), ('res2a_branch2a', (1, 64, 56, 56)), ('res2a_branch2b', (1, 64, 56, 56)), ('res2a', (1, 64, 56, 56)), ('res2a_res2a_relu_0_split_0', (1, 64, 56, 56)), ('res2a_res2a_relu_0_split_1', (1, 64, 56, 56)), ('res2b_branch2a', (1, 64, 56, 56)), ('res2b_branch2b', (1, 64, 56, 56)), ('res2b', (1, 64, 56, 56)), ('res2b_res2b_relu_0_split_0', (1, 64, 56, 56)), ('res2b_res2b_relu_0_split_1', (1, 64, 56, 56)), ('res3a_branch1', (1, 128, 28, 28)), ('res3a_branch2a', (1, 128, 28, 28)), ('res3a_branch2b', (1, 128, 28, 28)), ('res3a', (1, 128, 28, 28)), ('res3a_res3a_relu_0_split_0', (1, 128, 28, 28)), ('res3a_res3a_relu_0_split_1', (1, 128, 28, 28)), ('res3b_branch2a', (1, 128, 28, 28)), ('res3b_branch2b', (1, 128, 28, 28)), ('res3b', (1, 128, 28, 28)), ('res3b_res3b_relu_0_split_0', (1, 128, 28, 28)), ('res3b_res3b_relu_0_split_1', (1, 128, 28, 28)), ('res4a_branch1', (1, 256, 14, 14)), ('res4a_branch2a', (1, 256, 14, 14)), ('res4a_branch2b', (1, 256, 14, 14)), ('res4a', (1, 256, 14, 14)), ('res4a_res4a_relu_0_split_0', (1, 256, 14, 14)), ('res4a_res4a_relu_0_split_1', (1, 256, 14, 14)), ('res4b_branch2a', (1, 256, 14, 14)), ('res4b_branch2b', (1, 256, 14, 14)), ('res4b', (1, 256, 14, 14)), ('res4b_res4b_relu_0_split_0', (1, 256, 14, 14)), ('res4b_res4b_relu_0_split_1', (1, 256, 14, 14)), ('res5a_branch1', (1, 512, 7, 7)), ('res5a_branch2a', (1, 512, 7, 7)), ('res5a_branch2b', (1, 512, 7, 7)), ('res5a', (1, 512, 7, 7)), ('res5a_res5a_relu_0_split_0', (1, 512, 7, 7)), ('res5a_res5a_relu_0_split_1', (1, 512, 7, 7)), ('res5b_branch2a', (1, 512, 7, 7)), ('res5b_branch2b', (1, 512, 7, 7)), ('res5b', (1, 512, 7, 7)), ('fc2019-10-22_15-02-46', (1, 10, 1, 1)), ('prob', (1, 10, 1, 1))] set image transformer : About to compute detect the points : len(args) : 1 Inside predict_points step exec : nb paths : 1 treate image : temp/1748569210_1112947_987515173_91fa471b1a04f95b356afdbaf021f623.jpg size of numpy array img : 2408584 scale method : caffe/skimage size of numpy array img_scale : 2408584 (448, 448, 3) nb_h 8 nb_w 8 size of sub images : (224, 224, 3) size of caffe_input : 38535320 (64, 3, 224, 224) time to do the preprocess : 0.04096817970275879 time to do a prediction : 0.3303260803222656 dict_keys(['prob']) shape of output (64, 10, 1, 1) shape of the out_put heatmap (10, 8, 8) number of sub_photos vertical and horizon 8 8 size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) size of heatmap : (8,8) After datou_step_exec type output : time spend for datou_step_exec : 1.677257776260376 time spend to save output : 5.459785461425781e-05 total time spend for step 1 : 1.6773123741149902 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {987515173: [(987515173, 1982, 'Autre_Environement', 112, -1, 112, -1, 6.271815965880334e-12), (987515173, 1982, 'Autre_Environement', 144, -1, 112, -1, 2.4844125021128427e-11), (987515173, 1982, 'Autre_Environement', 176, -1, 112, -1, 1.0650548887269906e-08), (987515173, 1982, 'Autre_Environement', 208, -1, 112, -1, 4.448247921118309e-07), (987515173, 1982, 'Autre_Environement', 240, -1, 112, -1, 1.923155878102989e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 112, -1, 3.779579492402263e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 112, -1, 0.00012279744260013103), (987515173, 1982, 'Autre_Environement', 336, -1, 112, -1, 2.9485890991054475e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 144, -1, 2.374503260682559e-08), (987515173, 1982, 'Autre_Environement', 144, -1, 144, -1, 2.2141357192140276e-08), (987515173, 1982, 'Autre_Environement', 176, -1, 144, -1, 1.3738434745391714e-07), (987515173, 1982, 'Autre_Environement', 208, -1, 144, -1, 1.4727493180544116e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 144, -1, 1.1291859664197545e-05), (987515173, 1982, 'Autre_Environement', 272, -1, 144, -1, 0.00015815612277947366), (987515173, 1982, 'Autre_Environement', 304, -1, 144, -1, 0.0004434996226336807), (987515173, 1982, 'Autre_Environement', 336, -1, 144, -1, 6.545944779645652e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 176, -1, 1.3293632719069137e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 176, -1, 1.6162027804966783e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 176, -1, 2.526713387851487e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 176, -1, 1.615999963178183e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 176, -1, 6.257173481571954e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 176, -1, 8.64791072672233e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 176, -1, 0.0003271224850323051), (987515173, 1982, 'Autre_Environement', 336, -1, 176, -1, 0.0003056900459341705), (987515173, 1982, 'Autre_Environement', 112, -1, 208, -1, 1.860383417806588e-05), (987515173, 1982, 'Autre_Environement', 144, -1, 208, -1, 7.921114047348965e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 208, -1, 2.7036321625928394e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 208, -1, 1.801614234864246e-05), (987515173, 1982, 'Autre_Environement', 240, -1, 208, -1, 2.3417936972691678e-05), (987515173, 1982, 'Autre_Environement', 272, -1, 208, -1, 1.694694765319582e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 208, -1, 4.544725925370585e-06), (987515173, 1982, 'Autre_Environement', 336, -1, 208, -1, 8.81922915141331e-06), (987515173, 1982, 'Autre_Environement', 112, -1, 240, -1, 6.098133781051729e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 240, -1, 1.6467932937302976e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 240, -1, 1.963200020327349e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 240, -1, 1.4263438288253383e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 240, -1, 7.862996426410973e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 240, -1, 1.286251972487662e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 240, -1, 9.291064088756684e-06), (987515173, 1982, 'Autre_Environement', 336, -1, 240, -1, 2.1700687284464948e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 272, -1, 3.8262664929789025e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 272, -1, 2.5499746243440313e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 272, -1, 2.96101507046842e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 272, -1, 2.7530711577128386e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 272, -1, 4.317397269915091e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 272, -1, 8.171764420694672e-06), (987515173, 1982, 'Autre_Environement', 304, -1, 272, -1, 1.1662354154395871e-05), (987515173, 1982, 'Autre_Environement', 336, -1, 272, -1, 3.91181129089091e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 304, -1, 1.238445929629961e-05), (987515173, 1982, 'Autre_Environement', 144, -1, 304, -1, 1.595929279574193e-05), (987515173, 1982, 'Autre_Environement', 176, -1, 304, -1, 3.3600517781451344e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 304, -1, 0.00015436304965987802), (987515173, 1982, 'Autre_Environement', 240, -1, 304, -1, 0.0002585948968771845), (987515173, 1982, 'Autre_Environement', 272, -1, 304, -1, 0.0001877176546258852), (987515173, 1982, 'Autre_Environement', 304, -1, 304, -1, 0.00021357914374675602), (987515173, 1982, 'Autre_Environement', 336, -1, 304, -1, 0.00016418426821473986), (987515173, 1982, 'Autre_Environement', 112, -1, 336, -1, 4.553502094495343e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 336, -1, 1.733617318677716e-05), (987515173, 1982, 'Autre_Environement', 176, -1, 336, -1, 4.9303725973004475e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 336, -1, 0.0001215945158037357), (987515173, 1982, 'Autre_Environement', 240, -1, 336, -1, 0.00019625374989118427), (987515173, 1982, 'Autre_Environement', 272, -1, 336, -1, 0.00018786026339512318), (987515173, 1982, 'Autre_Environement', 304, -1, 336, -1, 0.00012362762936390936), (987515173, 1982, 'Autre_Environement', 336, -1, 336, -1, 0.0002712809364311397), (987515173, 1982, 'Carton', 112, -1, 112, -1, 1.576355685983799e-07), (987515173, 1982, 'Carton', 144, -1, 112, -1, 4.09165568271419e-06), (987515173, 1982, 'Carton', 176, -1, 112, -1, 6.995676812948659e-06), (987515173, 1982, 'Carton', 208, -1, 112, -1, 0.0008727198000997305), (987515173, 1982, 'Carton', 240, -1, 112, -1, 0.0026501729153096676), (987515173, 1982, 'Carton', 272, -1, 112, -1, 0.003380689537152648), (987515173, 1982, 'Carton', 304, -1, 112, -1, 0.03130565211176872), (987515173, 1982, 'Carton', 336, -1, 112, -1, 0.05585074424743652), (987515173, 1982, 'Carton', 112, -1, 144, -1, 0.00012447437620721757), (987515173, 1982, 'Carton', 144, -1, 144, -1, 0.00020915831555612385), (987515173, 1982, 'Carton', 176, -1, 144, -1, 0.0003673181345220655), (987515173, 1982, 'Carton', 208, -1, 144, -1, 0.006835674401372671), (987515173, 1982, 'Carton', 240, -1, 144, -1, 0.01591036096215248), (987515173, 1982, 'Carton', 272, -1, 144, -1, 0.009403361938893795), (987515173, 1982, 'Carton', 304, -1, 144, -1, 0.00975574180483818), (987515173, 1982, 'Carton', 336, -1, 144, -1, 0.022150758653879166), (987515173, 1982, 'Carton', 112, -1, 176, -1, 0.02188420481979847), (987515173, 1982, 'Carton', 144, -1, 176, -1, 0.19292433559894562), (987515173, 1982, 'Carton', 176, -1, 176, -1, 0.09658340364694595), (987515173, 1982, 'Carton', 208, -1, 176, -1, 0.12378031760454178), (987515173, 1982, 'Carton', 240, -1, 176, -1, 0.5330907702445984), (987515173, 1982, 'Carton', 272, -1, 176, -1, 0.4615733027458191), (987515173, 1982, 'Carton', 304, -1, 176, -1, 0.7711007595062256), (987515173, 1982, 'Carton', 336, -1, 176, -1, 0.8663382530212402), (987515173, 1982, 'Carton', 112, -1, 208, -1, 0.8503047227859497), (987515173, 1982, 'Carton', 144, -1, 208, -1, 0.9844670295715332), (987515173, 1982, 'Carton', 176, -1, 208, -1, 0.9847484230995178), (987515173, 1982, 'Carton', 208, -1, 208, -1, 0.9919507503509521), (987515173, 1982, 'Carton', 240, -1, 208, -1, 0.9993792772293091), (987515173, 1982, 'Carton', 272, -1, 208, -1, 0.9994138479232788), (987515173, 1982, 'Carton', 304, -1, 208, -1, 0.9995881915092468), (987515173, 1982, 'Carton', 336, -1, 208, -1, 0.9992272853851318), (987515173, 1982, 'Carton', 112, -1, 240, -1, 0.9276415705680847), (987515173, 1982, 'Carton', 144, -1, 240, -1, 0.9810431003570557), (987515173, 1982, 'Carton', 176, -1, 240, -1, 0.966120719909668), (987515173, 1982, 'Carton', 208, -1, 240, -1, 0.9677486419677734), (987515173, 1982, 'Carton', 240, -1, 240, -1, 0.996390163898468), (987515173, 1982, 'Carton', 272, -1, 240, -1, 0.9994217157363892), (987515173, 1982, 'Carton', 304, -1, 240, -1, 0.9997863173484802), (987515173, 1982, 'Carton', 336, -1, 240, -1, 0.9996683597564697), (987515173, 1982, 'Carton', 112, -1, 272, -1, 0.9895250797271729), (987515173, 1982, 'Carton', 144, -1, 272, -1, 0.995465874671936), (987515173, 1982, 'Carton', 176, -1, 272, -1, 0.9854716658592224), (987515173, 1982, 'Carton', 208, -1, 272, -1, 0.9733673334121704), (987515173, 1982, 'Carton', 240, -1, 272, -1, 0.9974767565727234), (987515173, 1982, 'Carton', 272, -1, 272, -1, 0.9992011189460754), (987515173, 1982, 'Carton', 304, -1, 272, -1, 0.9995144605636597), (987515173, 1982, 'Carton', 336, -1, 272, -1, 0.9991315007209778), (987515173, 1982, 'Carton', 112, -1, 304, -1, 0.9977748990058899), (987515173, 1982, 'Carton', 144, -1, 304, -1, 0.9977632761001587), (987515173, 1982, 'Carton', 176, -1, 304, -1, 0.9955496788024902), (987515173, 1982, 'Carton', 208, -1, 304, -1, 0.9927307963371277), (987515173, 1982, 'Carton', 240, -1, 304, -1, 0.9920262694358826), (987515173, 1982, 'Carton', 272, -1, 304, -1, 0.9835942387580872), (987515173, 1982, 'Carton', 304, -1, 304, -1, 0.9819222688674927), (987515173, 1982, 'Carton', 336, -1, 304, -1, 0.9809139370918274), (987515173, 1982, 'Carton', 112, -1, 336, -1, 0.9924526810646057), (987515173, 1982, 'Carton', 144, -1, 336, -1, 0.9762071967124939), (987515173, 1982, 'Carton', 176, -1, 336, -1, 0.99116450548172), (987515173, 1982, 'Carton', 208, -1, 336, -1, 0.9869438409805298), (987515173, 1982, 'Carton', 240, -1, 336, -1, 0.9085468053817749), (987515173, 1982, 'Carton', 272, -1, 336, -1, 0.9452228546142578), (987515173, 1982, 'Carton', 304, -1, 336, -1, 0.9365962147712708), (987515173, 1982, 'Carton', 336, -1, 336, -1, 0.9808369278907776), (987515173, 1982, 'Kraft', 112, -1, 112, -1, 1.957737527646941e-09), (987515173, 1982, 'Kraft', 144, -1, 112, -1, 1.720074749300693e-08), (987515173, 1982, 'Kraft', 176, -1, 112, -1, 9.63709453571937e-07), (987515173, 1982, 'Kraft', 208, -1, 112, -1, 3.138272586511448e-05), (987515173, 1982, 'Kraft', 240, -1, 112, -1, 4.440563134266995e-05), (987515173, 1982, 'Kraft', 272, -1, 112, -1, 0.00020620097348000854), (987515173, 1982, 'Kraft', 304, -1, 112, -1, 0.0010808930965140462), (987515173, 1982, 'Kraft', 336, -1, 112, -1, 0.0008296957821585238), (987515173, 1982, 'Kraft', 112, -1, 144, -1, 2.638400110299699e-05), (987515173, 1982, 'Kraft', 144, -1, 144, -1, 6.993338956817752e-06), (987515173, 1982, 'Kraft', 176, -1, 144, -1, 3.622401436587097e-06), (987515173, 1982, 'Kraft', 208, -1, 144, -1, 3.556726733222604e-05), (987515173, 1982, 'Kraft', 240, -1, 144, -1, 6.707842112518847e-05), (987515173, 1982, 'Kraft', 272, -1, 144, -1, 8.681361214257777e-05), (987515173, 1982, 'Kraft', 304, -1, 144, -1, 0.00012179985060356557), (987515173, 1982, 'Kraft', 336, -1, 144, -1, 0.00011247207294218242), (987515173, 1982, 'Kraft', 112, -1, 176, -1, 0.0004987759166397154), (987515173, 1982, 'Kraft', 144, -1, 176, -1, 0.00012291625898797065), (987515173, 1982, 'Kraft', 176, -1, 176, -1, 9.114220301853493e-05), (987515173, 1982, 'Kraft', 208, -1, 176, -1, 5.1748036639764905e-05), (987515173, 1982, 'Kraft', 240, -1, 176, -1, 0.00011549148621270433), (987515173, 1982, 'Kraft', 272, -1, 176, -1, 0.00043214470497332513), (987515173, 1982, 'Kraft', 304, -1, 176, -1, 0.0009234515018761158), (987515173, 1982, 'Kraft', 336, -1, 176, -1, 0.0014272574335336685), (987515173, 1982, 'Kraft', 112, -1, 208, -1, 6.90045126248151e-05), (987515173, 1982, 'Kraft', 144, -1, 208, -1, 1.8495431504561566e-05), (987515173, 1982, 'Kraft', 176, -1, 208, -1, 2.5861496396828443e-05), (987515173, 1982, 'Kraft', 208, -1, 208, -1, 3.5416676837485284e-05), (987515173, 1982, 'Kraft', 240, -1, 208, -1, 3.73744624084793e-05), (987515173, 1982, 'Kraft', 272, -1, 208, -1, 8.635753329144791e-05), (987515173, 1982, 'Kraft', 304, -1, 208, -1, 0.00012362583947833627), (987515173, 1982, 'Kraft', 336, -1, 208, -1, 0.0003905165649484843), (987515173, 1982, 'Kraft', 112, -1, 240, -1, 0.0003079257730860263), (987515173, 1982, 'Kraft', 144, -1, 240, -1, 4.163364792475477e-05), (987515173, 1982, 'Kraft', 176, -1, 240, -1, 1.2242019693076145e-05), (987515173, 1982, 'Kraft', 208, -1, 240, -1, 7.337544502661331e-06), (987515173, 1982, 'Kraft', 240, -1, 240, -1, 2.2982125301496126e-05), (987515173, 1982, 'Kraft', 272, -1, 240, -1, 5.8157253079116344e-05), (987515173, 1982, 'Kraft', 304, -1, 240, -1, 6.560523615917191e-05), (987515173, 1982, 'Kraft', 336, -1, 240, -1, 0.00018724572146311402), (987515173, 1982, 'Kraft', 112, -1, 272, -1, 0.0014629343058913946), (987515173, 1982, 'Kraft', 144, -1, 272, -1, 0.0006902696331962943), (987515173, 1982, 'Kraft', 176, -1, 272, -1, 0.0002742304641287774), (987515173, 1982, 'Kraft', 208, -1, 272, -1, 4.357844591140747e-05), (987515173, 1982, 'Kraft', 240, -1, 272, -1, 3.340828698128462e-05), (987515173, 1982, 'Kraft', 272, -1, 272, -1, 8.349671406904235e-05), (987515173, 1982, 'Kraft', 304, -1, 272, -1, 0.00011392419401090592), (987515173, 1982, 'Kraft', 336, -1, 272, -1, 0.0004221333365421742), (987515173, 1982, 'Kraft', 112, -1, 304, -1, 0.001014246023260057), (987515173, 1982, 'Kraft', 144, -1, 304, -1, 0.0009221484651789069), (987515173, 1982, 'Kraft', 176, -1, 304, -1, 0.0006190584390424192), (987515173, 1982, 'Kraft', 208, -1, 304, -1, 0.001081658760085702), (987515173, 1982, 'Kraft', 240, -1, 304, -1, 0.0017836974002420902), (987515173, 1982, 'Kraft', 272, -1, 304, -1, 0.004672076087445021), (987515173, 1982, 'Kraft', 304, -1, 304, -1, 0.0046897633001208305), (987515173, 1982, 'Kraft', 336, -1, 304, -1, 0.012484777718782425), (987515173, 1982, 'Kraft', 112, -1, 336, -1, 0.002183066913858056), (987515173, 1982, 'Kraft', 144, -1, 336, -1, 0.005705210845917463), (987515173, 1982, 'Kraft', 176, -1, 336, -1, 0.0008308480028063059), (987515173, 1982, 'Kraft', 208, -1, 336, -1, 0.001263151061721146), (987515173, 1982, 'Kraft', 240, -1, 336, -1, 0.007801832631230354), (987515173, 1982, 'Kraft', 272, -1, 336, -1, 0.012551669031381607), (987515173, 1982, 'Kraft', 304, -1, 336, -1, 0.017930345609784126), (987515173, 1982, 'Kraft', 336, -1, 336, -1, 0.007753178011626005), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 112, -1, 1.4981969831406872e-10), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 112, -1, 8.321308087033685e-09), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 112, -1, 5.521291654986271e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 112, -1, 5.514806161954766e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 112, -1, 8.237828296842054e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 112, -1, 3.563874997780658e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 112, -1, 8.149821951519698e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 112, -1, 4.2202686017844826e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 144, -1, 3.643475281478459e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 144, -1, 1.3091824939692742e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 144, -1, 2.465325906086946e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 144, -1, 3.4592671909194905e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 144, -1, 5.710682671633549e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 144, -1, 2.059975486190524e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 144, -1, 3.9010625187074766e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 144, -1, 1.3684768418897875e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 176, -1, 3.892207303124451e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 176, -1, 2.55041777563747e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 176, -1, 1.2033554412482772e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 176, -1, 7.819895472493954e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 176, -1, 6.353224762278842e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 176, -1, 2.6373883883934468e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 176, -1, 3.922960240743123e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 176, -1, 1.8472530427970923e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 208, -1, 1.7999934698309517e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 208, -1, 5.661302111548139e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 208, -1, 2.1980697511025937e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 208, -1, 1.7350848793284968e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 208, -1, 3.996210864443128e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 208, -1, 2.027948937666224e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 208, -1, 9.864364614031729e-08), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 208, -1, 8.694066622183527e-08), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 240, -1, 1.822134095164074e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 240, -1, 6.324926289380528e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 240, -1, 7.821980716471444e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 240, -1, 6.381894195328641e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 240, -1, 7.713734362368996e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 240, -1, 5.179261961529846e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 240, -1, 1.772643258846074e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 240, -1, 1.4453388530455413e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 272, -1, 4.1830210761872877e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 272, -1, 3.170695208609686e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 272, -1, 5.096800350656849e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 272, -1, 7.328864057853934e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 272, -1, 4.0205608797805326e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 272, -1, 3.2233640467893565e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 272, -1, 2.121068121141434e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 272, -1, 3.4728486753010657e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 304, -1, 3.321688666346745e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 304, -1, 3.2591808007964573e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 304, -1, 8.38333505726041e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 304, -1, 3.2999407721945317e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 304, -1, 4.413991973706288e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 304, -1, 5.0191547416034155e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 304, -1, 8.209121915569995e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 304, -1, 5.568050255533308e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 336, -1, 4.1875182432704605e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 336, -1, 8.223066743084928e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 336, -1, 7.151177214836935e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 336, -1, 1.9496640106808627e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 336, -1, 3.911284238711232e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 336, -1, 2.4560574729548534e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 336, -1, 3.435417056607548e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 336, -1, 4.8369684009230696e-06), (987515173, 1982, 'Metal', 112, -1, 112, -1, 7.40447078650952e-11), (987515173, 1982, 'Metal', 144, -1, 112, -1, 1.9585983945802354e-09), (987515173, 1982, 'Metal', 176, -1, 112, -1, 7.308526619453914e-07), (987515173, 1982, 'Metal', 208, -1, 112, -1, 9.458058229938615e-06), (987515173, 1982, 'Metal', 240, -1, 112, -1, 2.6567093300400302e-05), (987515173, 1982, 'Metal', 272, -1, 112, -1, 0.00016929887351579964), (987515173, 1982, 'Metal', 304, -1, 112, -1, 0.0007552222232334316), (987515173, 1982, 'Metal', 336, -1, 112, -1, 0.0003030860680155456), (987515173, 1982, 'Metal', 112, -1, 144, -1, 5.366252366911795e-07), (987515173, 1982, 'Metal', 144, -1, 144, -1, 4.3793150439341844e-07), (987515173, 1982, 'Metal', 176, -1, 144, -1, 3.176887958034058e-06), (987515173, 1982, 'Metal', 208, -1, 144, -1, 9.592667083779816e-06), (987515173, 1982, 'Metal', 240, -1, 144, -1, 1.658712062635459e-05), (987515173, 1982, 'Metal', 272, -1, 144, -1, 7.65814766054973e-05), (987515173, 1982, 'Metal', 304, -1, 144, -1, 0.00013093012967146933), (987515173, 1982, 'Metal', 336, -1, 144, -1, 4.670989801525138e-05), (987515173, 1982, 'Metal', 112, -1, 176, -1, 5.880497610633029e-06), (987515173, 1982, 'Metal', 144, -1, 176, -1, 4.011907549283933e-06), (987515173, 1982, 'Metal', 176, -1, 176, -1, 1.0134302101505455e-05), (987515173, 1982, 'Metal', 208, -1, 176, -1, 4.159005129622528e-06), (987515173, 1982, 'Metal', 240, -1, 176, -1, 6.656815003225347e-06), (987515173, 1982, 'Metal', 272, -1, 176, -1, 5.4316413297783583e-05), (987515173, 1982, 'Metal', 304, -1, 176, -1, 0.00012912174861412495), (987515173, 1982, 'Metal', 336, -1, 176, -1, 6.557787855854258e-05), (987515173, 1982, 'Metal', 112, -1, 208, -1, 8.146424988808576e-06), (987515173, 1982, 'Metal', 144, -1, 208, -1, 2.247384600195801e-06), (987515173, 1982, 'Metal', 176, -1, 208, -1, 6.634919373027515e-06), (987515173, 1982, 'Metal', 208, -1, 208, -1, 4.596822236635489e-06), (987515173, 1982, 'Metal', 240, -1, 208, -1, 1.7126073998952052e-06), (987515173, 1982, 'Metal', 272, -1, 208, -1, 1.3694709650735604e-06), (987515173, 1982, 'Metal', 304, -1, 208, -1, 5.784029326605378e-07), (987515173, 1982, 'Metal', 336, -1, 208, -1, 8.349181257472083e-07), (987515173, 1982, 'Metal', 112, -1, 240, -1, 2.9274015105329454e-06), (987515173, 1982, 'Metal', 144, -1, 240, -1, 7.392775955850084e-07), (987515173, 1982, 'Metal', 176, -1, 240, -1, 6.081426136006485e-07), (987515173, 1982, 'Metal', 208, -1, 240, -1, 6.215839221113129e-07), (987515173, 1982, 'Metal', 240, -1, 240, -1, 1.1986741128566791e-06), (987515173, 1982, 'Metal', 272, -1, 240, -1, 6.961149097151065e-07), (987515173, 1982, 'Metal', 304, -1, 240, -1, 3.145585480979207e-07), (987515173, 1982, 'Metal', 336, -1, 240, -1, 4.122880170598364e-07), (987515173, 1982, 'Metal', 112, -1, 272, -1, 2.620731265778886e-06), (987515173, 1982, 'Metal', 144, -1, 272, -1, 8.054418572100985e-07), (987515173, 1982, 'Metal', 176, -1, 272, -1, 8.229471291087975e-07), (987515173, 1982, 'Metal', 208, -1, 272, -1, 4.01219210743875e-07), (987515173, 1982, 'Metal', 240, -1, 272, -1, 2.7388938406147645e-07), (987515173, 1982, 'Metal', 272, -1, 272, -1, 3.310874774342665e-07), (987515173, 1982, 'Metal', 304, -1, 272, -1, 3.4722987152235874e-07), (987515173, 1982, 'Metal', 336, -1, 272, -1, 1.4728271935382509e-06), (987515173, 1982, 'Metal', 112, -1, 304, -1, 2.352227966184728e-06), (987515173, 1982, 'Metal', 144, -1, 304, -1, 1.9818360215140274e-06), (987515173, 1982, 'Metal', 176, -1, 304, -1, 4.541454472928308e-06), (987515173, 1982, 'Metal', 208, -1, 304, -1, 1.3329069588507991e-05), (987515173, 1982, 'Metal', 240, -1, 304, -1, 5.641319603455486e-06), (987515173, 1982, 'Metal', 272, -1, 304, -1, 5.678331945091486e-06), (987515173, 1982, 'Metal', 304, -1, 304, -1, 6.368874437612249e-06), (987515173, 1982, 'Metal', 336, -1, 304, -1, 7.368290880549466e-06), (987515173, 1982, 'Metal', 112, -1, 336, -1, 7.137274224078283e-06), (987515173, 1982, 'Metal', 144, -1, 336, -1, 3.321443364256993e-05), (987515173, 1982, 'Metal', 176, -1, 336, -1, 4.011798955616541e-05), (987515173, 1982, 'Metal', 208, -1, 336, -1, 7.578751683467999e-05), (987515173, 1982, 'Metal', 240, -1, 336, -1, 0.00012133477139286697), (987515173, 1982, 'Metal', 272, -1, 336, -1, 3.714204649440944e-05), (987515173, 1982, 'Metal', 304, -1, 336, -1, 2.9102686312398873e-05), (987515173, 1982, 'Metal', 336, -1, 336, -1, 2.8008123990730383e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 112, -1, 0.9999953508377075), (987515173, 1982, 'Papier_Magazine', 144, -1, 112, -1, 0.9999924898147583), (987515173, 1982, 'Papier_Magazine', 176, -1, 112, -1, 0.9999085664749146), (987515173, 1982, 'Papier_Magazine', 208, -1, 112, -1, 0.9949527978897095), (987515173, 1982, 'Papier_Magazine', 240, -1, 112, -1, 0.9937595725059509), (987515173, 1982, 'Papier_Magazine', 272, -1, 112, -1, 0.9866683483123779), (987515173, 1982, 'Papier_Magazine', 304, -1, 112, -1, 0.8921955227851868), (987515173, 1982, 'Papier_Magazine', 336, -1, 112, -1, 0.8720824122428894), (987515173, 1982, 'Papier_Magazine', 112, -1, 144, -1, 0.9998136162757874), (987515173, 1982, 'Papier_Magazine', 144, -1, 144, -1, 0.9997422099113464), (987515173, 1982, 'Papier_Magazine', 176, -1, 144, -1, 0.9994150400161743), (987515173, 1982, 'Papier_Magazine', 208, -1, 144, -1, 0.9911978840827942), (987515173, 1982, 'Papier_Magazine', 240, -1, 144, -1, 0.9780967831611633), (987515173, 1982, 'Papier_Magazine', 272, -1, 144, -1, 0.9000454545021057), (987515173, 1982, 'Papier_Magazine', 304, -1, 144, -1, 0.5362882614135742), (987515173, 1982, 'Papier_Magazine', 336, -1, 144, -1, 0.8107313513755798), (987515173, 1982, 'Papier_Magazine', 112, -1, 176, -1, 0.97757488489151), (987515173, 1982, 'Papier_Magazine', 144, -1, 176, -1, 0.8067144155502319), (987515173, 1982, 'Papier_Magazine', 176, -1, 176, -1, 0.9026957154273987), (987515173, 1982, 'Papier_Magazine', 208, -1, 176, -1, 0.8749292492866516), (987515173, 1982, 'Papier_Magazine', 240, -1, 176, -1, 0.46545782685279846), (987515173, 1982, 'Papier_Magazine', 272, -1, 176, -1, 0.5244843363761902), (987515173, 1982, 'Papier_Magazine', 304, -1, 176, -1, 0.16661480069160461), (987515173, 1982, 'Papier_Magazine', 336, -1, 176, -1, 0.11046022921800613), (987515173, 1982, 'Papier_Magazine', 112, -1, 208, -1, 0.1494138389825821), (987515173, 1982, 'Papier_Magazine', 144, -1, 208, -1, 0.015275389887392521), (987515173, 1982, 'Papier_Magazine', 176, -1, 208, -1, 0.014782627113163471), (987515173, 1982, 'Papier_Magazine', 208, -1, 208, -1, 0.007588740438222885), (987515173, 1982, 'Papier_Magazine', 240, -1, 208, -1, 0.0003462004824541509), (987515173, 1982, 'Papier_Magazine', 272, -1, 208, -1, 0.00017021365056280047), (987515173, 1982, 'Papier_Magazine', 304, -1, 208, -1, 5.368440906750038e-05), (987515173, 1982, 'Papier_Magazine', 336, -1, 208, -1, 6.279405351961032e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 240, -1, 0.0716143548488617), (987515173, 1982, 'Papier_Magazine', 144, -1, 240, -1, 0.01876274310052395), (987515173, 1982, 'Papier_Magazine', 176, -1, 240, -1, 0.03377354517579079), (987515173, 1982, 'Papier_Magazine', 208, -1, 240, -1, 0.03220600262284279), (987515173, 1982, 'Papier_Magazine', 240, -1, 240, -1, 0.003492899937555194), (987515173, 1982, 'Papier_Magazine', 272, -1, 240, -1, 0.00034153219894506037), (987515173, 1982, 'Papier_Magazine', 304, -1, 240, -1, 2.9719671147176996e-05), (987515173, 1982, 'Papier_Magazine', 336, -1, 240, -1, 1.6387210052926093e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 272, -1, 0.008499102666974068), (987515173, 1982, 'Papier_Magazine', 144, -1, 272, -1, 0.003649174701422453), (987515173, 1982, 'Papier_Magazine', 176, -1, 272, -1, 0.014067606069147587), (987515173, 1982, 'Papier_Magazine', 208, -1, 272, -1, 0.02651386708021164), (987515173, 1982, 'Papier_Magazine', 240, -1, 272, -1, 0.0024477692786604166), (987515173, 1982, 'Papier_Magazine', 272, -1, 272, -1, 0.0006472524837590754), (987515173, 1982, 'Papier_Magazine', 304, -1, 272, -1, 0.0002737916074693203), (987515173, 1982, 'Papier_Magazine', 336, -1, 272, -1, 0.00015876929683145136), (987515173, 1982, 'Papier_Magazine', 112, -1, 304, -1, 0.0009783600689843297), (987515173, 1982, 'Papier_Magazine', 144, -1, 304, -1, 0.0009607610409148037), (987515173, 1982, 'Papier_Magazine', 176, -1, 304, -1, 0.0031997081823647022), (987515173, 1982, 'Papier_Magazine', 208, -1, 304, -1, 0.005035818088799715), (987515173, 1982, 'Papier_Magazine', 240, -1, 304, -1, 0.0035042220260947943), (987515173, 1982, 'Papier_Magazine', 272, -1, 304, -1, 0.004052830394357443), (987515173, 1982, 'Papier_Magazine', 304, -1, 304, -1, 0.00725314998999238), (987515173, 1982, 'Papier_Magazine', 336, -1, 304, -1, 0.001950954901985824), (987515173, 1982, 'Papier_Magazine', 112, -1, 336, -1, 0.004847684409469366), (987515173, 1982, 'Papier_Magazine', 144, -1, 336, -1, 0.015508859418332577), (987515173, 1982, 'Papier_Magazine', 176, -1, 336, -1, 0.00679472042247653), (987515173, 1982, 'Papier_Magazine', 208, -1, 336, -1, 0.00797965843230486), (987515173, 1982, 'Papier_Magazine', 240, -1, 336, -1, 0.01901329681277275), (987515173, 1982, 'Papier_Magazine', 272, -1, 336, -1, 0.003658717731013894), (987515173, 1982, 'Papier_Magazine', 304, -1, 336, -1, 0.006297337356954813), (987515173, 1982, 'Papier_Magazine', 336, -1, 336, -1, 0.004447813145816326), (987515173, 1982, 'Plastique', 112, -1, 112, -1, 5.617829046400402e-08), (987515173, 1982, 'Plastique', 144, -1, 112, -1, 8.208448889490683e-07), (987515173, 1982, 'Plastique', 176, -1, 112, -1, 6.256026244955137e-05), (987515173, 1982, 'Plastique', 208, -1, 112, -1, 0.0035499518271535635), (987515173, 1982, 'Plastique', 240, -1, 112, -1, 0.003166872775182128), (987515173, 1982, 'Plastique', 272, -1, 112, -1, 0.007486181333661079), (987515173, 1982, 'Plastique', 304, -1, 112, -1, 0.054866764694452286), (987515173, 1982, 'Plastique', 336, -1, 112, -1, 0.05967129021883011), (987515173, 1982, 'Plastique', 112, -1, 144, -1, 2.791625547615695e-06), (987515173, 1982, 'Plastique', 144, -1, 144, -1, 1.903504744404927e-05), (987515173, 1982, 'Plastique', 176, -1, 144, -1, 0.00017590851348359138), (987515173, 1982, 'Plastique', 208, -1, 144, -1, 0.0015014332020655274), (987515173, 1982, 'Plastique', 240, -1, 144, -1, 0.005060270428657532), (987515173, 1982, 'Plastique', 272, -1, 144, -1, 0.08462032675743103), (987515173, 1982, 'Plastique', 304, -1, 144, -1, 0.4091968834400177), (987515173, 1982, 'Plastique', 336, -1, 144, -1, 0.0710282176733017), (987515173, 1982, 'Plastique', 112, -1, 176, -1, 3.9275050767173525e-06), (987515173, 1982, 'Plastique', 144, -1, 176, -1, 9.517036960460246e-05), (987515173, 1982, 'Plastique', 176, -1, 176, -1, 0.0003402219736017287), (987515173, 1982, 'Plastique', 208, -1, 176, -1, 0.0003888011851813644), (987515173, 1982, 'Plastique', 240, -1, 176, -1, 0.0006235309410840273), (987515173, 1982, 'Plastique', 272, -1, 176, -1, 0.00892091915011406), (987515173, 1982, 'Plastique', 304, -1, 176, -1, 0.037280965596437454), (987515173, 1982, 'Plastique', 336, -1, 176, -1, 0.0020717554725706577), (987515173, 1982, 'Plastique', 112, -1, 208, -1, 7.61425617383793e-05), (987515173, 1982, 'Plastique', 144, -1, 208, -1, 4.0941191400634125e-05), (987515173, 1982, 'Plastique', 176, -1, 208, -1, 8.232118125306442e-05), (987515173, 1982, 'Plastique', 208, -1, 208, -1, 3.834024028037675e-05), (987515173, 1982, 'Plastique', 240, -1, 208, -1, 8.352932127309032e-06), (987515173, 1982, 'Plastique', 272, -1, 208, -1, 1.1998071386187803e-05), (987515173, 1982, 'Plastique', 304, -1, 208, -1, 1.2090074051229749e-05), (987515173, 1982, 'Plastique', 336, -1, 208, -1, 4.101475042261882e-06), (987515173, 1982, 'Plastique', 112, -1, 240, -1, 0.00014464902051258832), (987515173, 1982, 'Plastique', 144, -1, 240, -1, 3.536247095325962e-05), (987515173, 1982, 'Plastique', 176, -1, 240, -1, 2.1159092284506187e-05), (987515173, 1982, 'Plastique', 208, -1, 240, -1, 7.3538108154025394e-06), (987515173, 1982, 'Plastique', 240, -1, 240, -1, 4.867225925409002e-06), (987515173, 1982, 'Plastique', 272, -1, 240, -1, 2.76307764579542e-06), (987515173, 1982, 'Plastique', 304, -1, 240, -1, 7.9772405570111e-07), (987515173, 1982, 'Plastique', 336, -1, 240, -1, 3.538978887718258e-07), (987515173, 1982, 'Plastique', 112, -1, 272, -1, 3.368564648553729e-05), (987515173, 1982, 'Plastique', 144, -1, 272, -1, 1.1664958947221749e-05), (987515173, 1982, 'Plastique', 176, -1, 272, -1, 1.341758525086334e-05), (987515173, 1982, 'Plastique', 208, -1, 272, -1, 5.420888555818237e-06), (987515173, 1982, 'Plastique', 240, -1, 272, -1, 9.547452464175876e-07), (987515173, 1982, 'Plastique', 272, -1, 272, -1, 6.88738509779796e-07), (987515173, 1982, 'Plastique', 304, -1, 272, -1, 5.724524498873507e-07), (987515173, 1982, 'Plastique', 336, -1, 272, -1, 7.429482025145262e-07), (987515173, 1982, 'Plastique', 112, -1, 304, -1, 3.3706512567732716e-06), (987515173, 1982, 'Plastique', 144, -1, 304, -1, 3.355880380695453e-06), (987515173, 1982, 'Plastique', 176, -1, 304, -1, 6.5082558649010025e-06), (987515173, 1982, 'Plastique', 208, -1, 304, -1, 8.736983545531984e-06), (987515173, 1982, 'Plastique', 240, -1, 304, -1, 3.3825183436420048e-06), (987515173, 1982, 'Plastique', 272, -1, 304, -1, 4.746831564261811e-06), (987515173, 1982, 'Plastique', 304, -1, 304, -1, 4.505414835875854e-06), (987515173, 1982, 'Plastique', 336, -1, 304, -1, 2.416980123598478e-06), (987515173, 1982, 'Plastique', 112, -1, 336, -1, 1.5473795428988524e-05), (987515173, 1982, 'Plastique', 144, -1, 336, -1, 5.35884391865693e-05), (987515173, 1982, 'Plastique', 176, -1, 336, -1, 3.427254705457017e-05), (987515173, 1982, 'Plastique', 208, -1, 336, -1, 2.372994458710309e-05), (987515173, 1982, 'Plastique', 240, -1, 336, -1, 3.8594636862399057e-05), (987515173, 1982, 'Plastique', 272, -1, 336, -1, 2.5748187908902764e-05), (987515173, 1982, 'Plastique', 304, -1, 336, -1, 3.3239921322092414e-05), (987515173, 1982, 'Plastique', 336, -1, 336, -1, 4.180555333732627e-05), (987515173, 1982, 'Sol_Environement', 112, -1, 112, -1, 2.930527999087107e-12), (987515173, 1982, 'Sol_Environement', 144, -1, 112, -1, 5.622760035350893e-10), (987515173, 1982, 'Sol_Environement', 176, -1, 112, -1, 4.922356993120047e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 112, -1, 1.0155593372473959e-05), (987515173, 1982, 'Sol_Environement', 240, -1, 112, -1, 9.177741958410479e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 112, -1, 5.625330231850967e-05), (987515173, 1982, 'Sol_Environement', 304, -1, 112, -1, 0.0003788383910432458), (987515173, 1982, 'Sol_Environement', 336, -1, 112, -1, 0.0001651871862122789), (987515173, 1982, 'Sol_Environement', 112, -1, 144, -1, 1.0347219614459391e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 144, -1, 5.797551807518175e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 144, -1, 1.428378482160042e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 144, -1, 5.890514785278356e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 144, -1, 1.824912214942742e-05), (987515173, 1982, 'Sol_Environement', 272, -1, 144, -1, 0.00012608319229912013), (987515173, 1982, 'Sol_Environement', 304, -1, 144, -1, 0.00032353654387407005), (987515173, 1982, 'Sol_Environement', 336, -1, 144, -1, 9.647644037613645e-05), (987515173, 1982, 'Sol_Environement', 112, -1, 176, -1, 4.182305133326736e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 176, -1, 1.0174305771215586e-06), (987515173, 1982, 'Sol_Environement', 176, -1, 176, -1, 6.5862955125339795e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 176, -1, 1.7199390640598722e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 176, -1, 3.7283095934981247e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 176, -1, 2.8019858291372657e-05), (987515173, 1982, 'Sol_Environement', 304, -1, 176, -1, 0.00015382831043098122), (987515173, 1982, 'Sol_Environement', 336, -1, 176, -1, 0.0002425080310786143), (987515173, 1982, 'Sol_Environement', 112, -1, 208, -1, 4.007958523288835e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 208, -1, 7.480698513973039e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 208, -1, 1.3315323030838044e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 208, -1, 9.053310918716306e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 208, -1, 6.944654842300224e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 208, -1, 1.081097934729769e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 208, -1, 1.442125835637853e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 208, -1, 2.321799456694862e-06), (987515173, 1982, 'Sol_Environement', 112, -1, 240, -1, 4.6070831558608916e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 240, -1, 3.7703117072851455e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 240, -1, 2.486881101049221e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 240, -1, 2.1575910125193332e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 240, -1, 5.582672315540549e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 240, -1, 8.621477718406823e-07), (987515173, 1982, 'Sol_Environement', 304, -1, 240, -1, 6.640090646214958e-07), (987515173, 1982, 'Sol_Environement', 336, -1, 240, -1, 7.835926112420566e-07), (987515173, 1982, 'Sol_Environement', 112, -1, 272, -1, 2.9966649890411645e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 272, -1, 6.880858904878551e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 272, -1, 5.021562969886872e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 272, -1, 2.0082485718830867e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 272, -1, 1.3152288147466606e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 272, -1, 2.6994132440449903e-07), (987515173, 1982, 'Sol_Environement', 304, -1, 272, -1, 1.635313680026229e-07), (987515173, 1982, 'Sol_Environement', 336, -1, 272, -1, 4.836798552787513e-07), (987515173, 1982, 'Sol_Environement', 112, -1, 304, -1, 7.164017574723403e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 304, -1, 4.1286557461717166e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 304, -1, 8.938050086726435e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 304, -1, 2.3328091174334986e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 304, -1, 3.0542189506377326e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 304, -1, 2.666334466994158e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 304, -1, 1.4987853091952275e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 304, -1, 1.6114091749841464e-06), (987515173, 1982, 'Sol_Environement', 112, -1, 336, -1, 4.634302399608714e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 336, -1, 1.127728523897531e-06), (987515173, 1982, 'Sol_Environement', 176, -1, 336, -1, 6.802097800573392e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 336, -1, 1.4341401310957735e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 336, -1, 3.3432154395995894e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 336, -1, 2.1192113308643457e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 336, -1, 2.2980780158832204e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 336, -1, 3.3749192880350165e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 112, -1, 4.571711997414241e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 112, -1, 2.4769290121184895e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 112, -1, 1.7669475710135885e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 112, -1, 0.000475196196930483), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 112, -1, 0.00011565273598534986), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 112, -1, 0.00020995673548895866), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 112, -1, 0.0014421100495383143), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 112, -1, 0.0020227127242833376), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 144, -1, 3.082058537984267e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 144, -1, 1.7206231859745458e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 144, -1, 2.0752373529830948e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 144, -1, 0.00027424466679804027), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 144, -1, 0.00046559632755815983), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 144, -1, 0.0005290906992740929), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 144, -1, 0.00018949910008814186), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 144, -1, 0.00020546613086480647), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 176, -1, 1.5011509276519064e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 176, -1, 6.764404588466277e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 176, -1, 2.106750434904825e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 176, -1, 2.4099981601466425e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 176, -1, 3.5370754631003365e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 176, -1, 7.897792238509282e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 176, -1, 0.00010101871885126457), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 176, -1, 0.00037821478326804936), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 208, -1, 1.506751050328603e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 208, -1, 1.4607620641982066e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 208, -1, 4.731698481919011e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 208, -1, 3.825424755632412e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 208, -1, 3.646920276878518e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 208, -1, 1.1670057574519888e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 208, -1, 1.3331959053175524e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 208, -1, 5.995051105855964e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 240, -1, 4.2267558455932885e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 240, -1, 5.323747700458625e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 240, -1, 3.905935500370106e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 240, -1, 2.409000217085122e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 240, -1, 5.77363744014292e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 240, -1, 2.10927701118635e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 240, -1, 1.7813898011809215e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 240, -1, 2.2721291315974668e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 272, -1, 0.00020094112551305443), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 272, -1, 6.691193266306072e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 272, -1, 4.36419177276548e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 272, -1, 1.4556288078892976e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 272, -1, 7.054786692606285e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 272, -1, 1.5496978448936716e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 272, -1, 1.5584806533297524e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 272, -1, 0.00010369653318775818), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 304, -1, 9.524247434455901e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 304, -1, 0.00010966919944621623), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 304, -1, 0.00015802381676621735), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 304, -1, 0.0005423129769042134), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 304, -1, 0.0023449906148016453), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 304, -1, 0.007443359587341547), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 304, -1, 0.005889120511710644), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 304, -1, 0.0044502816163003445), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 336, -1, 0.00024103432951960713), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 336, -1, 0.0020014706533402205), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 336, -1, 0.000750622944906354), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 336, -1, 0.0033519910648465157), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 336, -1, 0.0641678050160408), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 336, -1, 0.03821588680148125), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 336, -1, 0.03885292261838913), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 336, -1, 0.005882319528609514), (987515173, 1982, 'autre_refus', 112, -1, 112, -1, 5.819347781432782e-10), (987515173, 1982, 'autre_refus', 144, -1, 112, -1, 2.0962945157521062e-08), (987515173, 1982, 'autre_refus', 176, -1, 112, -1, 1.4469102325165295e-06), (987515173, 1982, 'autre_refus', 208, -1, 112, -1, 9.239284554496408e-05), (987515173, 1982, 'autre_refus', 240, -1, 112, -1, 0.00021754215413238853), (987515173, 1982, 'autre_refus', 272, -1, 112, -1, 0.0017496768850833178), (987515173, 1982, 'autre_refus', 304, -1, 112, -1, 0.017770571634173393), (987515173, 1982, 'autre_refus', 336, -1, 112, -1, 0.009003167040646076), (987515173, 1982, 'autre_refus', 112, -1, 144, -1, 9.2359130121622e-07), (987515173, 1982, 'autre_refus', 144, -1, 144, -1, 2.9226066544651985e-06), (987515173, 1982, 'autre_refus', 176, -1, 144, -1, 1.0294600542692933e-05), (987515173, 1982, 'autre_refus', 208, -1, 144, -1, 0.0001347130601061508), (987515173, 1982, 'autre_refus', 240, -1, 144, -1, 0.0003480661252979189), (987515173, 1982, 'autre_refus', 272, -1, 144, -1, 0.004933568183332682), (987515173, 1982, 'autre_refus', 304, -1, 144, -1, 0.043510738760232925), (987515173, 1982, 'autre_refus', 336, -1, 144, -1, 0.09554944932460785), (987515173, 1982, 'autre_refus', 112, -1, 176, -1, 1.5122335753403604e-05), (987515173, 1982, 'autre_refus', 144, -1, 176, -1, 0.0001271346991416067), (987515173, 1982, 'autre_refus', 176, -1, 176, -1, 0.00023703543411102146), (987515173, 1982, 'autre_refus', 208, -1, 176, -1, 0.0008105260203592479), (987515173, 1982, 'autre_refus', 240, -1, 176, -1, 0.0006539791356772184), (987515173, 1982, 'autre_refus', 272, -1, 176, -1, 0.0043151527643203735), (987515173, 1982, 'autre_refus', 304, -1, 176, -1, 0.023329641669988632), (987515173, 1982, 'autre_refus', 336, -1, 176, -1, 0.018692055717110634), (987515173, 1982, 'autre_refus', 112, -1, 208, -1, 8.862425602274016e-05), (987515173, 1982, 'autre_refus', 144, -1, 208, -1, 0.00018532731337472796), (987515173, 1982, 'autre_refus', 176, -1, 208, -1, 0.00031896625296212733), (987515173, 1982, 'autre_refus', 208, -1, 208, -1, 0.00035758939338847995), (987515173, 1982, 'autre_refus', 240, -1, 208, -1, 0.00019877831800840795), (987515173, 1982, 'autre_refus', 272, -1, 208, -1, 0.00028634705813601613), (987515173, 1982, 'autre_refus', 304, -1, 208, -1, 0.00020239698642399162), (987515173, 1982, 'autre_refus', 336, -1, 208, -1, 0.0002434736379655078), (987515173, 1982, 'autre_refus', 112, -1, 240, -1, 0.00023386807879433036), (987515173, 1982, 'autre_refus', 144, -1, 240, -1, 0.00010855076106963679), (987515173, 1982, 'autre_refus', 176, -1, 240, -1, 6.48176865070127e-05), (987515173, 1982, 'autre_refus', 208, -1, 240, -1, 2.5300205379608087e-05), (987515173, 1982, 'autre_refus', 240, -1, 240, -1, 7.287785410881042e-05), (987515173, 1982, 'autre_refus', 272, -1, 240, -1, 0.00013982862583361566), (987515173, 1982, 'autre_refus', 304, -1, 240, -1, 8.934963989304379e-05), (987515173, 1982, 'autre_refus', 336, -1, 240, -1, 8.184897160390392e-05), (987515173, 1982, 'autre_refus', 112, -1, 272, -1, 0.0002684956125449389), (987515173, 1982, 'autre_refus', 144, -1, 272, -1, 0.0001116798011935316), (987515173, 1982, 'autre_refus', 176, -1, 272, -1, 0.0001248639600817114), (987515173, 1982, 'autre_refus', 208, -1, 272, -1, 5.112284634378739e-05), (987515173, 1982, 'autre_refus', 240, -1, 272, -1, 2.9130251277820207e-05), (987515173, 1982, 'autre_refus', 272, -1, 272, -1, 4.2736210161820054e-05), (987515173, 1982, 'autre_refus', 304, -1, 272, -1, 6.920337909832597e-05), (987515173, 1982, 'autre_refus', 336, -1, 272, -1, 0.00014184493920765817), (987515173, 1982, 'autre_refus', 112, -1, 304, -1, 0.00011816414189524949), (987515173, 1982, 'autre_refus', 144, -1, 304, -1, 0.00022210566385183483), (987515173, 1982, 'autre_refus', 176, -1, 304, -1, 0.00042709894478321075), (987515173, 1982, 'autre_refus', 208, -1, 304, -1, 0.0004272170481272042), (987515173, 1982, 'autre_refus', 240, -1, 304, -1, 6.55979456496425e-05), (987515173, 1982, 'autre_refus', 272, -1, 304, -1, 3.1694882636656985e-05), (987515173, 1982, 'autre_refus', 304, -1, 304, -1, 1.1630776498350315e-05), (987515173, 1982, 'autre_refus', 336, -1, 304, -1, 1.8845976228476502e-05), (987515173, 1982, 'autre_refus', 112, -1, 336, -1, 0.00024746061535552144), (987515173, 1982, 'autre_refus', 144, -1, 336, -1, 0.00047122384421527386), (987515173, 1982, 'autre_refus', 176, -1, 336, -1, 0.0003342038835398853), (987515173, 1982, 'autre_refus', 208, -1, 336, -1, 0.0002366719563724473), (987515173, 1982, 'autre_refus', 240, -1, 336, -1, 0.00010693204967537895), (987515173, 1982, 'autre_refus', 272, -1, 336, -1, 9.549219976179302e-05), (987515173, 1982, 'autre_refus', 304, -1, 336, -1, 0.000131260123453103), (987515173, 1982, 'autre_refus', 336, -1, 336, -1, 0.000730274710804224)]} result thcl : {'987515187': [('987515187', 'Carton', 0.9811443, 1927, '1528'), 'temp/1748569143_1112947_987515187_9f62f98efd3caca0b9c17d27f5c70440.jpg'], '987515246': [('987515246', 'Carton', 0.9992336, 1927, '1528'), 'temp/1748569143_1112947_987515246_671a708f67f2efa19004b8257fc7b9c8.jpg'], '987515247': [('987515247', 'Carton', 0.9996687, 1927, '1528'), 'temp/1748569143_1112947_987515247_e47b65403df916ba909bc9c439b0af73.jpg'], '987515248': [('987515248', 'Carton', 0.98129404, 1927, '1528'), 'temp/1748569143_1112947_987515248_a70ad88462a22fb62a120721a42b2d42.jpg'], '987515249': [('987515249', 'Carton', 0.98130006, 1927, '1528'), 'temp/1748569143_1112947_987515249_a70ad88462a22fb62a120721a42b2d42.jpg'], '987515250': [('987515250', 'Carton', 0.98082167, 1927, '1528'), 'temp/1748569143_1112947_987515250_b2827c9639df69656f23abcc7f2f82d9.jpg'], '987515224': [('987515224', 'Carton', 0.90838104, 1927, '1528'), 'temp/1748569143_1112947_987515224_e8747b400e713ecbd08d5b75db4d7568.jpg'], '987515226': [('987515226', 'Papier_Magazine', 0.9869912, 1927, '1528'), 'temp/1748569143_1112947_987515226_a18048dca1a77ae086b62cf07759f704.jpg'], '987515239': [('987515239', 'Carton', 0.99978286, 1927, '1528'), 'temp/1748569143_1112947_987515239_b3fa6f29636080b5138c8d8c33fea309.jpg'], '987515240': [('987515240', 'Carton', 0.9995203, 1927, '1528'), 'temp/1748569143_1112947_987515240_7829b9b15f1bf128ea4e2c1a39b9f0dd.jpg'], '987515241': [('987515241', 'Carton', 0.9821817, 1927, '1528'), 'temp/1748569143_1112947_987515241_073420d938f5f010ffd5b4353c064e09.jpg'], '987515242': [('987515242', 'Carton', 0.93587774, 1927, '1528'), 'temp/1748569143_1112947_987515242_327abb5215d6fd1f0aad51f53ed8c324.jpg'], '987515243': [('987515243', 'Papier_Magazine', 0.87416136, 1927, '1528'), 'temp/1748569143_1112947_987515243_4375283f3bc5cdaa431c2fc6f17f53a4.jpg'], '987515244': [('987515244', 'Papier_Magazine', 0.81743693, 1927, '1528'), 'temp/1748569143_1112947_987515244_419530eaef5ef868f75c758b94eea4b4.jpg'], '987515245': [('987515245', 'Carton', 0.8658532, 1927, '1528'), 'temp/1748569143_1112947_987515245_757d9d208d5bd4375c5f21f68b699148.jpg'], '987515235': [('987515235', 'Papier_Magazine', 0.8919676, 1927, '1528'), 'temp/1748569143_1112947_987515235_87075955a2f76b3948b47ffe1825ecd9.jpg'], '987515236': [('987515236', 'Papier_Magazine', 0.53694963, 1927, '1528'), 'temp/1748569143_1112947_987515236_8b44a98b1aceadad73ed000d65836a9a.jpg'], '987515237': [('987515237', 'Carton', 0.770268, 1927, '1528'), 'temp/1748569143_1112947_987515237_1183dfa371a457f11ce2b622c7cf9467.jpg'], '987515238': [('987515238', 'Carton', 0.99957377, 1927, '1528'), 'temp/1748569143_1112947_987515238_e6292cb81e05894cfeb4b99f21a1d3f8.jpg'], '987515188': [('987515188', 'Carton', 0.9956579, 1927, '1528'), 'temp/1748569143_1112947_987515188_4116f9906657a69bb76c2fda982037b9.jpg'], '987515189': [('987515189', 'Carton', 0.99780136, 1927, '1528'), 'temp/1748569143_1112947_987515189_8e8590a26f72249d4c2116dffd0cf668.jpg'], '987515190': [('987515190', 'Carton', 0.9762947, 1927, '1528'), 'temp/1748569143_1112947_987515190_d56932bfc6ba2a8c974c691108755017.jpg'], '987515227': [('987515227', 'Papier_Magazine', 0.9008627, 1927, '1528'), 'temp/1748569143_1112947_987515227_e9c45a0e576ec9e44c1379c3fc5fec7c.jpg'], '987515228': [('987515228', 'Papier_Magazine', 0.52163786, 1927, '1528'), 'temp/1748569143_1112947_987515228_9f1759f20c9e603bccb9f9879d2f0d54.jpg'], '987515230': [('987515230', 'Carton', 0.99940526, 1927, '1528'), 'temp/1748569143_1112947_987515230_846ad925884264181565c81d152a2e94.jpg'], '987515231': [('987515231', 'Carton', 0.9994222, 1927, '1528'), 'temp/1748569143_1112947_987515231_dbf4cafa71b6db4771c5c8f0c25e9cda.jpg'], '987515232': [('987515232', 'Carton', 0.9992473, 1927, '1528'), 'temp/1748569143_1112947_987515232_38db7950cdb3c674ee0ad65915b021f3.jpg'], '987515233': [('987515233', 'Carton', 0.98351353, 1927, '1528'), 'temp/1748569143_1112947_987515233_a92514bed0e8c5724f2d032d3ab1e2ad.jpg'], '987515234': [('987515234', 'Carton', 0.94495124, 1927, '1528'), 'temp/1748569143_1112947_987515234_2eca3480aed0f8b876242675ad99b666.jpg'], '987515202': [('987515202', 'Carton', 0.9911075, 1927, '1528'), 'temp/1748569143_1112947_987515202_3314bd90d1404f31b827d8925abf2d62.jpg'], '987515204': [('987515204', 'Papier_Magazine', 0.9950631, 1927, '1528'), 'temp/1748569143_1112947_987515204_9779c4f9d44360a9c80499e3b01e8a09.jpg'], '987515205': [('987515205', 'Papier_Magazine', 0.99087125, 1927, '1528'), 'temp/1748569143_1112947_987515205_fd4b136d0b3a9a1a347942d7191f6fea.jpg'], '987515207': [('987515207', 'Papier_Magazine', 0.87419236, 1927, '1528'), 'temp/1748569143_1112947_987515207_de216ddb041e249524b0fb2b949064a5.jpg'], '987515208': [('987515208', 'Carton', 0.99174726, 1927, '1528'), 'temp/1748569143_1112947_987515208_a2b90cb74908aa64bbc4aae58f0c5ae8.jpg'], '987515209': [('987515209', 'Carton', 0.96787965, 1927, '1528'), 'temp/1748569143_1112947_987515209_02dfe1ae39f51994652f4a8538844aea.jpg'], '987515211': [('987515211', 'Carton', 0.9733934, 1927, '1528'), 'temp/1748569143_1112947_987515211_72cc7664d45bd40477351b9b764f1500.jpg'], '987515192': [('987515192', 'Papier_Magazine', 0.9999114, 1927, '1528'), 'temp/1748569143_1112947_987515192_b661073b218f5f056833d6af1c617153.jpg'], '987515193': [('987515193', 'Papier_Magazine', 0.9993962, 1927, '1528'), 'temp/1748569143_1112947_987515193_1a97fceb4dcbf5821d783b2e00b52fe6.jpg'], '987515195': [('987515195', 'Carton', 0.98459876, 1927, '1528'), 'temp/1748569143_1112947_987515195_30ccb89dfe410c445878a7f2819ddc36.jpg'], '987515196': [('987515196', 'Carton', 0.9846154, 1927, '1528'), 'temp/1748569143_1112947_987515196_30ccb89dfe410c445878a7f2819ddc36.jpg'], '987515198': [('987515198', 'Carton', 0.9661438, 1927, '1528'), 'temp/1748569143_1112947_987515198_599e80f444c876f407e94b533c89360b.jpg'], '987515200': [('987515200', 'Carton', 0.9859518, 1927, '1528'), 'temp/1748569143_1112947_987515200_978964436b5d5fb0eeda17e3bfafe889.jpg'], '987515201': [('987515201', 'Carton', 0.9954549, 1927, '1528'), 'temp/1748569143_1112947_987515201_b224d2acdc7fa2bbb134c09db6bca7ce.jpg'], '987515222': [('987515222', 'Carton', 0.99747145, 1927, '1528'), 'temp/1748569143_1112947_987515222_067a027bc7402f969b6277d0dcb47eaa.jpg'], '987515223': [('987515223', 'Carton', 0.99210656, 1927, '1528'), 'temp/1748569143_1112947_987515223_ebb57f09941cd11d7ee45a9368a883c1.jpg'], '987515175': [('987515175', 'Papier_Magazine', 0.99981445, 1927, '1528'), 'temp/1748569143_1112947_987515175_8b398cba2f448622cd9657f5eb3f9796.jpg'], '987515176': [('987515176', 'Papier_Magazine', 0.9998135, 1927, '1528'), 'temp/1748569143_1112947_987515176_8b398cba2f448622cd9657f5eb3f9796.jpg'], '987515177': [('987515177', 'Papier_Magazine', 0.9771309, 1927, '1528'), 'temp/1748569143_1112947_987515177_4a54e9967227806219ddf45d256539d8.jpg'], '987515178': [('987515178', 'Carton', 0.857373, 1927, '1528'), 'temp/1748569143_1112947_987515178_298b3d2bfe0fda6787b59a78e2e68867.jpg'], '987515179': [('987515179', 'Carton', 0.9270931, 1927, '1528'), 'temp/1748569143_1112947_987515179_f7d4d1757a470f4c96dc3541eac88b9e.jpg'], '987515212': [('987515212', 'Carton', 0.98691887, 1927, '1528'), 'temp/1748569143_1112947_987515212_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg'], '987515213': [('987515213', 'Carton', 0.98693603, 1927, '1528'), 'temp/1748569143_1112947_987515213_b0a038fcb9678ebfd60d9b1f6ec1fc17.jpg'], '987515215': [('987515215', 'Papier_Magazine', 0.9939308, 1927, '1528'), 'temp/1748569143_1112947_987515215_902ef348a7eebb9a8b87f42927347936.jpg'], '987515216': [('987515216', 'Papier_Magazine', 0.97749305, 1927, '1528'), 'temp/1748569143_1112947_987515216_4f7dc21f1d2cd3fcabadc4a6755921e1.jpg'], '987515217': [('987515217', 'Carton', 0.52847713, 1927, '1528'), 'temp/1748569143_1112947_987515217_78877bb2c5760be28518d17f77d1c609.jpg'], '987515219': [('987515219', 'Carton', 0.9993692, 1927, '1528'), 'temp/1748569143_1112947_987515219_c2d417a5ba6ccf7c84527636f8d5eef9.jpg'], '987515220': [('987515220', 'Carton', 0.99638546, 1927, '1528'), 'temp/1748569143_1112947_987515220_e729f316c4c3b32049adfbaaa336d95c.jpg'], '987515180': [('987515180', 'Carton', 0.98998404, 1927, '1528'), 'temp/1748569143_1112947_987515180_776a5d7d8486ee2961bbe3a0d90f95b5.jpg'], '987515181': [('987515181', 'Carton', 0.99777883, 1927, '1528'), 'temp/1748569143_1112947_987515181_1738c2798fb31152809ecb443ac286d6.jpg'], '987515182': [('987515182', 'Carton', 0.99242836, 1927, '1528'), 'temp/1748569143_1112947_987515182_fe7f29bf6d13e08c3e985f91b5232178.jpg'], '987515183': [('987515183', 'Papier_Magazine', 0.99999213, 1927, '1528'), 'temp/1748569143_1112947_987515183_6aab9ca0421398b4899892c10c2594c6.jpg'], '987515184': [('987515184', 'Papier_Magazine', 0.99973184, 1927, '1528'), 'temp/1748569143_1112947_987515184_19c8c2177209a285df6014d95fe53f2c.jpg'], '987515185': [('987515185', 'Papier_Magazine', 0.7978397, 1927, '1528'), 'temp/1748569143_1112947_987515185_e172d54457cabee9d7f02ee1300f3ae9.jpg'], '987515186': [('987515186', 'Carton', 0.9847239, 1927, '1528'), 'temp/1748569143_1112947_987515186_797def426440b544aa80dbd63a19234a.jpg']} result detect_point : {987515173: [(987515173, 1982, 'Autre_Environement', 112, -1, 112, -1, 6.271815965880334e-12), (987515173, 1982, 'Autre_Environement', 144, -1, 112, -1, 2.4844125021128427e-11), (987515173, 1982, 'Autre_Environement', 176, -1, 112, -1, 1.0650548887269906e-08), (987515173, 1982, 'Autre_Environement', 208, -1, 112, -1, 4.448247921118309e-07), (987515173, 1982, 'Autre_Environement', 240, -1, 112, -1, 1.923155878102989e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 112, -1, 3.779579492402263e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 112, -1, 0.00012279744260013103), (987515173, 1982, 'Autre_Environement', 336, -1, 112, -1, 2.9485890991054475e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 144, -1, 2.374503260682559e-08), (987515173, 1982, 'Autre_Environement', 144, -1, 144, -1, 2.2141357192140276e-08), (987515173, 1982, 'Autre_Environement', 176, -1, 144, -1, 1.3738434745391714e-07), (987515173, 1982, 'Autre_Environement', 208, -1, 144, -1, 1.4727493180544116e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 144, -1, 1.1291859664197545e-05), (987515173, 1982, 'Autre_Environement', 272, -1, 144, -1, 0.00015815612277947366), (987515173, 1982, 'Autre_Environement', 304, -1, 144, -1, 0.0004434996226336807), (987515173, 1982, 'Autre_Environement', 336, -1, 144, -1, 6.545944779645652e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 176, -1, 1.3293632719069137e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 176, -1, 1.6162027804966783e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 176, -1, 2.526713387851487e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 176, -1, 1.615999963178183e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 176, -1, 6.257173481571954e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 176, -1, 8.64791072672233e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 176, -1, 0.0003271224850323051), (987515173, 1982, 'Autre_Environement', 336, -1, 176, -1, 0.0003056900459341705), (987515173, 1982, 'Autre_Environement', 112, -1, 208, -1, 1.860383417806588e-05), (987515173, 1982, 'Autre_Environement', 144, -1, 208, -1, 7.921114047348965e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 208, -1, 2.7036321625928394e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 208, -1, 1.801614234864246e-05), (987515173, 1982, 'Autre_Environement', 240, -1, 208, -1, 2.3417936972691678e-05), (987515173, 1982, 'Autre_Environement', 272, -1, 208, -1, 1.694694765319582e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 208, -1, 4.544725925370585e-06), (987515173, 1982, 'Autre_Environement', 336, -1, 208, -1, 8.81922915141331e-06), (987515173, 1982, 'Autre_Environement', 112, -1, 240, -1, 6.098133781051729e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 240, -1, 1.6467932937302976e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 240, -1, 1.963200020327349e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 240, -1, 1.4263438288253383e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 240, -1, 7.862996426410973e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 240, -1, 1.286251972487662e-05), (987515173, 1982, 'Autre_Environement', 304, -1, 240, -1, 9.291064088756684e-06), (987515173, 1982, 'Autre_Environement', 336, -1, 240, -1, 2.1700687284464948e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 272, -1, 3.8262664929789025e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 272, -1, 2.5499746243440313e-06), (987515173, 1982, 'Autre_Environement', 176, -1, 272, -1, 2.96101507046842e-06), (987515173, 1982, 'Autre_Environement', 208, -1, 272, -1, 2.7530711577128386e-06), (987515173, 1982, 'Autre_Environement', 240, -1, 272, -1, 4.317397269915091e-06), (987515173, 1982, 'Autre_Environement', 272, -1, 272, -1, 8.171764420694672e-06), (987515173, 1982, 'Autre_Environement', 304, -1, 272, -1, 1.1662354154395871e-05), (987515173, 1982, 'Autre_Environement', 336, -1, 272, -1, 3.91181129089091e-05), (987515173, 1982, 'Autre_Environement', 112, -1, 304, -1, 1.238445929629961e-05), (987515173, 1982, 'Autre_Environement', 144, -1, 304, -1, 1.595929279574193e-05), (987515173, 1982, 'Autre_Environement', 176, -1, 304, -1, 3.3600517781451344e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 304, -1, 0.00015436304965987802), (987515173, 1982, 'Autre_Environement', 240, -1, 304, -1, 0.0002585948968771845), (987515173, 1982, 'Autre_Environement', 272, -1, 304, -1, 0.0001877176546258852), (987515173, 1982, 'Autre_Environement', 304, -1, 304, -1, 0.00021357914374675602), (987515173, 1982, 'Autre_Environement', 336, -1, 304, -1, 0.00016418426821473986), (987515173, 1982, 'Autre_Environement', 112, -1, 336, -1, 4.553502094495343e-06), (987515173, 1982, 'Autre_Environement', 144, -1, 336, -1, 1.733617318677716e-05), (987515173, 1982, 'Autre_Environement', 176, -1, 336, -1, 4.9303725973004475e-05), (987515173, 1982, 'Autre_Environement', 208, -1, 336, -1, 0.0001215945158037357), (987515173, 1982, 'Autre_Environement', 240, -1, 336, -1, 0.00019625374989118427), (987515173, 1982, 'Autre_Environement', 272, -1, 336, -1, 0.00018786026339512318), (987515173, 1982, 'Autre_Environement', 304, -1, 336, -1, 0.00012362762936390936), (987515173, 1982, 'Autre_Environement', 336, -1, 336, -1, 0.0002712809364311397), (987515173, 1982, 'Carton', 112, -1, 112, -1, 1.576355685983799e-07), (987515173, 1982, 'Carton', 144, -1, 112, -1, 4.09165568271419e-06), (987515173, 1982, 'Carton', 176, -1, 112, -1, 6.995676812948659e-06), (987515173, 1982, 'Carton', 208, -1, 112, -1, 0.0008727198000997305), (987515173, 1982, 'Carton', 240, -1, 112, -1, 0.0026501729153096676), (987515173, 1982, 'Carton', 272, -1, 112, -1, 0.003380689537152648), (987515173, 1982, 'Carton', 304, -1, 112, -1, 0.03130565211176872), (987515173, 1982, 'Carton', 336, -1, 112, -1, 0.05585074424743652), (987515173, 1982, 'Carton', 112, -1, 144, -1, 0.00012447437620721757), (987515173, 1982, 'Carton', 144, -1, 144, -1, 0.00020915831555612385), (987515173, 1982, 'Carton', 176, -1, 144, -1, 0.0003673181345220655), (987515173, 1982, 'Carton', 208, -1, 144, -1, 0.006835674401372671), (987515173, 1982, 'Carton', 240, -1, 144, -1, 0.01591036096215248), (987515173, 1982, 'Carton', 272, -1, 144, -1, 0.009403361938893795), (987515173, 1982, 'Carton', 304, -1, 144, -1, 0.00975574180483818), (987515173, 1982, 'Carton', 336, -1, 144, -1, 0.022150758653879166), (987515173, 1982, 'Carton', 112, -1, 176, -1, 0.02188420481979847), (987515173, 1982, 'Carton', 144, -1, 176, -1, 0.19292433559894562), (987515173, 1982, 'Carton', 176, -1, 176, -1, 0.09658340364694595), (987515173, 1982, 'Carton', 208, -1, 176, -1, 0.12378031760454178), (987515173, 1982, 'Carton', 240, -1, 176, -1, 0.5330907702445984), (987515173, 1982, 'Carton', 272, -1, 176, -1, 0.4615733027458191), (987515173, 1982, 'Carton', 304, -1, 176, -1, 0.7711007595062256), (987515173, 1982, 'Carton', 336, -1, 176, -1, 0.8663382530212402), (987515173, 1982, 'Carton', 112, -1, 208, -1, 0.8503047227859497), (987515173, 1982, 'Carton', 144, -1, 208, -1, 0.9844670295715332), (987515173, 1982, 'Carton', 176, -1, 208, -1, 0.9847484230995178), (987515173, 1982, 'Carton', 208, -1, 208, -1, 0.9919507503509521), (987515173, 1982, 'Carton', 240, -1, 208, -1, 0.9993792772293091), (987515173, 1982, 'Carton', 272, -1, 208, -1, 0.9994138479232788), (987515173, 1982, 'Carton', 304, -1, 208, -1, 0.9995881915092468), (987515173, 1982, 'Carton', 336, -1, 208, -1, 0.9992272853851318), (987515173, 1982, 'Carton', 112, -1, 240, -1, 0.9276415705680847), (987515173, 1982, 'Carton', 144, -1, 240, -1, 0.9810431003570557), (987515173, 1982, 'Carton', 176, -1, 240, -1, 0.966120719909668), (987515173, 1982, 'Carton', 208, -1, 240, -1, 0.9677486419677734), (987515173, 1982, 'Carton', 240, -1, 240, -1, 0.996390163898468), (987515173, 1982, 'Carton', 272, -1, 240, -1, 0.9994217157363892), (987515173, 1982, 'Carton', 304, -1, 240, -1, 0.9997863173484802), (987515173, 1982, 'Carton', 336, -1, 240, -1, 0.9996683597564697), (987515173, 1982, 'Carton', 112, -1, 272, -1, 0.9895250797271729), (987515173, 1982, 'Carton', 144, -1, 272, -1, 0.995465874671936), (987515173, 1982, 'Carton', 176, -1, 272, -1, 0.9854716658592224), (987515173, 1982, 'Carton', 208, -1, 272, -1, 0.9733673334121704), (987515173, 1982, 'Carton', 240, -1, 272, -1, 0.9974767565727234), (987515173, 1982, 'Carton', 272, -1, 272, -1, 0.9992011189460754), (987515173, 1982, 'Carton', 304, -1, 272, -1, 0.9995144605636597), (987515173, 1982, 'Carton', 336, -1, 272, -1, 0.9991315007209778), (987515173, 1982, 'Carton', 112, -1, 304, -1, 0.9977748990058899), (987515173, 1982, 'Carton', 144, -1, 304, -1, 0.9977632761001587), (987515173, 1982, 'Carton', 176, -1, 304, -1, 0.9955496788024902), (987515173, 1982, 'Carton', 208, -1, 304, -1, 0.9927307963371277), (987515173, 1982, 'Carton', 240, -1, 304, -1, 0.9920262694358826), (987515173, 1982, 'Carton', 272, -1, 304, -1, 0.9835942387580872), (987515173, 1982, 'Carton', 304, -1, 304, -1, 0.9819222688674927), (987515173, 1982, 'Carton', 336, -1, 304, -1, 0.9809139370918274), (987515173, 1982, 'Carton', 112, -1, 336, -1, 0.9924526810646057), (987515173, 1982, 'Carton', 144, -1, 336, -1, 0.9762071967124939), (987515173, 1982, 'Carton', 176, -1, 336, -1, 0.99116450548172), (987515173, 1982, 'Carton', 208, -1, 336, -1, 0.9869438409805298), (987515173, 1982, 'Carton', 240, -1, 336, -1, 0.9085468053817749), (987515173, 1982, 'Carton', 272, -1, 336, -1, 0.9452228546142578), (987515173, 1982, 'Carton', 304, -1, 336, -1, 0.9365962147712708), (987515173, 1982, 'Carton', 336, -1, 336, -1, 0.9808369278907776), (987515173, 1982, 'Kraft', 112, -1, 112, -1, 1.957737527646941e-09), (987515173, 1982, 'Kraft', 144, -1, 112, -1, 1.720074749300693e-08), (987515173, 1982, 'Kraft', 176, -1, 112, -1, 9.63709453571937e-07), (987515173, 1982, 'Kraft', 208, -1, 112, -1, 3.138272586511448e-05), (987515173, 1982, 'Kraft', 240, -1, 112, -1, 4.440563134266995e-05), (987515173, 1982, 'Kraft', 272, -1, 112, -1, 0.00020620097348000854), (987515173, 1982, 'Kraft', 304, -1, 112, -1, 0.0010808930965140462), (987515173, 1982, 'Kraft', 336, -1, 112, -1, 0.0008296957821585238), (987515173, 1982, 'Kraft', 112, -1, 144, -1, 2.638400110299699e-05), (987515173, 1982, 'Kraft', 144, -1, 144, -1, 6.993338956817752e-06), (987515173, 1982, 'Kraft', 176, -1, 144, -1, 3.622401436587097e-06), (987515173, 1982, 'Kraft', 208, -1, 144, -1, 3.556726733222604e-05), (987515173, 1982, 'Kraft', 240, -1, 144, -1, 6.707842112518847e-05), (987515173, 1982, 'Kraft', 272, -1, 144, -1, 8.681361214257777e-05), (987515173, 1982, 'Kraft', 304, -1, 144, -1, 0.00012179985060356557), (987515173, 1982, 'Kraft', 336, -1, 144, -1, 0.00011247207294218242), (987515173, 1982, 'Kraft', 112, -1, 176, -1, 0.0004987759166397154), (987515173, 1982, 'Kraft', 144, -1, 176, -1, 0.00012291625898797065), (987515173, 1982, 'Kraft', 176, -1, 176, -1, 9.114220301853493e-05), (987515173, 1982, 'Kraft', 208, -1, 176, -1, 5.1748036639764905e-05), (987515173, 1982, 'Kraft', 240, -1, 176, -1, 0.00011549148621270433), (987515173, 1982, 'Kraft', 272, -1, 176, -1, 0.00043214470497332513), (987515173, 1982, 'Kraft', 304, -1, 176, -1, 0.0009234515018761158), (987515173, 1982, 'Kraft', 336, -1, 176, -1, 0.0014272574335336685), (987515173, 1982, 'Kraft', 112, -1, 208, -1, 6.90045126248151e-05), (987515173, 1982, 'Kraft', 144, -1, 208, -1, 1.8495431504561566e-05), (987515173, 1982, 'Kraft', 176, -1, 208, -1, 2.5861496396828443e-05), (987515173, 1982, 'Kraft', 208, -1, 208, -1, 3.5416676837485284e-05), (987515173, 1982, 'Kraft', 240, -1, 208, -1, 3.73744624084793e-05), (987515173, 1982, 'Kraft', 272, -1, 208, -1, 8.635753329144791e-05), (987515173, 1982, 'Kraft', 304, -1, 208, -1, 0.00012362583947833627), (987515173, 1982, 'Kraft', 336, -1, 208, -1, 0.0003905165649484843), (987515173, 1982, 'Kraft', 112, -1, 240, -1, 0.0003079257730860263), (987515173, 1982, 'Kraft', 144, -1, 240, -1, 4.163364792475477e-05), (987515173, 1982, 'Kraft', 176, -1, 240, -1, 1.2242019693076145e-05), (987515173, 1982, 'Kraft', 208, -1, 240, -1, 7.337544502661331e-06), (987515173, 1982, 'Kraft', 240, -1, 240, -1, 2.2982125301496126e-05), (987515173, 1982, 'Kraft', 272, -1, 240, -1, 5.8157253079116344e-05), (987515173, 1982, 'Kraft', 304, -1, 240, -1, 6.560523615917191e-05), (987515173, 1982, 'Kraft', 336, -1, 240, -1, 0.00018724572146311402), (987515173, 1982, 'Kraft', 112, -1, 272, -1, 0.0014629343058913946), (987515173, 1982, 'Kraft', 144, -1, 272, -1, 0.0006902696331962943), (987515173, 1982, 'Kraft', 176, -1, 272, -1, 0.0002742304641287774), (987515173, 1982, 'Kraft', 208, -1, 272, -1, 4.357844591140747e-05), (987515173, 1982, 'Kraft', 240, -1, 272, -1, 3.340828698128462e-05), (987515173, 1982, 'Kraft', 272, -1, 272, -1, 8.349671406904235e-05), (987515173, 1982, 'Kraft', 304, -1, 272, -1, 0.00011392419401090592), (987515173, 1982, 'Kraft', 336, -1, 272, -1, 0.0004221333365421742), (987515173, 1982, 'Kraft', 112, -1, 304, -1, 0.001014246023260057), (987515173, 1982, 'Kraft', 144, -1, 304, -1, 0.0009221484651789069), (987515173, 1982, 'Kraft', 176, -1, 304, -1, 0.0006190584390424192), (987515173, 1982, 'Kraft', 208, -1, 304, -1, 0.001081658760085702), (987515173, 1982, 'Kraft', 240, -1, 304, -1, 0.0017836974002420902), (987515173, 1982, 'Kraft', 272, -1, 304, -1, 0.004672076087445021), (987515173, 1982, 'Kraft', 304, -1, 304, -1, 0.0046897633001208305), (987515173, 1982, 'Kraft', 336, -1, 304, -1, 0.012484777718782425), (987515173, 1982, 'Kraft', 112, -1, 336, -1, 0.002183066913858056), (987515173, 1982, 'Kraft', 144, -1, 336, -1, 0.005705210845917463), (987515173, 1982, 'Kraft', 176, -1, 336, -1, 0.0008308480028063059), (987515173, 1982, 'Kraft', 208, -1, 336, -1, 0.001263151061721146), (987515173, 1982, 'Kraft', 240, -1, 336, -1, 0.007801832631230354), (987515173, 1982, 'Kraft', 272, -1, 336, -1, 0.012551669031381607), (987515173, 1982, 'Kraft', 304, -1, 336, -1, 0.017930345609784126), (987515173, 1982, 'Kraft', 336, -1, 336, -1, 0.007753178011626005), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 112, -1, 1.4981969831406872e-10), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 112, -1, 8.321308087033685e-09), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 112, -1, 5.521291654986271e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 112, -1, 5.514806161954766e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 112, -1, 8.237828296842054e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 112, -1, 3.563874997780658e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 112, -1, 8.149821951519698e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 112, -1, 4.2202686017844826e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 144, -1, 3.643475281478459e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 144, -1, 1.3091824939692742e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 144, -1, 2.465325906086946e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 144, -1, 3.4592671909194905e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 144, -1, 5.710682671633549e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 144, -1, 2.059975486190524e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 144, -1, 3.9010625187074766e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 144, -1, 1.3684768418897875e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 176, -1, 3.892207303124451e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 176, -1, 2.55041777563747e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 176, -1, 1.2033554412482772e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 176, -1, 7.819895472493954e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 176, -1, 6.353224762278842e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 176, -1, 2.6373883883934468e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 176, -1, 3.922960240743123e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 176, -1, 1.8472530427970923e-05), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 208, -1, 1.7999934698309517e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 208, -1, 5.661302111548139e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 208, -1, 2.1980697511025937e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 208, -1, 1.7350848793284968e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 208, -1, 3.996210864443128e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 208, -1, 2.027948937666224e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 208, -1, 9.864364614031729e-08), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 208, -1, 8.694066622183527e-08), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 240, -1, 1.822134095164074e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 240, -1, 6.324926289380528e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 240, -1, 7.821980716471444e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 240, -1, 6.381894195328641e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 240, -1, 7.713734362368996e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 240, -1, 5.179261961529846e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 240, -1, 1.772643258846074e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 240, -1, 1.4453388530455413e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 272, -1, 4.1830210761872877e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 272, -1, 3.170695208609686e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 272, -1, 5.096800350656849e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 272, -1, 7.328864057853934e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 272, -1, 4.0205608797805326e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 272, -1, 3.2233640467893565e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 272, -1, 2.121068121141434e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 272, -1, 3.4728486753010657e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 304, -1, 3.321688666346745e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 304, -1, 3.2591808007964573e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 304, -1, 8.38333505726041e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 304, -1, 3.2999407721945317e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 304, -1, 4.413991973706288e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 304, -1, 5.0191547416034155e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 304, -1, 8.209121915569995e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 304, -1, 5.568050255533308e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 112, -1, 336, -1, 4.1875182432704605e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 144, -1, 336, -1, 8.223066743084928e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 176, -1, 336, -1, 7.151177214836935e-07), (987515173, 1982, 'Lointain_Papier_Magazine', 208, -1, 336, -1, 1.9496640106808627e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 240, -1, 336, -1, 3.911284238711232e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 272, -1, 336, -1, 2.4560574729548534e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 304, -1, 336, -1, 3.435417056607548e-06), (987515173, 1982, 'Lointain_Papier_Magazine', 336, -1, 336, -1, 4.8369684009230696e-06), (987515173, 1982, 'Metal', 112, -1, 112, -1, 7.40447078650952e-11), (987515173, 1982, 'Metal', 144, -1, 112, -1, 1.9585983945802354e-09), (987515173, 1982, 'Metal', 176, -1, 112, -1, 7.308526619453914e-07), (987515173, 1982, 'Metal', 208, -1, 112, -1, 9.458058229938615e-06), (987515173, 1982, 'Metal', 240, -1, 112, -1, 2.6567093300400302e-05), (987515173, 1982, 'Metal', 272, -1, 112, -1, 0.00016929887351579964), (987515173, 1982, 'Metal', 304, -1, 112, -1, 0.0007552222232334316), (987515173, 1982, 'Metal', 336, -1, 112, -1, 0.0003030860680155456), (987515173, 1982, 'Metal', 112, -1, 144, -1, 5.366252366911795e-07), (987515173, 1982, 'Metal', 144, -1, 144, -1, 4.3793150439341844e-07), (987515173, 1982, 'Metal', 176, -1, 144, -1, 3.176887958034058e-06), (987515173, 1982, 'Metal', 208, -1, 144, -1, 9.592667083779816e-06), (987515173, 1982, 'Metal', 240, -1, 144, -1, 1.658712062635459e-05), (987515173, 1982, 'Metal', 272, -1, 144, -1, 7.65814766054973e-05), (987515173, 1982, 'Metal', 304, -1, 144, -1, 0.00013093012967146933), (987515173, 1982, 'Metal', 336, -1, 144, -1, 4.670989801525138e-05), (987515173, 1982, 'Metal', 112, -1, 176, -1, 5.880497610633029e-06), (987515173, 1982, 'Metal', 144, -1, 176, -1, 4.011907549283933e-06), (987515173, 1982, 'Metal', 176, -1, 176, -1, 1.0134302101505455e-05), (987515173, 1982, 'Metal', 208, -1, 176, -1, 4.159005129622528e-06), (987515173, 1982, 'Metal', 240, -1, 176, -1, 6.656815003225347e-06), (987515173, 1982, 'Metal', 272, -1, 176, -1, 5.4316413297783583e-05), (987515173, 1982, 'Metal', 304, -1, 176, -1, 0.00012912174861412495), (987515173, 1982, 'Metal', 336, -1, 176, -1, 6.557787855854258e-05), (987515173, 1982, 'Metal', 112, -1, 208, -1, 8.146424988808576e-06), (987515173, 1982, 'Metal', 144, -1, 208, -1, 2.247384600195801e-06), (987515173, 1982, 'Metal', 176, -1, 208, -1, 6.634919373027515e-06), (987515173, 1982, 'Metal', 208, -1, 208, -1, 4.596822236635489e-06), (987515173, 1982, 'Metal', 240, -1, 208, -1, 1.7126073998952052e-06), (987515173, 1982, 'Metal', 272, -1, 208, -1, 1.3694709650735604e-06), (987515173, 1982, 'Metal', 304, -1, 208, -1, 5.784029326605378e-07), (987515173, 1982, 'Metal', 336, -1, 208, -1, 8.349181257472083e-07), (987515173, 1982, 'Metal', 112, -1, 240, -1, 2.9274015105329454e-06), (987515173, 1982, 'Metal', 144, -1, 240, -1, 7.392775955850084e-07), (987515173, 1982, 'Metal', 176, -1, 240, -1, 6.081426136006485e-07), (987515173, 1982, 'Metal', 208, -1, 240, -1, 6.215839221113129e-07), (987515173, 1982, 'Metal', 240, -1, 240, -1, 1.1986741128566791e-06), (987515173, 1982, 'Metal', 272, -1, 240, -1, 6.961149097151065e-07), (987515173, 1982, 'Metal', 304, -1, 240, -1, 3.145585480979207e-07), (987515173, 1982, 'Metal', 336, -1, 240, -1, 4.122880170598364e-07), (987515173, 1982, 'Metal', 112, -1, 272, -1, 2.620731265778886e-06), (987515173, 1982, 'Metal', 144, -1, 272, -1, 8.054418572100985e-07), (987515173, 1982, 'Metal', 176, -1, 272, -1, 8.229471291087975e-07), (987515173, 1982, 'Metal', 208, -1, 272, -1, 4.01219210743875e-07), (987515173, 1982, 'Metal', 240, -1, 272, -1, 2.7388938406147645e-07), (987515173, 1982, 'Metal', 272, -1, 272, -1, 3.310874774342665e-07), (987515173, 1982, 'Metal', 304, -1, 272, -1, 3.4722987152235874e-07), (987515173, 1982, 'Metal', 336, -1, 272, -1, 1.4728271935382509e-06), (987515173, 1982, 'Metal', 112, -1, 304, -1, 2.352227966184728e-06), (987515173, 1982, 'Metal', 144, -1, 304, -1, 1.9818360215140274e-06), (987515173, 1982, 'Metal', 176, -1, 304, -1, 4.541454472928308e-06), (987515173, 1982, 'Metal', 208, -1, 304, -1, 1.3329069588507991e-05), (987515173, 1982, 'Metal', 240, -1, 304, -1, 5.641319603455486e-06), (987515173, 1982, 'Metal', 272, -1, 304, -1, 5.678331945091486e-06), (987515173, 1982, 'Metal', 304, -1, 304, -1, 6.368874437612249e-06), (987515173, 1982, 'Metal', 336, -1, 304, -1, 7.368290880549466e-06), (987515173, 1982, 'Metal', 112, -1, 336, -1, 7.137274224078283e-06), (987515173, 1982, 'Metal', 144, -1, 336, -1, 3.321443364256993e-05), (987515173, 1982, 'Metal', 176, -1, 336, -1, 4.011798955616541e-05), (987515173, 1982, 'Metal', 208, -1, 336, -1, 7.578751683467999e-05), (987515173, 1982, 'Metal', 240, -1, 336, -1, 0.00012133477139286697), (987515173, 1982, 'Metal', 272, -1, 336, -1, 3.714204649440944e-05), (987515173, 1982, 'Metal', 304, -1, 336, -1, 2.9102686312398873e-05), (987515173, 1982, 'Metal', 336, -1, 336, -1, 2.8008123990730383e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 112, -1, 0.9999953508377075), (987515173, 1982, 'Papier_Magazine', 144, -1, 112, -1, 0.9999924898147583), (987515173, 1982, 'Papier_Magazine', 176, -1, 112, -1, 0.9999085664749146), (987515173, 1982, 'Papier_Magazine', 208, -1, 112, -1, 0.9949527978897095), (987515173, 1982, 'Papier_Magazine', 240, -1, 112, -1, 0.9937595725059509), (987515173, 1982, 'Papier_Magazine', 272, -1, 112, -1, 0.9866683483123779), (987515173, 1982, 'Papier_Magazine', 304, -1, 112, -1, 0.8921955227851868), (987515173, 1982, 'Papier_Magazine', 336, -1, 112, -1, 0.8720824122428894), (987515173, 1982, 'Papier_Magazine', 112, -1, 144, -1, 0.9998136162757874), (987515173, 1982, 'Papier_Magazine', 144, -1, 144, -1, 0.9997422099113464), (987515173, 1982, 'Papier_Magazine', 176, -1, 144, -1, 0.9994150400161743), (987515173, 1982, 'Papier_Magazine', 208, -1, 144, -1, 0.9911978840827942), (987515173, 1982, 'Papier_Magazine', 240, -1, 144, -1, 0.9780967831611633), (987515173, 1982, 'Papier_Magazine', 272, -1, 144, -1, 0.9000454545021057), (987515173, 1982, 'Papier_Magazine', 304, -1, 144, -1, 0.5362882614135742), (987515173, 1982, 'Papier_Magazine', 336, -1, 144, -1, 0.8107313513755798), (987515173, 1982, 'Papier_Magazine', 112, -1, 176, -1, 0.97757488489151), (987515173, 1982, 'Papier_Magazine', 144, -1, 176, -1, 0.8067144155502319), (987515173, 1982, 'Papier_Magazine', 176, -1, 176, -1, 0.9026957154273987), (987515173, 1982, 'Papier_Magazine', 208, -1, 176, -1, 0.8749292492866516), (987515173, 1982, 'Papier_Magazine', 240, -1, 176, -1, 0.46545782685279846), (987515173, 1982, 'Papier_Magazine', 272, -1, 176, -1, 0.5244843363761902), (987515173, 1982, 'Papier_Magazine', 304, -1, 176, -1, 0.16661480069160461), (987515173, 1982, 'Papier_Magazine', 336, -1, 176, -1, 0.11046022921800613), (987515173, 1982, 'Papier_Magazine', 112, -1, 208, -1, 0.1494138389825821), (987515173, 1982, 'Papier_Magazine', 144, -1, 208, -1, 0.015275389887392521), (987515173, 1982, 'Papier_Magazine', 176, -1, 208, -1, 0.014782627113163471), (987515173, 1982, 'Papier_Magazine', 208, -1, 208, -1, 0.007588740438222885), (987515173, 1982, 'Papier_Magazine', 240, -1, 208, -1, 0.0003462004824541509), (987515173, 1982, 'Papier_Magazine', 272, -1, 208, -1, 0.00017021365056280047), (987515173, 1982, 'Papier_Magazine', 304, -1, 208, -1, 5.368440906750038e-05), (987515173, 1982, 'Papier_Magazine', 336, -1, 208, -1, 6.279405351961032e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 240, -1, 0.0716143548488617), (987515173, 1982, 'Papier_Magazine', 144, -1, 240, -1, 0.01876274310052395), (987515173, 1982, 'Papier_Magazine', 176, -1, 240, -1, 0.03377354517579079), (987515173, 1982, 'Papier_Magazine', 208, -1, 240, -1, 0.03220600262284279), (987515173, 1982, 'Papier_Magazine', 240, -1, 240, -1, 0.003492899937555194), (987515173, 1982, 'Papier_Magazine', 272, -1, 240, -1, 0.00034153219894506037), (987515173, 1982, 'Papier_Magazine', 304, -1, 240, -1, 2.9719671147176996e-05), (987515173, 1982, 'Papier_Magazine', 336, -1, 240, -1, 1.6387210052926093e-05), (987515173, 1982, 'Papier_Magazine', 112, -1, 272, -1, 0.008499102666974068), (987515173, 1982, 'Papier_Magazine', 144, -1, 272, -1, 0.003649174701422453), (987515173, 1982, 'Papier_Magazine', 176, -1, 272, -1, 0.014067606069147587), (987515173, 1982, 'Papier_Magazine', 208, -1, 272, -1, 0.02651386708021164), (987515173, 1982, 'Papier_Magazine', 240, -1, 272, -1, 0.0024477692786604166), (987515173, 1982, 'Papier_Magazine', 272, -1, 272, -1, 0.0006472524837590754), (987515173, 1982, 'Papier_Magazine', 304, -1, 272, -1, 0.0002737916074693203), (987515173, 1982, 'Papier_Magazine', 336, -1, 272, -1, 0.00015876929683145136), (987515173, 1982, 'Papier_Magazine', 112, -1, 304, -1, 0.0009783600689843297), (987515173, 1982, 'Papier_Magazine', 144, -1, 304, -1, 0.0009607610409148037), (987515173, 1982, 'Papier_Magazine', 176, -1, 304, -1, 0.0031997081823647022), (987515173, 1982, 'Papier_Magazine', 208, -1, 304, -1, 0.005035818088799715), (987515173, 1982, 'Papier_Magazine', 240, -1, 304, -1, 0.0035042220260947943), (987515173, 1982, 'Papier_Magazine', 272, -1, 304, -1, 0.004052830394357443), (987515173, 1982, 'Papier_Magazine', 304, -1, 304, -1, 0.00725314998999238), (987515173, 1982, 'Papier_Magazine', 336, -1, 304, -1, 0.001950954901985824), (987515173, 1982, 'Papier_Magazine', 112, -1, 336, -1, 0.004847684409469366), (987515173, 1982, 'Papier_Magazine', 144, -1, 336, -1, 0.015508859418332577), (987515173, 1982, 'Papier_Magazine', 176, -1, 336, -1, 0.00679472042247653), (987515173, 1982, 'Papier_Magazine', 208, -1, 336, -1, 0.00797965843230486), (987515173, 1982, 'Papier_Magazine', 240, -1, 336, -1, 0.01901329681277275), (987515173, 1982, 'Papier_Magazine', 272, -1, 336, -1, 0.003658717731013894), (987515173, 1982, 'Papier_Magazine', 304, -1, 336, -1, 0.006297337356954813), (987515173, 1982, 'Papier_Magazine', 336, -1, 336, -1, 0.004447813145816326), (987515173, 1982, 'Plastique', 112, -1, 112, -1, 5.617829046400402e-08), (987515173, 1982, 'Plastique', 144, -1, 112, -1, 8.208448889490683e-07), (987515173, 1982, 'Plastique', 176, -1, 112, -1, 6.256026244955137e-05), (987515173, 1982, 'Plastique', 208, -1, 112, -1, 0.0035499518271535635), (987515173, 1982, 'Plastique', 240, -1, 112, -1, 0.003166872775182128), (987515173, 1982, 'Plastique', 272, -1, 112, -1, 0.007486181333661079), (987515173, 1982, 'Plastique', 304, -1, 112, -1, 0.054866764694452286), (987515173, 1982, 'Plastique', 336, -1, 112, -1, 0.05967129021883011), (987515173, 1982, 'Plastique', 112, -1, 144, -1, 2.791625547615695e-06), (987515173, 1982, 'Plastique', 144, -1, 144, -1, 1.903504744404927e-05), (987515173, 1982, 'Plastique', 176, -1, 144, -1, 0.00017590851348359138), (987515173, 1982, 'Plastique', 208, -1, 144, -1, 0.0015014332020655274), (987515173, 1982, 'Plastique', 240, -1, 144, -1, 0.005060270428657532), (987515173, 1982, 'Plastique', 272, -1, 144, -1, 0.08462032675743103), (987515173, 1982, 'Plastique', 304, -1, 144, -1, 0.4091968834400177), (987515173, 1982, 'Plastique', 336, -1, 144, -1, 0.0710282176733017), (987515173, 1982, 'Plastique', 112, -1, 176, -1, 3.9275050767173525e-06), (987515173, 1982, 'Plastique', 144, -1, 176, -1, 9.517036960460246e-05), (987515173, 1982, 'Plastique', 176, -1, 176, -1, 0.0003402219736017287), (987515173, 1982, 'Plastique', 208, -1, 176, -1, 0.0003888011851813644), (987515173, 1982, 'Plastique', 240, -1, 176, -1, 0.0006235309410840273), (987515173, 1982, 'Plastique', 272, -1, 176, -1, 0.00892091915011406), (987515173, 1982, 'Plastique', 304, -1, 176, -1, 0.037280965596437454), (987515173, 1982, 'Plastique', 336, -1, 176, -1, 0.0020717554725706577), (987515173, 1982, 'Plastique', 112, -1, 208, -1, 7.61425617383793e-05), (987515173, 1982, 'Plastique', 144, -1, 208, -1, 4.0941191400634125e-05), (987515173, 1982, 'Plastique', 176, -1, 208, -1, 8.232118125306442e-05), (987515173, 1982, 'Plastique', 208, -1, 208, -1, 3.834024028037675e-05), (987515173, 1982, 'Plastique', 240, -1, 208, -1, 8.352932127309032e-06), (987515173, 1982, 'Plastique', 272, -1, 208, -1, 1.1998071386187803e-05), (987515173, 1982, 'Plastique', 304, -1, 208, -1, 1.2090074051229749e-05), (987515173, 1982, 'Plastique', 336, -1, 208, -1, 4.101475042261882e-06), (987515173, 1982, 'Plastique', 112, -1, 240, -1, 0.00014464902051258832), (987515173, 1982, 'Plastique', 144, -1, 240, -1, 3.536247095325962e-05), (987515173, 1982, 'Plastique', 176, -1, 240, -1, 2.1159092284506187e-05), (987515173, 1982, 'Plastique', 208, -1, 240, -1, 7.3538108154025394e-06), (987515173, 1982, 'Plastique', 240, -1, 240, -1, 4.867225925409002e-06), (987515173, 1982, 'Plastique', 272, -1, 240, -1, 2.76307764579542e-06), (987515173, 1982, 'Plastique', 304, -1, 240, -1, 7.9772405570111e-07), (987515173, 1982, 'Plastique', 336, -1, 240, -1, 3.538978887718258e-07), (987515173, 1982, 'Plastique', 112, -1, 272, -1, 3.368564648553729e-05), (987515173, 1982, 'Plastique', 144, -1, 272, -1, 1.1664958947221749e-05), (987515173, 1982, 'Plastique', 176, -1, 272, -1, 1.341758525086334e-05), (987515173, 1982, 'Plastique', 208, -1, 272, -1, 5.420888555818237e-06), (987515173, 1982, 'Plastique', 240, -1, 272, -1, 9.547452464175876e-07), (987515173, 1982, 'Plastique', 272, -1, 272, -1, 6.88738509779796e-07), (987515173, 1982, 'Plastique', 304, -1, 272, -1, 5.724524498873507e-07), (987515173, 1982, 'Plastique', 336, -1, 272, -1, 7.429482025145262e-07), (987515173, 1982, 'Plastique', 112, -1, 304, -1, 3.3706512567732716e-06), (987515173, 1982, 'Plastique', 144, -1, 304, -1, 3.355880380695453e-06), (987515173, 1982, 'Plastique', 176, -1, 304, -1, 6.5082558649010025e-06), (987515173, 1982, 'Plastique', 208, -1, 304, -1, 8.736983545531984e-06), (987515173, 1982, 'Plastique', 240, -1, 304, -1, 3.3825183436420048e-06), (987515173, 1982, 'Plastique', 272, -1, 304, -1, 4.746831564261811e-06), (987515173, 1982, 'Plastique', 304, -1, 304, -1, 4.505414835875854e-06), (987515173, 1982, 'Plastique', 336, -1, 304, -1, 2.416980123598478e-06), (987515173, 1982, 'Plastique', 112, -1, 336, -1, 1.5473795428988524e-05), (987515173, 1982, 'Plastique', 144, -1, 336, -1, 5.35884391865693e-05), (987515173, 1982, 'Plastique', 176, -1, 336, -1, 3.427254705457017e-05), (987515173, 1982, 'Plastique', 208, -1, 336, -1, 2.372994458710309e-05), (987515173, 1982, 'Plastique', 240, -1, 336, -1, 3.8594636862399057e-05), (987515173, 1982, 'Plastique', 272, -1, 336, -1, 2.5748187908902764e-05), (987515173, 1982, 'Plastique', 304, -1, 336, -1, 3.3239921322092414e-05), (987515173, 1982, 'Plastique', 336, -1, 336, -1, 4.180555333732627e-05), (987515173, 1982, 'Sol_Environement', 112, -1, 112, -1, 2.930527999087107e-12), (987515173, 1982, 'Sol_Environement', 144, -1, 112, -1, 5.622760035350893e-10), (987515173, 1982, 'Sol_Environement', 176, -1, 112, -1, 4.922356993120047e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 112, -1, 1.0155593372473959e-05), (987515173, 1982, 'Sol_Environement', 240, -1, 112, -1, 9.177741958410479e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 112, -1, 5.625330231850967e-05), (987515173, 1982, 'Sol_Environement', 304, -1, 112, -1, 0.0003788383910432458), (987515173, 1982, 'Sol_Environement', 336, -1, 112, -1, 0.0001651871862122789), (987515173, 1982, 'Sol_Environement', 112, -1, 144, -1, 1.0347219614459391e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 144, -1, 5.797551807518175e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 144, -1, 1.428378482160042e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 144, -1, 5.890514785278356e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 144, -1, 1.824912214942742e-05), (987515173, 1982, 'Sol_Environement', 272, -1, 144, -1, 0.00012608319229912013), (987515173, 1982, 'Sol_Environement', 304, -1, 144, -1, 0.00032353654387407005), (987515173, 1982, 'Sol_Environement', 336, -1, 144, -1, 9.647644037613645e-05), (987515173, 1982, 'Sol_Environement', 112, -1, 176, -1, 4.182305133326736e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 176, -1, 1.0174305771215586e-06), (987515173, 1982, 'Sol_Environement', 176, -1, 176, -1, 6.5862955125339795e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 176, -1, 1.7199390640598722e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 176, -1, 3.7283095934981247e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 176, -1, 2.8019858291372657e-05), (987515173, 1982, 'Sol_Environement', 304, -1, 176, -1, 0.00015382831043098122), (987515173, 1982, 'Sol_Environement', 336, -1, 176, -1, 0.0002425080310786143), (987515173, 1982, 'Sol_Environement', 112, -1, 208, -1, 4.007958523288835e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 208, -1, 7.480698513973039e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 208, -1, 1.3315323030838044e-06), (987515173, 1982, 'Sol_Environement', 208, -1, 208, -1, 9.053310918716306e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 208, -1, 6.944654842300224e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 208, -1, 1.081097934729769e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 208, -1, 1.442125835637853e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 208, -1, 2.321799456694862e-06), (987515173, 1982, 'Sol_Environement', 112, -1, 240, -1, 4.6070831558608916e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 240, -1, 3.7703117072851455e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 240, -1, 2.486881101049221e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 240, -1, 2.1575910125193332e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 240, -1, 5.582672315540549e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 240, -1, 8.621477718406823e-07), (987515173, 1982, 'Sol_Environement', 304, -1, 240, -1, 6.640090646214958e-07), (987515173, 1982, 'Sol_Environement', 336, -1, 240, -1, 7.835926112420566e-07), (987515173, 1982, 'Sol_Environement', 112, -1, 272, -1, 2.9966649890411645e-06), (987515173, 1982, 'Sol_Environement', 144, -1, 272, -1, 6.880858904878551e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 272, -1, 5.021562969886872e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 272, -1, 2.0082485718830867e-07), (987515173, 1982, 'Sol_Environement', 240, -1, 272, -1, 1.3152288147466606e-07), (987515173, 1982, 'Sol_Environement', 272, -1, 272, -1, 2.6994132440449903e-07), (987515173, 1982, 'Sol_Environement', 304, -1, 272, -1, 1.635313680026229e-07), (987515173, 1982, 'Sol_Environement', 336, -1, 272, -1, 4.836798552787513e-07), (987515173, 1982, 'Sol_Environement', 112, -1, 304, -1, 7.164017574723403e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 304, -1, 4.1286557461717166e-07), (987515173, 1982, 'Sol_Environement', 176, -1, 304, -1, 8.938050086726435e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 304, -1, 2.3328091174334986e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 304, -1, 3.0542189506377326e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 304, -1, 2.666334466994158e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 304, -1, 1.4987853091952275e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 304, -1, 1.6114091749841464e-06), (987515173, 1982, 'Sol_Environement', 112, -1, 336, -1, 4.634302399608714e-07), (987515173, 1982, 'Sol_Environement', 144, -1, 336, -1, 1.127728523897531e-06), (987515173, 1982, 'Sol_Environement', 176, -1, 336, -1, 6.802097800573392e-07), (987515173, 1982, 'Sol_Environement', 208, -1, 336, -1, 1.4341401310957735e-06), (987515173, 1982, 'Sol_Environement', 240, -1, 336, -1, 3.3432154395995894e-06), (987515173, 1982, 'Sol_Environement', 272, -1, 336, -1, 2.1192113308643457e-06), (987515173, 1982, 'Sol_Environement', 304, -1, 336, -1, 2.2980780158832204e-06), (987515173, 1982, 'Sol_Environement', 336, -1, 336, -1, 3.3749192880350165e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 112, -1, 4.571711997414241e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 112, -1, 2.4769290121184895e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 112, -1, 1.7669475710135885e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 112, -1, 0.000475196196930483), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 112, -1, 0.00011565273598534986), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 112, -1, 0.00020995673548895866), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 112, -1, 0.0014421100495383143), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 112, -1, 0.0020227127242833376), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 144, -1, 3.082058537984267e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 144, -1, 1.7206231859745458e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 144, -1, 2.0752373529830948e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 144, -1, 0.00027424466679804027), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 144, -1, 0.00046559632755815983), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 144, -1, 0.0005290906992740929), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 144, -1, 0.00018949910008814186), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 144, -1, 0.00020546613086480647), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 176, -1, 1.5011509276519064e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 176, -1, 6.764404588466277e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 176, -1, 2.106750434904825e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 176, -1, 2.4099981601466425e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 176, -1, 3.5370754631003365e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 176, -1, 7.897792238509282e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 176, -1, 0.00010101871885126457), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 176, -1, 0.00037821478326804936), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 208, -1, 1.506751050328603e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 208, -1, 1.4607620641982066e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 208, -1, 4.731698481919011e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 208, -1, 3.825424755632412e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 208, -1, 3.646920276878518e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 208, -1, 1.1670057574519888e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 208, -1, 1.3331959053175524e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 208, -1, 5.995051105855964e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 240, -1, 4.2267558455932885e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 240, -1, 5.323747700458625e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 240, -1, 3.905935500370106e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 240, -1, 2.409000217085122e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 240, -1, 5.77363744014292e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 240, -1, 2.10927701118635e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 240, -1, 1.7813898011809215e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 240, -1, 2.2721291315974668e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 272, -1, 0.00020094112551305443), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 272, -1, 6.691193266306072e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 272, -1, 4.36419177276548e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 272, -1, 1.4556288078892976e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 272, -1, 7.054786692606285e-06), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 272, -1, 1.5496978448936716e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 272, -1, 1.5584806533297524e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 272, -1, 0.00010369653318775818), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 304, -1, 9.524247434455901e-05), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 304, -1, 0.00010966919944621623), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 304, -1, 0.00015802381676621735), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 304, -1, 0.0005423129769042134), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 304, -1, 0.0023449906148016453), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 304, -1, 0.007443359587341547), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 304, -1, 0.005889120511710644), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 304, -1, 0.0044502816163003445), (987515173, 1982, 'Teint_Dans_La_Masse', 112, -1, 336, -1, 0.00024103432951960713), (987515173, 1982, 'Teint_Dans_La_Masse', 144, -1, 336, -1, 0.0020014706533402205), (987515173, 1982, 'Teint_Dans_La_Masse', 176, -1, 336, -1, 0.000750622944906354), (987515173, 1982, 'Teint_Dans_La_Masse', 208, -1, 336, -1, 0.0033519910648465157), (987515173, 1982, 'Teint_Dans_La_Masse', 240, -1, 336, -1, 0.0641678050160408), (987515173, 1982, 'Teint_Dans_La_Masse', 272, -1, 336, -1, 0.03821588680148125), (987515173, 1982, 'Teint_Dans_La_Masse', 304, -1, 336, -1, 0.03885292261838913), (987515173, 1982, 'Teint_Dans_La_Masse', 336, -1, 336, -1, 0.005882319528609514), (987515173, 1982, 'autre_refus', 112, -1, 112, -1, 5.819347781432782e-10), (987515173, 1982, 'autre_refus', 144, -1, 112, -1, 2.0962945157521062e-08), (987515173, 1982, 'autre_refus', 176, -1, 112, -1, 1.4469102325165295e-06), (987515173, 1982, 'autre_refus', 208, -1, 112, -1, 9.239284554496408e-05), (987515173, 1982, 'autre_refus', 240, -1, 112, -1, 0.00021754215413238853), (987515173, 1982, 'autre_refus', 272, -1, 112, -1, 0.0017496768850833178), (987515173, 1982, 'autre_refus', 304, -1, 112, -1, 0.017770571634173393), (987515173, 1982, 'autre_refus', 336, -1, 112, -1, 0.009003167040646076), (987515173, 1982, 'autre_refus', 112, -1, 144, -1, 9.2359130121622e-07), (987515173, 1982, 'autre_refus', 144, -1, 144, -1, 2.9226066544651985e-06), (987515173, 1982, 'autre_refus', 176, -1, 144, -1, 1.0294600542692933e-05), (987515173, 1982, 'autre_refus', 208, -1, 144, -1, 0.0001347130601061508), (987515173, 1982, 'autre_refus', 240, -1, 144, -1, 0.0003480661252979189), (987515173, 1982, 'autre_refus', 272, -1, 144, -1, 0.004933568183332682), (987515173, 1982, 'autre_refus', 304, -1, 144, -1, 0.043510738760232925), (987515173, 1982, 'autre_refus', 336, -1, 144, -1, 0.09554944932460785), (987515173, 1982, 'autre_refus', 112, -1, 176, -1, 1.5122335753403604e-05), (987515173, 1982, 'autre_refus', 144, -1, 176, -1, 0.0001271346991416067), (987515173, 1982, 'autre_refus', 176, -1, 176, -1, 0.00023703543411102146), (987515173, 1982, 'autre_refus', 208, -1, 176, -1, 0.0008105260203592479), (987515173, 1982, 'autre_refus', 240, -1, 176, -1, 0.0006539791356772184), (987515173, 1982, 'autre_refus', 272, -1, 176, -1, 0.0043151527643203735), (987515173, 1982, 'autre_refus', 304, -1, 176, -1, 0.023329641669988632), (987515173, 1982, 'autre_refus', 336, -1, 176, -1, 0.018692055717110634), (987515173, 1982, 'autre_refus', 112, -1, 208, -1, 8.862425602274016e-05), (987515173, 1982, 'autre_refus', 144, -1, 208, -1, 0.00018532731337472796), (987515173, 1982, 'autre_refus', 176, -1, 208, -1, 0.00031896625296212733), (987515173, 1982, 'autre_refus', 208, -1, 208, -1, 0.00035758939338847995), (987515173, 1982, 'autre_refus', 240, -1, 208, -1, 0.00019877831800840795), (987515173, 1982, 'autre_refus', 272, -1, 208, -1, 0.00028634705813601613), (987515173, 1982, 'autre_refus', 304, -1, 208, -1, 0.00020239698642399162), (987515173, 1982, 'autre_refus', 336, -1, 208, -1, 0.0002434736379655078), (987515173, 1982, 'autre_refus', 112, -1, 240, -1, 0.00023386807879433036), (987515173, 1982, 'autre_refus', 144, -1, 240, -1, 0.00010855076106963679), (987515173, 1982, 'autre_refus', 176, -1, 240, -1, 6.48176865070127e-05), (987515173, 1982, 'autre_refus', 208, -1, 240, -1, 2.5300205379608087e-05), (987515173, 1982, 'autre_refus', 240, -1, 240, -1, 7.287785410881042e-05), (987515173, 1982, 'autre_refus', 272, -1, 240, -1, 0.00013982862583361566), (987515173, 1982, 'autre_refus', 304, -1, 240, -1, 8.934963989304379e-05), (987515173, 1982, 'autre_refus', 336, -1, 240, -1, 8.184897160390392e-05), (987515173, 1982, 'autre_refus', 112, -1, 272, -1, 0.0002684956125449389), (987515173, 1982, 'autre_refus', 144, -1, 272, -1, 0.0001116798011935316), (987515173, 1982, 'autre_refus', 176, -1, 272, -1, 0.0001248639600817114), (987515173, 1982, 'autre_refus', 208, -1, 272, -1, 5.112284634378739e-05), (987515173, 1982, 'autre_refus', 240, -1, 272, -1, 2.9130251277820207e-05), (987515173, 1982, 'autre_refus', 272, -1, 272, -1, 4.2736210161820054e-05), (987515173, 1982, 'autre_refus', 304, -1, 272, -1, 6.920337909832597e-05), (987515173, 1982, 'autre_refus', 336, -1, 272, -1, 0.00014184493920765817), (987515173, 1982, 'autre_refus', 112, -1, 304, -1, 0.00011816414189524949), (987515173, 1982, 'autre_refus', 144, -1, 304, -1, 0.00022210566385183483), (987515173, 1982, 'autre_refus', 176, -1, 304, -1, 0.00042709894478321075), (987515173, 1982, 'autre_refus', 208, -1, 304, -1, 0.0004272170481272042), (987515173, 1982, 'autre_refus', 240, -1, 304, -1, 6.55979456496425e-05), (987515173, 1982, 'autre_refus', 272, -1, 304, -1, 3.1694882636656985e-05), (987515173, 1982, 'autre_refus', 304, -1, 304, -1, 1.1630776498350315e-05), (987515173, 1982, 'autre_refus', 336, -1, 304, -1, 1.8845976228476502e-05), (987515173, 1982, 'autre_refus', 112, -1, 336, -1, 0.00024746061535552144), (987515173, 1982, 'autre_refus', 144, -1, 336, -1, 0.00047122384421527386), (987515173, 1982, 'autre_refus', 176, -1, 336, -1, 0.0003342038835398853), (987515173, 1982, 'autre_refus', 208, -1, 336, -1, 0.0002366719563724473), (987515173, 1982, 'autre_refus', 240, -1, 336, -1, 0.00010693204967537895), (987515173, 1982, 'autre_refus', 272, -1, 336, -1, 9.549219976179302e-05), (987515173, 1982, 'autre_refus', 304, -1, 336, -1, 0.000131260123453103), (987515173, 1982, 'autre_refus', 336, -1, 336, -1, 0.000730274710804224)]} ############################### TEST certificat_qualite_papier ################################ TEST certificat qualite papier Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=1848 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=1848 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 1848 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=1848 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) All sons are already in current list ! All sons are already in current list ! All sons are already in current list ! All sons are already in current list ! All sons are already in current list ! All sons are already in current list ! DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! Step 4442 tile have less inputs used (1) than in the step definition (3) : maybe we manage optionnal inputs ! WARNING : number of outputs for step 4441 detect_points is not consistent : 2 used against 1 in the step definition ! WARNING : number of inputs for step 4443 count_percent_refus is not consistent : 4 used against 3 in the step definition ! Step 4444 send_mail_dechet have less inputs used (3) than in the step definition (5) : maybe we manage optionnal inputs ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! WARNING : output 1 of step 4440 have datatype=1 whereas input 0 of step 4443 have datatype=2 WARNING : type of output 1 of step 4441 doesn't seem to be define in the database( WARNING : type of input 4 of step 4443 doesn't seem to be define in the database( DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : init_dechet, tile, detect_points, count_percent_refus, brightness, blur_detection, send_mail_dechet list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT ph.photo_id, ph.url FROM MTRBack.photos ph WHERE ph.photo_id IN (SELECT mtr_photo_id from MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (1902940) and hide_status = 0 ) ORDER BY ph.photo_id DESC LIMIT 0, 10000 Catched exception ! Connect or reconnect ! We have 1 , {} SELECT mtr_photo_id, mtr_portfolio_id FROM MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (1902940) AND hide_status = 0 ORDER by mtr_photo_id desc LIMIT 0, 10000 list_result: [{'photo_id': 987321136, 'portfolio_id': 1902940}] map_portfolio_id_photo_id: {1902940: [987321136]} ##### Call download_photos : nb_thread : 5 begin to download photo : 987321136 download finish for photo 987321136 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.22869372367858887 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 7 step1:init_dechet Fri May 30 03:40:12 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} debut step init detect dechets input : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg ON MODIFIE NB AVEC LE INPUT map photo id path extension : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg scale : 0.9481481481481482 FIN step init dechet After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True saveOutput not yet implemented for datou_step.type : init_dechet we use saveGeneral [987321136] map_info['map_portfolio_photo'] : {1902940: [987321136]} final : False mtd_id 1848 list_pids : [987321136] Looping around the photos to save general results len do output : 1 /987321136Didn't retrieve data .Didn't retrieve data .Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('1848', None, None, None, None, None, None, None, None) ('1848', '1902940', '987321136', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 4 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('1848', '1902940', '987321136', 'None', None, None, None, None, None)] time used for this insertion : 0.015553712844848633 save_final save missing photos in datou_result : time spend for datou_step_exec : 0.00016689300537109375 time spend to save output : 0.01582789421081543 total time spend for step 1 : 0.015994787216186523 step2:tile Fri May 30 03:40:13 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'987321136': ()} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 0 We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg',) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {0: 987321136} verbose : True param_json : {'token': '78d09a0790ec6ecbf119343125a81fdc', 'portfolio_name': 'tile_correct_upm', 'ETA': 86400, 'new_width': 1500, 'new_height': 20000, 'host': 'www.fotonower.com', 'protocol': 'https', 'photo_tile_type': 1522, 'option_bande': 'True'} type(crop_hashtag_type) : type(crop_hashtag_type_tiled) : We consider crop_hashtag_type is an integer ! map_chi_type_to_chi_type_cropped : {406: 410} map_filenames : {987321136: 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg'} list_pids : 1 list_pids : 2 list_subpids to replace list_pids : 1 batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 987321136,987321136,987321136) and `type` in (406) Loaded 0 chid ids of type : 0 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in () https://marlene.fotonower.com/api/v1/secured/portfolio/new?name=tile_correct_upm&access_token=78d09a0790ec6ecbf119343125a81fdc created feed_id_new_photos : 23444770 with name tile_correct_upm feed_id_new_photos : 23444770 filename : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg photo_id : 987321136 height_image_input : 439 width_image_input : 562 new_width : 1500 new_height : 20000 stride : 0 stride_relative : 0.1 chi to copy from the main photo to the tiled photo input_chi_for_this_image_as_chi : 0 list_bib_to_crops : 1 [(0, 562, 0, 439, 0)] calcul des nouveaux crops pour le tile x0:0,x1:562,y0:0,y1:439 chi selectionnes : [] new_crops_tiles : 1 crop_transformed : 0 insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) [(987321136, 2090988864, 1522, 0, 562, 0, 439, 1.0)] list_photo_ids_cropped : [987321136] batch 1 select photo_id, hashtag_id, `type`, x0, x1, y0, y1, score, id from MTRPhoto.crop_hashtag_ids where photo_id in ( 987321136) and `type` in (1522) Loaded 1 chid ids of type : 1522 SELECT crop_hashtag_id, points FROM MTRPhoto.crop_polygon_points WHERE crop_hashtag_id in (1608847328) SELECT * FROM MTRPhoto.crop_segments WHERE crop_hashtag_id in (1608847328) SELECT * FROM MTRPhoto.crop_sum_segments WHERE crop_hashtag_id in (1608847328) treat the image : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg , 0 before upload mediasElapsed time : 0.010347127914428711 About to upload 1 photos upload in portfolio : 23444770 Result OK ! uploaded one batch 0 Elapsed time : 4.944097518920898 upload mediasElapsed time : 4.954502582550049 , 0insert ignore into MTRPhoto.crop_sub_photo_ids (crop_hashtag_id, sub_photo_id, mtr_user_id) VALUES (%s,%s,%s) : [(1608847328, 1361688639, 0)] Saving 0 CHIs. list_chi_tile : [] end of tileElapsed time : 4.967602014541626 map_pid_results : {'1361688639': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg']} After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True saveOutput not yet implemented for datou_step.type : tile we use saveGeneral [987321136, 987321136, '1361688639'] map_info['map_portfolio_photo'] : {1902940: [987321136]} final : False mtd_id 1848 list_pids : [987321136, 987321136, '1361688639'] Looping around the photos to save general results len do output : 1 /1361688639Didn't retrieve data . before output type Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('1848', None, None, None, None, None, None, None, None) ('1848', '1902940', '987321136', None, None, None, None, None, None) ('1848', None, None, None, None, None, None, None, None) ('1848', '1902940', '987321136', None, None, None, None, None, None) ('1848', None, None, None, None, None, None, None, None) ('1848', None, '1361688639', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 4 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('1848', None, '1361688639', 'None', None, None, None, None, None), ('1848', '1902940', '987321136', None, None, None, None, None, None)] time used for this insertion : 0.013514041900634766 save_final save missing photos in datou_result : time spend for datou_step_exec : 11.71670413017273 time spend to save output : 0.013674020767211914 total time spend for step 2 : 11.730378150939941 step3:detect_points Fri May 30 03:40:24 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'1361688639': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg']} input_args_next_step : {'1361688639': ()} output_args : {'1361688639': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg']} args : 1361688639 depend.output_id : 0 complete output_args for input 1 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'1361688639': ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg',), '987321136': ()} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 2 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :2, first value : ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg',) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136, 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg': 1361688639} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}, 1361688639: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg'}} map_subphoto_mainphoto : {0: 987321136, 1361688639: 987321136} Beginning of datou step predict points ! Inside try reload ! classes : ['Autre_Environement', 'Carton', 'Kraft', 'Lointain_Papier_Magazine', 'Metal', 'Papier_Magazine', 'Plastique', 'Sol_Environement', 'Teint_Dans_La_Masse', 'autre_refus'] pht : 1927 model_name : learn_refus_upm_blanches_1924 {'id': 1528, 'mtr_user_id': 31, 'name': 'learn_refus_upm_blanches_1924', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'Autre_Environement,Carton,Kraft,Lointain_Papier_Magazine,Metal,Papier_Magazine,Plastique,Sol_Environement,Teint_Dans_La_Masse,autre_refus', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 1927, 'photo_desc_type': 4421, 'type_classification': 'caffe', 'hashtag_id_list': '2107752388,492774966,493202403,2107752389,492628673,2107752386,492725882,2107752387,2107752385,2107752406'} gpu_mode in detect_points : False To load net FromThcl() model_param file didn't exist model_name : learn_refus_upm_blanches_1924 model_type : caffe list file need : ['caffemodel', 'deploy_conv_normal.prototxt', 'deploy_fc.prototxt', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file exist in s3 : ['caffemodel', 'deploy.prototxt', 'mean.npy', 'synset_words.txt'] file manque in s3 : ['deploy_conv_normal.prototxt', 'deploy_fc.prototxt'] local folder : /data/models_weight/learn_refus_upm_blanches_1924 /data/models_weight/learn_refus_upm_blanches_1924/caffemodel size_local : 45774543 size in s3 : 45774543 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 caffemodel already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/deploy.prototxt size_local : 17312 size in s3 : 17312 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:03 deploy.prototxt already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/mean.npy size_local : 1572992 size in s3 : 1572992 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:05 mean.npy already exist and didn't need to update /data/models_weight/learn_refus_upm_blanches_1924/synset_words.txt size_local : 218 size in s3 : 218 create time local : 2021-08-09 05:29:53 create time in s3 : 2021-08-06 19:36:04 synset_words.txt already exist and didn't need to update reshape net's input to : (224, 224) origin shape : (10, 3, 224, 224) after reshape : (1, 3, 224, 224) [('data', (1, 3, 224, 224)), ('conv1', (1, 64, 112, 112)), ('pool1', (1, 64, 56, 56)), ('pool1_pool1_0_split_0', (1, 64, 56, 56)), ('pool1_pool1_0_split_1', (1, 64, 56, 56)), ('res2a_branch1', (1, 64, 56, 56)), ('res2a_branch2a', (1, 64, 56, 56)), ('res2a_branch2b', (1, 64, 56, 56)), ('res2a', (1, 64, 56, 56)), ('res2a_res2a_relu_0_split_0', (1, 64, 56, 56)), ('res2a_res2a_relu_0_split_1', (1, 64, 56, 56)), ('res2b_branch2a', (1, 64, 56, 56)), ('res2b_branch2b', (1, 64, 56, 56)), ('res2b', (1, 64, 56, 56)), ('res2b_res2b_relu_0_split_0', (1, 64, 56, 56)), ('res2b_res2b_relu_0_split_1', (1, 64, 56, 56)), ('res3a_branch1', (1, 128, 28, 28)), ('res3a_branch2a', (1, 128, 28, 28)), ('res3a_branch2b', (1, 128, 28, 28)), ('res3a', (1, 128, 28, 28)), ('res3a_res3a_relu_0_split_0', (1, 128, 28, 28)), ('res3a_res3a_relu_0_split_1', (1, 128, 28, 28)), ('res3b_branch2a', (1, 128, 28, 28)), ('res3b_branch2b', (1, 128, 28, 28)), ('res3b', (1, 128, 28, 28)), ('res3b_res3b_relu_0_split_0', (1, 128, 28, 28)), ('res3b_res3b_relu_0_split_1', (1, 128, 28, 28)), ('res4a_branch1', (1, 256, 14, 14)), ('res4a_branch2a', (1, 256, 14, 14)), ('res4a_branch2b', (1, 256, 14, 14)), ('res4a', (1, 256, 14, 14)), ('res4a_res4a_relu_0_split_0', (1, 256, 14, 14)), ('res4a_res4a_relu_0_split_1', (1, 256, 14, 14)), ('res4b_branch2a', (1, 256, 14, 14)), ('res4b_branch2b', (1, 256, 14, 14)), ('res4b', (1, 256, 14, 14)), ('res4b_res4b_relu_0_split_0', (1, 256, 14, 14)), ('res4b_res4b_relu_0_split_1', (1, 256, 14, 14)), ('res5a_branch1', (1, 512, 7, 7)), ('res5a_branch2a', (1, 512, 7, 7)), ('res5a_branch2b', (1, 512, 7, 7)), ('res5a', (1, 512, 7, 7)), ('res5a_res5a_relu_0_split_0', (1, 512, 7, 7)), ('res5a_res5a_relu_0_split_1', (1, 512, 7, 7)), ('res5b_branch2a', (1, 512, 7, 7)), ('res5b_branch2b', (1, 512, 7, 7)), ('res5b', (1, 512, 7, 7)), ('fc2019-10-22_15-02-46', (1, 10, 1, 1)), ('prob', (1, 10, 1, 1))] set image transformer : About to compute detect the points : len(args) : 2 Inside predict_points step exec : nb paths : 1 treate image : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg size of numpy array img : 2960752 scale method : caffe/skimage size of numpy array img_scale : 2655880 (416, 532, 3) nb_h 7 nb_w 11 size of sub images : (224, 224, 3) size of caffe_input : 46362776 (77, 3, 224, 224) time to do the preprocess : 0.11371445655822754 time to do a prediction : 16.21740484237671 dict_keys(['prob']) shape of output (77, 10, 1, 1) shape of the out_put heatmap (10, 7, 11) number of sub_photos vertical and horizon 7 11 size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) size of heatmap : (7,11) After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True Inside savePoints : final : False verbose : True threshold to save the result : 0.05 maximun points to save in the table mtr_datou_result for each class : 100 output flattener 5 example : {1361688639: [(1361688639, 1945, 'Autre_Environement', 185, -1, 118, -1, 1.8337410438107327e-05), (1361688639, 1945, 'Autre_Environement', 320, -1, 118, -1, 0.0001291327498620376), (1361688639, 1945, 'Autre_Environement', 388, -1, 151, -1, 1.3445685908664018e-05), (1361688639, 1945, 'Autre_Environement', 286, -1, 185, -1, 0.00013466527161654085), (1361688639, 1945, 'Autre_Environement', 118, -1, 219, -1, 9.86390750767896e-06), (1361688639, 1945, 'Autre_Environement', 219, -1, 219, -1, 0.00038494516047649086), (1361688639, 1945, 'Autre_Environement', 354, -1, 219, -1, 0.0003431888180784881), (1361688639, 1945, 'Autre_Environement', 421, -1, 253, -1, 1.494663450785083e-07), (1361688639, 1945, 'Autre_Environement', 185, -1, 286, -1, 1.814520658172114e-07), (1361688639, 1945, 'Autre_Environement', 320, -1, 286, -1, 4.115241154067917e-06), (1361688639, 1945, 'Autre_Environement', 118, -1, 320, -1, 2.5274063397695556e-10), (1361688639, 1945, 'Autre_Environement', 253, -1, 320, -1, 1.0447603199237321e-10), (1361688639, 1945, 'Autre_Environement', 388, -1, 320, -1, 4.5192118136583304e-07), (1361688639, 1945, 'Carton', 151, -1, 118, -1, 0.9897055625915527), (1361688639, 1945, 'Carton', 286, -1, 118, -1, 0.8210961222648621), (1361688639, 1945, 'Carton', 421, -1, 118, -1, 0.4851985573768616), (1361688639, 1945, 'Carton', 219, -1, 151, -1, 0.9841372966766357), (1361688639, 1945, 'Carton', 118, -1, 185, -1, 0.9978604912757874), (1361688639, 1945, 'Carton', 185, -1, 219, -1, 0.7996194958686829), (1361688639, 1945, 'Carton', 354, -1, 219, -1, 0.13047637045383453), (1361688639, 1945, 'Carton', 286, -1, 253, -1, 0.05070793256163597), (1361688639, 1945, 'Carton', 118, -1, 286, -1, 0.31989040970802307), (1361688639, 1945, 'Carton', 219, -1, 286, -1, 0.0034983144141733646), (1361688639, 1945, 'Carton', 421, -1, 286, -1, 0.01660206913948059), (1361688639, 1945, 'Carton', 354, -1, 320, -1, 0.0016258988762274384), (1361688639, 1945, 'Kraft', 185, -1, 118, -1, 0.004793279338628054), (1361688639, 1945, 'Kraft', 286, -1, 118, -1, 0.0023909551091492176), (1361688639, 1945, 'Kraft', 421, -1, 118, -1, 0.002102428814396262), (1361688639, 1945, 'Kraft', 354, -1, 151, -1, 0.05212335288524628), (1361688639, 1945, 'Kraft', 118, -1, 185, -1, 0.0017166697653010488), (1361688639, 1945, 'Kraft', 253, -1, 185, -1, 0.04599893465638161), (1361688639, 1945, 'Kraft', 185, -1, 219, -1, 0.021099306643009186), (1361688639, 1945, 'Kraft', 388, -1, 219, -1, 6.061792737455107e-05), (1361688639, 1945, 'Kraft', 286, -1, 253, -1, 0.0037522290367633104), (1361688639, 1945, 'Kraft', 118, -1, 286, -1, 0.010319208726286888), (1361688639, 1945, 'Kraft', 219, -1, 286, -1, 5.351284926291555e-05), (1361688639, 1945, 'Kraft', 421, -1, 286, -1, 9.467339623370208e-06), (1361688639, 1945, 'Kraft', 320, -1, 320, -1, 0.00017124204896390438), (1361688639, 1945, 'Lointain_Papier_Magazine', 185, -1, 118, -1, 2.352082447032444e-06), (1361688639, 1945, 'Lointain_Papier_Magazine', 320, -1, 118, -1, 1.991412864299491e-05), (1361688639, 1945, 'Lointain_Papier_Magazine', 253, -1, 151, -1, 1.1851832823595032e-05), (1361688639, 1945, 'Lointain_Papier_Magazine', 354, -1, 185, -1, 8.877575601218268e-05), (1361688639, 1945, 'Lointain_Papier_Magazine', 118, -1, 219, -1, 2.1015976017224602e-06), (1361688639, 1945, 'Lointain_Papier_Magazine', 219, -1, 219, -1, 7.86672972026281e-05), (1361688639, 1945, 'Lointain_Papier_Magazine', 421, -1, 219, -1, 5.019426794206083e-07), (1361688639, 1945, 'Lointain_Papier_Magazine', 320, -1, 253, -1, 8.220934978453442e-05), (1361688639, 1945, 'Lointain_Papier_Magazine', 185, -1, 286, -1, 3.6812457437918056e-07), (1361688639, 1945, 'Lointain_Papier_Magazine', 388, -1, 286, -1, 3.3782250739022857e-06), (1361688639, 1945, 'Lointain_Papier_Magazine', 118, -1, 320, -1, 3.5824958555252806e-09), (1361688639, 1945, 'Lointain_Papier_Magazine', 286, -1, 320, -1, 1.2866975396264024e-07), (1361688639, 1945, 'Metal', 185, -1, 118, -1, 7.473808364011347e-05), (1361688639, 1945, 'Metal', 286, -1, 118, -1, 3.432366065680981e-05), (1361688639, 1945, 'Metal', 118, -1, 151, -1, 1.1519473446242046e-06), (1361688639, 1945, 'Metal', 354, -1, 151, -1, 0.001217490527778864), (1361688639, 1945, 'Metal', 253, -1, 185, -1, 0.001836584648117423), (1361688639, 1945, 'Metal', 421, -1, 185, -1, 3.084278432652354e-05), (1361688639, 1945, 'Metal', 185, -1, 219, -1, 0.0011127882171422243), (1361688639, 1945, 'Metal', 118, -1, 253, -1, 2.7215228328714147e-05), (1361688639, 1945, 'Metal', 354, -1, 253, -1, 0.0005037166411057115), (1361688639, 1945, 'Metal', 219, -1, 286, -1, 1.657771281315945e-05), (1361688639, 1945, 'Metal', 421, -1, 286, -1, 2.2439740860136226e-05), (1361688639, 1945, 'Metal', 151, -1, 320, -1, 1.393576087860282e-10), (1361688639, 1945, 'Metal', 320, -1, 320, -1, 3.1560623028781265e-05), (1361688639, 1945, 'Papier_Magazine', 118, -1, 118, -1, 0.001665871823206544), (1361688639, 1945, 'Papier_Magazine', 253, -1, 118, -1, 0.28050497174263), (1361688639, 1945, 'Papier_Magazine', 185, -1, 151, -1, 0.003942570183426142), (1361688639, 1945, 'Papier_Magazine', 421, -1, 151, -1, 0.9828073978424072), (1361688639, 1945, 'Papier_Magazine', 354, -1, 185, -1, 0.7690255045890808), (1361688639, 1945, 'Papier_Magazine', 286, -1, 219, -1, 0.9288092851638794), (1361688639, 1945, 'Papier_Magazine', 118, -1, 253, -1, 0.04313919320702553), (1361688639, 1945, 'Papier_Magazine', 219, -1, 253, -1, 0.8978631496429443), (1361688639, 1945, 'Papier_Magazine', 388, -1, 253, -1, 0.9886879324913025), (1361688639, 1945, 'Papier_Magazine', 151, -1, 320, -1, 0.9999996423721313), (1361688639, 1945, 'Papier_Magazine', 253, -1, 320, -1, 0.9999960660934448), (1361688639, 1945, 'Papier_Magazine', 354, -1, 320, -1, 0.9942159056663513), (1361688639, 1945, 'Plastique', 118, -1, 118, -1, 1.236031312146224e-05), (1361688639, 1945, 'Plastique', 219, -1, 118, -1, 0.00030184732167981565), (1361688639, 1945, 'Plastique', 320, -1, 118, -1, 0.0002493929350748658), (1361688639, 1945, 'Plastique', 253, -1, 185, -1, 0.007031592074781656), (1361688639, 1945, 'Plastique', 354, -1, 185, -1, 0.032503753900527954), (1361688639, 1945, 'Plastique', 185, -1, 219, -1, 0.050375860184431076), (1361688639, 1945, 'Plastique', 421, -1, 219, -1, 0.00012226666149217635), (1361688639, 1945, 'Plastique', 118, -1, 253, -1, 0.003930176142603159), (1361688639, 1945, 'Plastique', 286, -1, 253, -1, 0.0025490771513432264), (1361688639, 1945, 'Plastique', 219, -1, 286, -1, 6.120907346485183e-05), (1361688639, 1945, 'Plastique', 354, -1, 286, -1, 0.00538312504068017), (1361688639, 1945, 'Plastique', 151, -1, 320, -1, 1.8926894773674263e-10), (1361688639, 1945, 'Plastique', 421, -1, 320, -1, 0.00020204381144139916), (1361688639, 1945, 'Sol_Environement', 185, -1, 118, -1, 9.372543900099117e-06), (1361688639, 1945, 'Sol_Environement', 320, -1, 118, -1, 2.7338848667568527e-05), (1361688639, 1945, 'Sol_Environement', 118, -1, 151, -1, 2.5881729470711434e-07), (1361688639, 1945, 'Sol_Environement', 253, -1, 185, -1, 0.00011454988998593763), (1361688639, 1945, 'Sol_Environement', 354, -1, 185, -1, 0.00020838991622440517), (1361688639, 1945, 'Sol_Environement', 185, -1, 219, -1, 9.349620813736692e-05), (1361688639, 1945, 'Sol_Environement', 421, -1, 219, -1, 1.1348908657282664e-07), (1361688639, 1945, 'Sol_Environement', 118, -1, 253, -1, 7.004572921687213e-07), (1361688639, 1945, 'Sol_Environement', 320, -1, 253, -1, 5.724640504922718e-05), (1361688639, 1945, 'Sol_Environement', 219, -1, 286, -1, 3.869550369017816e-08), (1361688639, 1945, 'Sol_Environement', 388, -1, 286, -1, 6.792529802623903e-06), (1361688639, 1945, 'Sol_Environement', 151, -1, 320, -1, 2.6225014184994705e-14), (1361688639, 1945, 'Sol_Environement', 286, -1, 320, -1, 1.5886031690115487e-07), (1361688639, 1945, 'Teint_Dans_La_Masse', 185, -1, 118, -1, 0.002272234996780753), (1361688639, 1945, 'Teint_Dans_La_Masse', 286, -1, 118, -1, 0.001813237671740353), (1361688639, 1945, 'Teint_Dans_La_Masse', 388, -1, 118, -1, 0.04538803920149803), (1361688639, 1945, 'Teint_Dans_La_Masse', 118, -1, 151, -1, 0.00010940170614048839), (1361688639, 1945, 'Teint_Dans_La_Masse', 253, -1, 185, -1, 0.0056123328395187855), (1361688639, 1945, 'Teint_Dans_La_Masse', 354, -1, 185, -1, 0.15166763961315155), (1361688639, 1945, 'Teint_Dans_La_Masse', 185, -1, 219, -1, 0.0013750138459727168), (1361688639, 1945, 'Teint_Dans_La_Masse', 118, -1, 253, -1, 6.252125604078174e-05), (1361688639, 1945, 'Teint_Dans_La_Masse', 286, -1, 253, -1, 0.0018639108166098595), (1361688639, 1945, 'Teint_Dans_La_Masse', 388, -1, 253, -1, 0.001438076258637011), (1361688639, 1945, 'Teint_Dans_La_Masse', 219, -1, 286, -1, 1.016586884361459e-05), (1361688639, 1945, 'Teint_Dans_La_Masse', 151, -1, 320, -1, 3.425693932967988e-07), (1361688639, 1945, 'Teint_Dans_La_Masse', 320, -1, 320, -1, 0.0020001486409455538), (1361688639, 1945, 'Teint_Dans_La_Masse', 421, -1, 320, -1, 8.311669807881117e-05), (1361688639, 1945, 'autre_refus', 185, -1, 118, -1, 0.028516046702861786), (1361688639, 1945, 'autre_refus', 354, -1, 118, -1, 0.0014720888575538993), (1361688639, 1945, 'autre_refus', 118, -1, 151, -1, 3.824138184427284e-05), (1361688639, 1945, 'autre_refus', 253, -1, 185, -1, 0.07516990602016449), (1361688639, 1945, 'autre_refus', 185, -1, 219, -1, 0.010234796442091465), (1361688639, 1945, 'autre_refus', 354, -1, 219, -1, 0.04460597038269043), (1361688639, 1945, 'autre_refus', 118, -1, 253, -1, 0.00015751634782645851), (1361688639, 1945, 'autre_refus', 286, -1, 253, -1, 0.01393966656178236), (1361688639, 1945, 'autre_refus', 219, -1, 286, -1, 4.3672833271557465e-05), (1361688639, 1945, 'autre_refus', 388, -1, 286, -1, 0.0030952864326536655), (1361688639, 1945, 'autre_refus', 151, -1, 320, -1, 1.5080686699420198e-10), (1361688639, 1945, 'autre_refus', 320, -1, 320, -1, 0.0030849112663418055)]} hashtag or score ? = 0.9897055625915527 hashtag or score ? = 0.8210961222648621 hashtag or score ? = 0.4851985573768616 hashtag or score ? = 0.9841372966766357 hashtag or score ? = 0.9978604912757874 hashtag or score ? = 0.7996194958686829 hashtag or score ? = 0.13047637045383453 hashtag or score ? = 0.05070793256163597 hashtag or score ? = 0.31989040970802307 hashtag or score ? = 0.05212335288524628 hashtag or score ? = 0.28050497174263 hashtag or score ? = 0.9828073978424072 hashtag or score ? = 0.7690255045890808 hashtag or score ? = 0.9288092851638794 hashtag or score ? = 0.8978631496429443 hashtag or score ? = 0.9886879324913025 hashtag or score ? = 0.9999996423721313 hashtag or score ? = 0.9999960660934448 hashtag or score ? = 0.9942159056663513 hashtag or score ? = 0.050375860184431076 hashtag or score ? = 0.15166763961315155 hashtag or score ? = 0.07516990602016449 insert ignore into MTRPhoto.crop_hashtag_ids (photo_id, hashtag_id, `type`,x0,x1,y0,y1,score) VALUES (%s,%s,%s,%s,%s,%s,%s,%s) [('1361688639', '492774966', '1945', '151', '-1', '118', '-1', '0.9897055625915527'), ('1361688639', '492774966', '1945', '286', '-1', '118', '-1', '0.8210961222648621'), ('1361688639', '492774966', '1945', '421', '-1', '118', '-1', '0.4851985573768616'), ('1361688639', '492774966', '1945', '219', '-1', '151', '-1', '0.9841372966766357'), ('1361688639', '492774966', '1945', '118', '-1', '185', '-1', '0.9978604912757874'), ('1361688639', '492774966', '1945', '185', '-1', '219', '-1', '0.7996194958686829'), ('1361688639', '492774966', '1945', '354', '-1', '219', '-1', '0.13047637045383453'), ('1361688639', '492774966', '1945', '286', '-1', '253', '-1', '0.05070793256163597'), ('1361688639', '492774966', '1945', '118', '-1', '286', '-1', '0.31989040970802307'), ('1361688639', '493202403', '1945', '354', '-1', '151', '-1', '0.05212335288524628'), ('1361688639', '2107752386', '1945', '253', '-1', '118', '-1', '0.28050497174263'), ('1361688639', '2107752386', '1945', '421', '-1', '151', '-1', '0.9828073978424072'), ('1361688639', '2107752386', '1945', '354', '-1', '185', '-1', '0.7690255045890808'), ('1361688639', '2107752386', '1945', '286', '-1', '219', '-1', '0.9288092851638794'), ('1361688639', '2107752386', '1945', '219', '-1', '253', '-1', '0.8978631496429443'), ('1361688639', '2107752386', '1945', '388', '-1', '253', '-1', '0.9886879324913025'), ('1361688639', '2107752386', '1945', '151', '-1', '320', '-1', '0.9999996423721313'), ('1361688639', '2107752386', '1945', '253', '-1', '320', '-1', '0.9999960660934448'), ('1361688639', '2107752386', '1945', '354', '-1', '320', '-1', '0.9942159056663513'), ('1361688639', '492725882', '1945', '185', '-1', '219', '-1', '0.050375860184431076'), ('1361688639', '2107752385', '1945', '354', '-1', '185', '-1', '0.15166763961315155'), ('1361688639', '2107752406', '1945', '253', '-1', '185', '-1', '0.07516990602016449')] final : False save missing photos in datou_result : time spend for datou_step_exec : 17.744869232177734 time spend to save output : 0.05796003341674805 total time spend for step 3 : 17.802829265594482 step4:count_percent_refus Fri May 30 03:40:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'987321136': ()} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 1 complete output_args for input 1 : {'1361688639': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg']} input_args_next_step : {'987321136': (987321136,), '1361688639': ()} output_args : {'1361688639': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg']} args : 1361688639 depend.output_id : 0 complete output_args for input 2 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'987321136': (987321136,), '1361688639': ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg',)} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 2 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :2, first value : (987321136, 0.9481481481481482) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136, 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg': 1361688639} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}, 1361688639: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg'}} map_subphoto_mainphoto : {0: 987321136, 1361688639: 987321136} debut step count percent refus args : {'987321136': (987321136, 0.9481481481481482), '1361688639': ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg',)} (987321136, 0.9481481481481482) ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg',) on trouve le portfolio_id = 1902940 list_photo : [987321136] list_photo_correc : [1361688639] debut step count percent refus Treating photo_id : 987321136 Calcul du count_res count res : ((492774966, 3), (2107752386, 7)) Hashtag_id : 492774966 Hashtag_id : 2107752386 We have 2 classes in this image After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True map_info[mapportfolio_photo] : {1902940: [987321136]} dans le for photo id : 987321136 output[photo_id] : [({'carton': 3, 'Papier_Magazine': 7}, [1361688639], {'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}, {'refus_total': 61.64383561643836, 'carton': 61.64383561643836, 'Papier_Magazine': 38.35616438356164}, 1902940)] begin to insert list_values into mtr_datou_result : length of list_values in save_final : 6 insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_surface',1902940,30.0,'refus_total',1945) on duplicate key update value= 30.0 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945]] insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_gravi',1902940,61.64383561643836,'refus_total',1945) on duplicate key update value= 61.64383561643836 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'refus_total', 1945]] insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_surface',1902940,30.0,'carton',1945) on duplicate key update value= 30.0 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'refus_total', 1945], ['0', 'qualipapia_surface', 1902940, 30.0, 'carton', 1945]] insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_gravi',1902940,61.64383561643836,'carton',1945) on duplicate key update value= 61.64383561643836 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'refus_total', 1945], ['0', 'qualipapia_surface', 1902940, 30.0, 'carton', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'carton', 1945]] insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_surface',1902940,70.0,'Papier_Magazine',1945) on duplicate key update value= 70.0 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'refus_total', 1945], ['0', 'qualipapia_surface', 1902940, 30.0, 'carton', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'carton', 1945], ['0', 'qualipapia_surface', 1902940, 70.0, 'Papier_Magazine', 1945]] insert into MTRLabel.upm_carac (ol,type_carac,portfolio_id,value,material,hashtag_type) values (0,'qualipapia_gravi',1902940,38.35616438356164,'Papier_Magazine',1945) on duplicate key update value= 38.35616438356164 list_values : [['0', 'qualipapia_surface', 1902940, 30.0, 'refus_total', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'refus_total', 1945], ['0', 'qualipapia_surface', 1902940, 30.0, 'carton', 1945], ['0', 'qualipapia_gravi', 1902940, 61.64383561643836, 'carton', 1945], ['0', 'qualipapia_surface', 1902940, 70.0, 'Papier_Magazine', 1945], ['0', 'qualipapia_gravi', 1902940, 38.35616438356164, 'Papier_Magazine', 1945]] time used for this insertion : 0.05565595626831055 save missing photos in datou_result : time spend for datou_step_exec : 0.01675271987915039 time spend to save output : 0.05588340759277344 total time spend for step 4 : 0.07263612747192383 step5:brightness Fri May 30 03:40:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'987321136': ()} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 0 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg',) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136, 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg': 1361688639} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}, 1361688639: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg'}} map_subphoto_mainphoto : {0: 987321136, 1361688639: 987321136} inside step calcul brightness treat image : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg pour la photo_id : -0.39870825574700136, le score de luminosite est de 987321136 brightness_score : {987321136: [(987321136, -0.39870825574700136, 496442774)]} After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True select photo_hashtag_type from MTRDatou.classification_theme where id = 1154 begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 1 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 0.007861137390136719 begin to insert list_values into photo_hahstag_ids : length of list_valuse in save_photo_hashtag_id_type : 1 insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) first line : ('987321136', '496442774', '1426') ... last line : ('987321136', '496442774', '1426') time used for this insertion : 0.010180950164794922 save missing photos in datou_result : time spend for datou_step_exec : 0.06232190132141113 time spend to save output : 0.022822856903076172 total time spend for step 5 : 0.0851447582244873 step6:blur_detection Fri May 30 03:40:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} input_args_next_step : {'987321136': ()} output_args : {'987321136': ['temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 987321136, 0.9481481481481482]} args : 987321136 depend.output_id : 0 VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ('temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg',) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136, 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg': 1361688639} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}, 1361688639: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg'}} map_subphoto_mainphoto : {0: 987321136, 1361688639: 987321136} inside step blur_detection score_blur_detection : {} methode: ratio et variance treat image : temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg resize: (439, 562) 987321136 -5.392404060312662 score_blur_detection : {987321136: [(987321136, -5.392404060312662, 492609224)]} After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : False verbose : True select photo_hashtag_type from MTRDatou.classification_theme where id = 1055 begin to insert list_values into class_photo_scores : length of list_valuse in save_photo_hashtag_id_thcl_score : 1 insert into MTRPhoto.class_photo_score (thcl, photo_id, hashtag_id, score) values (%s,%s,%s,%s) on duplicate key update score = values(score) time used for this insertion : 0.00795292854309082 begin to insert list_values into photo_hahstag_ids : length of list_valuse in save_photo_hashtag_id_type : 1 insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) insert ignore into MTRBack.photo_hashtag_ids (photo_id, hashtag_id, type) values (%s,%s,%s) first line : ('987321136', '492609224', '1294') ... last line : ('987321136', '492609224', '1294') time used for this insertion : 0.009061813354492188 save missing photos in datou_result : time spend for datou_step_exec : 0.09866714477539062 time spend to save output : 0.021359682083129883 total time spend for step 6 : 0.12002682685852051 step7:send_mail_dechet Fri May 30 03:40:42 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec Currently we do not manage missing dependencies information, that could maybe be correctly interpreted with default behavior Some of the step done at execution of the step could be done before when the tree of execution is build and the dependencies of different step analysed complete output_args for input 0 : {987321136: [(987321136, -5.392404060312662, 492609224)]} input_args_next_step : {987321136: ()} output_args : {987321136: [(987321136, -5.392404060312662, 492609224)]} args : 987321136 depend.output_id : 0 complete output_args for input 1 : {987321136: [(987321136, -0.39870825574700136, 496442774)]} input_args_next_step : {987321136: ((987321136, -5.392404060312662, 492609224),)} output_args : {987321136: [(987321136, -0.39870825574700136, 496442774)]} args : 987321136 depend.output_id : 0 complete output_args for input 2 : {987321136: [({'carton': 3, 'Papier_Magazine': 7}, [1361688639], {'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}, {'refus_total': 61.64383561643836, 'carton': 61.64383561643836, 'Papier_Magazine': 38.35616438356164}, 1902940)]} input_args_next_step : {987321136: ((987321136, -5.392404060312662, 492609224), (987321136, -0.39870825574700136, 496442774))} output_args : {987321136: [({'carton': 3, 'Papier_Magazine': 7}, [1361688639], {'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}, {'refus_total': 61.64383561643836, 'carton': 61.64383561643836, 'Papier_Magazine': 38.35616438356164}, 1902940)]} args : 987321136 depend.output_id : 0 We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! We should have FATAL ERROR but same_nb_input_output==True : this should be an optionnal input ! VR 22-3-18 : For now we do not clean correctly the datou structure input_args_next_step, len :1, first value : ((987321136, -5.392404060312662, 492609224), (987321136, -0.39870825574700136, 496442774), ({'carton': 3, 'Papier_Magazine': 7}, [1361688639], {'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}, {'refus_total': 61.64383561643836, 'carton': 61.64383561643836, 'Papier_Magazine': 38.35616438356164}, 1902940)) After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg': 987321136, 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg': 1361688639} map_photo_id_path_extension : {987321136: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea.jpg', 'extension': 'jpg'}, 1361688639: {'path': 'temp/1748569212_1112947_987321136_6a08497399a24a3041045c21475a90ea_0.jpg'}} map_subphoto_mainphoto : {0: 987321136, 1361688639: 987321136} dans la step send mail dechet list_name : ['one', 'sample', 'debug', 'board', 'détect', 'port'] corps du mail : La photo est trop sombre et nette, merci de reprendre la photo
Lien affichage photo


Dans ces conditions de prise de photo, les résultats sur le tas sont les suivants :
Le pourcentage de matière impropre est de 61.64 %.

Pour plus de détails:

Teint Dans La Masse: 0%.

carton: 61.64%.

metal: 0%.

plastique: 0%.

senders@fotonower.com retour de l'envoi du mail : None After datou_step_exec type output : map_portfolio_photo : len 1 keys : dict_keys([1902940]) Inside saveOutput : final : True verbose : True saveOutput not yet implemented for datou_step.type : send_mail_dechet we use saveGeneral [987321136, 987321136, '1361688639'] map_info['map_portfolio_photo'] : {1902940: [987321136]} final : True mtd_id 1848 list_pids : [987321136, 987321136, '1361688639'] Looping around the photos to save general results len do output : 1 /987321136. before output type Here is an output not treated by saveGeneral : Managing all output in save final without adding information in the mtr_datou_result ('1848', None, None, None, None, None, None, None, None) ('1848', '1902940', '987321136', None, None, None, None, None, None) ('1848', None, None, None, None, None, None, None, None) ('1848', '1902940', '987321136', None, None, None, None, None, None) ('1848', None, None, None, None, None, None, None, None) ('1848', None, '1361688639', None, None, None, None, None, None) begin to insert list_values into mtr_datou_result : length of list_values in save_final : 4 insert ignore into MTRPhoto.mtr_datou_result (mtd_id, mtr_portfolio_id,mtr_photo_id,result,result_long,result_double,hashtag_id,proba, mtr_current_id) values (%s,%s,%s,%s,%s,%s,%s,%s,%s) on duplicate key update mtr_portfolio_id = mtr_portfolio_id list_values : [('1848', '1902940', '987321136', "{'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}", None, None, None, None, None), ('1848', None, '1361688639', None, None, None, None, None, None)] time used for this insertion : 0.016046762466430664 save_final save missing photos in datou_result : time spend for datou_step_exec : 0.6210830211639404 time spend to save output : 0.01642322540283203 total time spend for step 7 : 0.6375062465667725 caffe_path_current : About to save ! 2 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 7 output : {987321136: (-110, -0.39870825574700136, -5.392404060312662, 30.0, 61.64383561643836, {'carton': 3, 'Papier_Magazine': 7}, {'refus_total': 30.0, 'carton': 30.0, 'Papier_Magazine': 70.0}, {'refus_total': 61.64383561643836, 'carton': 61.64383561643836, 'Papier_Magazine': 38.35616438356164}, 0.6164383561643836)} ############################### TEST image_temperature_detection ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=1807 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=1807 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 1807 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=1807 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : image_temperature_detection list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT photo_id, url FROM MTRBack.photos ph WHERE photo_id IN (984484223) Found this number of photos: 1 ##### Call download_photos : nb_thread : 5 begin to download photo : 984484223 download finish for photo 984484223 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 1 ; length of list_pids : 1 ; length of list_args : 1 ##### After load_data_input time to download the photos : 0.18254399299621582 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:image_temperature_detection Fri May 30 03:40:43 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569243_1112947_984484223_2e25dc219a9a57a9f85bcae482a80c35.jpg': 984484223} map_photo_id_path_extension : {984484223: {'path': 'temp/1748569243_1112947_984484223_2e25dc219a9a57a9f85bcae482a80c35.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} inside step blanche_jaune_detection treat image : temp/1748569243_1112947_984484223_2e25dc219a9a57a9f85bcae482a80c35.jpg 984484223 1.004309911525615 After datou_step_exec type output : time spend for datou_step_exec : 0.2318260669708252 time spend to save output : 6.318092346191406e-05 total time spend for step 1 : 0.2318892478942871 caffe_path_current : About to save ! 0 After save, about to update current ! datou_cur_ids : [] len(datou.list_steps) : 1 output : {984484223: [(984484223, 1.004309911525615, 492630606)]} {984484223: [(984484223, 1.004309911525615, 492630606)]} ############################### TEST broca ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=4041 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=4041 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 4041 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=4041 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! We are managing only one step so we do not consider checkConsistencyNbInputNbOutput ! We are managing only one step so we do not consider checkConsistencyTypeOutputInput ! no param json to modify List Step Type Loaded in datou : split_time_score list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT ph.photo_id, ph.url FROM MTRBack.photos ph WHERE ph.photo_id IN (SELECT mtr_photo_id from MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (5205529) and hide_status = 0 ) ORDER BY ph.photo_id DESC LIMIT 0, 10000 We have 1 , {} SELECT mtr_photo_id, mtr_portfolio_id FROM MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (5205529) AND hide_status = 0 ORDER by mtr_photo_id desc LIMIT 0, 10000 list_result: [{'photo_id': 1064921404, 'portfolio_id': 5205529}, {'photo_id': 1064921402, 'portfolio_id': 5205529}, {'photo_id': 1064921401, 'portfolio_id': 5205529}, {'photo_id': 1064921201, 'portfolio_id': 5205529}, {'photo_id': 1064921196, 'portfolio_id': 5205529}, {'photo_id': 1064919876, 'portfolio_id': 5205529}, {'photo_id': 1064919873, 'portfolio_id': 5205529}, {'photo_id': 1064919869, 'portfolio_id': 5205529}, {'photo_id': 1064919862, 'portfolio_id': 5205529}, {'photo_id': 1064919858, 'portfolio_id': 5205529}, {'photo_id': 1064919856, 'portfolio_id': 5205529}, {'photo_id': 1064919752, 'portfolio_id': 5205529}, {'photo_id': 1064919748, 'portfolio_id': 5205529}, {'photo_id': 1064919745, 'portfolio_id': 5205529}, {'photo_id': 1064919741, 'portfolio_id': 5205529}, {'photo_id': 1064919737, 'portfolio_id': 5205529}, {'photo_id': 1064919730, 'portfolio_id': 5205529}, {'photo_id': 1064919660, 'portfolio_id': 5205529}] map_portfolio_id_photo_id: {5205529: [1064921404, 1064921402, 1064921401, 1064921201, 1064921196, 1064919876, 1064919873, 1064919869, 1064919862, 1064919858, 1064919856, 1064919752, 1064919748, 1064919745, 1064919741, 1064919737, 1064919730, 1064919660]} ##### Call download_photos : nb_thread : 5 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos ##### After load_data_input time to download the photos : 0.018633604049682617 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load Calling datou_exec Inside datou_exec : verbose : True number of steps : 1 step1:split_time_score Fri May 30 03:40:43 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {} map_photo_id_path_extension : {} map_subphoto_mainphoto : {} split portfolio by speed calcul order for each photo with time calcul time for a portfolio query : SELECT photo_id, text FROM MTRBack.photos where photo_id in (SELECT mtr_photo_id FROM MTRUser.mtr_portfolio_photos where mtr_portfolio_id = 5205529); result : ((1064919660, 'image_01122021_10_11_30_014389.jpg'), (1064919730, 'image_01122021_10_12_17_665202.jpg'), (1064919737, 'image_01122021_10_11_40_031052.jpg'), (1064919741, 'image_01122021_10_11_34_021658.jpg'), (1064919745, 'image_01122021_10_11_32_018001.jpg'), (1064919748, 'image_01122021_10_12_27_027057.jpg'), (1064919752, 'image_01122021_10_12_24_005017.jpg'), (1064919856, 'image_01122021_10_13_13_399843.jpg'), (1064919858, 'image_01122021_10_13_04_729164.jpg'), (1064919862, 'image_01122021_10_12_56_581019.jpg'), (1064919869, 'image_01122021_10_12_29_030603.jpg'), (1064919873, 'image_01122021_10_13_30_005720.jpg'), (1064919876, 'image_01122021_10_13_22_147712.jpg'), (1064921196, 'image_01122021_10_16_18_114975.jpg'), (1064921201, 'image_01122021_10_16_14_925132.jpg'), (1064921401, 'image_01122021_10_16_57_981306.jpg'), (1064921402, 'image_01122021_10_16_53_913663.jpg'), (1064921404, 'image_01122021_10_16_47_889875.jpg')) INSERT ignore into MTRUser.mtr_portfolio_photos (`mtr_portfolio_id`, `mtr_photo_id`, `order`) VALUES (%s, %s, %s) on duplicate key update `order`=VALUES(`order`); first line : (5205529, 1064919660, 1098136690) ... last line : (5205529, 1064921404, 1098137007) 2021-12-01 10:11:30 2021-12-01 10:11:32 2021-12-01 10:11:30 2021-12-01 10:11:34 2021-12-01 10:11:32 2021-12-01 10:11:40 2021-12-01 10:11:34 2021-12-01 10:12:17 2021-12-01 10:11:40 2021-12-01 10:12:24 2021-12-01 10:12:17 2021-12-01 10:12:27 2021-12-01 10:12:24 2021-12-01 10:12:29 2021-12-01 10:12:27 2021-12-01 10:12:56 2021-12-01 10:12:29 2021-12-01 10:13:04 2021-12-01 10:12:56 2021-12-01 10:13:13 2021-12-01 10:13:04 2021-12-01 10:13:04 distance 1.4513659170185111 2021-12-01 10:13:13 2021-12-01 10:13:22 2021-12-01 10:13:13 2021-12-01 10:13:30 2021-12-01 10:13:22 2021-12-01 10:16:14 2021-12-01 10:13:30 2021-12-01 10:13:30 distance 8.382409567451603 2021-12-01 10:16:14 2021-12-01 10:16:18 2021-12-01 10:16:14 2021-12-01 10:16:47 2021-12-01 10:16:18 2021-12-01 10:16:53 2021-12-01 10:16:47 2021-12-01 10:16:47 distance 8.03396608896571 2021-12-01 10:16:53 2021-12-01 10:16:57 2021-12-01 10:16:53 dict_time_useful: {0: [1098136690, 1098136784, 48.864288393888884, 2.19199505125, [datetime.datetime(2021, 12, 1, 10, 11, 30), datetime.datetime(2021, 12, 1, 10, 13, 4), 94]], 1: [1098136974, 1098137007, 48.86291258986111, 2.19361357125, [datetime.datetime(2021, 12, 1, 10, 16, 14), datetime.datetime(2021, 12, 1, 10, 16, 47), 33]]} len of dic_time_useful : 2 get gps info of PAV SELECT id,Y_WGS84,X_WGS84 FROM MTRLabel.info_PAV; get gps info of PAV SELECT id,Y_WGS84,X_WGS84 FROM MTRLabel.info_PAV WHERE type_pav = "CS"; get gps info of PAV SELECT id,Y_WGS84,X_WGS84 FROM MTRLabel.info_PAV WHERE type_pav = "OM"; select cs_nb_photo / nb_photo, om_nb_photo / nb_photo from (select sum(1) as nb_photo,sum(if (tags= "[CS]",1,0)) as cs_nb_photo, sum(if (tags= "[OM]",1,0)) as om_nb_photo from MTRBack.photos where photo_id in ()) t1; select cs_nb_photo / nb_photo, om_nb_photo / nb_photo from (select sum(1) as nb_photo,sum(if (tags= "[CS]",1,0)) as cs_nb_photo, sum(if (tags= "[OM]",1,0)) as om_nb_photo from MTRBack.photos where photo_id in (1064919660, 1064919745, 1064919741, 1064919737, 1064919730, 1064919752, 1064919748, 1064919869, 1064919862, 1064919858)) t1; distance: RUEIL14CS [48.864288393888884, 2.19199505125] 16.57008455321128 (23444771, 48.864288393888884, 2.19199505125, 10, 1064919752, [datetime.datetime(2021, 12, 1, 10, 11, 30), datetime.datetime(2021, 12, 1, 10, 13, 4), 94.0], 5205529) After datou_step_exec type output : time spend for datou_step_exec : 0.15579628944396973 time spend to save output : 0.00011110305786132812 total time spend for step 1 : 0.15590739250183105 caffe_path_current : About to save ! 0 After save, about to update current ! {15: [(23444771, 48.864288393888884, 2.19199505125, 10, 1064919752, [datetime.datetime(2021, 12, 1, 10, 11, 30), datetime.datetime(2021, 12, 1, 10, 13, 4), 94.0], 5205529)]} résultat du premier test BROCA : True True ############################### TEST crop_conditional ################################ t Inside batchDatouExec : verbose : True ##### chargement datou SELECT name, created_at,limit_max FROM MTRDatou.mtr_datou WHERE id=719 SELECT mtd.id, mtdt.`type`, mtd.`param`, mtd.param_json, mtdt.nb_input, mtdt.nb_output, mtdt.prod, mtdt.is_local, mtdt.is_datou_depend, mtdt.is_photo_id_local FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_step_types mtdt WHERE mtdt.`id`=mtd.`type` AND mtd.mtd_id=719 SELECT mtd.id, mtd.mtd_id, mdsdt.id, mdsdt.name, mdsdt.description, msid.output_or_input, msid.data_order_id, mdsdt.type FROM MTRDatou.mtr_datou_step mtd, MTRDatou.mtr_datou_steptype_io_datatypes msid, MTRDatou.mtr_datou_step_data_types mdsdt WHERE mtd.`type`=msid.`mtr_datou_step_type` AND mtd.mtd_id= 719 AND msid.data_type=mdsdt.id SELECT mts_id_output, id_output, mts_id_input, id_input FROM MTRDatou.mtr_datou_step_by_step WHERE mtd_id=719 # VR 17-11-17 : to create in DB ! Here we check the datou graph and we reorder steps ! Tree builded and cycle checked, now we need to re-order the steps ! We have currenlty an error because there is no dependence between the last step for the case tile - detect - glue We can either keep the depence of, it is better to keep an order compatible with the id of steps if we do not have sons, so a lexical order : (number_son, step_id) DONE and to test : checkNoCycle ! Here we check the consistency of inputs/outputs number between the given ones and the db ! eke 1-6-18 : checkConsistencyNbInputNbOutput should be processed after step reordering ! WARNING : step 1335 frcnn is not linked in the step_by_step architecture ! WARNING : step 1336 crop_condition is not linked in the step_by_step architecture ! Number of inputs / outputs for each step checked ! Here we check the consistency of outputs/inputs types during steps connections eke 1-6-18 : checkConsistencyTypeOutputInput should be processed after checkConsistencyNbInputNbOutput ! DataTypes for each output/input checked ! no param json to modify List Step Type Loaded in datou : frcnn, crop_condition list_input_json : [] ##### fin chargement datou ##### chargement data ##### Call load_data_input : nb_thread : 5 origin SELECT ph.photo_id, ph.url FROM MTRBack.photos ph WHERE ph.photo_id IN (SELECT mtr_photo_id from MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (1981316) and hide_status = 0 ) ORDER BY ph.photo_id DESC LIMIT 0, 10000 We have 1 , {} SELECT mtr_photo_id, mtr_portfolio_id FROM MTRUser.mtr_portfolio_photos WHERE mtr_portfolio_id in (1981316) AND hide_status = 0 ORDER by mtr_photo_id desc LIMIT 0, 10000 list_result: [{'photo_id': 950003838, 'portfolio_id': 1981316}, {'photo_id': 950003813, 'portfolio_id': 1981316}, {'photo_id': 950003812, 'portfolio_id': 1981316}, {'photo_id': 950003696, 'portfolio_id': 1981316}, {'photo_id': 950003695, 'portfolio_id': 1981316}, {'photo_id': 926687666, 'portfolio_id': 1981316}] map_portfolio_id_photo_id: {1981316: [950003838, 950003813, 950003812, 950003696, 950003695, 926687666]} ##### Call download_photos : nb_thread : 5 begin to download photo : 950003838 begin to download photo : 950003812 begin to download photo : 950003695 download finish for photo 950003812 begin to download photo : 950003696 download finish for photo 950003838 begin to download photo : 950003813 download finish for photo 950003695 begin to download photo : 926687666 download finish for photo 950003696 download finish for photo 926687666 download finish for photo 950003813 we have missing 0 photos in the step downloads : photo missing : [] try to delete the photos missing in DB ##### After download_photos length of list_filenames : 6 ; length of list_pids : 6 ; length of list_args : 6 ##### After load_data_input time to download the photos : 0.3799450397491455 #### fin chargement data Blocking on flush ? No conitnuing About to test input to load we should then remove the video here, and this would fix the bug of datou_current ! Calling datou_exec Inside datou_exec : verbose : True number of steps : 2 step1:frcnn Fri May 30 03:40:44 2025 VR 17-11-17 : now, only for linear exec dependencies tree, some output goes to fill the input of the next VR 22-3-18 : now we test the dependencies tree, but keep two separate code for datou_prepare_output_input until the code is correctly tested, clean and works in both case VR 22-3-18 : but we use the first code for the first step id = -1, build in the code of datou_exec VR 22-3-18 : we should manage here the case when we are at the first step instead of building this step before datou_exec After prepare type args : Here we display some param of map_info ! map_filenames : {'temp/1748569244_1112947_950003812_3dbffe9f441f7d28d087f3e571769e74.jpg': 950003812, 'temp/1748569244_1112947_950003696_11e3a77b72af4b332d366d98984039c7.jpg': 950003696, 'temp/1748569244_1112947_950003695_22b4110c9a86b12e1542ec2bb977f6a8.jpg': 950003695, 'temp/1748569244_1112947_926687666_a8bc8c1fad77748c62ca641ceb29ad9c.jpg': 926687666, 'temp/1748569244_1112947_950003838_e480bc28e6ceabc2f5995246a6af6b46.jpg': 950003838, 'temp/1748569244_1112947_950003813_e28be02dfcce79cce594a390a9911a0a.jpg': 950003813} map_photo_id_path_extension : {950003812: {'path': 'temp/1748569244_1112947_950003812_3dbffe9f441f7d28d087f3e571769e74.jpg', 'extension': 'jpg'}, 950003696: {'path': 'temp/1748569244_1112947_950003696_11e3a77b72af4b332d366d98984039c7.jpg', 'extension': 'jpg'}, 950003695: {'path': 'temp/1748569244_1112947_950003695_22b4110c9a86b12e1542ec2bb977f6a8.jpg', 'extension': 'jpg'}, 926687666: {'path': 'temp/1748569244_1112947_926687666_a8bc8c1fad77748c62ca641ceb29ad9c.jpg', 'extension': 'jpg'}, 950003838: {'path': 'temp/1748569244_1112947_950003838_e480bc28e6ceabc2f5995246a6af6b46.jpg', 'extension': 'jpg'}, 950003813: {'path': 'temp/1748569244_1112947_950003813_e28be02dfcce79cce594a390a9911a0a.jpg', 'extension': 'jpg'}} map_subphoto_mainphoto : {} Beginning of datou step Faster rcnn ! Inside try reload ! classes : ['background', 'retroviseur', 'roue', 'capot', 'pare-brise', 'vitre', 'phare', 'feu-antibrouillard', 'feu-arriere', 'poignee', 'porte', 'radiateur', 'logo-marque', 'cache-reservoir', 'plaque-immatriculation', 'pot-echappement', 'info-modele', 'essuie-glace', 'pare-choc', 'coffre', 'carrosserie-autre', 'toit', 'logo-roue', 'aile-avant', 'aile-arriere', 'autre'] pht : 757 caffemodel_name (should be vgg16_immat_307 but not used because net loaded outside in the fonction) : {'id': 685, 'mtr_user_id': 31, 'name': 'learn_piece_voiture_0808_v2', 'pb_hashtag_id': 0, 'live': b'\x00', 'list_hashtags': 'background,retroviseur,roue,capot,pare-brise,vitre,phare,feu-antibrouillard,feu-arriere,poignee,porte,radiateur,logo-marque,cache-reservoir,plaque-immatriculation,pot-echappement,info-modele,essuie-glace,pare-choc,coffre,carrosserie-autre,toit,logo-roue,aile-avant,aile-arriere,autre', 'svm_portfolios_learning': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0', 'photo_hashtag_type': 757, 'photo_desc_type': 3800, 'type_classification': 'caffe_faster_rcnn', 'hashtag_id_list': '0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0'} To loadFromThcl() model_param file didn't exist model_name : learn_piece_voiture_0808_v2 model_type : caffe_faster_rcnn list file need : ['caffemodel', 'test.prototxt'] file exist in s3 : ['caffemodel', 'test.prototxt'] file manque in s3 : [] local folder : /data/models_weight/learn_piece_voiture_0808_v2 /data/models_weight/learn_piece_voiture_0808_v2/caffemodel size_local : 350215080 size in s3 : 350215080 create time local : 2021-08-09 05:30:22 create time in s3 : 2021-08-06 19:24:16 caffemodel already exist and didn't need to update /data/models_weight/learn_piece_voiture_0808_v2/test.prototxt size_local : 7166 size in s3 : 7166 create time local : 2021-08-09 05:30:22 create time in s3 : 2021-08-06 19:24:16 test.prototxt already exist and didn't need to update prototxt : /data/models_weight/learn_piece_voiture_0808_v2/test.prototxt caffemodel : /data/models_weight/learn_piece_voiture_0808_v2/caffemodel Loaded network /data/models_weight/learn_piece_voiture_0808_v2/caffemodel About to compute detect_faster_rcnn : len(args) : 6 Inside frcnn step exec : nb paths : 6 image_path : temp/1748569244_1112947_950003812_3dbffe9f441f7d28d087f3e571769e74.jpg image_size (480, 614, 3) [[[ 44 44 44] [ 49 51 51] [ 42 44 44] ... [ 8 10 10] [ 8 10 10] [ 8 10 10]] [[ 43 43 43] [ 36 38 38] [ 39 41 41] ... [ 5 7 7] [ 5 7 7] [ 5 7 7]] [[ 70 70 70] [ 40 42 42] [ 41 43 43] ... [ 4 6 6] [ 4 6 6] [ 4 6 6]] ... [[103 101 101] [110 108 108] [ 61 59 59] ... [ 8 10 10] [ 8 10 10] [ 8 10 10]] [[ 98 96 96] [115 113 113] [ 73 71 71] ... [ 0 2 2] [ 11 13 13] [ 21 23 23]] [[ 92 90 90] [114 112 112] [ 87 82 83] ... [ 10 12 12] [ 18 20 20] [ 25 27 27]]] Detection took 0.129s for 300 object proposals c : aile-arriere list_crops.shape (30, 5) proba : 0.13241859 (133.2964, 172.39214, 280.8272, 321.49292) proba : 0.013275191 (153.19482, 232.50722, 369.76282, 359.7831) c : aile-avant list_crops.shape (35, 5) proba : 0.95039564 (133.96971, 146.16376, 305.77844, 344.4903) proba : 0.020751996 (128.15636, 143.01431, 188.9175, 305.02066) proba : 0.014926507 (294.24365, 331.83704, 412.49805, 415.32684) proba : 0.011201178 (533.057, 198.27655, 583.49744, 268.45612) proba : 0.010010731 (0.0, 114.213425, 47.591232, 234.00322) c : autre list_crops.shape (36, 5) c : cache-reservoir list_crops.shape (35, 5) proba : 0.0256786 (426.7074, 57.94274, 496.61603, 125.31121) proba : 0.025326086 (342.5213, 266.61554, 421.34277, 330.5551) proba : 0.01725696 (306.41083, 352.4264, 382.6065, 421.99768) c : capot list_crops.shape (32, 5) proba : 0.9904727 (196.74986, 110.3707, 532.50415, 291.96805) proba : 0.22531678 (40.24803, 25.96585, 253.16167, 48.717735) proba : 0.0584295 (318.88617, 286.61154, 481.59198, 427.8967) proba : 0.017968638 (64.42415, 38.41092, 331.29376, 64.35657) proba : 0.010640398 (116.146935, 36.753754, 368.63474, 155.92374) c : carrosserie-autre list_crops.shape (29, 5) proba : 0.030300563 (429.34903, 22.063953, 497.8182, 122.47893) c : coffre list_crops.shape (29, 5) proba : 0.1230198 (421.339, 15.521957, 501.77576, 119.22037) proba : 0.10120325 (112.73723, 46.268753, 395.27368, 193.28386) proba : 0.062833354 (247.78433, 171.97992, 517.5138, 335.85544) c : essuie-glace list_crops.shape (37, 5) proba : 0.77170855 (178.35667, 109.0693, 389.54022, 147.70059) proba : 0.14876041 (343.28015, 268.29117, 419.46075, 332.34402) proba : 0.100799516 (364.45364, 261.52032, 501.23184, 317.86395) proba : 0.047933362 (446.58145, 21.589863, 508.33017, 88.695724) proba : 0.03781434 (328.7596, 360.46265, 389.8558, 423.33307) proba : 0.01470514 (85.602455, 24.01562, 308.1059, 43.347435) proba : 0.013854571 (267.03516, 240.3375, 412.9615, 348.48758) proba : 0.012837622 (427.0822, 58.162174, 496.34943, 124.402145) c : feu-antibrouillard list_crops.shape (37, 5) proba : 0.21099111 (307.2797, 354.31027, 381.7961, 421.6066) proba : 0.056967083 (342.8655, 269.36563, 420.86002, 329.44168) proba : 0.028142316 (427.13022, 58.8235, 496.1465, 124.55002) proba : 0.023515554 (362.9285, 264.44678, 504.0895, 316.8405) c : feu-arriere list_crops.shape (34, 5) proba : 0.09635342 (303.1506, 259.64258, 431.2908, 316.4461) proba : 0.04133272 (428.95325, 25.621979, 497.48468, 104.485794) proba : 0.03446671 (305.47357, 351.5556, 379.51233, 423.9369) c : info-modele list_crops.shape (36, 5) proba : 0.04286909 (306.73846, 353.5112, 383.1764, 421.5568) proba : 0.035217952 (426.5307, 57.958996, 497.43683, 125.61151) proba : 0.03343222 (342.96042, 267.55997, 422.6783, 330.89572) proba : 0.016011817 (363.58508, 261.26358, 507.57068, 317.39316) proba : 0.010811276 (445.9002, 19.326252, 507.99664, 91.46188) c : logo-marque list_crops.shape (36, 5) proba : 0.10543189 (327.43307, 361.328, 387.9539, 423.0327) proba : 0.02840331 (344.96402, 265.39206, 421.6273, 329.04623) proba : 0.025367392 (370.8595, 261.33926, 500.6148, 316.38693) proba : 0.017487604 (427.60837, 57.653343, 496.78586, 125.72262) c : logo-roue list_crops.shape (34, 5) proba : 0.014206147 (327.66193, 360.72095, 388.79962, 424.03906) c : pare-brise list_crops.shape (30, 5) proba : 0.95463026 (113.606415, 42.325085, 417.9238, 147.556) proba : 0.106999345 (294.6766, 24.32906, 399.7059, 129.18121) proba : 0.054095212 (258.47925, 161.2626, 521.5874, 295.05353) proba : 0.027617512 (424.62775, 20.867443, 499.57227, 97.22647) proba : 0.01762031 (72.58977, 41.74168, 127.25343, 168.79155) c : pare-choc list_crops.shape (25, 5) proba : 0.94855964 (247.46376, 265.4706, 552.9343, 445.27597) proba : 0.09525801 (194.32129, 225.2338, 348.43427, 419.22818) proba : 0.019614356 (422.45078, 329.61032, 549.5321, 424.81088) proba : 0.014919346 (455.54987, 26.12846, 592.8155, 120.20906) c : phare list_crops.shape (35, 5) proba : 0.67815506 (318.13898, 264.64725, 489.91864, 310.62466) proba : 0.403282 (261.63992, 234.11765, 408.51498, 331.02524) proba : 0.095903926 (301.1037, 356.41315, 385.4995, 425.2309) proba : 0.014656689 (525.6668, 198.5462, 575.21747, 289.94843) proba : 0.013314316 (427.9991, 23.336739, 501.3008, 100.64818) proba : 0.011495674 (277.73596, 207.81818, 557.26605, 296.85703) c : plaque-immatriculation list_crops.shape (36, 5) proba : 0.19053088 (491.39795, 294.22552, 563.82385, 390.1781) proba : 0.05778302 (439.5147, 291.40808, 531.0498, 407.167) proba : 0.0229528 (298.98117, 257.4155, 447.53647, 319.17233) proba : 0.01134137 (309.84048, 356.2036, 386.15775, 421.93097) c : poignee list_crops.shape (34, 5) proba : 0.025670175 (327.50812, 360.9723, 388.47064, 424.01715) proba : 0.02421679 (342.31424, 266.89, 422.49307, 331.5276) proba : 0.021988707 (426.4331, 57.759518, 497.0761, 125.795135) proba : 0.013407836 (559.0073, 0.020687103, 613.0, 71.29167) c : porte list_crops.shape (27, 5) proba : 0.99229604 (51.414944, 40.436646, 152.79457, 306.38666) proba : 0.97203076 (4.568203, 51.358734, 72.87775, 242.46848) proba : 0.05866451 (425.32352, 19.077072, 501.1229, 130.88013) proba : 0.030524034 (132.08403, 46.461075, 397.16632, 213.57866) proba : 0.014542678 (360.3008, 244.28137, 551.6101, 395.6222) c : pot-echappement list_crops.shape (36, 5) proba : 0.05058703 (328.16937, 360.6976, 387.76764, 423.72873) proba : 0.020146824 (427.16434, 57.469746, 496.29074, 125.83998) c : radiateur list_crops.shape (35, 5) c : retroviseur list_crops.shape (35, 5) proba : 0.43161234 (427.41595, 56.783222, 495.94183, 123.6676) proba : 0.18612465 (344.11072, 266.8148, 421.86273, 331.86935) proba : 0.11663672 (446.86646, 19.437637, 506.46387, 90.94862) proba : 0.04802851 (316.70703, 363.19122, 377.6407, 430.79016) proba : 0.027539348 (150.8237, 116.115486, 381.60434, 150.22705) proba : 0.012777411 (366.31247, 260.31393, 504.92538, 318.49088) proba : 0.011191455 (77.32788, 81.656006, 130.17639, 174.47607) c : roue list_crops.shape (37, 5) proba : 0.93756336 (166.73476, 261.1287, 275.2211, 420.81113) proba : 0.14076139 (307.7933, 351.3524, 385.35587, 430.96262) proba : 0.114036255 (4.556036, 161.28508, 50.79972, 253.06636) proba : 0.045009546 (513.9557, 252.55635, 573.92126, 393.48242) proba : 0.037106495 (526.6661, 191.77722, 581.601, 303.90338) c : toit list_crops.shape (33, 5) proba : 0.7964302 (58.201828, 31.45493, 316.00787, 55.59937) c : vitre list_crops.shape (33, 5) proba : 0.9803726 (69.39569, 49.349873, 135.8343, 142.42229) proba : 0.8937804 (14.668219, 41.21925, 66.174065, 115.17732) proba : 0.117509164 (205.48627, 48.102364, 379.27405, 127.906136) proba : 0.043755464 (428.1306, 21.90049, 499.21252, 97.15434) proba : 0.017626302 (308.70743, 352.47455, 385.6868, 423.24506) proba : 0.012153623 (338.94946, 258.5371, 419.8844, 323.73865) We are managing local photo_id image_path : temp/1748569244_1112947_950003696_11e3a77b72af4b332d366d98984039c7.jpg image_size (2160, 3264, 3) [[[168 165 161] [168 165 161] [168 165 161] ... [ 47 59 63] [ 48 60 64] [ 48 60 64]] [[168 165 161] [168 165 161] [168 165 161] ... [ 47 59 63] [ 47 59 63] [ 48 60 64]] [[168 165 161] [168 165 161] [168 165 161] ... [ 47 59 63] [ 47 59 63] [ 47 59 63]] ... [[167 164 160] [167 164 160] [167 164 160] ... [ 44 59 61] [ 44 59 61] [ 44 59 61]] [[165 162 158] [165 162 158] [165 162 158] ... [ 45 60 62] [ 45 60 62] [ 45 60 62]] [[164 161 157] [164 161 157] [164 161 157] ... [ 45 60 62] [ 45 60 62] [ 45 60 62]]] Detection took 1.283s for 300 object proposals c : aile-arriere list_crops.shape (52, 5) proba : 0.020721467 (396.2277, 192.55856, 703.2373, 482.5087) proba : 0.013662464 (2606.4824, 1489.7375, 2954.395, 2131.393) proba : 0.010280686 (2270.1775, 1651.4276, 2706.3865, 2138.8582) proba : 0.010018697 (2922.7305, 683.2705, 3263.0, 1188.8912) c : aile-avant list_crops.shape (55, 5) proba : 0.02582956 (2616.2043, 1475.3534, 2938.8386, 2144.0322) c : autre list_crops.shape (53, 5) proba : 0.013096714 (405.69415, 192.22011, 689.091, 476.50537) c : cache-reservoir list_crops.shape (50, 5) proba : 0.029478546 (2629.4731, 1461.0208, 2954.7158, 2148.8457) proba : 0.024540521 (299.21423, 89.51213, 716.6609, 498.37225) proba : 0.020117803 (2232.053, 676.7636, 2590.5251, 1171.2189) proba : 0.018791022 (2298.3862, 1682.803, 2783.6602, 2156.6445) proba : 0.012400504 (2317.1357, 1436.0543, 2614.4814, 1924.9572) proba : 0.012134908 (2925.146, 677.3063, 3258.894, 1209.0381) c : capot list_crops.shape (44, 5) proba : 0.018012634 (2645.9849, 973.0453, 3214.732, 1642.0939) proba : 0.016425302 (77.74625, 795.9651, 694.0112, 1481.112) c : carrosserie-autre list_crops.shape (45, 5) proba : 0.029314028 (251.7171, 0.0, 712.3475, 634.9905) proba : 0.013689069 (2284.6985, 1569.4866, 2687.3674, 2159.0) c : coffre list_crops.shape (42, 5) proba : 0.13940582 (283.03708, 0.0, 631.68713, 633.40295) proba : 0.013912742 (1720.6101, 507.72974, 2227.865, 1191.366) c : essuie-glace list_crops.shape (56, 5) proba : 0.018926969 (320.886, 110.23988, 600.9978, 462.47266) proba : 0.013813608 (2187.6116, 749.4853, 2497.323, 1214.4537) proba : 0.013557544 (2629.3281, 1462.932, 2950.8657, 2139.6462) proba : 0.013343526 (74.38254, 721.1958, 391.73373, 1112.9584) c : feu-antibrouillard list_crops.shape (52, 5) proba : 0.060854994 (372.1004, 242.48886, 759.43317, 515.92163) proba : 0.0183085 (2928.3777, 694.31165, 3257.6501, 1206.6466) proba : 0.018102482 (2299.5417, 1693.8593, 2784.1243, 2158.6943) proba : 0.015854767 (2230.3699, 698.5495, 2591.1804, 1175.3088) proba : 0.01435617 (2630.748, 1480.992, 2954.4736, 2151.7498) proba : 0.011737246 (383.55563, 1437.1194, 937.1349, 1845.1277) c : feu-arriere list_crops.shape (48, 5) proba : 0.6467057 (354.8506, 213.37936, 730.1616, 532.0514) proba : 0.12032885 (2225.6653, 661.19995, 2605.2288, 1235.9197) proba : 0.079391986 (256.1928, 24.239136, 540.2498, 478.11493) proba : 0.060691223 (1807.4395, 656.71423, 2155.8953, 1147.34) proba : 0.033559885 (2308.1853, 1667.5717, 2786.756, 2159.0) proba : 0.032259706 (2930.4404, 638.0879, 3237.7637, 1243.1892) proba : 0.018529493 (619.9441, 619.2528, 953.74536, 1223.5864) proba : 0.017301189 (2630.6829, 1496.9249, 2955.3792, 2159.0) proba : 0.016658014 (2317.2385, 1411.0967, 2618.2644, 1961.6982) proba : 0.015818492 (770.62537, 939.7186, 1053.5703, 1456.1454) proba : 0.010596907 (2489.4482, 424.4956, 2939.1895, 1234.4342) proba : 0.010258745 (828.07855, 1.9279022, 1175.0165, 369.92242) proba : 0.010093563 (14.196922, 430.89215, 165.10394, 734.0317) c : info-modele list_crops.shape (49, 5) proba : 0.07010075 (403.72833, 194.21487, 693.0529, 475.90414) proba : 0.022200566 (2299.254, 1684.2527, 2787.1133, 2154.6646) proba : 0.019054431 (2232.6118, 682.5421, 2592.748, 1170.8453) proba : 0.017020311 (2630.581, 1468.2198, 2959.7212, 2151.3442) proba : 0.010883385 (1.7524109, 575.7824, 261.89764, 885.71014) c : logo-marque list_crops.shape (51, 5) proba : 0.051885925 (408.4953, 200.25745, 688.273, 481.55743) proba : 0.034075037 (2304.7776, 1690.4414, 2793.887, 2159.0) proba : 0.018234424 (2631.4712, 1489.253, 2967.995, 2159.0) proba : 0.013037474 (2236.6926, 683.90155, 2594.6047, 1180.4988) proba : 0.011358403 (2319.9644, 1440.4084, 2621.3857, 1933.8372) c : logo-roue list_crops.shape (52, 5) proba : 0.03021933 (402.78076, 191.06743, 692.7543, 478.266) proba : 0.020899136 (2230.934, 677.7543, 2592.73, 1174.01) proba : 0.018412674 (2512.7612, 492.68985, 2833.5454, 843.99023) proba : 0.018071637 (2298.6833, 1681.3528, 2786.0842, 2156.1255) proba : 0.013551067 (2610.3896, 558.7997, 3007.5342, 1178.0139) proba : 0.013279904 (1762.8013, 675.0884, 2234.712, 1109.9471) proba : 0.012411377 (2317.8042, 1434.3539, 2615.2627, 1928.4259) proba : 0.011194086 (2783.5854, 1527.3981, 3036.1206, 2077.7427) c : pare-brise list_crops.shape (46, 5) proba : 0.05498699 (2411.5818, 398.67117, 2957.3645, 1171.5482) proba : 0.04502579 (2240.3513, 634.90924, 2579.2712, 1217.4434) proba : 0.02693876 (1424.8811, 304.92975, 2074.7837, 1161.5789) proba : 0.025700279 (312.5487, 97.246704, 605.9203, 472.59015) proba : 0.02420132 (2954.8013, 604.75757, 3255.3604, 1264.4807) proba : 0.021959618 (2056.87, 93.77356, 2392.5225, 713.7262) proba : 0.020822493 (323.81378, 1090.6993, 787.1589, 1678.9515) proba : 0.01857964 (103.620514, 794.4295, 696.2057, 1524.2798) proba : 0.017082958 (2701.8086, 949.3706, 3158.2378, 1362.8354) proba : 0.011479067 (2240.9038, 3.3504028, 2720.1997, 772.5817) proba : 0.010908261 (1694.8157, 659.701, 2162.8735, 1116.5872) c : pare-choc list_crops.shape (38, 5) proba : 0.106185585 (347.37988, 1367.4277, 1073.0182, 1857.0481) proba : 0.016682357 (1310.4026, 1570.4385, 2576.6174, 2051.2693) proba : 0.010981751 (0.0, 545.1247, 997.969, 1254.1414) proba : 0.01014831 (2850.2627, 1465.3717, 3233.6704, 2099.7495) c : phare list_crops.shape (49, 5) proba : 0.06944582 (2930.6892, 682.5454, 3257.6663, 1220.24) proba : 0.0641104 (310.48932, 118.323166, 715.94586, 508.53906) proba : 0.028806487 (324.56912, 1179.3406, 758.80994, 1634.1941) proba : 0.013004558 (2226.5781, 695.24585, 2603.4463, 1195.8197) proba : 0.011205127 (83.39986, 743.9847, 419.85736, 1117.4833) proba : 0.010501311 (2318.079, 1449.4436, 2631.6812, 1922.1353) c : plaque-immatriculation list_crops.shape (56, 5) proba : 0.057377104 (399.62717, 207.96106, 679.60803, 493.95428) proba : 0.045757234 (1744.0883, 665.86084, 2221.1555, 1071.6482) proba : 0.023891006 (2294.9841, 1697.7523, 2792.1926, 2132.5796) proba : 0.023446029 (2538.795, 1521.5745, 2898.875, 2073.3608) proba : 0.013996713 (2783.4604, 1564.8704, 3034.891, 2056.8032) proba : 0.013541733 (2245.3367, 695.26514, 2573.1736, 1168.442) proba : 0.0113009075 (2945.9404, 676.897, 3251.0166, 1190.9637) c : poignee list_crops.shape (48, 5) proba : 0.057073828 (2629.5083, 1457.6089, 2958.7676, 2152.9631) proba : 0.023785373 (2299.0864, 1682.4332, 2785.608, 2157.3928) proba : 0.018129984 (2232.1729, 677.85364, 2592.1064, 1173.5947) proba : 0.014422222 (403.02356, 191.02992, 692.26697, 478.38232) c : porte list_crops.shape (42, 5) proba : 0.1267709 (2295.5757, 1531.3809, 2685.6675, 2159.0) proba : 0.044953022 (2857.5166, 1450.7072, 3263.0, 2032.9554) proba : 0.041269533 (2466.142, 1663.9409, 3030.1992, 2159.0) proba : 0.028290616 (390.16098, 1195.2065, 897.9453, 1796.6548) proba : 0.023545831 (1444.4948, 1494.099, 2537.7588, 2103.853) proba : 0.017271917 (311.3313, 15.9531555, 609.7392, 659.75183) proba : 0.012647296 (2938.577, 622.98663, 3247.8054, 1300.4646) c : pot-echappement list_crops.shape (49, 5) proba : 0.03307391 (2632.8499, 1456.623, 2954.7917, 2152.6843) proba : 0.030935042 (2302.486, 1680.4297, 2781.5725, 2157.6838) proba : 0.015973596 (405.30927, 189.9841, 691.14386, 477.74838) proba : 0.011729476 (67.23358, 695.8216, 394.5915, 1122.9194) proba : 0.011400241 (1748.3762, 537.0152, 2217.035, 1143.0588) proba : 0.011030858 (2926.5417, 674.9753, 3258.5608, 1211.8645) c : radiateur list_crops.shape (50, 5) c : retroviseur list_crops.shape (52, 5) proba : 0.08365738 (2235.2085, 682.11597, 2591.2153, 1175.4144) proba : 0.059490494 (2321.9434, 1443.8877, 2613.5674, 1922.5825) proba : 0.043869615 (2937.0728, 682.27795, 3252.6143, 1205.03) proba : 0.042290155 (2632.6487, 1459.1227, 2953.6682, 2146.31) proba : 0.034093555 (71.958206, 699.83203, 393.3144, 1120.522) proba : 0.03188299 (308.78094, 91.360214, 716.1595, 497.18713) proba : 0.025059666 (2306.4976, 1681.2109, 2777.251, 2150.37) proba : 0.01979089 (2394.7373, 1122.3645, 2636.1602, 1534.456) proba : 0.018268555 (2293.5127, 971.5016, 2576.6206, 1378.5731) proba : 0.016394297 (2956.3313, 1594.3464, 3251.536, 2041.8997) proba : 0.01573069 (10.90258, 573.52124, 252.95627, 882.50757) proba : 0.01546782 (2710.9604, 884.6477, 3001.4067, 1337.1257) proba : 0.014101689 (2619.5166, 564.03314, 2997.3726, 1174.4978) proba : 0.013077765 (162.06506, 998.61273, 907.79175, 1740.7051) proba : 0.012829485 (2117.2534, 1467.1431, 2431.4888, 2134.755) proba : 0.012180149 (2070.6821, 153.2333, 2426.627, 695.58154) c : roue list_crops.shape (48, 5) proba : 0.08209363 (2921.2795, 638.37646, 3254.5295, 1258.6852) proba : 0.058019917 (2507.5432, 1513.9802, 2926.2498, 2083.4543) proba : 0.05729837 (2746.1458, 1692.6167, 3205.8572, 2127.3545) proba : 0.038228214 (2267.4563, 1589.3506, 2719.9377, 2144.457) proba : 0.02919623 (2521.586, 448.81555, 3069.2075, 1139.6471) proba : 0.026402874 (416.58722, 1400.4666, 896.45624, 1839.0698) proba : 0.022059439 (258.3818, 30.3945, 777.48206, 579.22766) proba : 0.018163174 (269.50745, 1082.4205, 809.20325, 1696.9222) proba : 0.016240455 (814.9617, 0.0, 1180.1952, 343.61127) proba : 0.014416294 (1708.7706, 636.1117, 2206.2307, 1093.4082) proba : 0.011404351 (2210.2778, 646.74316, 2617.3486, 1202.6804) c : toit list_crops.shape (51, 5) c : vitre list_crops.shape (47, 5) proba : 0.17969458 (2201.5254, 706.0773, 2505.8062, 1225.61) proba : 0.13708237 (2332.1968, 1420.2529, 2614.771, 1932.4644) proba : 0.0920719 (2941.3245, 639.27264, 3238.242, 1233.3057) proba : 0.047085464 (321.50052, 101.23656, 604.449, 480.0899) proba : 0.04115849 (2545.2346, 540.12836, 2972.1355, 1045.7681) proba : 0.039571848 (2362.6438, 1470.2211, 2901.422, 2152.2378) proba : 0.030477682 (2784.097, 1531.0607, 3025.3718, 2079.1077) proba : 0.017642446 (1748.734, 547.4203, 2193.0603, 1162.6349) proba : 0.015789824 (2879.6533, 1427.3745, 3199.2295, 2141.8555) proba : 0.010812808 (2403.897, 706.26624, 2720.5034, 1277.6042) proba : 0.010702644 (2683.0461, 868.76556, 2997.154, 1348.875) We are managing local photo_id image_path : temp/1748569244_1112947_950003695_22b4110c9a86b12e1542ec2bb977f6a8.jpg image_size (2160, 3840, 3) [[[111 118 91] [113 120 93] [115 120 93] ... [ 23 40 37] [ 23 40 37] [ 24 41 38]] [[111 118 91] [112 119 92] [115 120 93] ... [ 23 40 37] [ 23 40 37] [ 23 40 37]] [[113 118 91] [114 119 92] [115 120 93] ... [ 22 39 36] [ 23 40 37] [ 23 40 37]] ... [[120 125 94] [119 124 93] [118 123 92] ... [ 22 36 34] [ 22 36 34] [ 23 37 35]] [[119 124 93] [119 124 93] [118 123 92] ... [ 22 36 34] [ 22 36 34] [ 22 36 34]] [[118 123 91] [117 122 90] [117 122 91] ... [ 22 36 34] [ 22 36 34] [ 22 36 34]]] Detection took 1.171s for 300 object proposals c : aile-arriere list_crops.shape (45, 5) proba : 0.01638491 (3266.7026, 1031.8992, 3839.0, 1714.322) proba : 0.016076561 (16.38173, 491.1, 406.76904, 799.84924) proba : 0.012597205 (1997.7281, 259.50204, 2506.4885, 848.93933) proba : 0.011021228 (51.401093, 1690.1267, 479.94168, 1966.6375) c : aile-avant list_crops.shape (41, 5) c : autre list_crops.shape (46, 5) c : cache-reservoir list_crops.shape (46, 5) c : capot list_crops.shape (38, 5) c : carrosserie-autre list_crops.shape (44, 5) c : coffre list_crops.shape (31, 5) c : essuie-glace list_crops.shape (49, 5) c : feu-antibrouillard list_crops.shape (44, 5) proba : 0.014232134 (3282.567, 1219.7703, 3793.5493, 1821.4849) proba : 0.012605798 (27.94963, 498.8993, 410.0074, 789.6398) c : feu-arriere list_crops.shape (44, 5) proba : 0.07466318 (7.595749, 454.29453, 376.21747, 791.5864) proba : 0.025204122 (3284.6729, 1151.9507, 3782.7202, 1834.2451) proba : 0.011054892 (17.656479, 1586.6859, 282.00952, 1995.753) c : info-modele list_crops.shape (44, 5) proba : 0.018715309 (35.900284, 482.71948, 414.19495, 791.46704) c : logo-marque list_crops.shape (46, 5) proba : 0.017158717 (41.253128, 486.37634, 409.7342, 792.7191) c : logo-roue list_crops.shape (45, 5) c : pare-brise list_crops.shape (38, 5) proba : 0.016601741 (24.73471, 0.0, 388.85196, 684.50256) c : pare-choc list_crops.shape (35, 5) c : phare list_crops.shape (48, 5) c : plaque-immatriculation list_crops.shape (47, 5) proba : 0.025557566 (26.288193, 479.65518, 404.91602, 771.0746) proba : 0.01656386 (28.250778, 1620.7365, 299.3711, 1974.6) proba : 0.014695564 (18.401062, 3.5820465, 386.7264, 255.02422) c : poignee list_crops.shape (43, 5) c : porte list_crops.shape (40, 5) proba : 0.11927567 (1872.8767, 10.286285, 2442.9949, 862.1929) proba : 0.11464668 (2136.5327, 52.91455, 2855.9678, 815.3088) proba : 0.024782706 (3234.8447, 69.61813, 3823.1128, 847.86084) proba : 0.013809913 (109.564865, 1835.4814, 459.84363, 2159.0) proba : 0.01138681 (3315.0093, 1115.6017, 3779.0215, 1883.9008) proba : 0.011383201 (1348.9734, 1049.3528, 1927.9495, 1777.0039) proba : 0.0107546 (2525.6648, 168.78912, 3531.4797, 924.1529) c : pot-echappement list_crops.shape (43, 5) proba : 0.033426173 (5.1139374, 1746.7147, 343.4391, 2019.2792) proba : 0.0106583275 (117.19896, 1862.9652, 459.03418, 2155.1729) c : radiateur list_crops.shape (43, 5) c : retroviseur list_crops.shape (47, 5) proba : 0.018039485 (3296.5825, 1199.9968, 3788.0024, 1822.6375) proba : 0.013667412 (13.038696, 1748.9717, 340.37943, 2015.7627) proba : 0.012989895 (124.44766, 1867.7485, 456.31857, 2150.5208) c : roue list_crops.shape (45, 5) proba : 0.58398265 (3132.7415, 1107.543, 3839.0, 1925.343) proba : 0.045884173 (3481.315, 1409.904, 3814.2905, 1997.3789) proba : 0.037301708 (38.59784, 1751.9702, 339.5256, 2013.6733) proba : 0.029187175 (3244.3767, 40.531647, 3721.5984, 739.198) proba : 0.018179495 (3167.0742, 383.71725, 3484.2456, 1016.3723) proba : 0.012429726 (229.06554, 0.0, 685.78656, 557.6493) proba : 0.010356307 (2689.1248, 224.95532, 3238.7478, 934.29224) c : toit list_crops.shape (48, 5) c : vitre list_crops.shape (44, 5) proba : 0.018343285 (23.037186, 0.0, 379.90704, 299.26855) proba : 0.014216392 (3321.227, 1194.3063, 3770.21, 1805.963) We are managing local photo_id image_path : temp/1748569244_1112947_926687666_a8bc8c1fad77748c62ca641ceb29ad9c.jpg image_size (480, 640, 3) [[[36 41 44] [36 41 44] [35 40 43] ... [ 8 10 10] [ 8 10 10] [ 8 10 10]] [[37 42 45] [36 41 44] [35 40 43] ... [ 5 7 7] [ 5 7 7] [ 5 7 7]] [[37 42 45] [36 41 44] [35 40 43] ... [ 3 5 5] [ 4 6 6] [ 4 6 6]] ... [[42 47 50] [41 46 49] [40 45 48] ... [ 8 10 10] [ 8 10 10] [ 8 10 10]] [[41 46 49] [41 46 49] [40 45 48] ... [ 0 2 2] [10 12 12] [22 24 24]] [[40 45 48] [40 45 48] [40 45 48] ... [10 12 12] [17 19 19] [26 28 28]]] Detection took 0.044s for 300 object proposals c : aile-arriere list_crops.shape (32, 5) proba : 0.12727642 (160.8552, 172.85149, 306.19534, 321.39288) proba : 0.01153998 (538.5317, 190.82932, 613.9661, 298.88852) proba : 0.010253043 (197.33284, 160.36844, 490.47534, 311.2734) c : aile-avant list_crops.shape (40, 5) proba : 0.9481688 (161.78468, 149.54091, 330.64685, 343.4181) proba : 0.09027749 (20.761093, 105.88254, 54.775368, 195.62827) proba : 0.02217484 (152.80583, 143.27791, 217.16954, 306.25964) proba : 0.01375517 (320.70038, 330.9067, 439.328, 414.44528) proba : 0.011396097 (557.7659, 199.67122, 608.66583, 270.50854) c : autre list_crops.shape (36, 5) proba : 0.011426695 (458.26996, 15.298569, 521.144, 98.99303) c : cache-reservoir list_crops.shape (37, 5) proba : 0.029750206 (470.56717, 21.584175, 531.4649, 89.20078) proba : 0.025544558 (368.27335, 265.85626, 446.89816, 331.26587) proba : 0.014927192 (353.33075, 360.4899, 412.48865, 423.13107) proba : 0.014713106 (451.11234, 57.614006, 523.15857, 125.93281) c : capot list_crops.shape (33, 5) proba : 0.9944273 (211.27557, 115.1434, 555.7746, 300.74063) proba : 0.23368108 (65.79129, 24.148893, 285.73267, 46.316933) proba : 0.034851067 (343.23456, 279.91388, 508.2165, 431.91486) proba : 0.017473526 (89.42491, 37.517334, 357.51343, 63.62861) proba : 0.010149155 (424.37653, 246.66765, 564.62024, 357.00552) c : carrosserie-autre list_crops.shape (35, 5) proba : 0.028548518 (454.5429, 25.08237, 523.52234, 121.411545) c : coffre list_crops.shape (31, 5) proba : 0.12966235 (446.42865, 19.30341, 528.3393, 117.90968) proba : 0.10255872 (139.8904, 45.98333, 419.5232, 192.66495) proba : 0.030857768 (269.2946, 162.00217, 553.9408, 348.94202) c : essuie-glace list_crops.shape (40, 5) proba : 0.789503 (203.86551, 108.23105, 416.7576, 147.35785) proba : 0.14472471 (369.06165, 267.52997, 445.0334, 332.95746) proba : 0.09797766 (391.05792, 261.1877, 527.0127, 318.13113) proba : 0.07720719 (470.47858, 22.210804, 532.0738, 88.0226) proba : 0.035417486 (354.38446, 360.35986, 414.20453, 422.97754) proba : 0.01414922 (114.834564, 22.671364, 335.36716, 42.910583) proba : 0.0122261755 (254.58006, 94.724976, 456.8227, 288.56046) proba : 0.01093061 (353.83667, 285.56967, 484.79468, 414.69156) c : feu-antibrouillard list_crops.shape (39, 5) proba : 0.21193348 (341.7836, 363.48325, 404.85013, 430.6583) proba : 0.05587425 (368.6619, 268.62704, 446.44247, 330.0911) proba : 0.025236126 (470.42923, 21.90839, 531.39514, 89.25714) proba : 0.023563012 (389.44647, 264.04642, 529.788, 316.96017) proba : 0.01647149 (451.48853, 58.843243, 522.91144, 125.208786) proba : 0.010561288 (467.00406, 343.5473, 573.7648, 420.203) c : feu-arriere list_crops.shape (39, 5) proba : 0.10164161 (329.67056, 259.26117, 456.71976, 315.55115) proba : 0.042246886 (454.30267, 28.763565, 522.9127, 104.09871) proba : 0.025247272 (333.5108, 352.84415, 405.0608, 425.97427) proba : 0.010688223 (475.56393, 341.53702, 575.4437, 420.15268) c : info-modele list_crops.shape (37, 5) proba : 0.0704632 (470.48972, 21.814632, 532.0423, 89.57272) proba : 0.038774293 (353.82794, 361.5416, 412.9959, 422.8533) proba : 0.03335563 (368.70746, 266.81305, 448.2409, 331.6281) proba : 0.023050936 (450.8426, 57.474094, 524.0709, 126.34169) proba : 0.015797336 (390.14044, 260.8508, 533.16907, 317.53494) c : logo-marque list_crops.shape (39, 5) proba : 0.11936622 (352.96826, 361.31464, 412.53125, 422.70123) proba : 0.032974806 (470.55682, 20.779263, 530.9589, 88.80194) proba : 0.027516948 (370.81317, 264.58362, 447.20575, 329.64008) proba : 0.025024619 (397.47058, 260.8251, 526.3081, 316.55185) proba : 0.01432975 (584.1553, 0.74874496, 639.0, 70.73624) c : logo-roue list_crops.shape (37, 5) proba : 0.014608637 (353.3217, 360.72946, 413.24722, 423.60703) proba : 0.013224714 (470.34485, 21.601093, 532.065, 89.58136) c : pare-brise list_crops.shape (34, 5) proba : 0.9569799 (141.0971, 42.466076, 444.09308, 147.77231) proba : 0.105130985 (319.4548, 22.014011, 424.0319, 129.1307) proba : 0.06655231 (453.85498, 19.660057, 523.2705, 91.545456) proba : 0.050993863 (287.21542, 164.05997, 547.31995, 296.73648) proba : 0.023291383 (95.59754, 40.147087, 158.49188, 162.53784) c : pare-choc list_crops.shape (29, 5) proba : 0.9453821 (272.89435, 257.50577, 580.2391, 444.4462) proba : 0.2197559 (233.49008, 224.18582, 397.7679, 411.54916) proba : 0.028820626 (487.97293, 17.564262, 613.4275, 120.57682) proba : 0.028741293 (435.9593, 307.78696, 588.67584, 426.57657) c : phare list_crops.shape (38, 5) proba : 0.77762496 (326.83984, 251.89102, 477.7768, 312.86313) proba : 0.076984026 (328.2209, 359.63678, 410.24304, 426.96674) proba : 0.038830616 (292.80902, 225.13937, 466.7425, 392.34268) proba : 0.033105355 (538.2378, 197.21037, 600.2174, 303.44336) proba : 0.026036425 (466.65118, 20.84036, 531.7361, 84.11878) proba : 0.018858818 (305.61005, 209.47272, 597.71405, 307.11682) proba : 0.014257204 (478.92026, 344.12628, 578.6293, 416.75385) proba : 0.011292008 (96.66905, 66.98535, 163.911, 147.69135) c : plaque-immatriculation list_crops.shape (38, 5) proba : 0.23006286 (518.5571, 294.1658, 582.3375, 390.75748) proba : 0.03376684 (438.44525, 271.75998, 568.41895, 386.56113) proba : 0.025865281 (347.18628, 259.7807, 452.88037, 315.94507) proba : 0.021144493 (470.38928, 23.621984, 531.611, 85.37917) c : poignee list_crops.shape (35, 5) proba : 0.054377556 (583.3567, 0.030948639, 639.0, 70.98572) proba : 0.034034412 (470.454, 21.598267, 531.67426, 89.77441) proba : 0.026380457 (353.17023, 360.95868, 412.89447, 423.6001) proba : 0.024487995 (368.0542, 266.1219, 448.05353, 332.26685) c : porte list_crops.shape (28, 5) proba : 0.9854092 (78.16664, 43.078415, 169.65276, 306.44586) proba : 0.9672486 (33.64362, 44.17742, 90.64882, 241.29745) proba : 0.091565005 (452.36984, 19.606167, 521.08856, 113.535324) proba : 0.03417715 (158.70538, 46.139755, 421.3357, 212.15456) proba : 0.012429091 (436.9688, 223.37668, 575.05945, 403.3306) c : pot-echappement list_crops.shape (37, 5) proba : 0.055537377 (353.8206, 360.7027, 412.23123, 423.3391) proba : 0.018789351 (458.05484, 14.546093, 521.2203, 99.670425) proba : 0.012938654 (584.3939, 0.0, 639.0, 71.1006) proba : 0.011320591 (466.14304, 340.33652, 575.4017, 420.8793) c : radiateur list_crops.shape (38, 5) c : retroviseur list_crops.shape (39, 5) proba : 0.23813507 (452.17776, 56.80104, 522.1677, 124.12688) proba : 0.20780696 (471.15387, 21.540287, 531.2205, 89.2724) proba : 0.19334657 (369.90195, 266.01996, 447.44455, 332.586) proba : 0.04561949 (342.49402, 362.45786, 404.12836, 431.03366) proba : 0.026582334 (176.92133, 115.94922, 409.85126, 150.13248) proba : 0.021461492 (584.3822, 0.0, 639.0, 70.68931) proba : 0.013623883 (392.94077, 259.88467, 530.6085, 318.6445) proba : 0.010497375 (101.943436, 73.78637, 164.67805, 158.12628) c : roue list_crops.shape (38, 5) proba : 0.94177043 (187.7864, 271.24283, 306.51013, 426.33023) proba : 0.13538104 (334.873, 358.32187, 400.07733, 428.79782) proba : 0.049787898 (538.71277, 250.9963, 599.99365, 395.28146) proba : 0.037391514 (12.769609, 136.75076, 64.15765, 244.90955) proba : 0.034819603 (549.3066, 192.3883, 605.7956, 305.7133) proba : 0.015744282 (479.6858, 335.9129, 572.48254, 421.26593) c : toit list_crops.shape (34, 5) proba : 0.7919578 (83.63141, 31.277332, 341.39157, 55.135128) c : vitre list_crops.shape (37, 5) proba : 0.980471 WARNING: Logging before InitGoogleLogging() is written to STDERR F0530 03:40:49.543936 1112947 math_functions.cu:79] Check failed: error == cudaSuccess (700 vs. 0) an illegal memory access was encountered *** Check failure stack trace: *** Command terminated by signal 6 86.60user 46.09system 5:25.34elapsed 40%CPU (0avgtext+0avgdata 6355164maxresident)k 6073320inputs+44144outputs (6420major+6490020minor)pagefaults 0swaps