Updating Datasets
Updating Deep Lake datasets
How to make updates to Deep Lake datasets
This tutorial is also available as a Colab Notebook
Create a Representative Deep Lake Dataset
import deeplake
import pandas as pd
import os
from PIL import Image
images_directory = '/damaged_cars_tutorial' # Path to the COCO images directory
annotation_file = '/damaged_cars_tutorial/COCO_mul_val_annos.json' # Path to the COCO annotations file
deeplake_path = '/damaged_cars_dataset' # Path to the Deep Lake dataset
ds = deeplake.ingest_coco(images_directory, annotation_file, deeplake_path,
key_to_tensor_mapping={'category_id': 'labels'}, # Rename category_id to labels
ignore_keys=['area', 'image_id', 'id', 'segmentation', 'image_id', 'bbox', 'iscrowd'])Add Data to a New Tensor
1. Iterate through the Deep Lake samples and append data
Iterate through the supplemental data and add data at the corresponding Deep Lake index
Update Existing Rows without TQL

Delete Samples
PreviousTraining an Object Detection and Segmentation Model in PyTorchNextData Processing Using Parallel Computing
Was this helpful?