Active provides value to you in a number of use cases. This page lists a few. If you have other use cases you would like to explore with Active, contact us.

Data Cleansing/Curation

Alex, a DataOps manager at self-dr-AI-ving, faces challenges in managing and curating data for self-driving cars. Alex’s team struggles with scattered data, overwhelming amounts of data, unclear workflows, and an inefficient data curation processes. Alex is currently a big user of Encord Annotate, but would like to provide better datasets for annotation.
  1. Initial setup: Alex gathers a large number of images and gets them imported into Active. Alex then logs into Encord and navigates to Active (freemium).
  2. First collection: Alex opens the Project and after searching, sorting, and filtering the data, selects the images and clicks Add to a Collection and then clicks New Collection. Alex names the Collection RoadSigns as the Collection is designed for annotating road signs for the team.
  3. Data curation: Alex then further bulk-finds traffic sign images using the embeddings and similarity search. Alex then clicks Add to a Collection and then clicks Existing Collection and adds them images to the RoadSigns Collection in a matter of clicks.
  4. Labeling workflow: Thinking about different use-cases (for example, “Labeling” and “Data quality”) Alex assigns good quality road signs for labeling, and bad quality road signs for “Data quality” and future “Deletion”. In the future Alex might use “Active learning” to prioritize the data for labeling.
  5. Sent to Annotate: Alex goes to the Collections page, selects the Roadsigns Collection and clicks Create Dataset. Active creates the dataset and a new Project in Annotate. Alex then configures the workflow, annotators, and reviewers for the Project in Annotate.
  6. Review and insights: At the end of the week, Alex reviews the RoadSigns Project in Annotate. The dataset has been annotated. Alex goes to Active, clicks the More button on the Project then clicks Sync Project Data. Alex then clicks the Project, clicks Analytics and then Model Predictions where Alex gains insights into:
    • Number of labels per image
    • Quality of annotations
    • Distribution of annotations across metric
The process is seamless and fast, and Alex can focus on more strategic tasks while her team enjoys a much-improved, streamlined data curation workflow.

Label Correction/Validation

Chris, a Machine Learning Engineer at a micro-mobility startup, has been working with Encord Annotate. His team is dealing with a large set of scooter images that need accurate labeling for obstacle detection. After an initial round of annotations, Chris notices that some labels are incorrect or ambiguous. This has a significant impact on the performance of the ML model.
  1. Access Collections: Chris logs into Encord Active. Chris opens the Scooters project that was imported from Annotate. Chris goes to the Collections page for the project. Chris browses the existing Collections to see if he should create a new Collection or add to an existing Collection.
  2. Data exploration: Chris searches, sorts, and filters the previously annotated scooter images and identifies those that need re-labelling.
  3. Create re-labeling Collection: Chris selects the images and clicks Add to a Collection. Chris then clicks New Collection and names the Collection Re-label - Scooters.
  4. Initiate re-labelling: With the Collection ready, Chris returns to the Collections page, selects the Re-label - Scooters Collections and clicks Create Dataset. The Collection is sent to Annotate.
  5. Assigning in Annotate: In Annotate, Chris assigns re-labelling tasks to specific annotators. Annotators then complete their relabelling tasks.
  6. Quality check in Active: After the re-labeling tasks are completed, Chris clicks Sync Project Data. The updated labels then sync back to the Project. Chris reviews the changes, confirms the label quality, and plans for model re-training.
The “Collections” feature has simplified the task of identifying and re-labeling inaccurate or ambiguous data, streamlining the entire data annotation and quality control process for Chris and his team.

Model/Prediction Evaluation

Dana is a Machine Learning Engineer at a robotics company. Her team wants to improve the performance of their machine learning model used by their robots. She has ground truth labels created by her team and predictions from her model.
  1. Import ground truth labels and predictions: Dana uses Encord’s SDK to import ground truth labels and model predictions into an Encord Annotate Project. After the import, Dana imports the Annotate Project into Active.
  2. Prepare for Evaluation: Dana “imports” the predictions for Active. Active then performs comparison and optimization calculations between predictions and ground truth labels.
  3. Assess Model Evaluation Analytics: Dana navigates to the Analytics View page and creates several charts to compare her ground truth labels against the model predictions. Dana also navigates to the Model Evaluation page and also compares model performance against the ground truth labels. Dana is able to identify a number of areas where her model’s performance can improve. For example, using the Model Evaluation charts, Dana discovers that her model has poor performance on dark and low contrast images and videos.
  4. Curate, Relabel, and Retrain: Dana imports a number of images and videos with dark and low contrast characteristics to Index. She creates a Collection and then a Dataset of these images and videos. Dana then adds the Dataset to her Annotate Project and her team creates new ground truth labels to train the model. Dana exports the labels and trains her model using the new labels.
  5. Import and Evaluate:
Dana repeats the import and evaluation steps with the latest predictions.
  1. Reassess Model Performance: Dana navigates to the Model Evaluation page and creates several charts to compare her ground truth labels against the latest model predictions and the previous model performance. Dana can then continue the cycle until her model reaches the performance she needs.