Challenge Overview

Challenge Overview

  • The goal of this challenge to build a multiclass image classifier.

  • The classifier should return precisely three probabilities: 

    • Probability that ‘old  device’ is in the image

    • Probability that ‘new device’ is in the image

    • Probability that neither new device nor old device is present in the image.

  • The returned probabilities should be a float in the range [0, 1].

  • The provided training data contains a mixture of images and videos. Participants can extract images/frames from the videos and use it for training.

  • The amount of training data is relatively not that large, so participants are encouraged to use data augmentation techniques if required, and possibly generate similar images with altered lighting, brightness, skewness, rotation etc to ensure that your trained model generalizes well.

  • The participants are free to use any technique of their choice, be it machine learning, deep learning or even basic rule-based computer vision techniques. 

  • The path of training images, testing images and configurations should not be hard coded into the code. They should be either passed via the command line arguments, or should be configurable via a dedicated JSON based config.json file.

  • The tool should be a command line tool, with separate commands to run the train, validate and test steps. 

  • The Code should be implemented in Python 3.7 (or 3.6 in case some important library is not available in Python 3.7)

  • In addition to the usual Tensorflow, Pytorch or other kinds of model, the submission should also include the trained model in the ONNX format. Some libraries like Pytorch support native rendering to ONNX, while others might require converters like the following:

 

Data access & description

  • The training dataset can be found in the Forum. 

  • The top two winning submissions of the last challenge can be found in the forum. It is strongly recommended that members go through these submissions and ensure that the best practices of these submissions are incorporated in submissions to this challenge. Participants can also choose to use any one of the code as a starting point for their work for this challenge.

Output Format

The output generated by the algorithm should be a CSV file containing the filenames and its corresponding probabilities of all the image files in the target test folder. The csv file should be preferably named 'output.csv', and should have headers: 

filename, prob_old_device, prob_new_device,prob_no_device
 

An Important Note on Preprocessing and output layer

After this challenge, we will take the ONNX model created as a result of this challenge, convert it into an Apple CoreML model, and integrate it into an iPhone Native App.

So it is important that whatever pre-processing that needs to be done on a test image should be very clearly documented, such that after reading it, an iOS developer can easily implemented those pre-processing in the iPhone app. So in addition to the documentation, please include a document named ‘preprocessing_output_details’ detailing exactly the operations used in pre-processing, such as normalization, color scale changes, image resizing etc. Also, the line numbers of the pre-processing steps should be included in the document, so that in case of a confusion, the iOS developer can quickly refer to this documentation.

In addition to preprocessing, add details about the output layer that is used in the neural network model. This is required because the ONNX model at times does not contain the final layer, and it needs to be implemented manually. The details of the output layer should also be added to the file preprocessing_output_details.

Note - Ideally, if changing it is not required, the pre-processing used in the 1st placed submission of the last challenge can be maintained without any changes. This is because we already have a tested implementation of the last challenge winner’s pre-processing. 

Although it should be noted that this is not a strict requirement and if essential, the participants are free to change the pre-processing if necessary.

 

Expected Technology

Participants are free to use whatever library and tools they find useful, such as Tensorflow, PyTorch etc, and also include the extra ONNX format model as discussed above.

 

Submission Evaluation

  • Important - The performance will be evaluated using a custom function built upon the binary classification metric AUC ROC SCORE. We will use a custom metric to find binary classification performance of old_device vs new_device, old_device vs no device and new_device vs no device and combine them by taking a weighted mean of these AUC ROC scores. The weights here are not shared for now to ensure that the performance is not emphasised in any one of the binary classifications.

  • The submission will be evaluated subjectively, but broadly the rating will be determined by the following:

    • 70-80% - model performance in terms of ROC AUC score. 

    • 10-20% - Code quality and issues related to the code

    • 10-20% - Documentation and Ease of deployment, following solely the deployment guide.

 



Final Submission Guidelines

What to Submit

  • The code

  • A PDF/Word/Markup format based report detailing the techniques and algorithms used to achieve the goals.

  • A PDF/Word/Markup/Text file named ‘preprocessing_output_details’ detailing the preprocessing steps and the output layer as discussed in the section ‘An Important Note on Preprocessing and output layer’ above.

  • A README.md file detailing the deployment instructions, along with example and steps to verify the various capabilities of the Code.

  • Optional - screenshots or video to make it easier to quickly verify and test your code.
     

Technology Stack

  • Python

REVIEW STYLE:

Final Review:

Community Review Board

Approval:

User Sign-Off

SHARE:

ID: 30100196