SageMaker's project setup process isn't straightforward. Isaac found it tedious and frustrating. He also had to set up other AWS services like S3 for uploads and figure out the right AWS permissions – this was painful.
In comparison, BOLT's interface is intuitive and easy to use. With BOLT, we were able to set up the same project within just a few minutes. We designed and built BOLT for simplicity so our users can focus on what they need: annotated data.
Based on our experience, SageMaker doesn't provide feedback between you and the annotation process. Once SageMaker Ground Truth completes your task annotation tasks, the only action you're allowed to take is to export the data. It's up to you to analyze the quality of your tasks. The platform doesn't provide insights or information that will help you accelerate the learning and iteration process.
Insights and learnings are a collaborative process on BOLT. The feedback between our users and annotators helps to progressively improve annotation quality and reduce label noise with every iteration.
BOLT is designed specifically to improve the efficiency of iterating based on insights. Our annotators don’t just annotate, they add value to your data by highlighting issues such as taxonomy conflicts and edge cases within your data. When they find an issue with a task, they will also put the task on hold and provide feedback to avoid inconsistent data quality.
BOLT provides you with actionable insights from our professional annotators. We help you surface discrepancies and anomalies in your data so each batch on BOLT just gets better.
In SageMaker, there aren't any tools for analyzing and evaluating annotations. You have to export your data to a separate platform in order to do this. Understanding your data is tedious and painful. It would be good to pair SageMaker with a platform like Voxel51 to helps you visualize and explore your annotated data.
BOLT provides a better data exploration system. Use our built-in task viewer to interactively sort, filter and view your data. Tag any task or collection of tasks at any time to note interesting observations for note taking, communicating with your team, or for improving your quality later.
When you want to perform detailed data explorations, just create sample datasets by applying tags or filters on attributes and metadata. BOLT can also create sample datasets for you by randomly extracting tasks using our smart sampling algorithm. If you would like to have a quantitative understanding of the quality for a dataset, perform QE using our interface and obtain a detailed breakdown of your dataset accuracy.
Pricing on SageMaker is confusing. AWS SageMaker Ground Truth charges $0.08 per image to use their workflow. While this might seem trivial, this $0.08 charge is often more than the labor cost for easier tasks like bounding box annotations or image classification. Your overall costs will quickly double and add up when handling tens of thousands of images.
On top of that, there is a bit of variance on the price of a HIT. You must experiment to find the optimal amount to pay workers.
Compared to SageMaker, BOLT’s annotation costs are transparent. Using our platform is free and our annotation fees are fixed – pay as you go, with no minimum spend or commitment.
SageMaker's work is performed by untrained crowdworkers on MTurk. Their platform relies heavily on getting quality through consensus – i.e. assigning the same tasks to multiple workers. When we tried this out, the results were inconsistent.
As we continued to experiment by increasing the number of workers on the same task, the results were better. We had to increase the number of annotators for each task to 4 before getting any usable data.
Here’s how much we paid
The common misconception of MTurk is that you can label large datasets for pennies is most definitely false. The same project above would cost USD$2.88 on our platform, assuming clear instructions on how to annotate the work.
BOLT pricing → 4 annotations per task x 4 tasks x $0.18 USD per polygon = $2.88 🚀