Description of necessary data at project start.
Supported file formats are text PDF's, scanned PDF's, Word, plain text, RTF, PNG, jpeg, tiff ... Other file formats can be added on request.
For scans, the DPI should be at least 150 to get good quality OCR. Lower DPI will have an impact on the accuracy of the OCR.
Smartphone photos of documents are supported but can reduce the accuracy of the OCR if they are very bad. It is sometimes useful to preprocess them using a smartphone app like CamScanner.
For document classification, you can add a list of filenames with the correct document class or type.
For entity extraction, it is best to use Metamaze for labelling the data.
Document type prediction is usually a very accurate step in the process. Still, the amount of labelled data needed depends on the amount of variation in your data.
For classifying a document type based on text (for example a pay slip, invoice, loan agreement, ...) you typically need at least 20 examples per document type.
If the document types are very close together, more will be needed for the model to learn how to distinguish them. So for example if you want to discriminate between 2nd hand car purchase orders vs new car purchase orders, more data might be needed.
For tasks like sentiment analysis, priority estimation, ... that have a wide variety of input cases, a custom data requirements exercise is needed depending on the output classes. These can quickly need at least >100 documents per class.
When you need to classify a document based on visual content instead of text, please contact an Metamaze ML Engineer.
Data requirements depend on the problem you are trying to solve. The Machine Learning models learn from context, so the more variety there is in context in production, the more data you need to annotate. The other way around, the more relevant context you have, the easier for the algorithm.
It is best to upload training data that is as similar to production data as possible. So if you want to build a production model that works for only 5 suppliers in one language, then only upload data from those 5 suppliers in that language. If you want to build a production model that needs to work for any supplier (e.g. thousands of different ones) in any language, it is best to upload the highest diversity: different suppliers, different languages, ...
The following properties make it harder for the model, so increase the amount of data needed:
- Every document is unique: no recurring templates in documents
- Bad quality scans will not be used as training data, so if the source data contains a lot of them you need more source data to retain an equal amount of good quality scans.
- Lots of interpretation needed due to subtle differences
- Little context to learn from
- No standardisation of terminology
A couple of examples