NetOwl Extractor
NetOwl Extractor provides highly accurate, fast and scalable entity extraction in multiple language using AI-based natural languages processing and machine learning technologies. NetOwl's named-entity recognition software is available on premises and in the cloud. This allows for a variety Big Data Text Analytics applications. NetOwl's named entity recognition software can be deployed on premises or in the cloud. It supports entity extraction from over 100 entities. It includes people, different types of organizations (e.g. companies, governments), various types of places (e.g. countries, cities), addresses and artifacts as well as phone numbers, titles and titles. This vast named entity recognition (NER), forms the foundation for advanced relationship extraction and event extraction. Domains include Finance, Politics and Homeland Security, Law Enforcement, Military, National Security and Social Media.
Learn more
Docparser
Docparser extracts data from Word, PDF and image-based documents. It uses Zonal OCR technology, advanced patterns recognition and anchor keywords. To set up your document parser, there are three steps. Upload your document directly, connect with cloud storage (Dropbox. Box. Google Drive. OneDrive), email your files in attachments, or use the REST API. Docparser can extract the data you need without any programming. Use the options that best suit your document type to select preset rules that are specific to your PDF and image documents. You can either download directly to Excel, CSV or JSON formats or connect Docparser with thousands of cloud applications such as Zapier and Workato. You can choose from a variety of Docparser templates or create your own custom document rule. You can extract important invoice data and then integrate it into your accounting system. Data such as line items, dates, totals, and reference numbers can be pulled.
Learn more
ScrapeStorm
ScrapeStorm, an AI-powered visual web scraping software, is available. No manual operation required for intelligent data identification. ScrapeStorm uses artificial intelligence algorithms to identify List Data, Tabular Data, and Pagination Buttons. No need to set any rules. Just enter the URLs. Automatically identify forms, lists, links, images and prices. You can simply click on the webpage following the prompts. This is exactly the same as manually browsing the webpage. It can generate complex scraping guidelines in just a few steps. Any webpage data can be easily scouted. Enter text, click, move mouse and drop-down box. Scroll page, wait for loading, then loop operation. Finally, evaluate conditions. The scraped data may be exported to a local or cloud file. You can use the following support types: Excel, CSV and TXT; HTML, MySQL, MongoDB; SQL Server; PostgreSQL; WordPress; and Google Sheets.
Learn more
PrecisionOCR
PrecisionOCR is an easy-to-use, secure and HIPAA-compliant cloud-based optical character recognition (OCR) platform that organizations and providers can user to extract medical meaning from unstructured health care documents.
Our OCR tooling leverages machine learning (ML) and natural language processing (NLP) to power semi-automatic and automated transformations of source material, such as pdfs and images, into structured data records. These records integrate seamlessly with EMR data using the HL7s FHIR standards to make the data searchable and centralized alongside other patient health information.
Our health OCR technology can be accessed directly in a simple web-UI or the tooling can be used via integrations with API and CLI support on our open healthcare platform.
We partner directly with PrecisionOCR customers to build and maintain custom OCR report extractors, which intelligently look for the most critical health data points in your health documents to cut through the noise that comes with pages of health information.
PrecisionOCR is also the only self-service capable health OCR tool, allowing teams to easily test the technology for their task workflows.
Learn more