Create Jira Ticket
This idea cannot be sent to Jira.
COVID Data Scraping Scheme
Set up a Hipaa compliant COVID testing web site that has Data Scrape functionality built in.
The plan
1. Create a simple html website with COVID Data Scraping Scheme owned by Public Health entity.
2. Anyone with a local computer device with patient/worker/school database, logs on securely to the Public Health COVID Data Scraper. Once there, they click to allow website Scraper to connect and parse the desired/required specific COVID reporting data elements.
3. Must have HTTPS encryption to protect privacy and security.
4. Parsed data is then further standardized (ie field length, language) for public health and data analysis purposes.
To further explain, it's like a computer "virus" that instead of blocking it, you allow the "data scrape program" in to pull out data for the public health.
Goal is to -
Easy for anyone to login and share their data.
Eliminate need to manually enter information.
Avoid data entry errors
Eliminate need for incoming scraped computer data to be standardized. One of the most difficult steps in data transfer is to create matched data fields to ensure transfer.
Standardize data elements for analysis by public health. By creating a tool to standardize incoming data, any computer or device in whatever configuration can provide useable data. The basic idea of web scraping is that we are taking existing HTML data, using a scraper to identify the data, and convert it into a useful format.
Full disclosure - I am not a Programmer. I have been working on smooth medical data transfer problems for most of my thirty year career in a large reference lab, and more recently in smaller molecular diagnostic labs. I've worked on large scale national database projects, including public health reporting. I'd like to work with a team of programmers interested in this concept if it seems at all feasible.
This is an interesting idea!
Who do you imagine would use the scraper? And can you give an example or two of the kind of website it would scrape?
Yeah, this is a great idea. Thank you for sharing it and validating interest.
We have some functionality similar to this in the SANER Relay network, in so far as software stack it’s based on can handle a 2 page layout in desktop mode, and we sometimes iFrame external sites in the second page. Use the first page as the encoded measure report, and the second page (iFramed) lets the user select the data elements from the page, and then map them to the report one item at a time. That would be workable. And it’s sort of similar to the ETL extraction project we did a few years ago.
Yeah, we could commit to building out this functionality. Maybe even put together a screen rendering later this week.
Trying to understand how your approach would address the following (In general) 1. How to collect AOEs, from ordering provider/patient/specimen collector. Seems to assume this is stored on patient computer or provider computer? Is that correct? If provider, then how will scraper know which patient/data needs to be scraped to ensure correct data is with correct patient?
2. Integrate into app/LIS or other information source for patient to be married to results of IVD test device/system (either lab performed or patient performed at home like pregnancy test) How would results and AOEs scraped be accurately LOINC and SNOMED CT coded to meet ELR & HHS encoding requirements?
3. All transmitted to public health (ELR)
4. All transmitted to HHS (may be met by 3).
How is this web-scraper going to work when data might be inside a database, Excel files etc..? How is it going to handle incremental data loads over weekdays?
Perhaps this website concept can be used for Hospitals/clinics to upload schema-compliant data files for a batch-load process (triggers when you upload your files).
One of the most difficult steps in data transfer is to create matched data fields to ensure transfer.
Area Rug Cleaning