MedTech Case Study
Pattern recognition algorithm testing on genomic data
4 years in MedTech among 2 different projects. Although different, they share common issues, challenges, testing tech stack.

What we did was introducing MLOps processes and ML features functional and non-functional testing
Pattern recognition and patient/doctor applications
Challenges
Industry-specific challenges in MedTech projects related to QA, Data and MLOps
  • Sensitive data and privacy
    • Obfuscation
    • Deliberate change of specific parts of the sensitive data for testing and development
    • Manual process of data collection setup with legal, scientific team and data engineering team
  • Big Data
    Human genome and other medical information can hold tens of gigabytes of text in a single file which cannot be reduced through the whole development cycle
    • Cloud infrastructure
    • Compressing medicine-specific formats like FASTQ
    • Reducing amount of data to train/test the tests to speed up development of the pipelines
    • Chained interdependent pipelines including data preparation and processing stages
  • Closed isolated environments
    • Emulation on small scale
    • Partially limited End-To-End validation on test environment
    • Inability to control test data on legally unavailable services
    • Testing using proxy and on gateway level
  • Testing in production
    • Environment limitations result into necessity to run part of end-to-end regression testing in production environment
    • QA engineers proving their own personal data to test specific feature on production for them selves
  • ML pipelines run 30+ hours
    • Setup MLOps practices
    • Test each chain segment of chained of the pipeline independently
    • Realtime reporting and logging of the pipeline before run has finished
    • Long release cycles (monthly, in some cases - yearly)
  • Long design and analytical phases cause constant urgency in development
    • Prepare test data and test pipelines on early stages
    • Development and QA involved in analytics phase to reduce communication ping-pong in development
    • Feature toggle on prod for partial releases
ML Testing
Automation of test regression
Value of Automated regression
So far all the work has been done by ML engineers and compliance officers. Why do they need regular software engineers

On each new iteration of algorithm we dont know how truth data set should look like. What we can know is how old truth looked like.
Since we can't test new functionality manually - only ML engineers can. And they also can't btw. They only can train their model and hope it would perform well in production. So the value is: to protect us from regression bugs in our ML algorithms.
AI effect
Modern AI agents can be stoped with bureaucracy, compliance and other regulation requirements. Especially in such sensitive areas as diagnostics. Other than that boost in pipeline test development in support was significant.

> 50%
what have we learned
Output
We have made great progress verifying ML algorithms through 2 years on the project. Here are some bullet points and diagrams on that.
Made on
Tilda