verified data inside a permissionless federated learning infrastructure
Imagine you're an insurance company and want to train predictive models on sensitive healthcare data. You approach a hospital for access to this data, but they are unwilling to disclose patient information. Instead, they offer to train the models themselves without revealing the data. However, how can you trust that the data used for training is real and that the training process is valid? We found a way to do this!
The solution: TEEs + data ZKPs. Using Self, proofs can be generated at the time of data creation to verify that the data comes from real individuals. The Trusted Execution Environment (TEE) provided by Oasis ensures that these proofs are verified during the training process.
This way, as an insurance company, you can be confident that your models are trained properly and on validated, trustworthy data—without ever seeing the raw data yourself.