Show HN: Openlayer – test, fix, and improve your ML models https://ift.tt/h7lBgSi

Show HN: Openlayer – test, fix, and improve your ML models Hey HN! My name is Vikas, and my cofounders Rish, Gabe and I are building [Openlayer]( https://ift.tt/slUaChW ). Openlayer is an ML testing, evaluation, and observability platform designed to help teams pinpoint and resolve issues in their models. We each met at Apple where we were ML engineers experiencing the struggle that goes into properly evaluating models, making them robust to the myriad of unexpected edge cases they will encounter in production, and understanding the reasons behind their mistakes. It was like playing an endless game of whack-a-mole with Jupyter notebooks and CSV files — fix one issue and another pops up. This shouldn’t be the case. Error analysis is vital to establishing guardrails for AI and ensuring fairness across model predictions. Traditional software testing platforms are designed for deterministic systems, where a given input produces an expected output. Since ML models are probabilistic, testing them reliably has been a challenge. What sets Openlayer apart from other companies in the space is our end-to-end approach to tackling both pre- and post-deployment stages of the ML pipeline. This "shift-left" approach emphasizes the importance of thorough validation in pre-production rather than solely focusing on monitoring after deployment. Having a strong evaluation process pre-ship means fewer bugs in production, shorter and more efficient dev-cycles, and lower chances of getting into a PR disaster or having to recall a model. Openlayer provides ML teams and individuals with a suite of powerful tools to understand your models and data beyond simple metrics. The platform offers insights about the quality of your training and validation sets, the performance of your model across subpopulations of that data, and much more. Each of these insights can be turned into a “goal.” As you commit new versions of your models and data, you can see how your model progresses towards these goals, as you guard against regressions you may have otherwise not picked up on and continue to raise the bar. Here's a quick rundown of Openlayer's workflow: 1. Add a hook in your training / data ingestion pipeline to upload your data and model predictions to Openlayer via our API 2. Explore insights about your models and data and create goals around them - *Integrity* goals track the quality of your validation and training sets - *Consistency* goals guard against drift between your datasets - *Performance* goals evaluate your model's performance across subpopulations of the data - *Robustness* goals stress-test your model using synthetic data to uncover edge cases - *Fairness* goals help you understand biases in your model on sensitive populations 3. Diagnose issues with the help of our platform, using powerful tools like explainability (e.g. SHAP values) to get actionable recommendations on how to improve 4. Track the progress over time towards your goals with our UI and API and create new ones to keep improving Openlayer works best for teams — technical and non-technical folk can all play a role in the process of understanding, goal-setting, and improving. The platform enables this through a user-friendly UI (and API) with comments and notifications built-in to help keep everyone on the same page. We're excited to offer a free sandbox for you to try out the platform today! [You can sign up here]( https://ift.tt/USItXMY ). We are also soon adding support for even more ML tasks, so please reach out if your use case is not supported and we can add you to a waitlist. Give Openlayer a spin and join us in revolutionizing ML development for greater efficiency and success. Let us know what you think, or if you have any questions about Openlayer or model evaluation in general. https://ift.tt/dDyIAP5 May 10, 2023 at 06:00PM
Post a Comment (0)
Previous Post Next Post