AI is everywhere and now it’s helping monitor kid’s activities to maintain safety across different schools and prevent school shooting incidents. An AI startup called Securly, co-founded by Vinay Mahadik and Bharath Madhusudan focusses on student safety with the help of features such as web filtering, cyberbullying monitoring, and self-harm alerts.
Its cloud-based web filters maintain an age-appropriate internet, monitors bullying and ensures that schools remain CIPA-compliant. Another feature called ‘auditor’ in Securly uses Google’s Gmail service to send alerts when the risk of bullying or self-harm is detected. There’s also a tipline feature that sends anonymous tips to students over phone, text, or email.
The machine learning algorithms used by Securly are trained by safety specialists using safe and unsafe content. So, once the algorithms find any content as disturbing, the 24×7 student safety experts evaluate further context behind the activity and reach out to schools and authorities as needed.
Securly raised $16m in Series B round of funding last month. The company managed to raise $24 million and was led by Defy Partners. The company now wants to use these funds to amp up their research and development further in K-12 safety.
Moreover, Mahadik is also focussing on technologies that can be used across schools without hampering kids’ privacy. He told Forbes, ”You could say show me something that happened on the playground where a bunch of kids punched or kicked a certain kid. If you can avoid personally identifying kids and handle the data responsibly, some tech like this could be beneficial”. Securly currently has over 2,000 paid school districts using its free Chromebook filtering and email auditing services.
However, public reaction to the news isn’t entirely positive as many people are criticizing the startup of shifting the focus from the real issue ( i.e. providing kids with much-needed counseling and psychological help, implementing family counseling programs, etc ) and is instead promoting tracking every kid’s move to make sure they don’t ever falter. Securly is not the only surveillance service that has received heavy criticism. Predictim, an online service that uses AI to analyze risks associated with a babysitter, also got under the spotlight, over the concerns of its biased algorithms and for violating the privacy of the babysitters.
Can you imagine if our kids’ schools became centers for public health and education for our entire communities instead of pre-prisons criminalizing kids and their families for things beyond their control?
— ashe dryden (@ashedryden) January 9, 2019
Imagine if we spent that $16M on providing more after school programs that coincided with free job training, access to adult literacy education, food banks, medical clinics, tenant and labor rights organizations… this is just stuff off the top of my head.
— ashe dryden (@ashedryden) January 9, 2019
“there was strong demand despite the ramifications of a product suite that constantly monitors thousands of children 24/7” – as if demand makes this morally right 🤦♀️
— Jennifer Shehane (@jennifershehane) January 9, 2019
Do you recall the AI-powered "babysitter evaluator" system announced a few months ago? Due to backlash, they put their launch on hold. I can only hope the same thing happens here. https://t.co/xQEOQz1mHH
— David Makogon @ Home (@dmakogon) January 9, 2019
Read Next
Babysitters now must pass Perdictim’s AI assessment to be “perfect” to get the job
Center for the governance of AI releases report on American attitudes and opinions on AI