- Google’s DeepMind develops 21 AI tools focused on life advice, planning, and tutoring.
- Collaboration with Scale AI, a $7.3 billion startup, to assess these AI tools.
- Concerns raised about the implications of seeking life advice from AI.
DeepMind, a leading AI subsidiary of Google, is reportedly in the process of creating an array of at least 21 AI tools dedicated to offering life advice, planning assistance, and tutoring services, according to a recent revelation by The New York Times.
In its continuing quest for AI excellence, Google’s DeepMind stands as a frontrunner, driving innovation at a rapid pace. The current project, aimed at providing tools for life guidance, is one such manifestation of this momentum.
However, it’s crucial to note that this endeavor comes on the heels of concerns from Google’s own AI safety team. Earlier reports suggest they warned company executives about potential risks, including compromised well-being and a sense of losing personal agency, associated with relying on AI for life advice.
To ensure the functionality and reliability of these tools, Google has initiated a partnership with Scale AI, a startup valued at $7.3 billion, whose core expertise lies in training and validating AI solutions. This ambitious project has already enlisted the talents of over a hundred Ph.D. experts. An essential component of this testing phase is to determine the AI’s capability in offering relationship advice and answering deeply personal questions.
For instance, as cited by the Times, one test prompt revolved around a dilemma of not being able to attend a close friend’s destination wedding due to financial constraints. The overarching goal is to gauge the AI’s proficiency in addressing such real-world issues.
It’s vital to understand that these tools, while cutting-edge, are not designed as therapeutic solutions. As an illustration of this distinction, Google’s Bard chatbot, accessible to the public, solely offers references to mental health resources when solicited for therapeutic counsel.
This demarcation is perhaps in response to past controversies surrounding the application of AI in therapeutic settings. Notably, the National Eating Disorder Association had to halt the operations of its Tessa chatbot following incidents where it provided detrimental advice on eating disorders. The medical community remains divided on the integration of AI in therapeutic realms, emphasizing that its deployment necessitates prudence.
Google DeepMind conveyed their commitment to product safety, stating, “We continually collaborate with multiple partners to assess our products and research. This rigorous evaluation is pivotal in delivering technology that’s both safe and beneficial.”
Read more on New York Times