• 6 Posts
  • 350 Comments
Joined 2 years ago
cake
Cake day: June 26th, 2023

help-circle



  • I do not agree with the idea that humans are being trained to act like robots. Any company with a customer service department is likely tracking the root causes of their customers’ issues. With enough data, they can identify the most common problems and their solutions. If the goal is to resolve a customer’s issue as quickly as possible (which seems like a reasonable assumption), it makes sense to guide the customer through the most common solutions first, as that will likely solve the problem.

    If someone works in customer service and repeats the same script daily, it’s understandable that they may come across as robotic due to sheer boredom. A skilled customer service representative can recognize when to use the script and when to deviate. However, if a company fails to hire the right people and does not offer a fair salary, those best suited for the role are unlikely to take the job.



  • It appears this was a Victim impact statement.

    A victim impact statement is a written or oral statement made as part of the judicial legal process, which allows crime victims the opportunity to speak during the sentencing of the convicted person or at subsequent parole hearings.

    From the article (emphasizes mine):

    But the use of AI for a victim impact statement appears novel, according to Maura Grossman, a professor at the University of Waterloo who has studied the applications of AI in criminal and civil cases. She added, that she did not see any major legal or ethical issues in Pelkey’s case.

    "Because this is in front of a judge, not a jury, and because the video wasn’t submitted as evidence per se, its impact is more limited," she told NPR via email.