As artificial intelligence reshapes the workplace, particularly in areas like payroll and employee management, understanding its impact on workers is more critical than ever. Today, I’m thrilled to sit down with Ling-Yi Tsai, a seasoned HRTech expert with decades of experience helping organizations navigate change through technology. With her deep expertise in HR analytics and the integration of tech in recruitment, onboarding, and talent management, Ling-Yi offers invaluable insights into the evolving role of AI in the workplace. In our conversation, we’ll explore employee apprehensions about AI in payroll processing, the importance of trust and transparency, and the broader implications of AI in hiring and management decisions.
How do you view the unease among U.S. workers regarding the use of AI in payroll processing?
I’m not surprised by the hesitation. Payroll is deeply personal—it’s about how people earn their livelihood. When you hear that 34% of workers are uncomfortable with AI calculating their wages, it likely stems from a fear of errors or a lack of understanding about how the technology works. People want assurance that their paychecks are accurate and handled with care, and AI can feel like a black box. There’s also a cultural element; many workers value the human touch in something as fundamental as getting paid. It’s less about rejecting technology and more about needing reassurance that it won’t mess up something so critical.
Why do you think such a significant number—45% of workers—oppose AI handling payroll inquiries instead of a human?
Payroll inquiries often come with urgency or emotional weight—think about a missed payment or a tax discrepancy. Workers want to speak to someone who can empathize and resolve issues quickly. AI, even at its best, can feel impersonal or rigid in these moments. That 45% opposition reflects a desire for human interaction, someone who can understand the nuances of their situation and offer tailored solutions. It’s not just about getting an answer; it’s about feeling heard and supported.
How crucial is trust between employees and employers when introducing AI for tasks like payroll?
Trust is absolutely foundational. Payroll isn’t just a transaction; it’s a signal of how much a company values its people. If workers don’t trust the technology—or the employer’s intentions behind using it—they’ll question the fairness and accuracy of the process. Without trust, even the most efficient AI system can backfire, leading to disengagement or resentment. Building that trust requires consistent communication and proof that the technology serves employees, not just the bottom line.
What steps can companies take to foster trust among employees regarding AI in payroll?
First, transparency is key. Companies need to explain why they’re using AI, how it works, and what safeguards are in place to prevent errors. Hosting town halls or creating simple guides can demystify the technology. Second, they should ensure there’s always a human point of contact for payroll issues. Workers need to know they can escalate concerns to a real person if something goes wrong. Finally, involving employees in the rollout process—maybe through feedback sessions—can make them feel like partners rather than subjects of a tech experiment.
Do you believe human oversight, as recommended by industry experts, is sufficient to ease employee concerns about AI in payroll?
It’s a critical starting point, but it’s not the whole solution. Human oversight reassures workers that there’s accountability—someone to catch mistakes or step in when needed. However, it must be paired with clear communication about the oversight process. Employees need to know who’s responsible and how they can reach them. Oversight alone won’t address deeper fears about job replacement or loss of personal connection. It’s a necessary layer, but companies must also focus on the human experience around the tech.
Payroll is often seen as a reflection of how much a company values its employees. How do you interpret this perspective?
I completely agree with that view. Payroll isn’t just about numbers; it’s about reliability and respect. When a paycheck is delayed or incorrect, it sends a message that the employee’s well-being isn’t a priority. On the flip side, consistent accuracy and timeliness show that the company cares about its people’s financial stability. It’s one of the most tangible ways an organization demonstrates its commitment to its workforce, and any technology used in this space must uphold that sense of value.
How does the accuracy and timing of payroll impact employees’ perception of their employer?
It’s huge. Imagine missing a paycheck by even a day—bills pile up, stress skyrockets. That kind of delay can erode trust instantly, making employees question whether the company has its act together. Accuracy is just as vital; errors in pay can feel like a personal slight, even if they’re unintentional. When payroll runs smoothly, it builds confidence in the employer’s competence and care. It’s often an unspoken contract—handle my pay right, and I’ll feel valued.
Can AI in payroll reflect a company’s values, or does it risk depersonalizing the process?
AI can reflect values if it’s designed with intention—prioritizing accuracy, speed, and accessibility shows a commitment to employee well-being. But there’s a real risk of depersonalization. Without human elements, like a friendly payroll team or personalized support, AI can make the process feel cold and transactional. The challenge is blending efficiency with empathy, ensuring the tech enhances rather than replaces the human connection that payroll often represents.
Shifting gears, how do you feel about AI’s growing role in hiring and employee management decisions, such as promotions or layoffs?
It’s a double-edged sword. AI can streamline hiring by analyzing vast amounts of data to identify strong candidates, and it can support management decisions with objective insights. But I’m wary of over-reliance. These are deeply human processes—deciding who gets hired, promoted, or let go involves nuance that AI might miss. Emotions, cultural fit, and individual potential aren’t always quantifiable. If not carefully managed, AI risks reducing people to data points, which can harm morale and trust.
Are you concerned about AI potentially introducing bias or overlooking qualified candidates in hiring processes?
Absolutely. AI systems are only as good as the data they’re trained on, and if that data reflects historical biases—like favoring certain demographics—those biases get baked into the outcomes. I’ve seen cases where AI screens out qualified candidates because they don’t fit a rigid profile. It’s a real problem, especially since many companies don’t fully audit their algorithms for fairness. Human judgment must remain in the loop to catch what the tech might miss and ensure diversity and equity aren’t compromised.
Given that only a small fraction of leaders have received ethical training on using AI for employee decisions, do you think companies are adopting this technology too quickly?
Yes, in many cases, the pace is alarming. AI is powerful, but without proper training on ethics and accountability, leaders risk making decisions that are unfair or damaging. Only 3 in 10 leaders having formal training is a red flag—it shows a gap in preparedness. Companies are eager to leverage AI for efficiency, but they’re often skipping the groundwork needed to use it responsibly. Slowing down to prioritize education and ethical guidelines isn’t just smart; it’s essential to avoid long-term harm.
How can companies effectively communicate transparency about AI use in payroll and other HR processes to ease worker concerns?
It starts with plain language. Companies should break down what AI does in payroll or HR—whether it’s calculating wages or screening resumes—and share that in accessible ways, like videos or FAQs. They should also be upfront about benefits and limitations, admitting where AI might fall short and how humans step in. Regular updates, like newsletters or Q&A sessions, can keep the dialogue open. The goal is to make employees feel informed and included, not blindsided by tech they don’t understand.
How important is it for employees to have a voice or at least a clear understanding of AI’s role in their workplace?
It’s incredibly important. When employees understand AI’s role, they’re less likely to fear it. More than that, giving them a voice—through surveys or focus groups—shows respect for their perspectives. It’s not just about reducing anxiety; it’s about building a culture of collaboration. If workers feel they have a say in how tech is used, they’re more likely to embrace it. Ignoring their input, on the other hand, can breed distrust and resistance.
Do you think AI in payroll and HR can ever completely replace human involvement, or should there always be a balance?
I’m a firm believer in balance. AI can handle repetitive tasks—calculations, data analysis—with speed and precision, freeing up humans for more strategic or empathetic roles. But completely removing human involvement is a mistake. People need people, especially in areas like payroll or HR where emotions and trust are central. AI should augment, not replace, the human element. Finding that sweet spot where tech and touch work together is the future I’d advocate for.
What is your forecast for the role of AI in payroll and employee management over the next decade?
I see AI becoming even more embedded in these areas, handling complex tasks like predictive analytics for payroll trends or personalized employee development plans. But I also predict a stronger push for ethical frameworks and regulations as concerns about bias and privacy grow. Companies will need to prioritize hybrid models—AI for efficiency, humans for oversight and connection. If done right, AI could transform HR into a more proactive, data-driven field, but only if trust and transparency remain at the core. We’re at a pivotal moment, and the next decade will show whether we can strike that balance.