Search
- Workers' compensation insurance provides benefits to employees who are injured or become ill as a result of their job. It is a mandatory insurance program in most states in the US and is designed to protect both employees and employers. In exchange for providing workers' compensation insurance, employers are typically protected from lawsuits by injured employees.