Abstracts
Abstract
As much attention is turned to regulating AI systems to minimize the risk of harm, including the one caused by discriminatory biased outputs, a better understanding of how commercial practices may contravene anti-discrimination law is critical. This article investigates the instances in which algorithmic price personalization, (i.e., setting prices based on consumers’ personal information with the objective of getting as close as possible to their maximum willingness to pay (APP)), may violate anti-discrimination law. It analyses cases whereby APP could constitute prima facie discrimination, while acknowledging the difficulty to detect this commercial practice. It discusses why certain commercial practice differentiations, even on prohibited grounds, do not necessarily lead to prima facie discrimination, offering a more nuanced account of the application of anti-discrimination law to APP. However once prima facie discrimination is established, APP will not be easily exempted under a bona fide requirement, given APP’s lack of a legitimate business purpose under the stringent test of anti-discrimination law, consistent with its quasi-constitutional status. This article bridges traditional anti-discrimination law with emerging AI governance regulation. Pointing to identified gaps in anti-discrimination law, it analyses how AI governance regulation could enhance anti-discrimination law and improve compliance.
Résumé
À l’heure où l’on s’affaire à réglementer les systèmes de l’IA afin de minimiser les risques, y compris ceux causés par des résultats biaisés et discriminatoires, il est essentiel de mieux comprendre comment certaines pratiques commerciales pourraient violer les lois anti-discrimination. Cet article examine les cas où la personnalisation algorithmique des prix, c’est-à-dire la fixation des prix en fonction des renseignements personnels des consommateurs dans le but de se rapprocher le plus possible de leur volonté de payer maximale (APP), pourrait violer les lois canadiennes anti-discrimination. Il analyse les cas dans lesquels le APP pourrait constituer une discrimination prima facie, tout en reconnaissant la difficulté de détecter cette pratique commerciale. Il discute des raisons pour lesquelles certaines différenciations de pratiques commerciales, même pour des motifs interdits, ne conduisent pas nécessairement à une discrimination prima facie, offrant ainsi une analyse plus nuancée de l’application de la loi anti-discrimination à l’APP. Cependant, une fois la discrimination prima facie établie, le APP ne sera pas facilement exempté en vertu d’une exigence bona fide, étant donné l’absence d’objectif commercial légitime de l’APP et en vertu du test rigoureux de la loi anti-discrimination, en accord avec son statut quasi-constitutionnel. Cet article fait le lien entre le droit anti-discrimination traditionnel et la réglementation émergente en matière de gouvernance de l’IA. En soulignant les lacunes identifiées dans le droit anti-discrimination, il analyse comment la réglementation de la gouvernance de l’IA pourrait renforcer le droit anti-discrimination et en améliorer la conformité.
Article body
Introduction
The expanding use of autonomous decision-making tools by governments, law enforcement authorities, in employment, or in commercial transactions, raises concerns that have given rise to policy debates and legislative reform worldwide.[1] The increased algorithmic personalization of e-commerce transactions and their compliance with anti-discrimination law fall within those concerns.[2] Algorithmic price personalization (APP) is a form of differential pricing practice whereby suppliers set prices based on consumers’ personal information, with the objective of getting as close as possible to their maximum willingness to pay.[3] This article delves into APP’s compliance with anti-discrimination law (i.e., the body of law comprising Canadian federal and provincial human rights codes).[4] The questions posed by APP’s compliance with anti-discrimination law tie in with pressing issues concerning the deployment of artificial intelligence (AI) systems and the need for proper regulation.[5]
In an anti-discrimination law context, other than the harsh consequences of denying access to credit or allowing it at prohibitive cost or charging higher insurance rates, online purchases of goods or services involving potential discrimination on prohibited grounds (e.g., race, age, gender, sexual orientation, or family status) receive relatively less attention. Discrimination arising from day-to-day purchases may at first glance seem trivial and undeserving of a human rights complaint or further academic inquiry. The consequences of paying a few extra cents or dollars on a product or service, even as a result of a prohibited ground of discrimination, pale in comparison with being denied a job or housing based on such prohibited grounds or failing to accommodate one’s disability in an educational environment. However, with the intensification of algorithmically driven decision-making in all spheres of e-commerce, the compounded effects of prohibited discriminatory practices, if left unaddressed, defeat the raison d’être of anti-discrimination law, especially when differential treatment leads to the further marginalization of a protected group. This article looks into APP’s compliance with anti-discrimination law from a broad range of e-commerce transactions.
In addition to questions of anti-discrimination law, the legality of APP has given rise to numerous studies, commentary and reports from the perspectives of privacy and personal data protection, competition, contract, and consumer law, which are beyond the scope of this article.[6] In these studies and reports, APP compliance with anti-discrimination law is often mentioned as one of the main legal concerns, without providing a detailed analysis of how APP may contravene such law.[7] Similarly, there is relatively little academic literature that scrutinizes the legality of APP under anti-discrimination law, with some exceptions.[8]
The main contribution of this article is to fill this gap by investigating the instances in which the commercial practice of APP, with its specificities, contravenes Canadian anti-discrimination law. As much attention is turned to regulating AI systems to minimize the risk of harm, including the harm caused by discriminatory biased outputs,[9] understanding what may or may not violate anti-discrimination law is critical. The additional contribution of this article is to bridge traditional anti-discrimination law with emerging AI governance regulation, using the gaps identified in anti-discrimination law to show how AI governance regulation could enhance anti-discrimination law and improve compliance.[10]
Part I of this article defines APP and its various applications in e-commerce. Part II examines the treatment of supplier pricing practices in anti-discrimination law. This analysis includes how APP may lead to prima facie discrimination, how human rights codes address potentially countervailing social and economic norms in commerce relative to, for example, age or gender, and how algorithmically generated discrimination may be difficult to detect and prove. It also scrutinizes how the bona fide requirement applies to APP, by which a prima facie discriminatory practice or standard is absolved from a human rights code violation, with reference to the specific case of insurance contracts. Part III lays out how the analysis of APP and anti-discrimination law presented here may inform emerging models of AI governance regulation. In turn, it queries how such new forms of AI governance regulation may complement and improve compliance with anti-discrimination law in the future. The article concludes with a reminder of the broader issues raised by APP and other forms of AI systems.
I. Algorithmic Price Personalization (APP)
APP refers to the commercial practice by which firms set prices according to a consumer’s personal characteristics, targeting as much as possible their maximum willingness to pay (or the reservation price).[11] Often referred to as “perfect price discrimination,” it contrasts with versioning (offering different prices for different versions of a good or service)[12] or group pricing (charging different prices to different groups of consumers based on a personal characteristic they share (e.g., age, gender, or student status).[13] Differential pricing on certain prohibited grounds of discrimination, such as age and discount price for seniors, or age and gender for insurance contracts, is an established exception in anti-discrimination law that has been justified by countervailing societal goals or accepted industry norms.[14]
APP should not be confused with dynamic pricing where prices vary based on offer and demand.[15] Dynamic pricing is a common practice that prevails in the airline and hospitality industries. Given the opacity of pricing techniques and the limited ability to distinguish between APP and dynamic pricing, the line between APP and dynamic pricing may be blurry at times.[16] APP also differs from price steering (tailoring the order with which offers for goods or services are listed) or targeted advertising (the selection of advertising displayed to the consumer). For those commercial practices, firms will differentiate between buyers by using their personal characteristics. However, such differentiation does not influence the price per se.[17]
While earlier economic studies have been guarded as to the extent to which APP occurs, notably due to a lack of substantiated empirical research and the traditional economic theory requirements for APP (or first-degree price discrimination) to take place, there is growing evidence that firms are resorting to APP in online transactions.[18] APP is also likely to occur in payment-less brick-and-mortar retail stores.[19] Studies of price personalization directly relevant to anti-discrimination law report differential treatment for consumer credit ratings and insurance premiums based on race.[20] In other instances, differential treatment may result from pricing relying on criteria other than prohibited grounds (e.g., postal code areas) while potentially contravening anti-discrimination law.[21] Sometimes, price personalization leads to the sheer exclusion of a person from the market, such as by the refusal to provide credit or personal insurance.
The traditional economic theory preconditions for APP to occur are (i) the ability to assess consumers’ individual willingness to pay, (ii) the absence of or limited arbitrage,[22] and (iii) the presence of market power.[23] These preconditions need to be reconsidered in the online environment. Increasingly powerful tools using personal data are deployed by suppliers to influence online consumer purchasing decisions.[24] This influence may lead to “micro-market place chambers,” where consumers’ judgments of competitive alternatives are blurred.[25] This phenomenon is amplified for customers of large retail or service platforms (such as Amazon and Uber) where market power and control may hide beneath seemingly competitive prices.[26] APP may even occur for goods or services susceptible to arbitrage (i.e., which can be resold) or in (imperfectly) competitive markets.[27]
Various surveys indicate a strong consumer dislike of APP, which is viewed as unfair.[28] Such negative perceptions go beyond instances whereby APP would differentiate on a prohibited ground. They include any form of price personalization based on consumer profiling through the use of their personal information.[29] As a result, one can reasonably predict that retailers will either refrain from the practice or conceal it so as not to upset their consumer base. In fact, the ability to hide APP is arguably another precondition for it effectively taking place.[30] This lack of transparency impacts the burden of proof in allegations of discrimination on prohibited grounds.[31]
II. Supplier Pricing Practices and Anti-Discrimination Law
A. What Constitutes Discrimination under Human Rights Codes?
Federal and provincial human rights codes provide the right for individuals to equal treatment through enumerated prohibited grounds of discrimination. Relevant to APP is the protection against discriminatory treatment in the provision of goods or services generally available to the public[32] or to contract on equal terms.[33] These prohibitions against discrimination include the refusal to sell goods or services, or differential terms, such as the refusal to provide personal financing, or to do so on prohibitive terms. The Canadian Human Rights Act [CHRA][34] applies to federally regulated undertakings (such as banks, airlines, rail enterprises, and Crown corporations).[35] The Human Rights Code of Ontario [HRCO][36] and similar legislation in other provinces[37] apply to individuals and organizations whether in the public or private sector, unless they fall under exclusive federal jurisdiction.[38] This legislation forms the main body of anti-discrimination law applicable to commercial undertakings and practices such as APP.[39]
Prohibited grounds of discrimination are similar among the federal and provincial human rights codes, with some variations.[40] For instance, the HRCO prohibits discrimination on the basis of “race, ancestry, place of origin, colour, ethnic origin, citizenship, creed, sex, sexual orientation, gender identity, gender expression, age, marital status, family status or disability.”[41] All those prohibited grounds are potentially relevant to APP, given the precision with which algorithms deployed in e-commerce can accurately determine our demographic information, thereby profiling our “biographical core.”[42]
The Supreme Court of Canada has reiterated the quasi-constitutional status of human rights codes on numerous occasions.[43] The rights they provide warrant a large and liberal interpretation and the exceptions thereto are narrow.[44] The codes prohibit direct and indirect discrimination (including constructive discrimination) irrespective of intent.[45] Unlike direct prohibited discrimination,[46] indirect discrimination may occur when otherwise neutral considerations or policies in contractual terms, or in the offer of goods or services, nevertheless negatively impact certain groups.[47]
The Supreme Court’s decision in Meiorin[48] has brought a unified approach to discrimination, whether direct or indirect.[49] The complainant bears the burden of establishing discrimination prima facie[50] by showing (1) that they have a characteristic protected from discrimination under the relevant human rights code; (2) that they experienced an adverse impact with respect to the provision of the good or service; and (3) that the protected characteristic was a factor in the adverse impact.[51] Despite this unified approach to discrimination, practical differences remain between direct and indirect discrimination, including potentially greater evidentiary hurdles to establish indirect discrimination, even if the requirement is only prima facie.[52]
The following section examines how commercial practices surrounding pricing terms or access to publicly available goods or services were held to be prima facie discrimination by courts or tribunals, with additional examples from research reports. It provides further examples of how APP may lead to direct or indirect discrimination. It then investigates why pricing practices or access to goods or services that appear discriminatory on their face may not necessarily violate human rights codes and how this is relevant to APP. It also raises some of the difficulties inherent to APP in establishing prohibited forms of discrimination.
Pricing practices involving direct prima facie discrimination include the refusal to offer personal financing or to do so on prohibitive terms based on race.[53] They also include charging higher automobile insurance rates according to age or gender, which is a long-established form of APP in the insurance industry. The fact that some human rights codes explicitly allow these forms of discrimination in insurance contracts[54] does not negate that discrimination occurs within the ambit of the relevant human rights code.[55] A provider of goods or services must nonetheless prove that such discrimination is justified under a bona fide requirement.[56] Any other form of APP where the prohibited ground is directly embedded as a determinant factor in the pricing algorithm could amount to prima facie discrimination. Other examples of direct discrimination surrounding pricing terms include a restaurant asking black patrons to prepay for their meals while not imposing a similar condition on others.[57] Differential levels of access to commercial venues or services for people with a disability have also been held to be prima facie direct discrimination.[58]
Indirect prima facie discrimination on pricing terms could include online concert ticket sales whereby accessible seats are sold at a higher price than non-accessible ones, given their specific locations in selected seating areas.[59] Similarly, indirect prohibited discrimination could occur when banks or other lenders charge more for their services in a “financial desert” due to the lack of competition and when they suspect that customers are unlikely to shop around. Discriminating on the basis of such geographic location could constitute indirect discrimination if it disproportionately affects protected (e.g., racialized) groups under human rights codes.[60]
Not all forms of differentiation, even on protected grounds, are prohibited under human rights codes.[61] In addition to built-in exemptions for insurance contracts[62] or differential treatment based on age,[63] countervailing social norms or the difficulty in establishing prejudice will bar a discrimination claim from succeeding. The section that follows explores the blurry contours of commercial practices such as the “pink tax,” “ladies night,” or single-gender clubs[64] and how they may or may not contravene anti-discrimination law. This exercise leads to a more nuanced view of what constitutes prohibited forms of discrimination, with important ramifications for APP.
The “pink tax” is a grey area of anti-discrimination law directly relevant to APP. It refers to commercial practices whereby women’s clothing and products are priced higher than similar clothing or products marketed to men. On its face, the “pink tax” could be a prohibited form of discrimination based on gender. However, in practice, legislative efforts and legal action attacking it have often been unsuccessful.[65] Similarly to the difficulty in distinguishing APP from dynamic pricing[66] or price versioning,[67] the lines between what constitutes prohibited discrimination and product differentiation may be blurry in the case of the “pink tax.” While detailed studies demonstrate the “pink tax” phenomenon,[68] lawmakers are not readily convinced that it amounts to a prohibited form of discrimination.[69] Two products might be identical in production and content, yet there can be differences impacting pricing, such as the higher marketing costs for women’s products. This argument was successfully made against the claim that the “pink tax” was a prohibited form of discrimination.[70] Similarly, courts have taken the debatable position that women are not constrained to solely purchase products and apparels marketed to them, concluding that the commercial practice did not violate anti-discrimination law.[71]
If the pink tax essentially results from assumptions that women are willing to pay more for certain goods and services,[72] and to the extent that it demonstrably amplifies wealth inequality between men and women,[73] then the pink tax could constitute prima facie discrimination. APP, which is about getting as close as possible to a buyer’s maximum willingness to pay, could exacerbate this phenomenon. That is, to expand the pink tax to gender-neutral goods or services, beyond women’s clothing and products. This could warrant legislative attention tackling the issue at a systemic level, rather than at the level of a single transaction, to overcome the burden to prove adverse impact that is more than de minimis.
There exist other commercial practices which at first glance are discriminatory on prohibited grounds, but do not necessarily contravene human rights codes, such as “ladies night” (i.e., offering a lower cover charge or other preferential treatment to women in bars or restaurants). Women-only facilities open to the public also fall in this category. Claims that these practices are prohibited forms of discrimination based on gender have been unsuccessful.[74] Even if, in such examples, men represent a protected group[75] and their gender is a factor of differentiation, it would be difficult for them to demonstrate that they have suffered an adverse impact under the prima facie discrimination test.[76] Such impact should not be de minimis. It includes humiliation, the perpetration of stigmatization, or arbitrary detrimental exclusion.[77] For instance, refusing to offer personal financing or to do so on prohibitive terms with respect to a racialized or traditionally marginalized group would likely result in such negative impact. In contrast, men will have a harder time establishing that they suffered such prejudice as a result of the exclusivity of women-only facilities or preferential pricing for women.[78] And where a complainant may be able to show injury, such as significant cost difference or prohibited access with no other reasonable alternative, the differential treatment may still be justified by countervailing social norms, or benefits to another protected group under the bona fide requirement.[79] In the case of APP, the necessity of showing a detriment that is more than de minimis (e.g., a small price variation) may pose a challenge in establishing prima facie discrimination, unless the cumulative effects of APP can be factored in.
There are other factors specific to APP that make its compliance with anti-discrimination law difficult to ascertain. First, APP offerings are continually subject to change. Discrimination will be difficult to detect, as there might be no base reference price against which to establish it. Second, APP can appear in the form of real-time dynamic pricing.[80] Third, and more generally, the lack of transparency surrounding confidential business information regarding algorithms, data sets, and the criteria upon which algorithms set prices, may impair the ability to make a successful claim of prima facie discrimination.[81] For these reasons, and to the extent that determining a buyer’s maximum willingness to pay may in some cases automatically entail differentiation on prohibited grounds, APP may give rise to even more human rights code violations than in standardized retail pricing environments.
In short, the commercial practice of APP may lead to direct or indirect discrimination contrary to human rights codes—potentially more so than in brick-and-mortar businesses—despite the difficulties in detecting, ascertaining and establishing prima facie discrimination. That said, not all forms of differentiation, even on prohibited grounds, contravene anti-discrimination law. In some cases, the need to demonstrate an adverse impact resulting from pricing practices or access differentiation will be a challenge in establishing prima facie discrimination. This reality reveals a more nuanced application of anti-discrimination law to APP.
B. Whether the Discriminatory Standard Is Based on a Bona fide Requirement
Assuming a prima facie case of discrimination is proven with respect to APP or another commercial practice, discriminatory practices may still be allowed if they meet the bona fide requirement exception.[82] This exception applies in the employment context or to suppliers of goods or services to the public with necessary variations, as dictated by the language of the relevant human rights code.[83] For instance, a variation of the bona fide exception applies to prima facie discriminatory practices in insurance contracts.[84]
In order to benefit from the exception, the respondent must establish on a balance of probabilities that (1) the supplier of goods or services adopted the standard for a purpose rationally connected to the provision of the good or service; (2) the standard was adopted in the good faith belief that it was necessary to the fulfillment of that legitimate purpose; and (3) the standard is reasonably necessary to the accomplishment of that legitimate purpose.[85] The respondent must therefore demonstrate the impossibility of accommodating the claimant without imposing undue hardship upon themselves as a goods or services provider.[86] This undue hardship could include prohibitive financial costs, increased safety risk, or undermining the positive benefits to women patrons at women-only fitness club facilities.[87] As an exception to anti-discrimination principles, the bona fide requirement must be interpreted restrictively.[88]
With APP, the question arises as to whether the respondent can prove that the prima facie discrimination is rationally connected to the provision of goods or services. As set out in Meiorin, the business practice cannot meet the bona fide requirement exception unless the purpose for the discriminatory standard is legitimate.[89] In the case of APP, discrimination would be exercised to get as close as possible to the buyer’s maximum willingness to pay.[90] The business rationale of maximizing supplier profits would be, on its own, insufficient to justify prima facie discrimination. Several decades of standardized retail pricing practices without recourse to profiling buyers through their personal data (which, at this stage of the analysis, would be considered prima facie discrimination), weaken the argument of a rationally connected legitimate purpose of the discriminatory requirement. This business rationale contrasts with a legitimate purpose based on the safety of a product or a reasonable customer service requirement. To claim that profit maximization is a legitimate purpose in itself would be contrary to the quasi-constitutional protection offered by human rights codes and would make the principles they enshrine meaningless. As a matter of fact, rare are businesses for which profit maximization is not a primary goal.[91]
In addition, APP would also in some cases fail the second element of the bona fide requirement exception, if the algorithm’s criteria were specifically created to differentiate price on a prohibited ground, leading to prima facie discrimination. Such criteria, coupled with the absence of a mechanism to prevent discrimination contrary to human rights codes, could be characterized as a discriminatory animus motivation.[92]
Insofar as the commercial practice of APP satisfies the first two elements of the bona fide requirement exception laid out in Meiorin, the supplier of goods or services would then have to prove that the discriminatory standard is reasonably necessary (that its removal, after investigating non-discriminatory alternatives, would cause undue hardship to its business).[93] For example, a respondent may argue that modifying algorithms to make pricing compliant under human rights law would engender significant prohibitive costs that would be seriously harmful to their business.[94] General “impressionistic” evidence of increased risks or costs entailed by changing the standard to accommodate the claimant will not satisfy the respondent’s burden of proof.[95] At this stage of the analysis, the burden is strictly a matter of bringing satisfactory evidence of high financial costs or other similar burdens.
As an earlier form of algorithmic price personalization, the case of insurance contracts[96] is instructive regarding the dilemma surrounding sound industry practices and respect for anti-discrimination law principles. The leading decision applying the bona fide exemption for insurance contracts is Zurich Insurance Co. v. Ontario (Human Rights Commission) [Zurich Insurance].[97] In this case, the Supreme Court dealt for the first time with the provision in the HRCO allowing insurance providers to discriminate “on reasonable and bona fide grounds because of age, sex, marital status, family status or [handicap].”[98] The Court acknowledged that the specificities of insurance contracts warrant an adaptation of the bona fide requirement.[99]
Insurance premiums are set by assessing risk among various population groups which sometimes overlap with prohibited grounds, as explicitly recognized by the legislative exemption in the HRCO.[100] The Court stated that an insurance practice is “reasonable” in the context of the specific human rights code exemption if “(a) it is based on a sound and accepted insurance practice” that is “one which it is desirable to adopt for the purpose of achieving the legitimate business objective of charging premiums that are commensurate with risk;” and (b) “there is no practical alternative,” which is a question of fact.[101] Furthermore, the bona fide test is met if the practice “was adopted honestly, in the interests of sound and accepted business practice and not for the purpose of defeating the rights protected under the Code.”[102]
After acknowledging that the practice was undoubtedly sound and based on accepted business practice,[103] relying on credible actuarial evidence for risk assessment, Justice Sopinka, for the majority, noted that assessing the reasonableness of the practice required more.[104] Simply allowing “statistically supportable” discrimination would defeat the intent of human rights legislation. It would “perpetuate traditional stereotypes with all of their invidious prejudices.”[105] The respondent had to demonstrate that it had no reasonable alternative. Justice Sopinka was satisfied on this front. Justice L’Heureux-Dubé and Justice McLachlin (as she then was) disagreed on that point in separate dissenting reasons.[106] For Justice McLachlin, the respondent did not meet the burden of proof through its own failure to collect the relevant data proving that a reasonable alternative in fact did not exist.[107] Confusing the absence of a reasonable alternative with the absence of proof of such an alternative is to shift the burden of proof from the person who is prima facie in violation of the Code to the complainant.[108] For Justice McLachlin, this approach encourages maintaining discriminatory practices rather than reform aimed at achieving the Code’s objectives.[109]
The tensions highlighted in Zurich Insurance between anti-discrimination law and the extent to which insurance providers should be exempted from human rights codes provide relevant insights for APP. They are also reminiscent of current concerns and legislative reform debates on how to regulate the risks of biased outputs produced by AI systems.[110] This includes requiring AI system producers and users to have ex ante mechanisms in place to prevent violations of anti-discrimination law.[111]
In an era of advanced personalized data analytics, should insurance companies still be allowed to build in prohibited grounds of discrimination in their actuarial assessments of risk and corresponding insurance rates? Or, on the contrary, should individuals be assessed on their own merits, freed from the large group risk characteristics of the class(es) to which they belong? The increased lack of uniformity across jurisdictions in Canada, the United States and other countries allowing gender, age, and other forms of discrimination in insurance contracts calls into question the “sound accepted business practice” argument for maintaining at least some of the statutory anti-discrimination exemptions in insurance contracts.[112] Furthermore, evolving societal norms on gender expression and fluidity question the value of gender as a meaningful and distinct categorisation. Some insurance companies have taken note of this evolution and offer different insurance rates to transgender and non-binary people who do not identify exclusively as either female or male.[113] In light of these developments, the impracticality raised by Justice Sopinka in Zurich Insurance, whereby insurance rates would be difficult to determine without reference to specific groups on prohibited grounds of discrimination, is less and less convincing.
Decisions concerning insurance services post-Zurich Insurance suggest that courts and tribunals are willing to accept current discriminatory practices based on age or gender.[114] Justice Sopinka’s warning to insurers to avoid incorporating prohibited grounds of discrimination in their actuarial calculations,[115] and the dissents’ harsh criticism of these practices, did little to prompt insurers to change their ways, or legislators to re-examine the exemptions in human rights codes, at least in Ontario. Thirty years later, the impugned provision of the HRCO is still in force, and other human rights codes contain similar exemptions for insurance contracts.[116]
In short, once a complainant establishes that a form of APP is prima facie discrimination (overcoming the hurdles raised above),[117] the supplier resorting to APP will not be easily exempted under a bona fide business requirement, leading to a violation of anti-discrimination law. Discriminating between consumers on prohibited grounds by profiling them through the collection of their personal information[118] for the sole purpose of maximizing suppliers’ profit would not on its own be a legitimate and reasonable purpose to set aside the right to equal treatment of designated groups enshrined in human rights law. Furthermore, evolving social norms on gender, and increasingly precise big data analytics, call for a re-examination of the justifiability of human rights code exemptions for insurance contracts. The next part examines how our findings may inform ongoing AI governance legislative reform and future regulation, and how such regulation may fill in turn the gaps identified in relation to evidentiary matters and compliance with anti-discrimination law.
III. Algorithmic Price Personalization (APP) and AI Governance Regulation
There are serious concerns supported by empirical research that algorithmic decision-making tools and AI systems perpetrate bias and discrimination on prohibited grounds, whether directly or indirectly.[119] Ongoing Canadian legislative reform on AI governance seeks to address discriminatory outputs by imposing obligations on firms using AI systems to identify, measure, and mitigate their risks of contravening anti-discrimination law, as well as by imposing duties to record and disclose the tasks generated by the AI systems.[120] Under such a regulatory regime, APP would likely be subject to the highest level of obligations, given its categorization as a “high-impact system.” [121] These obligations are applicable to systems that generate decisions on “the type or cost of services to be provided to an individual.”[122] The AI governance regime confers administrative and investigative powers to both the designated Minister and the AI and Data Commissioner, who are empowered to order AI systems to be shut down, with substantial fines for non-compliance.[123] Through its explicit reference to anti-discrimination law, the AI governance regime is meant to reinforce this body of law ex ante.
Without providing a detailed account of all the implications and potential shortcomings of this AI governance regime, the analysis presented here provides some insight into how this and similar regimes may indeed reinforce compliance with anti-discrimination law and its key objectives with regard to APP.
Overall, a successful AI governance regime should decrease the occurrence of prohibited forms of discrimination. Additionally, when claims based on anti-discrimination law arise, it should improve the evidentiary process for claimants. A successful regime will give due weight to the quasi-constitutional status of anti-discrimination law and strengthen compliance for the benefit of members of a protected group, while clarifying the path toward compliance for service providers.
In the case of APP, this article identified two broad categories of problems. Under the first category, it discussed the issues related to the assessment of discrimination. For instance, constantly fluctuating prices make the determination of a reference price difficult to establish, creating obstacles to proving prima facie discrimination. This article also discussed how the lines may be blurry between pricing goods or services on the basis of personal characteristics (which may lead to the violation of human rights codes) and dynamic pricing. Obligations in the proposed federal AI governance regime requiring users of AI systems to assess and mitigate the risk of biased output (i.e., output contrary to anti-discrimination law) entails the written recording of steps undertaken, regular testing, and adjustments toward such risk mitigation.[124] These obligations include ensuring the algorithms do not incorporate prohibited forms of discrimination, and should also include mechanisms by which service providers regularly monitor and investigate pricing trends that differentiate one or more members of protected groups. Such duties would provide the opportunity to rectify their pricing algorithms accordingly. The ways in which firms can actually and effectively implement legal parameters to avoid biased output and comply with anti-discrimination law is complex, giving rise to ample scholarly debates which are beyond the scope of this article.[125]
Under the second category of issues, this article discussed algorithmic transparency problems, leading to evidentiary issues for claimants seeking to establish prima facie discrimination. Obligations requiring service providers using AI systems to maintain an accountability framework,[126] and to disclose it during tribunal or court proceedings, or through a regulatory order (including demonstrating the effectiveness of their anti-discrimination law mechanisms) could alleviate the evidentiary burden of human rights violation complainants.[127] Moreover, such obligations would facilitate service providers’ ability to adequately and more efficiently respond to human rights claims. Overall, the powers to monitor, investigate, shutdown, and sanction non-compliant systems vested in the AI governance authority should encourage greater compliance with anti-discrimination law.
The effects of compliance with AI governance regimes could include a reduction of differentiation on prohibited grounds of discrimination even if, as discussed earlier, such differentiation is not always in contravention of anti-discrimination law.[128] In other words, the difficulty in implementing anti-discrimination law, with all its nuances and subtleties, could lead firms to more restraints on differentiation than what is legally required. Furthermore, at a time when some exemptions allowing discriminatory practices for insurance contracts no longer seem justified,[129] another side effect of firms’ implementation of AI governance regime requirements might be to revisit, if not eliminate, such long-established discriminatory practices altogether.
Conclusion
This article tackled an underexplored area of law: how APP may comply with or contravene anti-discrimination law in Canada. It examined several examples where APP can lead to direct or indirect discrimination. By its nature, APP may be hard to detect, therefore rendering prima facie discrimination difficult to establish. The analysis also showed that not all price differentiation, even if based on prohibited grounds, necessarily violates anti-discrimination law, presenting a more nuanced view on the scope of application of this body of law. This article also explored the long-established exception of discrimination for insurance contracts, and whether it remains justified in the current context of changing social norms and a world of highly personalized datafication.
To the extent that a claimant overcomes the obstacles to successfully establishing discrimination prima facie, APP will not be easily exempted under a bona fide requirement, given APP’s lack of a legitimate business purpose under the stringent test of anti-discrimination law and its quasi-constitutional status.
This article bridged traditional anti-discrimination law with emerging AI governance regulation. It used the gaps identified by applying anti-discrimination law to APP to show how AI governance regulation could enhance anti-discrimination law and ensure greater compliance.
APP and the intensification of personalization in e-commerce raise broader issues beyond anti-discrimination law, such as in personal data protection law, competition law, and consumer law.[130] Similarly, the commercial practice of APP is one among many uses of algorithmic automated decision-making that give rise to serious societal concerns which hopefully will be addressed, at least in part, by AI system governance regulation.
Appendices
Notes
-
[1]
In Canada, see e.g. Bill C-27, An Act to enact the Consumer Privacy Protection Act, the Personal Information and Data Protection Tribunal Act and the Artificial Intelligence and Data Act and to make consequential and related amendments to other Acts, 1st Sess, 44th Parl, 2022 (first reading 16 June 2022) (if enacted, leading to the implementation of inter alia the Artificial Intelligence and Data Act [AIDA]); see also Canada, Standing Committee on Industry and Technology, Annex B: Guide to Proposed Amendments to AIDA (Ottawa: ISED, 2023), online (pdf): <ourcommons.ca> [perma.cc/ S7NA-8ADS] [Proposed Amendments to AIDA]. At the date of completion of this article, Bill C-27 had not yet become law. See also Law Commission of Ontario, Accountable AI (Toronto: Law Commission of Ontario, 2022).
-
[2]
AIDA, supra note 1, cls 5(1), 8, 14; Proposed Amendments to AIDA, supra note 1, cls 9(1)(b), 10(1)(b), 11(1)(b) (obligating creators and users of artificial intelligence systems to comply with anti-discrimination law).
-
[3]
See Part I, below, further defining APP in contrast to other forms of price personalization.
-
[4]
On the nature and scope of anti-discrimination law in Canada, and on its application to commercial transactions, see Part II(A), below, introductory paragraphs.
-
[5]
AIDA, supra note 1; Proposed Amendments to AIDA, supra note 1. For a critical analysis of AIDA, see Teresa Scassa, “Regulating AI In Canada: A Critical Look at the Proposed Artificial Intelligence and Data Act” (2023) 101:1 Can Bar Rev 1.
-
[6]
See e.g. Akiva A Miller, “What Do We Worry About When We Worry About Price Discrimination? The Law and Ethics of Using Personal Information for Pricing” (2014) 19:1 J Tech L & Pol’y 41; Ariel Ezrachi & Maurice Stucke, Virtual Competition: The Promise and Perils of The Algorithm-Driven Economy (Cambridge, Mass: Harvard University Press, 2016); Frederik Zuiderveen Borgesius & Joost Poort, “Online Price Discrimination and EU Data Privacy Law” (2017) 40:3 J Consumer Pol’y 347; Christopher Townley, Eric Morrison & Karen Yeung, “Big Data and Personalized Price Discrimination in EU Competition Law” (2017) 36:1 YB Eur L 683; Pedro Gonzaga, Michael Donohue & Dries Cuijpers, Personalized Pricing In The Digital Era—Background Note By The Secretariat (Paris: Organisation for Economic Co-operation and Development, 2018) [OECD Competition Committee]; Pascale Chapdelaine, “Algorithmic Personalized Pricing” (2020) 17:1 NYU JL & Bus 1 [Chapdelaine APP]; Pascale Chapdelaine, “Algorithmic Personalized Pricing: a Personal Data Protection and Consumer Law Perspective” (2024) 102:1 Can Bar Rev 1.
-
[7]
See e.g. Ezrachi & Stucke, supra note 6 at 124–27; OECD Directorate for Financial & Enterprise Affairs Competition Committee, Personalized Pricing in The Digital Era – Note by The European Union, DAF/COMP/WD(2018)128 (Paris: OECD, 2018); Alan M Sears, “The Limits of Online Price Discrimination in Europe” (2019) 21:1 Colum Sci & Tech L Rev 1 at 27–36; Option Consommateurs, “Changes to Prices Advertised Online: Analysis of Business Practices and the Legal Framework in Canada” (June 2018) at 42–43, online (pdf): <option-consommateurs.org> [perma.cc/8WTF-BQG8].
-
[8]
For a notable exception from a European law perspective, see Frederik Zuiderveen Borgesius, “Price Discrimination, Algorithmic Decision-Making, and European Non-Discrimination Law” (2020) 31:3 Eur Bus L Rev 401. See also Sears, supra note 7 at 27–37.
-
[9]
See e.g. Cathy O’Neil, Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy (New York: Crown, 2016); Deborah Hellman, “Measuring Algorithmic Fairness” (2020) 106:4 Va L Rev 811; Anja Lambrecht & Catherine Tucker, “Algorithmic Bias? An Empirical Study into Apparent Gender-Based Discrimination in the Display of STEM Career Ads” (9 March 2018), online: <dx.doi.org> [perma.cc/ 252D-42Y7]; Ignacio N Cofone, “Algorithmic Discrimination is an Information Problem” (2019) 70:6 Hastings LJ 1389; Atin Jindal, “Misguided Artificial Intelligence: How Racial Bias is Built Into Clinical Models” (2022) 2:1 Brown Hospital Medicine 2831; Robert Bartlett et al, “Consumer-Lending Discrimination in the FinTech Era” (2022) 143:1 J Financial Econ 30; See also House of Lords, Select Committee on Artificial Intelligence, AI in the UK: Ready, Willing and Able? (London: Authority of the House of Lords, 2018) at 44.
-
[10]
See discussion in Part III, below.
-
[11]
OECD Competition Committee, supra note 6 at 9; Ezrachi & Stucke, supra note 6 at 85–86.
-
[12]
OECD Competition Committee, supra note 6 at 9; Miller, supra note 6 at 55; Townley, Morrison & Yeung, supra note 6.
-
[13]
OECD Competition Committee, supra note 6 at 9; see also Miller, supra note 6 at 55; Townley, Morrison & Yeung, supra note 6 at 690.
-
[14]
See e.g. Human Rights Code, RSO 1990, c H 19, s 15 [HRCO] (preferential treatment based on age, i.e. 65 and over). See also discussion on insurance contracts in Part II, below.
-
[15]
OECD Competition Committee, supra note 6 at 9; Ezrachi & Stucke, supra note 6 at 87–88.
-
[16]
Ezrachi & Stucke, supra note 6 at 87–88.
-
[17]
OECD Competition Committee, supra note 6 at 9–10; see also Ezrachi & Stucke, supra note 6 at 107–08.
-
[18]
Chapdelaine APP, supra note 6 at 12–14 (citing several studies on the existence of APP as a widespread commercial practice). See also Ethan Wilk, “An Old-Fashioned Economic Tool Can Tame Pricing Algorithms: Left Unchecked, Pricing Algorithms Might Unintentionally Discriminate and Collude to Fix Prices” (26 April 2022), online: <scientificamerican.com> [perma.cc/TJY8-BZV2].
-
[19]
Amazon Go brick-and-mortar retail stores are highly personalized payless stores. See Andria Cheng, “Why Amazon Go May Soon Change The Way We Shop” (13 January 2019), online <forbes.com> [perma.cc/7L4X-ZF23].
-
[20]
Gissela Moya & Vinhcent Le, “Algorithmic Bias Explained” (last visited 7 July 2024) at 19–20, 24, online (pdf): <greenlining.org> [perma.cc/PK98-RPDA] (referring to various research reports indicating racial discrimination through the use of fintech tools).
-
[21]
Ibid.
-
[22]
The limited ability of buyers to resell goods or services acquired from suppliers, such as non-transferable purchases (airline tickets, hotel bookings), which would otherwise create a market that competes with the suppliers’ market.
-
[23]
Ezrachi & Stucke, supra note 6 at 86–87; OECD Competition Committee, supra note 6 at 13; Gerhard Wagner & Horst Eidenmüller, “Down by Algorithms? Siphoning Rents, Exploiting Biases, and Shaping Preferences: Regulating the Dark Side of Personalized Transactions” (2019) 86:2 U Chicago L Rev 581 at 585–86; Oren Bar-Gill, “Algorithmic Price Discrimination: When Demand is a Function of Both Preferences and (Mis)perceptions” (2019) 86:2 U Chicago L Rev 217 at 227. To these traditional pre-conditions, one may add the ability to conceal the practice of APP from buyers (Chapdelaine APP, supra note 6 at 17–18).
-
[24]
Chapdelaine APP, supra note 6 at 12–18.
-
[25]
I make reference here to “micro-market-place chambers,” by analogy to the phenomenon of “echo chambers;” see Ezrachi & Stucke, supra note 6 at 108–09.
-
[26]
On the lack of transparency in Uber’s algorithmic surge price settings and market power leading to the illusion of a competitive price, see Ezrachi & Stucke, supra note 6 at 208–11.
-
[27]
Miller, supra note 6 at 54, 57; Andrew Odlyzko, “Privacy, Economics, and Price Discrimination on the Internet” (Paper delivered at the ICEC2003 Fifth International Conference on Electronic Commerce, Minneapolis, 27 July 2003) at 5 (describing the recognition within economic literature that price discrimination can arise in a competitive environment).
-
[28]
See European Commission, Consumer Market Study on Online Market Segmentation Through Personalized Pricing/Offers in the European Union (Brussels: European Commission, 2018) at 243; Citizens Advice, “A Price of One’s Own: An Investigation into Personalized Pricing in Essential Markets” (31 August 2018), online: <citizensadvice.or.uk> [perma.cc/4FJ8-Y3JJ]; Joost Poort & Frederik J Zuiderveen Borgesius, “Does Everyone Have a Price? Understanding People’s Attitude Towards Online and Offline Price Discrimination” (2019) 8:1 Internet Pol’y Rev 1 (analysis of two surveys conducted in the Netherlands, whereby a vast majority of consumers viewed the practice of online price discrimination as unfair); Gabriele Pizzi et al, “Privacy Concerns and Justice Perceptions with the Disclosure of Biometric versus Behavioral Data for Personalized Pricing: Tell Me Who You Are, I’ll Tell You How Much You Pay. Consumers’ Fairness and Privacy Perceptions with Personalized Pricing” (2022) 148 J Bus Research 420; on the practice of personalization more generally, see UK Department for Business, Energy & Industrial Strategy, Modernising Consumer Markets: Consumer Green Paper (London: Secretary of State for Business, Energy and Industrial Strategy, 2018) at para 124 (finding that 78% of UK internet users “perceive personalisation to be unfair” and “that online platforms should be regulated to limit the extent” of personalization).
-
[29]
Ibid. Often such collection of data will include “personal information” defined as “information about an identifiable individual” in personal data protection law. See e.g. Personal Information Protection and Electronic Documents Act, SC 2000, c 5, s 2(1) “personal information”.
-
[30]
Chapdelaine APP, supra note 6 at 17–18.
-
[31]
See discussion in Part II (A), below.
-
[32]
See e.g. Canadian Human Rights Act, RSC 1985, c H-6, s 5(2) [CHRA]; Human Rights Code, RSBC 1996, c 210, ss 8(1)(a)–(b) [HRCBC].
-
[33]
HRCO, supra note 14, s 3.
-
[34]
Supra note 32.
-
[35]
Ibid, ss 2, 66(1). See also CED 4th, Fundamentals of Human Rights Law in Canada, “Constitutional Aspects of Anti-Discrimination Enactments” at §132 (2017).
-
[36]
Supra note 14.
-
[37]
See e.g. Charter of human rights and freedoms, CQLR, c C-12 [Quebec Charter]; HRCBC, supra note 32; Alberta Human Rights Act, RSA 2000, c A-25.5 [AHRA].
-
[38]
CED 4th, supra note 35 at §§ 139–40; see Dominique Payette & Virginia Torrie, “AI Governance in Canadian Banking: Fairness, Credit Models, and Equality Rights” (2020) 36:1 BFLR 5 at 19 (regarding federally regulated undertakings also being subject to provincial human rights codes for activities that fall under provincial jurisdiction).
-
[39]
AI governance regulatory regimes such as Bill C-27’s AIDA, supra note 1, supplement this existing body of anti-discrimination law by requiring anti-discrimination law compliance in the development and deployment of AI systems: see Part III, below.
-
[40]
CED 4th, supra note 35 at §§ 8–101.
-
[41]
Supra note 14, s 1; see also HRCBC, supra note 32, s 1.
-
[42]
For an overview of the manner and type of personal data collected in e-commerce, see Chapdelaine APP, supra note 6 at 9–12; “biographical core” refers to the Supreme Court of Canada’s frequent reference to the realm of intimate personal information that deserves protection under the Canadian Charter of Rights and Freedoms, ss 7–8, Part 1 of the Constitution Act, 1982, being Schedule B to the Canada Act 1982 (UK), 1982, c 11; see R v Plant, [1993] 3 SCR 281 at 293, 1993 CanLII 70 (SCC).
-
[43]
Zurich Insurance Co v Ontario (Human Rights Commission), 1992 CanLII 67 at 45 (SCC), [1992] 2 SCR 321 [Zurich Insurance], citing Ontario Human Rights Commission v Simpsons-Sears, [1985] 2 SCR 536 at 547, 1985 CanLII 18 (SCC) [Simpsons-Sears].
-
[44]
Dickason v University of Alberta, [1992] 2 SCR 1103 at 1121, 1992 CanLII 30 (SCC); Zurich Insurance, supra note 43 at 24.
-
[45]
Human rights codes protect against various forms of indirect or constructive discrimination or against discrimination by association: see e.g. HRCBC, supra note 32, s 2; Human Rights Code, CCSM c H175, s 9(1.1); HRCO, supra note 14, ss 11(1), 12 (protection from constructive discrimination and discrimination by association respectively).
-
[46]
E.g. when a service provider would explicitly preclude a protected group from accessing a facility or service, or when a supplier would include in its quote criteria constituting a prohibited ground of discrimination, such as race, gender, family status or age; McMorris v Northlander Motel, 1988 CarswellOnt 936, 9 CHRR D/5271 (Ontario Board of Inquiry); see also CED 4th, supra note 35, “Statutory Definitions of Discrimination” at § 103.
-
[47]
See e.g. Québec (Commission des droits de la personne et des droits de la jeunesse) v Bombardier Inc (Bombardier Aerospace Training Center), 2015 SCC 39 at para 32; Stewart v Elk Valley Coal Corp, 2017 SCC 30 at para 24.
-
[48]
British Columbia (Public Service Employee Relations Commission) v BCGSEU, [1999] 3 SCR 3 at paras 50–54, 1999 CanLII 652 (SCC) [Meiorin].
-
[49]
Ibid; British Columbia (Superintendent of Motor Vehicles) v British Columbia (Council of Human Rights), 1999 CanLII 646 at paras 18–19 (SCC), [1999] 3 SCR 868 [Grismer]; Moore v British Columbia (Education), 2012 SCC 61 at para 61 [Moore]; Fraser v Canada (Attorney General), 2020 SCC 28 at para 49.
-
[50]
Simpsons-Sears, supra note 43 at 558 (i.e., the ability to support their allegations, which are deemed complete and sufficient if they are believed by the court, allowing it to issue a verdict in the complainant’s favour in the absence of an answer from the respondent).
-
[51]
Moore, supra note 49 at para 33.
-
[52]
See Andy Yu, “Direct Discrimination and Indirect Discrimination: A Distinction with a Difference” (2019) 9:2 Western J Leg Studies 1 (arguing that even subsequent to the Supreme Court’s decision in Meiorin, supra note 48, there are practical benefits in recognizing the distinction between direct and indirect discrimination for the complainant, as the distinction is likely to bring more consistent analyses).
-
[53]
Moya & Le, supra note 20 at 19–20, 24 (referring to various research reports indicating racial discrimination through the use of fintech tools).
-
[54]
See e.g. HRCO, supra note 14, s 22.
-
[55]
See e.g. Zurich Insurance, supra note 43 at 16 (whereby the parties conceded that such an insurance practice constituted direct discrimination); Co-operators General Insurance Company v Alberta Human Rights Commission, 1993 ABCA 305 at paras 31–45 [Co-operators General Insurance]; see Part II(B) on the bona fide requirement for insurance contracts, below.
-
[56]
Ibid; Zurich Insurance, supra note 43; Co-operators General Insurance, supra note 55.
-
[57]
Wickham v Hong Shing Chinese Restaurant, 2018 HRTO 500 (the Toronto restaurant owner was ordered to pay $10,000); see also Taylor Auerbach, “It’s Fried Price if You Can’t Speak Chinese: Restaurant Charging English Speakers 10 Per Cent Extra per Dish”, The Daily Telegraph (20 July 2014), online: <dailytelegraph.com.au> [perma.cc/3UDC-B8XC].
-
[58]
See e.g. Commission des droits de la personne et des droits de la jeunesse c 9185-2152 Québec inc (Radio Lounge Brossard), 2015 QCCA 577 [Radio Lounge Brossard] (a patron with a service dog for the visually impaired was asked to sit in a secluded part of a dancing bar); Laidlaw Transit Ltd v Alberta (Human Rights and Citizenship Commission), 2006 ABQB 874 at paras 88–89 [Laidlaw Transit] (regarding a taxi company practice of not making a sufficient number of accessible cars available at any given point in time); see also Laverne Jacobs, “The Universality of the Human Condition: Theorizing Transportation Inequality Claims by Persons with Disabilities in Canada, 1976-2016” (2018) 7:1 Can J HR 35 (for an analysis of Canadian human rights tribunal decisions involving discrimination claims about transportation on the basis of disability).
-
[59]
Sprague v Maple Leaf Sports & Entertainment Ltd, 2019 HRTO 1617 (interim decision ordering that one of the allegations of discrimination continue); Sprague v Maple Leaf Sports & Entertainment Ltd, 2023 HRTO 1524 (motion to dismiss the case as time-barred dismissed; case moving to judgment on the merits). The applicant alleged he was the victim of discrimination based on his disability. He uses a service dog for assistance. The price charged by the respondent was significantly higher for accessible seats compared to seats that were not accessible and situated immediately next to the accessible seats.
-
[60]
Moya & Le, supra note 20 at 20.
-
[61]
McGill University Health Centre (Montreal General Hospital) v Syndicat des employés de l’Hôpital général de Montréal, 2007 SCC 4 at paras 48–49, Abella J [McGill University Health Centre]; see also Andrews v Law Society of British Columbia, [1989] 1 SCR 143 at 174–75, 1989 CanLII 2 (SCC).
-
[62]
See e.g. HRCO, supra note 14, s 22.
-
[63]
Ibid, s 15 (preferential treatment based on age, i.e. 65 and over); for a discussion of the arguments in favour and against age-related group discounts in the US, see also Alex Praschma, “New-Age Discrimination: Determining Whether Tinder Plus’s Price Is Right” (2017) 17:2 J High Tech L 372.
-
[64]
Anti-discrimination law prohibitions generally apply to services offered to the public, in contrast with private clubs, where members are subject to a selection process and criteria which are not necessarily contrary to human rights codes: see Gould v Yukon Order of Pioneers, [1996] 1 SCR 571 at para 68, 1996 CanLII 231 (SCC).
-
[65]
Regarding allegations that higher prices for women’s antiperspirant compared to that of men’s violated Quebec anti-discrimination law, see Maxwell v Unilever Canada Inc (14 February 2017), Montreal, QCCS 500-06-000846-176 (application for authorization to institute class action); Maxwell v Unilever Canada Inc et al (29 April 2019), Montreal, QCCS 500-06-000846-176 (desisting order subsequently granted). In the US, see e.g. Been v Edgewell Personal Care Company et al, (9 July 2019) 4:19 CV 02601 (Missouri Eastern Dist) (motion); Schulte v Conopco Inc, 4:19 CV 2546 RWS (Mo Dist Ct 2020); Schulte v Conopco Inc, 997 F (3d) 823 (8th Cir 2021) (confirming that the commercial practice of charging a higher price for women’s antiperspirant versus antiperspirant marketed to men did not constitute discrimination under the Missouri Merchandising Practices Act (MMPA)) [Schulte v Conopco 2021].
-
[66]
For a discussion of prices set based on real time offer and demand rather than on personal information and characteristics, see Part I, above.
-
[67]
Price versioning refers to offering different prices for different versions of a good or service (see Part I, above).
-
[68]
For a review of studies and reports conducted on the pink tax, see Kenneth A Jacobsen, “Rolling Back the Pink Tax: Dim Prospects for Eliminating Gender-Based Price Discrimination in the Sale of Consumer Goods and Services” (2018) 54:2 Cal WL Rev 241 at Part II. See also Lane Gillespie, “The Pink Tax: Latest Updates and Statistics” (27 February 2023), online <bankrate.com> [perma.cc/8FCB-6BKQ].
-
[69]
Jacobsen, supra note 68; Gillespie, supra note 68.
-
[70]
Schulte v Conopco 2021, supra note 65.
-
[71]
Ibid.
-
[72]
Moira McCormick, “Why Women Pay More than Men For Products” (20 July 2023), online: <blog.blackcurve.com> [perma.cc/P6UG-BTVV] (citing marketing companies’ knowledge that women on average, are willing to pay more for certain products than men, as one reason explaining the “pink tax”).
-
[73]
Gillespie, supra note 68.
-
[74]
Maclean v The Barking Frog, 2013 HRTO 630 [Maclean] (case dismissed as having little likelihood of success where male claimant unsuccessfully argued that being charged a higher fee than women by the respondent restaurant and bar “was perpetuating a belief in society than men are less worthy than women” and made him feel unwelcomed, at paras 7–8); Roncali v Cabana Pool Bar, 2020 HRTO 275 [Roncali] (male claimant’s allegation of discrimination on the basis of women being charged lower rates than men at respondent’s bar unsuccessful as tribunal saw no evidence of such practice perpetrating negative stereotypes against men); Stopps v Just Ladies Fitness (Metrotown) Ltd, 2006 BCHRT 557 at paras 80–118 [Just Ladies] (tribunal not satisfied that the male complainant had suffered any adverse impact by being ineligible to join a women-only fitness club, given other co-ed options available that he had not even sought).
-
[75]
Zurich Insurance, supra note 43 at 16 (the parties conceded that charging different rates to men under the age of 25 versus women of the same age was prima facie discrimination under the HRCO, supra note 14).
-
[76]
Moore, supra note 49 at para 33.
-
[77]
McGill University Health Centre, supra note 61 at para 48; Shell v Whistler Mountain Resort Ltd Partnership, 2009 BCHRT 424 at para 24 (claim of age limit to mature student discount not prohibited discrimination under relevant human rights code as pricing policy is “not based on arbitrary or demeaning stereotypes and differential treatment through the discounts does not … promote the notion that older students are less capable or less deserving of respect”).
-
[78]
See Maclean, supra note 74; Roncali, supra note 74; Just Ladies, supra note 74; see also Melody Jahanzadeh, “Does ‘Reverse Discrimination’ Exist? According to these Cases, Probably Not” (9 February 2021), online: <rubinthomlinson.com> [perma.cc/J6AG-36U4]; see however Koire v Metro Car Wash, 40 Cal (3d) 24 (Cal Sup Ct 1985) (charging a lower price for women versus men for car wash services was found to be in violation of the Unruh Act); see also Heidi C Paulson, “Ladies’ Night Discounts: Should We Bar Them or Promote Them?” (1991) 32:2 Boston College L Rev 487 (arguing that legislatures and courts should move toward banning ladies night practices as they encourage paternalism toward women and maintain stereotypes about men and women).
-
[79]
Such as the psychological and physical benefits to women patrons of a women-only fitness club: Just Ladies, supra note 74 at paras 62–70, 132, 144, 147; Part II(B), below.
-
[80]
See Part I, above.
-
[81]
Moya & Le, supra note 20 at 6.
-
[82]
Meiorin, supra note 48 at para 54; Grismer, supra note 49 at paras 20–22; see also CED 4th, supra note 35 at §§ 107, 111, on the requirements for an employment or goods/services to the public standard to comply with human rights codes.
-
[83]
Radio Lounge Brossard, supra note 58 at para 45; see also CED 4th, supra note 35 at §§ 107, 112.
-
[84]
Zurich Insurance, supra note 43 at 28; see discussion in this section, below.
-
[85]
Meiorin, supra note 48 at para 54; Grismer, supra note 49 at paras 20–21.
-
[86]
Meiorin, supra note 48 at para 54.
-
[87]
Just Ladies, supra note 74 at paras 119–150 (accommodating the male complainant would cause undue hardship to the women members of the fitness club and to the owner; the tribunal relied on expert evidence that showed the significant benefits to women patrons who would otherwise not frequent a co-ed fitness centre and would be without alternatives, as they want to be free from the male gaze, given poor body image issues. Additionally, the fitness club chain owner would suffer undue hardship given that the club would likely lose a significant number of female members and breach some of its leases if it had to become a co-ed fitness centre).
-
[88]
See supra note 44.
-
[89]
Meiorin, supra note 48 at para 59.
-
[90]
See discussion in Part I, above.
-
[91]
See however Borgesius, supra note 8 at 411, 413–18 (making the argument, in a European context, that APP having the main and proven goal of increasing the profit of businesses could meet the legitimate interest and proportionality requirements for indirect discrimination under EU anti-discrimination law. Borgesius also argues that, because the primary goal of APP is to increase profit, and not to discriminate on a prohibited ground, that this would not constitute direct discrimination).
-
[92]
Meiorin, supra note 48 at para 60 (referring to a standard being motivated by a discriminatory animus as not meeting the bona fide requirement exception).
-
[93]
Meiorin, supra note 48 at paras 62–65.
-
[94]
Ibid at para 62; Council of Canadians with Disabilities v VIA Rail Canada Inc, 2007 SCC 15 at paras 122, 130.
-
[95]
Meiorin, supra note 48 at para 79; Grismer, supra note 49 at paras 41–42; Laidlaw Transit, supra note 58 at paras 96–107 (complainant alleged discrimination on the basis of physical disability given that respondent taxi companies failed to provide sufficient accessible taxi services, and where the court found that the respondents did not meet the required burden of proof).
-
[96]
Insurance contracts are subject to exceptions allowing discrimination under some human rights codes. See e.g. HRCO, supra note 14, s 22.
-
[97]
Supra note 43 (followed in Co-operators General Insurance, supra note 55, and in Robert Farquhar v Bank of Nova Scotia, 2019 HRTO 410 [Bank of Nova Scotia]).
-
[98]
HRCO, supra note 14, s 21, now s 22 (which now refers to “disability” in lieu of “handicap”). The parties conceded that the insurance practice of charging higher automobile insurance rates for single men under the age of 25 was discrimination under the HRCO: Zurich Insurance, supra note 43 at 16.
-
[99]
Zurich Insurance, supra note 43 at 25–28.
-
[100]
Ibid at 24.
-
[101]
Ibid at 27.
-
[102]
Ibid. See however dissenting reasons of Justice L’Heureux-Dubé, ibid at 40–70, who proposed a different test whereby the insurance provider needs to show a causal connection between the group discriminated upon and the risk assessed, not a mere correlation.
-
[103]
Ibid at 35. There being no evidence of bad faith on the part of the insurance company, the analysis focused on whether the discriminatory practice was reasonable.
-
[104]
Ibid.
-
[105]
Ibid at 34–35.
-
[106]
Ibid at 67–69 (dissenting reasons L’Heureux-Dubé J), 70–82 (dissenting reasons McLachlin J).
-
[107]
Ibid at 72 (dissenting reasons McLachlin J).
-
[108]
Ibid at 73 (dissenting reasons McLachlin J).
-
[109]
Ibid.
-
[110]
See Part III, below.
-
[111]
Ibid (regarding AIDA, supra note 1 and Proposed Amendments to AIDA, supra note 1).
-
[112]
See Brooke Smith, “Will There Ever Be Equality in Auto Insurance Rates?” (24 December 2021), online: <canadianunderwriter.ca> [perma.cc/6JLF-RNZ3] (on various US states banning gender-based discrimination for automobile insurance); gender-based discrimination in insurance contracts is illegal in the EU: Association belge des Consommateurs Test-Achats ASBL v Conseil des ministres, C-236/09, [2011] CJEU I-800; see also European Commission, News Release, “EU Rules on Gender-Neutral Pricing in Insurance Industry Enter Into Force” (20 December 2012).
-
[113]
Leah Golob, “How Do Insurance Companies Set Rates for Gender ‘X’ Drivers?” (17 January 2022), online: <lowestrates.ca> [perma.cc/6B9T-BTSS].
-
[114]
Co-operators General Insurance, supra note 55 at paras 79–112 (while acknowledging evidence about genderless automobile insurance rates in various US jurisdictions, upholding the status quo of gender-based discrimination based on the principle of fairness overall); Bank of Nova Scotia, supra note 97.
-
[115]
Zurich Insurance, supra note 43 at 39.
-
[116]
HRCO, supra note 14, s 22 (formerly s 21); Quebec Charter, supra note 37, s 20.1.
-
[117]
See Part II(A).
-
[118]
As defined in personal data protection law: see supra note 29.
-
[119]
See supra note 9 (empirical research and other academic work discussing biased outputs in algorithms and AI systems).
-
[120]
For a definition of “biased outputs” as outputs contravening the CHRA, see AIDA, supra note 1, cls 5(1), 8; CHRA, supra note 32, s 8; Proposed Amendments to AIDA, supra note 1, cls 9(1)(b), 10(1)(b), 11(1)(b).
-
[121]
AIDA, supra note 1, cl 5(1) “high-impact system”; Proposed Amendments to AIDA, supra note 1, Schedule 2, “High-Impact Systems – Uses” Class 2(b).
-
[122]
Ibid.
-
[123]
AIDA, supra note 1, cls 13–21, 29–40, as modified by Proposed Amendments to AIDA, supra note 1.
-
[124]
AIDA, supra note 1, cl 8; Proposed Amendments to AIDA, supra note 1, cls 9(1)(b), 10(1)(b), 11(1)(b).
-
[125]
Scholars and policy makers have made several recommendations to address this issue. Regarding the regulation of machine-learning data sets and AI outputs, see e.g. Cofone, supra note 9; Sandra Wachter, Brent Mittelstadt & Chris Russell, “Why Automated Fairness Cannot Be Bridging the Gap Between EU Non-Discrimination Law And AI” (2021) 41 Computer L & Security Rev 1 (for a European Union perspective on the difficulties putting the proper tools in place to bridge AI system compliance with EU anti-discrimination law).
-
[126]
Proposed Amendments to AIDA, supra note 1, cls 12(3), 12(5). For high-impact systems, such as “policies and procedures respecting the management of risks relating to the system,” see ibid.
-
[127]
AIDA, supra note 1, cls 8, 14; Proposed Amendments to AIDA, supra note 1, cl 12(3) (requirement to maintain a written accountability framework for high-impact systems), cl 8(1)(a) (obligation to make information available about certain AI systems on publicly available websites).
-
[128]
See discussion in Part II(A), above.
-
[129]
See discussion in Part II(B), above.
-
[130]
See supra note 6.