Monitoring algorithmic fairness

Henzinger TA, Karimi M, Kueffner K, Mallik K. 2023. Monitoring algorithmic fairness. Computer Aided Verification. CAV: Computer Aided Verification, LNCS, vol. 13965, 358–382.

Download
OA 2023_LNCS_CAV_HenzingerT.pdf 647.76 KB

Conference Paper | Published | English
Series Title
LNCS
Abstract
Machine-learned systems are in widespread use for making decisions about humans, and it is important that they are fair, i.e., not biased against individuals based on sensitive attributes. We present runtime verification of algorithmic fairness for systems whose models are unknown, but are assumed to have a Markov chain structure. We introduce a specification language that can model many common algorithmic fairness properties, such as demographic parity, equal opportunity, and social burden. We build monitors that observe a long sequence of events as generated by a given system, and output, after each observation, a quantitative estimate of how fair or biased the system was on that run until that point in time. The estimate is proven to be correct modulo a variable error bound and a given confidence level, where the error bound gets tighter as the observed sequence gets longer. Our monitors are of two types, and use, respectively, frequentist and Bayesian statistical inference techniques. While the frequentist monitors compute estimates that are objectively correct with respect to the ground truth, the Bayesian monitors compute estimates that are correct subject to a given prior belief about the system’s model. Using a prototype implementation, we show how we can monitor if a bank is fair in giving loans to applicants from different social backgrounds, and if a college is fair in admitting students while maintaining a reasonable financial burden on the society. Although they exhibit different theoretical complexities in certain cases, in our experiments, both frequentist and Bayesian monitors took less than a millisecond to update their verdicts after each observation.
Publishing Year
Date Published
2023-07-18
Proceedings Title
Computer Aided Verification
Acknowledgement
This work is supported by the European Research Council under Grant No.: ERC-2020-AdG101020093.
Volume
13965
Page
358–382
Conference
CAV: Computer Aided Verification
Conference Location
Paris, France
Conference Date
2023-07-17 – 2023-07-22
ISSN
eISSN
IST-REx-ID

Cite this

Henzinger TA, Karimi M, Kueffner K, Mallik K. Monitoring algorithmic fairness. In: Computer Aided Verification. Vol 13965. Springer Nature; 2023:358–382. doi:10.1007/978-3-031-37703-7_17
Henzinger, T. A., Karimi, M., Kueffner, K., & Mallik, K. (2023). Monitoring algorithmic fairness. In Computer Aided Verification (Vol. 13965, pp. 358–382). Paris, France: Springer Nature. https://doi.org/10.1007/978-3-031-37703-7_17
Henzinger, Thomas A, Mahyar Karimi, Konstantin Kueffner, and Kaushik Mallik. “Monitoring Algorithmic Fairness.” In Computer Aided Verification, 13965:358–382. Springer Nature, 2023. https://doi.org/10.1007/978-3-031-37703-7_17.
T. A. Henzinger, M. Karimi, K. Kueffner, and K. Mallik, “Monitoring algorithmic fairness,” in Computer Aided Verification, Paris, France, 2023, vol. 13965, pp. 358–382.
Henzinger TA, Karimi M, Kueffner K, Mallik K. 2023. Monitoring algorithmic fairness. Computer Aided Verification. CAV: Computer Aided Verification, LNCS, vol. 13965, 358–382.
Henzinger, Thomas A., et al. “Monitoring Algorithmic Fairness.” Computer Aided Verification, vol. 13965, Springer Nature, 2023, pp. 358–382, doi:10.1007/978-3-031-37703-7_17.
All files available under the following license(s):
Creative Commons Attribution 4.0 International Public License (CC-BY 4.0):
Main File(s)
File Name
Access Level
OA Open Access
Date Uploaded
2023-07-31
MD5 Checksum
ccaf94bf7d658ba012c016e11869b54c


Export

Marked Publications

Open Data ISTA Research Explorer

Sources

arXiv 2305.15979

Search this title in

Google Scholar
ISBN Search