Paper Abstract and Keywords |
Presentation |
2020-03-11 15:10
Fairness Causes Vulnerability to Adversarial Attacks Koki Wataoka, Takashi Matsubara, Kuniaki Uehara (Kobe Univ.) IBISML2019-48 |
Abstract |
(in Japanese) |
(See Japanese page) |
(in English) |
When using machine learning models in society, it is essential to be ensure classifiers are fair to race and gender. In recent yeas, many methods have been proposed to make classifiers fair. However, the security of fair classifiers has been rarely discussed. In the field of machine learning, there is an attack method called adversarial attacks that reduce the accuracy of classifiers. In this paper, fair classifiers are vulnerable to adversarial attacks. Our experiment has shown that fair classifiers are less robust against adversarial attacks than usual classifiers and hence worse classification accuracy and part of fairness performance.
Key words Fairness, Adversarial Attacks, Adversarial Training |
Keyword |
(in Japanese) |
(See Japanese page) |
(in English) |
Fairness / Adversarial Attacks / Adversarial Training / / / / / |
Reference Info. |
IEICE Tech. Rep., vol. 119, no. 476, IBISML2019-48, pp. 101-105, March 2020. |
Paper # |
IBISML2019-48 |
Date of Issue |
2020-03-03 (IBISML) |
ISSN |
Online edition: ISSN 2432-6380 |
Copyright and reproduction |
All rights are reserved and no part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopy, recording, or any information storage and retrieval system, without permission in writing from the publisher. Notwithstanding, instructors are permitted to photocopy isolated articles for noncommercial classroom use without fee. (License No.: 10GA0019/12GB0052/13GB0056/17GB0034/18GB0034) |
Download PDF |
IBISML2019-48 |
Conference Information |
Committee |
IBISML |
Conference Date |
2020-03-10 - 2020-03-11 |
Place (in Japanese) |
(See Japanese page) |
Place (in English) |
Kyoto University |
Topics (in Japanese) |
(See Japanese page) |
Topics (in English) |
Machine learning, etc. |
Paper Information |
Registration To |
IBISML |
Conference Code |
2020-03-IBISML |
Language |
Japanese |
Title (in Japanese) |
(See Japanese page) |
Sub Title (in Japanese) |
(See Japanese page) |
Title (in English) |
Fairness Causes Vulnerability to Adversarial Attacks |
Sub Title (in English) |
|
Keyword(1) |
Fairness |
Keyword(2) |
Adversarial Attacks |
Keyword(3) |
Adversarial Training |
Keyword(4) |
|
Keyword(5) |
|
Keyword(6) |
|
Keyword(7) |
|
Keyword(8) |
|
1st Author's Name |
Koki Wataoka |
1st Author's Affiliation |
Kobe University (Kobe Univ.) |
2nd Author's Name |
Takashi Matsubara |
2nd Author's Affiliation |
Kobe University (Kobe Univ.) |
3rd Author's Name |
Kuniaki Uehara |
3rd Author's Affiliation |
Kobe University (Kobe Univ.) |
4th Author's Name |
|
4th Author's Affiliation |
() |
5th Author's Name |
|
5th Author's Affiliation |
() |
6th Author's Name |
|
6th Author's Affiliation |
() |
7th Author's Name |
|
7th Author's Affiliation |
() |
8th Author's Name |
|
8th Author's Affiliation |
() |
9th Author's Name |
|
9th Author's Affiliation |
() |
10th Author's Name |
|
10th Author's Affiliation |
() |
11th Author's Name |
|
11th Author's Affiliation |
() |
12th Author's Name |
|
12th Author's Affiliation |
() |
13th Author's Name |
|
13th Author's Affiliation |
() |
14th Author's Name |
|
14th Author's Affiliation |
() |
15th Author's Name |
|
15th Author's Affiliation |
() |
16th Author's Name |
|
16th Author's Affiliation |
() |
17th Author's Name |
|
17th Author's Affiliation |
() |
18th Author's Name |
|
18th Author's Affiliation |
() |
19th Author's Name |
|
19th Author's Affiliation |
() |
20th Author's Name |
|
20th Author's Affiliation |
() |
Speaker |
Author-1 |
Date Time |
2020-03-11 15:10:00 |
Presentation Time |
25 minutes |
Registration for |
IBISML |
Paper # |
IBISML2019-48 |
Volume (vol) |
vol.119 |
Number (no) |
no.476 |
Page |
pp.101-105 |
#Pages |
5 |
Date of Issue |
2020-03-03 (IBISML) |
|