{"entities":{"Q4914":{"pageid":28048,"ns":120,"title":"Item:Q4914","lastrevid":89970,"modified":"2020-07-29T07:44:14Z","type":"item","id":"Q4914","labels":{"en":{"language":"en","value":"Fawkes: Protecting Privacy against Unauthorized Deep Learning Models"}},"descriptions":{"en":{"language":"en","value":"article presented at the 29th USENIX Security Symposium  in July 2020"}},"aliases":{},"claims":{"P3":[{"mainsnak":{"snaktype":"value","property":"P3","datavalue":{"value":{"entity-type":"item","numeric-id":210,"id":"Q210"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$37F2DC59-FEC1-4031-927A-B4F02D92725C","rank":"normal"}],"P15":[{"mainsnak":{"snaktype":"value","property":"P15","datavalue":{"value":"http://people.cs.uchicago.edu/%7Eravenben/publications/abstracts/fawkes-usenix20.html","type":"string"},"datatype":"url"},"type":"statement","id":"Q4914$d79720cc-421a-b75d-4625-08845e80db42","rank":"normal"}],"P100":[{"mainsnak":{"snaktype":"value","property":"P100","datavalue":{"value":{"time":"+2020-07-00T00:00:00Z","timezone":0,"before":0,"after":0,"precision":10,"calendarmodel":"http://www.wikidata.org/entity/Q1985727"},"type":"time"},"datatype":"time"},"type":"statement","id":"Q4914$3809ffb1-40e0-afbc-b8f8-f89e9553df7b","rank":"normal"}],"P110":[{"mainsnak":{"snaktype":"value","property":"P110","datavalue":{"value":{"entity-type":"item","numeric-id":2594,"id":"Q2594"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$6cc5b564-4ddc-bb44-fff8-663b4cfd4b35","rank":"normal"},{"mainsnak":{"snaktype":"value","property":"P110","datavalue":{"value":{"entity-type":"item","numeric-id":3826,"id":"Q3826"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$2980295e-419d-0cdc-c62c-ee3613ca47a2","rank":"normal"},{"mainsnak":{"snaktype":"value","property":"P110","datavalue":{"value":{"entity-type":"item","numeric-id":4915,"id":"Q4915"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$be3fcfe9-4a98-76f1-a512-3b16d80aa147","rank":"normal"},{"mainsnak":{"snaktype":"value","property":"P110","datavalue":{"value":{"entity-type":"item","numeric-id":3825,"id":"Q3825"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$b219b097-4432-683f-4dec-88739f74b7d4","rank":"normal"},{"mainsnak":{"snaktype":"value","property":"P110","datavalue":{"value":{"entity-type":"item","numeric-id":2984,"id":"Q2984"},"type":"wikibase-entityid"},"datatype":"wikibase-item"},"type":"statement","id":"Q4914$2b27a87a-4436-3c6d-9cdc-2f0b67dbcca4","rank":"normal"}],"P203":[{"mainsnak":{"snaktype":"value","property":"P203","datavalue":{"value":"Fawkes, a system that helps individuals inoculate their images against unauthorized facial recognition models. Fawkes achieves this by helping users add imperceptible pixel-level changes (we call them\u201ccloaks\u201d) to their own photos before releasing them.","type":"string"},"datatype":"string"},"type":"statement","id":"Q4914$4290429f-48c7-6085-aee2-7b2a02fa1b50","rank":"normal"}]},"sitelinks":{}}}}