{"id":552,"date":"2020-04-26T17:09:05","date_gmt":"2020-04-26T14:09:05","guid":{"rendered":"https:\/\/persona.qcri.org\/blog\/?p=552"},"modified":"2021-03-18T22:14:18","modified_gmt":"2021-03-18T19:14:18","slug":"analyzing-demographic-bias-in-artificially-generated-facial-pictures","status":"publish","type":"post","link":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/","title":{"rendered":"Analyzing Demographic Bias in Artificially Generated Facial Pictures"},"content":{"rendered":"<p><strong>Artificial generation of facial images is increasingly popular, with machine learning achieving photo-realistic results.<\/strong><\/p>\n<p>Yet, there is a concern that the generated images might not fairly represent all demographic groups. This has implications for persona development when approaching the goal of generating the facial pictures for the persona profiles automatically.<\/p>\n<figure id=\"attachment_557\" aria-describedby=\"caption-attachment-557\" style=\"width: 754px\" class=\"wp-caption aligncenter\"><img data-recalc-dims=\"1\" loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-557\" src=\"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?resize=754%2C623&#038;ssl=1\" alt=\"Analyzing Demographic Bias in Artificially Generated Facial Pictures\" width=\"754\" height=\"623\" srcset=\"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?w=754&amp;ssl=1 754w, https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?resize=300%2C248&amp;ssl=1 300w\" sizes=\"auto, (max-width: 754px) 100vw, 754px\" \/><figcaption id=\"caption-attachment-557\" class=\"wp-caption-text\"><a href=\"http:\/\/www.bernardjjansen.com\/uploads\/2\/4\/1\/8\/24188166\/3334480.3382791.pdf\" target=\"_blank\" rel=\"noopener noreferrer\">Analyzing Demographic Bias in Artificially Generated Facial Pictures<\/a><\/figcaption><\/figure>\n<p><span style=\"font-family: inherit;\">In research led by\u00a0<a href=\"https:\/\/jonisalminen.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">Joni Salminen<\/a>, we use a state-of-the-art method to generate 10,000 facial images and observe that the generated images are skewed towards young people, especially white women.<\/span><\/p>\n<p><span style=\"font-family: inherit;\"><!--more--><\/span><\/p>\n<p><span style=\"font-family: inherit;\">We provide recommendations to reduce demographic bias in artificial images.<\/span><\/p>\n<h3>Read Full Research<\/h3>\n<p><span style=\"font-family: inherit;\">Salminen, J., Jung, S.G., Chowdhury. S., and\u00a0Jansen, B. J. (2020) <a href=\"http:\/\/www.bernardjjansen.com\/uploads\/2\/4\/1\/8\/24188166\/3334480.3382791.pdf\" target=\"_blank\" rel=\"noopener noreferrer\">Analyzing Demographic Bias in Artificially Generated Facial Pictures<\/a>. ACM CHI Conference on Human Factors in Computing Systems (CHI&#8217;20) (Extended Abstract), Honolulu, HI, USA. 25\u201330 April, 1-8.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Artificial generation of facial images is increasingly popular, with machine learning achieving photo-realistic results. Yet, there is a concern that the generated images might not fairly represent all demographic groups. This has implications for persona development when approaching the goal of generating the facial pictures for the persona profiles automatically. In research led by\u00a0Joni Salminen, [&hellip;]<\/p>\n","protected":false},"author":3,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[41,16,37,59,7,2],"tags":[83,456,84,82,455,51],"class_list":["post-552","post","type-post","status-publish","format-standard","hentry","category-chi","category-persona-creation","category-persona-design","category-persona-development","category-persona-research","category-personas","tag-bias","tag-confusion-photos","tag-demographics","tag-fake-pictures","tag-persona-faces","tag-persona-pictures"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v26.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog\" \/>\n<meta property=\"og:description\" content=\"Artificial generation of facial images is increasingly popular, with machine learning achieving photo-realistic results. Yet, there is a concern that the generated images might not fairly represent all demographic groups. This has implications for persona development when approaching the goal of generating the facial pictures for the persona profiles automatically. In research led by\u00a0Joni Salminen, [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\" \/>\n<meta property=\"og:site_name\" content=\"The Persona Blog\" \/>\n<meta property=\"article:published_time\" content=\"2020-04-26T14:09:05+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-03-18T19:14:18+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg\" \/>\n<meta name=\"author\" content=\"Jim Jansen\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Jim Jansen\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"1 minute\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\"},\"author\":{\"name\":\"Jim Jansen\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/27d3999853d82c7a10b08504351811f5\"},\"headline\":\"Analyzing Demographic Bias in Artificially Generated Facial Pictures\",\"datePublished\":\"2020-04-26T14:09:05+00:00\",\"dateModified\":\"2021-03-18T19:14:18+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\"},\"wordCount\":151,\"publisher\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/#organization\"},\"image\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg\",\"keywords\":[\"bias\",\"confusion photos\",\"demographics\",\"fake pictures\",\"persona faces\",\"persona pictures\"],\"articleSection\":[\"CHI\",\"Persona Creation\",\"Persona Design\",\"Persona Development\",\"Persona Research\",\"Personas\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\",\"url\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\",\"name\":\"Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog\",\"isPartOf\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg\",\"datePublished\":\"2020-04-26T14:09:05+00:00\",\"dateModified\":\"2021-03-18T19:14:18+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage\",\"url\":\"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?fit=754%2C623&ssl=1\",\"contentUrl\":\"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?fit=754%2C623&ssl=1\",\"width\":754,\"height\":623,\"caption\":\"Analyzing Demographic Bias in Artificially Generated Facial Pictures\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/persona.qcri.org\/blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Analyzing Demographic Bias in Artificially Generated Facial Pictures\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#website\",\"url\":\"https:\/\/persona.qcri.org\/blog\/\",\"name\":\"The Persona Blog\",\"description\":\"All Things Personas!\",\"publisher\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/persona.qcri.org\/blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#organization\",\"name\":\"Persona Blog\",\"url\":\"https:\/\/persona.qcri.org\/blog\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/i1.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/APG-2020.png?fit=490%2C300&ssl=1\",\"contentUrl\":\"https:\/\/i1.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/APG-2020.png?fit=490%2C300&ssl=1\",\"width\":490,\"height\":300,\"caption\":\"Persona Blog\"},\"image\":{\"@id\":\"https:\/\/persona.qcri.org\/blog\/#\/schema\/logo\/image\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/27d3999853d82c7a10b08504351811f5\",\"name\":\"Jim Jansen\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/332c40ecd52b5094e0b04c262ec09399e1b96fbc0d8b633d350174144289e63c?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/332c40ecd52b5094e0b04c262ec09399e1b96fbc0d8b633d350174144289e63c?s=96&d=mm&r=g\",\"caption\":\"Jim Jansen\"},\"description\":\"Dr. Jansen is a Principal Scientist in the social computing group of the Qatar Computing Research Institute, and a professor with the College of Science and Engineering, Hamad bin Khalifa University, and an adjunct professor with the College of Information Sciences and Technology at The Pennsylvania State University. He is a graduate of West Point and has a Ph.D. in computer science from Texas A&amp;M University, along with master degrees from Texas A&amp;M (computer science) and Troy State (international relations). Dr. Jim Jansen served in the U.S. Army as an Infantry enlisted soldier and communication commissioned officer.\",\"sameAs\":[\"http:\/\/www.bernardjjansen.com\/\"],\"url\":\"https:\/\/persona.qcri.org\/blog\/author\/jim-jansen\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/","og_locale":"en_US","og_type":"article","og_title":"Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog","og_description":"Artificial generation of facial images is increasingly popular, with machine learning achieving photo-realistic results. Yet, there is a concern that the generated images might not fairly represent all demographic groups. This has implications for persona development when approaching the goal of generating the facial pictures for the persona profiles automatically. In research led by\u00a0Joni Salminen, [&hellip;]","og_url":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/","og_site_name":"The Persona Blog","article_published_time":"2020-04-26T14:09:05+00:00","article_modified_time":"2021-03-18T19:14:18+00:00","og_image":[{"url":"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg","type":"","width":"","height":""}],"author":"Jim Jansen","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Jim Jansen","Est. reading time":"1 minute"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#article","isPartOf":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/"},"author":{"name":"Jim Jansen","@id":"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/27d3999853d82c7a10b08504351811f5"},"headline":"Analyzing Demographic Bias in Artificially Generated Facial Pictures","datePublished":"2020-04-26T14:09:05+00:00","dateModified":"2021-03-18T19:14:18+00:00","mainEntityOfPage":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/"},"wordCount":151,"publisher":{"@id":"https:\/\/persona.qcri.org\/blog\/#organization"},"image":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage"},"thumbnailUrl":"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg","keywords":["bias","confusion photos","demographics","fake pictures","persona faces","persona pictures"],"articleSection":["CHI","Persona Creation","Persona Design","Persona Development","Persona Research","Personas"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/","url":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/","name":"Analyzing Demographic Bias in Artificially Generated Facial Pictures &#8211; The Persona Blog","isPartOf":{"@id":"https:\/\/persona.qcri.org\/blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage"},"image":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage"},"thumbnailUrl":"https:\/\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg","datePublished":"2020-04-26T14:09:05+00:00","dateModified":"2021-03-18T19:14:18+00:00","breadcrumb":{"@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#primaryimage","url":"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?fit=754%2C623&ssl=1","contentUrl":"https:\/\/i0.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/fake_pictures.jpg?fit=754%2C623&ssl=1","width":754,"height":623,"caption":"Analyzing Demographic Bias in Artificially Generated Facial Pictures"},{"@type":"BreadcrumbList","@id":"https:\/\/persona.qcri.org\/blog\/analyzing-demographic-bias-in-artificially-generated-facial-pictures\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/persona.qcri.org\/blog\/"},{"@type":"ListItem","position":2,"name":"Analyzing Demographic Bias in Artificially Generated Facial Pictures"}]},{"@type":"WebSite","@id":"https:\/\/persona.qcri.org\/blog\/#website","url":"https:\/\/persona.qcri.org\/blog\/","name":"The Persona Blog","description":"All Things Personas!","publisher":{"@id":"https:\/\/persona.qcri.org\/blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/persona.qcri.org\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/persona.qcri.org\/blog\/#organization","name":"Persona Blog","url":"https:\/\/persona.qcri.org\/blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/persona.qcri.org\/blog\/#\/schema\/logo\/image\/","url":"https:\/\/i1.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/APG-2020.png?fit=490%2C300&ssl=1","contentUrl":"https:\/\/i1.wp.com\/persona.qcri.org\/blog\/wp-content\/uploads\/2020\/04\/APG-2020.png?fit=490%2C300&ssl=1","width":490,"height":300,"caption":"Persona Blog"},"image":{"@id":"https:\/\/persona.qcri.org\/blog\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/27d3999853d82c7a10b08504351811f5","name":"Jim Jansen","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/persona.qcri.org\/blog\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/332c40ecd52b5094e0b04c262ec09399e1b96fbc0d8b633d350174144289e63c?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/332c40ecd52b5094e0b04c262ec09399e1b96fbc0d8b633d350174144289e63c?s=96&d=mm&r=g","caption":"Jim Jansen"},"description":"Dr. Jansen is a Principal Scientist in the social computing group of the Qatar Computing Research Institute, and a professor with the College of Science and Engineering, Hamad bin Khalifa University, and an adjunct professor with the College of Information Sciences and Technology at The Pennsylvania State University. He is a graduate of West Point and has a Ph.D. in computer science from Texas A&amp;M University, along with master degrees from Texas A&amp;M (computer science) and Troy State (international relations). Dr. Jim Jansen served in the U.S. Army as an Infantry enlisted soldier and communication commissioned officer.","sameAs":["http:\/\/www.bernardjjansen.com\/"],"url":"https:\/\/persona.qcri.org\/blog\/author\/jim-jansen\/"}]}},"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"jetpack-related-posts":[],"_links":{"self":[{"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/posts\/552","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/users\/3"}],"replies":[{"embeddable":true,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/comments?post=552"}],"version-history":[{"count":4,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/posts\/552\/revisions"}],"predecessor-version":[{"id":565,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/posts\/552\/revisions\/565"}],"wp:attachment":[{"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/media?parent=552"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/categories?post=552"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/persona.qcri.org\/blog\/wp-json\/wp\/v2\/tags?post=552"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}