{"id":669,"date":"2025-04-07T15:50:49","date_gmt":"2025-04-07T15:50:49","guid":{"rendered":"https:\/\/pacific.ai\/staging\/3667\/?p=669"},"modified":"2026-02-19T10:38:02","modified_gmt":"2026-02-19T10:38:02","slug":"testing-for-bias-of-large-language-models-in-clinical-applications","status":"publish","type":"post","link":"https:\/\/pacific.ai\/staging\/3667\/testing-for-bias-of-large-language-models-in-clinical-applications\/","title":{"rendered":"Testing for Bias of Large Language Models in Clinical Applications"},"content":{"rendered":"<div id=\"bsf_rt_marker\"><\/div>\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n    <iframe loading=\"lazy\" title=\"Testing for Bias of Large Language Models in Clinical Applications\" width=\"580\" height=\"326\" src=\"https:\/\/www.youtube.com\/embed\/A-zWcH5VKQo?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe>\n    <\/div><\/figure>\n\n\n<h2>FAQ<\/h2>\n<p><strong>How is bias measured in clinical LLMs?<\/strong><\/p>\n<p>Bias is evaluated using clinical vignettes and &#8220;counterfactual&#8221; variations (e.g., changing patient attributes) to observe differential responses, allowing detection of both performance disparities and fairness issues across demographic groups.<\/p>\n<p><strong>How common are demographic biases in healthcare LLM outputs?<\/strong><\/p>\n<p>Systematic reviews reveal pervasive demographic bias, especially across race, ethnicity, gender, age, and disability, affecting tasks like trial matching and question answering\u2014suggesting biased care recommendations.<\/p>\n<p><strong>What types of bias do LLMs exhibit in clinical decision-making?<\/strong><\/p>\n<p>Bias can manifest as allocative harm (e.g., fewer diagnostic tests for certain groups), representational bias (using stereotypes), and performance disparities\u2014like lower accuracy or recommendation quality for some demographics.<\/p>\n<p><strong>What methods exist to mitigate clinical LLM bias?<\/strong><\/p>\n<p>Techniques include prompt engineering, fine-tuning, contrastive learning frameworks like EquityGuard, and multi-agent chain-of-thought reasoning\u2014all shown to reduce bias in medical question answering and trial matching tasks.<\/p>\n<p><strong>Who should conduct bias testing of LLMs before clinical use?<\/strong><\/p>\n<p>Bias testing should be done by developers and healthcare institutions using structured protocols and benchmarks like CLIMB, DiversityMedQA, or CPV datasets to ensure robust validation across diverse patient populations.<\/p>\n\n\n<script type=\"application\/ld+json\">\n{\n  \"@context\": \"https:\/\/schema.org\",\n  \"@type\": \"FAQPage\",\n  \"mainEntity\": [\n    {\n      \"@type\": \"Question\",\n      \"name\": \"How is bias measured in clinical LLMs?\",\n      \"acceptedAnswer\": {\n        \"@type\": \"Answer\",\n        \"text\": \"Bias is evaluated using clinical vignettes and \u201ccounterfactual\u201d variations (e.g., changing patient attributes) to observe differential responses, allowing detection of both performance disparities and fairness issues across demographic groups.\"\n      }\n    },\n    {\n      \"@type\": \"Question\",\n      \"name\": \"How common are demographic biases in healthcare LLM outputs?\",\n      \"acceptedAnswer\": {\n        \"@type\": \"Answer\",\n        \"text\": \"Systematic reviews reveal pervasive demographic bias, especially across race, ethnicity, gender, age, and disability, affecting tasks like trial matching and question answering\u2014suggesting biased care recommendations.\"\n      }\n    },\n    {\n      \"@type\": \"Question\",\n      \"name\": \"What types of bias do LLMs exhibit in clinical decision-making?\",\n      \"acceptedAnswer\": {\n        \"@type\": \"Answer\",\n        \"text\": \"Bias can manifest as allocative harm (e.g., fewer diagnostic tests for certain groups), representational bias (using stereotypes), and performance disparities\u2014like lower accuracy or recommendation quality for some demographics.\"\n      }\n    },\n    {\n      \"@type\": \"Question\",\n      \"name\": \"What methods exist to mitigate clinical LLM bias?\",\n      \"acceptedAnswer\": {\n        \"@type\": \"Answer\",\n        \"text\": \"Techniques include prompt engineering, fine-tuning, contrastive learning frameworks like EquityGuard, and multi-agent chain-of-thought reasoning\u2014all shown to reduce bias in medical question answering and trial matching tasks.\"\n      }\n    },\n    {\n      \"@type\": \"Question\",\n      \"name\": \"Who should conduct bias testing of LLMs before clinical use?\",\n      \"acceptedAnswer\": {\n        \"@type\": \"Answer\",\n        \"text\": \"Bias testing should be done by developers and healthcare institutions using structured protocols and benchmarks like CLIMB, DiversityMedQA, or CPV datasets to ensure robust validation across diverse patient populations.\"\n      }\n    }\n  ]\n}\n<\/script>\n","protected":false},"excerpt":{"rendered":"<p>FAQ How is bias measured in clinical LLMs? Bias is evaluated using clinical vignettes and &#8220;counterfactual&#8221; variations (e.g., changing patient attributes) to observe differential responses, allowing detection of both performance disparities and fairness issues across demographic groups. How common are demographic biases in healthcare LLM outputs? Systematic reviews reveal pervasive demographic bias, especially across race, [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":758,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"nf_dc_page":"","content-type":"","inline_featured_image":false,"footnotes":""},"categories":[119,10],"tags":[],"class_list":["post-669","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-case-studies","category-video"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.3 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Testing for Bias of Large Language Models in Clinical Applications - Pacific AI<\/title>\n<meta name=\"description\" content=\"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions\" \/>\n<meta name=\"robots\" content=\"noindex, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Testing for Bias of Large Language Models in Clinical Applications - Pacific AI\" \/>\n<meta property=\"og:description\" content=\"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions\" \/>\n<meta property=\"og:url\" content=\"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/\" \/>\n<meta property=\"og:site_name\" content=\"Pacific AI\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/people\/Pacific-AI\/61566807347567\/\" \/>\n<meta property=\"article:published_time\" content=\"2025-04-07T15:50:49+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2026-02-19T10:38:02+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/pacific.ai\/wp-content\/uploads\/2024\/11\/web_1.webp\" \/>\n\t<meta property=\"og:image:width\" content=\"550\" \/>\n\t<meta property=\"og:image:height\" content=\"440\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/webp\" \/>\n<meta name=\"author\" content=\"David Talby\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"David Talby\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"2 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/\"},\"author\":{\"name\":\"David Talby\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#\\\/schema\\\/person\\\/8a2b4d5d75c8752d83ae6bb1d44e0186\"},\"headline\":\"Testing for Bias of Large Language Models in Clinical Applications\",\"datePublished\":\"2025-04-07T15:50:49+00:00\",\"dateModified\":\"2026-02-19T10:38:02+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/\"},\"wordCount\":205,\"publisher\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/web_1.webp\",\"articleSection\":[\"Case studies\",\"Video\"],\"inLanguage\":\"en\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/\",\"url\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/\",\"name\":\"Testing for Bias of Large Language Models in Clinical Applications - Pacific AI\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/web_1.webp\",\"datePublished\":\"2025-04-07T15:50:49+00:00\",\"dateModified\":\"2026-02-19T10:38:02+00:00\",\"description\":\"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#breadcrumb\"},\"inLanguage\":\"en\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#primaryimage\",\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/web_1.webp\",\"contentUrl\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2024\\\/11\\\/web_1.webp\",\"width\":550,\"height\":440,\"caption\":\"Portrait of a healthcare data science leader alongside the article title, illustrating testing for bias in large language models used in clinical applications, with a focus on fairness, patient safety, and responsible AI in healthcare.\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/testing-for-bias-of-large-language-models-in-clinical-applications\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/pacific.ai\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Testing for Bias of Large Language Models in Clinical Applications\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#website\",\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/\",\"name\":\"Pacific AI\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#organization\",\"name\":\"Pacific AI\",\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2025\\\/06\\\/site_logo.svg\",\"contentUrl\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2025\\\/06\\\/site_logo.svg\",\"width\":182,\"height\":41,\"caption\":\"Pacific AI\"},\"image\":{\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/people\\\/Pacific-AI\\\/61566807347567\\\/\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/pacific-ai\\\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/#\\\/schema\\\/person\\\/8a2b4d5d75c8752d83ae6bb1d44e0186\",\"name\":\"David Talby\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2025\\\/03\\\/David_portret-96x96.webp\",\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2025\\\/03\\\/David_portret-96x96.webp\",\"contentUrl\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/wp-content\\\/uploads\\\/2025\\\/03\\\/David_portret-96x96.webp\",\"caption\":\"David Talby\"},\"description\":\"David Talby is a CTO at Pacific AI, helping healthcare &amp; life science companies put AI to good use. David is the creator of Spark NLP \u2013 the world\u2019s most widely used natural language processing library in the enterprise. He has extensive experience building and running web-scale software platforms and teams \u2013 in startups, for Microsoft\u2019s Bing in the US and Europe, and to scale Amazon\u2019s financial systems in Seattle and the UK. David holds a PhD in computer science and master\u2019s degrees in both computer science and business administration.\",\"sameAs\":[\"https:\\\/\\\/www.linkedin.com\\\/in\\\/davidtalby\\\/\"],\"url\":\"https:\\\/\\\/pacific.ai\\\/staging\\\/3667\\\/author\\\/david\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Testing for Bias of Large Language Models in Clinical Applications - Pacific AI","description":"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions","robots":{"index":"noindex","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"og_locale":"en_US","og_type":"article","og_title":"Testing for Bias of Large Language Models in Clinical Applications - Pacific AI","og_description":"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions","og_url":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/","og_site_name":"Pacific AI","article_publisher":"https:\/\/www.facebook.com\/people\/Pacific-AI\/61566807347567\/","article_published_time":"2025-04-07T15:50:49+00:00","article_modified_time":"2026-02-19T10:38:02+00:00","og_image":[{"width":550,"height":440,"url":"https:\/\/pacific.ai\/wp-content\/uploads\/2024\/11\/web_1.webp","type":"image\/webp"}],"author":"David Talby","twitter_card":"summary_large_image","twitter_misc":{"Written by":"David Talby","Est. reading time":"2 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#article","isPartOf":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/"},"author":{"name":"David Talby","@id":"https:\/\/pacific.ai\/staging\/3667\/#\/schema\/person\/8a2b4d5d75c8752d83ae6bb1d44e0186"},"headline":"Testing for Bias of Large Language Models in Clinical Applications","datePublished":"2025-04-07T15:50:49+00:00","dateModified":"2026-02-19T10:38:02+00:00","mainEntityOfPage":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/"},"wordCount":205,"publisher":{"@id":"https:\/\/pacific.ai\/staging\/3667\/#organization"},"image":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#primaryimage"},"thumbnailUrl":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2024\/11\/web_1.webp","articleSection":["Case studies","Video"],"inLanguage":"en"},{"@type":"WebPage","@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/","url":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/","name":"Testing for Bias of Large Language Models in Clinical Applications - Pacific AI","isPartOf":{"@id":"https:\/\/pacific.ai\/staging\/3667\/#website"},"primaryImageOfPage":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#primaryimage"},"image":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#primaryimage"},"thumbnailUrl":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2024\/11\/web_1.webp","datePublished":"2025-04-07T15:50:49+00:00","dateModified":"2026-02-19T10:38:02+00:00","description":"Explore how Pacific AI evaluates and mitigates bias in large language models for clinical applications, ensuring safer, more equitable healthcare AI solutions","breadcrumb":{"@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#breadcrumb"},"inLanguage":"en","potentialAction":[{"@type":"ReadAction","target":["https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/"]}]},{"@type":"ImageObject","inLanguage":"en","@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#primaryimage","url":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2024\/11\/web_1.webp","contentUrl":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2024\/11\/web_1.webp","width":550,"height":440,"caption":"Portrait of a healthcare data science leader alongside the article title, illustrating testing for bias in large language models used in clinical applications, with a focus on fairness, patient safety, and responsible AI in healthcare."},{"@type":"BreadcrumbList","@id":"https:\/\/pacific.ai\/testing-for-bias-of-large-language-models-in-clinical-applications\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/pacific.ai\/"},{"@type":"ListItem","position":2,"name":"Testing for Bias of Large Language Models in Clinical Applications"}]},{"@type":"WebSite","@id":"https:\/\/pacific.ai\/staging\/3667\/#website","url":"https:\/\/pacific.ai\/staging\/3667\/","name":"Pacific AI","description":"","publisher":{"@id":"https:\/\/pacific.ai\/staging\/3667\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/pacific.ai\/staging\/3667\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en"},{"@type":"Organization","@id":"https:\/\/pacific.ai\/staging\/3667\/#organization","name":"Pacific AI","url":"https:\/\/pacific.ai\/staging\/3667\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/pacific.ai\/staging\/3667\/#\/schema\/logo\/image\/","url":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2025\/06\/site_logo.svg","contentUrl":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2025\/06\/site_logo.svg","width":182,"height":41,"caption":"Pacific AI"},"image":{"@id":"https:\/\/pacific.ai\/staging\/3667\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/people\/Pacific-AI\/61566807347567\/","https:\/\/www.linkedin.com\/company\/pacific-ai\/"]},{"@type":"Person","@id":"https:\/\/pacific.ai\/staging\/3667\/#\/schema\/person\/8a2b4d5d75c8752d83ae6bb1d44e0186","name":"David Talby","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2025\/03\/David_portret-96x96.webp","url":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2025\/03\/David_portret-96x96.webp","contentUrl":"https:\/\/pacific.ai\/staging\/3667\/wp-content\/uploads\/2025\/03\/David_portret-96x96.webp","caption":"David Talby"},"description":"David Talby is a CTO at Pacific AI, helping healthcare &amp; life science companies put AI to good use. David is the creator of Spark NLP \u2013 the world\u2019s most widely used natural language processing library in the enterprise. He has extensive experience building and running web-scale software platforms and teams \u2013 in startups, for Microsoft\u2019s Bing in the US and Europe, and to scale Amazon\u2019s financial systems in Seattle and the UK. David holds a PhD in computer science and master\u2019s degrees in both computer science and business administration.","sameAs":["https:\/\/www.linkedin.com\/in\/davidtalby\/"],"url":"https:\/\/pacific.ai\/staging\/3667\/author\/david\/"}]}},"_links":{"self":[{"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/posts\/669","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/comments?post=669"}],"version-history":[{"count":8,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/posts\/669\/revisions"}],"predecessor-version":[{"id":2048,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/posts\/669\/revisions\/2048"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/media\/758"}],"wp:attachment":[{"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/media?parent=669"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/categories?post=669"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/pacific.ai\/staging\/3667\/wp-json\/wp\/v2\/tags?post=669"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}