<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" encoding="UTF-8" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:admin="http://webns.net/mvcb/" xmlns:atom="http://www.w3.org/2005/Atom/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:fireside="http://fireside.fm/modules/rss/fireside">
  <channel>
    <fireside:hostname>web02.fireside.fm</fireside:hostname>
    <fireside:genDate>Thu, 23 Apr 2026 10:13:37 -0500</fireside:genDate>
    <generator>Fireside (https://fireside.fm)</generator>
    <title>Pipeline Conversations - Episodes Tagged with “Data Centric Ai”</title>
    <link>https://podcast.zenml.io/tags/data-centric-ai</link>
    <pubDate>Thu, 15 Sep 2022 10:00:00 +0200</pubDate>
    <description>Pipeline Conversations brings you interviews with platform engineers, ML practitioners, and technical leaders building production AI systems. We dig into the real challenges of MLOps and LLMOps: orchestrating complex workflows on Kubernetes, fine-tuning and evaluating models at scale, and shipping AI that actually works. From ZenML.
</description>
    <language>en-us</language>
    <itunes:type>episodic</itunes:type>
    <itunes:subtitle>MLOps and LLMOps, from the trenches</itunes:subtitle>
    <itunes:author>ZenML GmbH</itunes:author>
    <itunes:summary>Pipeline Conversations brings you interviews with platform engineers, ML practitioners, and technical leaders building production AI systems. We dig into the real challenges of MLOps and LLMOps: orchestrating complex workflows on Kubernetes, fine-tuning and evaluating models at scale, and shipping AI that actually works. From ZenML.
</itunes:summary>
    <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/4/4d525632-f8ef-47c1-9321-20f5c498b1ac/cover.jpg?v=3"/>
    <itunes:explicit>no</itunes:explicit>
    <itunes:keywords>machine-learning, machinelearning, mlops, deeplearning, ai, artificialintelligence, artificial-intelligence, technology, tech, mlops, llmops</itunes:keywords>
    <itunes:owner>
      <itunes:name>ZenML GmbH</itunes:name>
      <itunes:email>podcast@zenml.io</itunes:email>
    </itunes:owner>
<itunes:category text="Technology"/>
<item>
  <title>Data-centric Computer Vision with Eric Landau</title>
  <link>https://podcast.zenml.io/data-centric-computer-vision-eric-landau-encord</link>
  <guid isPermaLink="false">f7d61b52-02c8-4401-894b-92110dde2267</guid>
  <pubDate>Thu, 15 Sep 2022 10:00:00 +0200</pubDate>
  <author>ZenML GmbH</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/4d525632-f8ef-47c1-9321-20f5c498b1ac/f7d61b52-02c8-4401-894b-92110dde2267.mp3" length="38096700" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>2</itunes:season>
  <itunes:author>ZenML GmbH</itunes:author>
  <itunes:subtitle>This week I spoke with Eric Landau, co-founder of Encord, a platform for data-centric computer vision. This podcast contains a lot of geekery about annotation, and even though Encord aren't an annotation tool per se, Eric and his team have tackled a bunch of quite complicated problems relating to that domain.</itunes:subtitle>
  <itunes:duration>51:51</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/4/4d525632-f8ef-47c1-9321-20f5c498b1ac/episodes/f/f7d61b52-02c8-4401-894b-92110dde2267/cover.jpg?v=1"/>
  <description>This week I spoke with Eric Landau, co-founder of Encord, a platform for data-centric computer vision. This podcast contains a lot of geekery about annotation, and even though Encord aren't an annotation tool per se, Eric and his team have tackled a bunch of quite complicated problems relating to that domain.
We also discuss the much-used term 'data-centric AI' and consider where it's useful and where perhaps there's a little bit of hype. We also get into some of the technical tradeoffs and decisions that come when building a platform. I'm really excited to get to present this episode to you today as I really enjoyed the discussion. Special Guest: Eric Landau.
</description>
  <itunes:keywords>computer-vision, data-centric-ai, machine-learning, annotation, engineering</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>This week I spoke with Eric Landau, co-founder of Encord, a platform for data-centric computer vision. This podcast contains a lot of geekery about annotation, and even though Encord aren&#39;t an annotation tool per se, Eric and his team have tackled a bunch of quite complicated problems relating to that domain.</p>

<p>We also discuss the much-used term &#39;data-centric AI&#39; and consider where it&#39;s useful and where perhaps there&#39;s a little bit of hype. We also get into some of the technical tradeoffs and decisions that come when building a platform. I&#39;m really excited to get to present this episode to you today as I really enjoyed the discussion.</p><p>Special Guest: Eric Landau.</p><p>Links:</p><ul><li><a title="Eric Landau (LinkedIn)" rel="nofollow" href="https://www.linkedin.com/in/eric-landau-40992ab0/">Eric Landau (LinkedIn)</a></li><li><a title="Encord | The platform for data-centric computer vision" rel="nofollow" href="https://encord.com/">Encord | The platform for data-centric computer vision</a></li><li><a title="Encord blog" rel="nofollow" href="https://blog.encord.com/">Encord blog</a></li><li><a title="Encord (Github)" rel="nofollow" href="https://github.com/encord-team">Encord (Github)</a></li><li><a title="Encord (@encord_team) / Twitter" rel="nofollow" href="https://twitter.com/encord_team">Encord (@encord_team) / Twitter</a></li></ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>This week I spoke with Eric Landau, co-founder of Encord, a platform for data-centric computer vision. This podcast contains a lot of geekery about annotation, and even though Encord aren&#39;t an annotation tool per se, Eric and his team have tackled a bunch of quite complicated problems relating to that domain.</p>

<p>We also discuss the much-used term &#39;data-centric AI&#39; and consider where it&#39;s useful and where perhaps there&#39;s a little bit of hype. We also get into some of the technical tradeoffs and decisions that come when building a platform. I&#39;m really excited to get to present this episode to you today as I really enjoyed the discussion.</p><p>Special Guest: Eric Landau.</p><p>Links:</p><ul><li><a title="Eric Landau (LinkedIn)" rel="nofollow" href="https://www.linkedin.com/in/eric-landau-40992ab0/">Eric Landau (LinkedIn)</a></li><li><a title="Encord | The platform for data-centric computer vision" rel="nofollow" href="https://encord.com/">Encord | The platform for data-centric computer vision</a></li><li><a title="Encord blog" rel="nofollow" href="https://blog.encord.com/">Encord blog</a></li><li><a title="Encord (Github)" rel="nofollow" href="https://github.com/encord-team">Encord (Github)</a></li><li><a title="Encord (@encord_team) / Twitter" rel="nofollow" href="https://twitter.com/encord_team">Encord (@encord_team) / Twitter</a></li></ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Autonomous Shipping with Captain AI</title>
  <link>https://podcast.zenml.io/autonomous-shipping-gerard-kruisheer-captain-ai</link>
  <guid isPermaLink="false">6c9b65e9-4f39-430e-8cfb-fd17a76fe8d7</guid>
  <pubDate>Thu, 21 Jul 2022 10:00:00 +0200</pubDate>
  <author>ZenML GmbH</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/4d525632-f8ef-47c1-9321-20f5c498b1ac/6c9b65e9-4f39-430e-8cfb-fd17a76fe8d7.mp3" length="44221152" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>2</itunes:season>
  <itunes:author>ZenML GmbH</itunes:author>
  <itunes:subtitle>This week on the podcast I spoke with Gerard Kruisheer, the CTO and co-founder of Captain AI, a company based in the Netherlands working on autonomous shipping out of the busy Rotterdam port.</itunes:subtitle>
  <itunes:duration>1:00:22</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/4/4d525632-f8ef-47c1-9321-20f5c498b1ac/episodes/6/6c9b65e9-4f39-430e-8cfb-fd17a76fe8d7/cover.jpg?v=1"/>
  <description>This week on the podcast I spoke with Gerard Kruisheer, the CTO and co-founder of Captain AI (https://www.captainai.com/), a company based in the Netherlands working on autonomous shipping out of the busy Rotterdam port.
We discussed the unique problems that come with building autonomous vehicles, the extent to which the latest and greatest research informs their work, their production stack and how they handle deployment for their particular setup.
As always please let us know if you have guests you'd like me to speak to by sending a message to us on slack or by emailing podcast@zenml.io (podcast@zenml.io). Special Guest: Gerard Kruisheer.
</description>
  <itunes:keywords>computer-vision, data-centric-ai, machine-learning, edge-ml, shipping, autonomous, vehicles</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>This week on the podcast I spoke with Gerard Kruisheer, the CTO and co-founder of <a href="https://www.captainai.com/" rel="nofollow">Captain AI</a>, a company based in the Netherlands working on autonomous shipping out of the busy Rotterdam port.</p>

<p>We discussed the unique problems that come with building autonomous vehicles, the extent to which the latest and greatest research informs their work, their production stack and how they handle deployment for their particular setup.</p>

<p>As always please let us know if you have guests you&#39;d like me to speak to by sending a message to us on slack or by emailing [<a href="mailto:podcast@zenml.io" rel="nofollow">podcast@zenml.io</a>](<a href="mailto:podcast@zenml.io" rel="nofollow">podcast@zenml.io</a>).</p><p>Special Guest: Gerard Kruisheer.</p><p>Links:</p><ul><li><a title="Gerard Kruisheer (LinkedIn profile)" rel="nofollow" href="https://www.linkedin.com/in/gkruisheer/">Gerard Kruisheer (LinkedIn profile)</a></li><li><a title="Captain AI – Autonomous ships for autonomous ports" rel="nofollow" href="https://www.captainai.com/">Captain AI – Autonomous ships for autonomous ports</a></li><li><a title="Blog – Captain AI" rel="nofollow" href="https://www.captainai.com/blog/">Blog – Captain AI</a></li><li><a title="The ship which sails itself: arriving soon, thanks to Captain AI and Xsens motion tracking modules" rel="nofollow" href="https://www.xsens.com/cases/the-ship-which-sails-itself-arriving-soon-thanks-to-captain-ai-and-xsens-motion-tracking-modules">The ship which sails itself: arriving soon, thanks to Captain AI and Xsens motion tracking modules</a></li><li><a title="Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/channel/UC4vsUMnLs06MfApE4GOBp5w">Captain AI - YouTube</a></li><li><a title="National Geographic - Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=sLr_NhnYI88">National Geographic - Captain AI - YouTube</a></li><li><a title="Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=3Hg7iNaa-GA">Captain AI - YouTube</a></li></ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>This week on the podcast I spoke with Gerard Kruisheer, the CTO and co-founder of <a href="https://www.captainai.com/" rel="nofollow">Captain AI</a>, a company based in the Netherlands working on autonomous shipping out of the busy Rotterdam port.</p>

<p>We discussed the unique problems that come with building autonomous vehicles, the extent to which the latest and greatest research informs their work, their production stack and how they handle deployment for their particular setup.</p>

<p>As always please let us know if you have guests you&#39;d like me to speak to by sending a message to us on slack or by emailing [<a href="mailto:podcast@zenml.io" rel="nofollow">podcast@zenml.io</a>](<a href="mailto:podcast@zenml.io" rel="nofollow">podcast@zenml.io</a>).</p><p>Special Guest: Gerard Kruisheer.</p><p>Links:</p><ul><li><a title="Gerard Kruisheer (LinkedIn profile)" rel="nofollow" href="https://www.linkedin.com/in/gkruisheer/">Gerard Kruisheer (LinkedIn profile)</a></li><li><a title="Captain AI – Autonomous ships for autonomous ports" rel="nofollow" href="https://www.captainai.com/">Captain AI – Autonomous ships for autonomous ports</a></li><li><a title="Blog – Captain AI" rel="nofollow" href="https://www.captainai.com/blog/">Blog – Captain AI</a></li><li><a title="The ship which sails itself: arriving soon, thanks to Captain AI and Xsens motion tracking modules" rel="nofollow" href="https://www.xsens.com/cases/the-ship-which-sails-itself-arriving-soon-thanks-to-captain-ai-and-xsens-motion-tracking-modules">The ship which sails itself: arriving soon, thanks to Captain AI and Xsens motion tracking modules</a></li><li><a title="Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/channel/UC4vsUMnLs06MfApE4GOBp5w">Captain AI - YouTube</a></li><li><a title="National Geographic - Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=sLr_NhnYI88">National Geographic - Captain AI - YouTube</a></li><li><a title="Captain AI - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=3Hg7iNaa-GA">Captain AI - YouTube</a></li></ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Edge Computer Vision with Karthik Kannan</title>
  <link>https://podcast.zenml.io/edge-computer-vision-karthik-kannan</link>
  <guid isPermaLink="false">2f68e4a6-373b-416d-86be-327ac4f52ab4</guid>
  <pubDate>Thu, 30 Jun 2022 09:00:00 +0200</pubDate>
  <author>ZenML GmbH</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/4d525632-f8ef-47c1-9321-20f5c498b1ac/2f68e4a6-373b-416d-86be-327ac4f52ab4.mp3" length="34519703" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>2</itunes:season>
  <itunes:author>ZenML GmbH</itunes:author>
  <itunes:subtitle>This week I spoke with Karthik Kannan, cofounder and CTO of Envision, a company that builds on top of the Google Glass and using Augmented Reality features of phones to allow visually impaired people to better sense the environment or objects around them.</itunes:subtitle>
  <itunes:duration>46:53</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/4/4d525632-f8ef-47c1-9321-20f5c498b1ac/episodes/2/2f68e4a6-373b-416d-86be-327ac4f52ab4/cover.jpg?v=1"/>
  <description>This week I spoke with Karthik Kannan, cofounder and CTO of Envision (https://www.letsenvision.com/), a company that builds on top of the Google Glass and using Augmented Reality features of phones to allow visually impaired people to better sense the environment or objects around them.
Their software and devices are pretty popular and as you'll hear in this conversation, they've been on a real journey to get to where they are now.
In particular, I really enjoyed the parts where Karthik explained their development and deployment process in detail. It's not too often that you get a deep dive into the workflows and stacks of an embedded computer vision company and tool and so I think you're going to really enjoy this one. Special Guest: Karthik Kannan.
</description>
  <itunes:keywords>computer-vision, data-centric-ai, machine-learning, google-glass, edge-ml</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>This week I spoke with Karthik Kannan, cofounder and CTO of <a href="https://www.letsenvision.com/" rel="nofollow">Envision</a>, a company that builds on top of the Google Glass and using Augmented Reality features of phones to allow visually impaired people to better sense the environment or objects around them.</p>

<p>Their software and devices are pretty popular and as you&#39;ll hear in this conversation, they&#39;ve been on a real journey to get to where they are now.</p>

<p>In particular, I really enjoyed the parts where Karthik explained their development and deployment process in detail. It&#39;s not too often that you get a deep dive into the workflows and stacks of an embedded computer vision company and tool and so I think you&#39;re going to really enjoy this one.</p><p>Special Guest: Karthik Kannan.</p><p>Links:</p><ul><li><a title="Karthik Kannan (@meTheKarthik) / Twitter" rel="nofollow" href="https://twitter.com/meTheKarthik">Karthik Kannan (@meTheKarthik) / Twitter</a></li><li><a title="Envision - Hear what you want to see." rel="nofollow" href="https://www.letsenvision.com/">Envision - Hear what you want to see.</a></li><li><a title="Glass – Glass" rel="nofollow" href="https://www.google.com/glass/start/">Glass – Glass</a></li><li><a title="Introducing Envision Glasses: AI-powered smartglasses for the Blind &amp; Visually Impaired - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=9ehENnq2EFo">Introducing Envision Glasses: AI-powered smartglasses for the Blind &amp; Visually Impaired - YouTube</a></li><li><a title="Envision Blog" rel="nofollow" href="https://www.letsenvision.com/blog">Envision Blog</a></li></ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>This week I spoke with Karthik Kannan, cofounder and CTO of <a href="https://www.letsenvision.com/" rel="nofollow">Envision</a>, a company that builds on top of the Google Glass and using Augmented Reality features of phones to allow visually impaired people to better sense the environment or objects around them.</p>

<p>Their software and devices are pretty popular and as you&#39;ll hear in this conversation, they&#39;ve been on a real journey to get to where they are now.</p>

<p>In particular, I really enjoyed the parts where Karthik explained their development and deployment process in detail. It&#39;s not too often that you get a deep dive into the workflows and stacks of an embedded computer vision company and tool and so I think you&#39;re going to really enjoy this one.</p><p>Special Guest: Karthik Kannan.</p><p>Links:</p><ul><li><a title="Karthik Kannan (@meTheKarthik) / Twitter" rel="nofollow" href="https://twitter.com/meTheKarthik">Karthik Kannan (@meTheKarthik) / Twitter</a></li><li><a title="Envision - Hear what you want to see." rel="nofollow" href="https://www.letsenvision.com/">Envision - Hear what you want to see.</a></li><li><a title="Glass – Glass" rel="nofollow" href="https://www.google.com/glass/start/">Glass – Glass</a></li><li><a title="Introducing Envision Glasses: AI-powered smartglasses for the Blind &amp; Visually Impaired - YouTube" rel="nofollow" href="https://www.youtube.com/watch?v=9ehENnq2EFo">Introducing Envision Glasses: AI-powered smartglasses for the Blind &amp; Visually Impaired - YouTube</a></li><li><a title="Envision Blog" rel="nofollow" href="https://www.letsenvision.com/blog">Envision Blog</a></li></ul>]]>
  </itunes:summary>
</item>
<item>
  <title>Humans in the Loop with Iva Gumnishka</title>
  <link>https://podcast.zenml.io/humans-in-loop-iva-gumnishka</link>
  <guid isPermaLink="false">2b5720f5-ce03-4cd5-a077-87780513ee1d</guid>
  <pubDate>Thu, 23 Jun 2022 10:00:00 +0200</pubDate>
  <author>ZenML GmbH</author>
  <enclosure url="https://aphid.fireside.fm/d/1437767933/4d525632-f8ef-47c1-9321-20f5c498b1ac/2b5720f5-ce03-4cd5-a077-87780513ee1d.mp3" length="37416318" type="audio/mpeg"/>
  <itunes:episodeType>full</itunes:episodeType>
  <itunes:season>2</itunes:season>
  <itunes:author>ZenML GmbH</itunes:author>
  <itunes:subtitle>We were lucky to get to talk to [Iva Gumnishka](https://www.linkedin.com/in/ivagumnishka/), the founder of [Humans in the Loop](https://humansintheloop.org/). They are an organisation that provides data annotation and collection services. Their teams are primarily made up of those who have been affected by conflict and now are asylum seekers or refugees.</itunes:subtitle>
  <itunes:duration>50:55</itunes:duration>
  <itunes:explicit>no</itunes:explicit>
  <itunes:image href="https://media24.fireside.fm/file/fireside-images-2024/podcasts/images/4/4d525632-f8ef-47c1-9321-20f5c498b1ac/episodes/2/2b5720f5-ce03-4cd5-a077-87780513ee1d/cover.jpg?v=1"/>
  <description>In this episode, I'm really happy to be able to continue the dialogue we've been having with our users and community around the role of data annotation and labeling in MLOps.
We were lucky to get to talk to Iva Gumnishka (https://www.linkedin.com/in/ivagumnishka/), the founder of Humans in the Loop (https://humansintheloop.org/). They are an organisation that provides data annotation and collection services. Their teams are primarily made up of those who have been affected by conflict and now are asylum seekers or refugees.
Iva has a ton of experience working with annotation and has seen how different companies build this into their production machine learning lifecycles. We're continuing to work on a feature that will allow you to do this as part of your MLOps workflow when using ZenML, and I welcome any feedback you might have on the back of this podcast or the articles we've been publishing on the ZenML blog. Special Guest: Iva Gumnishka.
</description>
  <itunes:keywords>data-annotation, labeling, annotation, data, data-centric-ai, machine-learning</itunes:keywords>
  <content:encoded>
    <![CDATA[<p>In this episode, I&#39;m really happy to be able to continue the dialogue we&#39;ve been having with our users and community around the role of data annotation and labeling in MLOps.</p>

<p>We were lucky to get to talk to <a href="https://www.linkedin.com/in/ivagumnishka/" rel="nofollow">Iva Gumnishka</a>, the founder of <a href="https://humansintheloop.org/" rel="nofollow">Humans in the Loop</a>. They are an organisation that provides data annotation and collection services. Their teams are primarily made up of those who have been affected by conflict and now are asylum seekers or refugees.</p>

<p>Iva has a ton of experience working with annotation and has seen how different companies build this into their production machine learning lifecycles. We&#39;re continuing to work on a feature that will allow you to do this as part of your MLOps workflow when using ZenML, and I welcome any feedback you might have on the back of this podcast or the articles we&#39;ve been publishing on the ZenML blog.</p><p>Special Guest: Iva Gumnishka.</p><p>Links:</p><ul><li><a title="Humans in the Loop | Image annotation for ethical AI" rel="nofollow" href="https://humansintheloop.org/">Humans in the Loop | Image annotation for ethical AI</a></li><li><a title="Blog | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/resources/blog/">Blog | Humans in the Loop</a></li><li><a title="10 of the best open-source annotation tools for computer vision 2022 | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/10-of-the-best-open-source-annotation-tools-for-computer-vision-2022/">10 of the best open-source annotation tools for computer vision 2022 | Humans in the Loop</a></li><li><a title="zenml-io/awesome-open-data-annotation: Open Source Data Annotation &amp; Labeling Tools" rel="nofollow" href="https://github.com/zenml-io/awesome-open-data-annotation">zenml-io/awesome-open-data-annotation: Open Source Data Annotation &amp; Labeling Tools</a></li><li><a title="Need an open-source data annotation tool? We’ve got you covered! | ZenML Blog" rel="nofollow" href="https://blog.zenml.io/open-source-data-annotation-tools/">Need an open-source data annotation tool? We’ve got you covered! | ZenML Blog</a></li><li><a title="How to get the most out of data annotation | ZenML Blog" rel="nofollow" href="https://blog.zenml.io/data-labelling-annotation/">How to get the most out of data annotation | ZenML Blog</a></li><li><a title="Foundation | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/why-us/foundation/">Foundation | Humans in the Loop</a></li><li><a title="Your Data Needs a Human Touch. The story of Iva Gumnishka, a Bulgarian… | by Antoaneta Manko | womenintechglobal | Medium" rel="nofollow" href="https://medium.com/bulgarianwomenintech/your-data-needs-a-human-touch-5bc2ee70d548">Your Data Needs a Human Touch. The story of Iva Gumnishka, a Bulgarian… | by Antoaneta Manko | womenintechglobal | Medium</a></li></ul>]]>
  </content:encoded>
  <itunes:summary>
    <![CDATA[<p>In this episode, I&#39;m really happy to be able to continue the dialogue we&#39;ve been having with our users and community around the role of data annotation and labeling in MLOps.</p>

<p>We were lucky to get to talk to <a href="https://www.linkedin.com/in/ivagumnishka/" rel="nofollow">Iva Gumnishka</a>, the founder of <a href="https://humansintheloop.org/" rel="nofollow">Humans in the Loop</a>. They are an organisation that provides data annotation and collection services. Their teams are primarily made up of those who have been affected by conflict and now are asylum seekers or refugees.</p>

<p>Iva has a ton of experience working with annotation and has seen how different companies build this into their production machine learning lifecycles. We&#39;re continuing to work on a feature that will allow you to do this as part of your MLOps workflow when using ZenML, and I welcome any feedback you might have on the back of this podcast or the articles we&#39;ve been publishing on the ZenML blog.</p><p>Special Guest: Iva Gumnishka.</p><p>Links:</p><ul><li><a title="Humans in the Loop | Image annotation for ethical AI" rel="nofollow" href="https://humansintheloop.org/">Humans in the Loop | Image annotation for ethical AI</a></li><li><a title="Blog | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/resources/blog/">Blog | Humans in the Loop</a></li><li><a title="10 of the best open-source annotation tools for computer vision 2022 | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/10-of-the-best-open-source-annotation-tools-for-computer-vision-2022/">10 of the best open-source annotation tools for computer vision 2022 | Humans in the Loop</a></li><li><a title="zenml-io/awesome-open-data-annotation: Open Source Data Annotation &amp; Labeling Tools" rel="nofollow" href="https://github.com/zenml-io/awesome-open-data-annotation">zenml-io/awesome-open-data-annotation: Open Source Data Annotation &amp; Labeling Tools</a></li><li><a title="Need an open-source data annotation tool? We’ve got you covered! | ZenML Blog" rel="nofollow" href="https://blog.zenml.io/open-source-data-annotation-tools/">Need an open-source data annotation tool? We’ve got you covered! | ZenML Blog</a></li><li><a title="How to get the most out of data annotation | ZenML Blog" rel="nofollow" href="https://blog.zenml.io/data-labelling-annotation/">How to get the most out of data annotation | ZenML Blog</a></li><li><a title="Foundation | Humans in the Loop" rel="nofollow" href="https://humansintheloop.org/why-us/foundation/">Foundation | Humans in the Loop</a></li><li><a title="Your Data Needs a Human Touch. The story of Iva Gumnishka, a Bulgarian… | by Antoaneta Manko | womenintechglobal | Medium" rel="nofollow" href="https://medium.com/bulgarianwomenintech/your-data-needs-a-human-touch-5bc2ee70d548">Your Data Needs a Human Touch. The story of Iva Gumnishka, a Bulgarian… | by Antoaneta Manko | womenintechglobal | Medium</a></li></ul>]]>
  </itunes:summary>
</item>
  </channel>
</rss>
