<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
<record>
  <controlfield tag="001">112208</controlfield>
  <controlfield tag="005">20220510091957.0</controlfield>
  <datafield tag="037" ind1=" " ind2=" ">
    <subfield code="a">TAZ-TFM-2022-026</subfield>
  </datafield>
  <datafield tag="041" ind1=" " ind2=" ">
    <subfield code="a">eng</subfield>
  </datafield>
  <datafield tag="100" ind1="1" ind2=" ">
    <subfield code="a">Berriel Martins, Tomás</subfield>
  </datafield>
  <datafield tag="242" ind1="0" ind2="0">
    <subfield code="a">Learning disentangled representations of scenes from images.</subfield>
  </datafield>
  <datafield tag="245" ind1="0" ind2="0">
    <subfield code="a">Aprendizaje de representaciones desenredadas de escenas a partir de imágenes.</subfield>
  </datafield>
  <datafield tag="260" ind1=" " ind2=" ">
    <subfield code="a">Zaragoza</subfield>
    <subfield code="b">Universidad de Zaragoza</subfield>
    <subfield code="c">2022</subfield>
  </datafield>
  <datafield tag="506" ind1=" " ind2=" ">
    <subfield code="a">by-nc-sa</subfield>
    <subfield code="b">Creative Commons</subfield>
    <subfield code="c">3.0</subfield>
    <subfield code="u">http://creativecommons.org/licenses/by-nc-sa/3.0/</subfield>
  </datafield>
  <datafield tag="520" ind1=" " ind2=" ">
    <subfield code="a">Artificial intelligence is at the forefront of a technological revolution, in particular as a key component to build autonomous agents. However, not only training such agents come at a great computational cost, but they also end up lacking human basic abilities like generalization, information extrapolation, knowledge transfer between contexts, or improvisation. To overcome current limitations, agents need a deeper understanding of their environment, and more efficiently learning it from data. &lt;br />There are very recent works that propose novel approaches to learn representations of the world: instead of learning invariant object encodings, they learn to isolate, or disentangle, the different variable properties which form an object. This would not only enable agents to understand object changes as modifications of one of their properties, but also to transfer such knowledge on the properties between different categories. This Master Thesis aims to develop a new machine learning model for disentangling object properties on monocular images of scenes. Our model is based on a state-of-the-art architecture for disentangled representations learning, and our goal is to reduce the computational complexity of the base model while also improving its performance. To achieve this, we will replace a recursive unsupervised segmentation network by an encoder-decoder segmentation network. Furthermore, before training such overparametrized neural model without supervision, we will profit from transfer learning of pre-trained weights from a supervised segmentation task. After developing a first vanilla model, we have tuned it to improve its performance and generalization capability. Then, an experimental validation has been performed on two commonly used synthetic datasets, evaluating both its disentanglement performance and computational efficiency, and on a more realistic dataset to analyze the model capability on real data. The results show that our model outperforms the state of the art, while reducing its computational footprint. Nevertheless, further research is needed to bridge the gap with real world applications.&lt;br /></subfield>
  </datafield>
  <datafield tag="521" ind1=" " ind2=" ">
    <subfield code="a">Máster Universitario en Robótica, Gráficos y Visión por Computador</subfield>
  </datafield>
  <datafield tag="540" ind1=" " ind2=" ">
    <subfield code="a">Derechos regulados por licencia Creative Commons</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="a">Civera Sancho, Javier</subfield>
    <subfield code="e">dir.</subfield>
  </datafield>
  <datafield tag="710" ind1="2" ind2=" ">
    <subfield code="a">Universidad de Zaragoza</subfield>
    <subfield code="b">Informática e Ingeniería de Sistemas</subfield>
    <subfield code="c">Ingeniería de Sistemas y Automática</subfield>
  </datafield>
  <datafield tag="856" ind1="0" ind2=" ">
    <subfield code="f">718756@unizar.es</subfield>
  </datafield>
  <datafield tag="856" ind1="4" ind2=" ">
    <subfield code="s">19046623</subfield>
    <subfield code="u">http://zaguan.unizar.es/record/112208/files/TAZ-TFM-2022-026.pdf</subfield>
    <subfield code="y">Memoria (eng)</subfield>
  </datafield>
  <datafield tag="909" ind1="C" ind2="O">
    <subfield code="o">oai:zaguan.unizar.es:112208</subfield>
    <subfield code="p">driver</subfield>
    <subfield code="p">trabajos-fin-master</subfield>
  </datafield>
  <datafield tag="950" ind1=" " ind2=" ">
    <subfield code="a"></subfield>
  </datafield>
  <datafield tag="980" ind1=" " ind2=" ">
    <subfield code="a">TAZ</subfield>
    <subfield code="b">TFM</subfield>
    <subfield code="c">EINA</subfield>
  </datafield>
  <datafield tag="999" ind1=" " ind2=" ">
    <subfield code="a">20220128171613.CREATION_DATE</subfield>
  </datafield>
  <datafield tag="951" ind1=" " ind2=" ">
    <subfield code="a">deposita:2022-05-10</subfield>
  </datafield>
</record>
</collection>