Now showing 1 - 2 of 2
  • Publication
    Extracting GHZ states from linear cluster states
    ( 2023-11)
    Jong, J. de
    ;
    Hahn, F.
    ;
    Tcholtchev, Nikolay Vassilev
    ;
    ;
    Pappa, Anna
    Quantum information processing architectures typically only allow for nearest-neighbour entanglement creation. In many cases, this prevents the direct generation of states, which are commonly used for many communication and computation tasks. Here, we show how to obtain states between nodes in a network that are connected in a straight line, naturally allowing them to initially share linear cluster states. We prove a strict upper bound of ⌊(n+3)/2⌋ on the size of the set of nodes sharing a state that can be obtained from a linear cluster state of n qubits, using local Clifford unitaries, local Pauli measurements, and classical communication. Furthermore, we completely characterize all selections of nodes below this threshold that can share a state obtained within this setting. Finally, we demonstrate these transformations on the quantum device for linear cluster states of up to n=19 qubits.
  • Publication
    VisionKG: Unleashing the Power of Visual Datasets via Knowledge Graph
    ( 2023-09)
    Yuan, Jicheng
    ;
    Le-Tuan, Anh
    ;
    Nguyen-Duc, Manh
    ;
    Tran, Trung-Kien
    ;
    ;
    Phuoc, Danh Le
    The availability of vast amounts of visual data with heterogeneous features is a key factor for developing, testing, and benchmarking of new computer vision (CV) algorithms and architectures. Most visual datasets are created and curated for specific tasks or with limited image data distribution for very specific situations, and there is no unified approach to manage and access them across diverse sources, tasks, and taxonomies. This not only creates unnecessary overheads when building robust visual recognition systems, but also introduces biases into learning systems and limits the capabilities of data-centric AI. To address these problems, we propose the Vision Knowledge Graph (VisionKG), a novel resource that interlinks, organizes and manages visual datasets via knowledge graphs and Semantic Web technologies. It can serve as a unified framework facilitating simple access and querying of state-of-the-art visual datasets, regardless of their heterogeneous formats and taxonomies. One of the key differences between our approach and existing methods is that ours is knowledge-based rather than metadatabased. It enhances the enrichment of the semantics at both image and instance levels and offers various data retrieval and exploratory services via SPARQL. VisionKG currently contains 519 million RDF triples that describe approximately 40 million entities, and are accessible at https://vision.semkg.org and through APIs. With the integration of 30 datasets and four popular CV tasks, we demonstrate its usefulness across various scenarios when working with CV pipelines.