Now showing 1 - 4 of 4
  • Publication
    Extracting GHZ states from linear cluster states
    ( 2023-11)
    Jong, J. de
    ;
    Hahn, F.
    ;
    Tcholtchev, Nikolay Vassilev
    ;
    ;
    Pappa, Anna
    Quantum information processing architectures typically only allow for nearest-neighbour entanglement creation. In many cases, this prevents the direct generation of states, which are commonly used for many communication and computation tasks. Here, we show how to obtain states between nodes in a network that are connected in a straight line, naturally allowing them to initially share linear cluster states. We prove a strict upper bound of ⌊(n+3)/2⌋ on the size of the set of nodes sharing a state that can be obtained from a linear cluster state of n qubits, using local Clifford unitaries, local Pauli measurements, and classical communication. Furthermore, we completely characterize all selections of nodes below this threshold that can share a state obtained within this setting. Finally, we demonstrate these transformations on the quantum device for linear cluster states of up to n=19 qubits.
  • Publication
    VisionKG: Unleashing the Power of Visual Datasets via Knowledge Graph
    ( 2023-09)
    Yuan, Jicheng
    ;
    Le-Tuan, Anh
    ;
    Nguyen-Duc, Manh
    ;
    Tran, Trung-Kien
    ;
    ;
    Phuoc, Danh Le
    The availability of vast amounts of visual data with heterogeneous features is a key factor for developing, testing, and benchmarking of new computer vision (CV) algorithms and architectures. Most visual datasets are created and curated for specific tasks or with limited image data distribution for very specific situations, and there is no unified approach to manage and access them across diverse sources, tasks, and taxonomies. This not only creates unnecessary overheads when building robust visual recognition systems, but also introduces biases into learning systems and limits the capabilities of data-centric AI. To address these problems, we propose the Vision Knowledge Graph (VisionKG), a novel resource that interlinks, organizes and manages visual datasets via knowledge graphs and Semantic Web technologies. It can serve as a unified framework facilitating simple access and querying of state-of-the-art visual datasets, regardless of their heterogeneous formats and taxonomies. One of the key differences between our approach and existing methods is that ours is knowledge-based rather than metadatabased. It enhances the enrichment of the semantics at both image and instance levels and offers various data retrieval and exploratory services via SPARQL. VisionKG currently contains 519 million RDF triples that describe approximately 40 million entities, and are accessible at https://vision.semkg.org and through APIs. With the integration of 30 datasets and four popular CV tasks, we demonstrate its usefulness across various scenarios when working with CV pipelines.
  • Publication
    SemRob: Towards semantic stream reasoning for robotic operating systems
    ( 2022)
    Nguyen-Duc, Manh
    ;
    Le-Tuan, Anh
    ;
    ;
    Bowden, David
    ;
    Phuoc, Danh Le
    Stream processing and reasoning is getting considerable attention in various application domains such as IoT, Industry IoT and Smart Cities. In parallel, reasoning and knowledge-based features have attracted research into many areas of robotics, such as robotic mapping, perception and interaction. To this end, the Semantic Stream Reasoning (SSR) framework can unify the representations of symbolic/semantic streams with deep neural networks, to integrate high-dimensional data streams, such as video streams and LiDAR point clouds, with traditional graph or relational stream data. As such, this positioning and system paper will outline our approach to build a platform to facilitate semantic stream reasoning capabilities on a robotic operating system called SemRob.
  • Publication
    CQELS 2.0: Towards a unified framework for semantic stream fusion
    ( 2022)
    Le-Tuan, Anh
    ;
    Nguyen-Duc, Manh
    ;
    Le, Chien-Quang
    ;
    Tran, Trung-Kien
    ;
    ;
    Eiter, Thomas
    ;
    Phuoc, Danh Le
    We present CQELS 2.0, the second version of Continuous Query Evaluation over Linked Streams. CQELS 2.0 is a platform-agnostic federated execution framework towards semantic stream fusion. In this version, we introduce a novel neuralsymbolic stream reasoning component that enables specifying deep neural network (DNN) based data fusion pipelines via logic rules with learnable probabilistic degrees as weights. As a platform-agnostic framework, CQELS 2.0 can be implemented for devices with different hardware architectures (from embedded devices to cloud infrastructures). Moreover, this version also includes an adaptive federator that allows CQELS instances on different nodes in a network to coordinate their resources to distribute processing pipelines by delegating partial workloads to their peers via subscribing continuous queries.