Skip to content

ontology

A module adding additional functionality to owlready2.

If desirable some of these additions may be moved back into owlready2.

BlankNode

Represents a blank node.

A blank node is a node that is not a literal and has no IRI. Resources represented by blank nodes are also called anonumous resources. Only the subject or object in an RDF triple can be a blank node.

Source code in ontopy/ontology.py
class BlankNode:
    """Represents a blank node.

    A blank node is a node that is not a literal and has no IRI.
    Resources represented by blank nodes are also called anonumous resources.
    Only the subject or object in an RDF triple can be a blank node.
    """

    def __init__(self, onto: Union[World, Ontology], storid: int):
        """Initiate a blank node.

        Args:
            onto: Ontology or World instance.
            storid: The storage id of the blank node.
        """
        if storid >= 0:
            raise ValueError(
                f"A BlankNode is supposed to have a negative storid: {storid}"
            )
        self.onto = onto
        self.storid = storid

    def __repr__(self):
        return repr(f"_:b{-self.storid}")

    def __hash__(self):
        return hash((self.onto, self.storid))

    def __eq__(self, other):
        """For now blank nodes always compare true against each other."""
        return isinstance(other, BlankNode)

__init__(self, onto, storid) special

Initiate a blank node.

Parameters:

Name Type Description Default
onto Union[ontopy.ontology.World, ontopy.ontology.Ontology]

Ontology or World instance.

required
storid int

The storage id of the blank node.

required
Source code in ontopy/ontology.py
def __init__(self, onto: Union[World, Ontology], storid: int):
    """Initiate a blank node.

    Args:
        onto: Ontology or World instance.
        storid: The storage id of the blank node.
    """
    if storid >= 0:
        raise ValueError(
            f"A BlankNode is supposed to have a negative storid: {storid}"
        )
    self.onto = onto
    self.storid = storid

Ontology (Ontology)

A generic class extending owlready2.Ontology.

Source code in ontopy/ontology.py
class Ontology(owlready2.Ontology):  # pylint: disable=too-many-public-methods
    """A generic class extending owlready2.Ontology."""

    def __init__(self, *args, **kwargs):
        # Properties controlling what annotations that are considered by
        # get_by_label()
        super().__init__(*args, **kwargs)
        self._label_annotations = None
        self.prefix = None

    # Properties controlling what annotations that are considered by
    # get_by_label()
    label_annotations = property(
        fget=lambda self: self._label_annotations,
        doc="List of label annotation searched for by get_by_label().",
    )

    # Name of special unlabeled entities, like Thing, Nothing, etc...
    _special_labels = None

    # Some properties for customising dir() listing - useful in
    # interactive sessions...
    _dir_preflabel = isinteractive()
    _dir_label = isinteractive()
    _dir_name = False
    _dir_imported = isinteractive()
    dir_preflabel = property(
        fget=lambda self: self._dir_preflabel,
        fset=lambda self, v: setattr(self, "_dir_preflabel", bool(v)),
        doc="Whether to include entity prefLabel in dir() listing.",
    )
    dir_label = property(
        fget=lambda self: self._dir_label,
        fset=lambda self, v: setattr(self, "_dir_label", bool(v)),
        doc="Whether to include entity label in dir() listing.",
    )
    dir_name = property(
        fget=lambda self: self._dir_name,
        fset=lambda self, v: setattr(self, "_dir_name", bool(v)),
        doc="Whether to include entity name in dir() listing.",
    )
    dir_imported = property(
        fget=lambda self: self._dir_imported,
        fset=lambda self, v: setattr(self, "_dir_imported", bool(v)),
        doc="Whether to include imported ontologies in dir() listing.",
    )

    # Other settings
    _colon_in_label = False
    colon_in_label = property(
        fget=lambda self: self._colon_in_label,
        fset=lambda self, v: setattr(self, "_colon_in_label", bool(v)),
        doc="Whether to accept colon in name-part of IRI.  "
        "If true, the name cannot be prefixed.",
    )

    def __dir__(self):
        dirset = set(super().__dir__())
        lst = list(self.get_entities(imported=self._dir_imported))
        if self._dir_preflabel:
            dirset.update(
                _.prefLabel.first() for _ in lst if hasattr(_, "prefLabel")
            )
        if self._dir_label:
            dirset.update(_.label.first() for _ in lst if hasattr(_, "label"))
        if self._dir_name:
            dirset.update(_.name for _ in lst if hasattr(_, "name"))
        dirset.difference_update({None})  # get rid of possible None
        return sorted(dirset)

    def __getitem__(self, name):
        item = super().__getitem__(name)
        if not item:
            item = self.get_by_label(name)
        return item

    def __getattr__(self, name):
        attr = super().__getattr__(name)
        if not attr:
            attr = self.get_by_label(name)
        return attr

    def __contains__(self, other):
        if self.world[other]:
            return True
        try:
            self.get_by_label(other)
        except NoSuchLabelError:
            return False
        return True

    def __objclass__(self):
        # Play nice with inspect...
        pass

    def __hash__(self):
        """Returns a hash based on base_iri.
        This is done to keep Ontology hashable when defining __eq__.
        """
        return hash(self.base_iri)

    def __eq__(self, other):
        """Checks if this ontology is equal to `other`.

        This function compares the result of
        ``set(self.get_unabbreviated_triples(label='_:b'))``,
        i.e. blank nodes are not distinguished, but relations to blank
        nodes are included.
        """
        return set(self.get_unabbreviated_triples(blank="_:b")) == set(
            other.get_unabbreviated_triples(blank="_:b")
        )

    def get_unabbreviated_triples(
        self, subject=None, predicate=None, obj=None, blank=None
    ):
        """Returns all matching triples unabbreviated.

        If `blank` is given, it will be used to represent blank nodes.
        """
        # pylint: disable=invalid-name
        return _get_unabbreviated_triples(
            self, subject=subject, predicate=predicate, obj=obj, blank=blank
        )

    def get_by_label(
        self,
        label: str,
        label_annotations: str = None,
        prefix: str = None,
        imported: bool = True,
        colon_in_label: bool = None,
    ):
        """Returns entity with label annotation `label`.

        Args:
           label: label so serach for.
               May be written as 'label' or 'prefix:label'.
               get_by_label('prefix:label') ==
               get_by_label('label', prefix='prefix').
           label_annotations: a sequence of label annotation names to look up.
               Defaults to the `label_annotations` property.
           prefix: if provided, it should be the last component of
               the base iri of an ontology (with trailing slash (/) or hash
               (#) stripped off).  The search for a matching label will be
               limited to this namespace.
           imported: Whether to also look for `label` in imported ontologies.
           colon_in_label: Whether to accept colon (:) in a label or name-part
               of IRI.  Defaults to the `colon_in_label` property of `self`.
               Setting this true cannot be combined with `prefix`.

        If several entities have the same label, only the one which is
        found first is returned.Use get_by_label_all() to get all matches.

        Note, if different prefixes are provided in the label and via
        the `prefix` argument a warning will be issued and the
        `prefix` argument will take precedence.

        A NoSuchLabelError is raised if `label` cannot be found.
        """
        # pylint: disable=too-many-arguments,too-many-branches,invalid-name
        if not isinstance(label, str):
            raise TypeError(
                f"Invalid label definition, must be a string: {label!r}"
            )

        if self._label_annotations is None:
            for iri in DEFAULT_LABEL_ANNOTATIONS:
                try:
                    self.add_label_annotation(iri)
                except ValueError:
                    pass

        if colon_in_label is None:
            colon_in_label = self._colon_in_label
        if colon_in_label:
            if prefix:
                raise ValueError(
                    "`prefix` cannot be combined with `colon_in_label`"
                )
        else:
            splitlabel = label.split(":", 1)
            if len(splitlabel) == 2 and not splitlabel[1].startswith("//"):
                label = splitlabel[1]
                if prefix and prefix != splitlabel[0]:
                    warnings.warn(
                        f"Prefix given both as argument ({prefix}) "
                        f"and in label ({splitlabel[0]}). "
                        "Prefix given in argument takes presendence "
                    )
                if not prefix:
                    prefix = splitlabel[0]

        if prefix:
            entitylist = self.get_by_label_all(
                label,
                label_annotations=label_annotations,
                prefix=prefix,
            )
            if len(entitylist) > 0:
                return entitylist[0]

            raise NoSuchLabelError(
                f"No label annotations matches {label!r} with prefix "
                f"{prefix!r}"
            )

        # Label is a full IRI
        entity = self.world[label]
        if entity:
            return entity

        # First entity with matching label annotation
        annotation_ids = (
            (self._abbreviate(ann, False) for ann in label_annotations)
            if label_annotations
            else (ann.storid for ann in self.label_annotations)
        )
        get_triples = (
            self.world._get_data_triples_spod_spod
            if imported
            else self._get_data_triples_spod_spod
        )
        for annotation_id in annotation_ids:
            for s, _, _, _ in get_triples(None, annotation_id, label, None):
                return self.world[self._unabbreviate(s)]

        # Special labels
        if self._special_labels and label in self._special_labels:
            return self._special_labels[label]

        # Check if label is a name under base_iri
        entity = self.world[self.base_iri + label]
        if entity:
            return entity

        # Check if label is a name in any namespace
        for namespace in self._namespaces.keys():
            entity = self.world[namespace + label]
            if entity:
                return entity

        raise NoSuchLabelError(f"No label annotations matches '{label}'")

    def get_by_label_all(self, label, label_annotations=None, prefix=None):
        """Like get_by_label(), but returns a list with all matching labels.

        Returns an empty list if no matches could be found.

        Note
        ----
        The current implementation also supports "*" as a wildcard
        matching any number of characters. This may change in the future.
        """
        if not isinstance(label, str):
            raise TypeError(
                f"Invalid label definition, " f"must be a string: {label!r}"
            )
        if " " in label:
            raise ValueError(
                f"Invalid label definition, {label!r} contains spaces."
            )

        if label_annotations is None:
            annotations = (_.name for _ in self.label_annotations)
        else:
            annotations = (
                _.name if hasattr(_, "storid") else _ for _ in label_annotations
            )
        entity = self.world.search(**{next(annotations): label})
        for key in annotations:
            entity.extend(self.world.search(**{key: label}))

        if self._special_labels and label in self._special_labels:
            entity.append(self._special_labels[label])

        entity_accessed_directly = self.world[self.base_iri + label]
        if entity_accessed_directly and entity_accessed_directly not in entity:
            entity.append(entity_accessed_directly)

        if prefix:
            return [_ for _ in entity if _.namespace.ontology.prefix == prefix]
        return entity

    def add_label_annotation(self, iri):
        """Adds label annotation used by get_by_label().

        May be provided either as an IRI or as its owlready2 representation.
        """
        if self._label_annotations is None:
            self._label_annotations = []
        label_annotation = iri if hasattr(iri, "storid") else self.world[iri]
        if label_annotation is None:
            warnings.warn(f"adding new IRI to ontology: {iri}")
            name = iri.rsplit("/")[-1].rsplit("#")[-1]
            bases = (owlready2.AnnotationProperty,)
            with self:
                label_annotation = types.new_class(name, bases)
        if label_annotation not in self._label_annotations:
            self._label_annotations.append(label_annotation)

    def remove_label_annotation(self, iri):
        """Removes label annotation used by get_by_label().

        May be provided either as an IRI or as its owlready2 representation.
        """
        label_annotation = iri if hasattr(iri, "storid") else self.world[iri]
        if not label_annotation:
            raise ValueError(f"IRI not in ontology: {iri}")
        self._label_annotations.remove(label_annotation)

    def set_common_prefix(
        self,
        iri_base: str = "http://emmo.info/emmo",
        prefix: str = "emmo",
    ) -> None:
        """Set a common prefix for all imported ontologies
        with the same first part of the base_iri.

        Args:
            iri_base: The start of the base_iri to look for. Defaults to
                the emmo base_iri http://emmo.info/emmo
            prefix: the desired prefix. Defaults to emmo.
        """
        if self.base_iri.startswith(iri_base):
            self.prefix = prefix
        for onto in self.imported_ontologies:
            onto.set_common_prefix(iri_base=iri_base, prefix=prefix)

    def load(  # pylint: disable=too-many-arguments,arguments-renamed
        self,
        only_local=False,
        filename=None,
        format=None,  # pylint: disable=redefined-builtin
        reload=None,
        reload_if_newer=False,
        url_from_catalog=None,
        catalog_file="catalog-v001.xml",
        emmo_based=True,
        prefix=None,
        prefix_emmo=None,
        **kwargs,
    ):
        """Load the ontology.

        Parameters
        ----------
        only_local: bool
            Whether to only read local files.  This requires that you
            have appended the path to the ontology to owlready2.onto_path.
        filename: str
            Path to file to load the ontology from.  Defaults to `base_iri`
            provided to get_ontology().
        format: str
            Format of `filename`.  Default is inferred from `filename`
            extension.
        reload: bool
            Whether to reload the ontology if it is already loaded.
        reload_if_newer: bool
            Whether to reload the ontology if the source has changed since
            last time it was loaded.
        url_from_catalog: bool | None
            Whether to use catalog file to resolve the location of `base_iri`.
            If None, the catalog file is used if it exists in the same
            directory as `filename`.
        catalog_file: str
            Name of Protègè catalog file in the same folder as the
            ontology.  This option is used together with `only_local` and
            defaults to "catalog-v001.xml".
        emmo_based: bool
            Whether this is an EMMO-based ontology or not, default `True`.
        prefix: defaults to self.get_namespace.name if
        prefix_emmo: bool, default None. If emmo_based is True it
            defaults to True and sets the prefix of all imported ontologies
            with base_iri starting with 'http://emmo.info/emmo' to emmo
        kwargs:
            Additional keyword arguments are passed on to
            owlready2.Ontology.load().
        """
        # TODO: make sure that `only_local` argument is respected...

        if self.loaded:
            return self
        self._load(
            only_local=only_local,
            filename=filename,
            format=format,
            reload=reload,
            reload_if_newer=reload_if_newer,
            url_from_catalog=url_from_catalog,
            catalog_file=catalog_file,
            **kwargs,
        )

        # Enable optimised search by get_by_label()
        if self._special_labels is None and emmo_based:
            for iri in DEFAULT_LABEL_ANNOTATIONS:
                self.add_label_annotation(iri)
            top = self.world["http://www.w3.org/2002/07/owl#topObjectProperty"]
            self._special_labels = {
                "Thing": owlready2.Thing,
                "Nothing": owlready2.Nothing,
                "topObjectProperty": top,
                "owl:Thing": owlready2.Thing,
                "owl:Nothing": owlready2.Nothing,
                "owl:topObjectProperty": top,
            }
        # set prefix if another prefix is desired
        # if we do this, shouldn't we make the name of all
        # entities of the given ontology to the same?
        if prefix:
            self.prefix = prefix
        else:
            self.prefix = self.name

        if emmo_based and prefix_emmo is None:
            prefix_emmo = True
        if prefix_emmo:
            self.set_common_prefix()

        return self

    def _load(  # pylint: disable=too-many-arguments,too-many-locals,too-many-branches,too-many-statements
        self,
        only_local=False,
        filename=None,
        format=None,  # pylint: disable=redefined-builtin
        reload=None,
        reload_if_newer=False,
        url_from_catalog=None,
        catalog_file="catalog-v001.xml",
        **kwargs,
    ):
        """Help function for load()."""
        web_protocol = "http://", "https://", "ftp://"
        url = str(filename) if filename else self.base_iri.rstrip("/#")
        if url.startswith(web_protocol):
            baseurl = os.path.dirname(url)
            catalogurl = baseurl + "/" + catalog_file
        else:
            if url.startswith("file://"):
                url = url[7:]
            url = os.path.normpath(os.path.abspath(url))
            baseurl = os.path.dirname(url)
            catalogurl = os.path.join(baseurl, catalog_file)

        def getmtime(path):
            if os.path.exists(path):
                return os.path.getmtime(path)
            return 0.0

        # Resolve url from catalog file
        iris = {}
        dirs = set()
        if url_from_catalog or url_from_catalog is None:
            not_reload = not reload and (
                not reload_if_newer
                or getmtime(catalogurl)
                > self.world._cached_catalogs[catalogurl][0]
            )
            # get iris from catalog already in cached catalogs
            if catalogurl in self.world._cached_catalogs and not_reload:
                _, iris, dirs = self.world._cached_catalogs[catalogurl]
            # do not update cached_catalogs if url already in _iri_mappings
            # and reload not forced
            elif url in self.world._iri_mappings and not_reload:
                pass
            # update iris from current catalogurl
            else:
                try:
                    iris, dirs = read_catalog(
                        uri=catalogurl,
                        recursive=False,
                        return_paths=True,
                        catalog_file=catalog_file,
                    )
                except ReadCatalogError:
                    if url_from_catalog is not None:
                        raise
                    self.world._cached_catalogs[catalogurl] = (0.0, {}, set())
                else:
                    self.world._cached_catalogs[catalogurl] = (
                        getmtime(catalogurl),
                        iris,
                        dirs,
                    )
            self.world._iri_mappings.update(iris)
        resolved_url = self.world._iri_mappings.get(url, url)
        # Append paths from catalog file to onto_path
        for path in sorted(dirs, reverse=True):
            if path not in owlready2.onto_path:
                owlready2.onto_path.append(path)

        # Use catalog file to update IRIs of imported ontologies
        # in internal store and try to load again...
        if self.world._iri_mappings:
            for abbrev_iri in self.world._get_obj_triples_sp_o(
                self.storid, owlready2.owl_imports
            ):
                iri = self._unabbreviate(abbrev_iri)
                if iri in self.world._iri_mappings:
                    self._del_obj_triple_spo(
                        self.storid, owlready2.owl_imports, abbrev_iri
                    )
                    self._add_obj_triple_spo(
                        self.storid,
                        owlready2.owl_imports,
                        self._abbreviate(self.world._iri_mappings[iri]),
                    )

        # Load ontology
        try:
            self.loaded = False
            fmt = format if format else guess_format(resolved_url, fmap=FMAP)
            if fmt and fmt not in OWLREADY2_FORMATS:
                # Convert filename to rdfxml before passing it to owlready2
                graph = rdflib.Graph()
                try:
                    graph.parse(resolved_url, format=fmt)
                except URLError as err:
                    raise EMMOntoPyException(
                        "URL error", err, resolved_url
                    ) from err

                with tempfile.NamedTemporaryFile() as handle:
                    graph.serialize(destination=handle, format="xml")
                    handle.seek(0)
                    return super().load(
                        only_local=True,
                        fileobj=handle,
                        reload=reload,
                        reload_if_newer=reload_if_newer,
                        format="rdfxml",
                        **kwargs,
                    )
            elif resolved_url.startswith(web_protocol):
                return super().load(
                    only_local=only_local,
                    reload=reload,
                    reload_if_newer=reload_if_newer,
                    **kwargs,
                )

            else:
                with open(resolved_url, "rb") as handle:
                    return super().load(
                        only_local=only_local,
                        fileobj=handle,
                        reload=reload,
                        reload_if_newer=reload_if_newer,
                        **kwargs,
                    )
        except owlready2.OwlReadyOntologyParsingError:
            # Owlready2 is not able to parse the ontology - most
            # likely because imported ontologies must be resolved
            # using the catalog file.

            # Reraise if we don't want to read from the catalog file
            if not url_from_catalog and url_from_catalog is not None:
                raise

            warnings.warn(
                "Recovering from Owlready2 parsing error... might be deprecated"
            )

            # Copy the ontology into a local folder and try again
            with tempfile.TemporaryDirectory() as handle:
                output = os.path.join(handle, os.path.basename(resolved_url))
                convert_imported(
                    input_ontology=resolved_url,
                    output_ontology=output,
                    input_format=fmt,
                    output_format="xml",
                    url_from_catalog=url_from_catalog,
                    catalog_file=catalog_file,
                )

                self.loaded = False
                with open(output, "rb") as handle:
                    try:
                        return super().load(
                            only_local=True,
                            fileobj=handle,
                            reload=reload,
                            reload_if_newer=reload_if_newer,
                            format="rdfxml",
                            **kwargs,
                        )
                    except HTTPError as exc:  # Add url to HTTPError message
                        raise HTTPError(
                            url=exc.url,
                            code=exc.code,
                            msg=f"{exc.url}: {exc.msg}",
                            hdrs=exc.hdrs,
                            fp=exc.fp,
                        ).with_traceback(exc.__traceback__)

        except HTTPError as exc:  # Add url to HTTPError message
            raise HTTPError(
                url=exc.url,
                code=exc.code,
                msg=f"{exc.url}: {exc.msg}",
                hdrs=exc.hdrs,
                fp=exc.fp,
            ).with_traceback(exc.__traceback__)

    def save(
        self,
        filename=None,
        format=None,
        dir=".",
        mkdir=False,
        overwrite=False,
        recursive=False,
        squash=False,
        write_catalog_file=False,
        append_catalog=False,
        catalog_file="catalog-v001.xml",
    ):
        """Writes the ontology to file.

        Parameters
        ----------
        filename: None | str | Path
            Name of file to write to.  If None, it defaults to the name
            of the ontology with `format` as file extension.
        format: str
            Output format. The default is to infer it from `filename`.
        dir: str | Path
            If `filename` is a relative path, it is a relative path to `dir`.
        mkdir: bool
            Whether to create output directory if it does not exists.
        owerwrite: bool
            If true and `filename` exists, remove the existing file before
            saving.  The default is to append to an existing ontology.
        recursive: bool
            Whether to save imported ontologies recursively.  This is
            commonly combined with `filename=None`, `dir` and `mkdir`.
        squash: bool
            If true, rdflib will be used to save the current ontology
            together with all its sub-ontologies into `filename`.
            It make no sense to combine this with `recursive`.
        write_catalog_file: bool
            Whether to also write a catalog file to disk.
        append_catalog: bool
            Whether to append to an existing catalog file.
        catalog_file: str | Path
            Name of catalog file.  If not an absolute path, it is prepended
            to `dir`.
        """
        # pylint: disable=redefined-builtin,too-many-arguments
        # pylint: disable=too-many-statements,too-many-branches
        # pylint: disable=too-many-locals,arguments-renamed
        if not _validate_installed_version(
            package="rdflib", min_version="6.0.0"
        ) and format == FMAP.get("ttl", ""):
            from rdflib import (  # pylint: disable=import-outside-toplevel
                __version__ as __rdflib_version__,
            )

            warnings.warn(
                IncompatibleVersion(
                    "To correctly convert to Turtle format, rdflib must be "
                    "version 6.0.0 or greater, however, the detected rdflib "
                    "version used by your Python interpreter is "
                    f"{__rdflib_version__!r}. For more information see the "
                    "'Known issues' section of the README."
                )
            )

        revmap = {value: key for key, value in FMAP.items()}

        if filename is None:
            if format:
                fmt = revmap.get(format, format)
                filename = f"{self.name}.{fmt}"
            else:
                raise TypeError("`filename` and `format` cannot both be None.")
        filename = os.path.join(dir, filename)
        dir = Path(filename).resolve().parent

        if mkdir:
            outdir = Path(filename).parent.resolve()
            if not outdir.exists():
                outdir.mkdir(parents=True)

        if not format:
            format = guess_format(filename, fmap=FMAP)
        fmt = revmap.get(format, format)

        if overwrite and filename and os.path.exists(filename):
            os.remove(filename)

        EMMO = rdflib.Namespace(  # pylint:disable=invalid-name
            "http://emmo.info/emmo#"
        )

        if recursive:
            if squash:
                raise ValueError(
                    "`recursive` and `squash` should not both be true"
                )
            base = self.base_iri.rstrip("#/")
            for onto in self.imported_ontologies:
                obase = onto.base_iri.rstrip("#/")
                newdir = Path(dir) / os.path.relpath(obase, base)
                onto.save(
                    filename=None,
                    format=format,
                    dir=newdir.resolve(),
                    mkdir=mkdir,
                    overwrite=overwrite,
                    recursive=recursive,
                    squash=squash,
                    write_catalog_file=write_catalog_file,
                    append_catalog=append_catalog,
                    catalog_file=catalog_file,
                )

        if squash:
            from rdflib import (  # pylint:disable=import-outside-toplevel
                URIRef,
                RDF,
                OWL,
            )

            graph = self.world.as_rdflib_graph()
            graph.namespace_manager.bind("emmo", EMMO)

            # Remove anonymous namespace and imports
            graph.remove((URIRef("http://anonymous"), RDF.type, OWL.Ontology))
            imports = list(graph.triples((None, OWL.imports, None)))
            for triple in imports:
                graph.remove(triple)

            graph.serialize(destination=filename, format=format)
        elif format in OWLREADY2_FORMATS:
            super().save(file=filename, format=fmt)
        else:
            # The try-finally clause is needed for cleanup and because
            # we have to provide delete=False to NamedTemporaryFile
            # since Windows does not allow to reopen an already open
            # file.
            try:
                with tempfile.NamedTemporaryFile(
                    suffix=".owl", delete=False
                ) as handle:
                    tmpfile = handle.name
                super().save(tmpfile, format="rdfxml")
                graph = rdflib.Graph()
                graph.parse(tmpfile, format="xml")
                graph.serialize(destination=filename, format=format)
            finally:
                os.remove(tmpfile)

        if write_catalog_file:
            mappings = {}
            base = self.base_iri.rstrip("#/")

            def append(onto):
                obase = onto.base_iri.rstrip("#/")
                newdir = Path(dir) / os.path.relpath(obase, base)
                newpath = newdir.resolve() / f"{onto.name}.{fmt}"
                relpath = os.path.relpath(newpath, dir)
                mappings[onto.get_version(as_iri=True)] = str(relpath)
                for imported in onto.imported_ontologies:
                    append(imported)

            if recursive:
                append(self)
            write_catalog(
                mappings,
                output=catalog_file,
                directory=dir,
                append=append_catalog,
            )

    def get_imported_ontologies(self, recursive=False):
        """Return a list with imported ontologies.

        If `recursive` is `True`, ontologies imported by imported ontologies
        are also returned.
        """

        def rec_imported(onto):
            for ontology in onto.imported_ontologies:
                if ontology not in imported:
                    imported.add(ontology)
                    rec_imported(ontology)

        if recursive:
            imported = set()
            rec_imported(self)
            return list(imported)

        return self.imported_ontologies

    def get_entities(  # pylint: disable=too-many-arguments
        self,
        imported=True,
        classes=True,
        individuals=True,
        object_properties=True,
        data_properties=True,
        annotation_properties=True,
    ):
        """Return a generator over (optionally) all classes, individuals,
        object_properties, data_properties and annotation_properties.

        If `imported` is `True`, entities in imported ontologies will also
        be included.
        """
        generator = []
        if classes:
            generator.append(self.classes(imported))
        if individuals:
            generator.append(self.individuals(imported))
        if object_properties:
            generator.append(self.object_properties(imported))
        if data_properties:
            generator.append(self.data_properties(imported))
        if annotation_properties:
            generator.append(self.annotation_properties(imported))
        for entity in itertools.chain(*generator):
            yield entity

    def classes(self, imported=False):
        """Returns an generator over all classes.

        If `imported` is `True`, will imported classes are also returned.
        """
        if imported:
            return self.world.classes()
        return super().classes()

    def individuals(self, imported=False):
        """Returns an generator over all individuals.

        If `imported` is `True`, will imported individuals are also returned.
        """
        if imported:
            return self.world.individuals()
        return super().individuals()

    def object_properties(self, imported=False):
        """Returns an generator over all object properties.

        If `imported` is true, will imported object properties are also
        returned.
        """
        if imported:
            return self.world.object_properties()
        return super().object_properties()

    def data_properties(self, imported=False):
        """Returns an generator over all data properties.

        If `imported` is true, will imported data properties are also
        returned.
        """
        if imported:
            return self.world.data_properties()
        return super().data_properties()

    def annotation_properties(self, imported=False):
        """Returns a generator iterating over all annotation properties
        defined in the current ontology.

        If `imported` is true, annotation properties in imported ontologies
        will also be included.
        """
        if imported:
            return self.world.annotation_properties()
        return super().annotation_properties()

    def get_root_classes(self, imported=False):
        """Returns a list or root classes."""
        return [
            cls
            for cls in self.classes(imported=imported)
            if not cls.ancestors().difference(set([cls, owlready2.Thing]))
        ]

    def get_root_object_properties(self, imported=False):
        """Returns a list of root object properties."""
        props = set(self.object_properties(imported=imported))
        return [p for p in props if not props.intersection(p.is_a)]

    def get_root_data_properties(self, imported=False):
        """Returns a list of root object properties."""
        props = set(self.data_properties(imported=imported))
        return [p for p in props if not props.intersection(p.is_a)]

    def get_roots(self, imported=False):
        """Returns all class, object_property and data_property roots."""
        roots = self.get_root_classes(imported=imported)
        roots.extend(self.get_root_object_properties(imported=imported))
        roots.extend(self.get_root_data_properties(imported=imported))
        return roots

    def sync_python_names(self, annotations=("prefLabel", "label", "altLabel")):
        """Update the `python_name` attribute of all properties.

        The python_name attribute will be set to the first non-empty
        annotation in the sequence of annotations in `annotations` for
        the property.
        """

        def update(gen):
            for prop in gen:
                for annotation in annotations:
                    if hasattr(prop, annotation) and getattr(prop, annotation):
                        prop.python_name = getattr(prop, annotation).first()
                        break

        update(
            self.get_entities(
                classes=False,
                individuals=False,
                object_properties=False,
                data_properties=False,
            )
        )
        update(
            self.get_entities(
                classes=False, individuals=False, annotation_properties=False
            )
        )

    def rename_entities(
        self,
        annotations=("prefLabel", "label", "altLabel"),
    ):
        """Set `name` of all entities to the first non-empty annotation in
        `annotations`.

        Warning, this method changes all IRIs in the ontology.  However,
        it may be useful to make the ontology more readable and to work
        with it together with a triple store.
        """
        for entity in self.get_entities():
            for annotation in annotations:
                if hasattr(entity, annotation):
                    name = getattr(entity, annotation).first()
                    if name:
                        entity.name = name
                        break

    def sync_reasoner(
        self, reasoner="FaCT++", include_imported=False, **kwargs
    ):
        """Update current ontology by running the given reasoner.

        Supported values for `reasoner` are 'Pellet', 'HermiT' and 'FaCT++'.

        If `include_imported` is true, the reasoner will also reason
        over imported ontologies.  Note that this may be **very** slow
        with Pellet and HermiT.

        Keyword arguments are passed to the underlying owlready2 function.
        """
        if reasoner == "FaCT++":
            sync = sync_reasoner_factpp
        elif reasoner == "Pellet":
            sync = owlready2.sync_reasoner_pellet
        elif reasoner == "HermiT":
            sync = owlready2.sync_reasoner_hermit
        else:
            raise ValueError(
                f"unknown reasoner {reasoner!r}. Supported reasoners "
                'are "Pellet", "HermiT" and "FaCT++".'
            )

        # For some reason we must visit all entities once before running
        # the reasoner...
        list(self.get_entities())

        with self:
            if include_imported:
                sync(self.world, **kwargs)
            else:
                sync(self, **kwargs)

    def sync_attributes(  # pylint: disable=too-many-branches
        self,
        name_policy=None,
        name_prefix="",
        class_docstring="comment",
        sync_imported=False,
    ):
        """This method is intended to be called after you have added new
        classes (typically via Python) to make sure that attributes like
        `label` and `comments` are defined.

        If a class, object property, data property or annotation
        property in the current ontology has no label, the name of
        the corresponding Python class will be assigned as label.

        If a class, object property, data property or annotation
        property has no comment, it will be assigned the docstring of
        the corresponding Python class.

        `name_policy` specify wether and how the names in the ontology
        should be updated.  Valid values are:
          None          not changed
          "uuid"        `name_prefix` followed by a global unique id (UUID).
                        If the name is already valid accoridng to this standard
                        it will not be regenerated.
          "sequential"  `name_prefix` followed a sequantial number.
        EMMO conventions imply ``name_policy=='uuid'``.

        If `sync_imported` is true, all imported ontologies are also
        updated.

        The `class_docstring` argument specifies the annotation that
        class docstrings are mapped to.  Defaults to "comment".
        """
        for cls in itertools.chain(
            self.classes(),
            self.object_properties(),
            self.data_properties(),
            self.annotation_properties(),
        ):
            if not hasattr(cls, "prefLabel"):
                # no prefLabel - create new annotation property..
                with self:
                    # pylint: disable=invalid-name,missing-class-docstring
                    # pylint: disable=unused-variable
                    class prefLabel(owlready2.label):
                        pass

                cls.prefLabel = [locstr(cls.__name__, lang="en")]
            elif not cls.prefLabel:
                cls.prefLabel.append(locstr(cls.__name__, lang="en"))
            if class_docstring and hasattr(cls, "__doc__") and cls.__doc__:
                getattr(cls, class_docstring).append(
                    locstr(inspect.cleandoc(cls.__doc__), lang="en")
                )

        for ind in self.individuals():
            if not hasattr(ind, "prefLabel"):
                # no prefLabel - create new annotation property..
                with self:
                    # pylint: disable=invalid-name,missing-class-docstring
                    # pylint: disable=function-redefined
                    class prefLabel(owlready2.label):
                        iri = "http://www.w3.org/2004/02/skos/core#prefLabel"

                ind.prefLabel = [locstr(ind.name, lang="en")]
            elif not ind.prefLabel:
                ind.prefLabel.append(locstr(ind.name, lang="en"))

        chain = itertools.chain(
            self.classes(),
            self.individuals(),
            self.object_properties(),
            self.data_properties(),
            self.annotation_properties(),
        )
        if name_policy == "uuid":
            for obj in chain:
                try:
                    # Passing the following means that the name is valid
                    # and need not be regenerated.
                    if not obj.name.startswith(name_prefix):
                        raise ValueError
                    uuid.UUID(obj.name.lstrip(name_prefix), version=5)
                except ValueError:
                    obj.name = name_prefix + str(
                        uuid.uuid5(uuid.NAMESPACE_DNS, obj.name)
                    )
        elif name_policy == "sequential":
            for obj in chain:
                counter = 0
                while f"{self.base_iri}{name_prefix}{counter}" in self:
                    counter += 1
                obj.name = f"{name_prefix}{counter}"
        elif name_policy is not None:
            raise TypeError(f"invalid name_policy: {name_policy!r}")

        if sync_imported:
            for onto in self.imported_ontologies:
                onto.sync_attributes()

    def get_relations(self):
        """Returns a generator for all relations."""
        warnings.warn(
            "Ontology.get_relations() is deprecated. Use "
            "onto.object_properties() instead.",
            DeprecationWarning,
        )
        return self.object_properties()

    def get_annotations(self, entity):
        """Returns a dict with annotations for `entity`.  Entity may be given
        either as a ThingClass object or as a label."""
        warnings.warn(
            "Ontology.get_annotations(entity) is deprecated. Use "
            "entity.get_annotations() instead.",
            DeprecationWarning,
        )

        if isinstance(entity, str):
            entity = self.get_by_label(entity)
        res = {"comment": getattr(entity, "comment", "")}
        for annotation in self.annotation_properties():
            res[annotation.label.first()] = [
                obj.strip('"')
                for _, _, obj in self.get_triples(
                    entity.storid, annotation.storid, None
                )
            ]
        return res

    def get_branch(  # pylint: disable=too-many-arguments
        self,
        root,
        leaves=(),
        include_leaves=True,
        strict_leaves=False,
        exclude=None,
        sort=False,
    ):
        """Returns a set with all direct and indirect subclasses of `root`.
        Any subclass found in the sequence `leaves` will be included in
        the returned list, but its subclasses will not.  The elements
        of `leaves` may be ThingClass objects or labels.

        Subclasses of any subclass found in the sequence `leaves` will
        be excluded from the returned list, where the elements of `leaves`
        may be ThingClass objects or labels.

        If `include_leaves` is true, the leaves are included in the returned
        list, otherwise they are not.

        If `strict_leaves` is true, any descendant of a leaf will be excluded
        in the returned set.

        If given, `exclude` may be a sequence of classes, including
        their subclasses, to exclude from the output.

        If `sort` is True, a list sorted according to depth and label
        will be returned instead of a set.
        """

        def _branch(root, leaves):
            if root not in leaves:
                branch = {
                    root,
                }
                for cls in root.subclasses():
                    # Defining a branch is actually quite tricky.  Consider
                    # the case:
                    #
                    #      L isA R
                    #      A isA L
                    #      A isA R
                    #
                    # where R is the root, L is a leaf and A is a direct
                    # child of both.  Logically, since A is a child of the
                    # leaf we want to skip A.  But a strait forward imple-
                    # mentation will see that A is a child of the root and
                    # include it.  Requireing that the R should be a strict
                    # parent of A solves this.
                    if root in cls.get_parents(strict=True):
                        branch.update(_branch(cls, leaves))
            else:
                branch = (
                    {
                        root,
                    }
                    if include_leaves
                    else set()
                )
            return branch

        if isinstance(root, str):
            root = self.get_by_label(root)

        leaves = set(
            self.get_by_label(leaf) if isinstance(leaf, str) else leaf
            for leaf in leaves
        )
        leaves.discard(root)

        if exclude:
            exclude = set(
                self.get_by_label(e) if isinstance(e, str) else e
                for e in exclude
            )
            leaves.update(exclude)

        branch = _branch(root, leaves)

        # Exclude all descendants of any leaf
        if strict_leaves:
            descendants = root.descendants()
            for leaf in leaves:
                if leaf in descendants:
                    branch.difference_update(
                        leaf.descendants(include_self=False)
                    )

        if exclude:
            branch.difference_update(exclude)

        # Sort according to depth, then by label
        if sort:
            branch = sorted(
                sorted(branch, key=asstring),
                key=lambda x: len(x.mro()),
            )

        return branch

    def is_individual(self, entity):
        """Returns true if entity is an individual."""
        if isinstance(entity, str):
            entity = self.get_by_label(entity)
        return isinstance(entity, owlready2.Thing)

    # FIXME - deprecate this method as soon the ThingClass property
    #         `defined_class` works correct in Owlready2
    def is_defined(self, entity):
        """Returns true if the entity is a defined class.

        Deprecated, use the `is_defined` property of the classes
        (ThingClass subclasses) instead.
        """
        warnings.warn(
            "This method is deprecated.  Use the `is_defined` property of "
            "the classes instad.",
            DeprecationWarning,
            stacklevel=2,
        )
        if isinstance(entity, str):
            entity = self.get_by_label(entity)
        return hasattr(entity, "equivalent_to") and bool(entity.equivalent_to)

    def get_version(self, as_iri=False) -> str:
        """Returns the version number of the ontology as inferred from the
        owl:versionIRI tag or, if owl:versionIRI is not found, from
        owl:versionINFO.

        If `as_iri` is True, the full versionIRI is returned.
        """
        version_iri_storid = self.world._abbreviate(
            "http://www.w3.org/2002/07/owl#versionIRI"
        )
        tokens = self.get_triples(s=self.storid, p=version_iri_storid)
        if (not tokens) and (as_iri is True):
            raise TypeError(
                "No owl:versionIRI "
                f"in Ontology {self.base_iri!r}. "
                "Search for owl:versionInfo with as_iri=False"
            )
        if tokens:
            _, _, obj = tokens[0]
            version_iri = self.world._unabbreviate(obj)
            if as_iri:
                return version_iri
            return infer_version(self.base_iri, version_iri)

        version_info_storid = self.world._abbreviate(
            "http://www.w3.org/2002/07/owl#versionInfo"
        )
        tokens = self.get_triples(s=self.storid, p=version_info_storid)
        if not tokens:
            raise TypeError(
                "No versionIRI or versionInfo " f"in Ontology {self.base_iri!r}"
            )
        _, _, version_info = tokens[0]
        return version_info.strip('"').strip("'")

    def set_version(self, version=None, version_iri=None):
        """Assign version to ontology by asigning owl:versionIRI.

        If `version` but not `version_iri` is provided, the version
        IRI will be the combination of `base_iri` and `version`.
        """
        _version_iri = "http://www.w3.org/2002/07/owl#versionIRI"
        version_iri_storid = self.world._abbreviate(_version_iri)
        if self._has_obj_triple_spo(  # pylint: disable=unexpected-keyword-arg
            # For some reason _has_obj_triples_spo exists in both
            # owlready2.namespace.Namespace (with arguments subject/predicate)
            # and in owlready2.triplelite._GraphManager (with arguments s/p)
            # owlready2.Ontology inherits from Namespace directly
            # and pylint checks that.
            # It actually accesses the one in triplelite.
            # subject=self.storid, predicate=version_iri_storid
            s=self.storid,
            p=version_iri_storid,
        ):
            self._del_obj_triple_spo(s=self.storid, p=version_iri_storid)

        if not version_iri:
            if not version:
                raise TypeError(
                    "Either `version` or `version_iri` must be provided"
                )
            head, tail = self.base_iri.rstrip("#/").rsplit("/", 1)
            version_iri = "/".join([head, version, tail])

        self._add_obj_triple_spo(
            s=self.storid,
            p=self.world._abbreviate(_version_iri),
            o=self.world._abbreviate(version_iri),
        )

    def get_graph(self, **kwargs):
        """Returns a new graph object.  See  emmo.graph.OntoGraph.

        Note that this method requires the Python graphviz package.
        """
        # pylint: disable=import-outside-toplevel,cyclic-import
        from ontopy.graph import OntoGraph

        return OntoGraph(self, **kwargs)

    @staticmethod
    def common_ancestors(cls1, cls2):
        """Return a list of common ancestors for `cls1` and `cls2`."""
        return set(cls1.ancestors()).intersection(cls2.ancestors())

    def number_of_generations(self, descendant, ancestor):
        """Return shortest distance from ancestor to descendant"""
        if ancestor not in descendant.ancestors():
            raise ValueError("Descendant is not a descendant of ancestor")
        return self._number_of_generations(descendant, ancestor, 0)

    def _number_of_generations(self, descendant, ancestor, counter):
        """Recursive help function to number_of_generations(), return
        distance between a ancestor-descendant pair (counter+1)."""
        if descendant.name == ancestor.name:
            return counter
        try:
            return min(
                self._number_of_generations(parent, ancestor, counter + 1)
                for parent in descendant.get_parents()
                if ancestor in parent.ancestors()
            )
        except ValueError:
            return counter

    def closest_common_ancestors(self, cls1, cls2):
        """Returns a list with closest_common_ancestor for cls1 and cls2"""
        distances = {}
        for ancestor in self.common_ancestors(cls1, cls2):
            distances[ancestor] = self.number_of_generations(
                cls1, ancestor
            ) + self.number_of_generations(cls2, ancestor)
        return [
            ancestor
            for ancestor, distance in distances.items()
            if distance == min(distances.values())
        ]

    @staticmethod
    def closest_common_ancestor(*classes):
        """Returns closest_common_ancestor for the given classes."""
        mros = [cls.mro() for cls in classes]
        track = defaultdict(int)
        while mros:
            for mro in mros:
                cur = mro.pop(0)
                track[cur] += 1
                if track[cur] == len(classes):
                    return cur
                if len(mro) == 0:
                    mros.remove(mro)
        raise EMMOntoPyException(
            "A closest common ancestor should always exist !"
        )

    def get_ancestors(
        self,
        classes: "Union[List, ThingClass]",
        closest: bool = False,
        generations: int = None,
        strict: bool = True,
    ) -> set:
        """Return ancestors of all classes in `classes`.
        Args:
            classes: class(es) for which ancestors should be returned.
            generations: Include this number of generations, default is all.
            closest: If True, return all ancestors up to and including the
                closest common ancestor. Return all if False.
            strict: If True returns only real ancestors, i.e. `classes` are
                are not included in the returned set.
        Returns:
            Set of ancestors to `classes`.
        """
        if not isinstance(classes, Iterable):
            classes = [classes]

        ancestors = set()
        if not classes:
            return ancestors

        def addancestors(entity, counter, subject):
            if counter > 0:
                for parent in entity.get_parents(strict=True):
                    subject.add(parent)
                    addancestors(parent, counter - 1, subject)

        if closest:
            if generations is not None:
                raise ValueError(
                    "Only one of `generations` or `closest` may be specified."
                )

            closest_ancestor = self.closest_common_ancestor(*classes)
            for cls in classes:
                ancestors.update(
                    anc
                    for anc in cls.ancestors()
                    if closest_ancestor in anc.ancestors()
                )
        elif isinstance(generations, int):
            for entity in classes:
                addancestors(entity, generations, ancestors)
        else:
            ancestors.update(*(cls.ancestors() for cls in classes))

        if strict:
            return ancestors.difference(classes)
        return ancestors

    def get_descendants(
        self,
        classes: "Union[List, ThingClass]",
        generations: int = None,
        common: bool = False,
    ) -> set:
        """Return descendants/subclasses of all classes in `classes`.
        Args:
            classes: class(es) for which descendants are desired.
            common: whether to only return descendants common to all classes.
            generations: Include this number of generations, default is all.
        Returns:
            A set of descendants for given number of generations.
            If 'common'=True, the common descendants are returned
            within the specified number of generations.
            'generations' defaults to all.
        """

        if not isinstance(classes, Iterable):
            classes = [classes]

        descendants = {name: [] for name in classes}

        def _children_recursively(num, newentity, parent, descendants):
            """Helper function to get all children up to generation."""
            for child in self.get_children_of(newentity):
                descendants[parent].append(child)
                if num < generations:
                    _children_recursively(num + 1, child, parent, descendants)

        if generations == 0:
            return set()

        if not generations:
            for entity in classes:
                descendants[entity] = entity.descendants()
                # only include proper descendants
                descendants[entity].remove(entity)
        else:
            for entity in classes:
                _children_recursively(1, entity, entity, descendants)

        results = descendants.values()
        if common is True:
            return set.intersection(*map(set, results))
        return set(flatten(results))

    def get_wu_palmer_measure(self, cls1, cls2):
        """Return Wu-Palmer measure for semantic similarity.

        Returns Wu-Palmer measure for semantic similarity between
        two concepts.
        Wu, Palmer; ACL 94: Proceedings of the 32nd annual meeting on
        Association for Computational Linguistics, June 1994.
        """
        cca = self.closest_common_ancestor(cls1, cls2)
        ccadepth = self.number_of_generations(cca, self.Thing)
        generations1 = self.number_of_generations(cls1, cca)
        generations2 = self.number_of_generations(cls2, cca)
        return 2 * ccadepth / (generations1 + generations2 + 2 * ccadepth)

    def new_entity(
        self, name: str, parent: Union[ThingClass, Iterable]
    ) -> ThingClass:
        """Create and return new entity

        Makes a new entity in the ontology with given parent(s).
        Return the new entity.

        Throws exception if name consists of more than one word.
        """
        if " " in name:
            raise LabelDefinitionError(
                f"Error in label name definition '{name}': "
                f"Label consists of more than one word."
            )
        parents = tuple(parent) if isinstance(parent, Iterable) else (parent,)
        for thing in parents:
            if not isinstance(thing, owlready2.ThingClass):
                raise ThingClassDefinitionError(
                    f"Error in parent definition: "
                    f"'{thing}' is not an owlready2.ThingClass."
                )

        with self:
            entity = types.new_class(name, parents)
        return entity

colon_in_label property writable

Whether to accept colon in name-part of IRI. If true, the name cannot be prefixed.

dir_imported property writable

Whether to include imported ontologies in dir() listing.

dir_label property writable

Whether to include entity label in dir() listing.

dir_name property writable

Whether to include entity name in dir() listing.

dir_preflabel property writable

Whether to include entity prefLabel in dir() listing.

label_annotations property readonly

List of label annotation searched for by get_by_label().

add_label_annotation(self, iri)

Adds label annotation used by get_by_label().

May be provided either as an IRI or as its owlready2 representation.

Source code in ontopy/ontology.py
def add_label_annotation(self, iri):
    """Adds label annotation used by get_by_label().

    May be provided either as an IRI or as its owlready2 representation.
    """
    if self._label_annotations is None:
        self._label_annotations = []
    label_annotation = iri if hasattr(iri, "storid") else self.world[iri]
    if label_annotation is None:
        warnings.warn(f"adding new IRI to ontology: {iri}")
        name = iri.rsplit("/")[-1].rsplit("#")[-1]
        bases = (owlready2.AnnotationProperty,)
        with self:
            label_annotation = types.new_class(name, bases)
    if label_annotation not in self._label_annotations:
        self._label_annotations.append(label_annotation)

annotation_properties(self, imported=False)

Returns a generator iterating over all annotation properties defined in the current ontology.

If imported is true, annotation properties in imported ontologies will also be included.

Source code in ontopy/ontology.py
def annotation_properties(self, imported=False):
    """Returns a generator iterating over all annotation properties
    defined in the current ontology.

    If `imported` is true, annotation properties in imported ontologies
    will also be included.
    """
    if imported:
        return self.world.annotation_properties()
    return super().annotation_properties()

classes(self, imported=False)

Returns an generator over all classes.

If imported is True, will imported classes are also returned.

Source code in ontopy/ontology.py
def classes(self, imported=False):
    """Returns an generator over all classes.

    If `imported` is `True`, will imported classes are also returned.
    """
    if imported:
        return self.world.classes()
    return super().classes()

closest_common_ancestor(*classes) staticmethod

Returns closest_common_ancestor for the given classes.

Source code in ontopy/ontology.py
@staticmethod
def closest_common_ancestor(*classes):
    """Returns closest_common_ancestor for the given classes."""
    mros = [cls.mro() for cls in classes]
    track = defaultdict(int)
    while mros:
        for mro in mros:
            cur = mro.pop(0)
            track[cur] += 1
            if track[cur] == len(classes):
                return cur
            if len(mro) == 0:
                mros.remove(mro)
    raise EMMOntoPyException(
        "A closest common ancestor should always exist !"
    )

closest_common_ancestors(self, cls1, cls2)

Returns a list with closest_common_ancestor for cls1 and cls2

Source code in ontopy/ontology.py
def closest_common_ancestors(self, cls1, cls2):
    """Returns a list with closest_common_ancestor for cls1 and cls2"""
    distances = {}
    for ancestor in self.common_ancestors(cls1, cls2):
        distances[ancestor] = self.number_of_generations(
            cls1, ancestor
        ) + self.number_of_generations(cls2, ancestor)
    return [
        ancestor
        for ancestor, distance in distances.items()
        if distance == min(distances.values())
    ]

common_ancestors(cls1, cls2) staticmethod

Return a list of common ancestors for cls1 and cls2.

Source code in ontopy/ontology.py
@staticmethod
def common_ancestors(cls1, cls2):
    """Return a list of common ancestors for `cls1` and `cls2`."""
    return set(cls1.ancestors()).intersection(cls2.ancestors())

data_properties(self, imported=False)

Returns an generator over all data properties.

If imported is true, will imported data properties are also returned.

Source code in ontopy/ontology.py
def data_properties(self, imported=False):
    """Returns an generator over all data properties.

    If `imported` is true, will imported data properties are also
    returned.
    """
    if imported:
        return self.world.data_properties()
    return super().data_properties()

get_ancestors(self, classes, closest=False, generations=None, strict=True)

Return ancestors of all classes in classes.

Parameters:

Name Type Description Default
classes Union[List, ThingClass]

class(es) for which ancestors should be returned.

required
generations int

Include this number of generations, default is all.

None
closest bool

If True, return all ancestors up to and including the closest common ancestor. Return all if False.

False
strict bool

If True returns only real ancestors, i.e. classes are are not included in the returned set.

True

Returns:

Type Description
set

Set of ancestors to classes.

Source code in ontopy/ontology.py
def get_ancestors(
    self,
    classes: "Union[List, ThingClass]",
    closest: bool = False,
    generations: int = None,
    strict: bool = True,
) -> set:
    """Return ancestors of all classes in `classes`.
    Args:
        classes: class(es) for which ancestors should be returned.
        generations: Include this number of generations, default is all.
        closest: If True, return all ancestors up to and including the
            closest common ancestor. Return all if False.
        strict: If True returns only real ancestors, i.e. `classes` are
            are not included in the returned set.
    Returns:
        Set of ancestors to `classes`.
    """
    if not isinstance(classes, Iterable):
        classes = [classes]

    ancestors = set()
    if not classes:
        return ancestors

    def addancestors(entity, counter, subject):
        if counter > 0:
            for parent in entity.get_parents(strict=True):
                subject.add(parent)
                addancestors(parent, counter - 1, subject)

    if closest:
        if generations is not None:
            raise ValueError(
                "Only one of `generations` or `closest` may be specified."
            )

        closest_ancestor = self.closest_common_ancestor(*classes)
        for cls in classes:
            ancestors.update(
                anc
                for anc in cls.ancestors()
                if closest_ancestor in anc.ancestors()
            )
    elif isinstance(generations, int):
        for entity in classes:
            addancestors(entity, generations, ancestors)
    else:
        ancestors.update(*(cls.ancestors() for cls in classes))

    if strict:
        return ancestors.difference(classes)
    return ancestors

get_annotations(self, entity)

Returns a dict with annotations for entity. Entity may be given either as a ThingClass object or as a label.

Source code in ontopy/ontology.py
def get_annotations(self, entity):
    """Returns a dict with annotations for `entity`.  Entity may be given
    either as a ThingClass object or as a label."""
    warnings.warn(
        "Ontology.get_annotations(entity) is deprecated. Use "
        "entity.get_annotations() instead.",
        DeprecationWarning,
    )

    if isinstance(entity, str):
        entity = self.get_by_label(entity)
    res = {"comment": getattr(entity, "comment", "")}
    for annotation in self.annotation_properties():
        res[annotation.label.first()] = [
            obj.strip('"')
            for _, _, obj in self.get_triples(
                entity.storid, annotation.storid, None
            )
        ]
    return res

get_branch(self, root, leaves=(), include_leaves=True, strict_leaves=False, exclude=None, sort=False)

Returns a set with all direct and indirect subclasses of root. Any subclass found in the sequence leaves will be included in the returned list, but its subclasses will not. The elements of leaves may be ThingClass objects or labels.

Subclasses of any subclass found in the sequence leaves will be excluded from the returned list, where the elements of leaves may be ThingClass objects or labels.

If include_leaves is true, the leaves are included in the returned list, otherwise they are not.

If strict_leaves is true, any descendant of a leaf will be excluded in the returned set.

If given, exclude may be a sequence of classes, including their subclasses, to exclude from the output.

If sort is True, a list sorted according to depth and label will be returned instead of a set.

Source code in ontopy/ontology.py
def get_branch(  # pylint: disable=too-many-arguments
    self,
    root,
    leaves=(),
    include_leaves=True,
    strict_leaves=False,
    exclude=None,
    sort=False,
):
    """Returns a set with all direct and indirect subclasses of `root`.
    Any subclass found in the sequence `leaves` will be included in
    the returned list, but its subclasses will not.  The elements
    of `leaves` may be ThingClass objects or labels.

    Subclasses of any subclass found in the sequence `leaves` will
    be excluded from the returned list, where the elements of `leaves`
    may be ThingClass objects or labels.

    If `include_leaves` is true, the leaves are included in the returned
    list, otherwise they are not.

    If `strict_leaves` is true, any descendant of a leaf will be excluded
    in the returned set.

    If given, `exclude` may be a sequence of classes, including
    their subclasses, to exclude from the output.

    If `sort` is True, a list sorted according to depth and label
    will be returned instead of a set.
    """

    def _branch(root, leaves):
        if root not in leaves:
            branch = {
                root,
            }
            for cls in root.subclasses():
                # Defining a branch is actually quite tricky.  Consider
                # the case:
                #
                #      L isA R
                #      A isA L
                #      A isA R
                #
                # where R is the root, L is a leaf and A is a direct
                # child of both.  Logically, since A is a child of the
                # leaf we want to skip A.  But a strait forward imple-
                # mentation will see that A is a child of the root and
                # include it.  Requireing that the R should be a strict
                # parent of A solves this.
                if root in cls.get_parents(strict=True):
                    branch.update(_branch(cls, leaves))
        else:
            branch = (
                {
                    root,
                }
                if include_leaves
                else set()
            )
        return branch

    if isinstance(root, str):
        root = self.get_by_label(root)

    leaves = set(
        self.get_by_label(leaf) if isinstance(leaf, str) else leaf
        for leaf in leaves
    )
    leaves.discard(root)

    if exclude:
        exclude = set(
            self.get_by_label(e) if isinstance(e, str) else e
            for e in exclude
        )
        leaves.update(exclude)

    branch = _branch(root, leaves)

    # Exclude all descendants of any leaf
    if strict_leaves:
        descendants = root.descendants()
        for leaf in leaves:
            if leaf in descendants:
                branch.difference_update(
                    leaf.descendants(include_self=False)
                )

    if exclude:
        branch.difference_update(exclude)

    # Sort according to depth, then by label
    if sort:
        branch = sorted(
            sorted(branch, key=asstring),
            key=lambda x: len(x.mro()),
        )

    return branch

get_by_label(self, label, label_annotations=None, prefix=None, imported=True, colon_in_label=None)

Returns entity with label annotation label.

Parameters:

Name Type Description Default
label str

label so serach for. May be written as 'label' or 'prefix:label'. get_by_label('prefix:label') == get_by_label('label', prefix='prefix').

required
label_annotations str

a sequence of label annotation names to look up. Defaults to the label_annotations property.

None
prefix str

if provided, it should be the last component of the base iri of an ontology (with trailing slash (/) or hash (#) stripped off). The search for a matching label will be limited to this namespace.

None
imported bool

Whether to also look for label in imported ontologies.

True
colon_in_label bool

Whether to accept colon (:) in a label or name-part of IRI. Defaults to the colon_in_label property of self. Setting this true cannot be combined with prefix.

None

If several entities have the same label, only the one which is found first is returned.Use get_by_label_all() to get all matches.

Note, if different prefixes are provided in the label and via the prefix argument a warning will be issued and the prefix argument will take precedence.

A NoSuchLabelError is raised if label cannot be found.

Source code in ontopy/ontology.py
def get_by_label(
    self,
    label: str,
    label_annotations: str = None,
    prefix: str = None,
    imported: bool = True,
    colon_in_label: bool = None,
):
    """Returns entity with label annotation `label`.

    Args:
       label: label so serach for.
           May be written as 'label' or 'prefix:label'.
           get_by_label('prefix:label') ==
           get_by_label('label', prefix='prefix').
       label_annotations: a sequence of label annotation names to look up.
           Defaults to the `label_annotations` property.
       prefix: if provided, it should be the last component of
           the base iri of an ontology (with trailing slash (/) or hash
           (#) stripped off).  The search for a matching label will be
           limited to this namespace.
       imported: Whether to also look for `label` in imported ontologies.
       colon_in_label: Whether to accept colon (:) in a label or name-part
           of IRI.  Defaults to the `colon_in_label` property of `self`.
           Setting this true cannot be combined with `prefix`.

    If several entities have the same label, only the one which is
    found first is returned.Use get_by_label_all() to get all matches.

    Note, if different prefixes are provided in the label and via
    the `prefix` argument a warning will be issued and the
    `prefix` argument will take precedence.

    A NoSuchLabelError is raised if `label` cannot be found.
    """
    # pylint: disable=too-many-arguments,too-many-branches,invalid-name
    if not isinstance(label, str):
        raise TypeError(
            f"Invalid label definition, must be a string: {label!r}"
        )

    if self._label_annotations is None:
        for iri in DEFAULT_LABEL_ANNOTATIONS:
            try:
                self.add_label_annotation(iri)
            except ValueError:
                pass

    if colon_in_label is None:
        colon_in_label = self._colon_in_label
    if colon_in_label:
        if prefix:
            raise ValueError(
                "`prefix` cannot be combined with `colon_in_label`"
            )
    else:
        splitlabel = label.split(":", 1)
        if len(splitlabel) == 2 and not splitlabel[1].startswith("//"):
            label = splitlabel[1]
            if prefix and prefix != splitlabel[0]:
                warnings.warn(
                    f"Prefix given both as argument ({prefix}) "
                    f"and in label ({splitlabel[0]}). "
                    "Prefix given in argument takes presendence "
                )
            if not prefix:
                prefix = splitlabel[0]

    if prefix:
        entitylist = self.get_by_label_all(
            label,
            label_annotations=label_annotations,
            prefix=prefix,
        )
        if len(entitylist) > 0:
            return entitylist[0]

        raise NoSuchLabelError(
            f"No label annotations matches {label!r} with prefix "
            f"{prefix!r}"
        )

    # Label is a full IRI
    entity = self.world[label]
    if entity:
        return entity

    # First entity with matching label annotation
    annotation_ids = (
        (self._abbreviate(ann, False) for ann in label_annotations)
        if label_annotations
        else (ann.storid for ann in self.label_annotations)
    )
    get_triples = (
        self.world._get_data_triples_spod_spod
        if imported
        else self._get_data_triples_spod_spod
    )
    for annotation_id in annotation_ids:
        for s, _, _, _ in get_triples(None, annotation_id, label, None):
            return self.world[self._unabbreviate(s)]

    # Special labels
    if self._special_labels and label in self._special_labels:
        return self._special_labels[label]

    # Check if label is a name under base_iri
    entity = self.world[self.base_iri + label]
    if entity:
        return entity

    # Check if label is a name in any namespace
    for namespace in self._namespaces.keys():
        entity = self.world[namespace + label]
        if entity:
            return entity

    raise NoSuchLabelError(f"No label annotations matches '{label}'")

get_by_label_all(self, label, label_annotations=None, prefix=None)

Like get_by_label(), but returns a list with all matching labels.

Returns an empty list if no matches could be found.

Note

The current implementation also supports "*" as a wildcard matching any number of characters. This may change in the future.

Source code in ontopy/ontology.py
def get_by_label_all(self, label, label_annotations=None, prefix=None):
    """Like get_by_label(), but returns a list with all matching labels.

    Returns an empty list if no matches could be found.

    Note
    ----
    The current implementation also supports "*" as a wildcard
    matching any number of characters. This may change in the future.
    """
    if not isinstance(label, str):
        raise TypeError(
            f"Invalid label definition, " f"must be a string: {label!r}"
        )
    if " " in label:
        raise ValueError(
            f"Invalid label definition, {label!r} contains spaces."
        )

    if label_annotations is None:
        annotations = (_.name for _ in self.label_annotations)
    else:
        annotations = (
            _.name if hasattr(_, "storid") else _ for _ in label_annotations
        )
    entity = self.world.search(**{next(annotations): label})
    for key in annotations:
        entity.extend(self.world.search(**{key: label}))

    if self._special_labels and label in self._special_labels:
        entity.append(self._special_labels[label])

    entity_accessed_directly = self.world[self.base_iri + label]
    if entity_accessed_directly and entity_accessed_directly not in entity:
        entity.append(entity_accessed_directly)

    if prefix:
        return [_ for _ in entity if _.namespace.ontology.prefix == prefix]
    return entity

get_descendants(self, classes, generations=None, common=False)

Return descendants/subclasses of all classes in classes.

Parameters:

Name Type Description Default
classes Union[List, ThingClass]

class(es) for which descendants are desired.

required
common bool

whether to only return descendants common to all classes.

False
generations int

Include this number of generations, default is all.

None

Returns:

Type Description
set

A set of descendants for given number of generations. If 'common'=True, the common descendants are returned within the specified number of generations. 'generations' defaults to all.

Source code in ontopy/ontology.py
def get_descendants(
    self,
    classes: "Union[List, ThingClass]",
    generations: int = None,
    common: bool = False,
) -> set:
    """Return descendants/subclasses of all classes in `classes`.
    Args:
        classes: class(es) for which descendants are desired.
        common: whether to only return descendants common to all classes.
        generations: Include this number of generations, default is all.
    Returns:
        A set of descendants for given number of generations.
        If 'common'=True, the common descendants are returned
        within the specified number of generations.
        'generations' defaults to all.
    """

    if not isinstance(classes, Iterable):
        classes = [classes]

    descendants = {name: [] for name in classes}

    def _children_recursively(num, newentity, parent, descendants):
        """Helper function to get all children up to generation."""
        for child in self.get_children_of(newentity):
            descendants[parent].append(child)
            if num < generations:
                _children_recursively(num + 1, child, parent, descendants)

    if generations == 0:
        return set()

    if not generations:
        for entity in classes:
            descendants[entity] = entity.descendants()
            # only include proper descendants
            descendants[entity].remove(entity)
    else:
        for entity in classes:
            _children_recursively(1, entity, entity, descendants)

    results = descendants.values()
    if common is True:
        return set.intersection(*map(set, results))
    return set(flatten(results))

get_entities(self, imported=True, classes=True, individuals=True, object_properties=True, data_properties=True, annotation_properties=True)

Return a generator over (optionally) all classes, individuals, object_properties, data_properties and annotation_properties.

If imported is True, entities in imported ontologies will also be included.

Source code in ontopy/ontology.py
def get_entities(  # pylint: disable=too-many-arguments
    self,
    imported=True,
    classes=True,
    individuals=True,
    object_properties=True,
    data_properties=True,
    annotation_properties=True,
):
    """Return a generator over (optionally) all classes, individuals,
    object_properties, data_properties and annotation_properties.

    If `imported` is `True`, entities in imported ontologies will also
    be included.
    """
    generator = []
    if classes:
        generator.append(self.classes(imported))
    if individuals:
        generator.append(self.individuals(imported))
    if object_properties:
        generator.append(self.object_properties(imported))
    if data_properties:
        generator.append(self.data_properties(imported))
    if annotation_properties:
        generator.append(self.annotation_properties(imported))
    for entity in itertools.chain(*generator):
        yield entity

get_graph(self, **kwargs)

Returns a new graph object. See emmo.graph.OntoGraph.

Note that this method requires the Python graphviz package.

Source code in ontopy/ontology.py
def get_graph(self, **kwargs):
    """Returns a new graph object.  See  emmo.graph.OntoGraph.

    Note that this method requires the Python graphviz package.
    """
    # pylint: disable=import-outside-toplevel,cyclic-import
    from ontopy.graph import OntoGraph

    return OntoGraph(self, **kwargs)

get_imported_ontologies(self, recursive=False)

Return a list with imported ontologies.

If recursive is True, ontologies imported by imported ontologies are also returned.

Source code in ontopy/ontology.py
def get_imported_ontologies(self, recursive=False):
    """Return a list with imported ontologies.

    If `recursive` is `True`, ontologies imported by imported ontologies
    are also returned.
    """

    def rec_imported(onto):
        for ontology in onto.imported_ontologies:
            if ontology not in imported:
                imported.add(ontology)
                rec_imported(ontology)

    if recursive:
        imported = set()
        rec_imported(self)
        return list(imported)

    return self.imported_ontologies

get_relations(self)

Returns a generator for all relations.

Source code in ontopy/ontology.py
def get_relations(self):
    """Returns a generator for all relations."""
    warnings.warn(
        "Ontology.get_relations() is deprecated. Use "
        "onto.object_properties() instead.",
        DeprecationWarning,
    )
    return self.object_properties()

get_root_classes(self, imported=False)

Returns a list or root classes.

Source code in ontopy/ontology.py
def get_root_classes(self, imported=False):
    """Returns a list or root classes."""
    return [
        cls
        for cls in self.classes(imported=imported)
        if not cls.ancestors().difference(set([cls, owlready2.Thing]))
    ]

get_root_data_properties(self, imported=False)

Returns a list of root object properties.

Source code in ontopy/ontology.py
def get_root_data_properties(self, imported=False):
    """Returns a list of root object properties."""
    props = set(self.data_properties(imported=imported))
    return [p for p in props if not props.intersection(p.is_a)]

get_root_object_properties(self, imported=False)

Returns a list of root object properties.

Source code in ontopy/ontology.py
def get_root_object_properties(self, imported=False):
    """Returns a list of root object properties."""
    props = set(self.object_properties(imported=imported))
    return [p for p in props if not props.intersection(p.is_a)]

get_roots(self, imported=False)

Returns all class, object_property and data_property roots.

Source code in ontopy/ontology.py
def get_roots(self, imported=False):
    """Returns all class, object_property and data_property roots."""
    roots = self.get_root_classes(imported=imported)
    roots.extend(self.get_root_object_properties(imported=imported))
    roots.extend(self.get_root_data_properties(imported=imported))
    return roots

get_unabbreviated_triples(self, subject=None, predicate=None, obj=None, blank=None)

Returns all matching triples unabbreviated.

If blank is given, it will be used to represent blank nodes.

Source code in ontopy/ontology.py
def get_unabbreviated_triples(
    self, subject=None, predicate=None, obj=None, blank=None
):
    """Returns all matching triples unabbreviated.

    If `blank` is given, it will be used to represent blank nodes.
    """
    # pylint: disable=invalid-name
    return _get_unabbreviated_triples(
        self, subject=subject, predicate=predicate, obj=obj, blank=blank
    )

get_version(self, as_iri=False)

Returns the version number of the ontology as inferred from the owl:versionIRI tag or, if owl:versionIRI is not found, from owl:versionINFO.

If as_iri is True, the full versionIRI is returned.

Source code in ontopy/ontology.py
def get_version(self, as_iri=False) -> str:
    """Returns the version number of the ontology as inferred from the
    owl:versionIRI tag or, if owl:versionIRI is not found, from
    owl:versionINFO.

    If `as_iri` is True, the full versionIRI is returned.
    """
    version_iri_storid = self.world._abbreviate(
        "http://www.w3.org/2002/07/owl#versionIRI"
    )
    tokens = self.get_triples(s=self.storid, p=version_iri_storid)
    if (not tokens) and (as_iri is True):
        raise TypeError(
            "No owl:versionIRI "
            f"in Ontology {self.base_iri!r}. "
            "Search for owl:versionInfo with as_iri=False"
        )
    if tokens:
        _, _, obj = tokens[0]
        version_iri = self.world._unabbreviate(obj)
        if as_iri:
            return version_iri
        return infer_version(self.base_iri, version_iri)

    version_info_storid = self.world._abbreviate(
        "http://www.w3.org/2002/07/owl#versionInfo"
    )
    tokens = self.get_triples(s=self.storid, p=version_info_storid)
    if not tokens:
        raise TypeError(
            "No versionIRI or versionInfo " f"in Ontology {self.base_iri!r}"
        )
    _, _, version_info = tokens[0]
    return version_info.strip('"').strip("'")

get_wu_palmer_measure(self, cls1, cls2)

Return Wu-Palmer measure for semantic similarity.

Returns Wu-Palmer measure for semantic similarity between two concepts. Wu, Palmer; ACL 94: Proceedings of the 32nd annual meeting on Association for Computational Linguistics, June 1994.

Source code in ontopy/ontology.py
def get_wu_palmer_measure(self, cls1, cls2):
    """Return Wu-Palmer measure for semantic similarity.

    Returns Wu-Palmer measure for semantic similarity between
    two concepts.
    Wu, Palmer; ACL 94: Proceedings of the 32nd annual meeting on
    Association for Computational Linguistics, June 1994.
    """
    cca = self.closest_common_ancestor(cls1, cls2)
    ccadepth = self.number_of_generations(cca, self.Thing)
    generations1 = self.number_of_generations(cls1, cca)
    generations2 = self.number_of_generations(cls2, cca)
    return 2 * ccadepth / (generations1 + generations2 + 2 * ccadepth)

individuals(self, imported=False)

Returns an generator over all individuals.

If imported is True, will imported individuals are also returned.

Source code in ontopy/ontology.py
def individuals(self, imported=False):
    """Returns an generator over all individuals.

    If `imported` is `True`, will imported individuals are also returned.
    """
    if imported:
        return self.world.individuals()
    return super().individuals()

is_defined(self, entity)

Returns true if the entity is a defined class.

Deprecated, use the is_defined property of the classes (ThingClass subclasses) instead.

Source code in ontopy/ontology.py
def is_defined(self, entity):
    """Returns true if the entity is a defined class.

    Deprecated, use the `is_defined` property of the classes
    (ThingClass subclasses) instead.
    """
    warnings.warn(
        "This method is deprecated.  Use the `is_defined` property of "
        "the classes instad.",
        DeprecationWarning,
        stacklevel=2,
    )
    if isinstance(entity, str):
        entity = self.get_by_label(entity)
    return hasattr(entity, "equivalent_to") and bool(entity.equivalent_to)

is_individual(self, entity)

Returns true if entity is an individual.

Source code in ontopy/ontology.py
def is_individual(self, entity):
    """Returns true if entity is an individual."""
    if isinstance(entity, str):
        entity = self.get_by_label(entity)
    return isinstance(entity, owlready2.Thing)

load(self, only_local=False, filename=None, format=None, reload=None, reload_if_newer=False, url_from_catalog=None, catalog_file='catalog-v001.xml', emmo_based=True, prefix=None, prefix_emmo=None, **kwargs)

Load the ontology.

Parameters

bool

Whether to only read local files. This requires that you have appended the path to the ontology to owlready2.onto_path.

str

Path to file to load the ontology from. Defaults to base_iri provided to get_ontology().

str

Format of filename. Default is inferred from filename extension.

bool

Whether to reload the ontology if it is already loaded.

bool

Whether to reload the ontology if the source has changed since last time it was loaded.

bool | None

Whether to use catalog file to resolve the location of base_iri. If None, the catalog file is used if it exists in the same directory as filename.

str

Name of Protègè catalog file in the same folder as the ontology. This option is used together with only_local and defaults to "catalog-v001.xml".

bool

Whether this is an EMMO-based ontology or not, default True.

prefix: defaults to self.get_namespace.name if

bool, default None. If emmo_based is True it

defaults to True and sets the prefix of all imported ontologies with base_iri starting with 'http://emmo.info/emmo' to emmo

Kwargs

Additional keyword arguments are passed on to owlready2.Ontology.load().

Source code in ontopy/ontology.py
def load(  # pylint: disable=too-many-arguments,arguments-renamed
    self,
    only_local=False,
    filename=None,
    format=None,  # pylint: disable=redefined-builtin
    reload=None,
    reload_if_newer=False,
    url_from_catalog=None,
    catalog_file="catalog-v001.xml",
    emmo_based=True,
    prefix=None,
    prefix_emmo=None,
    **kwargs,
):
    """Load the ontology.

    Parameters
    ----------
    only_local: bool
        Whether to only read local files.  This requires that you
        have appended the path to the ontology to owlready2.onto_path.
    filename: str
        Path to file to load the ontology from.  Defaults to `base_iri`
        provided to get_ontology().
    format: str
        Format of `filename`.  Default is inferred from `filename`
        extension.
    reload: bool
        Whether to reload the ontology if it is already loaded.
    reload_if_newer: bool
        Whether to reload the ontology if the source has changed since
        last time it was loaded.
    url_from_catalog: bool | None
        Whether to use catalog file to resolve the location of `base_iri`.
        If None, the catalog file is used if it exists in the same
        directory as `filename`.
    catalog_file: str
        Name of Protègè catalog file in the same folder as the
        ontology.  This option is used together with `only_local` and
        defaults to "catalog-v001.xml".
    emmo_based: bool
        Whether this is an EMMO-based ontology or not, default `True`.
    prefix: defaults to self.get_namespace.name if
    prefix_emmo: bool, default None. If emmo_based is True it
        defaults to True and sets the prefix of all imported ontologies
        with base_iri starting with 'http://emmo.info/emmo' to emmo
    kwargs:
        Additional keyword arguments are passed on to
        owlready2.Ontology.load().
    """
    # TODO: make sure that `only_local` argument is respected...

    if self.loaded:
        return self
    self._load(
        only_local=only_local,
        filename=filename,
        format=format,
        reload=reload,
        reload_if_newer=reload_if_newer,
        url_from_catalog=url_from_catalog,
        catalog_file=catalog_file,
        **kwargs,
    )

    # Enable optimised search by get_by_label()
    if self._special_labels is None and emmo_based:
        for iri in DEFAULT_LABEL_ANNOTATIONS:
            self.add_label_annotation(iri)
        top = self.world["http://www.w3.org/2002/07/owl#topObjectProperty"]
        self._special_labels = {
            "Thing": owlready2.Thing,
            "Nothing": owlready2.Nothing,
            "topObjectProperty": top,
            "owl:Thing": owlready2.Thing,
            "owl:Nothing": owlready2.Nothing,
            "owl:topObjectProperty": top,
        }
    # set prefix if another prefix is desired
    # if we do this, shouldn't we make the name of all
    # entities of the given ontology to the same?
    if prefix:
        self.prefix = prefix
    else:
        self.prefix = self.name

    if emmo_based and prefix_emmo is None:
        prefix_emmo = True
    if prefix_emmo:
        self.set_common_prefix()

    return self

new_entity(self, name, parent)

Create and return new entity

Makes a new entity in the ontology with given parent(s). Return the new entity.

Throws exception if name consists of more than one word.

Source code in ontopy/ontology.py
def new_entity(
    self, name: str, parent: Union[ThingClass, Iterable]
) -> ThingClass:
    """Create and return new entity

    Makes a new entity in the ontology with given parent(s).
    Return the new entity.

    Throws exception if name consists of more than one word.
    """
    if " " in name:
        raise LabelDefinitionError(
            f"Error in label name definition '{name}': "
            f"Label consists of more than one word."
        )
    parents = tuple(parent) if isinstance(parent, Iterable) else (parent,)
    for thing in parents:
        if not isinstance(thing, owlready2.ThingClass):
            raise ThingClassDefinitionError(
                f"Error in parent definition: "
                f"'{thing}' is not an owlready2.ThingClass."
            )

    with self:
        entity = types.new_class(name, parents)
    return entity

number_of_generations(self, descendant, ancestor)

Return shortest distance from ancestor to descendant

Source code in ontopy/ontology.py
def number_of_generations(self, descendant, ancestor):
    """Return shortest distance from ancestor to descendant"""
    if ancestor not in descendant.ancestors():
        raise ValueError("Descendant is not a descendant of ancestor")
    return self._number_of_generations(descendant, ancestor, 0)

object_properties(self, imported=False)

Returns an generator over all object properties.

If imported is true, will imported object properties are also returned.

Source code in ontopy/ontology.py
def object_properties(self, imported=False):
    """Returns an generator over all object properties.

    If `imported` is true, will imported object properties are also
    returned.
    """
    if imported:
        return self.world.object_properties()
    return super().object_properties()

remove_label_annotation(self, iri)

Removes label annotation used by get_by_label().

May be provided either as an IRI or as its owlready2 representation.

Source code in ontopy/ontology.py
def remove_label_annotation(self, iri):
    """Removes label annotation used by get_by_label().

    May be provided either as an IRI or as its owlready2 representation.
    """
    label_annotation = iri if hasattr(iri, "storid") else self.world[iri]
    if not label_annotation:
        raise ValueError(f"IRI not in ontology: {iri}")
    self._label_annotations.remove(label_annotation)

rename_entities(self, annotations=('prefLabel', 'label', 'altLabel'))

Set name of all entities to the first non-empty annotation in annotations.

Warning, this method changes all IRIs in the ontology. However, it may be useful to make the ontology more readable and to work with it together with a triple store.

Source code in ontopy/ontology.py
def rename_entities(
    self,
    annotations=("prefLabel", "label", "altLabel"),
):
    """Set `name` of all entities to the first non-empty annotation in
    `annotations`.

    Warning, this method changes all IRIs in the ontology.  However,
    it may be useful to make the ontology more readable and to work
    with it together with a triple store.
    """
    for entity in self.get_entities():
        for annotation in annotations:
            if hasattr(entity, annotation):
                name = getattr(entity, annotation).first()
                if name:
                    entity.name = name
                    break

save(self, filename=None, format=None, dir='.', mkdir=False, overwrite=False, recursive=False, squash=False, write_catalog_file=False, append_catalog=False, catalog_file='catalog-v001.xml')

Writes the ontology to file.

Parameters

None | str | Path

Name of file to write to. If None, it defaults to the name of the ontology with format as file extension.

str

Output format. The default is to infer it from filename.

str | Path

If filename is a relative path, it is a relative path to dir.

bool

Whether to create output directory if it does not exists.

bool

If true and filename exists, remove the existing file before saving. The default is to append to an existing ontology.

bool

Whether to save imported ontologies recursively. This is commonly combined with filename=None, dir and mkdir.

bool

If true, rdflib will be used to save the current ontology together with all its sub-ontologies into filename. It make no sense to combine this with recursive.

bool

Whether to also write a catalog file to disk.

bool

Whether to append to an existing catalog file.

str | Path

Name of catalog file. If not an absolute path, it is prepended to dir.

Source code in ontopy/ontology.py
def save(
    self,
    filename=None,
    format=None,
    dir=".",
    mkdir=False,
    overwrite=False,
    recursive=False,
    squash=False,
    write_catalog_file=False,
    append_catalog=False,
    catalog_file="catalog-v001.xml",
):
    """Writes the ontology to file.

    Parameters
    ----------
    filename: None | str | Path
        Name of file to write to.  If None, it defaults to the name
        of the ontology with `format` as file extension.
    format: str
        Output format. The default is to infer it from `filename`.
    dir: str | Path
        If `filename` is a relative path, it is a relative path to `dir`.
    mkdir: bool
        Whether to create output directory if it does not exists.
    owerwrite: bool
        If true and `filename` exists, remove the existing file before
        saving.  The default is to append to an existing ontology.
    recursive: bool
        Whether to save imported ontologies recursively.  This is
        commonly combined with `filename=None`, `dir` and `mkdir`.
    squash: bool
        If true, rdflib will be used to save the current ontology
        together with all its sub-ontologies into `filename`.
        It make no sense to combine this with `recursive`.
    write_catalog_file: bool
        Whether to also write a catalog file to disk.
    append_catalog: bool
        Whether to append to an existing catalog file.
    catalog_file: str | Path
        Name of catalog file.  If not an absolute path, it is prepended
        to `dir`.
    """
    # pylint: disable=redefined-builtin,too-many-arguments
    # pylint: disable=too-many-statements,too-many-branches
    # pylint: disable=too-many-locals,arguments-renamed
    if not _validate_installed_version(
        package="rdflib", min_version="6.0.0"
    ) and format == FMAP.get("ttl", ""):
        from rdflib import (  # pylint: disable=import-outside-toplevel
            __version__ as __rdflib_version__,
        )

        warnings.warn(
            IncompatibleVersion(
                "To correctly convert to Turtle format, rdflib must be "
                "version 6.0.0 or greater, however, the detected rdflib "
                "version used by your Python interpreter is "
                f"{__rdflib_version__!r}. For more information see the "
                "'Known issues' section of the README."
            )
        )

    revmap = {value: key for key, value in FMAP.items()}

    if filename is None:
        if format:
            fmt = revmap.get(format, format)
            filename = f"{self.name}.{fmt}"
        else:
            raise TypeError("`filename` and `format` cannot both be None.")
    filename = os.path.join(dir, filename)
    dir = Path(filename).resolve().parent

    if mkdir:
        outdir = Path(filename).parent.resolve()
        if not outdir.exists():
            outdir.mkdir(parents=True)

    if not format:
        format = guess_format(filename, fmap=FMAP)
    fmt = revmap.get(format, format)

    if overwrite and filename and os.path.exists(filename):
        os.remove(filename)

    EMMO = rdflib.Namespace(  # pylint:disable=invalid-name
        "http://emmo.info/emmo#"
    )

    if recursive:
        if squash:
            raise ValueError(
                "`recursive` and `squash` should not both be true"
            )
        base = self.base_iri.rstrip("#/")
        for onto in self.imported_ontologies:
            obase = onto.base_iri.rstrip("#/")
            newdir = Path(dir) / os.path.relpath(obase, base)
            onto.save(
                filename=None,
                format=format,
                dir=newdir.resolve(),
                mkdir=mkdir,
                overwrite=overwrite,
                recursive=recursive,
                squash=squash,
                write_catalog_file=write_catalog_file,
                append_catalog=append_catalog,
                catalog_file=catalog_file,
            )

    if squash:
        from rdflib import (  # pylint:disable=import-outside-toplevel
            URIRef,
            RDF,
            OWL,
        )

        graph = self.world.as_rdflib_graph()
        graph.namespace_manager.bind("emmo", EMMO)

        # Remove anonymous namespace and imports
        graph.remove((URIRef("http://anonymous"), RDF.type, OWL.Ontology))
        imports = list(graph.triples((None, OWL.imports, None)))
        for triple in imports:
            graph.remove(triple)

        graph.serialize(destination=filename, format=format)
    elif format in OWLREADY2_FORMATS:
        super().save(file=filename, format=fmt)
    else:
        # The try-finally clause is needed for cleanup and because
        # we have to provide delete=False to NamedTemporaryFile
        # since Windows does not allow to reopen an already open
        # file.
        try:
            with tempfile.NamedTemporaryFile(
                suffix=".owl", delete=False
            ) as handle:
                tmpfile = handle.name
            super().save(tmpfile, format="rdfxml")
            graph = rdflib.Graph()
            graph.parse(tmpfile, format="xml")
            graph.serialize(destination=filename, format=format)
        finally:
            os.remove(tmpfile)

    if write_catalog_file:
        mappings = {}
        base = self.base_iri.rstrip("#/")

        def append(onto):
            obase = onto.base_iri.rstrip("#/")
            newdir = Path(dir) / os.path.relpath(obase, base)
            newpath = newdir.resolve() / f"{onto.name}.{fmt}"
            relpath = os.path.relpath(newpath, dir)
            mappings[onto.get_version(as_iri=True)] = str(relpath)
            for imported in onto.imported_ontologies:
                append(imported)

        if recursive:
            append(self)
        write_catalog(
            mappings,
            output=catalog_file,
            directory=dir,
            append=append_catalog,
        )

set_common_prefix(self, iri_base='http://emmo.info/emmo', prefix='emmo')

Set a common prefix for all imported ontologies with the same first part of the base_iri.

Parameters:

Name Type Description Default
iri_base str

The start of the base_iri to look for. Defaults to the emmo base_iri http://emmo.info/emmo

'http://emmo.info/emmo'
prefix str

the desired prefix. Defaults to emmo.

'emmo'
Source code in ontopy/ontology.py
def set_common_prefix(
    self,
    iri_base: str = "http://emmo.info/emmo",
    prefix: str = "emmo",
) -> None:
    """Set a common prefix for all imported ontologies
    with the same first part of the base_iri.

    Args:
        iri_base: The start of the base_iri to look for. Defaults to
            the emmo base_iri http://emmo.info/emmo
        prefix: the desired prefix. Defaults to emmo.
    """
    if self.base_iri.startswith(iri_base):
        self.prefix = prefix
    for onto in self.imported_ontologies:
        onto.set_common_prefix(iri_base=iri_base, prefix=prefix)

set_version(self, version=None, version_iri=None)

Assign version to ontology by asigning owl:versionIRI.

If version but not version_iri is provided, the version IRI will be the combination of base_iri and version.

Source code in ontopy/ontology.py
def set_version(self, version=None, version_iri=None):
    """Assign version to ontology by asigning owl:versionIRI.

    If `version` but not `version_iri` is provided, the version
    IRI will be the combination of `base_iri` and `version`.
    """
    _version_iri = "http://www.w3.org/2002/07/owl#versionIRI"
    version_iri_storid = self.world._abbreviate(_version_iri)
    if self._has_obj_triple_spo(  # pylint: disable=unexpected-keyword-arg
        # For some reason _has_obj_triples_spo exists in both
        # owlready2.namespace.Namespace (with arguments subject/predicate)
        # and in owlready2.triplelite._GraphManager (with arguments s/p)
        # owlready2.Ontology inherits from Namespace directly
        # and pylint checks that.
        # It actually accesses the one in triplelite.
        # subject=self.storid, predicate=version_iri_storid
        s=self.storid,
        p=version_iri_storid,
    ):
        self._del_obj_triple_spo(s=self.storid, p=version_iri_storid)

    if not version_iri:
        if not version:
            raise TypeError(
                "Either `version` or `version_iri` must be provided"
            )
        head, tail = self.base_iri.rstrip("#/").rsplit("/", 1)
        version_iri = "/".join([head, version, tail])

    self._add_obj_triple_spo(
        s=self.storid,
        p=self.world._abbreviate(_version_iri),
        o=self.world._abbreviate(version_iri),
    )

sync_attributes(self, name_policy=None, name_prefix='', class_docstring='comment', sync_imported=False)

This method is intended to be called after you have added new classes (typically via Python) to make sure that attributes like label and comments are defined.

If a class, object property, data property or annotation property in the current ontology has no label, the name of the corresponding Python class will be assigned as label.

If a class, object property, data property or annotation property has no comment, it will be assigned the docstring of the corresponding Python class.

name_policy specify wether and how the names in the ontology should be updated. Valid values are: None not changed "uuid" name_prefix followed by a global unique id (UUID). If the name is already valid accoridng to this standard it will not be regenerated. "sequential" name_prefix followed a sequantial number. EMMO conventions imply name_policy=='uuid'.

If sync_imported is true, all imported ontologies are also updated.

The class_docstring argument specifies the annotation that class docstrings are mapped to. Defaults to "comment".

Source code in ontopy/ontology.py
def sync_attributes(  # pylint: disable=too-many-branches
    self,
    name_policy=None,
    name_prefix="",
    class_docstring="comment",
    sync_imported=False,
):
    """This method is intended to be called after you have added new
    classes (typically via Python) to make sure that attributes like
    `label` and `comments` are defined.

    If a class, object property, data property or annotation
    property in the current ontology has no label, the name of
    the corresponding Python class will be assigned as label.

    If a class, object property, data property or annotation
    property has no comment, it will be assigned the docstring of
    the corresponding Python class.

    `name_policy` specify wether and how the names in the ontology
    should be updated.  Valid values are:
      None          not changed
      "uuid"        `name_prefix` followed by a global unique id (UUID).
                    If the name is already valid accoridng to this standard
                    it will not be regenerated.
      "sequential"  `name_prefix` followed a sequantial number.
    EMMO conventions imply ``name_policy=='uuid'``.

    If `sync_imported` is true, all imported ontologies are also
    updated.

    The `class_docstring` argument specifies the annotation that
    class docstrings are mapped to.  Defaults to "comment".
    """
    for cls in itertools.chain(
        self.classes(),
        self.object_properties(),
        self.data_properties(),
        self.annotation_properties(),
    ):
        if not hasattr(cls, "prefLabel"):
            # no prefLabel - create new annotation property..
            with self:
                # pylint: disable=invalid-name,missing-class-docstring
                # pylint: disable=unused-variable
                class prefLabel(owlready2.label):
                    pass

            cls.prefLabel = [locstr(cls.__name__, lang="en")]
        elif not cls.prefLabel:
            cls.prefLabel.append(locstr(cls.__name__, lang="en"))
        if class_docstring and hasattr(cls, "__doc__") and cls.__doc__:
            getattr(cls, class_docstring).append(
                locstr(inspect.cleandoc(cls.__doc__), lang="en")
            )

    for ind in self.individuals():
        if not hasattr(ind, "prefLabel"):
            # no prefLabel - create new annotation property..
            with self:
                # pylint: disable=invalid-name,missing-class-docstring
                # pylint: disable=function-redefined
                class prefLabel(owlready2.label):
                    iri = "http://www.w3.org/2004/02/skos/core#prefLabel"

            ind.prefLabel = [locstr(ind.name, lang="en")]
        elif not ind.prefLabel:
            ind.prefLabel.append(locstr(ind.name, lang="en"))

    chain = itertools.chain(
        self.classes(),
        self.individuals(),
        self.object_properties(),
        self.data_properties(),
        self.annotation_properties(),
    )
    if name_policy == "uuid":
        for obj in chain:
            try:
                # Passing the following means that the name is valid
                # and need not be regenerated.
                if not obj.name.startswith(name_prefix):
                    raise ValueError
                uuid.UUID(obj.name.lstrip(name_prefix), version=5)
            except ValueError:
                obj.name = name_prefix + str(
                    uuid.uuid5(uuid.NAMESPACE_DNS, obj.name)
                )
    elif name_policy == "sequential":
        for obj in chain:
            counter = 0
            while f"{self.base_iri}{name_prefix}{counter}" in self:
                counter += 1
            obj.name = f"{name_prefix}{counter}"
    elif name_policy is not None:
        raise TypeError(f"invalid name_policy: {name_policy!r}")

    if sync_imported:
        for onto in self.imported_ontologies:
            onto.sync_attributes()

sync_python_names(self, annotations=('prefLabel', 'label', 'altLabel'))

Update the python_name attribute of all properties.

The python_name attribute will be set to the first non-empty annotation in the sequence of annotations in annotations for the property.

Source code in ontopy/ontology.py
def sync_python_names(self, annotations=("prefLabel", "label", "altLabel")):
    """Update the `python_name` attribute of all properties.

    The python_name attribute will be set to the first non-empty
    annotation in the sequence of annotations in `annotations` for
    the property.
    """

    def update(gen):
        for prop in gen:
            for annotation in annotations:
                if hasattr(prop, annotation) and getattr(prop, annotation):
                    prop.python_name = getattr(prop, annotation).first()
                    break

    update(
        self.get_entities(
            classes=False,
            individuals=False,
            object_properties=False,
            data_properties=False,
        )
    )
    update(
        self.get_entities(
            classes=False, individuals=False, annotation_properties=False
        )
    )

sync_reasoner(self, reasoner='FaCT++', include_imported=False, **kwargs)

Update current ontology by running the given reasoner.

Supported values for reasoner are 'Pellet', 'HermiT' and 'FaCT++'.

If include_imported is true, the reasoner will also reason over imported ontologies. Note that this may be very slow with Pellet and HermiT.

Keyword arguments are passed to the underlying owlready2 function.

Source code in ontopy/ontology.py
def sync_reasoner(
    self, reasoner="FaCT++", include_imported=False, **kwargs
):
    """Update current ontology by running the given reasoner.

    Supported values for `reasoner` are 'Pellet', 'HermiT' and 'FaCT++'.

    If `include_imported` is true, the reasoner will also reason
    over imported ontologies.  Note that this may be **very** slow
    with Pellet and HermiT.

    Keyword arguments are passed to the underlying owlready2 function.
    """
    if reasoner == "FaCT++":
        sync = sync_reasoner_factpp
    elif reasoner == "Pellet":
        sync = owlready2.sync_reasoner_pellet
    elif reasoner == "HermiT":
        sync = owlready2.sync_reasoner_hermit
    else:
        raise ValueError(
            f"unknown reasoner {reasoner!r}. Supported reasoners "
            'are "Pellet", "HermiT" and "FaCT++".'
        )

    # For some reason we must visit all entities once before running
    # the reasoner...
    list(self.get_entities())

    with self:
        if include_imported:
            sync(self.world, **kwargs)
        else:
            sync(self, **kwargs)

World (World)

A subclass of owlready2.World.

Source code in ontopy/ontology.py
class World(owlready2.World):
    """A subclass of owlready2.World."""

    def __init__(self, *args, **kwargs):
        # Caches stored in the world
        self._cached_catalogs = {}  # maps url to (mtime, iris, dirs)
        self._iri_mappings = {}  # all iri mappings loaded so far
        super().__init__(*args, **kwargs)

    def get_ontology(self, base_iri="emmo-inferred"):
        """Returns a new Ontology from `base_iri`.

        The `base_iri` argument may be one of:
          - valid URL (possible excluding final .owl or .ttl)
          - file name (possible excluding final .owl or .ttl)
          - "emmo": load latest version of asserted EMMO
          - "emmo-inferred": load latest version of inferred EMMO
            (default)
          - "emmo-development": load latest inferred development version
            of EMMO. Until first stable release emmo-inferred and
            emmo-development will be the same.
        """
        base_iri = base_iri.as_uri() if isinstance(base_iri, Path) else base_iri

        if base_iri == "emmo":
            base_iri = (
                "http://emmo-repo.github.io/versions/1.0.0-beta4/emmo.ttl"
            )
        elif base_iri == "emmo-inferred":
            base_iri = (
                "https://emmo-repo.github.io/versions/1.0.0-beta4/"
                "emmo-inferred.ttl"
            )
        elif base_iri == "emmo-development":
            base_iri = (
                "https://emmo-repo.github.io/versions/1.0.0-beta4/"
                "emmo-inferred.ttl"
            )

        if base_iri in self.ontologies:
            onto = self.ontologies[base_iri]
        elif base_iri + "#" in self.ontologies:
            onto = self.ontologies[base_iri + "#"]
        elif base_iri + "/" in self.ontologies:
            onto = self.ontologies[base_iri + "/"]
        else:
            if os.path.exists(base_iri):
                iri = os.path.abspath(base_iri)
            elif os.path.exists(base_iri + ".ttl"):
                iri = os.path.abspath(base_iri + ".ttl")
            elif os.path.exists(base_iri + ".owl"):
                iri = os.path.abspath(base_iri + ".owl")
            else:
                iri = base_iri

            if iri[-1] not in "/#":
                iri += "#"
            onto = Ontology(self, iri)

        return onto

    def get_unabbreviated_triples(
        self, subject=None, predicate=None, obj=None, blank=None
    ):
        # pylint: disable=invalid-name
        """Returns all triples unabbreviated.

        If any of the `subject`, `predicate` or `obj` arguments are given,
        only matching triples will be returned.

        If `blank` is given, it will be used to represent blank nodes.
        """
        return _get_unabbreviated_triples(
            self, subject=subject, predicate=predicate, obj=obj, blank=blank
        )

get_ontology(self, base_iri='emmo-inferred')

Returns a new Ontology from base_iri.

The base_iri argument may be one of: - valid URL (possible excluding final .owl or .ttl) - file name (possible excluding final .owl or .ttl) - "emmo": load latest version of asserted EMMO - "emmo-inferred": load latest version of inferred EMMO (default) - "emmo-development": load latest inferred development version of EMMO. Until first stable release emmo-inferred and emmo-development will be the same.

Source code in ontopy/ontology.py
def get_ontology(self, base_iri="emmo-inferred"):
    """Returns a new Ontology from `base_iri`.

    The `base_iri` argument may be one of:
      - valid URL (possible excluding final .owl or .ttl)
      - file name (possible excluding final .owl or .ttl)
      - "emmo": load latest version of asserted EMMO
      - "emmo-inferred": load latest version of inferred EMMO
        (default)
      - "emmo-development": load latest inferred development version
        of EMMO. Until first stable release emmo-inferred and
        emmo-development will be the same.
    """
    base_iri = base_iri.as_uri() if isinstance(base_iri, Path) else base_iri

    if base_iri == "emmo":
        base_iri = (
            "http://emmo-repo.github.io/versions/1.0.0-beta4/emmo.ttl"
        )
    elif base_iri == "emmo-inferred":
        base_iri = (
            "https://emmo-repo.github.io/versions/1.0.0-beta4/"
            "emmo-inferred.ttl"
        )
    elif base_iri == "emmo-development":
        base_iri = (
            "https://emmo-repo.github.io/versions/1.0.0-beta4/"
            "emmo-inferred.ttl"
        )

    if base_iri in self.ontologies:
        onto = self.ontologies[base_iri]
    elif base_iri + "#" in self.ontologies:
        onto = self.ontologies[base_iri + "#"]
    elif base_iri + "/" in self.ontologies:
        onto = self.ontologies[base_iri + "/"]
    else:
        if os.path.exists(base_iri):
            iri = os.path.abspath(base_iri)
        elif os.path.exists(base_iri + ".ttl"):
            iri = os.path.abspath(base_iri + ".ttl")
        elif os.path.exists(base_iri + ".owl"):
            iri = os.path.abspath(base_iri + ".owl")
        else:
            iri = base_iri

        if iri[-1] not in "/#":
            iri += "#"
        onto = Ontology(self, iri)

    return onto

get_unabbreviated_triples(self, subject=None, predicate=None, obj=None, blank=None)

Returns all triples unabbreviated.

If any of the subject, predicate or obj arguments are given, only matching triples will be returned.

If blank is given, it will be used to represent blank nodes.

Source code in ontopy/ontology.py
def get_unabbreviated_triples(
    self, subject=None, predicate=None, obj=None, blank=None
):
    # pylint: disable=invalid-name
    """Returns all triples unabbreviated.

    If any of the `subject`, `predicate` or `obj` arguments are given,
    only matching triples will be returned.

    If `blank` is given, it will be used to represent blank nodes.
    """
    return _get_unabbreviated_triples(
        self, subject=subject, predicate=predicate, obj=obj, blank=blank
    )

flatten(items)

Yield items from any nested iterable.

Source code in ontopy/ontology.py
def flatten(items):
    """Yield items from any nested iterable."""
    for item in items:
        if isinstance(item, Iterable) and not isinstance(item, (str, bytes)):
            for sub_item in flatten(item):
                yield sub_item
        else:
            yield item

get_ontology(*args, **kwargs)

Returns a new Ontology from base_iri.

This is a convenient function for calling World.get_ontology().

Source code in ontopy/ontology.py
def get_ontology(*args, **kwargs):
    """Returns a new Ontology from `base_iri`.

    This is a convenient function for calling World.get_ontology()."""
    return World().get_ontology(*args, **kwargs)