The entropic vector or entropic function is a concept arising in information theory. It represents the possible values of Shannon's information entropy that subsets of one set of random variables may take. Understanding which vectors are entropic is a way to represent all possible inequalities between entropies of various subsets. For example, for any two random variables , their joint entropy (the entropy of the random variable representing the pair ) is at most the sum of the entropies of and of :
Attributes | Values |
---|---|
rdfs:label |
|
rdfs:comment |
|
dcterms:subject | |
Wikipage page ID |
|
Wikipage revision ID |
|
Link from a Wikipage to another Wikipage |
|
sameAs | |
dbp:wikiPageUsesTemplate | |
has abstract |
|
gold:hypernym | |
prov:wasDerivedFrom | |
page length (characters) of wiki page |
|
foaf:isPrimaryTopicOf | |
is Link from a Wikipage to another Wikipage of | |
is Wikipage redirect of | |
is foaf:primaryTopic of |