Join our meetup, learn, connect, share, and get to know your Toronto AI community.
Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.
Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.
In the recent update of GraphVite, we release a new large-scale knowledge graph dataset, along with new benchmarks of knowledge graph embedding methods. The dataset, Wikidata5m, contains 5 million entities and 21 million facts constructed from Wikidata and Wikipedia. Most of the entities come from the general domain or the scientific domain, such as celebrities, events, concepts and things.
To facilitate the usage of knowledge graph representations in semantic tasks, we provide a bunch of pre-trained embeddings from popular models, including TransE, DistMult, ComplEx, SimplE and RotatE. You can directly access these embeddings by natural language index, such as “machine learning”, “united states” or even abbreviations like “m.i.t.”. Check out these models here.
Here are the benchmarks of these models on Wikidata5m.
MR | MRR | HITS@1 | HITS@3 | HITS@10 | |
---|---|---|---|---|---|
TransE | 109370 | 0.253 | 0.170 | 0.311 | 0.392 |
DistMult | 211030 | 0.253 | 0.209 | 0.278 | 0.334 |
ComplEx | 244540 | 0.281 | 0.228 | 0.310 | 0.373 |
SimplE | 112754 | 0.296 | 0.252 | 0.317 | 0.377 |
RotatE | 89459 | 0.290 | 0.234 | 0.322 | 0.390 |
submitted by /u/kiddozhu
[link] [comments]