Focused crawls are collections of frequently-updated webcrawl data from narrow (as opposed to broad or wide) web crawls, often focused on a single domain or subdomain.
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
The official repository for our programming kitchen which consists of 50+ delicious programming recipes having all the interesting ingredients ranging from dynamic programming, graph theory, linked lists and much more. All the articles contain beautiful images and some gif/video at times to help clear important concepts.
I noticed that I forgot to update the API docs for a refactor that moved mathy.mathy to mathy.api.
Currently, you have to run:
cd libraries/website
sh tools/docs.sh
And it generates all the API doc files. This means it's too slow to run all the time. It should be updated to run as part of lint-staged and support passing individual files on the command-line, similar to th
I noticed that I forgot to update the API docs for a refactor that moved
mathy.mathytomathy.api.Currently, you have to run:
cd libraries/website sh tools/docs.shAnd it generates all the API doc files. This means it's too slow to run all the time. It should be updated to run as part of lint-staged and support passing individual files on the command-line, similar to th