Neil Ireson, Fabio Ciravegna, Marie Elaine Califf, Dayne Freitag, Nicholas Kushmerick, Alberto Lavelli: Evaluating Machine Learning for Information Extraction, 22nd International Conference on Machine Learning (ICML 2005), Bonn, Germany, 7-11 August, 2005
38. H-M. Haav, An Application of Inductive Concept Analysis to Construction of Domain-specific Ontologies, In: B. Thalheim, Gunar Fiedler (Eds), Emerging Database Research in East Europe, Proceedings of the Pre-conference Workshop of VLDB 2003, Computer Science Reports, Brandenburg University of Technology at Cottbus, 2003, 14/3, pp 63-67
Lush is an object-oriented programming language designed for researchers, experimenters, and engineers interested in large-scale numerical and graphic applications.
Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed. - GitHub - divamgupta/diffusionbee-stable-diffusion-ui: Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.
The aim of MLpedia is to provide comprehensive information on a range of machine learning methods and applications, written and maintained by researchers and practitioners. Find out how to take part.
Deep Learning has revolutionised Pattern Recognition and Machine Learning. It is about credit assignment in adaptive systems with long chains of potentially causal links between actions and consequences.
The mission of the Journal of Machine Learning Gossip (JMLG) is to provide an archival source of important information that is often discussed informally at conferences but is rarely, if ever, written down.
Engineer friends often ask me: Graph Deep Learning sounds great, but are there any big commercial success stories? Is it being deployed in practical applications? Besides the obvious ones–recommendation systems at Pinterest, Alibaba and Twitter–a slightly nuanced success story is the Transformer architecture, which has taken the NLP industry by storm. Through this post, I want to establish links between Graph Neural Networks (GNNs) and Transformers. I’ll talk about the intuitions behind model architectures in the NLP and GNN communities, make connections using equations and figures, and discuss how we could work together to drive progress.
The Knowledge Discovery Machine Learning (KDML) group focuses on the neighboring subfields of computer science known as knowledge discovery in databases (KDD, sometimes referred to simply as data mining) and machine learning (ML). For us, these fields include on the one hand the automated analysis of large data sets using intelligent algorithms that are capable of extracting from the collected data hidden knowledge in order to produce models that can be used for prediction and decision making. On the other hand, they also include algorithms and systems that are capable of learning from experience and adapting to their environment or their users.
Mahout currently has
Collaborative Filtering
User and Item based recommenders
K-Means, Fuzzy K-Means clustering
Mean Shift clustering
Dirichlet process clustering
Latent Dirichlet Allocation
Singular value decomposition
Parallel Frequent Pattern mining
Complementary Naive Bayes classifier
Random forest decision tree based classifier
High performance java collections (previously colt collections)
A vibrant community
and many more cool stuff to come by this summer thanks to Google summer of code
I'm interested in machine learning techniques (graphical models, kernel methods) applied to text understanding (entity and relation extraction, coreference resolution, document classification and clustering, confidence prediction, social network analysis, data mining).
Talk with Yves Raimond at the GPU Tech Conference on Marth 28, 2018 in San Jose, CA. Abstract: In this talk, we will survey how Deep Learning methods can be ap…