Apache's Hadoop project aims to solve these problems by providing a framework for running large data processing applications on clusters of commodity hardware. Combined with Amazon EC2 for running the application, and Amazon S3 for storing the data, we can run large jobs very economically. This paper describes how to use Amazon Web Services and Hadoop to run an ad hoc analysis on a large collection of web access logs that otherwise would have cost a prohibitive amount in either time or money.
explores the challenges of constructing a distributed e-business architecture based on the concept of Request Based Virtual OrganiZation (RBVO) and presents a solution based on ebXML, Open Source e-business component
Apache Mesos abstracts CPU, memory, storage, and other compute resources away from machines (physical or virtual), enabling fault-tolerant and elastic distributed systems to easily be built and run effectively.
for creating Linux and Windows communications applications that contain Dialogic/Intel NetStructure products. Includes media and network classes, multithreaded event handling, state machine design pattern and distributed execution.
Real-time image transmission is crucial to an emerging class of distributed embedded systems operating in open network environments. Examples include avionics mission re-planning over Link-16, security systems based on wireless camera networks, a
The starting point is a paper on "Non-blocking atomic commitment" by Babaoglu and Toueg, from Mulender's book on Distributed Systems. They present an algorithm for distributed atomic commitment that enjoys different sets of properties depending on what type of broadcast is used. They give an informal correctness proof which they claim is compositional, although I'd rather call it incremental (when the algorithm is modified, the proof is modified accordingly, some parts being reused).
Building a centralized database to process billions of open-ended queries per day is a mammoth undertaking. It appears that Google, who perhaps is the only company on the planet with enough imagination, incentive, and expertise to effectively build such a
DDM algorithms for mining large distributed data sources,DDM algorithms for monitoring data streams, Privacy preserving distributed multi-party data mining;Sensor networks, Grid mining, mobile/wireless;applications, privacy preserving security-related
A paper presented at the Second Workshop on Understanding Work and Designing Artefacts: Design for Collaboration. Communities Constructing Technology at King's Manor, University of York (March 1999).
Our current research interests are web page cacheability, distributed proxy servers, cluster-based web servers, internet QoS, router and web switch design, Layer 5 Switch, and web server benchmark designs using trace and execution simulation of RSIM and Simplescalar.
The goal of the Condor® Project is to develop, implement, deploy, and evaluate mechanisms and policies that support High Throughput Computing (HTC) on large collections of distributively owned computing resources. Guided by both the technological and sociological challenges of such a computing environment, the Condor Team has been building software tools that enable scientists and engineers to increase their computing throughput
YaCy ist eine Suchmaschinensoftware die sich jeder installieren kann um damit ein Suchportal zu errichten, das Intranet zu indexieren oder andere Daten mit einer Suchfunktion zu erweitern. YaCy beinhaltet einen Webcrawler, so dass es einfach ist, eine große Anzahl von Dokumenten aus dem Web sehr schnell zu erfassen. Dabei ist YaCy sehr hoch skalierbar, eine Installation kann mehr als 20 Millionen Webseiten erfassen und speichern, aber im Peer-to-Peer Verbund mit weiteren YaCy Installationen hat YaCy eine unbegrenzte Kapazität
Distributed Sage is a framework that allows one to do distributed computing from within Sage. It includes a server, client and workers as well as a set of classes that one can subclass from to write distributed computation jobs. It is designed to be used mainly for ‘coarsely’ distributed computations, i.e., computations where jobs do not have to communicate much with each other. This is also sometimes referred to as ‘grid’ computing.
science which solves a large problem by giving small parts of the problem to many computers to solve and then combining the solutions for the parts into a solution for the problem.
mainly marketing articlle by cofiiunder What if you didn't have to do any of this funny business to get scalability and reliability? What if the JVM had access to a service that you could plug into to make its heap durable, arbitrarily large, and shared with every other JVM in your application tier? Enter Terracotta, network-attached, durable virtual heap for the JVM. In the spirit of full-disclosure, I'm a co-founder of Terracotta and work there as a software developer. Terracotta is an infrastructure service that is deployed as a stand-alone server plus a library that plugs into your existing JVMs and transparently clusters your JVM's heap. Terracotta makes some of your JVM heap shared via a network connection to the Terracotta server so that a bunch of JVMs can all access the shared heap as if it were local heap. You can think of it like a network-attached filesystem, but for your object data; see Figure 1.
C. Ghidini, and L. Serafini. Modelling and Using Context -- Proceedings of the 2nd International and Interdisciplinary Conference, Context'99, volume 1688 of Lecture Notes in Artificial Intelligence, page 159--172. Springer Verlag - Heidelberg, (1999)
G. Pirro, C. Mastroianni, and D. Talia. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF GRID COMPUTING-THEORY METHODS AND APPLICATIONS, 26 (1):
38-49(January 2010)
L. Kagal, S. Cost, T. Finin, and Y. Peng. Proceedings of IJCAI-01 Workshop on Autonomy, Delegation and Control, (2001)http://citeseer.nj.nec.com/kagal01framework.html.