<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="de">
	<id>https://iccl.inf.tu-dresden.de/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Long+Cheng</id>
	<title>International Center for Computational Logic - Benutzerbeiträge [de]</title>
	<link rel="self" type="application/atom+xml" href="https://iccl.inf.tu-dresden.de/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Long+Cheng"/>
	<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/web/Spezial:Beitr%C3%A4ge/Long_Cheng"/>
	<updated>2026-04-17T14:40:11Z</updated>
	<subtitle>Benutzerbeiträge</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3035&amp;diff=22889</id>
		<title>Article3035</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3035&amp;diff=22889"/>
		<updated>2017-03-07T21:44:12Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Ilias Tachmazidis; Spyros Kotoulas; Grigoris Antoniou;  }} {{Art…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Ilias Tachmazidis; Spyros Kotoulas; Grigoris Antoniou; &lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Design and Evaluation of Small-Large Outer Joins in Cloud Computing Environments&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2017&lt;br /&gt;
|Journal=Journal of Parallel and Distributed Computing&lt;br /&gt;
|Publisher=Elsevier&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Large-scale analytics is a key application area for data processing and parallel computing research. One of the most common (and challenging) operations in this domain is the join. Though inner join approaches have been extensively evaluated in parallel and distributed systems, there is little published work providing analysis of outer joins, especially in the extremely popular cloud computing environments. A common type of outer join is the small-large outer join, where one relation is relatively small and the other is large. Conventional implementations on this condition, such as one based on hash redistribution, often incur significant network communication, while the duplication-based approaches are complex and inefficient. In this work, we present a new method called DDR (duplication and direct redistribution), which aims to enable efficient small-large outer joins in cloud computing environments while being easy to implement using existing predicates in data processing frameworks. We present the detailed implementation of our approach and evaluate its performance through extensive experiments over the widely used MapReduce and Spark platforms. We show that the proposed method is scalable and can achieve significant performance improvements over the conventional approaches. Compared to the state-of-art method, the DDR algorithm is shown to be easier to implement and can achieve very similar or better performance under different outer join workloads, and thus, can be considered as a new option for current data analysis applications. Moreover, our detailed experimental results also have provided insights of current small-large outer join implementations, thereby allowing system developers to make a more informed choice for their data analysis applications.&lt;br /&gt;
|DOI Name=10.1016/j.jpdc.2017.02.007&lt;br /&gt;
|Projekt=Cfaed, DIAMOND, HAEC, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3139/en&amp;diff=22678</id>
		<title>Inproceedings3139/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3139/en&amp;diff=22678"/>
		<updated>2017-02-08T21:31:17Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Inproceedings3139&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Inproceedings3139]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3139&amp;diff=22677</id>
		<title>Inproceedings3139</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3139&amp;diff=22677"/>
		<updated>2017-02-08T21:31:16Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Tao Li }} {{Inproceedings |Referiert=1 |Title=Efficient Data Red…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Tao Li&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Efficient Data Redistribution to Speedup Big Data Analytics in Large Systems&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Month=Dezember&lt;br /&gt;
|Booktitle=Proc. 23rd IEEE International Conference on High Performance Computing (HiPC&#039;16)&lt;br /&gt;
|Pages=91-100&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The performance of parallel data analytics systems becomes increasingly important with the rise of Big Data. An essential operation in such environment is parallel join, which always incurs significant cost on network communication. State-of-the-art approaches have achieved performance improvements over conventional implementations through minimizing network traffic or communication time. However, these approaches still face performance issues in the presence of big data and/or large-scale systems, due to their heavy overhead of data redistribution scheduling. In this paper, we propose near-join, a network-aware redistribution approach targeting to efficiently reduce both network traffic and communication time of join executions. Particularly, near-join is lightweight and adaptable to processing large datasets over large systems. We present the details of our algorithm and its implementation. The experiments performed on a cluster of up to 400 nodes and datasets of about 100GB have demonstrated that our scheduling algorithm is much faster than the state-of-the-art methods. Moreover, our join implementation can also achieve speedups over the conventional approaches.&lt;br /&gt;
|ISBN=978-1-5090-5411-4&lt;br /&gt;
|Download=PID4490265.pdf&lt;br /&gt;
|DOI Name=10.1109/HiPC.2016.020&lt;br /&gt;
|Projekt=Cfaed, DIAMOND, HAEC, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:PID4490265.pdf&amp;diff=22676</id>
		<title>Datei:PID4490265.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:PID4490265.pdf&amp;diff=22676"/>
		<updated>2017-02-08T21:29:09Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079&amp;diff=21059</id>
		<title>Inproceedings3079</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079&amp;diff=21059"/>
		<updated>2016-09-01T08:12:09Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas;&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Efficient Large Outer Joins over MapReduce&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Month=August&lt;br /&gt;
|Booktitle=Proc. 22nd International European Conference on Parallel Processing (Euro-Par&#039;16)&lt;br /&gt;
|Pages=334-346&lt;br /&gt;
|Publisher=Springer&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Big Data analytics largely rely on being able to execute large joins efficiently. Though inner join approaches have been extensively evaluated in parallel and distributed systems, there is little published work providing analysis of outer joins, especially on the extremely popular MapReduce platform. In this paper, we studied several current algorithms/techniques used in large outer joins. We find that some of them could meet performance bottlenecks in the presence of data skew, while others could be complex and incur significant coordination overheads when applied to the MapReduce framework. In this light, we propose a new algorithm, called POPI (Partial Outer join &amp;amp; Partial Inner join), which targets for efficient processing large outer joins, and most important, is lightweight and adapted to the processing model of MapReduce. We implement our method in Pig and evaluate its performance on a Hadoop cluster of up to 256 cores and datasets of 1 billion tuples. Experimental results show that our method is scalable, robust and outperforms current implementations, at least in the case of high skew.&lt;br /&gt;
|ISBN=978-3-319-43658-6&lt;br /&gt;
|ISSN=0302-9743&lt;br /&gt;
|Download=2016-europar.pdf&lt;br /&gt;
|DOI Name=10.1007/978-3-319-43659-3_25&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2016-europar.pdf&amp;diff=21058</id>
		<title>Datei:2016-europar.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2016-europar.pdf&amp;diff=21058"/>
		<updated>2016-09-01T08:09:54Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=21057</id>
		<title>Article3019</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=21057"/>
		<updated>2016-09-01T08:06:28Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas;&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Scale-Out Processing of Large RDF Datasets&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Month=Dezember&lt;br /&gt;
|Journal=IEEE Transactions on Big Data&lt;br /&gt;
|Volume=1&lt;br /&gt;
|Number=4&lt;br /&gt;
|Pages=138-150&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Distributed RDF data management systems become increasingly important with the growth of the Semantic Web. Regardless, current methods meet performance bottlenecks either on data loading or querying when processing large amounts of data. In this work, we propose efficient methods for processing RDF using dynamic data re-partitioning to enable rapid analysis of large datasets. Our approach adopts a two-tier index architecture on each computation node: (1) a lightweight primary index, to keep loading times low, and (2) a series of dynamic, multi-level secondary indexes, calculated as a by-product of query execution, to decrease or remove inter-machine data movement for subsequent queries that contain the same graph patterns. In addition, we propose methods to replace some secondary indexes with distributed filters, so as to decrease memory consumption. Experimental results on a commodity cluster with 16 nodes show that the method presents good scale-out characteristics and can indeed vastly improve loading speeds while remaining competitive in terms of performance. Specifically, our approach can load a dataset of 1.1 billion triples at a rate of 2.48 million triples per second and provide competitive performance to RDF-3X and 4store for expensive queries.&lt;br /&gt;
|Download=2015-tbd.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Semantische Technologien&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=21056</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=21056"/>
		<updated>2016-09-01T08:05:57Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web Data using X10&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Month=September&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Volume=27&lt;br /&gt;
|Number=9&lt;br /&gt;
|Pages=2603-2617&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].  The source code in X10 is available at: https://github.com/longcheng11/rdf_encoding&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Encodings.pdf&lt;br /&gt;
|DOI Name=10.1109/TPDS.2015.2496579&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Semantische Technologien&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=21055</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=21055"/>
		<updated>2016-09-01T08:05:30Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web Data using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Month=September&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Volume=27&lt;br /&gt;
|Number=9&lt;br /&gt;
|Pages=2603-2617&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].  The source code in X10 is available at: https://github.com/longcheng11/rdf_encoding&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Encodings.pdf&lt;br /&gt;
|DOI Name=10.1109/TPDS.2015.2496579&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Semantische Technologien&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=21054</id>
		<title>Article3019</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=21054"/>
		<updated>2016-09-01T08:04:16Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas;&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Scale-Out Processing of Large RDF Datasets&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Big Data&lt;br /&gt;
|Volume=1&lt;br /&gt;
|Number=4&lt;br /&gt;
|Pages=138-150&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Distributed RDF data management systems become increasingly important with the growth of the Semantic Web. Regardless, current methods meet performance bottlenecks either on data loading or querying when processing large amounts of data. In this work, we propose efficient methods for processing RDF using dynamic data re-partitioning to enable rapid analysis of large datasets. Our approach adopts a two-tier index architecture on each computation node: (1) a lightweight primary index, to keep loading times low, and (2) a series of dynamic, multi-level secondary indexes, calculated as a by-product of query execution, to decrease or remove inter-machine data movement for subsequent queries that contain the same graph patterns. In addition, we propose methods to replace some secondary indexes with distributed filters, so as to decrease memory consumption. Experimental results on a commodity cluster with 16 nodes show that the method presents good scale-out characteristics and can indeed vastly improve loading speeds while remaining competitive in terms of performance. Specifically, our approach can load a dataset of 1.1 billion triples at a rate of 2.48 million triples per second and provide competitive performance to RDF-3X and 4store for expensive queries.&lt;br /&gt;
|Download=2015-tbd.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Semantische Technologien&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=20741</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=20741"/>
		<updated>2016-07-11T20:05:15Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Wissensbasierte Systeme&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=1&lt;br /&gt;
|Telefon=+49 351 463 43510&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web.&lt;br /&gt;
|Info EN=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://www.win.tue.nl/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079/en&amp;diff=19354</id>
		<title>Inproceedings3079/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079/en&amp;diff=19354"/>
		<updated>2016-04-30T10:52:03Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Inproceedings3079&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Inproceedings3079]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079&amp;diff=19353</id>
		<title>Inproceedings3079</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3079&amp;diff=19353"/>
		<updated>2016-04-30T10:52:01Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Spyros Kotoulas;  }} {{Inproceedings |Referiert=1 |Title=Efficie…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; &lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Efficient Large Outer Joins over MapReduce&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Month=August&lt;br /&gt;
|Booktitle=Proc. 22nd International European Conference on Parallel Processing (Euro-Par&#039;16)&lt;br /&gt;
|Publisher=Springer&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Big Data analytics largely rely on being able to execute large joins efficiently. Though inner join approaches have been extensively evaluated in parallel and distributed systems, there is little published work providing analysis of outer joins, especially on the extremely popular MapReduce platform. In this paper, we studied several current algorithms/techniques used in large outer joins. We find that some of them could meet performance bottlenecks in the presence of data skew, while others could be complex and incur significant coordination overheads when applied to the MapReduce framework. In this light, we propose a new algorithm, called POPI (Partial Outer join &amp;amp; Partial Inner join), which targets for efficient processing large outer joins, and most important, is lightweight and adapted to the processing model of MapReduce. We implement our method in Pig and evaluate its performance on a Hadoop cluster of up to 256 cores and datasets of 1 billion tuples. Experimental results show that our method is scalable, robust and outperforms current implementations, at least in the case of high skew.&lt;br /&gt;
|Projekt=DIAMOND, HAEC B08&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019/en&amp;diff=17659</id>
		<title>Article3019/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019/en&amp;diff=17659"/>
		<updated>2015-12-02T09:31:41Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Article3019&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Article3019]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=17658</id>
		<title>Article3019</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3019&amp;diff=17658"/>
		<updated>2015-12-02T09:31:41Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Spyros Kotoulas;  }} {{Article |Referiert=0 |Title=Scale-Out Pro…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; &lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Scale-Out Processing of Large RDF Datasets&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2016&lt;br /&gt;
|Journal=IEEE Transactions on Big Data&lt;br /&gt;
|Note=In press&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Distributed RDF data management systems become increasingly important with the growth of the Semantic Web. Regardless, current methods meet performance bottlenecks either on data loading or querying when processing large amounts of data. In this work, we propose efficient methods for processing RDF using dynamic data re-partitioning to enable rapid analysis of large datasets. Our approach adopts a two-tier index architecture on each computation node: (1) a lightweight primary index, to keep loading times low, and (2) a series of dynamic, multi-level secondary indexes, calculated as a by-product of query execution, to decrease or remove inter-machine data movement for subsequent queries that contain the same graph patterns. In addition, we propose methods to replace some secondary indexes with distributed filters, so as to decrease memory consumption. Experimental results on a commodity cluster with 16 nodes show that the method presents good scale-out characteristics and can indeed vastly improve loading speeds while remaining competitive in terms of performance. Specifically, our approach can load a dataset of 1.1 billion triples at a rate of 2.48 million triples per second and provide competitive performance to RDF-3X and 4store for expensive queries.&lt;br /&gt;
|Download=2015-tbd.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-tbd.pdf&amp;diff=17657</id>
		<title>Datei:2015-tbd.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-tbd.pdf&amp;diff=17657"/>
		<updated>2015-12-02T09:29:37Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17501</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17501"/>
		<updated>2015-11-10T10:25:24Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].  The source code in X10 is available at: https://github.com/longcheng11/rdf_encoding &lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Encodings.pdf&lt;br /&gt;
|DOI Name=10.1109/TPDS.2015.2496579&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Encodings.pdf&amp;diff=17499</id>
		<title>Datei:Encodings.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Encodings.pdf&amp;diff=17499"/>
		<updated>2015-11-10T10:23:30Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17445</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17445"/>
		<updated>2015-10-29T11:05:17Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Tpds encoding.pdf&lt;br /&gt;
|DOI Name=10.1109/TPDS.2015.2496579&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17412</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17412"/>
		<updated>2015-10-27T22:51:57Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Tpds encoding.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Tpds_encoding.pdf&amp;diff=17411</id>
		<title>Datei:Tpds encoding.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Tpds_encoding.pdf&amp;diff=17411"/>
		<updated>2015-10-27T22:51:41Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17408</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17408"/>
		<updated>2015-10-27T13:08:31Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Download=Tpds encodings.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Tpds_encodings.pdf&amp;diff=17407</id>
		<title>Datei:Tpds encodings.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:Tpds_encodings.pdf&amp;diff=17407"/>
		<updated>2015-10-27T13:08:23Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4049&amp;diff=17406</id>
		<title>Inproceedings4049</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4049&amp;diff=17406"/>
		<updated>2015-10-27T12:57:18Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik; Spyros Kotoulas; Tomas E. Ward; Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Efficient Parallel Dictionary Encoding for RDF Data&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=Juni&lt;br /&gt;
|Booktitle=Proc. 17th International Workshop on the Web and Databases (WebDB&#039;14)&lt;br /&gt;
|Note=An extended version of this work is the journal article [[Article3018/en|Fast Compression of Large Semantic Web using X10]]&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The SemanticWeb comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of SemanticWeb applications that perform computations over large volumes of information. A typical method for alleviating the impact of this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding for this purpose is particularly prevalent in Semantic Web database systems. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we describe a straightforward but very efficient encoding algorithm and evaluate its performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art MapReduce algorithm, we demonstrate a speedup of 2.6 - 7.4x and excellent scalability.&lt;br /&gt;
|Download=2014-Efficient-Parallel.pdf&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17405</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17405"/>
		<updated>2015-10-27T12:53:59Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=This paper is the extended journal version of the article [[Inproceedings4049/en|Efficient Parallel Dictionary Encoding for RDF Data]].&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018/en&amp;diff=17404</id>
		<title>Article3018/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018/en&amp;diff=17404"/>
		<updated>2015-10-27T12:50:00Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Article3018&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Article3018]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17403</id>
		<title>Article3018</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3018&amp;diff=17403"/>
		<updated>2015-10-27T12:49:59Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theod…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik;  Spyros Kotoulas;  Tomas E. Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Fast Compression of Large Semantic Web using X10&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Parallel and Distributed Systems&lt;br /&gt;
|Note=in press&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The Semantic Web comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of applications that perform computations over large volumes of such information. A common approach to alleviate this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding is particularly prevalent in Semantic Web database systems for this purpose. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we propose an efficient algorithm for fast encoding large Semantic Web data. Specially, we present the detailed implementation of our approach based on the state-of-art asynchronous partitioned global address space (APGAS) parallel programming model. We evaluate performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art approach, we demonstrate a speed-up of 2.6 - 7.4x and excellent scalability. In the meantime, these results also illustrate the significant potential of the APGAS model for efficient implementation of dictionary encoding and contributes to the engineering of more efficient, larger scale Semantic Web applications.&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4049&amp;diff=17402</id>
		<title>Inproceedings4049</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4049&amp;diff=17402"/>
		<updated>2015-10-27T12:44:45Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Avinash Malik; Spyros Kotoulas; Tomas E. Ward; Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Efficient Parallel Dictionary Encoding for RDF Data&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=Juni&lt;br /&gt;
|Booktitle=Proc. 17th International Workshop on the Web and Databases (WebDB&#039;14)&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The SemanticWeb comprises enormous volumes of semi-structured data elements. For interoperability, these elements are represented by long strings. Such representations are not efficient for the purposes of SemanticWeb applications that perform computations over large volumes of information. A typical method for alleviating the impact of this problem is through the use of compression methods that produce more compact representations of the data. The use of dictionary encoding for this purpose is particularly prevalent in Semantic Web database systems. However, centralized implementations present performance bottlenecks, giving rise to the need for scalable, efficient distributed encoding schemes. In this paper, we describe a straightforward but very efficient encoding algorithm and evaluate its performance on a cluster of up to 384 cores and datasets of up to 11 billion triples (1.9 TB). Compared to the state-of-art MapReduce algorithm, we demonstrate a speedup of 2.6 - 7.4x and excellent scalability.&lt;br /&gt;
|Download=2014-Efficient-Parallel.pdf&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17401</id>
		<title>Article3016</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17401"/>
		<updated>2015-10-27T12:40:34Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas;&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Efficient Skew Handling for Outer Joins in a Cloud Computing Environment&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Cloud Computing&lt;br /&gt;
|Note=In press.&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Outer joins are ubiquitous in many workloads and Big Data systems. The question of how to best execute outer joins in large parallel systems is particularly challenging, as real world datasets are characterized by data skew leading to performance issues. Although skew handling techniques have been extensively studied for inner joins, there is little published work solving the corresponding problem for parallel outer joins, especially in the extremely popular Cloud computing environment. Conventional approaches to the problem such as ones based on hash redistribution often lead to load balancing problems while duplication-based approaches incur significant overhead in terms of network communication. In this paper, we propose a new approach for efficient skew handling in outer joins over a Cloud computing environment. We present an efficient implementation of our approach over the Spark framework. We evaluate the performance of our approach on a 192-core system with large test datasets in excess of 100GB and with varying skew. Experimental results show that our approach is scalable and, at least of in cases of high skew, significantly faster than the state-of-the-art.&lt;br /&gt;
|Download=2015-tcc-cheng.pdf&lt;br /&gt;
|DOI Name=10.1109/TCC.2015.2487965&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17180</id>
		<title>Article3016</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17180"/>
		<updated>2015-10-05T15:06:36Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas;&lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Efficient Skew Handling for Outer Joins in a Cloud Computing Environment&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Cloud Computing&lt;br /&gt;
|Note=In press.&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Outer joins are ubiquitous in many workloads and Big Data systems. The question of how to best execute outer joins in large parallel systems is particularly challenging, as real world datasets are characterized by data skew leading to performance issues. Although skew handling techniques have been extensively studied for inner joins, there is little published work solving the corresponding problem for parallel outer joins, especially in the extremely popular Cloud computing environment. Conventional approaches to the problem such as ones based on hash redistribution often lead to load balancing problems while duplication-based approaches incur significant overhead in terms of network communication. In this paper, we propose a new approach for efficient skew handling in outer joins over a Cloud computing environment. We present an efficient implementation of our approach over the Spark framework. We evaluate the performance of our approach on a 192-core system with large test datasets in excess of 100GB and with varying skew. Experimental results show that our approach is scalable and, at least of in cases of high skew, significantly faster than the state-of-the-art.&lt;br /&gt;
|Download=2015-tcc-cheng.pdf&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-tcc-cheng.pdf&amp;diff=17178</id>
		<title>Datei:2015-tcc-cheng.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-tcc-cheng.pdf&amp;diff=17178"/>
		<updated>2015-10-05T15:06:18Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016/en&amp;diff=17161</id>
		<title>Article3016/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016/en&amp;diff=17161"/>
		<updated>2015-10-02T11:09:12Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Article3016&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Article3016]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17160</id>
		<title>Article3016</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Article3016&amp;diff=17160"/>
		<updated>2015-10-02T11:09:12Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Spyros Kotoulas;  }} {{Article |Referiert=0 |Title=Efficient Ske…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; &lt;br /&gt;
}}&lt;br /&gt;
{{Article&lt;br /&gt;
|Referiert=0&lt;br /&gt;
|Title=Efficient Skew Handling for Outer Joins in a Cloud Computing Environment&lt;br /&gt;
|To appear=1&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Journal=IEEE Transactions on Cloud Computing&lt;br /&gt;
|Note=In press.&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Outer joins are ubiquitous in many workloads and Big Data systems. The question of how to best execute outer joins in large parallel systems is particularly challenging, as real world datasets are characterized by data skew leading to performance issues. Although skew handling techniques have been extensively studied for inner joins, there is little published work solving the corresponding problem for parallel outer joins, especially in the extremely popular Cloud computing environment. Conventional approaches to the problem such as ones based on hash redistribution often lead to load balancing problems while duplication-based approaches incur significant overhead in terms of network communication. In this paper, we propose a new approach for efficient skew handling in outer joins over a Cloud computing environment. We present an efficient implementation of our approach over the Spark framework. We evaluate the performance of our approach on a 192-core system with large test datasets in excess of 100GB and with varying skew. Experimental results show that our approach is scalable and, at least of in cases of high skew, significantly faster than the state-of-the-art.&lt;br /&gt;
|Projekt=DIAMOND, HAEC&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=HAEC&amp;diff=17159</id>
		<title>HAEC</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=HAEC&amp;diff=17159"/>
		<updated>2015-10-02T11:03:38Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Projekt&lt;br /&gt;
|Kurzname=HAEC&lt;br /&gt;
|Name=Highly Adaptive Energy-Efficient Computing (Sonderforschungsbereich 912)&lt;br /&gt;
|Name EN=Highly Adaptive Energy-Efficient Computing (Collaborative Research Centre SFB 912)&lt;br /&gt;
|Beschreibung DE=Die Art und Weise, wie wir heutzutage das Internet verwenden, hat einen enormen ökonomischen Einfluss auf den Energieverbrauch unserer modernen Gesellschaft. Nimmt man zum Beispiel alle Server des Internets zusammen und berechnet Ihren Stromverbrauch, dann kommt man auf eine Zahl, die ungefähr 2% des jährlichen Stromverbrauchs der USA1 und ca. 26% des Gesamtenergieverbrauchs in Deutschland entspricht. Die Mission des Sonderforschungsbereichs „Hochadaptive Energieeffiziente Systeme“ (HAEC) ist es daher hochgradig energieeffiziente Systeme für unsere moderne IT Infrastruktur zu entwickeln, ohne dabei auf das Leistungspotential dieser Systeme zu verzichten.&lt;br /&gt;
&lt;br /&gt;
Natürlich wäre es ein einfacher Weg den Energieverbrauch jeder einzelnen Systemkomponente zu minimieren. Allerdings arbeiten die meisten Hard- und Softwarekomponenten heute schon auf ihrem Energie/Leistungsoptimum. Zielführender ist es daher ein generelles Verständnis dafür zu entwickeln, wie Software auf Hardwarekomponenten angepasst werden kann und umgekehrt, wie sich Hardwarekomponenten an den schwankenden Ressourcenbedarf der Software anpassen können. Für eine solche Anpassung sind neue Methoden und Werkzeuge nötig, die es Programmierern erlauben energiebewusste Software zu entwickeln. Neue Interaktions- und Kommunikationslösungen werden gebraucht, die eine Vielzahl an Softwarekomponenten auf einem hochgradig parallelen System unterstützten. Mit anderen Worten, wir brauchen einen neuen integrierten Ansatz zur Entwicklung hochadaptiver energieeffizienter Systeme, der alle Ebenen moderner Systeme umfasst.&lt;br /&gt;
&lt;br /&gt;
Der Sonderforschungsbereich HAEC ist ein erster Versuch eines solchen integrativen Ansatzes. Auf Schaltkreisebene versuchen wir innovative Ideen für optische und kabellose Chip zu Chip Kommunikationsmedien zu entwickeln. In der nächsthöheren Netzwerkschicht erforschen wir sichere Hochleistungskodierungsschemata für Netzwerkströme. Kontroll- und Koordinierungslösungen werden untersucht, um die Anpassung der Hardware- und Softwarekomponenten zu steuern. Neue Techniken zur Softwareentwicklung werden durch energiebewusste Ausführungsumgebungen bereitgestellt, die eng mit den Ressourcen-, Strom- und Konfigurationsmanagement Algorithmen auf Betriebssystem -und Anwendungsebene zusammenspielen und neue Internet Applikationen werden durch energiebewusste Dienste ermöglicht. Darüber hinaus erforschen wir formale Methoden, die in der Lage sein werden, ein neues Level an Vertrauen in die entwickelten Systeme zu liefern, da sie die Einhaltung von Garantien formal beweisen können. Unser Ziel ist es die Ergebnisse all diese Entwicklungen in einem Prototypen zu demonstrieren – der HAEC Box – und somit die antreibende Kraft hinter der akademischen und industriellen Entwicklung zukünftiger energieeffizienter Systeme zu werden.&lt;br /&gt;
|Beschreibung EN=The way we use the Internet today has an enormous ecological impact on the increasing energy demand of modern society. For example, the electricity required by the servers that make up the Internet relates to about 2% of the overall electricity consumption in the US1 and to about 26% of the overall energy consumption in Germany. The mission of the collaborative research center “Highly Adaptive Energy-Efficient Computing” (HAEC) is to enable high energy efficiency in today’s computing systems without compromising on high performance.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Certainly a straightforward way for improving energy efficiency would be to reduce the energy consumption of every individual hardware component involved. However, for many components the optimal energy performance point has already been reached. More important is an understanding of how software can adapt to hardware components and vice versa to address the computational problems of modern society in an energy efficient way. This requires new methods and tools to write energy-aware programs, new ways of interaction between the individual pieces that collaborate to solve a problem, new communication technologies to enable this interaction between pieces that are spread across highly scalable parallel systems, and a new multi-layer coordination infrastructure to bring together these technologies. In other words, we need an integrated approach for highly adaptive energy-efficient computing to approach energy-efficient computing at all involved technology levels. &lt;br /&gt;
&lt;br /&gt;
The collaborative research center HAEC is a first attempt to achieve high adaptivity and energy efficiency in such an integrated approach. At the circuit level, we focus on innovative ideas for optical and wireless chip-to-chip communication. At the network level, we research secure, high performance network coding schemes for wired and wireless board-to-board communication. Innovative results at the hardware/software interface level will include energy control loops, which allow hardware to adapt to varying software requirements and vice versa. Software development in general is supported by energy-aware runtimes, energy-aware resource, stream and configuration management schemes and by an analysis framework for high performance/low energy applications. New internet applications are supported by innovations in energy-aware service execution. And, last but not least, formal methods are developed to offer a new quality of assurance in our systems of tomorrow. Demonstrating our results in a joint prototype - the HAEC Box - our goal is to become a pace setter for industry and academia on the design of future energy efficient-computing systems.&lt;br /&gt;
|Kontaktperson=Gerhard Fettweis&lt;br /&gt;
|URL=http://tu-dresden.de/sfb912&lt;br /&gt;
|Start=2011/07/01&lt;br /&gt;
|Ende=2019/06/30&lt;br /&gt;
|Finanziert von=DFG&lt;br /&gt;
|Projektstatus=aktiv&lt;br /&gt;
|Logo=HAEC Logo.png&lt;br /&gt;
|Person=Franz Baader, Markus Krötzsch, Anni-Yasmin Turhan, Alexander Krause, Veronika Thost, Long Cheng, &lt;br /&gt;
|Forschungsgruppe=Automatentheorie, Knowledge Systems&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Beschreibungslogiken&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Semantische Technologien&lt;br /&gt;
}}&lt;br /&gt;
{{Forschungsgebiet Auswahl&lt;br /&gt;
|Forschungsgebiet=Wissensrepräsentation und logisches Schließen&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3054/en&amp;diff=17038</id>
		<title>Inproceedings3054/en</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3054/en&amp;diff=17038"/>
		<updated>2015-09-17T20:52:19Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Page created automatically by parser function on page Inproceedings3054&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Inproceedings3054]]&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3054&amp;diff=17037</id>
		<title>Inproceedings3054</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings3054&amp;diff=17037"/>
		<updated>2015-09-17T20:52:18Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: Die Seite wurde neu angelegt: „{{Publikation Erster Autor |ErsterAutorVorname=Long |ErsterAutorNachname=Cheng |FurtherAuthors=Spyros Kotoulas; Tomas E Ward;  Georgios Theodoropoulos }} {{Inp…“&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; Tomas E Ward;  Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=High Throughput Indexing for Large-scale Semantic Web Data&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2015&lt;br /&gt;
|Month=April&lt;br /&gt;
|Booktitle=Proc. 30th ACM/SIGAPP Symposium On Applied Computing (SAC&#039;15)&lt;br /&gt;
|Pages=416-422&lt;br /&gt;
|Publisher=ACM&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Distributed RDF data management systems become increasingly important with the growth of the Semantic Web. Currently, several such systems have been proposed, however, their indexing methods meet performance bottlenecks either on data loading or querying when processing large amounts of data. In this work, we propose a high throughout index to enable rapid analysis of large datasets. We adopt a hybrid structure to combine the loading speed of similar-size based methods with the execution speed of graph-based approaches, using dynamic data repartitioning over query workloads. We introduce the design and detailed implementation of our method. Experimental results show that the proposed index can indeed vastly improve loading speeds while remaining competitive in terms of performance. Therefore, the method could be considered as a good choice for RDF analysis in large-scale distributed scenarios. &lt;br /&gt;
|Download=2015-High-Throughput.pdf&lt;br /&gt;
|Link=http://dl.acm.org/citation.cfm?doid=2695664.2695920&lt;br /&gt;
|DOI Name=10.1145/2695664.2695920&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-High-Throughput.pdf&amp;diff=17036</id>
		<title>Datei:2015-High-Throughput.pdf</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Datei:2015-High-Throughput.pdf&amp;diff=17036"/>
		<updated>2015-09-17T20:43:37Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4055&amp;diff=17035</id>
		<title>Inproceedings4055</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4055&amp;diff=17035"/>
		<updated>2015-09-17T20:38:33Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; Tomas E. Ward; Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Design and Evaluation of Parallel Hashing over Large-scale Data&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=Dezember&lt;br /&gt;
|Booktitle=Proc. 21st IEEE International Conference on High Performance Computing (HiPC&#039;14)&lt;br /&gt;
|Pages=1-10&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=High-performance analytical data processing systems often run on servers with large amounts of memory. A common data structure used in such environment is the hash tables. This paper focuses on investigating efficient parallel hash algorithms for processing large-scale data. Currently, hash tables on distributed architectures are accessed one key at a time by local or remote threads while shared-memory approaches focus on accessing a single table with multiple threads. A relatively straightforward “bulk-operation” approach seems to have been neglected by researchers. In this work, using such a method, we propose a high-level parallel hashing framework, Structured Parallel Hashing, targeting efficiently processing massive data on distributed memory.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
We present a theoretical analysis of the proposed method and describe the design of our hashing implementations. The evaluation reveals a very interesting result - the proposed straightforward method can vastly outperform distributed hashing methods and can even offer performance comparable with approaches based on shared memory supercomputers which use specialized hardware predicates. Moreover, we characterize the performance of our hash implementations through extensive experiments, thereby allowing system developers to make a more informed choice for their high-performance applications.&lt;br /&gt;
|Download=2014-Design-Evaluation.pdf&lt;br /&gt;
|Link=http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=7116909&lt;br /&gt;
|DOI Name=10.1109/HiPC.2014.7116909&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4052&amp;diff=17034</id>
		<title>Inproceedings4052</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4052&amp;diff=17034"/>
		<updated>2015-09-17T20:36:49Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Spyros Kotoulas; Tomas E. Ward; Georgios Theodoropoulos&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Robust and Skew-resistant Parallel Joins in Shared-nothing Systems&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=November&lt;br /&gt;
|Booktitle=Proc. 23rd ACM International Conference on Information and Knowledge Management (CIKM&#039;14)&lt;br /&gt;
|Pages=1399-1408&lt;br /&gt;
|Publisher=ACM&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=The performance of joins in parallel database management systems is critical for data intensive operations such as querying. Since data skew is common in many applications, poorly engineered join operations result in load imbalance and performance bottlenecks. State-of-the-art methods designed to handle this problem offer significant improvements over naive implementations. However, performance could be further improved by removing the dependency on global skew knowledge and broadcasting. In this paper, we propose PRPQ (partial redistribution &amp;amp; partial query), an efficient and robust join algorithm for processing large-scale joins over distributed systems. We present the detailed implementation and a quantitative evaluation of our method. The experimental results demonstrate that the proposed PRPQ algorithm is indeed robust and scalable under a wide range of skew conditions. Specially, compared to the state-of-art PRPD method, we achieve 16% - 167% performance improvement and 24% - 54% less network communication under different join workloads.&lt;br /&gt;
|Download=2014-Robust-Skew.pdf&lt;br /&gt;
|Link=http://dl.acm.org/citation.cfm?doid=2661829.2661888&lt;br /&gt;
|DOI Name=10.1145/2661829.2661888&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4053&amp;diff=17033</id>
		<title>Inproceedings4053</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4053&amp;diff=17033"/>
		<updated>2015-09-17T20:34:14Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Ilias&lt;br /&gt;
|ErsterAutorNachname=Tachmazidis&lt;br /&gt;
|FurtherAuthors=Long Cheng; Spyros Kotoulas; Grigoris Antoniou; Tomas E Ward&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Massively Parallel Reasoning under the Well-Founded Semantics using X10&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=November&lt;br /&gt;
|Booktitle=Proc. 26th IEEE International Conference on Tools with Artificial Intelligence (ICTAI&#039;14)&lt;br /&gt;
|Pages=162-169&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Academia and industry are investigating novel approaches for processing vast amounts of data coming from enterprises, the Web, social media and sensor readings in an area that has come to be known as Big Data. Logic programming has traditionally focused on complex knowledge structures/programs. The question arises whether and how it can be applied in the context of Big Data. In this paper, we study how the well-founded semantics can be computed over huge amounts of data using mass parallelization. Specifically, we propose and evaluate a parallel approach based on the X10 programming language. Our experiments demonstrate that our&lt;br /&gt;
approach has the ability to process up to 1 billion facts within minutes.&lt;br /&gt;
|Download=2014-Massively- Parallel.pdf&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4054&amp;diff=17032</id>
		<title>Inproceedings4054</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4054&amp;diff=17032"/>
		<updated>2015-09-17T20:32:16Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Long&lt;br /&gt;
|ErsterAutorNachname=Cheng&lt;br /&gt;
|FurtherAuthors=Yue Ma&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Investigating Distributed Approaches to Efficiently Extract Textual Evidences for Biomedical Ontologies&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=November&lt;br /&gt;
|Booktitle=Proc. 14th IEEE International Conference on BioInformatics and BioEngineering (BIBE&#039;14)&lt;br /&gt;
|Pages=220-225&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Heterogeneous data resources in biomedicine become available both in structured and unstructured formats, such as scientific publications, healthcare guidelines, controlled vocabularies, and formal ontologies. Bridging the gaps among these heterogeneous data is useful to discovery implicit knowledge. To make this happen, efficient computational approaches are a necessity for applications in such a knowledge- and data- intensive domain. In this paper, we first define a particular task, relation alignment, which is to identify textual evidences for biomedical ontologies. Then, we investigate two parallel approaches for this task over distributed systems and present the details of their implementations. Moreover, we characterize the performance of our methods through extensive experiments, thereby allowing researchers to make a more informed choice in the presence of large-scale biomedical data.&lt;br /&gt;
|Download=2014-Inves-Distributed.pdf&lt;br /&gt;
|Projekt=DIAMOND&lt;br /&gt;
|Forschungsgruppe=Automatentheorie, Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=16932</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=16932"/>
		<updated>2015-09-02T13:53:36Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 43510&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web.&lt;br /&gt;
|Info EN=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=16704</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=16704"/>
		<updated>2015-07-27T17:00:38Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web. &lt;br /&gt;
|Info EN=I am currently working as a  Post-Doctoral Researcher  in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at TU Dresden. My research interests mainly include:&lt;br /&gt;
&lt;br /&gt;
* Distributed computing&lt;br /&gt;
* Large-scale data processing&lt;br /&gt;
* Data management&lt;br /&gt;
* Semantic web. &lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9734</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9734"/>
		<updated>2015-02-19T20:02:51Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and an Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include Distributed computing, Large-scale data processing, Data management and Semantic web.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and an Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include Distributed computing, Large-scale data processing, Data management and Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9728</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9728"/>
		<updated>2015-02-18T20:51:27Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include Distributed computing, Large-scale data processing, Data management and Semantic web.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include Distributed computing, Large-scale data processing, Data management and Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9478</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9478"/>
		<updated>2014-12-12T10:13:43Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing, Distributed data management systems and Semantic web.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014).&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing, Distributed data management systems and Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9477</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9477"/>
		<updated>2014-12-12T10:13:02Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=APB 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems and Semantic web.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems and Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9114</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=9114"/>
		<updated>2014-11-17T09:02:37Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=INF 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems and Semantic web.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems and Semantic web.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=8935</id>
		<title>Long Cheng</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Long_Cheng&amp;diff=8935"/>
		<updated>2014-11-07T12:26:56Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Mitarbeiter&lt;br /&gt;
|Vorname=Long&lt;br /&gt;
|Nachname=Cheng&lt;br /&gt;
|Akademischer Titel=Dr.&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
|Stellung=Wissenschaftlicher Mitarbeiter&lt;br /&gt;
|Ehemaliger=0&lt;br /&gt;
|Telefon=+49 351 463 38043&lt;br /&gt;
|Fax=+49 351 463 37959&lt;br /&gt;
|Email=long.cheng@tu-dresden.de&lt;br /&gt;
|Raum=INF 3034&lt;br /&gt;
|Bild=Long Cheng.jpg&lt;br /&gt;
|Info=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems, Semantic web and Kownledge representation and reasoning.&lt;br /&gt;
|Info EN=I am currently a Postdoc Researcher in the [[Forschungsgruppe::Knowledge Systems]] Group led by Dr. [[Miterarbeiter:: Markus Krötzsch]] at the Technische Universität Dresden. Before this, I was a PhD student at the National University of Ireland Maynooth (2011-2014) and advised by [http://www.eeng.nuim.ie/~tward/ Dr. Tomas Ward], [http://www.gtheodoropoulos.com/ Prof. Georgios Theodoropoulos] and [http://researcher.watson.ibm.com/researcher/view.php?person=ie-Spyros.Kotoulas Dr. Spyros Kotoulas].&lt;br /&gt;
&lt;br /&gt;
I got my B.Eng degree from Harbin Insititute of Technology, China (2007) and received the M.Sc degree from Universität Duisburg-Essen, Germany (2010). Additionally, I worked as an Engineer in Huawei Technologies Germany (Düsseldorf) in 2011 and a Research Assistant in IBM Research Ireland (Dublin) during 2011-2014.&lt;br /&gt;
&lt;br /&gt;
I joined in TU Dresden in June 2014. My research interests include High performance computing, Large-scale data processing (especially for RDF data, Wikidata and Biomedical data), Distributed data management systems, Semantic web and Kownledge representation and reasoning.&lt;br /&gt;
|Google Scholar=http://scholar.google.de/citations?user=aI-bwLgAAAAJ&amp;amp;hl=en&lt;br /&gt;
|Alternative URI=http://lat.inf.tu-dresden.de/~lcheng/&lt;br /&gt;
|Publikationen anzeigen=1&lt;br /&gt;
|Abschlussarbeiten anzeigen=0&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
	<entry>
		<id>https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4053&amp;diff=8506</id>
		<title>Inproceedings4053</title>
		<link rel="alternate" type="text/html" href="https://iccl.inf.tu-dresden.de/w/index.php?title=Inproceedings4053&amp;diff=8506"/>
		<updated>2014-11-02T16:42:49Z</updated>

		<summary type="html">&lt;p&gt;Long Cheng: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{Publikation Erster Autor&lt;br /&gt;
|ErsterAutorVorname=Ilias&lt;br /&gt;
|ErsterAutorNachname=Tachmazidis&lt;br /&gt;
|FurtherAuthors=Long Cheng; Spyros Kotoulas; Grigoris Antoniou; Tomas E Ward&lt;br /&gt;
}}&lt;br /&gt;
{{Inproceedings&lt;br /&gt;
|Referiert=1&lt;br /&gt;
|Title=Massively Parallel Reasoning under the Well-Founded Semantics using X10&lt;br /&gt;
|To appear=0&lt;br /&gt;
|Year=2014&lt;br /&gt;
|Month=November&lt;br /&gt;
|Booktitle=Proc. 26th IEEE International Conference on Tools with Artificial Intelligence (ICTAI&#039;14)&lt;br /&gt;
|Publisher=IEEE&lt;br /&gt;
}}&lt;br /&gt;
{{Publikation Details&lt;br /&gt;
|Abstract=Academia and industry are investigating novel approaches for processing vast amounts of data coming from enterprises, the Web, social media and sensor readings in an area that has come to be known as Big Data. Logic programming has traditionally focused on complex knowledge structures/programs. The question arises whether and how it can be applied in the context of Big Data. In this paper, we study how the well-founded semantics can be computed over huge amounts of data using mass parallelization. Specifically, we propose and evaluate a parallel approach based on the X10 programming language. Our experiments demonstrate that our&lt;br /&gt;
approach has the ability to process up to 1 billion facts within minutes.&lt;br /&gt;
|Download=2014-Massively- Parallel.pdf&lt;br /&gt;
|Forschungsgruppe=Knowledge Systems&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Long Cheng</name></author>
	</entry>
</feed>