EUROFIT - Integration, Homogenisation and Extension of the Scope Anthropometric Data Stored in Large EU Pools
Time Slot: June 06, 2012 - 12:00
Presented by: Juan Carlos González García
EUROFIT: Integration, Homogenisation and Extension of the Scope Anthropometric Data Stored in Large EU Pools
Over the last decades, human body metrics have been used to improve human-product interaction. Along this period, the use of 1D-measurements in “classic” ergonomic applications (e.g. workplace design) has been extended to consumer goods industries such as the automotive, apparel, furniture or orthopaedic sectors.
New technologies for the gathering, storage and analysis of anthropometric data (i.e. 3D scanners) have boosted the availability of digital anthropometric resources. Since 1999, over 16 large-scale national body scanning surveys have been conducted around the world (six in Europe) gathering 3D shape data from over 120,000 subjects (~50,000 Europeans). The availability of these data pools has created the opportunity to exploit shape information beyond current 1D-measure use.
However, these data pools are dispersed and heterogeneous (e.g. obtained through different scanning technologies and owned by different organisations) and, above all, the exploitation of 3D data at industry level requires knowledge, skills and resources beyond the means of companies, especially SMEs. These barriers have until now confined the use of existing 3D shape data to scientific research.
Our overall aim is thus to implement an online platform and an open framework that enables:
-
Designers and industrialists to draw useful shape information and use it in their product development processes in an easy and direct way
-
Database owners to pool the data they already hold and obtain revenues from its exploitation
-
Third party IT companies to develop new services to reuse pooled data in unforeseen ways R&D work will focus on the systematisation of proven methods for 3D shape data aggregation and analysis in a reliable but economically sustainable way, as well as the development of sector-specific applications and user-friendly interfaces. Project work will include the aggregation of 4 national databases and the running of a four-month demonstration by industrial end-users.
GAPFILLER - GNSS DAta Pool for PerFormances PredIction and SimuLation of New AppLications for DevelopERs
Time Slot: June 06, 2012 - 12:08
Presented by: Pierre Gruyer
GAPFILLER : The Web Portal of the GNSS Community
With the increase of criticality level of the GPS applications, products manufacturers and developers are now requesting a relevant mean to predict performances and reliability of their future applications. Often SMEs have ideas that sometimes cannot be implemented because they depend on the availability of data resources or specialised tools that are too expensive to obtain.
In this context, GAPFILLER project aims at filling the gap between big manufacturers and SMEs by providing the researchers and developers’ community with a unique extensible data pool enabling performances prediction and simulation of new Global Navigation Satellite System (GNSS) based applications and algorithms. GAPFILLER consortium aim is to create quality content and innovative services and to allow people to access and use online content and services across GNSS technologies barriers.
The main objective of the GAPFILLER project is to define, build and disseminate a powerful data pool of worldwide GNSS measurements enabling low cost tests campaigns, performances predictions and simulations of new GNSS based applications. Main targeted benefits are:
-
The development of an innovative use of a common GNSS database for performances assessment that will drastically reduce the cost of tests campaigns and certification process.
-
The adaptation of the EGNOS aeronautical integrity concept to new application domains that will allow SMEs to gain the know-how for exploiting it and getting competitive advantage.
-
The knowledge transfer between technology providers/data brokers and content providers (especially SMEs) boosting development speed of EGNOS and GALILEO services.
By promoting adoption of EGNOS and emergence of a new data economy related to GALILEO, the consortium will maximize the Return Of Investment of one of the most important budget line of the European Commission that has already plan to invest 3.4 billion Euros for the implementation and launch of the GALILEO satellites constellation.
QTLaunchPad - Preparation and Launch of a Large-Scale Action for Quality Translation Technology
Time Slot: June 06, 2012 - 12:15
Presented by: Hans Uszkoreit
Preparation and Launch of a Large-Scale Action for Quality Translation Technology
QTLaunchPad will assemble the best European centres of Machine Translation (MT) research, translation technology users, providers of language technology and resources, as well as integrators, to provide data, tools and test suites for quality assessment and extend an existing online platform for resource sharing to the needs of quality MT research.
SimpleFleet - Democratizing Fleet Management
Time Slot: June 06, 2012 - 12:22
Presented by: Rüdiger Ebendt
Challenges and Objectives of Project SimpleFleet
GPS positioning devices are becoming a commodity sensor platform with the emergence and popularity of smartphones and ubiquitous networking. While the positioning capability has been exploited in location-based services, so has its spatiotemporal cousin, tracking, so far only been considered in costly and complex fleet management applications.
The proposed project, SimpleFleet will make it easy for SMEs, both, from a technological and business perspective, to create (Mobile) Web-based fleet management applications. For this purpose, we build a large data pool comprising base data such as maps and traffic data from dedicated providers. A simple interface will provide a means to connect user-contributed data streams to this pool. In addition and to increase the market potential of the data pool, we also want to address the related geomarketing domain, which uses travel information in various geo-statistical analysis methods as well as visualizations of the data to be used in online
and print publications.
An algorithmic framework dubbed “TrafficIntelligence” that includes map-matching algorithms, vehicle routing services and a statistics package will utilize the collected data and provide value-added service. SMEs will be able to access the data and services by means of a Web-based API, a Software Development Kit (SDK) wrapping API access for specific languages and environments and Application Frameworks for rapid application development for target platforms such as Web (JavaScript), and iPhone and Android mobile platforms.
GPS positioning devices are becoming a commodity sensor platform with the emergence and popularity of smartphones and ubiquitous networking. While the positioning capability has been exploited in location-based services, so has its spatiotemporal cousin, tracking, so far only been considered in costly and complex fleet management applications. The proposed project, SimpleFleet will make it easy for SMEs, both, from a technological and business perspective, to create (Mobile) Web-based fleet management applications. For this purpose, we build a large data pool comprising base data such as maps and traffic data from dedicated providers. A simple interface will provide a means to connect user-contributed data streams to this pool. In addition and to increase the market potential of the data pool, we also want to address the related geomarketing domain, which uses travel information in various geo-statistical analysis methods as well as visualizations of the data to be used in online and print publications. An algorithmic framework dubbed “TrafficIntelligence” that includes map-matching algorithms, vehicle routing services and a statistics package will utilize the collected data and provide value-added service. SMEs will be able to access the data and services by means of a Web-based API, a Software Development Kit (SDK) wrapping API access for specific languages and environments and Application Frameworks for rapid application development for target platforms such as Web (JavaScript), and iPhone and Android mobile platforms.
Vista-TV - Video Stream Analytics for Viewers in the TV Industry
Time Slot: June 06, 2012 - 12:30
Presented by: Guus Schreiber
Video Stream Analytics for Viewers in the TV Industry
Live video content is increasingly consumed over IP networks in addition to traditional broadcasting. The move to IP provides a huge opportunity to discover what people are watching in much greater breadth and depth than currently possible through interviews or set-top box based data gathering by rating organizations, because it allows direct analysis of consumer behavior via the logs they produce. The ViSTA-TV project will gather consumers’ anonymized viewing behavior as well as the actual video streams from broadcasters and IPTV transmitters to combine them with enhanced electronic program-guide information as the input for a holistic live-stream data mining analysis: the basis for an SME-driven market-place for TV viewing-behavior information.
SOPCAWIND - SOFTWARE FOR THE OPTIMAL PLACE CALCULATION FOR WIND-FARMS
Time Slot: June 06, 2012 - 12:37
Presented by: Jose María Cabero
Data pooling for the optimal localisation of wind-farms
The design of wind farms is a complex task with the optimal siting of each wind turbine as the most critical aspect for the productivity of the farm. There are a number of data which are critical to determine the best location of a wind farm. Among them there are the wind potential of a region, local environment characteristics, potential interference on communication systems, visual impact, the existence of archaeological sites, socio‐economic factors, etc. The data to evaluate the above‐mentioned issues come from different sources such as public administrations and private companies, with different approaches and interests, and data of different formats and nature. This is usually a handicap for SMEs that have to consider all these collateral but crucial aspects.
The main objective of SOPCAWIND is to create new services and products for SMEs based on the development of a software tool being able to optimise the location for a wind farm based on a large and heterogeneous set of data containing information from different fields (weather, geography, environment, archaeology, radiocommunications and socialeconomy), that will be treated, converted, validated, and maintained for this purpose.
DOPA - Data Supply Chains for Pools, Services and Analytics in Economics and Finance
Time Slot: June 06, 2012 - 12:45
Presented by: Kostas Tzoumas
Linking and Analyzing Big Data with Stratosphere and DOPA
In this talk we will present Stratosphere, a framework for massively parallel data processing, and showcase how we intend to use that framework in the DOPA project in order to create and link large data pools of structured and unstructured data. The value of integrated data pools is demonstrated by two examples that create novel applications in market intelligence and financial data analysis.
Stratosphere (www.stratosphere.eu) is a collaborative research unit funded by the German Science Foundation (DFG), the European Institute of Technology (EIT) and industry partners. Stratosphere provides a scalable, massively parallel data processor with a programming abstraction that extends the popular map/reduce programming model for big data analytics with additional second order functions, to enable processing joins, iterations, and time-series operations on huge data sets. The platform also offers higher-level language abstractions for information extraction and integration, and is used by EIT partners for research and teaching in data-intensive computing. The system is available open-source.
The DOPA project (www.dopa-project.eu) is an EU funded Seventh Framework Program STREP that focuses on linking large data pools of both structured and unstructured data using data supply chains. Here, the goal is to multiply the utility of each individual service while simultaneously sharing the costs between them. This way DOPA lowers the barrier of entry for SMEs that need to perform advanced analytics across multiple data pools, since the required input data as well as the processing environment do not have to be provided by the SME itself. Currently, a web crawl archive and a structured fact database are used as initial data pools. However, the extensible architecture of DOPA is meant to accommodate more data pools in the future. DOPA used Stratosphere as the query processor that generates data supply chains. Use cases in market intelligence and financial data analysis illustrate the potential of the DOPA vision.
BIOPOOL - Services associated to digitalised contents of tissues in biobanks across Europe
Time Slot: June 06, 2012 - 12:52
Presented by: Roberto Bilbao
Services associated to digitalised contents of tissues in biobanks across Europe – BIOPOOL
Nowadays it has become common practice to take digital images of thin slices of biopsies that are obtained for studying the composition of cells, glands, tissues and organs, and the possible pathologies that may affect them. These images are of high interest in medical diagnostics, research and education.
Pathology departments in hospitals and Biobanks are usually the facilities that provide archived biological samples for use in life sciences. They manage the tissue samples, the associated digital images and other complementary digital data (health information such as pathologies or treatments followed by the patient). Although most of the pathology departments and biobanks adequately capture and store digital images of the different biologic materials, it is not so usual for biobanks to adequately associate the representative health information to the digital images of their samples and it is even less usual sharing these images in a network. So, the digital images are usually spread all over different systems stored in different formats, databases and facilities belonging to different types of institutions and they are not easily identifiable and reachable. This creates a difficult environment for sharing and reusing this type of data between different interested organisations.
The BIOPOOL consortium project has been created to carry out a new approach, that arises from the need of pathology departments and biobanks of sharing, exchanging, processing, understanding and exploiting the digital histology images and the data associated to the biologic material stored in these institutions. The project will develop the needed technology to extract and gather this digital information from different pools, analyse it, and being able to compare it and to score images similar to one provided as a search pattern based on an innovative Content Based Image Retrieval (CBIR) system capable of searching histological images using different mixed text and image queries. BIOPOOL will establish a complete intelligent biobank and pathology department network, building a constructive basis for pan-European cooperation in diagnosis and medical research.
Seven partners from four different countries collaborate together in BIOPOOL. A great effort is done by the SMEs within this consortium project. The results coming from BIOPOOL will be translated into new services and products using the technologies and data pools provided by the other organisations allowing more precise actuations in digital pathology, especially in diagnosis.
Fusepool - Fusing and pooling information for product/service development and research
Time Slot: June 06, 2012 - 14:00
Presented by: Michael Kaschesky
Fusepool - Fusing and pooling information for product/service development and research
The Internet is becoming an integral element of the value chain supporting as well as enabling new business interactions and transactions. However, small and medium enterprises (SMEs) are not yet leveraging the full potential of the Internet for more efficient, digitally supported and data-driven business insights and actions. For example, product and service offerings lack efficient detection and matching with sourcing requirements resulting in significant search costs for potential customers. In research and development, technology intelligence lacks easy detectability of technological opportunities and threats combined with efficient matching to own competences and solutions.
The international joint project "Fusepool" develops automated tools and intelligent interfaces for efficiently matching data harvested from various sources to internal business needs, be they in research and development, product marketing or areas such as product branding, public opinion, and civil emergency support. The project builds on proven technologies developed by the project partners as well as external providers, such as Semantic Similarity Matching algorithms or the Good Relations Ontology for the publication and quotation of product information. Existing systems of knowledge representation are translated into reusable Semantic Web ontologies. For example, data from multiple, diverse sources are gathered, analyzed, and integrated in a reusable data format RDF (Resource Description Framework) and published as Linked Open Data (LOD).
The core benefit of project "Fusepool" concerns more efficient and effective detection and matching of data for real-time decision making based on rapid access to timely, context-aware, and needs-oriented information.
PortDial - Language Resources for Portable Multilingual Spoken Dialogue Systems
Time Slot: June 06, 2012 - 14:08
Presented by: Aris Karanikas
PortDial - Language Resources for Portable Multilingual Spoken Dialogue Systems
PortDial will set up a platform for creating, cleaning-up and publishing multilingual domain ontologies and grammars for prototyping of spoken dialogue systems in the domains of entertainment, banking and customer service. The platform will enable rapid and cost-effective porting of voice applications into new domains and languages, serving in particular the SMEs in the mobile application development industry, but also the research community.
smeSpire - A European Community of SMEs built on Environmental Digital Content and Languages
Time Slot: June 06, 2012 - 14:15
Presented by: Giacomo Martirano
smeSpire
The INSPIRE Directive 2007/2/EC, establishes an Infrastructure for Spatial Information in Europe, requiring large amounts of environmental digital content to be made accessible across Europe, resulting in a data pool that is expected to be of huge value for a myriad of value-added applications. The INSPIRE Implementing Rules Legal Acts outlines these data pools, but more work is needed.
Making data available according to the INSPIRE standards in 30 countries using 22 languages requires specific skill sets that few public authorities have. The management of this content represents an opportunity for SMEs active in this sector.
SMEs can enable countries to fulfil the Directive, creating new market opportunities with increased potential for innovation and new jobs. Due to the legal requirements, the INSPIRE implementation becomes the entry-point for crucial business opportunities, opening new, or reinforcing existing perspectives (including Linked Open Data, Sensor Web, cloud computing and other e-environment application domains).
SmeSpire’s purpose is to encourage and enable the participation of SMEs in the mechanisms of harmonising and making large scale environmental content available. smeSpire will:
-
assess market potential for SMEs in relation to INSPIRE as an integral component of the Digital Agenda for Europe, describing obstacles for SMEs to enter this market in terms of knowledge gaps
-
collate, translate and exploit a Best Practice Catalogue in the management of environmental content
-
Develop a multilingual package to train environmental data analysts in the maintenance and exploitation of environmental data commons, and
-
create a network capable of transferring result-driven knowledge throughout Europe with research centres, environmental agencies, progressive technology providers and digital content providers.
smeSpire offers the Commission an early integration of results for relevant programmes (ISA) directly into the knowledge base of geo-ICT SMEs.
CODE - Commercially empowered Linked Open Data Ecosystems in Research
Time Slot: June 06, 2012 - 14:22
Presented by: Stefanie Lindstaedt
Commercially empowered Linked Open Data Ecosystems in Research – The CODE Approach
Linked Open Data (LOD) shows enormous potential in becoming the next big evolutionary step of the WWW. However, this potential remains largely untapped due to missing usage and commercialisation strategies.
CODE’s vision is to establish the foundation for a web-based, commercially oriented ecosystem for Linked Open Data. This ecosystem establishes a sustainable and commercial value-creation-chain among traditional (e.g. data provider and consumer) and non-traditional (e.g. data analyst) roles in data marketplaces. Monetary incentives will motivate people to analyse, organise and integrate LOD with unstructured information sources thereby increasing data quality and quantity.
Our use case focuses on research papers as a source for mining facts and their integration into LOD repositories and light-weight ontologies. Hence, we will leverage the wealth of knowledge contained in research publications on a semantic, machine-readable level by creating the Linked Science Data cloud. This cloud will have an impact on innovation driven businesses by making scientific knowledge more accessible and transparent.
Therefore, we will research and develop:
-
crowd-sourcing enabled semantic enrichment & integration techniques for integrating facts contained in unstructured information into the LOD cloud
-
federated, provenance-enabled querying methods for fact discovery in LOD repositories
-
web-based visual analysis interfaces to support human based analysis, integration and organisation of facts
-
socio-economic factors - roles, revenue-models and value chains - realisable in the envisioned ecosystem.
Plan4Business - A service platform for aggregation, processing and analysis of urban and regional planning data
Time Slot: June 06, 2012 - 14:30
Presented by: Joachim Rix
Plan 4Business - A service platform for aggregation, processing and analysis of urban and regional planning data
Urban and Regional Planning data sets are not aggregated so far, and thus it is very difficult to use them for any other purpose than for printing of simple publishing by the authorities that they were created by. Creating time series or comparative analyses on these data sets is not yet possible; researchers, spatial planners and professionals from the real estate world and other disciplines, such as insurance industry, investors, or market-relevant activities related to urban development have a growing stake in such capabilities. The plan4business project consequently aims to develop a platform that can serve users a full catalogue of planning data such as transport infrastructure, regional plans, urban plans and zoning plans. The platform offers clients not just the data itself in integrated, harmonised and thus ready-to-use form, but it also offers rich analysis and visualisation services via an API and an interactive web frontend. Functions offered range from simple statistical analysis to complex trend detection and to 2D/3D representations of these. The two main challenges that have so far hindered usage of planning data in such a manner are the required integration and harmonisation, which needs to be highly automated, as well as the need for an ICT system that can efficiently answer complex queries over the diverse and complex planning data sets. The business model for the plan4business platform foresees several different groups of active stakeholders: data providers (planning authorities, engineering bureaus, researchers), data curators (who perform integration and quality assurance), clients and the data brokers who will be hosting and exploiting the plan4business portal. Revenue is to be generated via on-demand and subscription services to different customer groups ranging from environmental and planning authorities and companies to banks and real estate companies and developers.
EUCLID - EdUcational Curriculum for the usage of Linked Data
Time Slot: June 06, 2012 - 14:37
Presented by: Elena Simperl
EdUcational Curriculum for the usage of Linked Data
Linked Data has established itself as the de facto means for the publication of structured data over the Web, enjoying amazing growth in terms of the number of organizations committing to use its core principles for exposing and interlinking data sets for seamless exchange, integration, and reuse. More and more ICT ventures offer innovative data management services on top of Linked (Open) Data, creating a demand for data practitioners possessing skills and detailed knowledge in this area. Ensuring the availability of such expertise will prove crucial if European businesses are to reap the full benefits of these advanced data management technologies, and the know-how accumulated over the past years by researchers, technology enthusiasts and early adopters in various European Member States.
EUCLID will contribute to this goal by providing a comprehensive educational curriculum, supported by multi-modal learning materials and highly visible eLearning distribution channels, tailored to the real needs of data practitioners. Building upon the experience of the consortium accumulated in over 20 Linked Data projects with over 40 companies and public offices in more than 10 countries, complemented by feedback from more than 20 training events, and an in-depth analysis of the community discourse through mailing lists, discussion forums, Twitter, and the blogosphere, the curriculum will focus on techniques and software to integrate, query, and visualize Linked Data, as core areas in which practitioners state to require most assistance.
The curriculum will be realized as living learning materials on a community Web site, and will be evaluated, refined, and extended in a webinar series, face-to-face training, and through continuous community feedback and contributions coordinated by a designated community manager. A significant share of the materials will consist of examples referring to real-world data sets and application scenarios, code snippets and demos that developers can run on their machines, as well as best practices and how-tos. In its final form, the curriculum will be delivered as an eBook series distributed via iTunes U, complemented by the materials collected on the community Web site. By providing these key knowledge-transfer components, EUCLID will not only promote the industrial uptake of Linked Data best practices and technologies, but, perhaps more importantly, will contribute to their further development and consolidation and support the sustainability of the community - all essential aspects given the novelty of the field and the rapid pace at which it has recently advanced.