Powered by OpenAIRE graph
Found an issue? Give us feedback

Oracle (United States)

Oracle (United States)

Funder
Top 100 values are shown in the filters
Results number
arrow_drop_down
15 Projects, page 1 of 3
  • Funder: European Commission Project Code: 223996
    more_vert
  • Funder: UK Research and Innovation Project Code: EP/L016427/1
    Funder Contribution: 4,746,530 GBP

    Overview: We propose a Centre for Doctoral Training in Data Science. Data science is an emerging discipline that combines machine learning, databases, and other research areas in order to generate new knowledge from complex data. Interest in data science is exploding in industry and the public sector, both in the UK and internationally. Students from the Centre will be well prepared to work on tough problems involving large-scale unstructured and semistructured data, which are increasingly arising across a wide variety of application areas. Skills need: There is a significant industrial need for students who are well trained in data science. Skilled data scientists are in high demand. A report by McKinsey Global Institute cites a shortage of up to 190,000 qualified data scientists in the US; the situation in the UK is likely to be similar. A 2012 report in the Harvard Business Review concludes: "Indeed the shortage of data scientists is becoming a serious constraint in some sectors." A report on the Nature web site cited an astonishing 15,000% increase in job postings for data scientists in a single year, from 2011 to 2012. Many of our industrial partners (see letters of support) have expressed a pressing need to hire in data science. Training approach: We will train students using a rigorous and innovative four-year programme that is designed not only to train students in performing cutting-edge research but also to foster interdisciplinary interactions between students and to build students' practical expertise by interacting with a wide consortium of partners. The first year of the programme combines taught coursework and a sequence of small research projects. Taught coursework will include courses in machine learning, databases, and other research areas. Years 2-4 of the programme will consist primarily of an intensive PhD-level research project. The programme will provide students with breadth throughout the interdisciplinary scope of data science, depth in a specialist area, training in leadership and communication skills, and appreciation for practical issues in applied data science. All students will receive individual supervision from at least two members of Centre staff. The training programme will be especially characterized by opportunities for combining theory and practice, and for student-led and peer-to-peer learning.

    more_vert
  • Funder: UK Research and Innovation Project Code: EP/K01790X/1
    Funder Contribution: 618,883 GBP

    Traditionally, most software projects have been tackled using a single programming language. However, as our ambitions for software grow, this is increasingly unnatural: no single language, no matter how "good", is well-suited to everything. Increasingly, different communities have created or adopted non-traditional languages - often, though not always, under the banner of Domain Specific Languages (DSLs) - to satisfy their specific needs. Consider a large organisation. Its back-end software may utilise SQL and Java; its desktop software C#; its website back-end PHP and the front-end Javascript and HTML5; reports may be created using R; and some divisions may prototype software with Python or Haskell. Though the organisation makes use of different languages, each must execute in its own silo. We currently have few techniques to allow a single running program to be written using multiple languages. In the Cooler project, we call this the "runtime composition" problem: how can languages execute directly alongside each other, exchange data, call each other, optimise with respect to each other, etc.? The chief existing technique for composing language runtimes is to translate all languages in the composition down to a base language, most commonly the byte code for one of the "big" Virtual Machines (VMs) - Java's HotSpot or .NET's CLR. Though this works well in some cases, it has two major problems. Firstly, a VM will intentionally target a specific family of languages, and may not provide the primitives needed by languages outside that family. HotSpot, for example, does not support tail recursion or continuations, excluding many advanced languages. Secondly, the primitives that a VM exposes may not allow efficient execution of programs. For example, dynamically typed languages running on HotSpot run slower than their seemingly much less sophisticated "home brew" VMs. The Cooler project takes a new approach to the composition problem. It hypothesizes that meta-tracing will allow the efficient composition of arbitrary language runtimes. Meta-tracing is a recently developed technique that creates efficient VMs with custom Just-in-Time (JIT) compilers. Firstly, language designers write an interpreter for their chosen language. When that interpreter executes a user's program, hot paths in the code are recorded ("traced"), optimised, and converted into machine code; subsequent calls then use that fast machine code rather than the slow interpreter. Meta-tracing is distinct from partial evaluation: it records actual actions executed by the interpreter on a specific user program. Meta-tracing is an exciting new technique for three reasons. Firstly, it leads to fast VMs: the PyPy VM (a fully compatible reimplementation of Python) is over 5 times faster than CPython (the C-based Python VM) and Jython (Python on the JVM). Secondly, it requires few resources: a meta-tracing implementation of the Converge language was completed in less than 3 person months, and runs faster than CPython and Jython. Third, because the user writes the interpreter themselves, there is no bias to any particular family of languages. The Cooler project will initially design the first language specifically designed for meta-tracing (rather than, as existing systems, reusing an unsuitable existing language). This will enable the exploration of various aspects of language runtime composition. First, cross-runtime sharing: how can different paradigms (e.g. imperative and functional) exchange data and behaviour? Second, optimisation: how can programs written in multiple paradigms be optimised (space and time)? Finally, the limits of the approach will be explored through known hard problems: cross-runtime garbage collection; concurrency; and to what extent runtimes not designed for composition can be composed. Ultimately, the project will allow users to compose together runtimes and programs in ways that are currently unfeasible.

    more_vert
  • Funder: European Commission Project Code: 212785
    more_vert
  • Funder: UK Research and Innovation Project Code: ST/G003556/1
    Funder Contribution: 86,993 GBP

    In 2003. over 44,000 people in the UK were diagnosed with breast cancer. This is now the commonest cancer occurring in the UK. The lifetime risk of developing breast cancer is 1 in 9, and while most of the women who get breast cancer are past their menopause, almost 8,000 diagnosed each year are under 50 years old. Improving outcomes is a key challenge in treatment. High throughput genomic methods, such as expression profiling of frozen tissue samples using microarrays, have resulted in the discovery of many novel gene signatures that are correlated with clinical outcomes in cancer treatment. It is essential that these potential biomarkers are validated on large numbers of independent samples prior to clinical use. Tissue microarrays (TMA) created from paraffin-embedded tumour samples from large clinical trials are the ideal reagent for these validation experiments. However, the analysis and scoring of antibody-based markers on TMAs that may contain thousands of patient samples presents major challenges for pathology reporting and image handling. Our initial pilot PathGrid study (Oct 2007 to Oct 2008) is using a range of techniques which have been developed in astronomy to both analyse imaging data and to handle and manipulate the resulting data products. These have been applied to the challenges involved in analysing the TMA image data taken from the SEARCH study population-based study of breast cancer. We have utilised astronomy 'Virtual Observatory' components specifically adapted from those developed within the AstroGrid Virtual Observatory eScience programme (http://www.astrogrid.org), to facilitate secure data transport, resource discovery through appropriate metadata, data acquisition, ingression to a database system, and secure distributed access to those data and information resources. Image analysis has been applied to the input TMA data utilising a range of algorithms originally developed for diverse astronomical use cases. The resulting data products will in turn be interfaced (though work planned here) to the clinical trials systems developed through the CancerGrid system (see http://www.cancergrid.org). In our initial test case we have automated the scoring of Estrogen Receptor (ER) assessments. ER is an important regulator of mammary growth, but is also a key prognostic and therapeutic target in breast cancer. Assessing ER status at time of diagnosis of breast cancer, determines which treatment programmes should be followed by patients. In particular, those patients who have ER-positive breast cancer will be offered estrogen antagonist therapies such as tamoxifen. At CR-UK, breast cancer studies utilising genomics tools are underway to validate existing and new prognostic and/or predictive markers. TMAs have been created from a large population-based clinical trial (SEARCH; part of the Anglia Breast Cancer study) for analysis with a range of candidate markers. Immunohistochemistry is used to assess the level of nuclear ER expression. The focus of our initial pilot has been on the algorithm development and validation. For the miniPIPSS programme the work will move to increasing the utility of the analysis system by running all processing operations in a pipeline. This automation, the operational basis of our Pathgrid system, has been implemented in prototype by making use of the application-grid infrastructural components from AstroGrid This miniPIPSS project will facilitate the further interchange of ideas and technologies between the physical and medical sciences, and provide methods for the handling and processing of clinical image data in an open and extensible manner The interaction with Oracle represents the initial stage of a longer term partnership, aiming to further develop these analysis techniques such that they are available to he wider medical research community - and further ahead for use in a clinical environment.

    more_vert
  • chevron_left
  • 1
  • 2
  • 3
  • chevron_right

Do the share buttons not appear? Please make sure, any blocking addon is disabled, and then reload the page.

Content report
No reports available
Funder report
No option selected
arrow_drop_down

Do you wish to download a CSV file? Note that this process may take a while.

There was an error in csv downloading. Please try again later.