[ExI] Job openings on AI and xrisk at FHI
Anders Sandberg
anders at aleph.se
Wed Dec 9 11:11:51 UTC 2015
Sorry for dragging my job onto the list, but maybe I can get some
list-members into our office :-)
--
The Future of Humanity Institute at the University of Oxford invites
applications for four research positions. We seek outstanding applicants
with backgrounds that could include computer science, mathematics,
economics, technology policy, and/or philosophy.
The Future of Humanity Institute is a leading research centre in the
University of Oxford looking at big-picture questions for human
civilization. We seek to focus our work where we can make the greatest
positive difference. Our researchers regularly collaborate with
governments from around the world and key industry groups working on
artificial intelligence. To read more about the institute’s research
activities, please see http://www.fhi.ox.ac.uk/research/research-areas/.
*1. Research Fellow – AI – Strategic Artificial Intelligence Research
Centre, Future of Humanity Institute* (Vacancy ID# 121242). We are
seeking expertise in the technical aspects of AI safety, including a
solid understanding of present-day academic and industrial research
frontiers, machine learning development, and knowledge of academic and
industry stakeholders and groups. The fellow is expected to have the
knowledge and skills to advance the state of the art in proposed
solutions to the “control problem.” This person should have a technical
background, for example, in computer science, mathematics, or
statistics. Candidates with a very strong machine learning or
mathematics background are encouraged to apply even if they do not have
experience with AI safety topics, assuming they are willing to switch to
this subfield. Applications are due by Noon 6 January 2016. You can
apply for this position through the Oxford recruitment website at
http://bit.ly/1M11RbY.
*2. Research Fellow – AI Policy – Strategic Artificial Intelligence
Research Centre, Future of Humanity Institute* (Vacancy ID# 121241). We
are looking for someone with expertise relevant to assessing the
socio-economic and strategic impacts of future technologies, identifying
key issues and potential risks, and rigorously analysing policy options
for responding to these challenges. This person might have an economics,
political science, social science, or risk analysis background.
Applications are due by Noon 6 January 2016. You can apply for this
position through the Oxford recruitment website at http://bit.ly/1OfWd7Q.
*3. Research Fellow – AI Strategy – Strategic Artificial Intelligence
Research Centre, Future of Humanity Institute* (Vacancy ID# 121168). We
are looking for someone with a multidisciplinary science, technology, or
philosophy background and with outstanding analytical ability. The post
holder will investigate, understand, and analyse the capabilities and
plausibility of theoretically feasible but not yet fully developed
technologies that could impact AI development, and to relate such
analysis to broader strategic and systemic issues. The academic
background of the post-holder is unspecified, but could involve, for
example, computer science or economics. Applications are due by Noon 6
January 2016. You can apply for this position through the Oxford
recruitment website at http://bit.ly/1jM5Pic.
*4. Research Fellow – ERC UnPrEDICT Programme, Future of Humanity
Institute* (Vacancy ID# 121313). This Research Fellowship will work on a
new European Research Council-funded UnPrEDICT (Uncertainty and
Precaution: Ethical Decisions Involving Catastrophic Threats) programme,
hosted by the Future of Humanity Institute at the University of Oxford.
This is a research position for a strong generalist, and will focus on
topics related to existential risk, model uncertainty, the precautionary
principle, and other principles for handling technological progress. In
particular, this research fellow will help to develop decision
procedures for navigating empirical uncertainties related to existential
risk, including information hazards and situations where model or
structural uncertainty are the dominating form of uncertainty. The
research could take a decision-theoretic approach, although this is not
strictly necessary. We also expect the candidate to engage with the
research on specific existential risks, possibly including developing a
framework to evaluate uncertain risks in the context of nuclear weapons,
climate risks, dual use biotechnology, and/or the development of future
artificial intelligence. The successful candidate must demonstrate
evidence of, or the potential for producing, outstanding research in the
areas of relevance to the project, the ability to integrate
interdisciplinary research in philosophy, mathematics and/or economics,
and familiarity with both normative and empirical issues surrounding
existential risk. Applications are due by Noon 6 January 2016. You can
apply for this position through the Oxford recruitment website at
http://bit.ly/1HSCKgP.
Alternatively, please visit http://www.fhi.ox.ac.uk/vacancies/ or
https://www.recruit.ox.ac.uk/ and search using the above vacancy IDs for
more details.
--
--
Anders Sandberg
Future of Humanity Institute
Oxford Martin School
Oxford University
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20151209/9d6a96fe/attachment.html>
More information about the extropy-chat
mailing list