You are here

LifeCLEF 2018

logolifeclef

Schedule

  • 30th Nov 2017: registration opens for all LifeCLEF tasks (until 30.04.2018)
  • Jan 2018 - February 2018: training data release
  • March - April 2018: test data release
  • 1st May 2018: deadline for submission of runs by the participants
  • 15th May 2018: release of processed results by the task organizers
  • 15th June 2018: deadline for submission of working notes papers by the participants
  • 10th-14th Sept 2018: CLEF 2018 Avignon, France

Motivation

Building accurate knowledge of the identity, the geographic distribution and the evolution of living species is essential for a sustainable development of humanity as well as for biodiversity conservation. Unfortunately, such basic information is often only partially available for professional stakeholders, teachers, scientists and citizens, and often incomplete for ecosystems that possess the highest diversity. A noticeable cause and consequence of this sparse knowledge is that identifying living plants or animals is usually impossible for the general public, and often a difficult task for professionals, such as farmers, fish farmers or foresters and even also for the naturalists and specialists themselves. This taxonomic gap was actually identified as one of the main ecological challenges to be solved during Rio’s United Nations Conference in 1992. In this context, an ultimate ambition is to set up innovative information systems relying on the automated identification and understanding of living organisms as a mean to engage massive crowds of observers and boost the production of biodiversity and agro-biodiversity data.

The LifeCLEF 2018 lab proposes three data-oriented challenges related to this vision, in the continuity of the two previous editions of the lab, but with several consistent novelties intended to push the boundaries of the state-of-the-art in several research directions at the frontier of information retrieval, machine learning and knowledge engineering.

CLEF Conference and working notes

CLEF 2017 CLEF Initiative

LifeCLEF lab is part of the Conference and Labs of the Evaluation Forum: CLEF 2018. CLEF 2018 consists of independent peer-reviewed workshops on a broad range of challenges in the fields of multilingual and multimodal information access evaluation, and a set of benchmarking activities carried in various labs designed to test different aspects of mono and cross-language Information retrieval systems. More details about the conference can be found here. Also there is more information about the Clef Initiative.

Submitting a working note with the full description of the methods used in each run is mandatory. Any run that could not be reproduced thanks to its description in the working notes might be removed from the official publication of the results. Working notes are published within CEUR-WS proceedings, resulting in an assignment of an individual DOI (URN) and an indexing by many bibliography systems including DBLP. According to the CEUR-WS policies, a light review of the working notes will be conducted by LifeCLEF organizing committee to ensure quality. As an illustration, LifeCLEF 2017 working notes (task overviews and participant working notes) can be found within CLEF 2017 CEUR-WS proceedings.

Registration and data access

  1. Each participant has to register on (https://www.crowdai.org) with username, email and password. A representative team name should be used
    as username.
  2. In order to be compliant with the CLEF requirements, participants also have to fill in the following additional fields on their profile:
    • First name
    • Last name
    • Affiliation
    • Address
    • City
    • Country
  3. Participants can now access the dataset tab on the challenge's page. A LifeCLEF participant is considered as registered for a task as soon as he/she has downloaded a file of the task's dataset via the dataset tab of the challenge.
    Registrations are handled one a per-task basis. This means if a task has multiple challenges (subtasks), a participant can automatically access the data of all challenges in that task. There is one dataset per task, meaning we do not separate datasets on a per-challenge basis.

Contact

lifeclef-org@inria.fr

Coordination

  • Alexis Joly, INRIA Sophia-Antipolis - ZENITH team, LIRMM, University of Montpellier, France, alexis.joly(replace-by-an-arrobe)inria.fr
  • Henning Müller, University of Applied Sciences Western Switzerland in Sierre, Switzerland, henning.mueller(replace-by-an-arrobe)hevs.ch

GeoLifeCLEF

  • Christophe Botella, Inra – AMAP, Montpellier, France, Christophe Botella
  • Alexis Joly, INRIA Sophia-Antipolis - ZENITH team, LIRMM, University of Montpellier, France, alexis.joly(replace-by-an-arrobe)inria.fr
  • Pierre Bonnet, Cirad – AMAP, Montpellier, France, pierre.bonnet(replace-by-an-arrobe)cirad.fr

BirdCLEF

  • Hervé Glotin, University of Toulon, France, glotin(replace-by-an-arrobe)univ-tln.fr
  • Hervé Goëau, Cirad - AMAP, Montpellie, France.
  • Willem-Pier Vellinga, Xeno-Canto foundation for nature sounds, The Netherlands, wp(replace-by-an-arrobe)xeno-canto.org
  • Alexis Joly, INRIA Sophia-Antipolis - ZENITH team, LIRMM, University of Montpellier, France, alexis.joly(replace-by-an-arrobe)inria.fr

ExpertLifeCLEF

  • Hervé Goëau, Cirad - AMAP, Montpellier, France.
  • Alexis Joly, INRIA Sophia-Antipolis - ZENITH team, LIRMM, University of Montpellier, France, alexis.joly(replace-by-an-arrobe)inria.fr
  • Pierre Bonnet, Cirad – AMAP, Montpellier, France, pierre.bonnet(replace-by-an-arrobe)cirad.fr
  • Concetto Spampinato, University of Catania, Italy, cspampin(replace-by-an-arrobe)dieei.unict.it

Credits

  • Ivan Eggel, University of Applied Sciences Western Switzerland, Sierre, Switzerland, ivan.eggel(replace-by-an-arrobe)hevs.ch

Floris'tic      Inria          CNRS      Xeno-Canto           Sabiod      Cetamada