el corpus del espaņol

el corpus del espaņol


OVERVIEW (PDF) (ES)   English Espaņol 

Created by Mark Davies. Funded by the US National Endowment for the Humanities (2001-2002, 2015-2017).

    Corpus Size Created
1 Info Genre / Historical 100 million words 2001
2 Info Web / Dialects * 2 billion words 2016
3 Info NOW (2012 - 2019) 7.3 billion words 2018
4 Info Google Books n-grams 45 billion words 2011

This is our interface for the Google Books n-grams data. It is based on 45 billion words in tens of millions of books from the 1800s-2000s. The n-grams data does not allow the full range of queries that a normal corpus would, but you can still find the frequency of words and phrases over time, as well as finding the collocates of a given word (including the collocates over time, to see semantic change). And because it's based on 45 billion words, it is incredibly rich data.