common-crawl
There are 40 repositories under common-crawl topic.
ashvardanian/StringZilla
Up to 100x faster strings for C, C++, CUDA, Python, Rust, Swift, JS, & Go, leveraging NEON, AVX2, AVX-512, SVE, GPGPU, & SWAR to accelerate search, hashing, sorting, edit distances, sketches, and memory ops 🦖
commoncrawl/cc-pyspark
Process Common Crawl data with Python and Spark
commoncrawl/news-crawl
News crawling with StormCrawler - stores content as WARC
michaelharms/comcrawl
A python utility for downloading Common Crawl data
commoncrawl/cc-crawl-statistics
Statistics of Common Crawl monthly archives mined from URL index files
oscar-project/ungoliant
:spider: The pipeline for the OSCAR corpus
crissyfield/troll-a
Drill into WARC web archives
commoncrawl/cc-webgraph
Tools to construct and process Common Crawl webgraphs
oscar-project/goclassy
An asynchronous concurrent pipeline for classifying Common Crawl based on fastText's pipeline.
commoncrawl/cc-notebooks
Various Jupyter notebooks about Common Crawl data
IBM/cc-dbp
A dataset for knowledge base population research using Common Crawl and DBpedia.
bminixhofer/gerpt2
German small and large versions of GPT2.
cisnlp/GlotCC
🕸 GlotCC Dataset and Pipline -- NeurIPS 2024
oscar-project/oscar-website
The website of the Oscar Project
toimik/CommonCrawl
Common Crawl's processing tools
Mgosi/Big-Data-Analysis-using-MapReduce-in-Hadoop
We explore data by using Big Data Analysis and Visualization skills. To obtain this, we perform 3 main operations. i.e. i)Data Aggregation through different sources. ii) Big Data Analysis using MapReduce and iii) Visualization through Tableau. Data Analysis is very critical in understanding the data, and what we can do with the data. For small datasets it is easier to process and obtain the results. But as for big companies, it becomes crucial for them to obtain the trends of the company for any changes need to be made. Hence we introduce Big Data Analysis to solve this problem. In this lab, we collect close to 20000 tweets, 500 articles on New York Times and 500 articles on Common Crawl Data about Entertainment, which is our main topic of discussion. Using this data, we perform preprocessing and feed it to a MapReduce to find the Word Count and Word Co-Occurrence. Using this, we find the trend of the data collected in this topic. We have used Python to perform Data Analysis.Data Analysis is very critical in understanding the data, and what we can do with the data. For small datasets it is easier to process and obtain the results. But as for big companies, it becomes crucial for them to obtain the trends of the company for any changes need to be made. Hence we introduce Big Data Analysis to solve this problem. In this lab, we collect close to 20000 tweets, 500 articles on New York Times and 500 articles on Common Crawl Data about Entertainment, which is our main topic of discussion. Using this data, we perform preprocessing and feed it to a MapReduce to find the Word Count and Word Co-Occurrence. Using this, we find the trend of the data collected in this topic. We have used Python to perform Data Analysis.
alumik/common-crawl-downloader
Distributed download scripts for Common Crawl data
tokenmill/common-crawl-utils
Various Common Crawl utilities in Clojure.
code402/warc-benchmark
Sample code to grep Common Crawl WARC files in Go, Java, Node and Python.
HRN-Projects/common_crawl_with_scrapy
Parsing Huge Web Archive files from Common Crawl data index to fetch any required domain's data concurrently with Python and Scrapy.
hrbrmstr/cc
⛏Extract metadata of a specific target based on the results of "commoncrawl.org"
ilyankou/cc-gpx
CC-GPX: Extracting High-Quality Annotated Geospatial Data from Common Crawl
connor-marchand/gau-python
This library gets urls from AlienVault's Open Threat Exchange, the Wayback Machine, and Common Crawl. Inspired by Corbin Leo's gau
thunderpoot/cc-getpage
Lightweight Python utility for retrieving individual pages from the Common Crawl archives.
mwoss/mors
Application of topic models for information retrieval and search engine optimization.
neil-zt/common-crawl-client
A Common Crawl client example for scraping specific websites.
socket-var/nyt-twitter-cc-hadoop
Perform big data analysis on New york times, Twitter and Common Crawl APIs
bottomless-archive-project/common-crawl-client
This library is a very lightweight client to Common Crawl's WARC files.
fizerkhan/cdx-index-client
A command-line tool for using CommonCrawl Index API at http://index.commoncrawl.org/
fizerkhan/CommonCrawlDocumentDownload
A small tool which uses the CommonCrawl URL Index to download documents with certain file types or mime-types for mass-testing of frameworks like Apache POI and Apache Tika
fizerkhan/KeywordAnalysis
Word analysis, by domain, on the Common Crawl data set for the purpose of finding industry trends
bottomless-archive-project/url-collector
An application that crawls the Common Crawl corpus for URLs with the specified file extensions.
Dahouabdelhalim/Discourse-marksers-and-Web-crawling
Discourse Markers identification in French Language
hadrianw/abracabra
Eventually a search engine, but currently a filtering pipeline for HTML and soon WARC files.
skyler-myers-db/Common-Crawl-Analysis
Parsing the common crawl database using Scala and Spark
srmocher/fake-science
Analyzing Common Crawl data (specifically) to classify fake/real based on trained deep learning models (LSTM, CNN)