ai-tech-articles / README.md
siavava's picture
Upload README.md with huggingface_hub
d2139e2
|
raw
history blame
1.87 kB
---
language:
- en
license: mit
task_categories:
- text-generation
- feature-extraction
pretty_name: AI/Technology Articles
tags:
- temporal series data
- language data
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: id
dtype: int64
- name: year
dtype: int64
- name: title
dtype: string
- name: url
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 180820047
num_examples: 17092
download_size: 81702921
dataset_size: 180820047
---
# AI/Tech Dataset
This dataset is a collection of AI/tech articles scraped from the web.
It's hosted on [HuggingFace Datasets](https://huggingface.co/datasets/siavava/ai-tech-articles), so it is easier to load in and work with.
## To load the dataset
### 1. Install [HuggingFace Datasets](https://huggingface.co/docs/datasets/installation.html)
```bash
pip install datasets
```
### 2. Load the dataset
```python
from datasets import load_dataset
dataset = load_dataset("siavava/ai-tech-articles")
# optionally, convert it to a pandas dataframe:
df = dataset["train"].to_pandas()
```
You do not need to clone this repo.
HuggingFace will download the dataset for you, the first time that you load it,
and cache it locally so it does not need to re-download it again
(unless it detects a change upstream).
## File Structure
- [`analytics.ipynb`](analytics.ipynb) - Notebook containing some details about the dataset.
- [`example.ipynb`](example.ipynb) - A minimal notebook that loads in the dataset and converts to Pandas.
- [`raw.csv`](raw.csv) - The raw data, in CSV format.
- `data/*.parquet`- compressed [parquet](https://www.databricks.com/glossary/what-is-parquet) containing the data.
- For raw text files, see the [scraper repo](https://github.com/siavava/scrape.hs) on GitHub.