Datasets:
language: | |
- en | |
license: mit | |
task_categories: | |
- text-generation | |
- feature-extraction | |
pretty_name: AI/Technology Articles | |
tags: | |
- temporal series data | |
- language data | |
configs: | |
- config_name: default | |
data_files: | |
- split: train | |
path: data/train-* | |
dataset_info: | |
features: | |
- name: id | |
dtype: int64 | |
- name: year | |
dtype: int64 | |
- name: title | |
dtype: string | |
- name: url | |
dtype: string | |
- name: text | |
dtype: string | |
splits: | |
- name: train | |
num_bytes: 123475673 | |
num_examples: 2429 | |
download_size: 25153621 | |
dataset_size: 123475673 | |
# AI/Tech Dataset | |
This dataset is a collection of AI/tech articles scraped from the web. | |
It's hosted on [HuggingFace Datasets](https://huggingface.co/datasets/siavava/ai-tech-articles), so it is easier to load in and work with. | |
## To load the dataset | |
### 1. Install [HuggingFace Datasets](https://huggingface.co/docs/datasets/installation.html) | |
```bash | |
pip install datasets | |
``` | |
### 2. Load the dataset | |
```python | |
from datasets import load_dataset | |
dataset = load_dataset("siavava/ai-tech-articles") | |
# optionally, convert it to a pandas dataframe: | |
df = dataset["train"].to_pandas() | |
``` | |
> [!NOTE] | |
> You do not need to clone this repo. | |
> | |
> HuggingFace will download the dataset for you, the first time that you load it, | |
> and cache it so it does not need to re-download it again | |
> (unless it detects a change upstream). | |
## File Structure | |
- [analytics.ipynb](analytics.ipynb) - Notebook containing some details about the dataset and how to load it. | |
- [data/index.parquet](./index.csv) - compressed [parquet](https://www.databricks.com/glossary/what-is-parquet) containing the data. | |
- For raw text files, see the [scraper repo](https://github.com/siavava/scrape.hs) on GitHub. | |